MAE (Masked Autoencoders) is a self-supervised learning framework for visual representation learning using masked image modeling. It trains a Vision Transformer (ViT) by randomly masking a high percentage of image patches (typically 75%) and reconstructing the missing content from the remaining visible patches. This forces the model to learn semantic structure and global context without supervision. The encoder processes only the visible patches, while a lightweight decoder reconstructs the full image—making pretraining computationally efficient. After pretraining, the encoder serves as a powerful backbone for downstream tasks like image classification, segmentation, and detection, achieving top performance with minimal fine-tuning. The repository provides pretrained models, fine-tuning scripts, evaluation protocols, and visualization tools for reconstruction quality and learned features.

Features

  • Masked image modeling with random high-ratio patch masking
  • Efficient pretraining via encoder-decoder separation (encoder sees only visible patches)
  • Scalable Vision Transformer backbone for downstream vision tasks
  • Pretrained models and fine-tuning scripts for classification, detection, and segmentation
  • Visualization tools for reconstruction and representation analysis
  • Self-supervised training paradigm requiring no labeled data

Project Samples

Project Activity

See All Activity >

Categories

AI Models

License

MIT License

Follow MAE (Masked Autoencoders)

MAE (Masked Autoencoders) Web Site

You Might Also Like
MongoDB Atlas runs apps anywhere Icon
MongoDB Atlas runs apps anywhere

Deploy in 115+ regions with the modern database for every enterprise.

MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
Start Free
Rate This Project
Login To Rate This Project

User Reviews

Be the first to post a review of MAE (Masked Autoencoders)!

Additional Project Details

Programming Language

Python

Related Categories

Python AI Models

Registered

2025-10-06