About 124,000 results
Open links in new tab
  1. [2111.06377] Masked Autoencoders Are Scalable Vision Learners

    Nov 11, 2021 · Abstract: This paper shows that masked autoencoders (MAE) are scalable self-supervised learners for computer vision. Our MAE approach is simple: we mask random patches of the input image and reconstruct the missing pixels. It is based on two core designs.

  2. Masked Autoencoders in Deep Learning - GeeksforGeeks

    Jul 8, 2024 · Masked autoencoders are neural network models designed to reconstruct input data from partially masked or corrupted versions, helping the model learn robust feature representations. They are significant in deep learning for tasks such as data denoising, anomaly detection, and improving model generalization by training on incomplete data.

  3. PyTorch implementation of MAE https//arxiv.org/abs/2111.06377

    title = {Masked Autoencoders Are Scalable Vision Learners}, year = {2021}, The original implementation was in TensorFlow+TPU. This re-implementation is in PyTorch+GPU. This repo is a modification on the DeiT repo. Installation and preparation follow that repo.

  4. Masked Autoencoders Are Effective Tokenizers for Diffusion Models

    Feb 5, 2025 · Motivated by these insights, we propose MAETok, an autoencoder (AE) leveraging mask modeling to learn semantically rich latent space while maintaining reconstruction fidelity.

  5. Masked image modeling with Autoencoders - Keras

    Dec 20, 2021 · Inspired from the pretraining algorithm of BERT (Devlin et al.), they mask patches of an image and, through an autoencoder predict the masked patches. In the spirit of "masked language modeling", this pretraining task could be referred to as "masked image modeling".

  6. [2505.09160] A Multi-Task Foundation Model for Wireless Channel ...

    3 days ago · Current applications of self-supervised learning to wireless channel representation often borrow paradigms developed for text and image processing, without fully addressing the unique characteristics and constraints of wireless communications. Aiming to fill this gap, we first propose WiMAE (Wireless Masked Autoencoder), a transformer-based encoder-decoder foundation model pretrained on a ...

  7. How to Implement State-of-the-Art Masked AutoEncoders (MAE)

    Sep 16, 2024 · Today, I’m excited to delve into one of the most significant breakthroughs in Computer Vision post-Vision Transformers: Masked Autoencoders (MAE). This article serves as the practical implementation companion to my previous post: The Ultimate Guide to Masked Autoencoders (MAE)

  8. Attention-Guided Masked Autoencoders for Learning Image …

    TL;DR: We guide the reconstruction learning of a masked autoencoder with attention maps to learn image represenations with an improved high-level semantic understanding.

  9. Masked Autoencoders: The Hidden Puzzle Pieces of Modern AI

    Nov 21, 2024 · At its heart, a Masked Autoencoder is a self-supervised learning model designed to understand data by reconstructing its masked components. Initially inspired by the success of masked language models like BERT, the concept has since been adapted and optimized for computer vision and other domains.

  10. Spatial-Spectral Hierarchical Multiscale Transformer-Based Masked ...

    3 days ago · Spatial-Spectral Hierarchical Multiscale Transformer-Based Masked Autoencoder for Hyperspectral Image Classification ... Transformer, with its powerful long-range relationship modeling ability, has become a popular model; however, it usually requires a large number of labeled data for parameter training, which may be costly and impractical for ...

Refresh