Skip to content

Jian-Lang/awesome-modality-missing-learning

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Awesome Modality Missing Learning

This repository contains a collection of research studies centered on Modality Missing Learning (MML) (also referred to as Incomplete Multimodal Learning).

alt text

📢 News

  • 2026.04.14 🔨 We release the awesome MML, which collects the methods mainly published in conferences for MML.

📖 Contents

📃 Paper List

📚 Survey

Title Venue Year Paper Code
Deep Multimodal Learning with Missing Modality: A Survey TMLR 2026 Paper N/A
Multimodal Learning Under Imperfect Data Conditions: A Survey arxiv 2026 Paper N/A
Multimodal fusion on low-quality data: A comprehensive survey arxiv 2024 Paper N/A

🧩 Reconstruction-based Methods

Title Venue Year Paper Code
RAG4DMC: Retrieval-Augmented Generation for Data-Level Modality Completion ICLR 2026 Paper N/A
Inference-Time Dynamic Modality Selection for Incomplete Multimodal Classification ICLR 2026 Paper Stars
Sample-specific Modality Diagnosis and Cross-modal Enhancement for Incomplete Multimodal Representations AAAI 2026 Paper Stars
TMDC: A Two-Stage Modality Denoising and Complementation Framework for Multimodal Sentiment Analysis with Missing and Noisy Modalities AAAI 2026 Paper Stars
Recovering Coherent Affective Patterns: Addressing Modality Missing in Multimodal Sentiment Analysis AAAI 2026 Paper Stars
Tackling Dual-stage Missing Modalities in Brain Tumor Segmentation via Robust Modality Reconstruction and Prompt-guided Modality Adaptation AAAI 2026 Paper N/A
MCMoE: Completing Missing Modalities with Mixture of Experts for Incomplete Multimodal Action Quality Assessment AAAI 2026 Paper Stars
OMG-Agent: Toward Robust Missing Modality Generation with Decoupled Coarse-to-Fine Agentic Workflows AAAI 2026 Paper N/A
Unbiased Missing-modality Multimodal Learning ICCV 2025 Paper N/A
Knowledge Bridger: Towards Training-Free Missing Modality Completion CVPR 2025 Paper Stars
CyIN: Cyclic Informative Latent Space for Bridging Complete and Incomplete Multimodal Learning NeurIPS 2025 Paper Stars
IMOL: Incomplete-Modality-Tolerant Learning for Multi-Domain Fake News Video Detection ACL 2025 Paper Stars
Generating with Fairness: A Modality-Diffused Counterfactual Framework for Incomplete Multimodal Recommendations WWW 2025 Paper Stars
FedMobile: Enabling Knowledge Contribution-aware Multi-modal Federated Learning with Incomplete Modalities WWW 2025 Paper N/A
Learning Trimodal Relation for Audio-Visual Question Answering with Missing Modality ECCV 2024 Paper Stars
A Flexible Generative Model for Heterogeneous Tabular EHR with Missing Modality ICLR 2024 Paper N/A
LDS2AE: Local Diffusion Shared-Specific Autoencoder for Multimodal Remote Sensing Image Classification with Arbitrary Missing Modalities AAAI 2024 Paper Stars
Toward Robust Incomplete Multimodal Sentiment Analysis via Hierarchical Representation Learning NeurIPS 2024 Paper N/A
Towards Robust Multimodal Sentiment Analysis with Incomplete Data NeurIPS 2024 Paper Stars
Distribution-Consistent Modal Recovering for Incomplete Multimodal Learning ICCV 2023 Paper Stars
Contrastive Intra- and Inter-Modality Generation for Enhancing Incomplete Multimedia Recommendation MM 2023 Paper N/A
Towards Good Practices for Missing Modality Robust Action Recognition AAAI 2023 Paper Stars
Incomplete Multimodality-Diffused Emotion Recognition NeurIPS 2023 Paper Stars
Gcnet: Graph completion network for incomplete multimodal learning in conversation TPAMI 2023 Paper Stars
Client-Adaptive Cross-Model Reconstruction Network for Modality-Incomplete Multimodal Federated Learning MM 2023 Paper N/A
M3Care: Learning with Missing Modalities in Multimodal Healthcare Data KDD 2022 Paper Stars
Missing Modality Imagination Network for Emotion Recognition with Uncertain Missing Modalities ACL 2021 Paper Stars
SMIL: Multimodal Learning with Severely Missing Modality AAAI 2021 Paper Stars
Deep Adversarial Learning for Multi-Modality Missing Data Completion KDD 2018 Paper Stars
Semi-supervised Deep Generative Modelling of Incomplete Multi-Modality Emotional Data MM 2018 Paper N/A
Missing Modalities Imputation via Cascaded Residual Autoencoder CVPR 2017 Paper Stars

🔗 Multimodal Shared Representation Learning Methods

Title Venue Year Paper Code
SiMO: Single-Modality-Operable Multimodal Collaborative Perception ICLR 2026 Paper Stars
TMDC: A Two-Stage Modality Denoising and Complementation Framework for Multimodal Sentiment Analysis with Missing and Noisy Modalities AAAI 2026 Paper Stars
Proxy-Driven Robust Multimodal Sentiment Analysis with Incomplete Data ACL 2025 Paper N/A
T2DR: A Two-Tier Deficiency-Resistant Framework for Incomplete Multimodal Learning ACL Finding 2025 Paper Stars
DrFuse: Learning Disentangled Representation for Clinical Multi-Modal Fusion with Missing Modality and Modal Inconsistency AAAI 2024 Paper Stars
Multi-Modal Learning with Missing Modality via Shared-Specific Feature Modelling CVPR 2023 Paper Stars
Rethinking Missing Modality Learning: From a Decoding View MM 2023 Paper N/A
Found in Translation: Learning Robust Joint Representations by Cyclic Translations between Modalities AAAI 2019 Paper Stars

🎓 Teacher-Student Distillation and Alignment Methods

Title Venue Year Paper Code
Towards Unified Vision-Language Models with Incomplete Multi-Modal Inputs AAAI 2026 Paper N/A
CMAD: Correlation-Aware and Modalities-Aware Distillation for Multimodal Sentiment Analysis with Missing Modalities ICCV 2025 Paper Stars
OGP-Net: Optical Guidance Meets Pixel-Level Contrastive Distillation for Robust Multi-Modal and Missing Modality Segmentation AAAI 2025 Paper N/A
Multimodal Patient Representation Learning with Missing Modalities and Labels ICLR 2024 Paper Stars
Correlation-Decoupled Knowledge Distillation for Multimodal Sentiment Analysis with Incomplete Modalities CVPR 2024 Paper N/A
MaskMentor: Unlocking the Potential of Masked Self-Teaching for Missing Modality RGB-D Semantic Segmentation MM 2024 Paper N/A
A Unified Self-Distillation Framework for Multimodal Sentiment Analysis with Uncertain Missing Modalities AAAI 2024 Paper N/A
Probabilistic Conformal Distillation for Enhancing Missing Modality Robustness NeurIPS 2024 Paper Stars
MMANet: Margin-aware Distillation and Modality-aware Regularization for Incomplete Multimodal Learning CVPR 2023 Paper Stars
MissModal: Increasing Robustness to Missing Modality in Multimodal Sentiment Analysis TACL 2023 Paper Stars
Multimodal Learning with Incomplete Modalities by Knowledge Distillation KDD 2020 Paper Stars

🔎 Retrieval-Augmented Methods

Title Venue Year Paper Code
RAG4DMC: Retrieval-Augmented Generation for Data-Level Modality Completion ICLR 2026 Paper N/A
Retrieval-Augmented Dynamic Prompt Tuning for Incomplete Multimodal Learning AAAI 2025 Paper Stars
REDEEMing Modality Information Loss: Retrieval-Guided Conditional Generation for Severely Modality Missing Learning KDD 2025 Paper Stars
MissRAG: Addressing the Missing Modality Challenge in Multimodal Large Language Models ICCV 2025 Paper Stars
IMOL: Incomplete-Modality-Tolerant Learning for Multi-Domain Fake News Video Detection ACL 2025 Paper Stars

📊 Mixture-of-Experts Methods

Title Venue Year Paper Code
Taming Cascaded Mixture-of-Experts for Modality-missing Multi-modal Salient Object Detection AAAI 2026 Paper Stars
MCMoE: Completing Missing Modalities with Mixture of Experts for Incomplete Multimodal Action Quality Assessment AAAI 2026 Paper Stars
SimMLM: A Simple Framework for Multi-modal Learning with Missing Modality ICCV 2025 Paper Stars
Multimodal Emotion Recognition with Missing Modality via a Unified Multi-task Pre-training Framework MM 2025 Paper Stars
FuseMoE: Mixture-of-Experts Transformers for Fleximodal Fusion NeurIPS 2024 Paper Stars
Flex-MoE: Modeling Arbitrary Modality Combination via the Flexible Mixture-of-Experts NeurIPS 2024 Paper Stars
Leveraging Knowledge of Modality Experts for Incomplete Multimodal Learning MM 2024 Paper Stars

⚖️ Modality Enhancement Learning Methods

Title Venue Year Paper Code
BALM: A Model-Agnostic Framework for Balanced Multimodal Learning under Imbalanced Missing Rates CVPR 2026 Paper Stars
Plug, Play, and Fortify: A Low-Cost Module for Robust Multimodal Image Understanding Models ICLR 2026 Paper Stars
Cross-modal Prompting for Balanced Incomplete Multi-modal Emotion Recognition AAAI 2026 Paper Stars
Hyper-Modality Enhancement for Multimodal Sentiment Analysis with Missing Modalities NeurIPS 2025 Paper Stars
RedCore: Relative Advantage Aware Cross-Modal Representation Learning for Missing Modalities with Imbalanced Missing Rates AAAI 2024 Paper Stars
PASSION: Towards Effective Incomplete Multi-Modal Medical Image Segmentation with Imbalanced Missing Rates MM 2024 Paper Stars
Gradient-Guided Modality Decoupling for Missing-Modality Robustness AAAI 2024 Paper Stars

🔄 Continual Missing Modality Learning Methods

Title Venue Year Paper Code
DeLo: Dual Decomposed Low-Rank Experts Collaborationfor Continual Missing Modality Learning AAAI 2026 Paper Stars
Efficient Prompting for Continual Adaptation to Missing Modalities NAACL 2025 Paper N/A
Reconstruct before Query: Continual Missing Modality Learning with Decomposed Prompt Collaboration arxiv 2024 Paper Stars

⚙️ Multimodal Transformer-oriented Methods

Title Venue Year Paper Code
MoRA: Missing Modality Low-Rank Adaptation for Visual Recognition ICLR 2026 Paper Stars
Retrieval-Augmented Dynamic Prompt Tuning for Incomplete Multimodal Learning AAAI 2025 Paper Stars
REDEEMing Modality Information Loss: Retrieval-Guided Conditional Generation for Severely Modality Missing Learning KDD 2025 Paper Stars
Synergistic Prompting for Robust Visual Recognition with Missing Modalities ICCV 2025 Paper N/A
Enhancing Multimodal Model Robustness Under Missing Modalities via Memory-Driven Prompt Learning IJCAI 2025 Paper Stars
Deep Correlated Prompting for Visual Recognition with Missing Modalities NeurIPS 2024 Paper Stars
Multimodal Prompt Learning with Missing Modalities for Sentiment Analysis and Emotion Recognition ACL 2024 Paper Stars
Missing Modality Prediction for Unpaired Multimodal Learning via Joint Embedding of Unimodal Models ECCV 2024 Paper N/A
Multimodal Prompting with Missing Modalities for Visual Recognition CVPR 2023 Paper Stars
Are Multimodal Transformers Robust to Missing Modality? CVPR 2022 Paper N/A

🤖 Large Multimodal Model-oriented Methods

Title Venue Year Paper Code
MissRAG: Addressing the Missing Modality Challenge in Multimodal Large Language Models ICCV 2025 Paper Stars

🎯 Downstream Applications

🏥 Medical Applications

Title Venue Year Paper Code
Tackling Dual-stage Missing Modalities in Brain Tumor Segmentation via Robust Modality Reconstruction and Prompt-guided Modality Adaptation AAAI 2026 Paper N/A
MUST: Modality-Specific Representation-Aware Transformer for Diffusion-Enhanced Survival Prediction with Missing Modality CVPR 2026 Paper Stars
Semantic-guided Masked Mutual Learning for Multi-modal Brain Tumor Segmentation with Arbitrary Missing Modalities AAAI 2025 Paper N/A
Distilled Prompt Learning for Incomplete Multimodal Survival Prediction CVPR 2025 Paper Stars
Incomplete Multi-modal Brain Tumor Segmentation via Learnable Sorting State Space Model CVPR 2025 Paper N/A
KMD: Koopman Multi-modality Decomposition for Generalized Brain Tumor Segmentation under Incomplete Modalities CVPR 2025 Paper N/A
DrFuse: Learning Disentangled Representation for Clinical Multi-Modal Fusion with Missing Modality and Modal Inconsistency AAAI 2024 Paper Stars
A Flexible Generative Model for Heterogeneous Tabular EHR with Missing Modality ICLR 2024 Paper N/A
Multimodal Patient Representation Learning with Missing Modalities and Labels ICLR 2024 Paper Stars
PASSION: Towards Effective Incomplete Multi-Modal Medical Image Segmentation with Imbalanced Missing Rates MM 2024 Paper Stars
Modal-aware Visual Prompting for Incomplete Multi-modal Brain Tumor Segmentation MM 2023 Paper N/A
M3Care: Learning with Missing Modalities in Multimodal Healthcare Data KDD 2022 Paper Stars

😊 Multimodal Sentiment Analysis

Title Venue Year Paper Code
Cross-modal Prompting for Balanced Incomplete Multi-modal Emotion Recognition AAAI 2026 Paper Stars
Recovering Coherent Affective Patterns: Addressing Modality Missing in Multimodal Sentiment Analysis AAAI 2026 Paper Stars
TMDC: A Two-Stage Modality Denoising and Complementation Framework for Multimodal Sentiment Analysis with Missing and Noisy Modalities AAAI 2026 Paper Stars
Proxy-Driven Robust Multimodal Sentiment Analysis with Incomplete Data ACL 2025 Paper N/A
CMAD: Correlation-Aware and Modalities-Aware Distillation for Multimodal Sentiment Analysis with Missing Modalities ICCV 2025 Paper Stars
Multimodal Emotion Recognition with Missing Modality via a Unified Multi-task Pre-training Framework MM 2025 Paper Stars
Hyper-Modality Enhancement for Multimodal Sentiment Analysis with Missing Modalities NeurIPS 2025 Paper Stars
A Unified Self-Distillation Framework for Multimodal Sentiment Analysis with Uncertain Missing Modalities AAAI 2024 Paper N/A
Multimodal Prompt Learning with Missing Modalities for Sentiment Analysis and Emotion Recognition ACL 2024 Paper Stars
Correlation-Decoupled Knowledge Distillation for Multimodal Sentiment Analysis with Incomplete Modalities CVPR 2024 Paper N/A
Toward Robust Incomplete Multimodal Sentiment Analysis via Hierarchical Representation Learning NeurIPS 2024 Paper N/A
Towards Robust Multimodal Sentiment Analysis with Incomplete Data NeurIPS 2024 Paper Stars
Incomplete Multimodality-Diffused Emotion Recognition NeurIPS 2023 Paper Stars
MissModal: Increasing Robustness to Missing Modality in Multimodal Sentiment Analysis TACL 2023 Paper Stars
Missing Modality Imagination Network for Emotion Recognition with Uncertain Missing Modalities ACL 2021 Paper Stars
Semi-supervised Deep Generative Modelling of Incomplete Multi-Modality Emotional Data MM 2018 Paper N/A

🧪 Other Applications

Title Venue Year Paper Code
MCMoE: Completing Missing Modalities with Mixture of Experts for Incomplete Multimodal Action Quality Assessment AAAI 2026 Paper Stars
Taming Cascaded Mixture-of-Experts for Modality-missing Multi-modal Salient Object Detection AAAI 2026 Paper Stars
Towards Unified Vision-Language Models with Incomplete Multi-Modal Inputs AAAI 2026 Paper N/A
Inference-Time Dynamic Modality Selection for Incomplete Multimodal Classification ICLR 2026 Paper Stars
MoRA: Missing Modality Low-Rank Adaptation for Visual Recognition ICLR 2026 Paper Stars
Plug, Play, and Fortify: A Low-Cost Module for Robust Multimodal Image Understanding Models ICLR 2026 Paper Stars
SiMO: Single-Modality-Operable Multimodal Collaborative Perception ICLR 2026 Paper Stars
OGP-Net: Optical Guidance Meets Pixel-Level Contrastive Distillation for Robust Multi-Modal and Missing Modality Segmentation AAAI 2025 Paper N/A
IMOL: Incomplete-Modality-Tolerant Learning for Multi-Domain Fake News Video Detection ACL 2025 Paper Stars
Synergistic Prompting for Robust Visual Recognition with Missing Modalities ICCV 2025 Paper N/A
Enhancing Multimodal Model Robustness Under Missing Modalities via Memory-Driven Prompt Learning IJCAI 2025 Paper Stars
I3-MRec: Invariant Learning with Information Bottleneck for Incomplete Modality Recommendation MM 2025 Paper Stars
Generating with Fairness: A Modality-Diffused Counterfactual Framework for Incomplete Multimodal Recommendations WWW 2025 Paper Stars
LDS2AE: Local Diffusion Shared-Specific Autoencoder for Multimodal Remote Sensing Image Classification with Arbitrary Missing Modalities AAAI 2024 Paper Stars
Learning Trimodal Relation for Audio-Visual Question Answering with Missing Modality ECCV 2024 Paper Stars
MaskMentor: Unlocking the Potential of Masked Self-Teaching for Missing Modality RGB-D Semantic Segmentation MM 2024 Paper N/A
Towards Good Practices for Missing Modality Robust Action Recognition AAAI 2023 Paper Stars
Multimodal Prompting with Missing Modalities for Visual Recognition CVPR 2023 Paper Stars
Contrastive Intra- and Inter-Modality Generation for Enhancing Incomplete Multimedia Recommendation MM 2023 Paper N/A
Gcnet: Graph completion network for incomplete multimodal learning in conversation TPAMI 2023 Paper Stars

📦 Others

Title Venue Year Paper Code
ICYM2I: The illusion of multimodal informativeness under missingness ICLR 2026 Paper Stars

🏆 Benchmarks

🏥 Medical

Benchmark Modality Description Paper
BraTS2018 Flair + T1 + T1c + T2 MRI A standard benchmark for brain tumor subregion segmentation under incomplete MRI modalities. It is widely used to evaluate robustness when one or more MR sequences are missing at training or test time. Paper
BraTS2020 Flair + T1 + T1ce + T2 MRI A later BraTS benchmark with the same core multimodal MRI setting but a newer challenge split. It is often used to test whether missing-modality segmentation methods generalize beyond BraTS2018. Paper
MyoPS2020 bSSFP cine CMR + LGE CMR + T2 CMR A cardiac image segmentation benchmark built from complementary CMR sequences. It is useful for studying incomplete-modality learning because different sequences capture function, edema, and scar information. Paper
OpenI Chest X-ray image + clinical report text A radiology image-text benchmark pairing chest X-rays with reports. It is suitable for clinical multimodal classification or retrieval when either the image or the report is partially unavailable. Paper
MIMIC-IV Clinical EHR tables + vitals + labs + diagnoses + treatments A large ICU electronic health record benchmark with heterogeneous tabular and temporal clinical data. It is commonly used for outcome prediction and patient modeling under partially observed modalities. Paper
MIMIC-CXR-JPG Chest X-ray images + structured labels / reports A large-scale chest radiography benchmark aligned with reports and labels. It is a core medical vision-language resource for incomplete image-report fusion settings. Paper
eICU-CRD ICU EHR time series + diagnoses + treatments A multi-center critical care benchmark covering diverse ICU stays and structured clinical signals. It is often used to evaluate robustness when visits, measurements, or clinical views are missing. Paper
ADNI MRI / PET + clinical + genetics + biomarkers A canonical multimodal Alzheimer's benchmark combining imaging, cognitive, genomic, and biomarker information. It is widely used for diagnosis and prognosis under partial modality availability. Paper
ODIR Fundus images + age + diagnostic keywords An ophthalmology benchmark combining retinal fundus images with metadata and diagnostic tags. It supports multimodal ocular disease prediction when image or clinical attributes are incomplete. Paper
TCGA Pan-Cancer Pathology / genomics / clinical A multimodal oncology benchmark spanning pathology, molecular profiles, and clinical records across many cancer types. Different papers usually construct task-specific subsets for survival prediction or risk modeling under missing modalities. Paper

😊 Sentiment Analysis

Benchmark Modality Description Paper
CMU-MOSI Audio + video + text A classic multimodal sentiment benchmark built from opinion videos. It is widely used to study sentiment prediction when acoustic, visual, or textual cues are missing. Paper
CMU-MOSEI Audio + video + text A larger and more diverse successor to CMU-MOSI covering many speakers and topics. It is a standard benchmark for robust multimodal sentiment learning under missing modalities. Paper
IEMOCAP Audio + video + motion capture + text A multimodal emotion recognition benchmark recorded from dyadic acted conversations. It is commonly used to test robustness when one or more expressive modalities are absent. Paper
CH-SIMS Audio + video + text A Chinese multimodal sentiment benchmark with fine-grained modality annotations. It is useful for evaluating missing-modality methods beyond English-centric datasets. Paper

📦 Others

Benchmark Modality Description Paper
MM-IMDb Image + text A movie genre classification benchmark pairing posters with plot summaries. It is one of the most common image-text testbeds for incomplete multimodal classification. Paper
HateMemes Image + text A meme understanding benchmark where hateful intent often depends on combining image and text. It is challenging because either modality alone can be insufficient for correct prediction. Paper
Food101 Image + text A food understanding benchmark used with paired visual and textual recipe information. It is useful for studying incomplete image-text recognition and retrieval. Paper
Audiovision-MNIST (avMNIST) Image + audio A bimodal digit classification benchmark composed of MNIST images and spoken-digit audio represented by MFCCs. It is a simple but widely used testbed for missing-modality classification with independent visual and acoustic inputs. Paper
PolyMNIST Five image modalities A synthetic multimodal benchmark where the same digit is rendered in five modality-specific visual styles with different backgrounds. It is widely used to test scalability and arbitrary missing-modality combinations beyond two-modal settings. Paper

📮 Contact Us

If you find any missing work related to IML, please report it by creating an Issue in the repository to contribute the community together.

If you have other questions, please contact jian_lang@std.uestc.edu.cn.

About

This repository contains a collection of research studies centered on Modality Missing Learning (MML) (also referred to as Incomplete Multimodal Learning).

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors