Research2026-05-11
Modality Gap-Driven Subspace Alignment Training Paradigm For Multimodal Large Language Models
Source: Arxiv CS.AI
arXiv:2602.07026v2 Announce Type: replace-cross Abstract: Despite the success of multimodal contrastive learning in aligning visual and linguistic representations, a persistent geometric anomaly, the Modality Gap, remains: embeddings of distinct modalities expressing identical semantics occupy...
arxivpapersmultimodal