Research2026-05-06
G-Loss: Graph-Guided Fine-Tuning of Language Models
Source: Arxiv CS.AI
arXiv:2604.25853v2 Announce Type: replace-cross Abstract: Traditional loss functions, including cross-entropy, contrastive, triplet, and su pervised contrastive losses, used for fine-tuning pre-trained language models such as BERT, operate only within local neighborhoods and fail to account for the...
arxivpapersfine-tuning