Research2026-05-12
Echo-LoRA: Parameter-Efficient Fine-Tuning via Cross-Layer Representation Injection
Source: Arxiv CS.AI
arXiv:2605.08177v1 Announce Type: cross Abstract: Parameter-efficient fine-tuning (PEFT) has become a practical route for adapting large language models to downstream tasks, with LoRA-style methods being particularly attractive because they are inexpensive to train and easy to deploy. Most LoRA...
arxivpapersfine-tuning