arxiv
PublishedApril 24, 2026 at 4:00 AM
—neutral
Why Supervised Fine-Tuning Fails to Learn: A Systematic Study of Incomplete Learning in Large Language Models
Publisher summary· verbatim
arXiv:2604.10079v3 Announce Type: replace Abstract: Supervised Fine-Tuning (SFT) is the standard approach for adapting large language models (LLMs) to downstream tasks. However, we observe a persistent failure mode: even after convergence, models often fail to correctly reproduce a subset of their o
Discussion
No replies yet. Be first.
Originally published on arxiv ↗