arxiv
PublishedApril 24, 2026 at 4:00 AM
—neutral
MGDA-Decoupled: Geometry-Aware Multi-Objective Optimisation for DPO-based LLM Alignment
Publisher summary· verbatim
arXiv:2604.20685v1 Announce Type: new Abstract: Aligning large language models (LLMs) to desirable human values requires balancing multiple, potentially conflicting objectives such as helpfulness, truthfulness, and harmlessness, which presents a multi-objective optimisation challenge. Most alignment
Discussion
No replies yet. Be first.
Originally published on arxiv ↗