This research introduces an active exploration algorithm to enhance the efficiency of preference alignment in large language models (LLMs) by strategically selecting human feedback. The authors frame this as an active contextual dueling bandit problem, where the system actively chooses which "contexts" (prompts) and "actions" (LLM responses) to present to human evaluators. Their proposed method, AE-Borda, leverages uncertainty estimation and a generalized Borda function to identify the most informative data points for training, leading to faster learning and reduced data collection costs. The paper validates its theoretical guarantees with synthetic experiments and demonstrates practical improvements on LLM performance across various datasets, including two new contributions: Jeopardy! for factual correctness and Haikus for creative writing.
信息
- 节目
- 频率一周一更
- 发布时间2025年9月6日 UTC 04:32
- 长度15 分钟
- 分级儿童适宜