Models Recall What They Violate: Constraint Adherence in Multi-Turn LLM Ideation

·ArXiv cs.AI··

arXiv:2604.28031v1 Announce Type: cross Abstract: When researchers iteratively refine ideas with large language models, do the models preserve fidelity to the original objective? We introduce DriftBench, a benchmark for evaluating constraint adherence in multi-turn LLM-assisted scientific ideation. Across 2,146 scored benchmark runs spanning seven models from five providers (including two open-weight), four interaction conditions, and 38 research briefs from 24 scientific domains, we find that i...

Read full article →

Related Articles

Welfare Biology and AI: The Psychopath, the Nematode, and the Arahant
Dawn Drescher · EA Forum · 4d ago
Immigration changes are driving foreign researchers to leave the U.S. — or not come to begin with 
Andrew Joseph · STAT News · 4d ago
Looking for papers on general formalizations of "agency"
lovagrus · LessWrong · 5d ago
SFF’s HSEE grant round; human intelligence amplification projects I’d like to see by TsviBT
TsviBT · Nuno Sempere · 8d ago
The other paper that killed deep learning theory
LawrenceC · Alignment Forum · 11d ago