大型语言模型在多轮对话中容易迷失方向。
LLMs get lost in multi-turn conversation

原始链接: https://arxiv.org/abs/2505.06120

arXivLabs是一个框架,允许合作者直接在我们的网站上开发和分享新的arXiv功能。与arXivLabs合作的个人和组织都已接受并认同我们开放、社区、卓越和用户数据隐私的价值观。arXiv 致力于这些价值观,并且只与坚持这些价值观的合作伙伴合作。有提升arXiv社区价值的项目想法吗?了解更多关于arXivLabs的信息。

A Hacker News discussion highlights the challenges LLMs face in maintaining context during multi-turn conversations, confirming observations that long interactions can "poison" results. Users share experiences where LLMs struggle to recover from initial errors, requiring fresh starts. While some find LLMs helpful for compressing information and debugging complex issues like IPSEC configurations or PPP drivers, others note the models' tendency to mix versions, hallucinate details, and invent explanations. Many agree that LLMs lack introspection and often fail to ask for clarification when uncertain, unlike humans. Solutions discussed include prompt engineering to keep context clean, manually editing conversation history, and forking conversations to explore different directions. Users suggest that LLMs often provide initial responses without adequate information, sticking with incorrect answers even when subsequent information clarifies. Managing context effectively is thus crucial for reliable results.

原文

arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website.

Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them.

Have an idea for a project that will add value for arXiv's community? Learn more about arXivLabs.

联系我们 contact @ memedata.com