原文整理页

Andrej Karpathy 建议将大模型视为“模拟器”而非“实体”,通过模拟多方观点而非询问其“个人看法”来获得更好的结果

来源作者:Andrej Karpathy (@karpathy)原始来源:https://x.com/karpathy/status/1997731268969304070

中文导读

Andrej Karpathy 建议将大模型视为“模拟器”而非“实体”,通过模拟多方观点而非询问其“个人看法”来获得更好的结果。

正文 Markdown

Don't think of LLMs as entities but as simulators. For example, when exploring a topic, don't ask: "What do you think about xyz"? There is no "you". Next time try: "What would be a good group of people to explore xyz? What would they say?" The LLM can channel/simulate many perspectives but it hasn't "thought about" xyz for a while and over time and formed its own opinions in the way we're used to. If you force it via the use of "you", it will give you something by adopting a personality embedding vector implied by the statistics of its finetuning data and then simulate that. It's fine to do, but there is a lot less mystique to it than I find people naively attribute to "asking an AI".