Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

The answer is that it does know. Not exactly, but the "general shape" of the answer is known to the LLM before the very first token of the answer is emitted!

"Next token prediction" is often overstated - "pick the next token" is the exposed tip of a very large computational process.

And LLMs are very sharp at squeezing the context for every single bit of information available in it. Much less so at using it in the ways you want them to.

There's enough information at "no token emitted yet" for an LLM to start steering the output towards "here's the answer" or "I don't know the answer" or "I need to look up more information to give the answer" immediately. And if it fails to steer it right away? An LLM optimized for hallucination avoidance could still go "fuck consistency drive" and take a sharp pivot towards "no, I'm wrong" mid-sentence if it had to. For example, if you took control and forced a wrong answer by tampering with the tokens directly, then handed the control back to the LLM.



By "shape" of the answer, what do you mean? I always visualized token prediction as a vector pointing off into some sort of cloud of related tokens, and if that's a fair way to visualize it, I could understand how you could say, before even emitting the first token of the answer, "we are pointing towards generally the correct place where the answer is found". But when a single token can make or break an answer, I still don't see how you can truly know whether the answer is correct until the very last token is reached. Because of this, I'm still not convinced hallucination can be stopped.

Can you help correct where I'm going wrong?




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: