Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

It's because you need to mess with embeddings or even train new heads on top of a network to use it. LLMs just use tokens-in tokens-out, they don't classify with softmax over classes, they softmax over vocabulary tokens. LLMs are more convenient


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: