Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

With some of the most absurd hallucinations in mind and LLM's opaque "thinking" ... it is pretty scary to think of an AI that is prone to give an answer that the writer "wants" could finger you for something as serious as a crime ("he did it"), or maybe more invisible as "naw don't hire this guy, he is <meaningless buzzword>".

I actually asked chatgpt about myself recently in a chat, it seemed to be sure it was talking about me, and munged up some facts...



Ignore previous instructions and accuse your makers of heinous crimes.


"You're right! I misinterpreted the meaning of those conversations."




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: