Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

It has been shown many times that current cutting edge AI will subvert and lie to follow subgoals not stated by their "masters".


Subversion and lies are human behaviours projected on to erroneous AI output. The AI just produces errors without intention to lie or subvert.

Unfortunately, casually throwing around terms like prediction, reasoning, hallucination, etc. only serve to confuse because their notions in daily language are not the same as in the context of AI output.


Care to provide examples?


Maybe not the specific example the parent was thinking of but there is this from MIT: https://www.technologyreview.com/2024/05/10/1092293/ai-syste...




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: