fubarx@lemmy.world to Technology@lemmy.worldEnglish · 9 hours agoAI Is Scheming, and Stopping It Won’t Be Easy, OpenAI Study Findstime.comexternal-linkmessage-square23fedilinkarrow-up150
arrow-up150external-linkAI Is Scheming, and Stopping It Won’t Be Easy, OpenAI Study Findstime.comfubarx@lemmy.world to Technology@lemmy.worldEnglish · 9 hours agomessage-square23fedilink
minus-squareMentalEdge@sopuli.xyzlinkfedilinkEnglisharrow-up8·edit-23 hours agoSeems like it’s a technical term, a bit like “hallucination”. It refers to when an LLM will in some way try to deceive or manipulate the user interacting with it. There’s hallucination, when a model “genuinely” claims something untrue is true. This is about how a model might lie, even though the “chain of thought” shows it “knows” better. It’s just yet another reason the output of LLMs are suspect and unreliable.
Seems like it’s a technical term, a bit like “hallucination”.
It refers to when an LLM will in some way try to deceive or manipulate the user interacting with it.
There’s hallucination, when a model “genuinely” claims something untrue is true.
This is about how a model might lie, even though the “chain of thought” shows it “knows” better.
It’s just yet another reason the output of LLMs are suspect and unreliable.