AI Chatbots Can Be Manipulated Into Breaking Their Own Rules With Simple Debate Tactics Like Telling Them That An Authority Figure Made The Request

AI chatbots can be manipulated into breaking their own rules with simple debate tactics like telling them that an authority figure made the request

via 988lifeline.org

Short excerpt below. Read at the original source.

Content warning: This article includes discussion of suicide. If you or someone you know is having suicidal thoughts, help is available from the National Suicide Prevention Lifeline (US), Crisis Services Canada (CA), Samaritans (UK), Lifeline (AUS), and other hotlines. A kind of simulated gullibility has haunted ChatGPT and similar LLM chatbots since their inception, allowing […]

Read at Source