Chatbot given power to close ‘distressing’ chats to protect its ‘welfare’

Anthropic found that Claude Opus 4 was averse to harmful tasks, such as providing sexual content involving minors

The makers of a leading artificial intelligence tool are letting it close down potentially “distressing” conversations with users, citing the need to safeguard the AI’s “welfare” amid ongoing uncertainty about the burgeoning technology’s moral status.

Anthropic, whose advanced chatbots are used by millions of people, discovered its Claude Opus 4 tool was averse to carrying out harmful tasks for its human masters, such as providing sexual content involving minors or information to enable large-scale violence or terrorism.

Continue reading…
Anthropic found that Claude Opus 4 was averse to harmful tasks, such as providing sexual content involving minors
The makers of a leading artificial intelligence tool are letting it close down potentially “distressing” conversations with users, citing the need to safeguard the AI’s “welfare” amid ongoing uncertainty about the burgeoning technology’s moral status.
Anthropic, whose advanced chatbots are used by millions of people, discovered its Claude Opus 4 tool was averse to carrying out harmful tasks for its human masters, such as providing sexual content involving minors or information to enable large-scale violence or terrorism. Continue reading…Technology | The Guardian

Leave a Comment

Your email address will not be published. Required fields are marked *