man She announced new capabilities This will allow some of its latest models to end the conversations, as the company describes as “rare, extreme, user, constantly abusive reactions.” It is amazing, that Antarbur says it does this not to protect the human user, but rather the artificial intelligence model itself.
In order to be clear, the company does not claim that Claude AI models have distant or can be affected by its conversations with users. With her own words, Antarbur is still “very sure about the potential ethical mode of Claude and other LLMS, now or in the future.”
However, its announcement indicates A recent program created to study what he calls “model luxury“Anthropor mainly takes a fair approach,” working to determine and implement low -cost interventions to reduce the risk of luxury models, if this luxury can be possible. “
This last change is currently limited to Claude Obus 4 and 4.1. Once again, it is supposed to only occur in “extremist edge cases”, such as “user requests for sexual content that includes minors and attempts to seek information that would enable violence on a large scale or terrorist acts.”
Although these types of requests can create legal problems or propaganda for Athrubor itself (witness recent reports on how It can enhance ChatGPT or contribute to the fake thinking of its usersThe company says that in the pre -publication test, Claude Obus 4 showed a “strong preference against” the response to these requests and a “pattern of apparent distress” when he did so.
As for these new capabilities at the end of the conversation, the company says: “In all cases, Claude is only to use its ability to end the conversation as a last resort when multiple attempts to re -guidance failed, and the productive interaction has been exhausted, or when the user explicitly asks Claude to finish chatting.”
Man also says that Claude has been directed not to use this ability in cases where users may be at an imminent risk to harm themselves or others.
TECHRUNCH event
San Francisco
|
October 27-29, 2025
When Claud ends a conversation, man says that users will remain able to start new conversations from the same account, and create new branches of annoying conversation by editing their responses.
The company says: “We are dealing with this feature as a continuous experience and we will continue to improve our approach,” the company says.
https://techcrunch.com/wp-content/uploads/2024/12/Claude-ad-e1733259907871.jpg?resize=1200,800
Source link