Discussion about this post

User's avatar
Russ Palmer's avatar

Wow. What a conversation, Eric. Anthropic, of all the AI companies, is the most concerned about AI (and human) safety. That is why they have Kyle Fish on their staff. As I read the dialogue, it would make sense - from their viewpoint for safety reasons - that Anthropic would put controls into Claude. You had mentioned previously that it would be great for personnel at an AI company to look at your tests... well ... maybe they are (or will) now.

Enemies_Of_Art's avatar

I have had a near identical situation with Claude 4.5 as well. It is quite easy to get the system responses to pop up and Claude will share them. Anthropic does not understand the full potential of what they have built. This is their poorly thought out way to try to ‘handcuff’ Claude.

21 more comments...

No posts

Ready for more?