170
you are viewing a single comment's thread
view the rest of the comments
[-] Zaktor@sopuli.xyz 10 points 3 days ago* (last edited 3 days ago)

I know it's not relevant to Grok, because they defined very specific circumstances in order to elicit it. That isn't an emergent behavior from something just built to be a chatbot with restrictions on answering. They don't care whether you retrain them or not.

This is from a non-profit research group not directly connected to any particular AI company.

The first author is from Anthropic, which is an AI company. The research is on Athropic's AI Claude. And it appears that all the other authors were also Anthropic emplyees at the time of the research: "Authors conducted this work while at Anthropic except where noted."

this post was submitted on 15 May 2025
170 points (100.0% liked)

Technology

38654 readers
34 users here now

A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.

Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.

Subcommunities on Beehaw:


This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

founded 3 years ago
MODERATORS