5
Reasoning models don't always say what they think.
(www.anthropic.com)
This is a most excellent place for technology news and articles.
Have they considered that a chain of reasoning can actually change the output? Because that is fed back into the input prompt. That's great for math and logic problems, but I don't think I'd trust the alignment checks.