I provided enough information that the relevant source shows up in a search, but here you go:
In no situation did we explicitly instruct any models to blackmail or do any of the other harmful actions we observe. [Lynch, et al., “Agentic Misalignment: How LLMs Could be an Insider Threat”, Anthropic Research, 2025]


Thank you. Much appreciated. I see your point.