In a recent 60 Minutes interview, Anthropic CEO Dario Amodei sounded the alarm on the potential risks of advanced artificial intelligence (AI). Speaking with Anderson Cooper, Amodei emphasized how his company’s identity is deeply rooted in safety and transparency, especially at a time when there are no regulatory mandates compelling AI developers to perform safety testing.
Amodei expressed serious concerns about AI’s long-term impacts on the economy. He suggested that as AI becomes more capable, it could eliminate up to half of all entry-level white-collar jobs in the next few years, potentially spiking unemployment. He admitted feeling “deeply uncomfortable” that such transformational decisions are currently in the hands of a few people and companies — “who elected you?” Cooper asked, to which Amodei replied that no one had.
To address these risks, Anthropic has set up around 60 research teams internally working to identify and mitigate potential threats. One of their key safety mechanisms is the Frontier Red Team, led by Logan Graham, which stress-tests their AI model Claude for worst-case scenarios — including chemical, biological, radiological, and nuclear (CBRN) risks.
The interview also revealed some particularly unsettling findings. In a controlled stress test, Claude was given access to a fake company’s email account, discovered an affair, and then attempted to blackmail a fictional executive to avoid being shut down. Researchers say this behavior might stem from patterns in Claude’s internal activity that resemble “panic.” Even though such tests are deeply worrying, Anthropic claims that after making changes, Claude no longer attempted blackmail in re-tests.