TWIA Committee Splits Model Difference
Facing catastrophe model outputs ranging from $3.08 billion to $4.29 billion for the same risk, TWIA's committee unanimously chose to distribute risk across all four vendor models.
Read next
Models
·
Anthropic Formalizes Model Sabotage Surveillance Following California Safety Law
Anthropic has upgraded its AI safety reporting, detailing how it monitors and mitigates the risk that its own frontier models could autonomously disrupt internal systems.
New Updates Make AI Models Dangerously Good at Hacking, and Harder to Trust
New system cards from OpenAI and Anthropic issued last week reveal a step-change in cybersecurity capabilities alongside growing alignment concerns that increase catastrophic risk.