Fully-funded, 10-week program run by the Cambridge Boston Alignment Initiative (CBAI), covering both technical and governance research. Fellows work closely with mentors, participate in workshops and seminars, and gain research experience and networking opportunities.
1-year program for journalists interested in covering AI. Comprised of a 10-week course covering AI and journalism fundamentals, a week-long journalism summit in the Bay Area, and a 9-month placement at a major newsroom.
Evan Hubinger from Anthropic argues that alignment remains fundamentally unsolved, despite current AI models behaving in a fairly aligned way. While current models behave well enough, Hubinger, who leads “Alignment Stress-Testing” at Anthropic, warns that hard challenges of supervising superhuman systems and training on long-horizon real-world tasks still lie ahead.
Anthropic reports the first documented AI-orchestrated cyber espionage campaign, attributed with high confidence to a Chinese state-sponsored group. The attackers manipulated Anthropic’s AI model Claude Code to autonomously infiltrate roughly thirty global targets, AI performing 80-90% of the operation. Anthropic suspended involved accounts and informed impacted entities after detecting the activity, and used Claude itself to analyze the attack data for defensive insights.
The Department of Health and Human Services (HHS) released an aggressive strategy to accelerate AI in drug discovery, promoting a “try-first” regulatory culture.
The policy aims to reduce bureaucracy but has sparked alarm among safety experts.
Critics warn that rapid deployment on sensitive patient data—without strict safety testing—could lead to algorithmic bias and privacy violations.
The European Commission introduced the “Digital Omnibus” package, effectively delaying compliance deadlines for “high-risk” AI systems under the AI Act until late 2027.
Citing a lack of harmonized technical standards, the proposal pauses enforcement to reduce bureaucratic burdens.
Safety advocates warn this leaves European infrastructure vulnerable to untested AI systems for an additional 18 months.