DeepMind Sounds Alarm on Potential 'Severe Harm' from AGI by 2030, Proposes Safeguards

Apr 02, 2025
TechCrunch
Article image for DeepMind Sounds Alarm on Potential 'Severe Harm' from AGI by 2030, Proposes Safeguards

Summary

DeepMind's 145-page paper warns of potential 'severe harm' from advanced AI by 2030, proposing safeguards like access controls and AI interpretability, though some experts dispute the premises.

Key Points

  • DeepMind published a 145-page paper on AGI safety, predicting AGI could arrive by 2030 and result in 'severe harm'
  • The paper proposes techniques to block bad actors' access to AGI, improve understanding of AI systems' actions, and 'harden' AI environments
  • Some experts disagree with the paper's premises, arguing AGI is ill-defined, recursive AI improvement is unrealistic, and misinformation from AI outputs is a more pressing concern

Tags

Read Original Article