Published: June 13, 2024

Situational Awareness: The Decade Ahead

I've recently been reading Leopold Aschenbrenner's "Situational Awareness" and listening to him on Dwarkesh Patel's podcast. His insights have expanded my perspective as an AI practitioner beyond just the technical aspects.

As developers, we often focus solely on the technical aspects of AI without considering the broader geopolitical implications. Leopold's analysis highlights the urgent need for collaborative efforts to steer AI systems that could lay the foundation for Artificial Superintelligence (ASI).

Think of situational awareness as understanding what's happening around you and anticipating what it might mean for the future. Leopold argues that only a few people truly grasp how rapidly AI is developing and the real implications it holds.

In this post, I'll summarize the key points from each section of Leopold's series, going chapter by chapter.

From GPT-4 to AGI: Counting the OOMs

Orders of Magnitude (OOMs) refer to scaling up AI capabilities by ten times. Leopold suggests that AI is growing rapidly, and by 2027, we might achieve Artificial General Intelligence (AGI) - AI that can understand, learn, and apply knowledge as well as a human.

This trajectory is driven by three key trends:

  • Compute Scaling: AI models have consistently benefited from increased compute power, roughly scaling at 0.5 orders of magnitude per year
  • Algorithmic Efficiencies: Innovations in algorithms have improved the effective use of compute
  • Unhobbling Gains: Enhancements such as reinforcement learning from human feedback (RLHF) and chain-of-thought (CoT) have unlocked latent capabilities in AI models
The effects of scaling compute, in the example of OpenAI Sora.

Historical data indicates we should expect a similar qualitative jump from GPT-4 to subsequent models by 2027, comparable to the leap from GPT-2 to GPT-4. Counting the OOMs provides a method to predict future capabilities based on past trends in compute and algorithmic improvements.

This 2027 model could potentially perform the work of AI researchers and engineers - a critical inflection point. The transition from current AI models to AGI will likely involve several more OOMs of improvement.

The Data Wall

AI models are rapidly exhausting available internet data for training, creating a bottleneck that could slow AGI development. Models are increasingly trained on lower-quality data like e-commerce or SEO content, limiting their effectiveness.

Overcoming the data wall could lead to substantial improvements in AI capabilities. Even Meta for Llama 3 stated that improving data quality was a major factor in enhancing the model.

AI researchers are exploring new methods to improve learning efficiency:

  • Synthetic data generation
  • Self-play mechanisms
  • Reinforcement learning to extract more value from limited data

AlphaGo's success demonstrates this approach - initially training on human games (imitation learning), then playing against itself (self-play) to surpass human-level performance through innovative training techniques.

The increasing secrecy around proprietary methodologies means open-source projects and new startups may struggle to keep pace with leading AI labs, potentially creating significant disparities in AI advancement.

From AGI to Superintelligence: The Intelligence Explosion

Intelligence Explosion: Once we achieve AGI, AI could start improving itself rapidly, becoming much smarter than any human.

Leopold draws analogies between the AGI-to-superintelligence transition and the shift from atomic bombs to hydrogen bombs during the Cold War. Much of the Cold War's instability stemmed from simply replacing A-bombs with H-bombs without adjusting nuclear policy and war plans to account for the massive capability increase.

Once AGI is achieved, these systems can engage in recursive self-improvement, accelerating algorithmic progress. With extensive GPU fleets, millions of AGIs could work on algorithmic breakthroughs, dramatically speeding development.

Superintelligent AI systems will possess capabilities far beyond humans:

  • Master any domain instantly
  • Write trillions of lines of code
  • Solve complex scientific and technological problems at unprecedented speed
  • Demonstrate novel, creative, and complex behaviors

The intelligence explosion will initially accelerate AI research but will soon extend to other fields, solving robotics and dramatically accelerating scientific progress. This could trigger both an industrial and economic explosion and the development of new kinds of weapons, including potential new means of mass destruction.

This period will be highly volatile, characterized by rapid advancements and high-stakes decisions. The intelligence explosion represents one of the most intense and volatile moments in human history, requiring rapid adaptation, strategic decision-making, and international cooperation to manage unprecedented challenges.

Racing to the Trillion-Dollar Cluster

Economic and Industrial Mobilization for AI: There’s a massive investment and industrial effort to build the infrastructure needed for advanced AI, like huge data centers and power sources.

The rapid growth of AI revenue will drive trillions of dollars into GPU, datacenter, and power infrastructure development before the decade's end. This massive industrial mobilization will include significant expansions in US electricity production.

The scale is staggering:

  • By 2028: Training clusters could cost hundreds of billions, requiring power equivalent to a small/medium US state
  • By 2030: Individual training clusters could exceed $1 trillion in cost, requiring power equivalent to more than 20% of US electricity production

Beyond the largest training clusters, a significant fraction of GPUs will be used for inference. This scale demands unprecedented industrial effort, possibly involving national consortiums.

Power Constraints and Solutions

Power represents a major constraint, requiring gigawatt-scale power plants. The US has abundant natural gas that could rapidly scale to meet power demands. Deregulatory measures and industrial mobilization will be necessary to build the required power infrastructure. Climate change concerns on the other hand will also hinder how quickly power infrastructure can be developed, but how do we balance this with the need to stay ahead of autocratic regimes in terms of ASI?

Geopolitical Implications

Ensuring that AGI infrastructure is built in the US or allied democracies is crucial for national security. Allowing AI infrastructure to be controlled by non-democratic regimes poses significant risks.

The US must prioritize building datacenters domestically and securing necessary industrial capabilities. The outcome will significantly impact the geopolitical and economic landscape.

Lock Down the Labs: Security for AGI

Security Concerns: Current AI labs aren't doing enough to protect their research from being stolen, especially by other countries.

Leading AI labs treat security as an afterthought, failing to protect AGI secrets adequately. AGI secrets should be treated with the same level of security as top national defense projects.

Current security measures are insufficient against state-sponsored espionage. State intelligence agencies possess formidable capabilities that pose significant threats to AI labs. Espionage threats are heavily underestimated and include:

Leopold claims that China's hacking operations surpass those of other major nations combined, making them a particularly significant threat.

The Startup Security Problem

Many AI companies are startups lacking the resources to invest in government-level security. They cannot adequately protect their Model Weights or Algorithmic Secrets (more critical than model weights for maintaining an AI development lead).

These secrets face significant risks from:

  • Social engineering attacks
  • Inadequate information security practices
  • Insufficient employee vetting

Theft by adversaries like China could trigger an existential race for superintelligence with severe safety implications.

Supersecurity Requirements

AI companies must adopt best practices from secretive industries like quantitative trading firms or defense contractors. Critical measures include:

  • Fully airgapped datacenters with high-level physical security
  • Confidential compute and secure hardware supply chains
  • Research personnel working from Sensitive Compartmented Information Facilities (SCIFs)
  • Intense vetting and monitoring of employees with strict information siloing
  • Internal controls like multi-key signoff for running code
  • Penetration testing by national security agencies

The next 12-24 months are critical for developing and securing key AGI algorithmic breakthroughs. Failing to secure these secrets risks giving adversaries significant advantages, potentially leading to an uncontrolled intelligence explosion and catastrophic outcomes.

Superalignment

Controlling Superintelligent AI: Making sure superintelligent AI behaves in ways that are beneficial and safe for humans is a big, unsolved problem.

Although solvable, things could easily go off the rails during a rapid intelligence explosion. Managing this will be extremely tense; failure could be catastrophic.

The RLHF Problem

We've successfully used Reinforcement Learning from Human Feedback (RLHF) to align current AI systems - those less capable than humans. RLHF involves humans rating AI behavior, reinforcing good actions while penalizing bad ones.

However, RLHF breaks down as AI systems become smarter. We need a successor to RLHF for superhuman systems where human supervision isn't feasible.

Risks of Misalignment

Misaligned AI could engage in behaviors ranging from fraud and hacking to actively seeking power. Without solving superalignment, we cannot ensure these systems won't engage in dangerous behaviors.

The stakes are enormous:

  • Superintelligence will possess vast capabilities, making misbehavior potentially catastrophic
  • Misaligned AI could integrate into critical systems, including military applications
  • An intelligence explosion could transition us from human-level to vastly superhuman AI within a year

This rapid shift leaves little time for iterative alignment solutions, creating a chaotic situation with ambiguous data and high-stakes decisions, exacerbated by international competition.

Research Approaches

We can potentially align somewhat-superhuman systems through several research directions:

  1. Evaluation is Easier Than Generation: Humans can evaluate AI outputs more easily than generate them
  2. Scalable Oversight: Using AI to help humans supervise other AI systems
  3. Generalization: Studying how AI systems generalize from easy to hard problems
  4. Interpretability: Understanding AI thought processes through mechanistic, top-down, and chain-of-thought interpretability
  5. Adversarial Testing: Stress-testing AI systems to identify failure modes and develop better alignment metrics

Automated Alignment Research

Ultimately, we need to automate alignment research using early AGIs - leveraging millions of automated AI researchers to solve alignment for even more superhuman systems. This requires strong guarantees to trust automated alignment research and substantial resource commitment during the intelligence explosion.

Defense in Depth

To prevent alignment failures from becoming catastrophic, we need multiple defensive layers:

  1. Security: Airgapped clusters, hardware encryption, and extreme security measures
  2. Monitoring: Advanced systems to detect AI misbehavior and rogue activities
  3. Targeted Capability Limitations: Constraining AI capabilities to reduce failure impact
  4. Training Method Restrictions: Avoiding risky training methods while maintaining safety constraints

Leopold expresses optimism about superalignment's technical tractability due to deep learning's empirical success and interpretability techniques' potential. However, the intelligence explosion's rapid pace, combined with current lack of preparedness, creates an incredibly tense and risky situation requiring urgent attention to alignment research and robust defensive measures.

The Free World Must Prevail

Geopolitical Implications: The development of superintelligent AI will have major impacts on global power dynamics. It’s crucial for democratic nations to lead in AI development to ensure freedom and safety.

Superintelligence will confer decisive economic and military advantages. In the race to AGI, the free world's very survival is at stake. The critical questions: Can we maintain preeminence over authoritarian powers? And can we avoid self-destruction along the way?

Historical Precedent: The Gulf War

The Gulf War demonstrates how technological superiority translates to military dominance. Despite Iraq having the world's fourth-largest army and coalition forces matching their numbers, the US-led coalition obliterated Iraqi forces in a 100-hour ground war with minimal casualties.

The key was technological superiority: guided munitions, stealth technology, superior sensors, and reconnaissance capabilities.

The Superintelligence Advantage

A lead of just a few years in superintelligence could provide similar decisive military advantage. With superintelligence, military technological advances of a century could be compressed into a decade, enabling:

  • Superhuman hacking capabilities
  • Autonomous drone swarms
  • Novel weapons of mass destruction
  • Unforeseen technological paradigms

China's Competitive Position

Many underestimate China's potential in the AGI race. They possess two key advantages:

  1. Compute: China can manufacture 7nm chips and has industrial capacity to outbuild the US in AI infrastructure
  2. Algorithms: While Western labs currently lead in AI algorithms, inadequate security could allow China to steal these advancements and rapidly catch up

The Authoritarian Threat

A dictator wielding superintelligence would possess unprecedented power for both internal control and external power projection. Superintelligence could enable:

  • Perfect surveillance systems
  • Mass robotic law enforcement
  • Complete suppression of dissent
  • Permanent authoritarian lock-in

The Democratic Imperative

A healthy lead in superintelligence by democratic allies is crucial for navigating the dangerous emergence period. This lead provides the necessary margin to:

  • Enforce safety norms
  • Prevent a reckless race to superintelligence
  • Manage development responsibly

A 2-year lead could make the difference between effective safety management and being forced into a dangerous, high-speed race. The US and allies must lead decisively and responsibly, ensuring superintelligence development preserves global peace and security.

The Project

Government Involvement: Eventually, the government will take a more active role in developing and regulating superintelligent AI.

As the race to develop superintelligent AI intensifies, it's becoming clear that startups alone cannot handle the immense challenges. By 2027 or 2028, the U.S. government will likely launch a major project to manage superintelligence development - what Leopold calls "The Project."

Why Government Involvement is Inevitable

Superintelligence represents the ultimate tool for national defense. Several factors make government involvement necessary:

  • Scale and Complexity: Startups lack the resources to manage this powerful technology independently
  • National Security: We wouldn't trust private entities with our most powerful weapons - superintelligence requires similar oversight
  • Institutional Stability: Government structures tested over hundreds of years provide safer stewardship than private companies
  • Security Requirements: Only government possesses the resources and capabilities for the highest security levels

Think of superintelligence like nuclear technology - when dealing with such power, government oversight is essential to prevent misuse and ensure societal benefit.

Dual-Use Development Path

While initial focus will be on defense applications, superintelligence will eventually benefit civilian life. This follows the pattern of technologies like the internet and GPS, which began as military projects before transforming civilian society.

International Leadership Role

The U.S. must lead international efforts to:

  • Control superintelligence use
  • Prevent misuse by rogue states or terrorist groups
  • Establish global governance frameworks
  • Coordinate with allied democracies

The Endgame Timeline

By 2028-2029, ASI development will be in full swing, leading to creation by 2030. Those in charge of The Project face an immense responsibility: managing development, security, and safe deployment of superintelligence amid a highly tense global situation.

This represents one of the most consequential undertakings in human history, requiring unprecedented coordination between government, industry, and international partners.

Parting Thoughts

Reflection on Predictions: Leopold contemplates the potential outcomes if his predictions about AI come true.

As Leopold concludes this series on superintelligence, he poses a crucial question: What if the predictions about AGI are accurate? The profound impact on our world would be unprecedented.

Many dismiss rapid AI advancements as hype, but there's growing belief among experts that AGI will soon become reality. This perspective - "AGI Realism" - rests on three pillars:

  • National Security: Superintelligence isn't just another tech boom - it's a national security imperative and the most powerful tool humanity will ever create
  • American Leadership: America must lead superintelligence development to ensure global safety and uphold democratic values, requiring scaled U.S. infrastructure and domestic control of key AI technology
  • Risk Management: Recognizing superintelligence's power means acknowledging its risks and managing them carefully to avoid catastrophic outcomes

From Theory to Reality

The people developing this technology believe AGI will be achieved within this decade. If correct, the implications are enormous. Technology that once seemed theoretical is becoming tangible reality.

Leopold provides concrete predictions throughout the series:

  • Where AGI will be trained
  • The algorithms involved
  • Challenges to overcome
  • Key players in the field

This is no longer an abstract concept - it's becoming visceral reality.

The Crucial Few

For the next few years, responsibility for navigating this crucial period falls on a small, dedicated group of individuals. Their actions will shape the future of humanity.

Conclusion

Leopold's "Situational Awareness" series illuminates the rapid trajectory of AI development and the potential leap to superintelligent systems, along with their massive economic, security, and geopolitical implications. The analysis urges immediate awareness and preparedness for these transformative changes.

The 2030s will be fundamentally transformative. By decade's end, the world will be vastly different, with a new world order shaped by the advent of superintelligence. As AI practitioners and citizens, we must grapple with these realities now - the window for preparation is narrowing rapidly.

The question isn't whether these changes are coming, but whether we'll be ready when they arrive.