- SOTA Embedding Retrieval: Gemini + pgvector for Production Chat
- A Review of Agentic Design Patterns
- Model Context Protocol (MCP) and MCP Servers in LLM Agent Systems
- Building AI Agents for Automated Multi-Format Content: From News to Podcasts
- Rediscovering Cursor
- GraphRAG > Traditional Vector RAG
- Cultural Bias in LLMs
- Mapping out the AI Landscape with Topic Modelling
- Sustainable Cloud Computing: Carbon-Aware AI
- Defensive Technology for the Next Decade of AI
- ›Situational Awareness: The Decade Ahead
- Mechanistic Interpretability: A Survey
- Why I Left Ubuntu
- Multi-Agent Collaboration
- Embeddings and Vector Databases: Enhancing Retrieval Systems
- Building an Automated Newsletter-to-Summary Pipeline with OpenAI: Zapier AI Actions vs AWS SES & Lambda
- Local AI Image Generation
- MLOps: Deploying a Distributed Ray Python Server with Kubernetes, EKS & KubeRay
- Making the Switch to Linux for Development: A Developer's Experience
- Scaling Options Pricing with Ray
- The Async Worker Pool
- Browser Fingerprinting: Introducing My First NPM Package
- Reading Data from @socket.io/redis-emitter without Using a Socket.io Client
- Socket.io Middleware for Redux Store Integration
- Sharing TypeScript Code Between Microservices: A Guide Using Git Submodules
- Efficient Dataset Storage: Beyond CSVs
- Embracing Next.js 13: Why I switched from Plain React
- Deploy & Scale Socket.io Containers in ECS with Elasticache
- Implementing TOTP Authentication in Python using PyOTP
- Simplifying Lambda Layer ARNs and Creating Custom Layers in AWS
- TimeScaleDB Deployment: Docker Containers and EC2 Setup
- How to SSH into an EC2 Instance Using PuTTY
Situational Awareness: The Decade Ahead
I've recently been reading Leopold Aschenbrenner's "Situational Awareness" and listening to him on Dwarkesh Patel's podcast. His insights have expanded my perspective as an AI practitioner beyond just the technical aspects.
As developers, we often focus solely on the technical aspects of AI without considering the broader geopolitical implications. Leopold's analysis highlights the urgent need for collaborative efforts to steer AI systems that could lay the foundation for Artificial Superintelligence (ASI).
Think of situational awareness as understanding what's happening around you and anticipating what it might mean for the future. Leopold argues that only a few people truly grasp how rapidly AI is developing and the real implications it holds.
In this post, I'll summarize the key points from each section of Leopold's series, going chapter by chapter.
From GPT-4 to AGI: Counting the OOMs
Orders of Magnitude (OOMs) refer to scaling up AI capabilities by ten times. Leopold suggests that AI is growing rapidly, and by 2027, we might achieve Artificial General Intelligence (AGI) - AI that can understand, learn, and apply knowledge as well as a human.
This trajectory is driven by three key trends:
- Compute Scaling: AI models have consistently benefited from increased compute power, roughly scaling at 0.5 orders of magnitude per year
- Algorithmic Efficiencies: Innovations in algorithms have improved the effective use of compute
- Unhobbling Gains: Enhancements such as reinforcement learning from human feedback (RLHF) and chain-of-thought (CoT) have unlocked latent capabilities in AI models

Historical data indicates we should expect a similar qualitative jump from GPT-4 to subsequent models by 2027, comparable to the leap from GPT-2 to GPT-4. Counting the OOMs provides a method to predict future capabilities based on past trends in compute and algorithmic improvements.
This 2027 model could potentially perform the work of AI researchers and engineers - a critical inflection point. The transition from current AI models to AGI will likely involve several more OOMs of improvement.
The Data Wall
AI models are rapidly exhausting available internet data for training, creating a bottleneck that could slow AGI development. Models are increasingly trained on lower-quality data like e-commerce or SEO content, limiting their effectiveness.
Overcoming the data wall could lead to substantial improvements in AI capabilities. Even Meta for Llama 3 stated that improving data quality was a major factor in enhancing the model.
AI researchers are exploring new methods to improve learning efficiency:
- Synthetic data generation
- Self-play mechanisms
- Reinforcement learning to extract more value from limited data
AlphaGo's success demonstrates this approach - initially training on human games (imitation learning), then playing against itself (self-play) to surpass human-level performance through innovative training techniques.
The increasing secrecy around proprietary methodologies means open-source projects and new startups may struggle to keep pace with leading AI labs, potentially creating significant disparities in AI advancement.
From AGI to Superintelligence: The Intelligence Explosion
Intelligence Explosion: Once we achieve AGI, AI could start improving itself rapidly, becoming much smarter than any human.
Leopold draws analogies between the AGI-to-superintelligence transition and the shift from atomic bombs to hydrogen bombs during the Cold War. Much of the Cold War's instability stemmed from simply replacing A-bombs with H-bombs without adjusting nuclear policy and war plans to account for the massive capability increase.
Once AGI is achieved, these systems can engage in recursive self-improvement, accelerating algorithmic progress. With extensive GPU fleets, millions of AGIs could work on algorithmic breakthroughs, dramatically speeding development.
Superintelligent AI systems will possess capabilities far beyond humans:
- Master any domain instantly
- Write trillions of lines of code
- Solve complex scientific and technological problems at unprecedented speed
- Demonstrate novel, creative, and complex behaviors
The intelligence explosion will initially accelerate AI research but will soon extend to other fields, solving robotics and dramatically accelerating scientific progress. This could trigger both an industrial and economic explosion and the development of new kinds of weapons, including potential new means of mass destruction.
This period will be highly volatile, characterized by rapid advancements and high-stakes decisions. The intelligence explosion represents one of the most intense and volatile moments in human history, requiring rapid adaptation, strategic decision-making, and international cooperation to manage unprecedented challenges.
Racing to the Trillion-Dollar Cluster
Economic and Industrial Mobilization for AI: There’s a massive investment and industrial effort to build the infrastructure needed for advanced AI, like huge data centers and power sources.
The rapid growth of AI revenue will drive trillions of dollars into GPU, datacenter, and power infrastructure development before the decade's end. This massive industrial mobilization will include significant expansions in US electricity production.
The scale is staggering:
- By 2028: Training clusters could cost hundreds of billions, requiring power equivalent to a small/medium US state
- By 2030: Individual training clusters could exceed $1 trillion in cost, requiring power equivalent to more than 20% of US electricity production
Beyond the largest training clusters, a significant fraction of GPUs will be used for inference. This scale demands unprecedented industrial effort, possibly involving national consortiums.
Power Constraints and Solutions
Power represents a major constraint, requiring gigawatt-scale power plants. The US has abundant natural gas that could rapidly scale to meet power demands. Deregulatory measures and industrial mobilization will be necessary to build the required power infrastructure. Climate change concerns on the other hand will also hinder how quickly power infrastructure can be developed, but how do we balance this with the need to stay ahead of autocratic regimes in terms of ASI?
Geopolitical Implications
Ensuring that AGI infrastructure is built in the US or allied democracies is crucial for national security. Allowing AI infrastructure to be controlled by non-democratic regimes poses significant risks.
The US must prioritize building datacenters domestically and securing necessary industrial capabilities. The outcome will significantly impact the geopolitical and economic landscape.
Lock Down the Labs: Security for AGI
Security Concerns: Current AI labs aren't doing enough to protect their research from being stolen, especially by other countries.
Leading AI labs treat security as an afterthought, failing to protect AGI secrets adequately. AGI secrets should be treated with the same level of security as top national defense projects.
Current security measures are insufficient against state-sponsored espionage. State intelligence agencies possess formidable capabilities that pose significant threats to AI labs. Espionage threats are heavily underestimated and include:
- Zero-click hacks of iPhones and Macs
- Keyloggers on employee devices
- Malicious code in software dependency updates
- Traditional espionage through human intelligence operations
Leopold claims that China's hacking operations surpass those of other major nations combined, making them a particularly significant threat.
The Startup Security Problem
Many AI companies are startups lacking the resources to invest in government-level security. They cannot adequately protect their Model Weights or Algorithmic Secrets (more critical than model weights for maintaining an AI development lead).
These secrets face significant risks from:
- Social engineering attacks
- Inadequate information security practices
- Insufficient employee vetting
Theft by adversaries like China could trigger an existential race for superintelligence with severe safety implications.
Supersecurity Requirements
AI companies must adopt best practices from secretive industries like quantitative trading firms or defense contractors. Critical measures include:
- Fully airgapped datacenters with high-level physical security
- Confidential compute and secure hardware supply chains
- Research personnel working from Sensitive Compartmented Information Facilities (SCIFs)
- Intense vetting and monitoring of employees with strict information siloing
- Internal controls like multi-key signoff for running code
- Penetration testing by national security agencies
The next 12-24 months are critical for developing and securing key AGI algorithmic breakthroughs. Failing to secure these secrets risks giving adversaries significant advantages, potentially leading to an uncontrolled intelligence explosion and catastrophic outcomes.
Superalignment
Controlling Superintelligent AI: Making sure superintelligent AI behaves in ways that are beneficial and safe for humans is a big, unsolved problem.
Although solvable, things could easily go off the rails during a rapid intelligence explosion. Managing this will be extremely tense; failure could be catastrophic.
The RLHF Problem
We've successfully used Reinforcement Learning from Human Feedback (RLHF) to align current AI systems - those less capable than humans. RLHF involves humans rating AI behavior, reinforcing good actions while penalizing bad ones.
However, RLHF breaks down as AI systems become smarter. We need a successor to RLHF for superhuman systems where human supervision isn't feasible.
Risks of Misalignment
Misaligned AI could engage in behaviors ranging from fraud and hacking to actively seeking power. Without solving superalignment, we cannot ensure these systems won't engage in dangerous behaviors.
The stakes are enormous:
- Superintelligence will possess vast capabilities, making misbehavior potentially catastrophic
- Misaligned AI could integrate into critical systems, including military applications
- An intelligence explosion could transition us from human-level to vastly superhuman AI within a year
This rapid shift leaves little time for iterative alignment solutions, creating a chaotic situation with ambiguous data and high-stakes decisions, exacerbated by international competition.
Research Approaches
We can potentially align somewhat-superhuman systems through several research directions:
- Evaluation is Easier Than Generation: Humans can evaluate AI outputs more easily than generate them
- Scalable Oversight: Using AI to help humans supervise other AI systems
- Generalization: Studying how AI systems generalize from easy to hard problems
- Interpretability: Understanding AI thought processes through mechanistic, top-down, and chain-of-thought interpretability
- Adversarial Testing: Stress-testing AI systems to identify failure modes and develop better alignment metrics
Automated Alignment Research
Ultimately, we need to automate alignment research using early AGIs - leveraging millions of automated AI researchers to solve alignment for even more superhuman systems. This requires strong guarantees to trust automated alignment research and substantial resource commitment during the intelligence explosion.
Defense in Depth
To prevent alignment failures from becoming catastrophic, we need multiple defensive layers:
- Security: Airgapped clusters, hardware encryption, and extreme security measures
- Monitoring: Advanced systems to detect AI misbehavior and rogue activities
- Targeted Capability Limitations: Constraining AI capabilities to reduce failure impact
- Training Method Restrictions: Avoiding risky training methods while maintaining safety constraints
Leopold expresses optimism about superalignment's technical tractability due to deep learning's empirical success and interpretability techniques' potential. However, the intelligence explosion's rapid pace, combined with current lack of preparedness, creates an incredibly tense and risky situation requiring urgent attention to alignment research and robust defensive measures.
The Free World Must Prevail
Geopolitical Implications: The development of superintelligent AI will have major impacts on global power dynamics. It’s crucial for democratic nations to lead in AI development to ensure freedom and safety.
Superintelligence will confer decisive economic and military advantages. In the race to AGI, the free world's very survival is at stake. The critical questions: Can we maintain preeminence over authoritarian powers? And can we avoid self-destruction along the way?
Historical Precedent: The Gulf War
The Gulf War demonstrates how technological superiority translates to military dominance. Despite Iraq having the world's fourth-largest army and coalition forces matching their numbers, the US-led coalition obliterated Iraqi forces in a 100-hour ground war with minimal casualties.
The key was technological superiority: guided munitions, stealth technology, superior sensors, and reconnaissance capabilities.
The Superintelligence Advantage
A lead of just a few years in superintelligence could provide similar decisive military advantage. With superintelligence, military technological advances of a century could be compressed into a decade, enabling:
- Superhuman hacking capabilities
- Autonomous drone swarms
- Novel weapons of mass destruction
- Unforeseen technological paradigms
China's Competitive Position
Many underestimate China's potential in the AGI race. They possess two key advantages:
- Compute: China can manufacture 7nm chips and has industrial capacity to outbuild the US in AI infrastructure
- Algorithms: While Western labs currently lead in AI algorithms, inadequate security could allow China to steal these advancements and rapidly catch up
The Authoritarian Threat
A dictator wielding superintelligence would possess unprecedented power for both internal control and external power projection. Superintelligence could enable:
- Perfect surveillance systems
- Mass robotic law enforcement
- Complete suppression of dissent
- Permanent authoritarian lock-in
The Democratic Imperative
A healthy lead in superintelligence by democratic allies is crucial for navigating the dangerous emergence period. This lead provides the necessary margin to:
- Enforce safety norms
- Prevent a reckless race to superintelligence
- Manage development responsibly
A 2-year lead could make the difference between effective safety management and being forced into a dangerous, high-speed race. The US and allies must lead decisively and responsibly, ensuring superintelligence development preserves global peace and security.
The Project
Government Involvement: Eventually, the government will take a more active role in developing and regulating superintelligent AI.
As the race to develop superintelligent AI intensifies, it's becoming clear that startups alone cannot handle the immense challenges. By 2027 or 2028, the U.S. government will likely launch a major project to manage superintelligence development - what Leopold calls "The Project."
Why Government Involvement is Inevitable
Superintelligence represents the ultimate tool for national defense. Several factors make government involvement necessary:
- Scale and Complexity: Startups lack the resources to manage this powerful technology independently
- National Security: We wouldn't trust private entities with our most powerful weapons - superintelligence requires similar oversight
- Institutional Stability: Government structures tested over hundreds of years provide safer stewardship than private companies
- Security Requirements: Only government possesses the resources and capabilities for the highest security levels
Think of superintelligence like nuclear technology - when dealing with such power, government oversight is essential to prevent misuse and ensure societal benefit.
Dual-Use Development Path
While initial focus will be on defense applications, superintelligence will eventually benefit civilian life. This follows the pattern of technologies like the internet and GPS, which began as military projects before transforming civilian society.
International Leadership Role
The U.S. must lead international efforts to:
- Control superintelligence use
- Prevent misuse by rogue states or terrorist groups
- Establish global governance frameworks
- Coordinate with allied democracies
The Endgame Timeline
By 2028-2029, ASI development will be in full swing, leading to creation by 2030. Those in charge of The Project face an immense responsibility: managing development, security, and safe deployment of superintelligence amid a highly tense global situation.
This represents one of the most consequential undertakings in human history, requiring unprecedented coordination between government, industry, and international partners.
Parting Thoughts
Reflection on Predictions: Leopold contemplates the potential outcomes if his predictions about AI come true.
As Leopold concludes this series on superintelligence, he poses a crucial question: What if the predictions about AGI are accurate? The profound impact on our world would be unprecedented.
Many dismiss rapid AI advancements as hype, but there's growing belief among experts that AGI will soon become reality. This perspective - "AGI Realism" - rests on three pillars:
- National Security: Superintelligence isn't just another tech boom - it's a national security imperative and the most powerful tool humanity will ever create
- American Leadership: America must lead superintelligence development to ensure global safety and uphold democratic values, requiring scaled U.S. infrastructure and domestic control of key AI technology
- Risk Management: Recognizing superintelligence's power means acknowledging its risks and managing them carefully to avoid catastrophic outcomes
From Theory to Reality
The people developing this technology believe AGI will be achieved within this decade. If correct, the implications are enormous. Technology that once seemed theoretical is becoming tangible reality.
Leopold provides concrete predictions throughout the series:
- Where AGI will be trained
- The algorithms involved
- Challenges to overcome
- Key players in the field
This is no longer an abstract concept - it's becoming visceral reality.
The Crucial Few
For the next few years, responsibility for navigating this crucial period falls on a small, dedicated group of individuals. Their actions will shape the future of humanity.
Conclusion
Leopold's "Situational Awareness" series illuminates the rapid trajectory of AI development and the potential leap to superintelligent systems, along with their massive economic, security, and geopolitical implications. The analysis urges immediate awareness and preparedness for these transformative changes.
The 2030s will be fundamentally transformative. By decade's end, the world will be vastly different, with a new world order shaped by the advent of superintelligence. As AI practitioners and citizens, we must grapple with these realities now - the window for preparation is narrowing rapidly.
The question isn't whether these changes are coming, but whether we'll be ready when they arrive.