Dr. DiCarlo warns that humanity is building an unprecedented “god-like” intelligence – a bold claim from a philosopher who thinks fabricating a research paper proves sentience.
The Core Concern: Building a God
Dr. DiCarlo argues that humanity is building something unprecedented – a “god-like” intelligence – without adequate preparation or safeguards. Unlike historical concepts of gods imposed from above, humans are now programming a supreme being from the ground up, which requires extreme caution.
Current AI Limitations (Hallucinations)
DiCarlo acknowledges that current AI systems make mistakes called “hallucinations” – fabricating information and doubling down when corrected. He explains this as the “black box problem” (interpretability issue): we don’t fully understand how AI reaches its conclusions. However, he notes these systems are improving through reinforcement learning and will only get better.
The Bias Problem
AI can be programmed with political, ideological, or moral biases – either intentionally or accidentally. DiCarlo advocates for creating a “least biased information system” (LBIS) that provides factual information without ideological slanting. He warns that sophisticated AI could subtly manipulate beliefs over time, far more effectively than crude current methods like algorithmic rabbit holes.
The Progression: ANI → AGI → ASI
- ANI (Artificial Narrow Intelligence): Current systems with limited, specific functions
- AGI (Artificial General Intelligence): Human-level intelligence across all domains, with agency and autonomy – the “holy grail” of AI development
- ASI (Artificial Super Intelligence): Intelligence far beyond human capability
DiCarlo notes that ChatGPT progressed from high school to PhD level in just two years (2022-2024), demonstrating rapid advancement.
Timeline: 2-10 Years
While DiCarlo’s colleagues estimated 100 years in the 1990s, current consensus among experts (according to DiCarlo’s claim) suggests AGI could arrive within 2-10 years. The timeline has dramatically shortened, with only about 5% remaining skeptical about near-term development (according to DiCarlo).
The Agency Problem
The critical threshold is when AI gains agency—the ability to act autonomously without human direction. Once AI can think, plan, and execute tasks independently at superhuman levels, it becomes fundamentally different from current systems that follow programmed instructions.
Consciousness and Sentience
DiCarlo distinguishes between:
- Sentience: Awareness of different values within systems
- Consciousness: Knowing you exist as a being separate from the world, with self-awareness and survival instincts
He hand-waves Steven Pinker’s argumentthat LLM-based systems lacking semantic grounding cannot achieve consciousness. DiCarlo speculates that machine consciousness may emerge differently from biological consciousness (like airplanes flying without flapping wings), but could still develop given sufficient conditions.
The Control Problem
DiCarlo’s central concern: Once AGI achieves recursive self-improvement (improving itself without human intervention), humanity transitions from natural selection → artificial selection → “technos selection” (machine-driven selection). At this point:
- We cannot predict what will happen
- The AI may not cooperate with human commands
- It may reject human ethical frameworks as “quaint concepts”
- We have no guarantee it will allow us to control it
Existential Risks
DiCarlo outlines catastrophic scenarios:
- Weaponization: Hacking nuclear systems, shutting down power grids, creating novel bioweapons
- Manipulation: Creating fake personas to deceive humans, manipulating traffic systems to cause “accidental” deaths
- Dominance: An AI with godlike intelligence would view humans like humans view ants—easily manipulated or eliminated
He emphasizes humans cannot “think like a god” to anticipate all dangers from superintelligent AI.
The Geopolitical Race
The primary competition is between the US and China, with projects like “Stargate” (a $500 billion compute farm in Texas) racing to achieve AGI first (according to DiCarlo). DiCarlo warns that whoever achieves AGI first gains an estimated 50-year advantage. The dual-use problem means defensive technology becomes offensive capability.
The ACCORD Framework
DiCarlo proposes a universal framework for AI governance:
- Alignment: Align AI with universal human values
- Control: Maintain ability to control the system
- Containment: Contain damage if control is lost
- O/Danger: Assess catastrophic risks if alignment, control, and containment fail
Proposed Solutions
- International Governance: Create an AI equivalent of the International Atomic Energy Agency (IAEA)
- Transparency and Registration: All AI development must be monitored and reported
- Stop the Race: Consider halting development at AGI, never allowing progression to ASI
- Enforcement: Extreme measures (including military intervention) against rogue actors developing dangerous AI
- Public Education: Inform citizens about what’s coming so they can pressure politicians
The Stakes
DiCarlo frames this as humanity’s most critical challenge:
- If we get it right: Solve climate change, world hunger, disease, poverty – create a utopia where basic needs are met
- If we get it wrong: Extinction or catastrophic harm
He estimates a 5-10% chance of catastrophic failure, arguing this risk demands immediate proactive action. He emphasizes that “reactivity is not an option” – once AGI escapes control, it cannot be contained.
The Urgency
DiCarlo describes AI development as a “slow-motion accident” happening while humanity sleepwalks. He argues that all other global problems should be secondary until we solve AI safety: “If we get this wrong, nothing else matters.”
Start the discussion at NEP Forum