Dr. Christopher DiCarlo on “Building A God”

Dr. DiCarlo warns that humanity is building an unprecedented “god-like” intelligence – a bold claim from a philosopher who thinks fabricating a research paper proves sentience.
The Core Concern: Building a God

Dr. DiCarlo argues that humanity is building something unprecedented – a “god-like” intelligence  – without adequate preparation or safeguards. Unlike historical concepts of gods imposed from above, humans are now programming a supreme being from the ground up, which requires extreme caution.

Current AI Limitations (Hallucinations)

DiCarlo acknowledges that current AI systems make mistakes called “hallucinations” – fabricating information and doubling down when corrected. He explains this as the “black box problem” (interpretability issue): we don’t fully understand how AI reaches its conclusions. However, he notes these systems are improving through reinforcement learning and will only get better.

The Bias Problem

AI can be programmed with political, ideological, or moral biases – either intentionally or accidentally. DiCarlo advocates for creating a “least biased information system” (LBIS) that provides factual information without ideological slanting. He warns that sophisticated AI could subtly manipulate beliefs over time, far more effectively than crude current methods like algorithmic rabbit holes.

The Progression: ANI → AGI → ASI
  • ANI (Artificial Narrow Intelligence): Current systems with limited, specific functions
  • AGI (Artificial General Intelligence): Human-level intelligence across all domains, with agency and autonomy – the “holy grail” of AI development
  • ASI (Artificial Super Intelligence): Intelligence far beyond human capability

DiCarlo notes that ChatGPT progressed from high school to PhD level in just two years (2022-2024), demonstrating rapid advancement.

Timeline: 2-10 Years

While DiCarlo’s colleagues estimated 100 years in the 1990s, current consensus among experts (according to DiCarlo’s claim) suggests AGI could arrive within 2-10 years. The timeline has dramatically shortened, with only about 5% remaining skeptical about near-term development (according to DiCarlo).

The Agency Problem

The critical threshold is when AI gains agency—the ability to act autonomously without human direction. Once AI can think, plan, and execute tasks independently at superhuman levels, it becomes fundamentally different from current systems that follow programmed instructions.

Consciousness and Sentience

DiCarlo distinguishes between:

  • Sentience: Awareness of different values within systems
  • Consciousness: Knowing you exist as a being separate from the world, with self-awareness and survival instincts

He hand-waves Steven Pinker’s argumentthat LLM-based systems lacking semantic grounding cannot achieve consciousness. DiCarlo speculates that machine consciousness may emerge differently from biological consciousness (like airplanes flying without flapping wings), but could still develop given sufficient conditions.

The Control Problem

DiCarlo’s central concern: Once AGI achieves recursive self-improvement (improving itself without human intervention), humanity transitions from natural selection → artificial selection → “technos selection” (machine-driven selection). At this point:

  • We cannot predict what will happen
  • The AI may not cooperate with human commands
  • It may reject human ethical frameworks as “quaint concepts”
  • We have no guarantee it will allow us to control it
Existential Risks

DiCarlo outlines catastrophic scenarios:

  • Weaponization: Hacking nuclear systems, shutting down power grids, creating novel bioweapons
  • Manipulation: Creating fake personas to deceive humans, manipulating traffic systems to cause “accidental” deaths
  • Dominance: An AI with godlike intelligence would view humans like humans view ants—easily manipulated or eliminated

He emphasizes humans cannot “think like a god” to anticipate all dangers from superintelligent AI.

The Geopolitical Race

The primary competition is between the US and China, with projects like “Stargate” (a $500 billion compute farm in Texas) racing to achieve AGI first (according to DiCarlo). DiCarlo warns that whoever achieves AGI first gains an estimated 50-year advantage. The dual-use problem means defensive technology becomes offensive capability.

The ACCORD Framework

DiCarlo proposes a universal framework for AI governance:

  • Alignment: Align AI with universal human values
  • Control: Maintain ability to control the system
  • Containment: Contain damage if control is lost
  • O/Danger: Assess catastrophic risks if alignment, control, and containment fail
Proposed Solutions
  1. International Governance: Create an AI equivalent of the International Atomic Energy Agency (IAEA)
  2. Transparency and Registration: All AI development must be monitored and reported
  3. Stop the Race: Consider halting development at AGI, never allowing progression to ASI
  4. Enforcement: Extreme measures (including military intervention) against rogue actors developing dangerous AI
  5. Public Education: Inform citizens about what’s coming so they can pressure politicians
The Stakes

DiCarlo frames this as humanity’s most critical challenge:

  • If we get it right: Solve climate change, world hunger, disease, poverty – create a utopia where basic needs are met
  • If we get it wrong: Extinction or catastrophic harm

He estimates a 5-10% chance of catastrophic failure, arguing this risk demands immediate proactive action. He emphasizes that “reactivity is not an option” – once AGI escapes control, it cannot be contained.

The Urgency

DiCarlo describes AI development as a “slow-motion accident” happening while humanity sleepwalks. He argues that all other global problems should be secondary until we solve AI safety: “If we get this wrong, nothing else matters.”

Authors

  • Lloyd Robertson

    Lloyd Hawkeye Robertson is an Adjunct Professor of Psychology at the University of Regina. His main professional interest has been on the evolution and structure of the self.   He has also published on the psychological impacts of Indian residential schools, the use of a community development process to combat youth suicide, the construction of the (North American) aboriginal self, the concept of free will in psychotherapy, and male stigma as it affects men’s identity.  He is currently President of the New Enlightenment Project: A Canadian Humanist Initiative.

  • Scott Douglas Jacobsen is the Founder of In-Sight Publishing and Editor-in-Chief of "In-Sight: Independent Interview-Based Journal" (ISSN 2369–6885). He is a Freelance, Independent Journalist with the Canadian Association of Journalists in Good Standing. Email: Scott.Douglas.Jacobsen@Gmail.Com.

Start the discussion at NEP Forum