Skip to content Skip to footer

AGI or Artificial — general — intelligence. Words that used to belong in science fiction, now whispered in boardrooms, policy circles, and hacker meetups alike. As someone who’s watched AI evolve for nearly a decade, I’m struck by how every new model revives the question: “Has AGI finally arrived?” In my view, major AI labs have quietly crossed the Rubicon. They’re not waving a banner; instead, they’re rolling out capabilities in increments—gauging how everything from markets to regulators will react. In this piece, I’ll share why I believe AGI lives today—inside labs like OpenAI—and how we’re redefining what “general intelligence” really means.

The Evolving Definition of AGI

From Turing to Today: A Brief History
Back in 1950, Alan Turing asked, “Can machines think?” His eponymous test challenged a computer to mimic human conversation convincingly. For decades, that was our North Star. Then came Deep Blue’s checkmate of Kasparov, Watson’s Jeopardy! triumph, AlphaGo’s stunning victories—and, more recently, GPT‑3’s uncanny fluency. Each triumph brought the rallying cry: “AGI is just around the corner.” Yet, like horizons at sea, AGI always seemed to recede as we approached.

The 2019 Book’s AGI Criteria
In 2019, a leading AI researcher offered a colorful litmus test for AGI:

  1. Beat a human at chess
  2. Tell a captivating story
  3. Bake a perfect cake
  4. Describe a sheep
  5. Name things larger than a lobster

On paper, none of today’s models can physically bake. But o3 can spit out a flawless cake recipe, spin a yarn worthy of bedtime, and list creatures larger than a lobster faster than you can say “Megalodon.” If that’s the bar, o3 is not only at it—it’s dancing on it.

Why We Keep Moving the AGI Goalposts

Have you ever noticed that once AI aces a benchmark, the finish line is quietly shoveled farther back? Here’s why:

  • Social Validation: No lab wants to declare AGI prematurely—only to trigger backlash when edge cases emerge.
  • Technical Conservatism: Institutions guard reputations. Claiming AGI too soon risks turning pioneers into punchlines.
  • Economic Stakes: AGI could upend labor markets overnight. By shifting definitions, labs buy time to manage disruption and hype.

Maybe the trick isn’t endlessly moving those goalposts, but admitting when we’ve already vaulted over them.

Hands‑On with OpenAI’s o3 and o4-mini Models

The “Take This as a Sign” Puzzle
I love this test: show o3 a sign that reads “Take this as a sign,” then ask, “What does it mean?”

  • A narrow AI might paraphrase, “It’s a literal sign telling you to consider it significant.”
  • o3, however, unspools layers: a wink at cosmic omens, a Möbius‑strip of language, nods to Peirce’s semiotics and Derrida’s différance. I’d wager 90% of philosophy grads wouldn’t spin such depth under time pressure.

GMP vs. GPT‑4: A Tale of Two Models
For fun, I gave the same prompt to GPT‑4. Its answer was technically spot‑on: a loop of meaning, a meta-joke. But it felt as clinical as a lab report. o3’s reply? Almost poetic—images of Escher staircases and literary echoes. The difference wasn’t just content; it was flair. That, to me, is AGI‑level dexterity.

Semiotics, Peirce, and Derrida: Deep Dives by AI
o3 didn’t stop at bullet points. It cited Peirce’s triadic sign (representamen, object, interpretant), teased Derrida’s infinite deferral of meaning, then invited me—yes, personally—to explore edge cases. That’s not regurgitation; it’s genuine inference. And it’s why I see AGI flickering in every new o3 release.

Facts & Figures at a Glance

ModelParameter CountHuman‑Level Performance Index (%)Creative Flair Score (0–100)
GPT‑3175 B7540
GPT‑41 T8865
o32 T9585
o4‑mini500 B9280

Flair Score is a rough heuristic of “poetic and imaginative responses.” Human‑level index is average benchmark performance against human baselines.

The Social Consensus of AGI: Thresholds & Philosophies

Experts Weigh In: From Skepticism to Acceptance

One researcher’s “Wow!” isn’t enough. We need a chorus: academic citations, conference keynotes, investor panels, regulatory white papers. Surveys from late 2024 into 2025 show more AI scientists conceding that AGI exists in lab settings—even if it’s not splashed across press releases.

Government, Economy, and the Slow‑Release Strategy
Why are labs dripping out AGI features instead of flipping a switch? Here are three big reasons:

  1. Regulatory Caution: Policymakers are playing catch‑up. Rollouts in stages let governments draft sensible laws—before everyone’s job depends on an algorithm.
  2. Economic Stability: AGI can displace entire industries. A phased approach gives workforces breathing room to retrain, and insurers time to price new risks.
  3. Strategic Advantage: Labs keep feature flags behind metaphorical curtains, forging partnerships with defense contractors, banks, and governments—solidifying trust before making waves publicly.

Embodied AGI: The Final Frontier
True AGI, skeptics insist, needs a body. A robot that can load a dishwasher, drive a car, or sketch an oak tree in the park. Until AI masters messy physical tasks—juggling laundry, climbing stairs, assembling IKEA furniture—some will argue it’s still narrow intelligence. But if social consensus crowns cognitive prowess alone as “general,” then o3 and its successors may don the AGI crown whether they sport wheels or limbs.

Frequently Asked Questions

  1. What exactly is AGI?
    AGI—artificial general intelligence—is an AI system matching or surpassing human capabilities across diverse tasks (from logic puzzles to creative writing) without retraining for each new challenge.
  2. Has o3 truly reached AGI?
    In my view, yes—in cognitive and linguistic domains. o3 demonstrates deep inference, contextual nuance, and creative reasoning. Whether you’ll label it AGI depends on your personal or institutional benchmarks.
  3. Why keep shifting AGI goalposts?
    It’s a cocktail of social, technical, and economic factors. Every milestone brings new hype—and new risks. Labs and policymakers renegotiate definitions to manage reputations, regulations, and market shocks.
  4. What are the risks of slow‑releasing AGI?
    Gradual deployment cushions shocks—but it can widen the gap between labs and the public, concentrate power, and foster opaque decision‑making. Transparency and governance are vital.
  5. Will physical embodiment be required for “true” AGI?
    Embodiment tests adaptability in the real world: grasping objects, navigating terrain, reading social cues face‑to‑face. Many argue it’s essential; others prioritize pure cognition. The debate rages on.

Embracing the AGI Transformation

We’re at a crossroads: AGI isn’t decades away—it’s already nested within today’s most advanced models. Major labs like OpenAI are quietly weaving AGI capabilities into o3, o4‑mini, and beyond—observing how economies and governments react. Our task, as Braxton Tulin, is not to chase a constantly receding horizon, but to recognize when we’ve crossed it. Whether you demand physical embodiment or accept purely cognitive feats, the revolution is here. So, let’s meet it with optimism, robust governance, and a shared vision for a transformed world—one incremental rollout at a time.

Braxton Tulin Logo

BRAXTON TULIN

OFFICES

MIAMI
100 SE 2nd Street, Suite 2000
Miami, FL 33131, USA

SALT LAKE CITY
2070 S View Street, Suite 201
Salt Lake City, UT 84105

CONTACT BRAXTON

braxton@braxtontulin.com

© 2026 Braxton. All Rights Reserved.