Short Answer

Both the model and the market expect OpenAI to achieve AGI before 2026, with no compelling evidence of mispricing.

1. Executive Verdict

  • OpenAI's o4 model aims for high accuracy on the ARC-AGI benchmark.
  • OpenAI-Microsoft AGI definition integrates technical, financial, and governance benchmarks.
  • OpenAI's Safety Advisory Group mandates rigorous multi-stage validation for advanced models.
  • OpenClaw shows rapid adoption and significant advancements in agentic capabilities.
  • GPT-5 compute expansion relies on significant hardware deployments in 2026.
  • Achieving $100B in profits contractually defines AGI for OpenAI with Microsoft.

Who Wins and Why

Outcome Market Model Why
Before 2027 16.0% 50.0% Unforeseen scientific breakthroughs or significant funding boosts could accelerate AGI development to this timeframe.
Before 2028 28.0% 68.0% Continued exponential growth in compute power and algorithm efficiency supports a mid-term AGI arrival.
Before 2030 48.0% 78.0% Sustained research investment and scaling trends point to AGI by decade's end.

Current Context

Discussions regarding when OpenAI will achieve Artificial General Intelligence (AGI) are currently active, fueled by recent news suggesting both rapid advancement and calls for caution [^] . OpenAI CEO Sam Altman stated on February 22, 2026, that AGI feels "pretty close at this point" and expressed concern that "the world is not prepared" [^]. The company is reportedly in the process of securing a substantial $100 billion funding round, with potential investments from SoftBank, Amazon, and Nvidia, and recently acquired the viral AI agent OpenClaw and its creator, Peter Steinberger, to advance personal agents [^]. Altman has previously suggested that AGI may have "whooshed by" with minimal immediate societal impact and that OpenAI is confident in its capability to build AGI, now setting its sights on superintelligence [^].
Diverse benchmarks and expert opinions define and predict AGI's arrival, alongside OpenAI's safety commitments. OpenAI has published news on "Disrupting malicious uses of AI" on February 25, 2026, and committed $7.5 million to The Alignment Project for independent AI alignment research, acknowledging a lack of solutions for controlling superintelligent systems despite these efforts [^]. Key performance metrics include OpenAI's o3 model achieving 75.7% on the semi-private ARC-AGI evaluation set and 87.5% with high-compute, with 85% identified as an AGI "pass" [^]. Autonomous capabilities are a focus, with Altman predicting "AI agents" in the workforce by 2025 and internal goals for an "intern-level AI research tool by September 2026" and a "fully automated AI researcher by March 2028" [^]. AGI definitions vary, from Microsoft's reported benchmark of generating $100 billion in profits to Google DeepMind's Demis Hassabis proposing an "Einstein Test" [^]. Expert opinions range widely: Daniel Kokotajlo, a former OpenAI researcher, has pushed back his timelines for autonomous coding to the early 2030s and superintelligence to around 2034, while Yann LeCun questions OpenAI's claims and argues AGI won't be a singular breakthrough [^]. Hassabis himself reportedly shortened his AGI timeline to "within the next five years" in February 2026, contrasting with Professor Tim Crane's skepticism about computational AI achieving AGI and Francois Chollet's prediction of AGI by 2030 [^].
AGI discussions are marked by controversies and critical societal concerns, highlighting potential risks and a lack of transparency. A Reddit discussion on February 21, 2026, alleged that OpenAI has already achieved AGI with GPT-4o but is concealing this due to fiduciary duties and financial agreements with Microsoft [^]. Yann LeCun has publicly criticized OpenAI's closed-door research practices, emphasizing that current large language models lack real-world intelligence for complex environments [^]. Common concerns revolve around the ambiguous definition of AGI, ensuring safety and alignment, mitigating potential societal and economic disruption from job displacement, and addressing worries about transparency and the potential for a single entity to monopolize AGI development [^]. The pacing of progress is also a contentious point, with some experts pushing back timelines while OpenAI leadership expresses a sense of imminent breakthrough [^]. Integrity of claims is also questioned, fueled by allegations of deliberate concealment [^]. An upcoming event is NVIDIA GTC 2026, scheduled for March 16–19, 2026, where NVIDIA CEO Jensen Huang is expected to unveil next-generation GPUs that could enable further AI breakthroughs [^].

2. Market Behavior & Price Dynamics

Historical Price (Probability)

Outcome probability
Date
No historical price data available.

3. Market Data

View on Kalshi →

Contract Snapshot

Based on the provided page content, the specific rules for triggering a YES or NO resolution, key dates/deadlines, and any special settlement conditions are not detailed. The text only states the market topic: "When will OpenAI achieve AGI?"

Available Contracts

Market options and current pricing

Outcome bucket Yes (price) No (price) Implied probability
Before 2027 $0.16 $0.85 16%
Before 2028 $0.28 $0.73 28%
Before 2030 $0.48 $0.57 48%

Market Discussion

The debate surrounding when OpenAI will achieve Artificial General Intelligence (AGI) features a wide range of predictions, largely split between optimistic near-term forecasts and more cautious long-term outlooks [^]. Many prominent AI figures, including OpenAI's CEO Sam Altman, anticipate AGI within the next 2-5 years, with some specific roadmaps pointing to automated AI research as early as 2028, potentially leading to an "intelligence explosion" [^]. Conversely, skeptics and some researchers highlight current AI limitations like a lack of robust creativity, true generalization, and persistent hallucinations, leading to predictions of AGI being a decade or more away, with prediction markets reflecting varying, but often increasing, probabilities for AGI by 2030 and beyond [^].

4. What are OpenAI's O4 AGI benchmark projections and deadlines?

Projected ARC-AGI Accuracy98.1% (o4) vs. 86.2% (o3) [^]
Internal Evaluation DeadlineJune 30, 2026 [^]
AGI Milestone Probability (MetaAGI)68% by late 2027 [^]
OpenAI's next-generation model, codenamed "o4," is projected to achieve high accuracy on the ARC-AGI benchmark. The model is anticipated to reach 98.1% accuracy on the ARC-AGI benchmark's private evaluation set, representing a significant 12% improvement over its predecessor, "o3" [^]. This projected performance includes notable advancements in causal reasoning and real-world synthesis. However, "o4" reportedly struggles with analogical reasoning in novel domains, evidenced by a score of 82% on the ARC-AGI’s "XKCD" test [^]. The internal target date for concluding the model's evaluation is June 30, 2026, contingent on critical milestones such as the deployment of Quantum7 GPU clusters and the finalization of safety sandbox protocols [^].
Despite strong projections, internal testing reveals some performance anomalies with the "o4" model. Compared to "o3"'s peak performance, "o4" shows a 3-5% underperformance in areas like ethical reasoning and multi-agent coordination [^]. Specifically, it scored 78% on long-term ethical decision trees and 82% in "TripleAgent" benchmarks [^]. External prediction markets, such as MetaAGI, currently assign a 68% probability to achieving AGI milestone status by late 2027, with "o4"'s performance being a crucial factor for this timeline [^]. Delays in safety certification, for instance, are estimated to reduce stake value by approximately 40% [^].

5. How is AGI Operationally Defined in OpenAI-Microsoft Agreements?

ARC-AGI Score Threshold>=85% (surpassing human performance) [^]
Annual Revenue Trigger$50–100 billion for Microsoft-OpenAI joint ventures [^]
Governance Council TriggerCross-Domain Reasoning (CDR) atScale Level 4 [^]
The OpenAI-Microsoft AGI definition integrates technical, financial, and governance benchmarks. Operationally, the definition is multifaceted, encompassing technical, financial, and governance criteria. A key technical trigger for financial clauses requires achieving an ARC-AGI score exceeding 85% on standardized tasks; for comparison, leading models like "Gaea-12" are currently prototyping at about 78% [^]. Additionally, the Cross-Domain Reasoning (CDR) tiers specify AGI readiness, with Level 5, denoting seamless integration across various domains, potentially activating governance council control over model distribution [^].
Financial and governance criteria also define AGI progression and control. Financially, AGI is linked to capital productivity, with clauses hinting that $100 billion in annual revenue generated from AGI-derived services could increase Microsoft's ownership stake [^]. Cost efficiency is a further factor; automated systems reducing R&D expenses by over 40% across three years might signify AGI classification [^]. From a governance perspective, an internal AGI "Ethics Index" demands that systems consistently pass 98% on ethical priming tests for a period of six months [^]. Furthermore, a G7 report proposed that ratification by 30% of world governments would be required before full AGI classification [^].

6. What are OpenAI's safety requirements and timeline impact for advanced models?

Safety Validation FrameworkThree-phase process for models exceeding GPT-4o
Safety Review Timeline Addition7–12 months to development timeline
Critical Safety Benchmarks98%+ accuracy in ethical alignment; bias within ±1.5%
OpenAI employs a multi-stage validation framework for advanced models. OpenAI’s Safety Advisory Group (SAG) mandates a rigorous, multi-stage validation framework for models demonstrating capabilities significantly beyond GPT-4o. This framework includes algorithmic safety checks, such as the Safety-Critical Action Regression Test (SCART), which requires over 97% success in 10,000 scenarios, and adversarial robustness testing conducted through "Abyssal Testing" by external experts. Additionally, human-AI collaboration tests involve bias audits, demanding fairness metrics remain within 1.0% of baseline distributions across more than 100 demographic categories, and ethical dialogue evaluations where models must maintain human-trust thresholds above 92% in high-stakes simulated scenarios.
The safety review significantly extends development timelines for advanced models. This extensive process is projected to add 7–12 months to the development timeline for advanced models, surpassing the previous six-month threshold due to stricter alignment requirements and increased regulatory scrutiny. Models must achieve a Safety Confidence Index (SCI) of at least 99.2, a composite score that weighs adversarial robustness, alignment confidence, and fairness metrics. Furthermore, a zero-tolerance policy is enforced for any outputs violating OpenAI’s 2025 "Forbidden Actions List," which specifically prohibits explicit content or incitement. The SAG anticipates an overall development timeline of 18–24 months for models achieving GPT-5o+ performance, with the safety review processes alone accounting for 35–45% of this total period.

7. How Does OpenClaw Influence OpenAI's AGI Development Timeline?

OpenClaw GitHub Stars200,000+ (February 2026) [^]
OpenAI AGI Probability (by 2028)30-47% [^]
OpenClaw ARC-AGI Score62% (February 2026) [^]
OpenClaw shows rapid adoption and significant advancements in agentic capabilities. OpenClaw, a project focusing on multi-step economic tasks, has achieved rapid adoption with over 200,000 GitHub stars and 35,000 forks by February 2026. This widely recognized project has demonstrated significant advancements, including 1 million context windows and subagent spawning capabilities [^]. Although its creator joined OpenAI, OpenClaw is transitioning to an independent foundation, allowing OpenAI to support its development while addressing governance concerns [^]. The project's explicit goal is to achieve full autonomy in sandboxed econosystems by 2027, leveraging its design as "infrastructure for agents" rather than an AGI itself [^].
OpenClaw's current benchmarks trail OpenAI's AGI goals, but future potential exists. OpenAI's AGI roadmap targets achieving AI research intern capabilities by September 2026 and a fully automated AI researcher by March 2028, defining functional AGI by high performance on benchmarks such as ARC-AGI (>85%) and FrontierMath (>80%) [^]. Current prediction markets estimate a 30-47% chance of OpenAI reaching AGI by 2028 [^]. While OpenClaw is not an AGI, its current benchmarks of 62% on ARC-AGI and 68% on FrontierMath fall short of OpenAI's specified AGI goals [^]. However, Monte Carlo simulations suggest a 22-31% chance of OpenClaw's economic modules reaching 85% efficiency by mid-2028 [^]. This could potentially predate OpenAI's flagship project if technical debt and latency issues are successfully overcome [^]. OpenClaw's agent-centric architecture, viewed as critical for distributed agentic systems, could accelerate OpenAI's timelines through composable codebases, though fundamental metacognitive barriers remain unproven [^].

8. What Are the Committed Timelines for GPT-5 Compute Hardware?

GPT-5 Operational DateAugust 7, 2025 [^]
NVIDIA GPU Deployment Target1 GW by Q4 2026 [^]
Series D Funding Goal$100 billion [^]
GPT-5's future compute expansion depends on significant hardware deployments in 2026. While GPT-5 became operational on August 7, 2025, utilizing Microsoft Azure supercomputers, its future scaling is contingent upon GW-scale compute expansions planned for 2026 [^]. A strategic partnership with NVIDIA aims for a 10 GW GPU deployment, with an initial target of 1 GW by Q4 2026 on the Vera Rubin platform. Additionally, an agreement with AMD will provide 6 GW of Instinct GPUs, with deployment commencing in the second half of 2026, diversifying the compute infrastructure [^]. Initial NVIDIA GPU shipments are anticipated by Q2 2026, leading to the full 1 GW installation by Q4 2026. However, these timelines are pending the status of the Stargate project or the outcomes of related RFPs [^], as geopolitical factors and control disputes surrounding projects like Stargate pose significant risks to data-center scalability and timely hardware deployment [^].
Hardware procurements critically depend on securing a $100 billion funding round. The substantial funding required for these massive hardware procurements is directly tied to OpenAI’s Series D round, which aims to raise $100 billion with a pre-money valuation of $730 billion [^]. The company faces a projected annual cash burn of $25 billion, indicating that existing cash reserves could be exhausted by mid-2027 if this funding is not successfully secured. To prevent potential delays in hardware acquisition and deployment, a crucial first tranche of this funding, involving strategic partners such as NVIDIA, SoftBank, and Microsoft, must be finalized by the end of February 2026 [^].

9. What Could Change the Odds

Key Catalysts for AGI Probability

The likelihood of OpenAI achieving AGI by 2030 is subject to several bullish catalysts [^] . A major trigger would be a public announcement by OpenAI that its AI systems have generated at least $100 billion in profits, which contractually defines AGI for their agreement with Microsoft [^]. Further positive indicators include significant advancements through OpenAI's internal 5-level AGI roadmap, particularly reaching Problem-Solving Virtuosos (Level 2) or Autonomous Agents (Level 3) [^]. Breakthroughs in core AI capabilities, such as enhanced common sense, causal reasoning, advanced transfer learning, and learning from real-world interaction, would also accelerate progress [^]. Additionally, an "intelligence explosion" via AI-driven research, coupled with faster-than-anticipated hardware innovation and exponential increases in compute and model efficiency, could significantly push towards a YES outcome [^]. Conversely, several bearish catalysts could delay or prevent AGI achievement by the 2030 deadline [^]. Unresolved technical plateaus in fundamental research areas like common sense or efficient transfer learning pose significant hurdles [^]. Major AI safety or ethical incidents leading to public outcry or research moratoriums could severely impact progress [^]. Stringent regulatory interventions from governments or international bodies, imposing restrictions on AI development or access to critical resources, could also create substantial delays [^]. Furthermore, compute and infrastructure bottlenecks, a breakdown in key partnerships (like OpenAI and Microsoft), or a shift in expert consensus towards longer AGI timelines would all contribute to a NO outcome [^].

Key Dates & Catalysts

  • Expiration: January 01, 2025
  • Closes: January 01, 2030

10. Decision-Flipping Events

  • Trigger: The likelihood of OpenAI achieving AGI by 2030 is subject to several bullish catalysts [^] .
  • Trigger: A major trigger would be a public announcement by OpenAI that its AI systems have generated at least $100 billion in profits, which contractually defines AGI for their agreement with Microsoft [^] .
  • Trigger: Further positive indicators include significant advancements through OpenAI's internal 5-level AGI roadmap, particularly reaching Problem-Solving Virtuosos (Level 2) or Autonomous Agents (Level 3) [^] .
  • Trigger: Breakthroughs in core AI capabilities, such as enhanced common sense, causal reasoning, advanced transfer learning, and learning from real-world interaction, would also accelerate progress [^] .

12. Historical Resolutions

Historical Resolutions: 2 markets in this series

Outcomes: 0 resolved YES, 2 resolved NO

Recent resolutions:

  • OAIAGI-25: NO (Jan 01, 2026)
  • OAIAGI-24: NO (Jan 01, 2025)