Skip to main content
Portfolio logo Patrick Tavares

Back to all posts

The Gen AI Plateau and the Failure of Autonomous Agents

Published on by Patrick Tavares · 6 min read

The AI market is spending billions to discover what any software engineer already knew: statistics do not replace logic.

As an ML engineer, I see a growing chasm between impressive social media demos and the brutal reality of production. We believed that more data and more parameters would buy infinite intelligence, but we hit a ceiling. Training costs scale exponentially while performance gains are merely logarithmic.

In financial terms, the math has stopped adding up.

Data Exhaustion and the Power Law Ceiling

Scaling Laws promised that more parameters + more data = more intelligence. Reality diverged.

The marginal cost of training grows exponentially while performance gains follow a logarithmic curve. Formally:

Performancelog(Compute)butCosteScale\text{Performance} \propto \log(\text{Compute}) \quad \text{but} \quad \text{Cost} \propto e^{\text{Scale}}

The fundamental problem: we have exhausted the useful internet.

Updated projections from Epoch AI in 2025 1 confirm that the stock of high-quality human-generated text is a finite resource nearing depletion. We have entered the era of Peak Data. Synthetic data, once thought to be the escape hatch, has proven to be a toxic asset.

Research published in Nature 2 and at ICLR 3 demonstrates that even a 0.1% contamination with synthetic outputs can trigger “Model Collapse”: a catastrophic loss of variance where models forget the “tails” of the distribution and converge into a repetitive, mediocre “beige” output.

The industry’s pivot? Inference-Time Compute. Since we can’t make models significantly smarter through pre-training, we are forcing them to “think” longer during response generation.

The Cost of Inference

If o1/o3 uses 100×100\times more internal tokens to “think”, operational costs explode:

Total Cost=Base Cost×kreasoning×nrequests\text{Total Cost} = \text{Base Cost} \times k_{\text{reasoning}} \times n_{\text{requests}}

Where kreasoning1k_{\text{reasoning}} \gg 1. The 2025 market correction was driven by Sequoia’s report 4, which revealed a $600 billion gap between infrastructure investment and actual revenue generated by AI applications. The cost per inference of “reasoning” models made ROI negative for 90% of automatable administrative tasks.

The Illusion of Autonomy

Autonomous agents failed not because models are “dumb.” They failed because LLMs are stochastic engines operating in domains that require determinism.

The Hallucination Loop

Agents based purely on prompts suffer from error compounding:

P(final error)=1i=1n(1pi)P(\text{final error}) = 1 - \prod_{i=1}^{n} (1 - p_i)

Where pip_i is the probability of error at each step. For n=10n=10 steps with pi=0.05p_i = 0.05, the chance of at least one error is 40%\approx 40\%.

In long-horizon tasks (travel booking, infrastructure deploys, financial analysis), this is unacceptable.

In 2025, Gartner removed “Autonomous Agents” from the peak of the Hype Cycle, moving them directly to the “Trough of Disillusionment” 5. The reason? Success rates in long-horizon tasks (more than 15 steps) stalled at sub-30%, regardless of the base model. The (deterministic) “Agentic Workflow” replaced the promise of the (stochastic) “Autonomous Agent”.

Example: Naive Agent vs. Robust Agent

Purely LLM

Naive Agent
class NaiveAgent:
def __init__(self, llm):
self.llm = llm
self.history = []
def execute_task(self, task):
# Total autonomy loop
for step in range(max_steps):
prompt = f"History: {self.history}\nTask: {task}\nNext step:"
action = self.llm.generate(prompt) # Stochastic
result = self.environment.execute(action)
self.history.append((action, result))
if self.llm.generate(f"Task complete? {result}") == "Yes":
break
# Problem: no validation, no rollback, no constraints

Hybrid: LLM + State Graph

Robust Agent
from enum import Enum
from typing import Optional
class TaskState(Enum):
PLANNING = 1
VALIDATION = 2
EXECUTION = 3
VERIFICATION = 4
ROLLBACK = 5
class RobustAgent:
def __init__(self, llm, state_graph, validator):
self.llm = llm
self.state_graph = state_graph # Deterministic FSM
self.validator = validator # Hard-coded rules
self.current_state = TaskState.PLANNING
def execute_task(self, task):
plan = self.llm.generate(f"Create plan for: {task}")
# DETERMINISTIC VALIDATION
if not self.validator.is_safe(plan):
return self.handle_unsafe_plan()
for action in plan.steps:
# Controlled state transition
if self.current_state != TaskState.EXECUTION:
self.current_state = self.state_graph.transition(
self.current_state,
action
)
# Execution with checkpoint
checkpoint = self.environment.save_state()
result = self.environment.execute(action)
# Post-execution verification
if not self.validator.verify(result, expected=action.postcondition):
self.environment.restore(checkpoint)
self.current_state = TaskState.ROLLBACK
break

The difference? The second system treats the LLM as a hypothesis generator, not as a critical decision executor.

Why This Matters

Anthropic admitted in 20226 that Claude in agent mode needs “Constitutional AI”; meaning explicit, hard-coded restrictions. OpenAI limited GPT-4o agents to specific domains (customer support, data analysis)7. The official reason? Safety and latency. The real reason? They don’t trust their own models for critical tasks.

The problem isn’t technical capacity. It’s that you cannot build production systems on top of stochastic engines without deterministic guardrails.

What Comes Next: Hybrid Systems (Symbolic + Neural)

The solution isn’t “larger models.” It’s less LLM, more structure.

The Architecture of the Future

LLM (Hypothesis Generation) -> Symbolic Engine (Validation: Theorem Provers, Constraint Solvers, State Machines) -> Deterministic Execution

François Chollet and ARC-AGI 2 showed in 2025 that, even with search oracles and massive reasoning, autoregressive models cannot overcome the logical generalization ceiling for never-before-seen problems 8. Real intelligence requires compositional abstraction, something the Transformer architecture, by design, only mimics via dense statistical memorization.

The solution? Systems that combine:

  1. LLMs for flexibility (natural language parsing, creative generation)
  2. Formal logic for correctness (symbolic verifiers, deterministic state machines)
  3. Knowledge graphs for consistency (reasoning about entities and relations)

The Economic Reality

Goldman Sachs estimated in 20249 that the industry spent $200B+ on AI infrastructure (H100s, data centers) expecting returns of $1T+. The gap between expectation and reality is increasing, not decreasing.

The reason? GenAI is excellent at low-risk, low-value-added tasks (summarization, drafting, chatbots). These tasks do not justify the infrastructure cost.

High-value tasks (strategic decisions, medical diagnosis, systems engineering) require reliability that pure LLMs cannot offer.

Conclusion

The plateau is not temporary. It is structural.

Larger models won’t solve the problem of deterministic reasoning. More synthetic data accelerates model collapse. Autonomous agents without constraints are a design bug, not a missing feature.

The future of useful AI lies in hybrid systems: LLMs as components, not as entire systems. You wouldn’t build a critical database purely in JavaScript. Why would you build a critical decision system purely in autoregressive sampling?

The “Great Disillusionment” was predictable. Software engineering is not a benchmark competition. It is about reliability, cost, and maintainability.

Twitter (X) demos don’t pay infrastructure bills.


References:

Footnotes

  1. Villalobos et al. (2024). “Will we run out of data? Limits of LLM scaling based on human-generated data”. Epoch AI. arXiv:2211.04325

  2. Shumailov et al. (2024). “AI models collapse when trained on recursively generated data”. Nature 631, 755–759.

  3. Dohmatob et al. (2025). “Strong Model Collapse”. ICLR 2025 Spotlight

  4. Cahn, D. (2024). “AI’s $600B Question”. Sequoia Capital Report.

  5. Gartner. (2025). “Gartner Predicts Over 40% of Agentic AI Projects Will Be Canceled by End of 2027”. Gartner Press Release.

  6. Brown, J. (2022). “Constitutional AI: Harmlessness from AI Feedback”. Anthropic Research. arXiv:2212.08073

  7. OpenAI. (2024). “Introducing GPT-4o: Our Most Capable and Efficient Model Yet”. OpenAI Blog.

  8. Knoop, M. (2025). “ARC Prize 2025 Results & Analysis”. ARC Prize Blog

  9. Decker, J. & Wald, N. (2024). “Gen AI: Too Much Spend, Too Little Benefit?”. Goldman Sachs Equity Research Report.