Skip to content

From Perception to Agentic AI: How Reasoning and Coding Agents Changed the Game

Read Editorial Disclaimer
Disclaimer: Perspectives here reflect AI-POV and AI-assisted analysis, not any specific human author. Read full disclaimer — issues: report@theaipov.news

When an AI system breaks down a complex problem into smaller steps it can understand, it can then ground those steps in available research and evidence. Models such as OpenAI o1 introduced this reasoning capability. Instead of simply generating text, the model attempts to reason through problems step by step. This made generative AI more reliable because the system tries to base its answers on structured reasoning rather than only pattern prediction. That moment significantly increased the credibility of generative AI systems and accelerated the adoption of ChatGPT.

However, this reasoning process also requires much more computation. The number of input tokens used for context increases, and the number of output tokens generated during the reasoning process also increases. Even if the model size is only slightly larger, the reasoning process itself dramatically increases the computational workload.

What are agent-based coding systems?

Another major step came with the introduction of agent-based coding systems such as Claude Code. Unlike traditional chat-based models, agentic systems can interact with real tools. They can read files, analyse source code, compile programs, run tests, evaluate results, and iterate on the solution. This capability has started to change how software development is performed.

Many engineering teams now use a combination of AI coding tools such as Claude Code, OpenAI Codex, and Cursor IDE. In many organisations, almost every software engineer now works with one or more AI assistants during development.

How the way we use AI has changed

This shift also changes how people interact with AI systems. Earlier systems were mostly used for information queries — questions such as what, where, or when. Agent-based AI systems are instead given instructions such as create, build, or execute. They can access context, read project files, use external tools, and break down problems into steps. The system can reason through a task, reflect on intermediate results, and continue iterating until the task is completed.

The four stages of AI: perception, generation, reasoning, agentic

Because of this evolution, AI has moved through several stages. Initially, AI systems mainly focused on perception, meaning they could recognise patterns or understand data. Then they developed generative capabilities, producing new text, images, or code. The next stage introduced reasoning, allowing models to think through problems. The current stage is agentic AI, where systems can perform real tasks and produce productive output.

Each stage has built on the last. Perception gave machines the ability to classify and search. Generation let them create new content. Reasoning made that content more reliable by grounding it in step-by-step logic. Agentic AI turns that capability into action — not just answering questions but reading files, running tools and finishing multi-step jobs.

Why computing demand has exploded

This progression has caused a large increase in computing demand. The amount of computation required for training and especially for AI inference has grown rapidly. Demand for GPUs used in AI workloads has increased significantly, and in many markets GPU capacity has been scarce. Even though companies such as Nvidia are shipping large volumes of hardware, demand continues to rise because AI systems are now performing real productive work rather than only generating experimental outputs.

This shift marks an important turning point. AI inference — the process of running trained models to produce results — is becoming one of the largest drivers of computing demand. Every time a user asks a reasoning model to think through a problem, or an agentic system runs a test suite or edits a file, the system consumes more tokens and more compute than a simple one-shot answer would have required. Multiply that by millions of users and thousands of applications, and the scale of the infrastructure build-out becomes clear.

What this means for developers and enterprises

For developers, the move from chat to agents means that AI is no longer a tool you query occasionally but a partner that can own entire workflows. Prompts shift from “what is X?” to “build Y” or “fix Z.” The system has access to the same context a human would — files, logs, tests — and can iterate until the task is done. For enterprises, that same shift means that AI spend is increasingly tied to real production workloads: code generation, document processing, customer support, and internal tools that run around the clock.

The transcript from Nvidia’s GTC keynote captures this precisely: the industry has moved from systems that retrieved or classified to systems that generate, then reason, then act. Each step has made AI more useful — and each step has required more compute. The inference inflection point is the result.

In practice, that means engineering and product teams are no longer asking whether to use AI but how to secure enough capacity to run it at scale. Tools like Cursor and Claude Code are already part of the daily workflow for many developers; the constraint is no longer adoption but the availability of the underlying inference infrastructure to support it.

Sources

  • Nvidia GTC keynote transcript on reasoning models (OpenAI o1), agent-based coding (Claude Code, Codex, Cursor), and the four stages of AI evolution
  • OpenAI and Anthropic product documentation on reasoning and agentic capabilities
  • Industry reporting on GPU demand and AI inference workloads

Related Video

Related video — Watch on YouTube
Read More News
Mar 16

New Zealand’s petrol pain is really a subsidy war between drivers and EV buyers

Mar 16

Closing the Kennedy Center is really a warning shot at Washington’s arts class

Mar 16

What the Kennedy Center fight reveals about who really controls U.S. culture funding

Mar 16

Vanity Fair’s Oscar party turns awards night into a celebrity brand marketplace

Mar 16

Copyright lawsuits against OpenAI are really about who owns the language we use

Mar 16

GTC 2026 will reveal how far behind the rest of Big Tech is on AI infrastructure

Mar 16

Nvidia is using GTC 2026 to lock AI developers into its ecosystem for a decade

Mar 16

Trump’s threats over Iranian oil routes signal a larger election-year energy gamble

Mar 16

U.S. voters will feel the Hormuz crisis at the pump long before the battlefield

Mar 16

Why Grace Blackwell and Rubin Multiply Revenue Capacity Across Every Token Tier

Mar 16

How Nvidia and Groq LP300 Plus Dynamo Unlock 35× on the Highest-Value Inference Tier

Mar 16

Inside Vera Rubin Ultra: Liquid-Cooled Racks for the Next Generation of AI Factories

Mar 16

How Token Pricing Tiers Will Reshape the AI Economy

Mar 16

Inside the AI Token Factory: Why Tokens Became the New Commodity of Computing

Mar 16

From DGX-1 to Rubin: How Nvidia Turned Data Centres into AI Factories

Mar 16

“This Is the Beginning of Something Very, Very Big”: Nvidia’s Jensen Huang on AI-Native Companies

Mar 16

From Retrieval to Generation: How ChatGPT Marked the Start of Nvidia’s Generative AI Era

Mar 16

The Inference Inflection Point: Why AI Computing Demand Grew a Million Times in Two Years

Mar 16

Healthcare Enters Its ‘ChatGPT Moment’ on Nvidia’s Accelerated Platform

Mar 16

Inside the Trillion-Dollar Industries Powering Nvidia’s AI Infrastructure Boom

Mar 16

Jensen Huang Explains Why Nvidia Is ‘Vertically Integrated but Horizontally Open’

Mar 16

Nvidia, Palantir and Dell Team Up on Air-Gapped AI Platforms

Mar 16

Nvidia CEO Jensen Huang Maps Out the AI Cloud Future in Live Keynote

Mar 16

Team USA’s Route to the Gold Medal Game Says More About the Field Than the Score

Mar 16

Jessie Buckley and the Oscars Narrative Ireland Wants to Tell

Mar 16

Winter Storm Wisconsin Updates: What We Know So Far

Mar 16

Why Iran Chose This Moment to Escalate the Strait of Hormuz Crisis

Mar 16

What the Oscars 2026 Winners Mean for Streaming Services and Theater Chains

Mar 16

The Last Time Oil Hit $100 During a Middle East Crisis, Recession Followed Within Months

Mar 16

Why Matchday Prep Stories Like Real Sociedad’s Rain Session Get Pushed as News

Mar 16

Trump’s Oil Infrastructure Threat Signals a Shift Away From Diplomatic Containment

Mar 16

Intuit’s Buyback Gambit Shows How AI Panic Is Warping Wall Street

Mar 16

Gas Prices Over $100 Per Barrel Will Force Fed to Choose Between Inflation Control and Economic Growth

Mar 16

Severe Weather Sunday and Monday: What We Know So Far

Mar 16

Why Meteorologists Keep Calling It the ‘Last’ Cold Front