From Vibe Coding to Agentic Engineering: What 18 Months of AI Coding Progress Actually Means
Coding|

From Vibe Coding to Agentic Engineering: What 18 Months of AI Coding Progress Actually Means

Andrej Karpathy coined 'vibe coding' in early 2025. By mid-2026, it has evolved into agentic engineering. Here is the story of the most consequential shift in how software gets built — and where it goes next.

Reading time6 min
|
Words1,133
|
CategoryCoding
|
Vibe CodingAgentic EngineeringKarpathy

Advertisement

Google AdSense — ad code will be placed here after approval

In February 2025, Andrej Karpathy posted something on X that would name a movement: "There is a new kind of coding I call 'vibe coding,' where you fully give in to the vibes, embrace exponentials, and forget that the code even exists." He described typing a natural language prompt, accepting AI-generated diffs without reading them, and iterating purely by feel.

Eighteen months later, nobody serious calls it vibe coding anymore. The term that stuck is agentic engineering — and the distinction is not semantic. It represents the difference between "let the AI figure it out" and "architect a system where AI agents do the work under human supervision." Here is how we got here and what it means if you write software for a living.

The Timeline

DateEventSignificance
Feb 2025Karpathy coins "vibe coding"Named the emerging behavior of prompt-only development
Jul 2025Cursor usage surpasses 1M developersAI-native IDE goes mainstream
Sep 2025Claude Code launched with agentic featuresFirst terminal-based autonomous coding agent
Nov 2025GitHub Copilot adds Agent modeIncumbent embraces agentic paradigm
Jan 2026Claude Opus 4.6 ships with planning capabilityModels begin reasoning about multi-step engineering tasks
Feb 2026OpenAI Codex desktop app launchesDedicated agent orchestration tool, not IDE plugin
Mar 2026Karpathy introduces "Agentic Engineering"Distinguishes supervised multi-agent development from raw vibe coding
Apr 2026Cursor 3 ships with Agent-First interfaceThe IDE itself becomes an agent control plane
Apr 2026Claude Opus 4.7 achieves SWE-bench 87.6%AI autonomously resolves 87.6% of real GitHub issues
Apr 2026GPT-5.5 + Codex supports 7-10 hour autonomous operationAI agents run for a full workday without intervention

What Changed: 3 Fundamental Shifts

1. From Suggestions to Agents

In 2024, AI coding meant inline autocomplete. You typed, the AI guessed the next few lines, you accepted or ignored. GitHub Copilot and early Cursor defined this era.

In 2026, AI coding means agents that read your entire codebase, plan multi-file changes, execute them, run tests, and fix their own failures. Claude Code, Codex, and Cursor 3's Agent mode represent this shift.

The difference is not incremental — it is categorical. An autocomplete tool saves keystrokes. An agent reduces the number of decisions you need to make. An agent that self-verifies reduces the number of mistakes that reach production.

2. From Single Model to Multi-Agent

The most important architectural insight of 2026: one agent is good. Multiple agents working in parallel are transformative.

Codex pioneered this with sandboxed parallel workspaces — you assign one agent to build the API, another to write tests, and a third to review the code. They work simultaneously. They do not step on each other. Cursor 3 followed with /multitask and agent fleets. Claude Code added Agent Teams in April 2026.

The developer's role shifts from "write code" to "review diffs from N parallel agents and merge the good ones." This is not a marginal change in workflow. It is a different job.

3. From Stateless to Persistent

Early AI coding tools had no memory. Each session started fresh. You re-explained your project every time.

In 2026, Codex remembers your preferences across sessions through its Memory system. Claude Code learned persistent project context through .claude/settings.json and cross-session file system memory. Cursor indexes your entire codebase and maintains awareness across sessions.

The effect: the AI gets better at your specific project over time. It learns your conventions, your architectural preferences, and which patterns you tend to accept versus reject.

The Developer Sentiment Data

The numbers tell a striking story about adoption and preference:

  • 4% of all public GitHub commits are now generated by Claude Code (GitHub Octoverse, Q1 2026)
  • 46% of developers rank Claude Code as their #1 AI coding tool; Cursor at 19%, Copilot at 9% (r/MachineLearning survey, March 2026)
  • GitHub Copilot has 470 million paid subscribers, up 75% year-over-year (Microsoft earnings, Q1 2026)
  • Cursor is valued at $29 billion with over 1 million users and $2 billion+ ARR (March 2026)
  • 62% of professional developers report using at least 2 AI coding tools regularly (Stack Overflow survey, April 2026)

The Counterargument

Not everyone is convinced. The skeptics make three arguments worth taking seriously:

1. "Agentic coding produces superficial understanding." When developers accept AI-generated code without understanding it, they lose the mental model of their own codebase. This shows up later as debugging difficulty and architectural drift. Several senior engineers I respect have observed this pattern in their teams.

2. "The tooling creates vendor lock-in." Codex only works with OpenAI models. Cursor ties you to its agent runtime. Claude Code's Skills format is proprietary. Building your workflow around a specific tool's agent paradigm means switching costs later.

3. "Cost is unpredictable." Usage-based pricing for agent tools can produce surprising bills. One developer reported spending ~$2,000/week on Cursor's agent compute before switching to Claude Code's Max plan at a fraction of the cost. The pricing models are still evolving.

These are real concerns. They do not invalidate agentic engineering — but they argue for thoughtful adoption rather than wholesale surrender.

Where This Is Going

The trajectory from vibe coding to agentic engineering points toward a model where:

  1. Specs replace tickets. Instead of writing a Jira ticket, you write a specification that AI agents can execute against. Cursor's Design Mode and Claude Code's project configuration already support this.

  2. Review replaces writing. The developer's primary creative act shifts from writing code to reviewing and integrating AI-generated contributions. Code review becomes the job, not a step in the process.

  3. Orchestration replaces implementation. The most valuable skill becomes knowing how to decompose problems into tasks that AI agents can handle independently and in parallel — then assembling the results.

  4. Verification becomes the bottleneck. As AI generates code faster, the limiting factor is not writing speed — it is confidence that the code is correct. Claude Opus 4.7's self-verification and the emerging category of AI-powered testing tools address this directly.

None of this means developers are obsolete. It means the job is changing — from craftsman to conductor, from builder to architect, from coder to engineer in the fullest sense of the word. The developers I know who are thriving in 2026 are not the ones typing the fastest. They are the ones who have learned to manage a team of AI agents the way a senior engineer manages a team of junior developers: clear specifications, thorough review, and accountability for the result.

Last updated: May 4, 2026.

Advertisement

Google AdSense — ad code will be placed here after approval

Was this article helpful?

More in Coding

3 ARTICLES