The Rise of the AI Engineer

James Phoenix
James Phoenix

A new engineering role is emerging between ML research and software engineering, focused on building products with foundation models via APIs.

Source: Latent Space | Author: Swyx ([email protected]) | Date: June 2023


Core Thesis

Foundation models have created a “shift right” in applied AI. A new professional role, the AI Engineer, is emerging as distinct from traditional Machine Learning Engineers. AI Engineers build products by orchestrating foundation models through APIs rather than training models from scratch.

Tasks requiring 5 years and research teams in 2013 now take afternoons in 2023.

This follows the pattern of previous role innovations: Site Reliability Engineering, DevOps, Data Engineering, Analytics Engineering. Each emerged when a technology shift created enough surface area for a dedicated discipline.


AI Engineer vs ML Engineer

Dimension ML Engineer AI Engineer
Background PhD, research, PyTorch Software engineering, API fluency
Core skill Training models Orchestrating foundation models
Starting point Data collection, model architecture API docs, prompt engineering
Validation speed Weeks to months Hours to days
Language Python-centric Python + JavaScript
Product type Classifiers, recommenders, fraud detection Generative apps, agents, copilots

The author predicts a “flippening” in the job market. Currently ML Engineer roles outnumber AI Engineer roles 10:1 on job boards. Within five years, this ratio inverts based on growth trajectory analysis.

Udemy Bestseller

Learn Prompt Engineering

My O'Reilly book adapted for hands-on learning. Build production-ready prompts with practical exercises.

4.5/5 rating
306,000+ learners
View Course

Five Drivers of the AI Engineer Role

1. Foundation Models as Few-Shot Learners

Models exhibit emergent capabilities beyond their creators’ original intentions. Practitioners discover novel applications through experimentation, not training. The skill is knowing how to prompt and orchestrate, not how to build from scratch.

2. Concentrated Research Talent

~5,000 LLM researchers globally vs ~50 million software engineers. APIs serve as “AI Research as a Service.” The bottleneck is not model capability but product engineering on top of models.

3. GPU Economics

Hardware concentration forces startups to use APIs rather than train models. Billion-dollar funding rounds (Inflection $1.3B, Mistral $113M) emphasize GPU ownership. Most companies will consume AI, not produce it.

4. Rapid Prototyping (“Fire, Ready, Aim”)

LLM-based prototyping enables 10-100x faster product validation compared to traditional ML workflows. No data collection phase, no model training, no infrastructure setup. Prototype first, validate, then optimize.

5. Language Expansion

Tools now support JavaScript alongside Python. This doubles the potential developer audience and expands the total addressable market for AI tooling.


Software 1.0, 2.0, 3.0

Building on Andrej Karpathy’s “Software 2.0” concept:

Era Definition Example
Software 1.0 Hand-coded programming logic Traditional applications
Software 2.0 Neural networks approximating logic Trained ML models
Software 3.0 Human-written code orchestrating LLM power LangChain apps, AI agents, copilots

Software 3.0 is where AI Engineers operate. The moat is in orchestration code, not raw LLM access. This explains valuations like LangChain’s $200M+.


The “Code Core, LLM Shell” Pattern

Successful AI products use human code as the core with LLM capabilities wrapped around it. This pattern offers:

  • Defensibility against competitors who can access the same models
  • Reliability through deterministic code paths where possible
  • Security against prompt injection via code-level guardrails

This aligns directly with compound engineering principles: “Most ‘AI agents’ in production aren’t pure agentic systems. They’re predominantly deterministic code with targeted LLM decision-making.”


Notable AI Engineers

The article highlights practitioners building successful products without traditional ML credentials:

  • Simon Willison – Tools and exploration with LLMs
  • Riley Goodside (Scale AI) – Prompt engineering as a discipline
  • Pieter Levels – Photo AI, Interior AI (indie products)
  • Teams at Notion, Figma (via Diagram acquisition), Vercel

Compensation range: $300k-$900k at major labs, demonstrating market validation of the role.


Connection to Compound Engineering

The AI Engineer role maps directly to the compound engineering stack:

  1. Foundation model APIs replace model training (the “AI Research as a Service” layer)
  2. Orchestration code is the harness (deterministic code wrapping LLM decisions)
  3. Prompt engineering is context engineering (structuring information for models)
  4. Rapid prototyping aligns with the RALPH Loop (fast iteration with fresh context)

The AI Engineer is the practitioner. Compound engineering is the methodology. Context engineering is the core skill.


Key Insight

“This will likely be the highest-demand engineering job of the decade.”

The gap between ~5,000 LLM researchers and ~50 million software engineers means the leverage is in building on top of models, not in building the models themselves. Demand-and-supply economics will drive this role to dominance regardless of perception from traditional ML practitioners.


Related

Topics
Ai EngineerApi IntegrationFoundation ModelsLlm OrchestrationPrompt Engineering

More Insights

Cover Image for ASCII Previews Before Expensive Renders

ASCII Previews Before Expensive Renders

Image and video generation are among the most expensive API calls you can make. A single image render costs $0.02-0.20+, and video generation can cost dollars per clip. Before triggering these renders

James Phoenix
James Phoenix
Cover Image for The Six-Layer Lint Harness: What Actually Scales Agent-Written Code

The Six-Layer Lint Harness: What Actually Scales Agent-Written Code

Rules eliminate entire bug classes permanently. But rules alone aren’t enough. You need the three-legged stool: structural constraints, behavioral verification, and generative scaffolding.

James Phoenix
James Phoenix