Research Report · Language & Discourse

The Words That Ate Product Management

A deep study of how AI rewrote the vocabulary of product management in 2026 — what's new, what's dying, and the structural forces that explain it all.

Terms Catalogued
24 key terms
Sources Analysed
30+ sources
Time Period
2024 → 2026
1
The Context

A Language Under Reconstruction

Every major technology shift produces a vocabulary crisis. The words that helped us understand the old system become inadequate — sometimes actively misleading — for navigating the new one. That is exactly what is happening to product management in 2026.

The linguistic shift in AI PM isn't cosmetic. It reflects a genuine change in what PMs do, what they're responsible for, and what failure looks like. When 94% of product professionals use AI frequently and nearly half have it embedded in their daily workflows, the vocabulary has to evolve to match the operational reality.

"Speed without direction is just chaos in disguise. The new vocabulary isn't about keeping up with trends — it's about having the precision to talk about systems that didn't exist three years ago."

— Synthesised from 2026 Product School & Ant Murphy discourse
94%
of product professionals use AI frequently in 2026, per product school research
82%
of IT and data leaders say prompt engineering alone is no longer sufficient
1,445%
surge in multi-agent system inquiries from Q1 2024 to Q2 2025 (Gartner)
2/3
of all compute will be inference workloads in 2026, up from 1/3 in 2023
2
Arriving Vocabulary

The Terms That Now Define the Work

These are the terms that have moved from engineering Slack channels into product strategy decks, job descriptions, and stakeholder updates. They are no longer jargon — they are the operating vocabulary of the discipline.

Agentic & Orchestration
Infrastructure Economics
Quality & Evaluation
Culture & Discourse

Agentic & Orchestration Cluster

The vocabulary of AI that acts, not just answers

Agentic

Agentic Workflow

AI that executes multi-step tasks across tools — email, CRM, code repos, files — with minimal human input between steps. The defining shift from "AI that answers" to "AI that acts." In 2026, the industry is firmly post-copilot: 40% of enterprises plan to embed agents by year-end, up from near-zero in 2023.

🔺 In active use in virtually every enterprise AI strategy doc in 2026

Agentic

Orchestration

The coordination layer that manages which agents fire in what order, how they hand off to each other, and when to escalate to a human. As important as the individual models themselves — "success depends less on model capability and more on orchestration architecture."

📍 Gartner: 1,445% surge in multi-agent inquiries Q1 2024 → Q2 2025

Agentic

AgentOps

The emerging discipline for managing AI agents in production — parallel to DevOps and MLOps. Covers monitoring, reliability, cost governance, and escalation paths. Analyst firms expect enterprises to spin up dedicated AgentOps functions as they scale agents past pilot phase.

🔺 Emerging as a new PM specialty; expect job titles by 2027

Agentic

Human-in-the-Loop (HITL) — shifted

HITL: human reviews outputnew meaning

Critically, the meaning of HITL has shifted. In 2022 it meant a human reviewed a model's output before it shipped. In 2026 it means a human approves specific actions before an agent takes them. The same acronym now describes a different checkpoint in a different kind of system.

⚠️ Same term, different operational reality — causes confusion across orgs

Agentic

Compound AI

An AI system that combines multiple components — a retriever, a reasoning model, a tool execution layer, and a memory store — rather than relying on a single monolithic model. Accountability is diffuse in compound systems: when things go wrong, which component is responsible?

📍 Enterprise buyers increasingly asking vendors to disclose component architecture

Agentic

Agent-Washing

Marketing standard, linear automation workflows as "agents" when they lack genuine autonomy, reasoning, or adaptive behaviour. The 2026 equivalent of "AI washing" — a term PMs need to both detect in vendor pitches and avoid in their own communication.

⚠️ FTC has flagged this category; due diligence imperative for enterprise buyers

Infrastructure Economics Cluster

The vocabulary that turned infra costs into PM decisions

Infra Economics

Context Engineering

Prompt EngineeringContext Engineering

Where prompt engineering optimises how you ask a model, context engineering optimises what the model knows when answering — what you select to include, compress, isolate, and store across turns. It is persistent infrastructure, not per-interaction tuning. 82% of IT leaders say prompt engineering alone is now insufficient.

🔺 Displacing "prompt engineering" in serious technical PM discourse

Infra Economics

Token Economics

The discipline of managing cost-per-token across AI workflows — input tokens, output tokens, and the expensive category of reasoning tokens that represent hidden internal computation. In multi-agent coding systems, over 53% of token spend goes to re-consuming context rather than generating net-new output.

📍 "Can we afford to scale it?" has replaced "Can we build it?"

Infra Economics

Test-Time Compute / Long Thinking

Reasoning models generate hidden internal tokens before producing an output — "long thinking." This dramatically increases latency and cost: chain-of-thought can generate 10–20× more tokens than a standard inference pass. Enabling "reasoning mode" is now a product decision with direct budget implications.

📍 Inference compute exceeded training spend globally for the first time in early 2026

Infra Economics

Inference Economy

The macro-economic frame for the period now that spending on running models has overtaken spending on training them. Inference workloads will account for ~two-thirds of all compute in 2026. Intelligent routing can reduce inference cost by 30–60% in mixed workloads — a PM-level optimisation decision.

🔺 Spending on inference chips projected to reach $50B+ in 2026

Infra Economics

AI-Native (as prefix)

Products where AI is the core value engine — not a feature layered on top. "If the AI were removed, the product would not just cease to function as intended — it would cease to be useful at all." Seen applied broadly: AI-native company, AI-native workflow, AI-native PM. IDC projects AI-native enterprises will capture 60% of new SaaS market share.

🔺 Rapidly becoming the dominant frame for startup positioning in 2026

Infra Economics

World Models

AI systems that predict the next state of the world rather than the next token — learning physics, causality, and spatial relationships. Mostly research-stage, but entering PM vocabulary as an aspirational frame for what comes after LLMs: models that reason about reality, not just language.

📍 Appearing in frontier product strategy discussions; watch 2027–28 for productisation

Quality & Evaluation Cluster

The vocabulary of knowing whether your AI is actually working

Evaluation

Evals (AI Evaluations)

Structured test suites that measure AI system quality across multiple dimensions: accuracy, tone, safety, relevance. Unlike traditional software tests, evals are probabilistic. The OpenAI CPO named "writing evals" as the single most important skill for PMs in 2026. "AI features don't fail because of the model. They fail because nobody evaluated them."

🔺 Fastest-rising PM skill in job postings at Anthropic, OpenAI, Stripe, Linear, Figma

Evaluation

LLM-as-a-Judge

Using one language model to evaluate the outputs of another — a scaling technique for evals that removes the bottleneck of human review. The judge model applies structured criteria (a rubric) to score primary model outputs. The technique has limitations PMs must understand: a judge model can share the biases of the judged one.

📍 Standard practice at AI-native orgs; now diffusing into traditional product orgs

Evaluation

pass@k vs pass^k

Anthropic's eval framework for agents. pass@k: the agent succeeds at least once across k attempts. pass^k: it succeeds every time. An internal triage tool can live with pass@k; a customer-facing workflow cannot. The distinction defines whether an AI feature is safe to ship as-is or needs guardrails.

📍 Entering design-review vocabulary at AI product teams

Culture & Discourse Cluster

The vocabulary of the era's anxieties and attitudes

Culture

Vibe Coding

Coined by Andrej Karpathy in February 2025 and named Collins Dictionary Word of the Year 2025: describing to an AI what you want to build and letting it generate the code, "fully giving in to the vibes." Now matured into a legitimate development paradigm with two product categories — AI code editors (Cursor, Windsurf) and AI app builders (Bolt, Lovable). Collins made it official; Merriam-Webster listed it as slang in March 2025.

🔺 No longer ironic; "vibe PM" is also emerging as a frame for intent-driven product spec

Culture

Slop

AI-generated content produced at volume without human oversight or intent — "utter lack of purpose other than to farm engagement or fill space." Coined by programmer Simon Willison in 2024, formally inducted into the lexicon by Merriam-Webster and the American Dialect Society in January 2026. The slop family: workslop, promptslop, deckslop.

📍 Now a boardroom concern: "Is our AI output slop?" is a legitimate product quality question

Culture

Model Collapse

What happens when new AI models are trained on the outputs of previous AI models — degraded quality, amplified errors, loss of diversity and nuance. In 2026, early signs of benchmark saturation are appearing: models converging at near-maximum scores, collapsing measurable differences between them. A systemic risk, not just a technical one.

⚠️ Informing content strategy decisions: human-original data is now a premium asset

Culture

GEO (Generative Engine Optimisation)

The successor frame to SEO for a world where AI surfaces cite sources rather than rank pages. Slop is what you publish when you forgot to be a source worth citing. GEO-aware publishing — with statistics, quotes, structure, and honest FAQs — is how products maintain discoverability in AI-mediated search environments.

🔺 Moving from SEO/marketing into product content strategy decisions

3
Departing Vocabulary

The Terms Losing Their Grip

Not everything dies loudly. Some terms are being quietly retired — not banned, but hollowed out by overuse, superseded by more precise alternatives, or simply made irrelevant by the speed of change.

Prompt Engineering

Being supplanted by "context engineering" as the dominant frame for working with AI systems at scale. Prompt engineering describes a one-time, per-interaction intervention; context engineering describes ongoing infrastructure. The term still works for demos; it's insufficient for production.

→ Replaced by: Context Engineering, Context Design

AI-Powered

In 2023, "AI-powered" was a differentiator. In 2026 it is table stakes — so generic it communicates nothing. Serious product teams have stopped using it in internal docs. In external comms it persists, but sophisticated buyers discount it entirely.

→ Replaced by: specific capability claims, AI-native framing

Copilot (as aspiration)

The copilot frame — AI that assists a human who remains in the driver's seat — was the dominant paradigm of 2022–2024. It is now seen as underselling what agents can do, and increasingly as a constraint on product ambition. The industry has moved to fully agentic framing.

→ Replaced by: Agentic, Autonomous workflow, Orchestrator

Disruption (as neutral descriptor)

The Claytonism has been so thoroughly overused that it has lost precision. In 2026 PM discourse, particularly in AI-native circles, "disruption" has been replaced by more specific operational frames: "workflow automation," "category creation," "model displacement." The word still exists; serious PMs avoid it.

→ Replaced by: model displacement, workflow restructuring, category creation

LLM (in casual product discourse)

Not dying in technical writing — it remains precise and necessary there. But in product strategy, stakeholder communication, and leadership discourse, "LLM" is being replaced by the name of the specific model or system. "We're using Claude" communicates more than "we're using an LLM." The abstraction is losing its utility as products get specific.

→ Replaced by: model names, "foundation model," specific capability framing

Move Fast and Break Things

When AI systems touch real workflows, real data, and real customers at scale, the cultural permission to break things expires. The 2026 discourse is explicitly "pragmatism over hype." The mantra has been replaced by "eval before you ship" and "compute-aware design."

→ Replaced by: eval-driven development, governance-first, agentic accountability

4
The Shifts

Before / After: The Vocabulary in Transit

Language shifts are clearest when you hold the old term and the new one side by side. These transitions map the conceptual movement of the discipline.

2022–24

"We're building an AI-powered product"

2026

"Our product is AI-native — the model is the core value engine"

2022–24

"We're doing prompt engineering"

2026

"We're doing context engineering — building the knowledge architecture the model draws on"

2022–24

"AI with a human in the loop" (human reviews output)

2026

"Human-in-the-loop" (human approves an action before the agent takes it)

2022–24

"Can we build this with an LLM?"

2026

"Can we afford to scale this? What's the token cost at 10,000 users?"

2022–24

"We're shipping this AI feature"

2026

"We've run evals across 500 scenarios. Here are the pass@k metrics."

2022–24

"Developers are writing code"

2026

"Developers are orchestrating logic — vibe coding the architecture, agents doing the syntax"

5
Structural Forces

Why the Language Changed

Vocabulary doesn't change arbitrarily. The specific lexical shifts of 2026 are explained by four structural forces operating simultaneously across the industry.

🏭

Demo → Production

The era of impressive demos is over. AI systems are now in production — serving real customers, taking real actions, costing real money. Terms like "evals," "AgentOps," "token economics," and "inference cost" are production vocabulary. They exist because the failure modes are real now, not theoretical.

🤖

Agents Eating Operations

The shift from AI that answers to AI that acts has forced a vocabulary of action and accountability. "Orchestration," "HITL," "compound AI," and "agent-washing" are all terms that emerged because agents have real-world consequences. Language had to acquire precision it didn't need when AI was advisory-only.

💰

Cost as a Product Constraint

When inference compute exceeded training compute globally, cost stopped being an engineering concern and became a product decision. "Token economics," "test-time compute," "intelligent routing," and the inference economy are all terms that migrated upward from infrastructure because someone had to decide whether to enable reasoning mode on a $25/M-token model.

🗑️

The Great Slopification

The industrialisation of AI-generated content at scale — "slop" — has created a counter-vocabulary of quality and authenticity: GEO, model collapse, workslop, eval hygiene. When content can be produced infinitely and cheaply, the vocabulary of scarcity and craft migrated from editorial rooms into product thinking.

"In 2026, AI shifted from hype to pragmatism. The vocabulary followed. Terms got more specific because the failures got more concrete."

— Synthesised from TechCrunch, MIT Technology Review, IBM research, 2026
6
The PM Identity

How the Role Itself Is Being Renamed

The linguistic shift isn't just about external vocabulary — it's about how the PM role is being redefined from within. New titles are emerging, old expectations are being retired.

The PM That Was · Pre-2025

  • 💼 "I manage the backlog and sprint ceremonies"
  • 📋 "I write PRDs and user stories"
  • 🤝 "I coordinate between design and engineering"
  • 📊 "I define KPIs and track dashboards"
  • 🔮 "I understand the roadmap; the model is a black box"
  • ✍️ "I do prompt engineering for our AI features"

The PM That Is · 2026

  • 🎯 "I am an orchestrator of intelligent ecosystems"
  • 🧪 "I write evals and define what 'good' looks like for our AI"
  • 💰 "I make token economics decisions — we can't enable reasoning mode at this scale yet"
  • 🏗️ "I architect the context layer: what the agent knows and when"
  • 🔎 "AI-fluency is no longer a specialty — it's a baseline filter"
  • ⚖️ "I balance agent autonomy with governance, escalation paths, and accountability"
$159k
Average yearly pay for AI product managers in the US as of April 2026
80%
of AI PM roles are traditional products adding AI features — not AI-native orgs
more open roles in traditional companies than AI-native organisations

The job titles proliferating in 2026 — AI Product Manager, AI Strategist, Chief AI Officer, Head of AI, Platform PM, Application PM — reflect a genuine bifurcation: PMs who own the end-to-end user experience with AI, versus PMs who build the tools and infrastructure that other product teams use to build with AI. Both are needed. Neither is going away.