Strategy Work vs Execution Work — Where AI Eats and Where Humans Stay
Strategy Work vs Execution Work — Where AI Eats
TL;DR: Across multiple marketing domains in 2026, a consistent pattern is visible: current AI tooling has the clearest ROI on high-volume, structured, pattern-driven execution work (creative production, creator discovery, brief writing, metrics tracking, code execution). What stays human-leveraged is strategy, judgment, integration, and leadership — the work that requires taste, market understanding, and cross-functional translation. This isn’t a prediction; it’s a synthesis of what’s already shown up empirically in three independent domains. The implications for how to staff teams, where to invest your own skill-building, and how to value different categories of work are substantial.
The Pattern
Three independent data points from three different marketing domains, each showing the same shape:
Paid media — Eric Seufert, Mobile Dev Memo (2022 onward). Algorithm-driven ad platforms (Meta Advantage+, Google PMax, TikTok Smart+) automated bidding, targeting, and placement — the three levers media-buyers used to pull manually. Creative concept and variation became the dominant remaining human lever. The phrase: glossary/creative-is-new-targeting. This is the execution layer being eaten by automation.
Influencer marketing — Modash 2026 salary survey, n=499. Influencer marketers handle ~19 distinct weekly tasks. The salary data shows a +$14,830 premium for owning campaign strategy, +$4,743 for team management, +$4,378 for cross-department collaboration. “High-execution-style tasks correlated with some of the lowest salaries globally.” See marketing/influencer-marketing-task-overload. This is the market already pricing the same shift, with hard salary numbers attached.
Software development — Karpathy, February 2025. The “vibe coding” framing names a workflow where the human describes intent and the AI produces the code. Execution-layer typing-and-syntax work compresses dramatically; what remains scarce is architectural judgment and product strategy. See glossary/vibe-coding. This is the same shape, in a domain (software) much further along the AI-tooling curve.
In each domain, the tools, vocabulary, and surface details are different. The underlying mechanism is the same: automation eats the high-volume, structured, pattern-driven layer of work first. The strategic, judgment-heavy, integration-heavy layer of work doesn’t go away — and its market value rises, because the execution layer it sits on top of just got cheaper.
What “Execution” and “Strategy” Mean Here
The terms are doing real work in this analysis, so it’s worth being precise.
Execution work has these signatures:
- High-volume, repetitive output (many ads per week, many outreach emails, many lines of code)
- Structured input → structured output (a brief produces a draft; a creative formula produces a variant; a feature spec produces a function)
- Pattern-recognizable (a competent practitioner’s output is correlated with the average competent practitioner’s output for the same input)
- Documented or learnable from examples (the AI can train on past instances)
- The judgment per-instance is small; the volume per-week is large
Examples across domains: writing 100 ad variants from a winning formula; outreach to 200 creators with personalized hooks; producing weekly performance reports; writing CRUD code; transcribing interview audio; tagging metadata.
Strategy work has these signatures:
- One-shot or low-volume, high-stakes decisions (which market do we enter; what’s the campaign thesis; what’s the architecture)
- Unstructured inputs requiring synthesis (qualitative signals, market state, organizational context)
- Pattern-breaking is often the point (sameness as a competitor’s strategy is usually a losing strategy)
- Requires cross-functional translation and political navigation (selling the thesis to leadership, aligning teams)
- The judgment per-instance is large; the volume per-week is small
Examples: deciding which creator partnership is on-brand; choosing the campaign thesis; setting category positioning; deciding what NOT to ship; coaching a teammate through a hard decision.
The line between them isn’t always clean — many real tasks have both layers. But the cleaner the task is on one side, the cleaner the AI-fit signal:
| Execution-heavy | Strategy-heavy | |
|---|---|---|
| Current AI fit | High (this is what AI does well in 2026) | Low (humans still substantially better) |
| Time per instance | Minutes to hours | Hours to weeks |
| Volume per week | Many | Few |
| Variability tolerable | Low (consistency wins) | High (novelty often wins) |
| 2026 salary direction | Flat or compressing | Premium, expanding |
| Personal-skill direction | Diminishing-returns to specialize | Increasing-returns to specialize |
Empirical anchors
The pattern isn’t just theoretical. The hard numbers across the three domains:
Modash 2026 salary survey
Salary delta by task ownership (n=499 influencer marketers):
| Task ownership | Salary impact |
|---|---|
| Campaign strategy | +$14,830 |
| Team management | +$4,743 |
| Cross-department collaboration | +$4,378 |
| Ambassador program management | +$2,668 |
| Budget management | +$2,348 |
“High-execution-style tasks correlated with some of the lowest salaries globally.” — Modash, 2026
The +$14,830 strategy premium isn’t a Primores extrapolation; it’s a direct survey finding. Influencer marketers who own strategy earn nearly 30% more than the global median ($49,981) for the role. The market is already pricing strategy as the scarce resource.
Post-ATT performance marketing (Seufert et al., 2022-2026)
Three years of practitioner consensus across DTC, mobile-app, and ecommerce performance marketing, anchored by:
- Apple iOS 14.5 ATT rollout (April 2021) cut off third-party signal
- Meta Advantage+ shipped 2022; Google PMax shipped 2021; TikTok Smart+ shipped 2024
- All three platforms converged on the same pattern: hand the algorithm a goal + creative pool + budget, let it decide everything else
- Roles in performance-marketing teams shifted from media buyers (audience-targeting expertise) to creative strategists (brief-writing, formula extraction, AI-assisted variation production)
The empirical evidence isn’t a single peer-reviewed paper; it’s three years of platform documentation, agency hiring patterns, and practitioner writing all pointing the same direction.
AI-assisted software development (Valeo, etc.)
From automation/ai-developer-tools-cases and automation/ai-implementation-patterns:
- Valeo: 35% of code now AI-generated across 100,000 employees (Gemini Code Assist deployment)
- KPMG: 90% Gemini adoption in first month for repetitive research
- Adore Me: 20 hours → 20 minutes for product description batches (text generation)
- Galaxies: months → 48 hours for campaign testing (code-driven workflow)
These are the execution-layer compressions. What hasn’t moved: senior-engineering compensation, product-management compensation, design-leadership compensation. The execution-layer compression has increased the market value of the strategy/judgment layer that orchestrates it.
Peer-reviewed academic foundation
The strategy-vs-execution pattern now has direct empirical support from three randomized/quasi-experimental studies published in 2023, plus a theoretical foundation from earlier decision-science research. This is what was missing from the framing before.
| Study | N | Finding | Wiki anchor |
|---|---|---|---|
| Brynjolfsson, Li & Raymond (2023) Generative AI at Work, NBER WP 31161 | 5,179 customer-support agents | +14% issues/hour avg; +34% novices, ~0% experts | glossary/ai-skill-leveling |
| Noy & Zhang (2023) Science 381, 187–192 | 444 college-educated professionals | Time −40%, quality +18%; AI compresses rough-drafting, idea generation and editing remain or grow in relative share | glossary/ai-task-restructuring |
| Dell’Acqua et al. (2023) HBS WP 24-013 | 758 BCG consultants | Inside frontier: +12.2% tasks, +25.1% faster, +40% quality. Outside frontier: −19pp accuracy. The frontier is invisible from the outside. | glossary/jagged-frontier |
| Klein 1998 / Kahneman-Klein 2009 Sources of Power / American Psychologist 64(6) | Theoretical synthesis | Intuitive expertise reliable only when (a) environment is high-validity and (b) feedback is rapid and unambiguous. Predicts where AI pattern-matching will struggle for the same reasons | glossary/recognition-primed-decision |
Three findings the academic evidence sharpens:
-
Skill leveling is real and consistent. All three empirical studies independently find that low-skill workers gain disproportionately from AI tools. This is the labor-economics mechanism behind the bifurcation — it’s not just “execution gets cheap”; it’s “execution gets accessible to anyone.” Senior strategy hires get harder to substitute for; junior execution hires get easier to substitute for.
-
AI is asymmetric, not uniform. Dell’Acqua’s −19pp outside-frontier finding is the dark counterpart to all the productivity numbers. AI doesn’t just fail to help outside its frontier — it misleads. This sharpens the strategy-vs-execution claim from “AI handles execution” to “AI handles execution inside the frontier, and using it outside is actively dangerous.”
-
The Klein-Kahneman frame predicts where AI itself will struggle. Strategy work tends to live in lower-validity environments (long feedback loops, sparse data, novel contexts). The same conditions that make human intuition unreliable make AI pattern-matching unreliable. The reason to keep humans on strategy isn’t that humans are reliably accurate — it’s that neither is, and accountability needs to be located somewhere. This is a substantial refinement to the framing.
How to Tell If a Function Is on the Curve
Not every marketing function shows the pattern equally clearly. Three diagnostic signals to check:
Signal 1: Is there a high-volume, structured execution layer?
If yes, that layer is a candidate for AI compression. If the work is mostly bespoke judgment per-instance, it isn’t.
- Performance marketing: yes (creative variants, audience tests). Pattern visible.
- Influencer marketing: yes (creator discovery, outreach, briefs). Pattern visible.
- Email/CRM/lifecycle: yes (segment design, copy variation, send-time optimization). Pattern likely visible (open question, see questions/automation-eats-execution-next-domains).
- Brand-building (Sharp’s framework): mostly no (mental availability is built over years, not produced per-week). Pattern not visible. See marketing/brand-vs-content-layers.
- Investor relations / partnerships: mostly no (per-instance judgment dominates). Pattern not visible.
Signal 2: Has tooling matured enough that the execution layer is actually being automated, or is it still on the roadmap?
If automation is real today, the labor-economics shift is already visible. If automation is announced but not adopted, the shift is anticipated but not yet measurable.
- Paid media: real today. Six years of platform automation.
- Influencer marketing: partially real today. Tooling is maturing; some agencies have moved most of Tier 1 to AI, others haven’t started.
- Organic content/SEO: partially real today. AI-assisted content production is mainstream; agentic SEO tooling is earlier.
- Email automation: substantially real today. Klaviyo / Hubspot / Iterable have AI features in production.
Signal 3: Is the judgment-heavy layer well-defined and senior in the org chart?
If the strategic layer is well-defined and senior, it gets the rising premium. If the function lacks a clear senior strategy role, the compression collapses the function rather than splitting it.
- Performance marketing has clear strategy roles (creative direction, growth strategy). Compression bifurcates the function: senior strategy hires become more valuable, junior media-buyer hires get squeezed.
- Influencer marketing in 2026 doesn’t always have clear strategy roles — Modash’s data shows many marketers own end-to-end. The compression risks collapsing the role rather than splitting it. The strategic answer is staffing for the bifurcation explicitly: hire senior-strategy + AI tooling, not generalist mid-level execution.
Implications
For marketing operators, three concrete implications:
1. Where to invest your personal skill-building. If your weekly work is mostly Tier 1 execution in a domain where automation is real (creative production, creator outreach, ad copy variation), continuing to optimize execution speed has diminishing returns. The compounding investments are in the strategy layer: developing taste for which creators are on-brand, which campaign angles are differentiating, which brand positioning is defensible. The Modash data is uncomfortably specific: the +$14,830 premium goes to the strategy ownership, not the volume of execution shipped.
2. How to staff teams. Most marketing teams in 2026 are still staffed for the pre-automation labor profile: many junior execution-layer hires, a few senior leaders. The pattern argues for the inverse: a small number of senior strategy hires equipped with AI tooling that handles the execution layer. This requires actually trusting the AI tools, building review-and-quality processes for AI output, and resisting the temptation to add more execution headcount when budgets allow.
3. How to price work. For agencies and consultants, execution work is being commoditized as the AI tooling that produces it becomes available to clients directly. Pricing models pegged to execution volume (per-asset, per-creator, per-post) face downward pressure. Pricing models pegged to strategic outcomes (campaign performance, mental-availability metrics, attributed revenue) face upward pressure. The pricing-model shift is happening; the question is whether you’re shifting your offering ahead of it.
The pattern recurs at three layers (a fractal observation)
A useful way to test whether a framework is generalizable is to see if it shows up at multiple scales. The strategy-vs-execution pattern recurs at three layers, all simultaneously:
- Org chart layer — senior strategists make judgment calls; AI-augmented juniors handle volume execution. (This page’s main thesis.)
- Individual workflow layer — a human user retains routine judgment; an AI advisor is consulted selectively on hard decisions. See questions/ai-as-personal-advisor.
- Model architecture layer — Anthropic’s glossary/advisor-strategy (April 2026): a cheaper executor model (Sonnet/Haiku) drives tasks; a smarter advisor model (Opus) is consulted only on hard decisions. Benchmark numbers: Haiku + Opus advisor more than doubles Haiku’s BrowseComp accuracy at 85% lower cost than Sonnet alone.
The same architectural insight at three different scales. What it predicts is consistent: most decisions don’t require top-tier judgment; routing them to top-tier capability is wasteful; a small fraction of decisions are pivotal and worth top-tier; the system wins when escalation is selective and the executor is decent on its own.
This isn’t loose analogy. The model-architecture pattern works for the exact same reasons the org-chart pattern works: heterogeneous task difficulty + competent baseline + clear escalation criteria. The fractal property is empirical evidence that the strategy-vs-execution frame captures something structural about how cognitive work decomposes — not a 2026 incidental.
When the framing breaks down
Three honest limits:
Brand-building (Sharp’s framework) is not on this curve. Mental availability is built through years of consistent distinctive-asset deployment. The work is execution-heavy in one sense (lots of touchpoints over time) but strategy-heavy in another (the asset system, the consistency discipline, the long-horizon investment thesis are all judgment work). AI tools help with production efficiency but the underlying work doesn’t compress the way performance-marketing creative does. See marketing/brand-vs-content-layers for the full layer distinction.
Regulated and high-trust domains move more slowly. Healthcare marketing, financial services, legal advisory, and pharmaceutical work all have execution-layer tasks that look automatable but actually require regulatory or professional judgment per-instance. The same task name (e.g. “drafting outreach emails”) has very different AI-fit depending on whether the output ships under FDA review or directly to a creator’s inbox.
The pattern is descriptive, not predictive in detail. “Strategy stays human-leveraged” is a directional claim, not a permanent one. If multimodal models in 2027-2028 substantially close the strategic-judgment gap, the framing will need revision. The current confident reading is: the gap is real for the next 1-2 years; beyond that, it’s an open question.
Cross-functional creep is a real failure mode. Modash documented this specifically: 4 in 10 influencer marketers had social media management bolted onto their role, leading to 12% lower pay and 15% lower satisfaction. Adding more execution-layer responsibilities without specialization is value-destroying, not value-creating — for the marketer and the brand. The implication isn’t “do more”; it’s “specialize harder, with AI tooling underneath you doing the volume work that would otherwise pull you off-strategy.”
Related
Framework synthesis:
- glossary/automation-eats-execution — the named-framework glossary entry for this pattern (use this when citing in passing)
- questions/automation-eats-execution-next-domains — open question: which marketing functions are next on this curve?
Domain anchors:
- glossary/creative-is-new-targeting — the originating framing in paid media
- marketing/influencer-marketing-task-overload — the empirical anchor in influencer marketing (Modash 2026 survey)
- glossary/vibe-coding — the same pattern in software development (the accessibility/floor side)
- glossary/agent-engineering — the production/ceiling side of the same software-engineering layer; Karpathy’s Sequoia 2026 framing names the discipline that operates the strategy layer once vibe-coding has compressed the execution layer
- marketing/marketing-analytics-in-2026 — the pattern at the marketing-analytics layer. Execution (data prep, model fitting, multi-touch attribution table generation) gets automated by AI inside data clean rooms and MMM tools (Meridian, Robyn, Scenario Planner); strategy (which channels matter, how to allocate budget, when to trust attribution vs. incrementality) stays human-leveraged
Academic foundations (peer-reviewed):
- glossary/jagged-frontier — Dell’Acqua 2023, BCG × Harvard, n=758. AI is asymmetric: helps inside frontier, hurts outside.
- glossary/ai-skill-leveling — Brynjolfsson + Noy-Zhang + Dell’Acqua. Three independent studies showing AI raises low-performer productivity disproportionately.
- glossary/ai-task-restructuring — Noy & Zhang 2023, Science, n=444. AI shifts the bottleneck from drafting to framing and editing.
- glossary/recognition-primed-decision — Klein 1998 + Klein-Kahneman 2009. Theoretical foundation for why the frontier is jagged the way it is.
Architectural fractal:
- glossary/advisor-strategy — The same pattern at the model-architecture level. Anthropic’s April 2026 model-pairing pattern.
- questions/ai-as-personal-advisor — The same pattern at the individual-workflow level.
Adjacent:
- marketing/brand-vs-content-layers — the layer distinction (this analysis is about the content/performance layers; brand-building is its own layer)
- automation/finding-ai-use-cases — the TRIPS framework for scoring which tasks are AI-fit
- automation/ai-implementation-patterns — empirical anchor: 1,048 documented implementations show the same patterns
- marketing/preparing-for-agentic-ai — what changes when AI agents become the audience
- strategist-pattern — meta: how a wiki-as-thinking-partner amplifies strategic judgment
Key takeaways
- A consistent cross-domain pattern in 2026: AI tooling commoditizes high-volume, structured execution work first; strategy, judgment, and integration stay human-leveraged.
- Three independent data points anchor the pattern: paid media (Seufert / post-ATT writing), influencer marketing (Modash 2026 salary survey), software (Karpathy / vibe coding adoption).
- The Modash data is unusually specific: a +$14,830 salary premium for owning strategy, with execution-style tasks correlating with the lowest pay.
- Diagnostic signals for whether a function is on the curve: high-volume structured execution layer + matured tooling + well-defined strategy roles.
- Brand-building (Sharp’s framework) and regulated domains are explicitly not on this curve — different mechanism, different time horizon.
- Implication for individuals: invest skill-building in strategy, taste, integration. Implication for teams: staff for the bifurcation, not the legacy execution-heavy profile. Implication for agencies: shift pricing from execution-volume to strategic outcomes.
Sources
- Modash (2026). State of Influencer Marketing Salaries 2026. n=499 (399 in-house + 100 freelance). Source for the +$14,830 strategy premium and the “execution = lowest pay” finding. See marketing/influencer-marketing-task-overload for the full analysis.
- Eric Seufert, Mobile Dev Memo (2022-2026). Multi-year practitioner writing on the post-ATT performance-marketing shift. Source for the originating “creative is the new targeting” framing.
- Andrej Karpathy (February 2, 2025), tweet coining “vibe coding”. Source for the software-domain anchor.
- Google Cloud (April 2026), 1,048 AI implementations dataset. Empirical anchor for the execution-layer compression numbers (Valeo, Adore Me, Gelato, etc.). See automation/ai-implementation-patterns.
- Sharp, B. (2010). How Brands Grow. Source for the brand-building counter-example: not all marketing work is on this curve.
- Brynjolfsson, E., Li, D., & Raymond, L. (2023). Generative AI at Work. NBER Working Paper No. 31161. — n=5,179 customer-support agents. Source for the +14% / +34% novice / ~0% expert findings. See glossary/ai-skill-leveling.
- Noy, S., & Zhang, W. (2023). Experimental evidence on the productivity effects of generative artificial intelligence. Science, 381(6654), 187–192. — n=444 preregistered experiment. Source for time −40% / quality +18% / task restructuring findings. See glossary/ai-task-restructuring.
- Dell’Acqua, F. et al. (2023). Navigating the Jagged Technological Frontier. HBS WP 24-013. — n=758 BCG consultants, GPT-4 randomized field experiment. Source for the inside/outside-frontier asymmetry. See glossary/jagged-frontier.
- Kahneman, D., & Klein, G. (2009). Conditions for intuitive expertise: A failure to disagree. American Psychologist, 64(6), 515–526. — Theoretical foundation for when pattern-matching (human or AI) is reliable. See glossary/recognition-primed-decision.
- Primores synthesis (2026). The cross-domain pattern itself — five independent data points (three industry + three peer-reviewed academic) all showing the same shape — consolidates a working framework. Three of the five anchors are now peer-reviewed academic studies; the framework no longer rests only on practitioner observation.