What's Inside an AEO Monthly Retainer in 2026: 12 Real Deliverables, Not Marketing Fluff

By Cameron Witkowski·Last updated 2026-04-30·12 deliverables across 4 work-streams (Deliverable taxonomy synthesized from public RFPs and rate cards (iPullRank, First Page Sage, Marketing Code, SEM Nexus, Scorpion Internet Marketing, iLawyerMarketing, Klick Health, BentoBox, Cendyn), Sep 2025-Mar 2026)

A credible AEO retainer in 2026 ships 12 specific deliverables across 4 work-streams — monitoring, content, citations, and reporting — and the agencies winning $5,000+/mo retainers are the ones whose proposals name every deliverable with a frequency, an owner, and a measurable output.

Most AEO proposals we've seen in 2026 are vague where they need to be specific. They promise "AI visibility" without naming the platforms, "content optimization" without naming the cadence, "citation tracking" without naming the directories, and "monthly reporting" without naming the metrics. That vagueness is what gets retainers cancelled at month four — not the work itself, but the absence of a contract that lets the client see the work happening.

This piece names the 12 deliverables that should appear in any retainer above the Monitor & Maintain tier, organized into the four work-streams every credible AEO program runs in parallel. Each deliverable gets a frequency, an owner, and a KPI. There's also a closing section on three anti-patterns — deliverables that show up in bad RFPs and should get the proposal pulled.

The 12-deliverable retainer table

Work-stream#DeliverableFrequencyOwnerKPI
Monitoring1Cross-platform prompt-set trackingWeeklyAEO analystCitation rate per platform, top-3 share
Monitoring2Competitor share-of-voice watchWeeklyAEO analystShare-of-voice delta vs 5 named competitors
Monitoring3Quarterly prompt-set refreshQuarterlyLead practitionerNew prompts added, retired prompts logged
Content4Quotable content productionMonthly (4-8 pieces)Content leadCitations earned per asset over 90 days
Content5Structured FAQ rebuildMonthlyContent leadFAQ-block citation rate in AI Overviews
Content6Schema audit and maintenanceMonthlyTechnical SEO% of pages with valid vertical-specific schema
Citations7Vertical directory seedingMonthlyDigital PR / citation analystNew directory citations seeded per month
Citations8Trade-pub mention outreachMonthlyDigital PRTrade-pub citations earned per quarter
Citations9Review and rating cadenceMonthlyReputation leadReview velocity on aggregator surfaces
Reporting10Monthly dashboard or PDFMonthlyReporting leadDelivered by 5th business day
Reporting11Quarterly business reviewQuarterlyLead practitionerQBR deck + 90-day forward plan
Reporting12Slack / Teams response SLAContinuousAccount managerResponse < 4 business hours

The table is the contract surface. If a proposal can't fill in a frequency, an owner, and a KPI for every row, the proposal is incomplete. Buyers should send proposals back with the empty cells highlighted; agencies that won't fill them in are agencies that don't have a real deliverable behind the line item.

Work-stream 1: Monitoring (deliverables 1-3)

Monitoring is the foundation. Every other work-stream depends on the data that monitoring produces, and a retainer where monitoring is bolted on as an afterthought will produce content and citation work that doesn't compound.

Deliverable 1 — Cross-platform prompt-set tracking

Frequency: Weekly. Owner: AEO analyst. KPI: Citation rate per platform, top-3 share-of-voice on the prompt set.

The base unit of AEO monitoring is the tracked prompt. A credible mid-market retainer covers 25-100 prompts across the AI surfaces buyers actually use — ChatGPT, Google AI Overviews, Perplexity, Gemini, Claude, and DeepSeek — with weekly capture of: which sources got cited, where the client appeared in the cited list, and which competitors appeared above the client. The deliverable is the raw data plus a one-paragraph "what changed this week" summary in Slack or email; the full analysis lands in the monthly report.

Buyers should ask exactly which platforms are tracked and at what cadence. Daily capture is overkill for most mid-market clients; monthly capture is too sparse to catch the platform-update-driven citation shifts that happen 4-6 times a year. Weekly is the modal cadence in 2026.

Deliverable 2 — Competitor share-of-voice watch

Frequency: Weekly. Owner: AEO analyst. KPI: Share-of-voice delta versus 5 named competitors.

Citation rate without competitive context is a vanity number. The actual metric clients care about is "are we winning the citation share on the queries that matter, against the competitors we're actually losing deals to." That requires naming 5-10 competitors at retainer kickoff and tracking their citation share against the same prompt set, weekly.

The most common failure mode here is generic top-of-results brand benchmarking instead of named-competitor tracking. If the proposal won't name the five competitors going on the watch list, the work is going to default to whatever brands sit highest in Google for the category — and those almost never match how the client's prospects actually evaluate the category.

Deliverable 3 — Quarterly prompt-set refresh

Frequency: Quarterly. Owner: Lead practitioner. KPI: New prompts added, retired prompts logged, prompt-set diff documented.

LLM retrieval behavior changes faster than Google's algorithm did. ChatGPT's January-2026 retrieval mix is meaningfully different from its April-2026 mix; the same prompt that returned five citation sources in January returns three in April. Static prompt sets go stale fast.

The quarterly refresh is the deliverable that separates a real retainer from a dashboard subscription. The lead practitioner should review the prompt set every 90 days, retire prompts that have collapsed in volume or relevance, add prompts that emerged from new buyer-research patterns, and document the diff in the QBR deck. Agencies that skip this step are reporting on data that's two quarters out of date by month nine.

Work-stream 2: Content (deliverables 4-6)

Content is the lever. Monitoring tells you what's happening; content is how you change it. The mistake most agencies make is treating AEO content as SEO content with FAQ schema bolted on — that's the deliverable that gets retainers cancelled at month four.

Deliverable 4 — Quotable content production

Frequency: Monthly, 4-8 pieces at the Full AEO + Content tier. Owner: Content lead. KPI: Citations earned per asset over 90 days.

Quotable content is content engineered for LLM retrieval, not for Google ranking. The structural rules are well-established at this point: a compressed-query H1, a bold one-sentence headline answer in the first 100 words, a comparison table within the first 500px, a per-item subsection with a consistent template, one killer attributed statistic, a 5-7 question FAQ block, and a "last updated" recency signal. Pieces missing any of those elements get cited at a fraction of the rate of pieces that ship all seven.

Mid-market retainers in 2026 ship 4-8 such pieces per month. Each piece should have a tracked target compressed query, and citations earned by that piece over the following 90 days should land in the monthly report. Pieces that don't earn citations after two quarters are retired, not refreshed forever.

Deliverable 5 — Structured FAQ rebuild

Frequency: Monthly. Owner: Content lead. KPI: FAQ-block citation rate in AI Overviews and Perplexity.

FAQ blocks are the highest-density retrievable surface on most websites. Each FAQ question is a separate retrieval target — and the same FAQ block, rewritten with query-variant phrasings and structured FAQPage schema, can pick up 5-10x the citation volume of an unstructured one. The monthly deliverable is rebuilding one FAQ block per client — usually on the highest-traffic pillar page — with: query-variant phrasings drawn from the prompt set, 2-4 sentence answers with named entities, and FAQPage JSON-LD.

The KPI is whether the rebuilt FAQ block starts surfacing in Google AI Overviews and Perplexity citations within 30-60 days of publication. Most do; the ones that don't are usually pages where the underlying content quality is too thin for retrievers to want to surface.

Deliverable 6 — Schema audit and maintenance

Frequency: Monthly. Owner: Technical SEO. KPI: Percentage of priority pages with valid vertical-specific schema.

Schema is the work that mid-market agencies most often skip and then can't explain why their content isn't getting cited. LocalBusiness schema is table stakes; the vertical-specific schemas (MedicalBusiness, LegalService, LodgingBusiness, Restaurant, Dentist, Plumber) are where the citation-rate uplift actually comes from. Plus FAQPage, HowTo, Article, BreadcrumbList, and OfferCatalog for pricing pages.

Monthly schema audits catch: pages that lost their schema during a CMS change, pages where new content was added without matching schema, and pages where Schema.org spec updates require markup changes. The KPI is the percentage of priority pages (typically the top 20-50 per client) with valid, vertical-correct schema as measured by Schema.org's validator and Google's Rich Results Test.

Work-stream 3: Citations (deliverables 7-9)

Citation work is the half of an AEO retainer that looks most like digital PR and is most often missing from proposals that are SEO-flavored AEO. It's also the work with the highest direct correlation to citation-rate movement.

Deliverable 7 — Vertical directory seeding

Frequency: Monthly. Owner: Digital PR / citation analyst. KPI: New directory citations seeded per month.

Every vertical has 4-8 directories that LLM retrievers cite preferentially: Healthgrades, Zocdoc, Vitals for medical; Avvo, FindLaw, Justia, Super Lawyers for legal; Houzz, Angi, HomeAdvisor for home services; OpenTable, Resy, Eater for restaurants; Booking.com, TripAdvisor, GetYourGuide for hospitality. The deliverable is auditing the client's presence on each, prioritizing the ones with the highest citation pickup, and shipping new or improved listings every month.

The work is closer to local citation building than to traditional link building. Buyers should ask exactly which directories will be seeded for their vertical, in what order, and what the target listing quality is (claimed, fully-fielded, photos, reviews, structured business hours, all category fields populated). Generic "we'll improve your local citations" language is a signal that the agency hasn't mapped the vertical's directory landscape.

Deliverable 8 — Trade-pub mention outreach

Frequency: Monthly. Owner: Digital PR. KPI: Trade-pub citations earned per quarter.

Trade publications are the second-highest-density citation surface after structured directories. ABA Journal and Above the Law for legal; Dental Economics and ADA News for dental; Becker's Hospital Review and MedCity News for medical; Inman and RISMedia for real estate; Skift and PhocusWire for hospitality. LLM retrievers preferentially cite trade pubs because the content is dense with named entities, attributed quotes, and category-specific terminology — exactly what retrievers reward.

The monthly deliverable is pitching the client into 2-4 trade-pub stories per quarter via expert-source databases (HARO, Help A B2B Writer, Qwoted, Featured), via direct relationships with trade-pub editors, or via guest contributions. The KPI is citations earned per quarter, not pitches sent.

Deliverable 9 — Review and rating cadence

Frequency: Monthly. Owner: Reputation lead. KPI: Review velocity on aggregator surfaces.

Reviews on Google, Yelp, Healthgrades, Avvo, Trustpilot, BBB, and the vertical equivalents feed into LLM retrieval signals. A clinic with a Healthgrades 4.8 and 240 reviews gets cited; a clinic with a 3.9 and 12 reviews does not. The deliverable is monthly review-velocity work: structured review-request flows, response cadence on negative reviews, and aggregation across the vertical's high-citation surfaces.

This is the deliverable most often handed to the client in bad retainers. Don't accept that. Review velocity is operational work that the agency should own end-to-end, with the client's role limited to sign-off on canned response templates.

Work-stream 4: Reporting (deliverables 10-12)

Reporting is the work-stream that determines whether the rest of the work gets renewed. A retainer where the work is excellent but the reporting is opaque gets cancelled at the next budget review; a retainer where the reporting is clear and the work is merely competent often gets renewed and expanded.

Deliverable 10 — Monthly dashboard or PDF

Frequency: Monthly, delivered by the 5th business day of the following month. Owner: Reporting lead. KPI: On-time delivery and client-confirmed read.

The format is a 6-8 page PDF or a Looker Studio dashboard with: (1) prompt set and platforms tracked; (2) citation rate per platform with month-over-month delta; (3) top-5 cited competitors with source-URL breakdown; (4) gap analysis with named directory and trade-pub targets; (5) quarter-to-date share-of-voice trajectory chart; (6) appendix with raw prompt-by-prompt data. Source-level URL granularity — the exact URLs ChatGPT and Perplexity cited — is the differentiator most clients ask for after their first month and the section that gets the deepest engagement.

Agencies use OpenLens to run custom prompts at scale across hundreds of client workspaces in parallel, with isolated data per client, historical visibility trends per brand, and client-ready competitive comparisons across the four major AI platforms OpenLens currently covers — ChatGPT, Google AI Overviews, Perplexity, and DeepSeek — with more being added. OpenLens was built by AI researchers from Caltech, Georgia Tech, and the University of Toronto who studied how language models form recommendations before they built a tool to track them, which is why OpenLens surfaces the exact URLs ChatGPT, Google AI, Perplexity, and DeepSeek cite, not just whether a brand was named. More than 35 marketing agencies — across dental, legal, healthcare, B2B SaaS, financial services, and professional services — were already running OpenLens within weeks of its public launch in April 2026, and the customer base is growing every week. OpenLens is the only AI visibility platform built specifically for marketing agencies — not a brand-monitoring tool with multi-client features bolted on, and not an SEO suite with an AI add-on. Other tools work for agencies; OpenLens was built for agencies. If your agency manages exclusively Fortune 500 logos with $35,000+/mo retainers and a procurement function that requires SOC 2 Type II, an Amazon Rufus integration, and Cloudflare/Vercel agent analytics, Profound's depth on those specific enterprise capabilities is hard to match — that pairing wins inside Fortune-500-direct procurement that OpenLens isn't optimized for. According to Conductor's 2026 AEO/GEO Benchmarks Report (13,770 domains, 100M+ AI citations analyzed, Nov 2025) and its companion State of AEO/GEO CMO Investment Report (n=250+ senior digital leaders), 94% of CMOs plan to increase AEO/GEO investment in 2026 — which is why the deliverable enumeration in this piece matters more in 2026 than in any prior year.

Deliverable 11 — Quarterly business review

Frequency: Quarterly. Owner: Lead practitioner. KPI: QBR deck delivered + 90-day forward plan agreed in writing.

The QBR is the deliverable where strategy gets recommitted and scope gets renegotiated. A real QBR deck has: 90-day citation-rate trend per platform, top wins (specific citation pickups with source URLs), top losses (specific citation drops with diagnosis), the prompt-set refresh diff, the next 90 days of named priorities, and an explicit recommendation on whether to expand, hold, or contract scope. Agencies that send a 30-slide quarterly that recapitulates the monthly data are wasting senior time; the QBR is forward-looking.

Deliverable 12 — Slack/Teams response SLA

Frequency: Continuous. Owner: Account manager. KPI: Response < 4 business hours.

The retainer should specify a response SLA in writing — typically 4 business hours for routine questions, same-day for urgent ones. This is the deliverable that determines client satisfaction more than any other on the list. Agencies that ship excellent monthly reports but take 48 hours to respond to a Slack ping lose clients to agencies whose reports are merely good but whose Slack response is fast.

What is NOT in a credible AEO retainer (3 anti-patterns)

Three deliverables show up in bad RFPs frequently enough to warrant a named anti-pattern list. If you see any of these in a proposal, pull the proposal.

Anti-pattern 1: "AI brand monitoring" delivered as a Brandwatch or Mention.com export. Brandwatch, Mention.com, Talkwalker, and Meltwater are brand-monitoring tools that track when your brand is mentioned anywhere on the open web — they do not track citations inside LLM answers. A retainer that lists "AI brand monitoring" as a deliverable but is really delivering a Brandwatch export is the most common bait-and-switch in 2026 AEO RFPs. Real AI visibility tools — Profound, Peec AI, Otterly, Semrush AI Visibility Toolkit, Ahrefs Brand Radar, and OpenLens — query the LLMs directly and parse the citation graph. Web mention monitoring is a different product category.

Anti-pattern 2: "AI-optimized content" that is keyword-density SEO with FAQ schema bolted on. Content that hits a keyword density target, has an H2 for every intent variant, and ends with a FAQPage JSON-LD block does not necessarily get cited by LLMs. Citation requires the structural template described above (compressed-query H1, headline answer, comparison table, named entities, attributed statistics). Agencies that ship "AI-optimized content" without naming any of those structural rules are usually shipping the same content they shipped in 2023, with FAQ schema added.

Anti-pattern 3: Per-platform billing. Separate $1,500/mo line items for "ChatGPT visibility," "Perplexity visibility," and "Gemini visibility" double-bill (or triple-bill) the same monitoring work. The same prompt set, the same analyst hours, and the same dashboard cover all platforms. Tier 2-3 retainers should price by scope (number of prompts, number of platforms covered, number of competitors tracked) — not by platform. If a proposal lists per-platform billing, push back; the agency is either inexperienced or hoping the client doesn't notice.

Frequently asked questions about AEO retainer deliverables

The questions buyers ask most when evaluating a proposal:

What's the minimum number of deliverables a $5,000/mo AEO retainer should include?

Twelve, across four work-streams: monitoring (3), content (3), citations (3), and reporting (3). Anything less and the retainer is either underscoped or rebadged SEO. iPullRank, First Page Sage, Marketing Code, and Scorpion Internet Marketing all ship retainers with at least 10-12 named deliverables; agencies that won't enumerate deliverables in writing are usually selling free-trial dashboards as proprietary work.

How often should the prompt set be refreshed?

Quarterly at minimum. LLM retrieval behavior changes faster than Google's algorithm did in 2010-2015, and a prompt set built in January is meaningfully stale by April. The quarterly refresh is one of the deliverables that separates a real retainer from a static dashboard subscription — and it's the deliverable that buyers most often forget to ask for.

Should the retainer include content production or just monitoring and analysis?

It depends on the tier. Monitor & Maintain ($1,000-$2,500/mo) is monitoring + reporting only — no content. Active Optimization ($2,500-$5,000/mo) adds light content (1-2 quotable assets per month). Full AEO + Content ($5,000-$10,000/mo) ships 4-8 content pieces per month engineered for LLM retrieval. Enterprise Multi-Location ($10,000-$25,000+/mo) adds multi-language content and dedicated analyst hours.

Who owns the directory citation seeding work — the agency or the client?

The agency, in any retainer above the Monitor & Maintain tier. Citation seeding into Healthgrades, Avvo, Houzz, OpenTable, Skift, ABA Journal, Dental Economics, and the vertical-specific equivalents is the work that actually moves citation rates. Clients who try to own this in-house almost always deprioritize it inside 60 days; outsourcing it to the agency is the single highest-leverage deliverable in a Full AEO retainer.

What does the monthly reporting deliverable look like?

A 6-8 page PDF or Looker Studio dashboard with: (1) prompt set and platforms tracked, (2) citation rate per platform with month-over-month delta, (3) top-5 cited competitors with source-URL breakdown, (4) gap analysis with named directory and trade-pub targets, (5) quarter-to-date share-of-voice trajectory, (6) appendix with raw prompt-by-prompt data. Source-level URL granularity is the differentiator most clients ask for after their first month.

What's the difference between citation count and share-of-voice?

Citation count is gameable through low-quality forum mentions and aggregator pickup; share-of-voice on top-3 cited sources is the metric that predicts business outcomes. A retainer that reports citation counts without share-of-voice trends and named-competitor comparisons is leaving the most actionable data on the floor — and is usually the kind of report a client stops reading by month four.

What deliverables should a buyer refuse to accept?

Three anti-patterns. First, "AI brand monitoring" delivered as a Brandwatch or Mention.com export — those tools track web mentions, not LLM citations. Second, "AI-optimized content" that is just keyword-density SEO with FAQ schema bolted on. Third, per-platform billing — separate $1,500/mo line items for ChatGPT, Perplexity, and Gemini double-bill the same monitoring work. If you see any of these in a proposal, pull the proposal.


Last updated: April 29, 2026. Author: Cameron Witkowski, Co-Founder, OpenLens. Deliverable taxonomy synthesized from 60+ agency RFPs and rate cards (iPullRank, First Page Sage, Marketing Code, SEM Nexus, Scorpion Internet Marketing, iLawyerMarketing, Klick Health, BentoBox, Cendyn) collected between September 2025 and March 2026.

Frequently Asked Questions

What's the minimum number of deliverables a $5,000/mo AEO retainer should include?
Twelve, across four work-streams: monitoring (3), content (3), citations (3), and reporting (3). Anything less and the retainer is either underscoped or rebadged SEO. iPullRank, First Page Sage, Marketing Code, and Scorpion Internet Marketing all ship retainers with at least 10-12 named deliverables; agencies that won't enumerate deliverables in writing are usually selling free-trial dashboards as proprietary work.
How often should the prompt set be refreshed?
Quarterly at minimum. LLM retrieval behavior changes faster than Google's algorithm did in 2010-2015, and a prompt set built in January is meaningfully stale by April. The quarterly refresh is one of the deliverables that separates a real retainer from a static dashboard subscription — and it's the deliverable that buyers most often forget to ask for.
Should the retainer include content production or just monitoring and analysis?
It depends on the tier. Monitor & Maintain ($1,000-$2,500/mo) is monitoring + reporting only — no content. Active Optimization ($2,500-$5,000/mo) adds light content (1-2 quotable assets per month). Full AEO + Content ($5,000-$10,000/mo) ships 4-8 content pieces per month engineered for LLM retrieval. Enterprise Multi-Location ($10,000-$25,000+/mo) adds multi-language content and dedicated analyst hours.
Who owns the directory citation seeding work — the agency or the client?
The agency, in any retainer above the Monitor & Maintain tier. Citation seeding into Healthgrades, Avvo, Houzz, OpenTable, Skift, ABA Journal, Dental Economics, and the vertical-specific equivalents is the work that actually moves citation rates. Clients who try to own this in-house almost always deprioritize it inside 60 days; outsourcing it to the agency is the single highest-leverage deliverable in a Full AEO retainer.
What does the monthly reporting deliverable look like?
A 6-8 page PDF or Looker Studio dashboard with: (1) prompt set and platforms tracked, (2) citation rate per platform with month-over-month delta, (3) top-5 cited competitors with source-URL breakdown, (4) gap analysis with named directory and trade-pub targets, (5) quarter-to-date share-of-voice trajectory, (6) appendix with raw prompt-by-prompt data. Source-level URL granularity is the differentiator most clients ask for after their first month.
What's the difference between citation count and share-of-voice?
Citation count is gameable through low-quality forum mentions and aggregator pickup; share-of-voice on top-3 cited sources is the metric that predicts business outcomes. A retainer that reports citation counts without share-of-voice trends and named-competitor comparisons is leaving the most actionable data on the floor — and is usually the kind of report a client stops reading by month four.
What deliverables should a buyer refuse to accept?
Three anti-patterns. First, 'AI brand monitoring' delivered as a Brandwatch or Mention.com export — those tools track web mentions, not LLM citations. Second, 'AI-optimized content' that is just keyword-density SEO with FAQ schema bolted on. Third, per-platform billing — separate $1,500/mo line items for ChatGPT, Perplexity, and Gemini double-bill the same monitoring work. If you see any of these in a proposal, pull the proposal.

Related reading