Your client walks into their Q2 QBR expecting to see AI visibility trending up. You pull the dashboard and report a blended AEO Citation Score of 42 across all engines—a 23% improvement from Q1. The client nods. Then they ask: "But what about ChatGPT specifically?" You don't have that breakdown. They press: "Who's citing us?" You don't track sentiment. They ask: "How much of this is real vs the ones who ranked us in the 'avoid this' section?" And then, at the end of the meeting, they say they're shopping for a new agency.
This scenario plays out in retainer reviews across 200+ agencies right now. Not because the work isn't happening, but because the measurement and the framing are both broken. The 10 mistakes below are the patterns GenPicked Research Team sees repeating across agency portfolios—each one costs retainer value, each one is preventable, and most of them are invisible until the QBR conversation goes sideways.
Start your 14-day free trial
Growth plan free for 14 days. Five AI engines. Full agency dashboard.
Start free trialWhy now: three numbers that should reshape your AEO playbook
Conductor's 2026 State of AEO/GEO Report documents what every agency owner should be running toward right now:
Put those together: 94% of your clients' prospects are using AI when they research. ChatGPT captures 87.4% of the traffic from those sessions. But when Google's AI Overviews appear, the ranked page loses 58% of clicks to the answer. Being #1 in Google no longer means what it meant last year. Your clients know something is broken in your reporting. You can feel it in the QBR—they're asking harder questions, comparing you to competitors, and the retainer conversation is tighter than it used to be. The 10 mistakes below explain why. More importantly, they're all fixable before your next retainer review.
Mistake #1: Reporting One Averaged AI Visibility Score Across All Engines
The Problem
You pull an AEO dashboard, see a blended score of 42/100, and report it to the client. You just told them almost nothing. ChatGPT mentions brands in 73.6% of its answers. Claude mentions brands in 97.3%. The same brand can be invisible on ChatGPT and category-leading on Claude. The averaged number hides the gap that actually matters.
Why It Costs Retainers
A client with an averaged AEO score of 42 might have a ChatGPT subscore of 18 (invisible) and Claude subscore of 68 (competitive). The average of 43 masks that their prospects using ChatGPT—the 87.4% engine—see them as absent. You enter the QBR believing the number. The client asks "why am I not in ChatGPT?" and you don't have the answer because your dashboard never split the engines.
QBR Signal You'll Spot
Client: "How are we doing on AI?"
You: "Our AEO score is up 23%."
Client: "But what about ChatGPT specifically?"
You: [pause] "Let me check that."
The Fix
Report per-engine breakdown always. Use weighting to show which engines drive actual client value: ChatGPT 0.35, Perplexity 0.25, Gemini 0.25, Claude 0.15. This weighting mirrors Conductor's traffic distribution data. Tools like Semrush AI Visibility Toolkit and Conductor's AEO benchmarking enable this tracking. Build a dashboard grid showing each engine as its own column with subscore and mention frequency.
A single visibility score is a liability, not a metric. It masks the specific engine where your client is losing. Always report engine-by-engine, always.
Mistake #2: Ignoring the 58% CTR Cliff from AI Overviews
The Problem
A brand ranking #1 on Google now receives 58% fewer clicks when an AI Overview appears for that query. Your client is likely still budgeting for the old CTR model—the one where position #1 = X traffic. It doesn't anymore. This shift is not theoretical. It's happening now across 48% of queries tracked.
Why It Costs Retainers
A mid-market SaaS client with 1,000 monthly clicks on an informational query now gets ~420 clicks. That's $3,000-$8,000 in lost monthly revenue. You're reporting "ranked #1," the client sees traffic down 35%, and you don't have an explanation. The fix isn't SEO (the rank is still solid)—it's AEO. You need to be cited in the AI Overview, not just ranked beneath it.
QBR Signal You'll Spot
Client: "Our organic traffic dropped 35% but we're still ranking top-3. What happened?"
You: [nervous] "Let me investigate."
The Fix
Audit each key query for AI Overview presence using Ahrefs' tool. For every informational query you target: (1) Is an AI Overview present? (2) Are you cited in it? (3) If not, who is? If you're ranking #1 but not cited, your content is missing schema, or your messaging is buried, or you lack the domain authority the AI systems trust. Cited brands achieve 35% higher organic CTR and 91% higher paid CTR compared to non-cited brands on the same queries. Being cited is worth the work.
Mistake #3: Treating Brand Mentions and Backlinks as Equivalent
The Problem
You're allocating 70-80% of link-building budget to backlink acquisition, with 20-30% on earned brand mentions. Brand mentions correlate 0.664 with AI visibility; backlinks correlate 0.218 across a study of 75,000 brands by RivalHound and Ahrefs. That's roughly a 3:1 advantage for mentions over links. You're building the wrong assets for the channel.
Why It Costs Retainers
Your client has a strong backlink profile (50+ referring domains) but zero presence in "best of" comparison content, industry roundups, Reddit discussions, and YouTube mentions. Their citations remain flat even though their link profile is solid. The problem isn't your link-building—it's that you're not optimizing for the signals AI systems actually weight.
QBR Signal You'll Spot
Strong backlink data on your SEO audit, but AI citation count isn't moving. When you dig into the citations you do have, they're mostly on low-trust domains (directories, affiliate sites). Zero citations from tier-1 industry publications where your prospects actually research.
The Fix
Rebalance your external strategy 50/50 to brand mentions and links. Invest in original research + data reports (high mention value), participation in comparison guides and roundups, community contributions (Reddit, forums), and YouTube presence. YouTube mentions correlate 0.737 with AI visibility—the highest single signal per Ahrefs. Track unlinked mentions using Semrush Enterprise AIO or RivalHound.
Mistake #4: Using Generic Schema Markup (Which Underperforms No Schema)
The Problem
Why It Costs Retainers
Your dev team implements schema across 50 pages. Clients see the structured data tags and assume it's working. AI citations remain flat or decline. You can't explain the ROI.
QBR Signal You'll Spot
Your schema audit shows 100% coverage of JSON-LD markup. AI citations are flat or declining. When you examine the schema, it's generic (no pricing, no ratings, no product attributes).
The Fix
Delete generic schema implementations. Implement only attribute-rich types: (1) For ecommerce: Product + Review schema with populated pricing, ratings, availability. (2) For B2B: Organization schema + FAQPage schema. (3) For content sites: FAQPage schema on pillar topics. FAQ schema is 3.2× more likely in AI Overviews per Frase. Only implement schema where you have rich data to populate all attributes. Incomplete schema is worse than no schema.
Mistake #5: Not Tracking Reddit Citation Share Per Engine
The Problem
Reddit accounts for 46.7% of Perplexity's top 10 citations but only 11.3% of ChatGPT citations. You're treating Reddit as a uniform signal across all engines. It isn't. Perplexity is growing 15%+ month-over-month with highly engaged, early-stage research audiences. Zero Reddit presence means Perplexity essentially ignores your client.
Why It Costs Retainers
Your client is visible in ChatGPT (3-5 mentions/week) but invisible on Perplexity despite Perplexity's user base being your client's most valuable early-stage research audience. Competitors without your backlink budget beat you on Perplexity. You don't report this breakdown, so clients don't realize the gap.
QBR Signal You'll Spot
Brand visible in ChatGPT but Perplexity citations remain under 1 per week despite Perplexity's growth. When you check where the Perplexity citations come from, they're mostly generic comparison content, not your client.
The Fix
Identify 5-8 communities where your client's target audience researches. Monitor threads for mention opportunities. Create referenceable assets (comparisons, use-case libraries, FAQs). Track Perplexity-specific brand citations separately in your dashboard. Report Perplexity as a distinct engine, not lumped into "AI engines generally."
Mistake #6: Focusing Entirely on Top-of-Funnel AI Visibility
The Problem
You optimize for brand name visibility and high-intent keywords. You're ignoring mid-funnel queries where only 38% of AI-cited pages rank top-10 for the original query. Google's AI systems pull from fan-out queries (related keyword SERPs), not direct matches. Mid-funnel comparison, feature, and use-case queries drive 2-3× conversion lift per HubSpot.
Why It Costs Retainers
Your client has strong brand presence on high-volume keywords, but AI-driven inbound doesn't reflect expected conversion rates. You're measuring top-of-funnel visibility but not influence on purchase decisions. The retainer conversation shifts from "How visible are we?" to "Are these leads actually converting?"
QBR Signal You'll Spot
Strong brand presence on high-intent keywords in your AI visibility dashboard. Client reports AI-driven inbound in CRM but the conversion rates look odd (lower than organic or paid).
The Fix
Expand your content strategy to mid-funnel queries: comparisons, features, use-cases, alternatives. Optimize for FAQ schema on these pages (3.2× AI citation boost per Frase). Track AI citations on mid-funnel pages separately from top-of-funnel. Report both brand visibility AND mid-funnel influence in QBRs. Show the client that citations on "alternative to [competitor]" queries matter more than mentions on their own brand name.
Mistake #7: No Integration Between SEO Rankings and AI Citation Status
The Problem
28.3% of ChatGPT's most-cited pages have zero organic Google visibility. You're treating SEO and AEO as separate channels with separate metrics. You're missing opportunities.
Why It Costs Retainers
ChatGPT referral traffic comes from pages with no Google ranking data. You can't explain the traffic source to the client. These "ghost pages" target long-tail queries with minimal search volume but consistent ChatGPT citations. Opportunity to expand niche content goes undetected because you didn't connect the two dashboards.
QBR Signal You'll Spot
ChatGPT referral traffic appears in your attribution. You can't identify which pages are driving it because you haven't matched Google Search Console ranking data to AI citation data.
The Fix
Build a unified visibility dashboard showing Google ranking position + AI citation status for each page. Identify pages with citations but no organic ranking (invest in E-E-A-T signals to unlock Google rank). Identify pages ranking but not cited (missing schema, buried message, quality issues). Identify pages absent from both (deprioritize). Tools like ZipTie.dev or custom sheets pulling from Google Search Console + Semrush AI Visibility + Ahrefs enable this.
Mistake #8: Requesting llms.txt File Implementation (Wasting Dev Time)
The Problem
SE Ranking's analysis of 300,000 domains found zero measurable correlation between llms.txt and AI citation frequency. Yet agencies keep requesting it. It's a dev task consuming hours with zero measured uplift.
Why It Costs Retainers
Client sees llms.txt implemented. You bill for the work. AI citations don't improve. Client questions ROI on "AI optimization." You default to "it's a best practice." But you don't have citation data before/after to prove it.
QBR Signal You'll Spot
Client asks: "What's llms.txt doing for us?" You don't have citation data before/after. You say it's a "best practice," but you can't back that up with client-specific results.
The Fix
Don't implement llms.txt. If a client insists, A/B test it on a small cluster and measure citations. Redirect dev time to: (1) FAQ schema implementation (proven 3.2× boost). (2) Attribute-rich Product/Review schema (proven 61.7% citation rate per Growth Marshal). (3) Content expansion on mid-funnel queries. (4) Brand mention generation (proven 0.664 correlation per RivalHound). Only 10.13% of domains use llms.txt with zero measured impact.
Audit your current AEO recommendations. If you're recommending llms.txt without citation data, remove it. Your credibility is worth more than a checkbox implementation.
Mistake #9: Reporting AI Visibility Without Sentiment or Context
The Problem
You report: "Brand appeared in 47 AI citations this month." The client sees 47 as progress. But 30 of those citations were in "avoid this competitor" contexts where the client lost. Mention counts without sentiment are visibility debt, not assets.
Why It Costs Retainers
Citation counts overstate performance. You're reporting positive signal, the client is hearing competitive loss on the sales floor ("prospects told us about this cheaper alternative"), and the disconnect destroys trust in your reporting.
QBR Signal You'll Spot
Your report shows 60 monthly ChatGPT citations (up 40% YoY). Client says: "But sales is telling us prospects are comparing us to [competitor] more, not less."
The Fix
Track three metrics instead of one: (1) Mention count (raw citations). (2) Sentiment distribution (% positive / neutral / negative). (3) Citation type (brand mentions vs comparison rankings vs feature references). Use Semrush AI Visibility Index (includes sentiment drivers) or RivalHound (includes competitor context). Report sentiment split in every QBR. This turns "47 citations" into "31 positive brand mentions, 12 neutral features, 4 competitive losses."
Mistake #10: Treating 6-Month AI Visibility Timelines as Industry Standard
The Problem
You promise clients AI citation growth by Q3, then deliver flat results by Q2 and lose the retainer. AI citation building is slower than organic SEO—it requires content freshness, entity verification, and distributed brand mentions. Content freshness: 40% of Perplexity's weighting per ConvertMate. Entity verification takes 2-3 months. Brand mention campaigns take 8-12 weeks.
Why It Costs Retainers
Clients see flat results in months 1-4, assume failure. Agencies lose ~8% of clients in months 1-6. Churn is highest at the 4-6 month mark when the client attends Q2 QBR expecting "40% citations growth" and you report "still building baseline."
QBR Signal You'll Spot
Client attends Q2 QBR expecting citations "up 40%" (per your initial pitch). You report: "still building baseline." They ask: "So what exactly are we paying for?" And suddenly the retainer is at risk.
The Fix
Set realistic timelines and report input metrics, not month-1 guesses: (1) Months 1-2: Baseline + content audit. Output: baseline ACS + gap map. (2) Months 2-4: Content optimization + schema. Output: 20-30% schema coverage, 5-10 refreshed pillar pages. (3) Months 4-8: Citation building (mentions + roundups). Output: 30+ new earned mentions, 3-5 video placements. (4) Months 8-12: Entity verification + ongoing program. Output: 40-50% citation growth. Report input progress ("completed schema on 8 pages") and early signals ("2 new earned mentions on roundup sites") so the client sees momentum even before citations shift.
Start your 14-day free trial
Growth plan free for 14 days. Five AI engines. Full agency dashboard.
Start free trial