ESSAY № 04
How AI 'Query Fan-Out' Works — and Why Ranking #1 Isn't Enough Anymore
When a buyer asks ChatGPT "what's the best protein powder for lactose intolerance," ChatGPT doesn't run one search. It runs five. Your store has to rank for all five to appear in the answer.
AI engines don't take a buyer's question and hand it to a search engine. They break it apart — into 3 to 7 sub-queries — run each separately, then synthesize the results into a single paragraph. If your store ranks for the broad question but misses the specific sub-queries, you're invisible on the answer that actually gets shown. This is the mechanic most AI-SEO advice ignores, and it's why merchants with strong rank-tracking reports still don't show up in AI shopping answers.
"AI systems use 'query fan out' — breaking user requests into multiple sub-queries sent to search engines, then synthesizing results."
— Shopify Enterprise, "The New Rules of Generative Engine Optimization (GEO)"
What query fan-out is
Query fan-out is what happens between "buyer types a question" and "model generates an answer." Instead of treating the question as a single retrieval target, the engine expands it into a set of narrower sub-queries — each covering a different facet of the intent — runs each sub-query against its index independently, then concatenates the top passages as the context the language model sees. The final answer is synthesized from that combined context, not from any single retrieval pass.
Here's the expansion for the protein-powder question. One shopper question, five parallel retrievals:
- Broad query: "What's the best protein powder for lactose intolerance?"
- "Lactose-free protein powder brands" — surfaces which brands explicitly market a lactose-free product.
- "Whey isolate lactose content" — surfaces technical claims about residual lactose per serving.
- "Third-party lab tested protein powder" — surfaces verification and certification signals.
- "Protein powder reviews lactose intolerance" — surfaces first-person reviews from people with the condition.
- "Best-tasting dairy-free protein" — surfaces adjacent preferences the buyer didn't name but cares about.
The model sees passages from all five retrievals. If your product page has a paragraph about lactose-free sourcing but nothing about taste, nothing about lab testing, and no first-person reviews off-site, you're present in one retrieval. The other four pull from competitors. The synthesized answer quotes whichever store has the most coverage across the five — not the one that ranks highest for the original question. In one sentence: AI recommends the store with the broadest sub-query coverage, not the one with the single best page.
Why AI fans out (freshness, specificity, triangulation)
Fan-out isn't stylistic — it solves three problems single-query retrieval can't.
Freshness. A single retrieval pass can only pull from what the query vector matches. Shopper questions often combine a stable concept (lactose intolerance) with a volatile one (which brands launched lactose-free products this year). Fan-out runs one sub-query against the stable, high-confidence index and another against fresher web search, then combines them. Single-shot retrieval can't do that.
Specificity. The broad buyer question is usually too vague to match any single page well. "Best protein powder for lactose intolerance" isn't in any product title. Fan-out decomposes it into concrete attributes that actually appear on pages — "under 0.5g lactose per serving," "whey isolate," "NSF Certified." Each sub-query matches precise content. The synthesis step is what lets the AI sound like it answered the vague question while retrieving against specific ones.
Triangulation. AI engines don't trust any single source, especially a brand's own site. Fan-out pulls corroborating passages from independent sources — product page, Reddit, Trustpilot, review articles — and synthesizes only claims that show up in more than one. This is why a brand can rank #1 on-site and still not get cited: the AI couldn't triangulate externally. See off-site signals for AI recommendations for which external sources get weighted.
How to see fan-out in action (free test)
You can watch this happen in about five minutes. Perplexity is the easiest engine to see fan-out on because it shows its work — every answer lists the sources it retrieved, often grouped by the sub-query that fetched them. ChatGPT with web search and Gemini do the same thing internally but show less of it.
- Open Perplexity in an incognito window. Switch on "Pro Search" if you have it — that's the mode that exposes the fan-out steps. Ask a category shopping question in your own vertical — not your brand name, a buyer-shaped question. "What's the best [category] for [specific need]?"
- Read the answer, then scroll to the sources list. You'll see 5-12 cited sources, often grouped by the sub-query that fetched them.
- In Pro Search, expand the "Steps" section above the answer — Perplexity shows the exact search strings it expanded your question into. This is the fan-out, exposed. (Free Perplexity hides this; if you only have free, you'll see the source list but not the sub-query strings.)
- Count how many of the sub-queries your own store appears in. Zero? One? Five? That's your sub-query coverage score for this intent.
- Repeat with three different buyer questions in your category. The patterns repeat. Your gaps are usually the same across questions — titles too broad, descriptions without numbers, no third-party review density.
Pair this with the brand-awareness diagnostic in how to test what ChatGPT knows about your store. The two tests answer different questions — brand-awareness asks "does AI know we exist"; fan-out asks "for the questions we want to be cited on, how many sub-queries do we cover." You need both.
Three reasons your products miss sub-queries
When we audit stores that score well on on-page SEO but show up poorly in AI answers, the gap almost always traces to one of three patterns. Each maps to a specific sub-query type the store fails to match.
Titles too broad. A title like "Premium Whey Protein Powder" matches the broad question but none of the sub-queries. "Lactose-free" isn't in it. "Whey isolate" isn't distinguished from "whey concentrate." No variant, no size, no attribute. Fan-out's sub-queries are attribute-shaped ("whey isolate lactose content"), and a title with no attributes retrieves for none of them. The fix is the Merchant-Center-style title pattern — [Brand] [Product Type] [Key Attribute] [Variant] [Size]. See the Shopify product title format guide. Every attribute in the title is a sub-query you become eligible for.
Descriptions without specific attributes. Even with a tight title, a description written as brand voice — "When it comes to finding the right protein…" — contains no retrievable claims. The specificity sub-queries ("under 0.5g lactose," "NSF Certified") retrieve passages with specific numbers and named certifications. Vague prose matches nothing. The fix is the Claim → Evidence → Qualifier pattern in product descriptions AI will quote.
No FAQ chunks. A large share of sub-queries are question-shaped: "is [product] good for [condition]," "how does [product] compare to [alternative]," "can [demographic] use [product]." These match FAQ-format content far better than prose descriptions, because the question itself embeds close to the sub-query vector. A store with no FAQ schema and no Q&A on product pages is invisible to the question-shaped half of the fan-out. Fix: structured FAQ content on the page, wrapped in FAQPage schema — four to six Q&A pairs per product, answers written as standalone chunks.
How to diagnose your sub-query coverage (20 minutes, no tools)
You can score your own coverage in twenty minutes. No tools — a pen, a doc, one browser tab. The goal is to see which sub-queries your top product can answer and which it can't. Run this once per top SKU; the patterns repeat so you don't need to cover your whole catalog.
- Pick one SKU. Your top seller. Write the broad question a buyer would ask an AI to find it — "What's the best [your product] for [your audience]?" Not a branded query; a problem-shaped one.
- Brainstorm five sub-queries. Imagine an AI engine expanding the broad question. One for brand names, one for a technical attribute, one for verification, one for reviews, one for an adjacent preference. Write them down.
- Search your product page. Cmd-F each sub-query phrase. Does the page contain the exact words, the specific number, the named certification? If nothing matches, the page isn't retrievable for that sub-query.
- Search your whole domain. For any missed sub-query, try
site:yourdomain.comfor the phrase. FAQ pages, blog posts, and collection pages all count. Nothing on your domain containing the phrase is a gap. - Check off-site. Google the sub-query plus your brand. Does any third-party source — reviews, press, Reddit — contain both? External corroboration is what triangulation sub-queries retrieve.
- Score 0-5. Anything under 4 means the synthesis step is pulling more from competitors than from you on this intent.
- Fix the biggest gap first. Missed verification? Add a certification line. Missed reviews? Wire up post-purchase review collection. Missed an attribute? Rewrite the title. One gap per week beats a fifty-item audit you never start.
For Shopify merchants: the four touch points
Sub-query coverage on Shopify comes down to four on-page touch points, plus off-site reinforcement. These are the same levers the tactical guides in this series cover in depth — the insight is that they're not four separate projects, they're four slots in a single fan-out, each winning a different kind of sub-query.
Product titles in Merchant-Center format — [Brand] [Product Type] [Key Attribute] [Variant] [Size] — win attribute sub-queries. Every attribute in the title is a sub-query you become eligible for. Covered in the title format guide.
additionalProperty schema — JSON-LD entries for material, dimensions, weight, allergens, certifications — wins structured-attribute sub-queries. Machine-readable attributes don't require parsing prose. See the additionalProperty schema guide.
FAQPage schema — on-page Q&A paragraphs wrapped in FAQPage schema — wins the question-shaped sub-queries that make up roughly a third of any fan-out. One FAQ block per product page, four to six Q&A pairs, is the cheapest coverage gain available.
Product descriptions in Claim → Evidence → Qualifier shape win the specificity sub-queries that ask for exact numbers and named certifications. Each retrievable paragraph matches one sub-query; five good paragraphs cover five.
Off-site is the fifth lever on top of all four — external sources are what triangulation sub-queries need to corroborate your claims. On-page wins eligibility; off-site wins trust.
StoreAudit's Full AI Audit already runs a Query Deep Dive: it generates realistic buyer questions, captures live AI answers across enabled providers, checks organic search and Shopping evidence, and reports where your store appears, where competitors outrank you, and which fixes map back to the missed query. The longer-term roadmap is to expand this into full fan-out coverage: for each simulated question, expose the 3-7 sub-queries the engine fanned into, score your coverage, and map gaps back to the four touch points. Until that ships, the 20-minute manual diagnostic above gets 80% of the signal. Pick one top SKU, fix the biggest gap, repeat next week.
Single-query rank tracking is obsolete for AI. The unit of measurement is sub-query coverage across a buyer intent — and the unit of work is making sure every one of the five retrievals in a fan-out finds something from your store worth quoting. Run a free audit on your store →