Boost your skills with Growth Memo’s weekly expert insights. Subscribe for free!

Most teams talk about “AI visibility” like it’s one thing. New data on 3.7 million citations across ChatGPT, Perplexity, and Google AI Overviews suggests it isn’t. And the gap between the three engines is wider (and more strategically important) than your dashboard likely admits.

Today’s memo breaks down:

  • Why a blended AEO score hides the only finding that matters.
  • Which page types and domains actually travel across engines.
  • The shift from measuring AI presence to measuring portability.

One of the biggest differences between AEO and SEO is that AEO plays on more platforms.

Omnia data shows across multiple samples that only 2.35% to 2.45% of cited URLs appeared in ChatGPT, Perplexity, and Google AI Overviews for the same prompt. 91% of citations appeared in only one engine.

Conclusion: AI visibility is not a single leaderboard. Instead, it’s three different distribution systems that sometimes overlap and usually do not.

Only 2% Of URLs Get Cited By All 3 Engines

Most people would guess that if a URL gets cited by one major AI engine, it has a reasonable shot at appearing in the others.

But the 20,000 prompt sample shows only 2.37% of cited URLs show up across all three engines for the same prompt.

Meanwhile, 91.07% show up in only one. Those two numbers belong next to each other because they explain each other. The remaining ~7% overlap in pairs, which means engines are drawing from largely disjoint pools rather than ranking the same pool differently.

Image Credit: Kevin Indig

For AEO/SEO teams, that means a single composite visibility score is the wrong unit of measurement. Averaged AEO scores hide this. A brand can look strong in aggregate and be invisible in 2 of 3 engines. Teams chasing one blended AI visibility number are compressing three ranking systems into one metric and calling it strategy.

The 2% Holds Across Every Cut

The ~2% overlap rate and ~91% exclusive rate stay almost perfectly flat across four samples.

Image Credit: Kevin Indig

That consistency matters more than the exact decimal point. The consensus gap is not an artifact of one query set or one time window. It looks structural.

In Q3 2025, universal overlap was 2.2%. In Q4 2025 and Q1 2026, it rose to 2.7%. Engine-exclusive citations fell from 90.1% to about 88%. So yes, a small amount of convergence. But even after that shift, fragmentation still dominates.

Commercial Prompts Don’t Converge Either

The intent split is one of the quietest but most useful parts of the dataset. You could argue that commercial queries should produce more consensus. When someone searches for [the best CRM], [best running shoes], or [best project management software], the pool of acceptable sources feels narrower than it does for broad informational prompts.

Surprisingly, the data does not support a big difference.

Image Credit: Kevin Indig

Commercial prompts show 2.4% universal overlap. Informational prompts show 2.0%. Even when the query should narrow the answer set, the engines still choose different sources most of the time.

That pushes against a common instinct in SEO and content strategy. Teams often assume high-intent queries are where shared authority will show up. The opposite looks closer to true. Even in commercial territory, each engine’s own retrieval logic, what sources it trusts, what formats it prefers, is doing most of the work.

Guides Beat Homepages By 2x

The page type breakdown below shows guides and tutorials have the highest cross-engine overlap at 2.3%, followed by blogs at 1.8%, category pages at 1.6%, product pages at 1.2%, and homepages at 1.1%.

Image Credit: Kevin Indig

Two lessons:

  1. First, explanatory content travels better than brand or transactional assets. If you want the best shot at showing up across engines, the strongest candidate is not the homepage and not the product page. It is the page that helps, explains, compares, or teaches, but keep in mind that these are also content formats that AIs can answer directly well.
  2. Second, even the best page types perform badly in absolute terms. Guides are not winning across engines in any meaningful sense. The right read on this is not “publish more guides and you will win everywhere.” It’s simpler than that: Helpful content travels better than brand content.

Visibility Is Not The Same As Portability

One of the easiest mistakes in this space is to confuse citation frequency with citation portability. Wikipedia is the cleanest example. It appears 16,073 times in the dataset, but only 1.3% of those appearances are universal across engines. Reddit appears 14,267 times, but only 0.1% are universal. Reuters shows up 1,202 times and still lands at 0.0% universal overlap.

Image Credit: Kevin Indig

That is why an important metric is portability. A domain can show up all over one engine and barely travel, which means a brand looking dominant in an aggregate dashboard may be one platform’s habit away from invisibility. Presence tells you whether you are visible. Portability tells you whether that visibility is resilient.

What This Means For Operators

The practical implication is simple: Stop treating AI visibility as one thing. Examine the comprehensive visibility of your domain by measuring:

1. Presence, the % of your tracked prompts where your domain appears in any engine. Presence tells you whether you’re visible.

2. Portability, the % of your cited URLs that appear in all three engines. Portability tells you whether that visibility is resilient.

3. Concentration, the % of your citations that come from a single engine. Concentration tells you which engine your current dashboard is secretly built on.

If the overlap between engines is this low, a single AEO strategy is too abstract to be useful.

When we approach AI visibility from a holistic perspective, it forces sharper questions:

  • Which engine matters most for us?
  • Which of our assets travel across engines, and which only work in one?
  • Are we measuring presence when we should be measuring portability?

This also changes how brand teams should think about diagnostics. A weak homepage across engines may not be a homepage problem. It is a symptom of something broader: Engines favor utility over brand centrality. In that world, visibility comes less from being the official source and more from being the useful source.

The strategic question is no longer, “How do we rank in AI?” We should instead be asking ourselves, “How do we build assets that survive different engine preferences?” That is a narrower question. It is also a better one.

Methodology

There are a few caveats to this analysis:

  • The dataset is skewed toward Omnia’s customer base.
  • The intent and page-type cuts rely on regex classification, which is useful for directional analysis but not perfect taxonomy work.

Those caveats do not weaken the main finding much. The biggest signal is not precision at the edges. It is consistency at the center. No matter how the cuts change, the same pattern resurfaces: very little overlap, very high engine-specificity, and only modest differences by time, intent, or page type.

Dataset Size And Time Window

The analysis draws on four prompt samples. Three cohorts of 5,000 prompts each, tracked from Jan. 1, 2025; July 1, 2025; and Jan. 1, 2026. A separate 20,000-prompt random sample underpins the headline 2.37% and 91.07% figures. The time-view cut spans Q3 2025 through Q1 2026 (to date) and covers 3.7 million URL citations in total. Commercial/Informational/Other intent splits are drawn from roughly 2.6 million URLs across the combined sample. Page-type splits span 4.1 million URL appearances.

How Prompts Were Selected

The 20,000 prompts are drawn as a random sample from Omnia’s live prompt monitoring pool. The pool reflects what real marketing teams chose to track, weighted toward Omnia’s customer geography (Spain-heavy, plus UK, Nordics, and other EU markets). Each prompt runs in its country’s primary language, so Spanish is overrepresented versus a U.S.-only dataset. Industry mix is fintech/insurtech, travel, SaaS, B2B services. Treat findings as directional for European AI search.

Engine Coverage

The study covers three engines: ChatGPT, Perplexity, and Google AI Overviews. Each fires the same prompt concurrently within the same minute, twice a day, with country localization, and each engine queried in its default web-enabled, unauthenticated state. Perplexity tracking runs on Sonar, while ChatGPT and Google AI Overviews use each vendor’s default production model for logged-out web browsing (which neither OpenAI nor Google pins publicly to a specific version).

Classification Methodology

Intent and page type are assigned by regex. Intent buckets are Commercial, Informational, and Other. Page-type buckets are Guide/tutorial, Article/blog, Category page, Product page, Homepage, Wikipedia, and Other. The rules are keyword- and URL-pattern-based, which makes them fast enough for a multi-million-URL dataset but coarse at the edges. Edge cases fall into Other, which is why Other carries a high share in both the intent and page-type tables. Treat the regex cuts as directional, not authoritative.

More Resources:


Featured Image: FGC/Shutterstock; Paulo Bobita/Search Engine Journal



Source link