AI Overviews don't rank pages — they extract claims. Google's generative search, ChatGPT's web mode, Perplexity, and Claude's search tool all pull short, attributable snippets from a small set of source pages and stitch them into a synthesized answer. Understanding which pages get pulled and why is the entire game in generative engine optimization. After building the GEO Crash Test and running it against hundreds of URLs, five factors do most of the work — and a sixth, structured data coverage, sits underneath all of them as the technical foundation.

The shift: from ten blue links to a handful of citations

Traditional SEO optimized for a ranked list. A user saw ten results and chose one. AI Overviews collapse that interface. The model produces one answer, and a small number of sources — usually three to seven — get cited as the supporting evidence.

This changes what "winning" means. You're no longer competing for position one. You're competing to be one of the few pages a language model deems worth extracting from. The bar is different, and most SEO playbooks haven't caught up.

The shift is also why zero-click rates are climbing across the board. When the answer is rendered in the interface, the citation is courtesy, not destination. Your job is to make sure that courtesy mentions you by name.

Factor 1: Extractability

The single most important property of a citable page is that a model can lift a clean, self-contained claim out of it without context collapse.

This sounds obvious. It isn't. Most content is written as an argument — a thread of reasoning where paragraph three depends on paragraph two. Models don't extract arguments. They extract sentences and short paragraphs that stand alone.

Pages built like that have a recognizable shape: declarative opening sentences, one idea per paragraph, definitions before elaboration, and a willingness to state the conclusion before showing the work. This is closer to encyclopedia writing than essay writing, and it's a real adjustment for content teams trained on storytelling.

When I rewrote my own articles to lead with the definition in the first 60 words, citation frequency in test queries climbed noticeably.

Factor 2: Entity authority

AI systems don't cite URLs in a vacuum. They cite sources, and a source is an entity — a person, an organization, a publication — with a stable identity across the web.

Entity authority comes from three places: structured data that declares who you are (Organization and Person schema), consistent cross-platform identity (the same name, same bio, same links on LinkedIn, GitHub, Substack, your site), and sameAs properties that connect those identities into a single graph.

Pages from entities the model can resolve get cited more often than pages from entities it can't. This is why a no-name blog and a personal site with full schema can outperform a content-farm page that has more backlinks. The model is asking "who is this?" before "what does this say?"

If your structured data is incomplete, you're invisible at the entity layer no matter how good the content is.

Factor 3: Topical depth

Models prefer sources that demonstrate depth on the specific topic of the query. A site with one article on a topic looks accidental. A site with a cluster — pillar piece, supporting articles, a tool, a methodology page, cross-links between them — looks like an authority.

This is the same logic as topic clusters in traditional SEO, but the consequence is sharper. With ten ranked results, a thin source could still appear. With three to seven citations, thin sources get filtered out in favor of pages that sit inside a recognizable expertise neighborhood.

The practical implication: don't publish one article on a topic you want to own. Publish four. Link them. Give the model a coherent neighborhood to point at.

Factor 4: Freshness signals

AI Overviews lean toward content that signals it's alive. Recent dateModified values, current data, references to recent developments, and active publishing cadence on the surrounding site all push a page up the citation list.

This isn't about chasing news. It's about not being a tombstone. A definitional article on GEO from 2022 will lose to a definitional article from 2026 even if the older one is technically more authoritative, because the model has no way to know the older claims still hold.

The cheapest version of this work is a quarterly content refresh — update the date, add one new paragraph reflecting what's changed, re-publish. It's a low-effort signal that does meaningful work.

Factor 5: Originality of frame

The factor that's hardest to game and the one that matters most over time: did this page contribute a frame, a definition, a number, or a methodology that didn't exist before it?

Aggregator content — "the top 10 tips for X" — is the worst possible target for AI citation. It contains nothing the model can't synthesize from its training data plus three other sources. Why cite an aggregator when you can cite the originals?

Original frames are different. A named framework ("the 2026 GEO prioritization framework"), a specific number from original analysis, a methodology with documented steps — these are things a model has to attribute because they can't be derived without the source.

This is the moat. Tools, original research, named frameworks, and documented methodologies are durable citation assets. Most other content is replaceable.

What AI Overviews ignore

It's worth naming what doesn't move the needle.

Keyword density doesn't matter. Word count doesn't matter past the threshold of "enough to make a coherent claim." Meta descriptions don't influence citation. Backlink quantity matters less than it did for traditional ranking — quality citations from entity-resolved sources matter more than volume.

And the biggest one: traffic doesn't matter. A page with 40 monthly users can outcite a page with 40,000 if the smaller page has clearer extractability, sharper entity signals, and an original frame.

How to apply this

Audit your top five pages for the five factors. Score each one on extractability, entity authority, topical depth, freshness signals, and frame originality. The lowest score is where the work is.

Most sites discover the same thing I did: extractability and entity authority are the cheap wins, and frame originality is the long game. Start with the cheap wins. Build the moat in parallel.

If you want to measure your own pages against these factors, that's exactly what the GEO Crash Test was built for — point it at a URL and see how the citation signals score. For the definitional reference on what gets measured, see What is GEO Score?.

FAQ

How do AI Overviews decide which sources to cite?

AI Overviews evaluate five primary factors: extractability of self-contained claims, entity authority of the source, topical depth across the site, freshness signals, and originality of frame. Structured data coverage sits underneath all of them as the technical foundation.

How many sources do AI Overviews typically cite?

AI-generated answers usually cite three to seven sources. This is a much smaller competitive set than traditional search results, which is why citation in AI Overviews is more selective than ranking in a list of ten links.

What is the most important factor for AI citation?

Extractability is the single most important property. A language model has to be able to lift a clean, self-contained claim from the page without surrounding context. Pages written as long arguments are harder to cite than pages written as discrete, declarative claims.

Does backlink count still matter for AI citation?

Backlink quantity matters less for AI citation than it did for traditional ranking. Quality citations from entity-resolved sources matter more than volume. AI systems weight entity authority and structural signals more heavily than raw link counts.

Can a low-traffic page outperform a high-traffic page in AI citation?

Yes. A page with 40 monthly users can outcite a page with 40,000 if the smaller page has clearer extractability, sharper entity signals, and an original frame. Traffic is not a primary input to citation behavior.