Wednesday, February 25, 2026

Visibility Is an Engineering Problem

7 min read

Visibility Engineering

The search landscape didn't just shift in the past year. It fractured.

Google AI Overviews peaked at roughly 25% of searches in mid-2025 and continue to appear on a significant share of queries. Over 60% of searches now end without a click. On queries where AI summaries show up, organic click-through rates have dropped by 61%. And that's just Google. ChatGPT, Perplexity, and Gemini are pulling answers from the web and serving them directly, bypassing your site entirely.

The SEO industry's response? Write more content. Publish faster. Add another blog post. Spin up an AI content pipeline and flood the index.

That response misses the point entirely.

The content-first fallacy

For years, the SEO playbook was straightforward: research keywords, write content, build links, rank. It worked because Google's algorithm was fundamentally a document retrieval system: match a query to the best document.

That model is breaking down.

AI-powered search doesn't retrieve documents. It synthesises answers. It retrieves passages from multiple sources, weighs them against its training data, and generates a response. Research shows that the content most likely to be selected tends to be semantically complete and entity-rich. Not because AI explicitly checks a rubric, but because comprehensive, well-structured content produces stronger retrieval signals. Your perfectly optimised 2,000-word blog post doesn't compete with other blog posts anymore. It competes with the combined knowledge of the entire web.

In this environment, producing more content without fixing the underlying architecture is like pouring water into a leaking bucket. No amount of content will compensate for a foundation that can't support visibility.

What actually drives AI citations

Recent research has mapped the factors that determine whether AI systems cite your content. The findings should make every content-first SEO team uncomfortable.

Semantic completeness is the single strongest predictor of whether an AI system will cite a page (correlation r=0.87). Content that scores 8.5 out of 10 or higher on semantic completeness is 4.2 times more likely to be cited. This isn't about word count or keyword density. It's about whether your content fully covers the entities, relationships, and concepts that define a topic.

Entity density matters more than keyword density. Pages with 15 or more recognised entities show a 4.8 times higher probability of being selected by AI systems. Entities are people, organisations, concepts, locations, and products: the building blocks of a knowledge graph, not the building blocks of a keyword list.

Structured data is the bridge. Google's AI Overviews pull from its existing search index, where structured data directly influences which pages are selected as sources. For other AI systems like ChatGPT and Perplexity, well-structured content with clear semantic markup makes information easier to extract, even when those crawlers process it as raw HTML rather than parsing schema directly.

Server performance correlates with citations. An SE Ranking study of 129,000 domains found that the fastest sites average 6.7 AI citations, while the slowest drop to 2.1. AI crawlers don't render pages like browsers do. They fetch raw HTML. What matters is server response time: if your server is slow to respond, crawlers operating on a processing budget move on.

Freshness is non-negotiable. Multiple analyses suggest content not updated quarterly is significantly more likely to lose AI citations. AI systems favour recency because they're trying to give accurate, current answers, and content with recent timestamps tends to get retrieved over stale alternatives.

None of these are content strategy problems. Every single one is an engineering problem.

The architecture gap most businesses don't see

Here's what most businesses get wrong about AI-era search visibility: they think the problem is on the page. It's not. The problem is underneath the page.

Consider what happens when an AI system encounters your website:

Your content enters AI systems through multiple paths, and each one depends on technical foundations.

Google AI Overviews pull from Google's existing search index, built by Googlebot. If your pages aren't properly crawlable and indexed by Google, they can't appear in AI Overviews either. ChatGPT Search relies primarily on Bing's index, with occasional real-time browsing via GPTBot. Perplexity crawls pages directly in real time with PerplexityBot. And all major AI models are trained on web-scale datasets where the most technically accessible, well-structured content gets the cleanest signal.

What these paths have in common: none of them execute JavaScript. AI crawlers like GPTBot and PerplexityBot fetch raw HTML. They operate more like a 2010 search bot than a modern browser. Content rendered client-side via React, Angular, or Vue without server-side rendering is invisible to them.

Most websites fail here. Their content is locked behind JavaScript rendering, buried in flat URL structures, or presented without any structured data. The crawler, whether Googlebot indexing for AI Overviews or GPTBot fetching for ChatGPT, can't extract what it needs. The competitor whose technical foundation makes extraction easy becomes the cited source instead.

This is the architecture gap. And no amount of new content will close it.

Why "more content" is the wrong response

The industry's reflex toward content volume is understandable. For fifteen years, more content meant more indexed pages, which meant more ranking opportunities. But the maths have changed.

AI Overviews aren't selecting the best document for a query. They're synthesising from multiple sources. Being one of those sources requires something fundamentally different from ranking number one in a traditional SERP.

Here's the uncomfortable truth: 85% of brand mentions in AI responses originate from third-party pages, not your own website. On some AI platforms, community sources dominate: Perplexity draws nearly half its citations from Reddit, while Wikipedia accounts for close to half of ChatGPT's. Your owned content alone isn't enough. What matters is how the broader web talks about you, structured by the technical signals you control.

Meanwhile, Google's December 2025 core update continued the trend of penalising surface-level content. Wikipedia, the internet's most prolific content producer, lost over 435 visibility points, making it the biggest loser. The signal is clear: volume without depth, structure, and genuine authority is a liability, not an asset.

What an engineering-first approach looks like

Treating visibility as an engineering problem means starting from the foundation, not the surface. It means asking different questions:

Instead of "What keywords should we target?" ask "What entities does our business need to own in the knowledge graph?" Entity coverage determines whether AI systems associate your brand with a topic. Without entity authority, even perfectly written content gets overlooked.

Instead of "How many blog posts do we need?" ask "Is our site architecture designed for how AI systems parse and extract information?" This means evaluating your URL structure, internal linking architecture, structured data implementation, and whether your content is organised in a way that machines, not just humans, can navigate.

Instead of "How do we get more backlinks?" ask "Are we building citable assets that AI systems can reference?" Citable assets are original research, proprietary data, and structured knowledge bases that AI systems can extract specific facts from. They are fundamentally different from generic thought leadership.

Instead of "How do we rank for this query?" ask "How do we get cited across Google, ChatGPT, Perplexity, and Gemini simultaneously?" Optimising for one surface is no longer enough. AI crawlers from different platforms weight different signals. A unified strategy built on solid technical architecture works across all of them.

The compounding advantage of systems over deliverables

There's a second dimension to the engineering approach that matters: systems compound, deliverables don't.

A blog post is a deliverable. Once it's published, its value starts decaying. An AI content pipeline with quality scoring, entity enrichment, and automated freshness monitoring is a system. It compounds over time because every new piece of content benefits from the infrastructure already in place.

A one-off technical SEO audit is a deliverable. A structured data implementation that automatically generates JSON-LD for every page type (products, services, locations, people) is a system. It scales with the business rather than requiring proportionally more effort.

This distinction matters because the businesses winning in AI-era search aren't doing more work. They're building better systems. Their technical foundation handles the complexity, freeing them to focus on genuine expertise and original insight: the things AI systems actually value.

The measurement problem (and why it matters)

One more reason the industry's response has been wrong: we're measuring the wrong things.

Traditional SEO metrics (rankings, impressions, click-through rate) assume a deterministic search environment. You rank third, you get a predictable percentage of clicks. But AI search is probabilistic. You might be cited in one AI Overview and not in the next, even for the same query.

The new metrics that matter are:

  • Share of voice across AI surfaces: how often does your brand appear in AI-generated responses relative to competitors?
  • Citation rate: what percentage of relevant AI responses cite your content?
  • Entity coverage: how many topic-relevant entities does your content cover compared to the sources AI systems currently prefer?
  • AI referral traffic quality: early data from Semrush suggests AI search traffic converts at 4.4 times the rate of traditional Google traffic. The volume is still small (under 1% of total web traffic), but the intent signal is significantly stronger. Are you positioned to capture it as it grows?

These metrics require different tooling, different infrastructure, and a fundamentally different approach to optimisation. They can't be bolted onto a content-first strategy. They need to be built into the technical foundation from the start.

The bottom line

The SEO industry has spent the past 18 months responding to the AI shift with more content. The data says that response is failing. Google search traffic to publishers dropped by a third globally in 2025. Zero-click searches keep climbing. And the businesses that are winning aren't producing the most content. They're building the best technical foundations.

Visibility in 2026 is an engineering problem. It requires architecture that AI systems can parse, structured data they can extract, entities they can map, and systems that compound over time. It requires thinking about search as a multi-surface challenge where Google, ChatGPT, Perplexity, and Gemini all matter.

More content isn't the answer. Better engineering is.

Frequently asked questions

Related glossary terms

Entities

Entities in SEO are uniquely identifiable, well-defined concepts that search engines recognise through structured knowledge bases, enabling semantic understanding rather than keyword matching.

Structured Data

Structured data in SEO/GEO is standardized Schema.org markup that enables search engines and AI systems to understand page content, creating rich results and improving AI citation accuracy.

E-E-A-T

E-E-A-T stands for Experience, Expertise, Authoritativeness, and Trustworthiness: Google's quality assessment framework used by human raters to evaluate content credibility, particularly for YMYL topics.

Keep reading

More articles about generative engine optimisation.

The Ranking Apocalypse: Why Position One Means Nothing in AI Search
Tuesday, March 24, 2026
Traditional SEO rankings have collapsed as predictors of AI visibility, with only 37% of AI Overview citations now coming from top-10 pages.
Jenoff Van Hulle
Entity Optimisation for AI Overviews: Staying Visible in Breaking News
Monday, March 23, 2026
When AI Overviews do appear for news queries, the publications that get cited aren't necessarily the top-ranking ones in traditional search.
Jenoff Van Hulle
AI Search
Tuesday, February 3, 2026
AI referral traffic grew 357% in a single year. Traditional SEO still matters, but it's no longer enough. Learn how AI systems find, trust, and cite content, and what your brand must do to stay visible across ChatGPT, Gemini, Perplexity, and Google AI Overviews.
Jenoff Van Hulle