user@None

I Didn't Know What AEO Was. Then ChatGPT Started Sending Me Traffic.

The First Time ChatGPT Showed Up in Google Analytics It was when I first saw ChatGPT as a traffic source in Google Analytics. At the time, I was not intentionally optimizing for AI, so it was an...

I Didn't Know What AEO Was. Then ChatGPT Started Sending Me Traffic.
Published Apr 17, 2026 · Updated Apr 18, 2026 · 6 min read

It was when I first saw ChatGPT as a traffic source in Google Analytics.

At the time, I was not intentionally optimizing for AI, so it was an interesting novelty, and certainly an indicator of what was to come.

After some digging, I noticed that a few product pages were the landing pages for this new traffic. Query strings came in empty. I couldn't see what people had asked ChatGPT to arrive there.

The Boring Pages Won

It was a B2B website for HVAC parts. The product catalog was very technical: long pages with lists and bullets. No UGC, no fresh content. Just boring lists of industrial parts and products. Detailed technical specs, part sizes, dimensions.

Nothing you'd put in a content marketing case study.

If you think about it: Schema tags, structured data, JSON-LD: it is really not that hard to see why AI crawlers would find that type of content. It is like ChatGPT was caught in my net of tags and structured content. Those pages were easy for machines to parse because they were factual, organized, and unambiguous. AI "sees" extractable structure more than creativity or marketing polish.

For many commercial and technical searches, the best-performing AI-source content may be the most boring page on the site. The question is whether your competitors have figured that out yet.

What I Changed After That

I had no idea what RAG, AEO, or LLM Visibility meant at the time. I just knew the machines were picking the pages where the answer was already prepared.

From that moment on, I would always try to sneak in blog posts, technical charts, tables, bullet lists. Then I worked to improve the lists that were already getting AI traffic:

  • Clearer hierarchy, consistent H1/H2 sections
  • Standardized spec tables
  • Compatibility blocks: dimensions, materials, colors
  • Installation notes and use cases

AI systems seem to list pages where the answer is already prepared.

Another thing I started doing: mixing technical content with normal text. I'd add content above and below the specs. A short plain-English summary at the top: what this part does, where it's used, who it's for, then detailed specs, then FAQs, comparisons, troubleshooting, and related models. That helps both humans and AI systems understand context.

All that content would feed very detailed schema tags and structured data wherever possible: Product, FAQ, HowTo, Organization, Breadcrumbs, availability, identifiers. The easier it is to classify, the easier it is to retrieve.

The Attribution Problem

Initially, I was following it in Google Analytics as traffic from known AI platforms: ChatGPT, Copilot, Perplexity. I could see it growing. But I was not certain if the growth was due to my recent updates or the fact that these AI platforms were starting and growing at the same time.

There was no clean control group. Query strings arrived empty. Attribution was muddy.

Only after reading more on the subject and going deeper did I start to see other possible KPIs and metrics to follow. In fact, I even created my own AI Visibility tool to help me: (a) understand it, and (b) start measuring, monitoring, and comparing pages.

It gives me crawler access, AEO score, LLM visibility, GEO information, and some technical signals. Users paste any URL to check what AI crawlers can fetch, how well the page is structured for AI citation, and how LLMs answer queries on that topic. Includes AEO scoring, GEO recommendations, sitemap discovery, and a robots.txt builder.

What an AEO Audit Actually Shows

Here is what a real audit looks like on an underperforming page: 1,404 words of content, one H2 heading, two H3s, no FAQ section, no direct answer sentence at the top.

SignalValueAssessment
Retrieval Readiness62/100Structurally weak
AEO Score42/100, Grade DPoor extraction probability
H2 headings1 for 1,404 wordsFar too flat
Words per H2 section1,404Unparseable block
FAQ sectionNoneMissing
Answer blockMissingNo entry point
Lists / tablesPresentOne strength
Structured dataMultiple typesAnother strength
JS render riskLowAccessible

The gap between Retrieval Readiness (62) and AEO Score (42) on the same page tells you something. A page can have decent raw content: lists, low render risk, good word count, and still score poorly because the structure isn't giving AI systems clean entry points to extract from.

What the audit surfaced as high-priority fixes:

  • Add a direct answer sentence in the first paragraph. AI systems seem to prefer pages that define the topic upfront.
  • Add comprehensive heading structure. Only 1 H2 for 1,404 words, that averages 1,404 words per section, far too dense for AI parsing.
  • Add an FAQ section. AI systems appear to favor FAQ content for specific questions.
  • Break long sections into more specific subsections with descriptive headings.

These are structural fixes, not content rewrites. The words were already there.

Who Is and Isn't Prepared

I worked inside large organizations long enough to know how content decisions actually get made. Homepage layout: a committee of editorial, upper management, and IT. Marketing usually not in the room. 

Nobody asking whether a crawler could read the page, because nobody was thinking about crawlers beyond Google. AI visibility wasn't on the radar then. For most of those organizations, it still isn't. The decision-making structure hasn't changed. The crawlers have.

I ran Amazon through the same tool: a standard product page, a bamboo cutting board set. Score: 43 out of 100.

Not because the page is poorly written. Amazon product pages have reasonable structure, decent word count, lists, and tables. The problem is their robots.txt. Many major AI crawlers are blocked. (as of 04/17/2026)

AI CrawlerAmazon Access
GPTBotBlocked
ChatGPT-UserBlocked
ClaudeBotBlocked
Google-ExtendedBlocked
PerplexityBotBlocked
CCBotBlocked
BytespiderBlocked

This is prioritization, not ignorance. Amazon has its own AI strategy and its own data. They don't need ChatGPT sending traffic to product pages they'd rather control directly. Fair enough.

But Amazon's decision reveals something about everyone else.

The companies most at risk are not the giants. The giants are either making deliberate choices or are already embedded deeply enough in LLM training data that brand-level citations happen regardless. Their brand authority predates the optimization problem.

The vulnerable ones sit in the middle: companies with real content, real web presence, and real marketing budgets, but no visibility into whether AI systems can actually read their pages. Big enough to have slow, committee-built websites. Not big enough to have brand authority baked into model training data.

AI wasn't on the radar inside large organizations. Legal owned content review. IT owned the CMS. Nobody had mapped what the crawlers could and couldn't reach. For most organizations, that's still true. 

And the companies that move first on structure will be the ones that own those citations when the market catches up.

For smaller businesses, the problem is often more fundamental: thin content, no schema, aging WordPress installations. AEO strategy is premature when the foundation isn't there.

The companies that should be paying attention now are the ones in between.

What to Measure If You're Starting From Zero

SignalWhat to CheckGoodPoor
Crawler accessrobots.txt allows AI botsAll major crawlers allowedAny blocked
Heading structureH2 count vs word countRecommended:
1 H2 per 300-400 words
1 H2 per 1,400+ words
Answer blockDirect answer in first paragraphPresent and specificMissing or buried
FAQ sectionStructured Q&A on pagePresentAbsent
Structured dataJSON-LD schema types presentProduct, FAQ, HowToNone detected

None of these require a new content strategy. Most can be addressed by restructuring pages that already exist.

The content is often there. The structure usually isn't. Every week without these fixes is a potential missed opportunity for AI citations.

The Underlying Pattern

Most updates that improved AI citation probability on those HVAC pages also improved the page for human readers. Clearer structure, faster access to the answer, less ambiguity about what the page covers.

Differently than Google and the elusive “Content Quality” guidelines, I believe AI systems are focused less on reading your content for its quality. They are scanning it for structure, specificity, and extractability. Pages that prepare the answer in advance get cited. Pages that make the reader work to find it don't.

Structure precedes everything else.

Frequently Asked Questions

What is AEO (Answer Engine Optimization)?
AEO refers to structuring web content so that AI systems, chatbots, and generative search engines can extract and cite specific answers. Unlike traditional SEO, which targets ranking positions in search results, AEO targets citation probability in AI-generated responses. The core variables are content structure, schema markup, heading hierarchy, and the presence of direct answer formats.
How is AEO different from traditional SEO?
Traditional SEO optimizes for crawler indexing and keyword relevance as signals for ranking. AEO optimizes for extractability: can an AI system isolate a useful answer from this page? A page can rank well in traditional search and score poorly on AEO if its content is unstructured, keyword-dense, or buried under navigation. The signals that matter most for AEO are heading structure, FAQ sections, answer blocks, and structured data.
Does blocking AI crawlers in robots.txt affect AEO score?
Yes, and it is the most direct AEO failure mode. If AI crawlers cannot access the page, no amount of structural optimization matters. Many sites have inadvertently blocked AI crawlers through broad Disallow directives that predate the current generation of AI search tools. Auditing robots.txt access is the first step in any AEO review.
What is the minimum viable AEO structure for a product or service page?
At minimum: a direct answer sentence in the first paragraph, at least one H2 per 400 words, a FAQ section addressing common questions, and basic structured data (Product or Article schema). These four changes address the most common AEO failure patterns without requiring a full content rewrite.
How do you track and measure AI referral traffic from ChatGPT and other AI platforms?
AI referral traffic can initially be tracked in Google Analytics by identifying traffic sources from known AI platforms like ChatGPT, Copilot, and Perplexity. However, attribution is challenging because query strings often come in empty and it's difficult to distinguish whether traffic growth comes from your optimizations or the platforms' own growth. This attribution gap is a known challenge in AI-First SEO measurement.
Why do boring technical pages rank well in AI search results?
Boring technical pages rank well because AI systems prioritize extractable structure over creativity or marketing polish. Pages with detailed technical specs, organized lists, dimensions, and standardized data are easy for machines to parse. Schema tags and structured data act like a net for AI crawlers, making factual and unambiguous content ideal for AI citation and retrieval.
What schema markup types should you use for AI SEO optimization?
For AI SEO optimization, implement schema markup types including Product, FAQ, HowTo, Organization, and Breadcrumbs, along with structured data for availability, identifiers, dimensions, materials, and use cases. The more thoroughly you classify your content with schema tags, the easier it is for AI systems to retrieve and cite it in their responses.
How do you combine technical content with readable text for AI and human audiences?
The recommended approach is to add a short plain-English summary at the top of the page explaining what the product or topic is, where it's used, and who it's for. Follow that with detailed technical specs, then layer in FAQs, comparisons, troubleshooting guides, and related models below. This sandwich structure helps both AI systems understand context and human readers find the information they need.