An AEO audit is a pass/fail review of whether AI search engines can crawl your B2B site, extract clean answers from your pages, parse your schema, verify your authority, and find your brand mentioned off-site. This checklist gives you 30 specific checks across five categories, each with a 60-second test. According to Forrester's 2026 Buyer Insights, 94% of B2B buyers now use GenAI for self-guided research, so any check you fail is leaking pipeline to a competitor that passes it.

How do I audit my B2B site for AI search visibility?

Run this 30-point checklist top to bottom. Each item is binary: pass or fail, with a method to verify in 60 seconds or less.

The audit is organized into five categories that mirror how a citation actually happens:

  1. Crawlability -- can AI bots reach your pages at all?
  2. Extractability -- is your content shaped so a model can lift a clean answer?
  3. Schema -- are you sending machine-readable signals?
  4. Authority -- do AI engines trust the source?
  5. Distribution -- are you mentioned in the places AI engines weight heavily?

Tools you'll need: a browser, Google's Rich Results Test, one logged-out ChatGPT or Perplexity tab, and 90 minutes. No engineering help required for ~70% of the checks. Score each check 1 (pass) or 0 (fail), tally by category, and you have a heat map of where to spend the next sprint.

What does the 30-point AEO audit checklist cover?

The 30 checks map to the five citation factors that AI engines weight most heavily in 2026. Distribution by category:

Category # of checks What it tests
Crawlability 6 Robots.txt, sitemaps, JS rendering, llms.txt, status codes, WAF rules
Extractability 6 TL;DR, question H2s, section length, tables, citations, answer-first structure
Schema 6 Article, FAQPage, Organization, HowTo, Product, validation
Authority 6 Wikidata, named authors, primary citations, brand consistency, co-mentions, press
Distribution 6 Reddit, LinkedIn, YouTube, directories, citation tracking, refresh cycle

A score of 25+/30 means your site is AEO-ready. 18-24 means you have meaningful gaps. Below 18, your brand is effectively invisible to ChatGPT, Perplexity, and Google AI Overviews.

Crawlability: can AI crawlers actually read your site?

Crawlability is the first failure point. If GPTBot or PerplexityBot can't fetch your page, nothing else matters. Six checks:

1. AI crawlers allowed in robots.txt

  • Pass: GPTBot, OAI-SearchBot, ChatGPT-User, ClaudeBot, PerplexityBot, and Google-Extended are not disallowed.
  • 60-second check: visit yourdomain.com/robots.txt and search for each user-agent. No Disallow: / lines under those agents.

2. XML sitemap is live and submitted

  • Pass: sitemap returns 200, lists priority URLs, and is referenced in robots.txt.
  • 60-second check: visit yourdomain.com/sitemap.xml. Submit in Google Search Console.

3. Critical content renders without JavaScript

  • Pass: main copy is in initial HTML, not loaded via client-side JS.
  • 60-second check: right-click the page, View Source, search for a sentence from your H1. If it's there, you pass. AI training crawlers historically skip JS-rendered content.

4. Cloudflare / WAF rules aren't blocking AI bots

  • Pass: AI bot user-agents are not on a 403 / challenge list.
  • 60-second check: in Cloudflare > Security > Bots, confirm AI Crawl Control is set to allow (not block) the bots you care about.

5. llms.txt file exists at root

6. No 404 or redirect loops on priority URLs

  • Pass: every URL in your sitemap returns 200, not 301-301-200 or 404.
  • 60-second check: run Screaming Frog free version on your top 100 URLs. AI training data freezes; broken canonicals stay broken in the model.

Extractability: is your content shaped for AI extraction?

AI engines lift answers, not articles. If your page buries the answer or rambles, you don't get cited even if the bot reaches you. Six checks:

7. Direct answer in the first 50-100 words

  • Pass: the H1's question is answered in the opening paragraph, no preamble.
  • 60-second check: read the first paragraph. Does it state the answer? 90% of top-cited sources do.

8. TL;DR or summary box at the top

  • Pass: a visible summary block with 3-5 bullets above the fold.
  • 60-second check: page loads -- can you see a summary without scrolling? If not, fail.

9. Question-shaped H2s

  • Pass: every H2 reads as a question a human would type into ChatGPT ("What is X?", "How does Y work?").
  • 60-second check: outline the page. Count question-shaped H2s vs cute headings. Aim for 80%+ questions.

10. Sections are 120-180 words between headings

  • Pass: most sections fall in the 120-180 word band.
  • 60-second check: per SE Ranking's 2025 study of 129,000 domains, pages with 120-180 word sections receive 70% more ChatGPT citations. Scan for walls of text or 30-word stubs.

11. Tables for multi-attribute comparisons

  • Pass: any "X vs Y" or feature comparison uses an HTML table.
  • 60-second check: search the page for <table> or markdown pipes. AI engines parse tables cleanly; prose comparisons get skipped.

12. Statistics are sourced inline with hyperlinks

  • Pass: every number names the source, year, and links to the original.
  • 60-second check: Cmd-F for "studies show" or "research shows". Each instance is a fail. Per the Princeton GEO study, adding cited statistics boosts AI visibility ~40%.

Schema: are AI engines getting structured signals?

Schema markup is how a page tells an AI engine what it is. Done well, it lifts citation rates substantially. Done badly, it can hurt. Six checks:

13. Article schema with author, datePublished, dateModified

  • Pass: every blog post has Article JSON-LD with all three properties.
  • 60-second check: paste the URL into Google's Rich Results Test. Confirm Article appears with all dates.

14. FAQPage schema on FAQ blocks

15. Organization schema sitewide

  • Pass: homepage and footer carry Organization schema with sameAs links to LinkedIn, Crunchbase, G2, Wikidata.
  • 60-second check: View Source on homepage, search for "@type":"Organization".

16. HowTo schema on process / step-by-step content

  • Pass: any "how to do X" article uses HowTo with named steps.
  • 60-second check: Rich Results Test on a how-to URL.

17. Product schema on product / pricing pages

  • Pass: product pages carry Product schema with offers, reviews, and aggregateRating.
  • 60-second check: Rich Results Test on /product or /pricing.

18. Schema validates with no errors

  • Pass: zero errors, zero warnings on priority templates.
  • 60-second check: Rich Results Test. Per a Growth Marshal study (n=730 citations), attribute-rich schema earns a 61.7% citation rate, but generic minimally-populated schema underperforms no schema at all (41.6% vs 59.8%). Quality matters more than presence.

Authority: do AI engines trust your brand?

AI engines weight source authority heavily. ChatGPT pulls 47.9% of top citations from Wikipedia. If your brand has no entity footprint, you're not in the model. Six checks:

19. Wikipedia or Wikidata entry exists

  • Pass: a Wikidata Q-number exists for your company, with instance of, industry, founded, headquarters, and official website populated.
  • 60-second check: search Wikidata for your brand. No entry = highest-leverage fix on this list.

20. Named author with bio and credentials

  • Pass: every article has an author byline linking to a bio page with credentials, LinkedIn, and Person schema.
  • 60-second check: click an author name. Do you land on a real bio?

21. Primary sources cited inline with hyperlinks

  • Pass: every claim links to a study, doc, or report -- not another marketing blog.
  • 60-second check: open three articles, count outbound citations. Aim for 3-5 per piece. Per Princeton, adding authoritative external citations lifts visibility up to 115% for lower-ranked content.

22. Brand consistency across G2, Crunchbase, LinkedIn, Apollo

  • Pass: company name, tagline, category, and founding year match across all major B2B databases.
  • 60-second check: spot-check three platforms. Inconsistencies confuse entity resolution and tank citation rates.

23. 5+ third-party co-mentions per priority page

  • Pass: each pillar page has 5 or more third-party pages mentioning your brand alongside the topic.
  • 60-second check: Google "yourbrand" "keyword" minus your own domain. Count results.

24. Recent third-party press or industry mention

  • Pass: at least one mention from a credible publication or analyst report in the last 90 days.
  • 60-second check: news search your brand name. Per Perplexity's freshness bias, recency is weighted heavily.

Distribution: are you visible where AI engines actually look?

AI engines pull from off-site sources at high rates. 5W's 2026 AI Platform Citation Source Index found Reddit accounts for roughly 40% of citations across LLMs and ~24% of Perplexity citations alone. Owned content is half the game. Six checks:

25. Active presence in 3+ relevant subreddits

  • Pass: employees post substantive answers (not link drops) in subreddits where buyers ask category questions.
  • 60-second check: search Reddit for site:reddit.com "yourbrand". Count organic mentions.

26. LinkedIn distribution for every new article

  • Pass: every article gets a thoughtful LinkedIn post within 24 hours of publish.
  • 60-second check: check your last 10 articles vs your last 10 LinkedIn posts.

27. YouTube or podcast presence

  • Pass: at least one YouTube channel or podcast appearance per quarter where your brand is named in transcripts.
  • 60-second check: per GEORaiser, YouTube overtook Reddit as the #1 AI citation source for some categories in 2026. Search YouTube for your brand.

28. Listed in 5+ industry directories (G2, Capterra, TrustRadius, Gartner Peer Insights, Product Hunt)

  • Pass: claimed and populated profiles with reviews on each.
  • 60-second check: search each directory.

29. AI citation tracking installed

  • Pass: a tool like Profound, Otterly, or Peec.ai is monitoring your brand citation rate weekly.
  • 60-second check: if no one on the team can show this week's citation report, fail.

30. 13-week content refresh cycle in place

  • Pass: priority pages have a dateModified newer than 90 days.
  • 60-second check: sort your blog by last-modified date. Anything older than 13 weeks needs review.

Which AEO checks have the highest impact-to-effort ratio?

Five checks deliver disproportionate citation lift relative to the work involved. Hit these first:

  1. Allow AI crawlers in robots.txt (Check #1) -- 5 minutes of work, unlocks everything downstream. The most common silent killer.
  2. TL;DR + answer-first intro (Checks #7-8) -- 30 minutes per priority page; 90% of top-cited sources answer in the first 100 words.
  3. FAQPage schema on every FAQ block (Check #14) -- one template change, 60% more likely to feature in AI Overviews per Wellows.
  4. Wikidata entry (Check #19) -- 90 minutes once; permanent entity grounding.
  5. Reddit and YouTube distribution (Checks #25, #27) -- ongoing, but Reddit alone drives ~24% of Perplexity citations per 5W's 2026 Index.

A team with no engineering help can land all five in under a week. That alone moves most B2B sites from invisible to citable in ChatGPT and Perplexity within 3-5 publishing cycles.

Citation Lift by AEO Optimization (Princeton GEO Study)
Citing authoritative sources (low-rank pages)
115%
120-180 word sections (vs walls of text)
70%
Adding inline statistics
41%
Comprehensive schema markup
36%
FAQ schema in AI Overviews
60%
Source: Princeton/Georgia Tech GEO Study (2024), SE Ranking 2025, Wellows 2026

How long does a full AEO audit take?

A solo marketer can complete the 30-point audit in 60 to 90 minutes if the site is under 500 pages. Breakdown:

  • Crawlability (6 checks): 15 min -- robots.txt, sitemap, source view, WAF, llms.txt, status codes
  • Extractability (6 checks): 20 min -- spot-check 5 priority URLs against checks 7-12
  • Schema (6 checks): 15 min -- run priority templates through Rich Results Test
  • Authority (6 checks): 20 min -- Wikidata, author bios, brand consistency spot checks
  • Distribution (6 checks): 15 min -- Reddit, LinkedIn, YouTube, directory inventory

Fixing what the audit surfaces is the longer job. Per Fountain City's B2B GEO guide, full implementation typically runs 2-3 months: technical fundamentals in 1-2 weeks, content restructure in 2-4 weeks, and authority/entity work as a rolling effort. AI engines pull new content into citation pools within 3-5 business days, so quick wins compound fast.

What's the first thing to check if my brand isn't being cited by ChatGPT?

Check robots.txt first. It's the #1 silent killer of AI search visibility for B2B sites.

A surprising number of B2B marketing teams inherited a Disallow: / rule for GPTBot from a 2023 IP-protection panic, never reverted it, and now wonder why ChatGPT can't see them. Visit yourdomain.com/robots.txt right now. If you see User-agent: GPTBot followed by Disallow: /, that's your answer.

Second check: Wikidata. ChatGPT relies heavily on Wikipedia for entity grounding. No Wikidata entry, no clean entity association, lower citation probability across every prompt.

Third check: are your pages JS-rendered? If your homepage renders client-side via React/Vue with no SSR, AI training crawlers may have indexed an empty shell. View Source and search for a sentence from your H1 -- if it's not in the raw HTML, you have a rendering problem to fix before any other AEO work matters.

Which AEO checks can a marketer do without engineering help?

21 of the 30 checks are doable by a marketer with no developer support. The 9 checks that typically need engineering are flagged below:

Check Marketer-only?
#1 robots.txt config Engineering
#2 Sitemap Engineering
#3 SSR rendering Engineering
#4 WAF rules Engineering
#5 llms.txt file Engineering
#6 Status codes Engineering (audit yes, fix no)
#7-12 Content/extractability Marketer
#13-18 Schema Engineering for install, Marketer for audit
#19-24 Authority Marketer
#25-30 Distribution Marketer

A marketer can run the full audit solo, then ship a prioritized ticket to engineering for the technical fixes. Most B2B teams find that 60-70% of their citation lift comes from the marketer-doable checks: TL;DRs, question H2s, FAQ schema (via plugin), Wikidata, and Reddit distribution.

#CategoryCheck60-Second MethodPass Criteria
1CrawlabilityAI crawlers allowed in robots.txtVisit /robots.txt, search for GPTBot, ClaudeBot, PerplexityBotNo Disallow: / under any AI user-agent
2CrawlabilityXML sitemap is live and submittedVisit /sitemap.xml, check Google Search ConsoleReturns 200, referenced in robots.txt
3CrawlabilityContent renders without JavaScriptView Source, search for H1 sentenceFound in raw HTML
4CrawlabilityWAF doesn't block AI botsCloudflare > Security > BotsAI bots set to allow
5Crawlabilityllms.txt exists at rootVisit /llms.txtReturns 200 with markdown index
6CrawlabilityNo broken redirects on priority URLsRun Screaming Frog on top 100 URLsAll return 200, no chains
7ExtractabilityDirect answer in first 50-100 wordsRead opening paragraphAnswer stated, no preamble
8ExtractabilityTL;DR / summary block above the foldLoad page, look without scrollingVisible summary block
9ExtractabilityQuestion-shaped H2sOutline the page80%+ H2s are questions
10ExtractabilitySections of 120-180 wordsSpot-check section lengthsMost sections in 120-180 band
11ExtractabilityTables for multi-attribute comparisonsSearch for <table> in sourceComparisons use tables
12ExtractabilityStatistics sourced inline with linksCmd-F for 'studies show'Zero matches; every stat linked
13SchemaArticle schema with dates + authorRich Results TestArticle appears with author, datePublished, dateModified
14SchemaFAQPage schema on FAQ blocksRich Results Test on FAQ pageFAQPage detected
15SchemaOrganization schema sitewideView Source on homepage@type:Organization with sameAs
16SchemaHowTo schema on step-by-step contentRich Results Test on how-to URLHowTo detected
17SchemaProduct schema on product pagesRich Results Test on /productProduct with offers + ratings
18SchemaSchema validates with zero errorsRich Results Test on priority templatesNo errors, no warnings
19AuthorityWikidata entry existsSearch wikidata.org for brandQ-number with populated properties
20AuthorityNamed author with bio + credentialsClick author bylineReal bio page with credentials
21AuthorityPrimary sources cited with hyperlinksCount outbound citations on3 articles3-5 primary citations per piece
22AuthorityBrand consistency across databasesSpot-check G2, Crunchbase, LinkedInName, category, year match
23Authority5+ third-party co-mentions per pillarGoogle '"brand" "keyword"' minus own domain5+ results
24AuthorityRecent press / analyst mention (90 days)News search brand name1+ credible mention in 90 days
25DistributionActive in 3+ relevant subredditsSearch reddit.com for brandOrganic, substantive mentions
26DistributionLinkedIn distribution per articleCompare last 10 posts to last 10 articles1:1 ratio within 24 hours
27DistributionYouTube or podcast presence quarterlySearch YouTube for brand1+ named mention per quarter
28Distribution5+ industry directory listingsCheck G2, Capterra, TrustRadius, Gartner Peer Insights, Product HuntClaimed and populated on each
29DistributionAI citation tracking installedAsk team for this week's citation reportActive monitoring tool in use
30Distribution13-week content refresh cycleSort blog by dateModifiedPriority pages updated within 90 days