40%
of information-seeking queries now start in an AI assistant (Gartner 2025)
Research
Industry benchmarks and data on how businesses appear in AI-generated answers.
The AI search landscape has fundamentally shifted. In 2026, an estimated 40% of information-seeking queries now begin in an AI assistant rather than a traditional search engine. This report examines how businesses across industries are adapting — or failing to adapt — to this new reality. Based on data from thousands of website audits and AI crawler interactions, we present the first comprehensive benchmarks for AI visibility performance.
of information-seeking queries now start in an AI assistant (Gartner 2025)
of B2B buyers use AI assistants during purchasing research (Forrester 2025)
weekly active users on ChatGPT alone (OpenAI 2025)
of websites score below 50/100 on AI visibility audits (Appear internal data)
average increase in AI referral traffic after optimization (Appear client data)
median time to measurable AI visibility improvement (Appear client data)
Average AI visibility scores by industry, out of 100. Based on thousands of audits run through Appear's AI visibility scoring engine.
Highest average scores due to text-heavy, well-structured documentation sites. SaaS companies tend to publish detailed product pages, technical docs, and comparison content that AI crawlers can parse effectively.
Heavy JavaScript rendering and dynamic pricing pages make most stores invisible to AI. Product descriptions are often short, duplicated across variants, and trapped inside client-side frameworks.
Regulatory content tends to be well-structured, but patient-facing sites often rely on JavaScript portals. Provider directories and appointment booking systems are typically invisible to AI crawlers.
Compliance-driven content provides good structured data foundations, but marketing sites lag. Product comparison pages and rate information are frequently rendered client-side.
Most law firm websites are template-based with thin, duplicated content across practice areas. Attorney bios and case results are rarely structured in a way AI systems can parse.
The worst-performing category. Most local business sites are JavaScript-rendered templates with minimal content. Service pages are often one paragraph with a contact form.
Listing-heavy sites built on MLS integrations that render entirely client-side. Property data, neighborhood information, and agent pages are almost always invisible to AI crawlers.
Menu and booking widgets dominate, leaving almost no crawlable content. Hours, menus, and location details are trapped in third-party embeds that AI systems cannot read.
AI crawlers are the bots that major AI platforms use to read, index, and learn from your website. Understanding their behavior is essential to managing your AI visibility.
The most aggressive AI crawler, averaging 340 requests/day to optimized domains. Respects robots.txt. Powers both ChatGPT's training data pipeline and its live browsing feature for real-time answers.
Second most active, averaging 180 requests/day. Focuses heavily on structured data and clean HTML. Claude's reasoning-first approach means it rewards pages with explicit definitions, precise language, and logical hierarchy.
Averaging 120 requests/day, with a strong preference for FAQ-structured content and citation-ready formats. Because Perplexity surfaces visible source links, page clarity directly affects whether your brand gets cited.
Growing rapidly, now averaging 200 requests/day. Benefits from existing Googlebot infrastructure. Sites that already perform well in Google's traditional index tend to have a head start with Gemini's AI surfaces.
Averaging 90 requests/day. Less transparent about data usage but increasingly important in Asian markets. Brands targeting international audiences should monitor Bytespider activity alongside the Western-focused crawlers.
The most frequent technical and content issues preventing websites from appearing in AI-generated answers.
62% of sites serve empty HTML shells to AI crawlers. When your content only appears after client-side JavaScript executes, most AI bots see a blank page instead of your product descriptions, pricing, and key messaging.
71% of business websites lack JSON-LD schema markup. Without structured data, AI systems have to guess what your business is, what you sell, and how you relate to your category — and they often guess wrong.
34% of sites have no robots.txt file, leaving AI crawlers without indexing guidance. While this doesn't block crawling, it means you're missing an opportunity to direct crawler attention to your most important pages.
45% of sites have fewer than 500 words of extractable text on key pages. AI systems need substance to work with. Pages with a headline, a sentence, and a form give crawlers nothing to cite or recommend.
18% of sites actively block AI crawlers in robots.txt, making themselves invisible by design. Some businesses do this intentionally, but many are running default configurations that block bots they actually want to be discovered by.
The data in this report comes from thousands of AI visibility audits. If you want to know how your website scores across ChatGPT, Perplexity, Claude, and Gemini — and what it would take to improve — we can show you in a live walkthrough.
Install in minutes. Keep your stack. Improve AI visibility.