DEV Community

Cover image for ChatGPT vs Claude vs Gemini: How Each AI Actually Searches the Web (With Real Data)
William C.
William C.

Posted on

ChatGPT vs Claude vs Gemini: How Each AI Actually Searches the Web (With Real Data)

Not all AI platforms search the same way. After intercepting the actual web requests from ChatGPT, Claude, and Gemini across hundreds of sessions, I have hard data on how each platform discovers, evaluates, and cites web content.

The differences are significant — and if you're trying to get your content cited by AI, you need to understand them.

The Experiment

I built a Chrome extension that overrides window.fetch to intercept the real search queries and source URLs from each AI platform. No API simulation — this captures the actual network requests and Server-Sent Event streams.

For the study: 500+ browsing sessions, same questions asked across all three platforms, data collected from February 2025 to February 2026.

The Core Numbers

Metric ChatGPT Claude Gemini
Avg queries per prompt 8.2 5.4 6.8
Sources consulted 14 8 11
Sources cited 4 3 5
Cite/consult ratio 28% 37% 45%
Reformulation Gap 52% 38% 44%
Avg response time 12s 8s 10s

Every platform approaches search differently. Let me break down what these numbers mean.

ChatGPT: The Aggressive Researcher

ChatGPT is the most aggressive searcher. It generates the most queries (8.2 per prompt), consults the most sources (14), but has the lowest citation rate (28%). It reads a lot but credits very few.

Search Engine

ChatGPT uses Bing as its search backend. This has implications:

  • Content needs to be indexed in Bing (not just Google)
  • Bing Webmaster Tools is critical for ChatGPT visibility
  • Bing's ranking algorithm differs from Google's — it weights social signals more heavily

Query Strategy

ChatGPT is an aggressive reformulator (52% Reformulation Gap). It takes a simple question and expands it into a research project:

User: "How do I optimize images for the web?"

ChatGPT generates:

  1. web image optimization best practices 2026
  2. WebP AVIF format comparison performance
  3. lazy loading images JavaScript implementation
  4. responsive images srcset sizes attribute
  5. image CDN cloudflare cloudinary comparison
  6. core web vitals LCP image optimization
  7. image compression quality vs file size benchmark
  8. next-gen image formats browser support 2026

Eight queries for a simple question. It's thorough but unfocused.

Citation Behavior

ChatGPT tends to cite:

  • High-authority domains (MDN, official docs, major publications)
  • Pages with clear, extractable answers near the top
  • Recent content (strong recency bias)

It tends to skip citing:

  • Forum threads (even if it heavily uses them)
  • Pages that are informative but lack specific data points
  • Content behind heavy ad/popup layouts

Claude: The Selective Scholar

Claude is the most conservative searcher. Fewer queries (5.4), fewer sources consulted (8), but the highest cite-per-search efficiency. When Claude finds a good source, it's more likely to credit it.

Search Engine

Claude uses its own internal search infrastructure. The queries are sent to Anthropic's backend, which handles web search differently from Bing or Google.

Query Strategy

Claude has the lowest Reformulation Gap (38%). It stays closer to the user's original intent:

User: "How do I optimize images for the web?"

Claude generates:

  1. image optimization web performance
  2. modern image formats WebP AVIF
  3. responsive images implementation guide
  4. image compression tools comparison
  5. lazy loading images best practice

Five focused queries. No tangential exploration.

Citation Behavior

Claude has a distinct citation style:

  • Cites fewer sources but gives them more context
  • More likely to cite niche/specialized sources over general authority
  • Prefers technical documentation and primary sources
  • Less recency bias than ChatGPT — quality over freshness

SSE Format

For developers: Claude uses standard SSE format with input_json_delta chunks. It's cleaner to parse than ChatGPT's JSON Patch operations:

event: content_block_delta
data: {"type":"content_block_delta","delta":{"type":"input_json_delta","partial_json":"..."}}
Enter fullscreen mode Exit fullscreen mode

Gemini: The Balanced Citator

Gemini sits between ChatGPT and Claude on most metrics but has the best cite/consult ratio (45%). Nearly half the sources it reads get cited.

Search Engine

Gemini uses Google Search — the same index that powers traditional Google search results. This means:

  • If you rank on Google, you're already in Gemini's search pool
  • Google Search Console data is directly relevant to Gemini visibility
  • Gemini inherits Google's quality signals (E-E-A-T, Core Web Vitals)

Query Strategy

Gemini has a unique approach: it runs queries in two phases.

Phase 1 (fast): 2-3 broad queries appear quickly
Phase 2 (delayed): 3-5 more specific queries appear after initial results are processed

This dual-phase approach means Gemini's later queries are influenced by what it found in the first batch. It adapts its research based on early results.

Citation Behavior

Gemini is the most generous citator:

  • Higher citation rate across all domain authority levels
  • More likely to cite multiple competing sources on the same claim
  • Includes "also see" style references more frequently
  • Better at citing code-heavy and technical content

Technical Architecture

Gemini uses Service Workers and Web Workers for its web requests. This is important for developers because it means standard window.fetch interception doesn't work. The requests bypass the main thread entirely.

To intercept Gemini's queries, you need to hook into the Service Worker registration or use different interception techniques — a significantly harder technical challenge.

What This Means for Your Content

If You Want ChatGPT Citations:

  1. Get indexed in Bing (submit via Bing Webmaster Tools)
  2. Front-load your best content — ChatGPT scans many pages quickly and decides fast
  3. Include specific data points — numbers, benchmarks, dates
  4. Update content frequently — ChatGPT has strong recency bias

If You Want Claude Citations:

  1. Be the primary source — Claude prefers original research over aggregation
  2. Write in-depth — Claude reads deeper into content than ChatGPT
  3. Technical accuracy matters — Claude seems to evaluate factual consistency
  4. Niche expertise wins — Claude cites specialized sources more readily

If You Want Gemini Citations:

  1. Optimize for Google — Gemini uses Google's search index
  2. Include code examples — Gemini cites technical content at higher rates
  3. Cover topics comprehensively — Gemini's two-phase search rewards thorough content
  4. Schema.org markup — Gemini, being Google, weights structured data heavily

Universal Strategies:

  • Schema.org markup increases citations across all platforms
  • Author credibility signals (bio pages, credentials) help everywhere
  • Extractable, specific claims beat vague statements on all platforms
  • AI crawler access in robots.txt is a prerequisite for all

How I Collected This Data

All data was collected using AI Query Revealer, a Chrome extension I built that intercepts the actual network requests from each platform. It works by:

  1. Injecting a MAIN world content script that overrides window.fetch
  2. Parsing each platform's specific streaming format (JSON Patch for ChatGPT, standard SSE for Claude, Service Worker interception for Gemini)
  3. Extracting queries, source URLs, and citation decisions from the stream
  4. Calculating metrics like Reformulation Gap and cite/consult ratios

Everything runs client-side. No data leaves your browser.

The Bottom Line

There's no single "best" AI platform for citation. Each has its own search strategy, citation logic, and technical architecture. The platforms that cite your content depend heavily on:

  • Where you're indexed (Bing vs Google vs Anthropic's crawler)
  • How your content is structured
  • Whether you're a primary source or aggregator
  • How recently you've updated

The landscape is still shifting. These platforms update their search behavior regularly, which is why I built the monitoring tooling — to track when things change.


Which AI platform cites your content the most? Any surprising differences you've noticed? I'm particularly interested in hearing from people in specialized niches where citation patterns might differ from the mainstream.

Top comments (0)