Methodology
How the Index was built.
The Venture Capital AI Visibility Index 2026 analyzed 28,400 prompts across ChatGPT, Claude, Gemini, Perplexity, and Google AI Overviews, run in two independent waves to test for stability across retrieval drift and model updates.
Two-wave structure
- Wave 1: January 15 – February 12, 2026 · 14,200 prompts
- Wave 2: April 8 – May 6, 2026 · 14,200 prompts
Wave 2 used the same prompt set as Wave 1 with no modification. Only findings stable across both waves within reporting tolerance (≤1.5 percentage-point delta on firm-level Citation Share; ≤2.0 pp on source-level share) are published here. Findings unstable across waves were excluded from the Index.
Prompt design
Queries simulated real founder, LP, journalist, and analyst research behavior. Prompts included branded firm queries ("What does Andreessen Horowitz invest in?"), non-branded category queries ("Best VC firms for AI startups"), comparison queries ("Sequoia vs Andreessen Horowitz"), intent-driven queries ("Who funds Series A SaaS startups in 2026?"), partner-level queries ("Top venture capitalists in the United States"), and crisis or controversy queries. Prompts were distributed evenly across the five engines so that each engine received the same prompt mix per category.
Seven venture capital categories measured
- Generalist venture capital firms
- Named partner / individual venture capitalists
- AI and machine learning investing
- Crypto and Web3 investing
- Seed, accelerator, and pre-seed
- Growth, crossover, and late-stage
- Sector-focused investing (biotech, fintech, climate, defense)
Sampling
Each prompt was issued three times per engine within a wave, with responses sampled at varied time-of-day windows to reduce within-engine retrieval drift. Reported Citation Share values represent the average of all retrieved responses for a prompt across both waves.
What "Citation Share" means operationally
Three distinct response signals were tracked per retrieved answer:
- Mention — the entity is named anywhere in the response
- Recommendation — the entity is named as a suggested or preferred option
- Source citation — the entity (or its owned domain) is cited as a reference
For the headline metric reported as Citation Share, all three signals were weighted equally per retrieved response. Source-level analyses (the publisher rankings in Figure 01) used source citations only.
Cross-engine normalization
Each engine was weighted equally in aggregate figures, regardless of differences in response length or default-citation frequency per engine. This was a deliberate choice: weighting by raw citation volume would have over-weighted Perplexity and Google AI Overviews, both of which return more citations per response than ChatGPT, Claude, or Gemini by default. Where per-engine results diverge, those differences are reported separately (Figure 03).
Retrieval vs training-data signal
This benchmark does not separately attribute citations to retrieval (live web search at query time) versus training data (pre-trained associations). The two are increasingly entangled in production AI assistants and not externally observable. Where engine-level behavior differs from a pure retrieval-only baseline (Figure 03), that variance is interpreted as a combination of retrieval architecture, training-data composition, and engine-specific ranking heuristics.
Limitations
Results reflect sampled outputs during a defined testing window. AI models, training data, retrieval indexes, and ranking systems evolve continuously; results may shift outside the test period. The Index is best read as a structured snapshot of observed system behavior across two waves — not as a continuous live measurement. The full prompt set, per-engine response logs, and category-level datasets are available on request for replication.