Tactical guide · Updated 2026-05-07

How to Rank on ChatGPT.
The 10 Factors That Actually Move Citations.

ChatGPT cites a fraction of the open web. The brands inside that fraction did ten specific things. This is the playbook — Bing index strategy, Wikipedia presence, schema, recency, prompt-audit measurement — plus a 12-point self-audit checklist.

See AEO Services
Written by Aditya Kathotia, Founder of Nico Digital · 175+ brands tracked across ChatGPT
Short answer

To rank on ChatGPT, get into Bing's index and rank top-10 there for your seed queries (ChatGPT Search retrieves from Bing). Build a Wikidata entity and earn corroborating mentions across Reddit, news, and tier-1 publishers (the sources OpenAI's training data weights heaviest). Allow GPTBot and OAI-SearchBot in robots.txt, ship a clean Organization + Article + FAQPage schema graph, restructure pillar content around H2-as-question + 40-80-word answer passages, and keep dateModified fresh. Track citation share monthly through a controlled prompt audit. Most brands see early citations inside 90 days; meaningful share-of-voice gains in 6 to 9 months.

How ChatGPT actually picks the sources it cites

ChatGPT does not have one ranking algorithm. It has two layers, and the tactics that win each are different.

Layer 1

Training-data weighting

GPT-4 / GPT-5 was pre-trained on a curated open-web corpus. Sources that appear repeatedly across high-trust domains (Wikipedia, Reddit via the OpenAI-Reddit deal, mainstream news, .edu, .gov, GitHub, large industry publishers) become high-confidence facts the model emits without retrieval.

How to win this layer: Wikidata entity, Wikipedia article (where notable), Reddit footprint, sustained editorial pickups in tier-1 publishers your category respects.

Layer 2

Live retrieval (ChatGPT Search & Browse)

When ChatGPT Search is active, the model issues queries to Bing, retrieves a candidate set, and the reranker selects passages to cite. Cited URLs tend to share four traits: top-10 Bing rank, clean schema, recent dateModified, and direct passage answers.

How to win this layer: Bing Webmaster Tools setup, allow OAI-SearchBot, FAQPage / Article schema, H2-as-question passage structure, quarterly content refresh.

The 10 ranking factors that move ChatGPT citations

In rough order of leverage, based on prompt-audit data across our retainer book over the last 12 months.

01Foundational

Bing index + ranking

ChatGPT Search retrieves from Bing. Submit your sitemap to Bing Webmaster Tools, fix Bing-specific crawl errors, and benchmark your top-10 share for seed queries on Bing — not Google.

02High

Wikipedia / Wikidata presence

Disproportionately weighted in training data. Wikidata entity with a complete sameAs graph is the fastest realistic win; a notability-grade Wikipedia article is the long-game upgrade.

03High

Reddit & forum citations

OpenAI signed a content-licensing deal with Reddit in May 2024. Brand mentions in relevant subreddits — earned organically through useful contribution, never astroturfed — feed both training data and ChatGPT Search retrieval.

04High

Editorial mentions in tier-1 publishers

Publisher mentions are corroborating signals the model treats as ground truth. Run digital PR against the trade and tier-1 publishers your category respects.

05Medium-high

Schema graph (Organization + Article + FAQPage)

Nested schema with sameAs links to Wikidata, LinkedIn, Crunchbase and authoritative profiles makes you a defined entity. FAQPage and Article unlock direct passage extraction.

06Medium

GPTBot + OAI-SearchBot allowlisting

Default-allow in robots.txt. If a CDN bot-management rule is blocking by user agent, OpenAI sees a 403 and the page never enters the training set.

07Medium

Content recency + dateModified

ChatGPT Search retrieval prefers recently-updated pages. Refresh pillar content quarterly and update both datePublished and dateModified in JSON-LD.

08Low (early)

llms.txt + AI declarations

Cheap to ship, signal value still proving out. Belt-and-braces — implement and move on. Both /llms.txt and meta name=ai-content-declaration are live on this site.

09Medium

Passage structure (Q&A scaffolding)

H2-as-question + 40-to-80-word direct answer is the format the model extracts most reliably. The same structure that wins featured snippets wins ChatGPT citations.

10Medium-high

Topical authority depth

ChatGPT prefers sources that have written extensively on a topic over one-off pages. Cluster depth (pillar + 8-12 supporting pages) measurably outperforms isolated content.

12-point ChatGPT self-audit checklist

Run through this in a single afternoon. Anything you cannot tick is leakage you can fix without waiting for the next algorithm shift.

1Submit sitemap to Bing Webmaster Tools and verify domain
2Confirm GPTBot and OAI-SearchBot are not blocked in robots.txt or CDN bot rules
3Ship /llms.txt and /llms-full.txt route handlers
4Add meta name="ai-content-declaration" content="permitted-with-attribution" to root <head>
5Implement Organization JSON-LD with sameAs to Wikidata, LinkedIn, Crunchbase, X
6Create or update Wikidata entity (Q-number)
7Audit pillar pages for FAQPage + Article schema
8Restructure top-10 pages around H2-as-question + 40-80-word answer format
9Refresh dateModified on top-20 pages quarterly
10Set up a 50-prompt monthly audit across ChatGPT (free + Plus + Search)
11Earn 3+ tier-1 editorial mentions per quarter through digital PR
12Build sustained presence in 3-5 high-relevance subreddits (organic, useful contribution only)

Want the audit run for you?

We benchmark all 10 factors across your domain plus 4 competitors, run a 50-prompt audit across ChatGPT free / Plus / Search, and deliver a 90-day priority sequence. Free.

Frequently asked questions

Eight questions we get most from founders trying to make sense of ChatGPT visibility.

ChatGPT cites sources at two layers: the training corpus (which decides whether the model knows your brand exists at all) and the live retrieval layer that powers ChatGPT Search and Browse (which decides which specific URL gets cited at query time). To rank in both, the brand needs (1) presence in Bing's index — ChatGPT's retrieval layer is Microsoft Bing, not Google; (2) corroborating mentions across high-trust open-web sources OpenAI's training data heavily weighted — Wikipedia, Reddit, news, GitHub, .edu and tier-1 publishers; (3) a clean, schema-marked website with FAQPage and Article schema; (4) a configured llms.txt and unblocked GPTBot in robots.txt; (5) recency signals — datePublished and dateModified fields, refreshed content. Brands that get all five layers right see citation share build inside 90 days.

Yes. GPTBot is OpenAI's training crawler and OAI-SearchBot is its real-time retrieval crawler. Check your server logs for those user agents — both should be visible if your robots.txt is permissive. The default Next.js robots.ts in this stack already allows them. The fastest sanity check is grep "GPTBot\|OAI-SearchBot" on your access logs over a 30-day window. If neither shows up, your robots.txt is blocking, your CDN is blocking by user-agent rule (Cloudflare's bot fight mode does this), or your site is too new and small to be in OpenAI's crawl rotation. The first two are fixable in minutes; the third resolves once authority signals build.

It is the single biggest lever, but not the whole story. ChatGPT Search retrieves from Bing's index at query time, so being indexed in Bing and ranking in the top 10 for your seed queries is non-negotiable. But citations are not just a function of Bing rank — OpenAI's reranker biases toward sources that already exist in the training data (Wikipedia, Reddit, large publishers) and toward content with clear question-answer structure. Brands that rank top-3 in Bing but have no Wikipedia presence and no Reddit mentions still under-cite. The hierarchy: Bing index + ranking → corroborating training-data mentions → on-page schema and structure.

More than almost any other single signal. Wikipedia and Wikidata are disproportionately weighted in OpenAI's training data — the model treats them as ground truth. Brands with a Wikipedia article are cited at materially higher rates than brands without one for the same category queries. The threshold for a stand-alone Wikipedia article (notability, independent secondary coverage) is a real bar, but a Wikidata entity with a clean sameAs graph is achievable for almost any operating brand and lifts citation rates noticeably on its own. We treat Wikidata entry as a 90-day deliverable on every GEO retainer.

llms.txt is a plain-text manifest at /llms.txt that tells LLMs which URLs on your site are canonical, what each one is about, and which sections to prioritise. Anthropic and several others have signalled they will read it; OpenAI has not committed but historically follows community standards once they reach critical mass. The cost to implement is hours, the upside is non-zero — it costs nothing to be early. Both /llms.txt and /llms-full.txt are already shipping on this site (auto-generated from blog content). If you are running a content site of any size, ship it.

Recency matters for ChatGPT Search and Browse — the live retrieval modes — and it matters less for default ChatGPT, which leans on training data. Pages with explicit datePublished and dateModified schema, recent content updates, and freshness signals (new publication dates, current-year references in body copy) are preferred by the retrieval layer. For brands competing in fast-moving categories (AI, fintech, ecommerce trends), refreshing pillar content quarterly and updating the dateModified field measurably lifts ChatGPT Search citation rates.

Indirectly, yes. The structural shifts that win citations — short, declarative summary paragraphs at the top of each section, clear H2 questions followed by 40-to-80 word answers, definition blocks before context, lists and tables for comparative data — are the same shifts that improve human readability. Writing for the model and writing for the reader converge on the same prose discipline. What we explicitly do not recommend: the kind of awkward keyword-stuffed prose that early SEO tools used to recommend for AI Overviews. ChatGPT's reranker penalises synthetic, low-information prose just like Google's quality systems do.

Yes — through controlled prompt audits. Pick a fixed set of 50 to 200 category-defining prompts, run them monthly across ChatGPT (free + Plus + Search), record whether the brand was cited, the position of the citation, and the surrounding context. Aggregate the data into a share-of-voice score versus competitors. Paid tools (Profound, Otterly, Goodie, Daydream) automate this; the manual version is feasible at smaller scale. Track the same prompt set across ChatGPT, Gemini, Claude and Perplexity so you can see which engine is moving and which is lagging. We bundle this measurement into every GEO retainer.

Find out where you actually stand on ChatGPT.

Free audit. We benchmark your domain across all 10 ranking factors, run a 50-prompt citation audit, and map the 90-day moves with the highest leverage. Across India, US and UK markets.