ChatGPT Citation Verification

ChatGPT Citation Checker

Verify whether ChatGPT's citations are real. Paste text generated by GPT-4o, GPT-5, or any ChatGPT version and see which references resolve to actual published papers — and which are fabricated.

Why ChatGPT invents citations

ChatGPT does not look up papers — it predicts tokens. When a user asks for sources, the model can generate strings that match the statistical shape of real citations: plausible author names, a recent publication year, a journal that publishes in the relevant field, a DOI prefix that corresponds to a real publisher. The result reads as authoritative but is generated, not retrieved.

Three hallucination patterns appear repeatedly in ChatGPT output:

  • Chimera citations. Real authors who work in the claimed field are attached to a paper title they never wrote, published in a journal where they have no record. Each component is plausible; the combination does not exist.
  • DOI fabrication. The citation includes a DOI-formatted string (10.xxxx/yyyy) that fails to resolve. CrossRef returns nothing; the DOI was pattern-matched, not retrieved.
  • Temporal drift. A paper's actual publication year is replaced with a more recent year, or a pre-print date is cited as a final publication. This is particularly common when ChatGPT is asked for "recent" sources.

How Citely catches ChatGPT hallucinations

1

Extract every citation

Paste ChatGPT output — in-text citations, reference list, or full draft. Citely parses each reference into structured fields: DOI, title, author, year, journal.

2

Resolve against primary sources

Each field is checked independently against CrossRef (for DOI and publisher metadata), PubMed (biomedical), arXiv (preprints), and OpenAlex (broad coverage). A chimera fails on author-title pairing; a fabricated DOI fails on resolution.

3

Report with evidence

For each citation: verified (with authoritative record shown), partial match (metadata fields disagree), or fabricated. You see exactly which component failed — not just a pass/fail.

All ChatGPT versions, one checker

Citation hallucinations have been observed across every ChatGPT generation. Newer models may reduce the frequency for some prompts, but the underlying failure mode — plausibility without ground truth — has not been eliminated. Independent verification remains necessary regardless of which model produced the text.

GPT-3.5 Turbo
GPT-4
GPT-4 Turbo
GPT-4o
GPT-5
ChatGPT with browsing
Custom GPTs
ChatGPT Team / Enterprise

Citely verifies the output text, not the model — so the same check works regardless of which ChatGPT version produced the citations.

Frequently asked questions

Does ChatGPT actually make up citations?

Yes. Large language models like ChatGPT generate text token-by-token from statistical patterns rather than retrieving from a live database, so they can produce references that sound plausible but do not correspond to published papers. The rate varies by topic and prompt — niche subfields and recent literature are particularly affected — and the phenomenon has been documented in medical, legal, and academic contexts.

What's different about ChatGPT's hallucinations compared to Claude or Gemini?

ChatGPT tends to produce 'chimera' citations — combining a real author, a real journal, and a plausible title that does not correspond to any actual paper. It also frequently invents DOIs that follow CrossRef's format but resolve to nothing. Claude more often misattributes real papers to wrong authors; Gemini more often fabricates plausible journal names. Citely detects all three patterns the same way: by resolving each citation against authoritative records.

Can I just ask ChatGPT to verify its own citations?

No. This is the most common mistake. Asking ChatGPT 'are these citations real?' is asking the system that produced the hallucination to detect itself. The model has no access to the CrossRef database, no DOI resolver, and no way to confirm publication metadata. A self-check will often produce a confident 'yes, these are real' for citations that do not exist. Independent verification against academic databases is the only reliable method.

Does ChatGPT with browsing or search plugins solve this?

Partially, but unreliably. ChatGPT's search and browsing features can surface real papers, but the underlying generation still introduces drift — the cited paper may differ from the linked source, the DOI may not match the URL, and the retrieved snippet may be misattributed. Verification remains necessary even when ChatGPT claims to have searched.

What ChatGPT models does Citely work with?

All of them. GPT-3.5, GPT-4, GPT-4o, GPT-4 Turbo, GPT-5, ChatGPT with browsing, Custom GPTs, and any fine-tuned variant — Citely verifies the output, not the model. Paste the references and the verification runs identically.

Is it enough to check just the DOI?

No. A fabricated citation can include a DOI that happens to resolve to a real but unrelated paper, or a DOI that follows the correct format but resolves to nothing. Citely cross-checks DOI against title, title against author list, and author list against journal — so a citation that passes one dimension but fails the others is flagged, not approved.

Verify ChatGPT's citations before you submit

Paste the reference list. See which citations are real, which are chimeras, and which are pure fabrications.

Check ChatGPT Citations