GPT-5.5 has the highest absolute accuracy ever recorded on a major AI benchmark. It also has an 86% hallucination rate on citation-sensitive tasks. These two facts sound like they contradict each other. They don't.
Researcher Karo Zieminski published the first structured benchmark comparison for GPT-5.5 on April 25, 2026, using the AA-Omniscience benchmark, designed to penalize confident wrong answers. The findings reframe what OpenAI's "reliability-first" positioning means in practice, and what it means for any brand that needs AI to represent it accurately.
GPT-5.5 rolled out to ChatGPT Plus, Pro, Business, and Enterprise on April 23. The API went live on April 24. Our post from April 24 covered the citation pool compression implications of the reliability-first design. This post covers a different problem: what happens to your brand when the model doesn't have enough to go on.
AA-Omniscience Benchmark — Karo Zieminski (April 25, 2026)
Hallucination rate by model
Measures how often each model provides a confident wrong answer rather than abstaining on uncertain queries. Lower is better.
Highest abstention rate — hedges when uncertain
Recommended for citation-sensitive and regulatory tasks
Highest accuracy ever recorded — 57% on AA-Omniscience
The accuracy paradox
Per-fact accuracy improvement vs. prior models
Overall response accuracy improvement vs. prior models
GPT-5.5 packs more facts into every response. Individual facts are more accurate, but more facts per response means more chances for at least one error to appear. The response-level improvement is nearly flat despite the per-fact gain.
Source: Karo Zieminski, Substack (April 25, 2026). AA-Omniscience benchmark penalizes confident wrong answers over appropriate abstention.
What 57% accuracy actually measures
The 57% absolute accuracy figure is real. It is the highest the AA-Omniscience benchmark has recorded for any model. The benchmark measures total performance across a wide question set, including questions where the model has strong training data coverage.
When GPT-5.5 knows something, it is right more often than any previous model. The 57% figure covers all queries in aggregate. For topics, companies, and facts that appear frequently in GPT-5.5's training data, the accuracy improvement is genuine.
One important constraint: 57% accuracy tells you about performance in cases where the model produces a measurable answer. It says nothing about what the model does when it is uncertain.
What 86% hallucination rate actually measures
The hallucination rate measures a specific failure mode: what happens when GPT-5.5 does not have reliable information?
Zieminski's framing is precise. "It doesn't mean the model is wrong 86% of the time. It means: when GPT-5.5 doesn't know, it almost never tells you. It guesses, in the same confident tone it uses when it's correct."
AA-Omniscience is built to detect this. It includes questions the model should abstain from answering because the information is obscure, contested, or outside its reliable knowledge. A model with good calibration hedges on those questions. GPT-5.5 hedges on 14% of them.
The cross-model comparison from the benchmark:
| Model | Hallucination rate |
|---|---|
| Grok 4.20 | 17% |
| Claude Opus 4.7 | 36% |
| Gemini 3.1 Pro Preview | 50% |
| GPT-5.5 | 86% |
Source: Karo Zieminski, Substack, April 25, 2026
GPT-5.5 has the highest accuracy score and the highest hallucination rate of the four models tested. Grok 4.20 leads on calibration at 17%. Claude Opus 4.7 sits at 36%. Gemini is mid-table at 50%.
Why more facts per response makes the math worse
OpenAI's system card for GPT-5.5 explains the apparent paradox. The model generates longer responses with more individual facts per answer compared to GPT-5.4.
Two numbers from the system card:
- •Individual fact accuracy improved approximately 23% versus prior models
- •Overall response accuracy improved approximately 3%
The math matters here. GPT-5.5 packs more facts into every response. Each individual fact is more likely to be correct than before. But when you multiply more facts by even a fractionally lower per-fact error rate, the chance that at least one error appears in a complete response barely moves. More facts times a slightly smaller error probability still equals more total errors per response.
The practical result: GPT-5.5 is more accurate per claim. It is not meaningfully more accurate per response, because there are more claims per response to go wrong.
That dynamic is part of what the 86% hallucination rate captures. A model producing more content per answer encounters uncertain territory more frequently. When GPT-5.5 reaches uncertain territory, it does not pull back to what it is confident in. It continues at full speed.
What is GPT-5.5 saying about your brand right now?
We run brand description audits across GPT-5.5 to identify any fabricated claims, compare them against your GPT-5.4 baseline, and build the off-site footprint that anchors accurate AI representation going forward.
Book a Discovery CallWhat this means for brands not well-covered in training data
The accuracy improvement accrues to brands that are well-documented in GPT-5.5's training data. If your company has years of high-authority third-party coverage, G2 reviews with specificity, editorial placements, and analyst mentions, the per-fact accuracy improvement likely improves how GPT-5.5 represents you.
The hallucination risk flows in the opposite direction. Evertune's research on AI hallucination patterns identified the categories most exposed: niche topics, lesser-known brands, recent events, and attributes requiring precise verification. These are exactly the cases where the model's training data coverage is thin or inconsistent.
For those brands, GPT-5.5's behavior is a specific liability. When the model encounters a brand it does not have deep knowledge of, it does not say so. It generates a confident description using vocabulary and patterns associated with that type of company, interpolated from adjacent examples in its training data. The resulting output may be plausible-sounding but wrong in specifics that matter: incorrect product capabilities, invented integrations, made-up customers, or wrong pricing structures.
The concerning shift from GPT-5.4 to GPT-5.5: the confident-wrong failure mode was always present, but earlier models were somewhat less inclined to produce long, fact-dense responses on uncertain topics. GPT-5.5's tendency toward thoroughness means it generates more content about brands it knows less about, not less. The 86% hallucination rate is the empirical result of that combination.
This reframes the G2, press, and Reddit content strategy. As we covered in the training data piece, two-thirds of ChatGPT answers about brands already come from training data, not live web retrieval. With GPT-5.5, that share is likely higher. The same program that builds citation frequency now has a second function: it builds protection against brand description fabrication. High-authority off-site coverage teaches the model what your brand actually does, which is the only real defense against confident hallucination.
How to check your brand under GPT-5.5 right now
GPT-5.5 is the live default for Plus, Pro, Business, and Enterprise ChatGPT users starting April 23. API access opened April 24. Whatever GPT-5.5 says about your brand is what buyers are reading in research sessions happening this week.
The check is direct. In a ChatGPT session, disable web search by turning off the search toggle, then ask the model to describe your company, your product capabilities, your customers, and your category position. The knowledge-only response reveals what the model carries in training data about you.
Compare that output to what appears in standard sessions with web search enabled. Discrepancies between the two show how much your citation presence depends on live retrieval versus training data representation. Large gaps between them indicate over-indexing on fresh content optimization and under-indexing on the off-site signals that feed training data.
A second check targets fabrication: ask about specific product features, integrations, and technical details. If the model generates confident descriptions that are incorrect, those are active brand representation problems that buyers encounter when they ask similar questions.
GPT-5.5 is freshly deployed. The next two weeks are the window to catch any changes while you can still compare against your GPT-5.4 baseline from before April 23.
Which model fits which task
Zieminski's conclusion from the benchmark data is direct: GPT-5.5 is the right tool for coding, reasoning, agentic planning, and 1M-token context work. For citation-heavy tasks, regulatory references, source claims, or any work where confident-wrong is the failure mode, Claude Opus 4.7 is the better choice.
That division maps onto a practical framework for B2B brands thinking about platform optimization priorities:
| Use case | Model | Rationale |
|---|---|---|
| Vendor shortlist research | Claude Opus 4.7 | 36% hallucination rate; abstains when uncertain |
| Technical documentation research | Claude Opus 4.7 | Lower risk of confident wrong answer on precise claims |
| Coding, architecture, planning | GPT-5.5 | 57% accuracy; strong on reasoning tasks |
| Long-context document analysis | GPT-5.5 | 1M token context strength |
| Competitive analysis research | Grok 4.20 | 17% hallucination rate; highest abstention |
The implication for GEO strategy: citation optimization priorities should track where buyers research vendors, not where they do technical work. Procurement and vendor evaluation workflows tend toward lower-hallucination models because the cost of a confident wrong answer is high. Our platform prioritization guide covers the research by buyer type, but the benchmark data adds a new input: for B2B evaluation queries, Claude and Grok are increasingly the research tools of choice among buyers who understand the hallucination tradeoffs.
The G2 and press program now does two things simultaneously
Before GPT-5.5, brand authority was the strongest predictor of AI citations. The Digital Bloom research found a 0.664 correlation between off-site brand mentions and AI citation frequency, the highest of any factor measured.
GPT-5.5 adds a second function to the same work. Building G2 reviews with specificity, editorial coverage in sector publications, analyst mentions, and LinkedIn content that generates practitioner discussion now produces two outcomes: more citation frequency for well-documented brands, and more accurate brand representation when the model reaches uncertain territory about you.
The brands most exposed to GPT-5.5's hallucination risk are those that built AI visibility primarily through owned content: well-optimized blog posts, product pages, and press releases. Those surfaces matter for live-retrieval citations. They do not build training data representation in the same way that third-party coverage does. A brand with 200 structured blog posts but limited G2 presence and sparse editorial coverage is well-positioned for the live-retrieval share of ChatGPT responses, and poorly positioned for the training-data share. With GPT-5.5, the training-data share of responses is larger, and the hallucination risk when training data is thin is higher.
Conductor's 2026 benchmarks found that ChatGPT still drives 87.4% of all AI referral traffic. GPT-5.5 is a significant share of that. The hallucination rate does not change the platform's importance. It changes what kind of optimization actually defends your brand representation on it.
FAQ
What is the AA-Omniscience benchmark?
AA-Omniscience is an AI benchmark designed to measure epistemic calibration: whether models abstain appropriately when they do not know an answer, rather than generating confident wrong answers. It penalizes models that provide incorrect information with high confidence rather than acknowledging uncertainty. GPT-5.5 achieves the highest accuracy score ever recorded at 57%, and the highest hallucination rate at 86%, because it maximizes performance on questions it knows while almost never abstaining on questions it does not.
How can GPT-5.5 have both the highest accuracy and the highest hallucination rate?
They measure different things. Absolute accuracy covers what percentage of total answers are correct, averaged across all questions including areas of model strength. Hallucination rate measures what percentage of uncertain questions produce confident wrong answers rather than appropriate abstention. GPT-5.5 improved per-fact accuracy by 23% but produces more facts per response. The result: higher individual accuracy, roughly flat overall response accuracy, and a high hallucination rate because more content per response means more confident coverage of topics where the model is reaching beyond what it reliably knows.
Which brands face the highest GPT-5.5 hallucination risk?
Brands with thin, inconsistent, or contradictory training data coverage face the highest risk. This includes newer companies without multi-year editorial coverage, niche products with limited third-party documentation, and any brand that built AI visibility primarily through owned content rather than off-site signals. When GPT-5.5 encounters these brands, it draws on adjacent patterns to construct a response rather than abstaining. The confident description it produces may be plausible in tone but factually wrong in specifics that matter: product features, integrations, customers, pricing.
What is the fastest way to check if GPT-5.5 is misrepresenting my brand?
Disable web search in a ChatGPT session by turning off the search toggle, then ask specific questions about your product capabilities, integrations, customer base, and pricing. Compare the training-data-only response against your actual product description. Any confident incorrect claims represent active exposure. Then run the same session with web search enabled to see how live retrieval changes the output. The gap between the two responses shows how much your current citation presence depends on owned content versus third-party training data coverage.
Should I deprioritize ChatGPT optimization if GPT-5.5 has an 86% hallucination rate?
No. The 86% rate applies to uncertain-query behavior, not to overall output. For queries where your brand has strong training data coverage, GPT-5.5 is more accurate than prior models. ChatGPT still drives 87.4% of all AI referral traffic according to Conductor's 2026 benchmarks. GPT-5.5 also remains the leading platform for agentic workflows and long-context analysis. The hallucination rate changes what you optimize for: less emphasis on content freshness alone, more emphasis on building the third-party training data signals that prevent fabrication and anchor accurate brand representation.
The audit window is open
GPT-5.5 went live for most ChatGPT users this week. Whatever errors or accurate representations the model has adopted for your brand are measurable right now, before those outputs propagate through more buyer research sessions.
The brands that check now get a clean comparison against the GPT-5.4 baseline. The ones that wait find that incorrect descriptions have already influenced an unknown number of buyer research sessions, and the next training data update may entrench those errors further.
A full AI visibility audit that includes training-data representation testing captures both dimensions: citation frequency and description accuracy. GPT-5.5 made the difference between the two more consequential than any prior model release.
Run a GPT-5.5 brand description audit before the baseline closes.
We test your brand representation across GPT-5.5's training-data and live-retrieval modes, identify any fabricated claims, compare against your GPT-5.4 baseline, and build the off-site footprint that anchors accurate AI representation long-term.
Get Your AI Visibility AuditFramework
Learn the CITE framework behind our GEO and AEO work
See how Comprehend, Influence, Track, and Evolve turn AI visibility into an operating system.
Services
Explore our managed GEO services and AEO execution model
Audit, prompt discovery, content execution, and ongoing monitoring tied to AI search outcomes.
GEO Agency
See what a managed GEO agency should actually do
Compare real GEO operating work against generic reporting or tool-only approaches.
Audit
Start with an AI visibility audit before execution
Understand prompt coverage, recommendation gaps, source mix, and where competitors are winning.