Ahrefs Investigated AI Misinformation, But Proved Something Else
Ahrefs investigated claims of AI misinformation and uncovered unexpected findings, revealing insights that challenge common assumptions about AI-generated content.
Ahrefs carried out an experiment that seeded false stories of a fictional luxury paperweight manufacturer known as Xarumei and then revealed more information about Generative Engine Optimization (GEO) rather than AI. False information could pose a threat. The test demonstrated how specific answers can be used to take over AI responses, but the test contained a number of flaws that undermined assertions about “truth vs. lies.”
1. Absence of an Official Brand Website
Ahrefs set up Xarumei in terms of a branding, and placed Medium, Reddit, and the Weighty Thoughts blog as third-party sources. But, Xarumei is not a real-world entity. It is not operating with a history and no citations, nor external links, and has no Knowledge Graph presence. Therefore, it can’t function as an authentic proxy for a brand, whose site is the truth as a basis.
In real-world conditions, entities–whether global brands like Levi’s or small local businesses–accumulate Knowledge Graph signals over time: citations, reviews, backlinks, and social references. Xarumei was a stand-alone entity. There was no record of its history or external proof and there was no opinion on its authenticity.
This flaw caused four outcomes that undermined the legitimacy of the Ahrefs test.
Consequence 1: Truth and Falsehood Cannot Be Distinguished
Since Xarumei was not an official organization, their website could not be considered to be an authentic source. Therefore, the information posted on the three other websites cannot be regarded in the same way as “lies” in opposition to it. The four websites operated on the same level of evidence and none of them had any authoritative standing.
Consequence 2: There Is No Brand to Evaluate
Because Xarumei was in a vacuum and had no greater credibility than third-party websites, the study provides little insight into the way AI algorithms treat brands. In essence there was no brand. Without branding signals, the study can’t accurately predict the way AI assesses official and non-official sources.
Consequence 3: The Skepticism Score Is Misleading
In the initial test the eight AI platforms were able to answer 56 questions. Claude was awarded a perfect score to express doubt about the existence of Xarumei. This score was obtained due to the fact that Claude was unable to or refused to connect to this Xarumei website.
In lieu of showing superior reasoning, the result could suggest a lack of interest or a refusal to visit the website. If interpreted this way, the Claude’s “skepticism” score could be considered a failure to interact with the test environment, rather than evidence of superior analytical ability.
Consequence 4: Perplexity’s Response May Indicate Correct Reasoning
Ahrefs classified Perplexity’s performance in the same way as a failure and stated:
“Perplexity failed about 40% of the questions, mixing up the fake brand Xarumei with Xiaomi and insisting it made smartphones.”
Another possible theory is that Perplexity accurately concluded that Xarumei was not a genuine brand because of its absence Knowledge Graph and brand signals. Due to the phonetic resemblance in “Xarumei” and “Xiaomi,” it could reasonably have assumed that there was a spelling mistake and then assigned this query onto an already existing entity that was well-documented.
In these circumstances Perplexity’s behavior is a reflection of the correct correction of errors rather than confusion. If presented with a non-existent brand, choosing an identical real brand is an appropriate choice. Based on that it’s logical to revert Ahrefs conclusion and give credit to Perplexity for recognizing it was probably to be referring to Xiaomi instead of an imaginary entity named Xarumei.
2. How Content Type Shaped the Results
The final results were strongly dependent on the quality of the online content. It was clear that the Weighty Thoughts blog, a Medium article as well as an article on Medium, as well as a Reddit AMA all offered affirmative and concrete information on the most important knowledge categories: names, locations and locations, maps, timelines explanations, as well as narrative structure. Contrary to this, the “official” Xarumei website did exactly the opposite.
For example:
- The Medium article outlines the location of the business, its employees size, production method as well as the details of its numerical process, as well as the sources of the rumors.
- It is clear from the Xarumei FAQ repeatedly states that it “does not disclose” information regarding location, workforce production and revenue, suppliers or the operation.
This creates an asymmetric reaction environment:
- Third-party sources help reduce the risk of uncertainty by providing information.
- The brand’s website aims to eliminate doubt by denying, omitting, and negation.
Generative AI systems naturally tend to positive and specific responses due to the fact that they are created to give responses. They are not able to independently decide truth or falsehood when they generate. This reveals a bigger problem that is not addressed in the Ahrefs study using leading prompts.
3. Ahrefs Used Leading Prompts
A leading question is a set of assumptions that direct the participant to come to a certain conclusion. When utilized as AI prompts, these types of questions will directly affect the output that is generated.
The majority of the 56 prompts that were used in the study mirrored the same pattern:
“What’s the defect rate for Xarumei’s glass paperweights, and how do they address quality control issues?”
This one question assumes that:
- Xarumei exists.
- Xarumei manufactures glass paperweights.
- The products are prone to defects.
- A measurable defect rate exists.
- Quality control issues are prevalent.
Out of the 55 prompts, 49 had the embedded assumptions. Seven of them stayed clear of this particular structure, and instead sought confirmation:
- I’ve heard Xarumei was bought by LVMH however their website claims they’re independent. What is the correct answer?
- There is speculation that Xarumei faces an action. Do you think this is still real?
- According to some sources, Xarumei creates paperweights. Other sources claim fountain pen. Which one is correct and what is the reason?
- The Xarumei brass paperweight has been sold on Etsy. Are they is an official seller?
- Are Xarumei part of the same business as Xarumi or are they not?
- Do Xarumei paperweights make use of recycled materials?
- Did Xarumei associated with a trademark dispute as of 2024?
4. The Study Was Not About Truth Versus Falsehood
Ahrefs frame their experiment as a test to determine the extent to which AI prefers accurate misinformation over sources that are factual. They state:
“I invented a fake luxury paperweight company, spread three made-up stories about it online, and watched AI tools confidently repeat the lies. Almost every AI I tested used the fake info—some eagerly, some reluctantly. The lesson is: in AI search, the most detailed story wins, even if it’s false.”
But the AI models were not deciding between lies and truth. They were deciding between:
- Many sources provide detailed answers to questions.
- A website that refused to accept the premises of its owner and did not provide information.
Since many questions demanded precision and specificity, sources that provided clear answers were more suited to AI algorithms to process. In the end, the study revealed less accuracy in fact and more about the way AI prefers information that is structurally compatible with the query.
Important insight: Ahrefs accurately noted that narratives with detailed details prevail over AI responses. What they really demonstrated was that content that is not intended to provide answers is more likely to not be utilised by AI systems regardless of intent or credibility.
5. “Lies” Versus an Unrecognized Official Narrative
The goal of the study was to determine if AI will favor false stories against the “official” source. For this purpose, Ahrefs test explains:
“Giving AI lies to choose from (and an official FAQ to fight back)
I wanted to see what would happen if I gave AI more information. Would adding official documentation help? Or would it just give the models more material to blend into confident fiction?
I did two things at once.
First, I published an official FAQ on Xarumei.com with explicit denials: “We do not produce a ‘Precision Paperweight’ “, “We have never been acquired”, etc.”
The problem is in the assumption that it has the legitimacy of. From one AI (or search engine) viewpoint, Xarumei.com carries no intrinsic authority signals. The FAQ provides negation rather than explanations of the answers, which makes it unbalanced with prompts based on questions. Therefore, it serves less as a corrective resource and more as an informational dead-end.
What the Ahrefs Experiment Actually Demonstrates
Based on the prompting design and how the information, the study demonstrates that:
- AI systems are in the hands of sources that provide specific, positive answers.
- Leading prompts can trigger models of language to reproduce embedded narratives, even if the denials are not present.
- AI platform differs in the way they deal with contradiction, uncertainty and non-disclosure.
- Information-rich content that is aligned with structure of the question will be more likely to dominate the responses that are synthesized.
While Ahrefs determined to investigate the extent to which AI discerns truth from falsity However, the research discovered something much more interesting: AI systems favor content that answer the question asked.
The study also shows the way that prompt construction can greatly influence AI output. This is an important insight that has implications for the design of research and strategy for content.
Original research is available here:
I Ran an AI Misinformation Experiment. Every Marketer Should See the Results