Trusting AI Tools Like Grok, Gemini, ChatGPT, and Perplexity for Citations is Unwise

The Shift in Information Retrieval: AI’s Influence on Search Engines
Over the years, our interaction with the internet has transformed significantly, often resulting in less control over the information we encounter. This change is especially visible in social media, where algorithm-driven content delivery has taken over traditional methods of information seeking.
The Rise of AI Curation in Search Engines
While traditional search engines allow users to directly search for specific information, the rise of artificial intelligence (AI) suggests a shift towards algorithm-driven curation as a standard practice in retrieving information. A recent development from Google, known as the ‘AI Mode,’ hints at a future where AI dominates this landscape. This trend is mirrored across various AI-powered search tools, such as Perplexity, Grok, and ChatGPT. Current statistics indicate that around one in four Americans is now opting for AI instead of conventional search engines.
Study Highlights Flaws in AI Search Tools
A recent report from the Tow Center for Digital Journalism at Columbia University evaluated multiple AI search engines, including ChatGPT and Grok. The researchers selected ten articles from twenty different publishers and presented snippets from these articles to the AI tools. The goal was to see how accurately these tools could identify the original headline, publisher, publication date, and URL.
The results were alarming. The study revealed that over 60% of the responses generated by these AI search engines were incorrect. Specifically, Perplexity made errors on 37% of its responses, while Grok had a staggering 94% incorrect answers. This raises significant concerns about the reliability of AI as a search tool.
Key Findings from the Study
- High Error Rate: More than 60% of responses were incorrect, with specific platforms like Grok performing particularly poorly.
- Citation Issues: Many AI tools provided citations for invalid or broken links, diminishing user trust considerably.
- Source Attribution Failures: The AIs often failed to trace back to the original source. For example, although Perplexity has partnerships with the Texas Tribune, it incorrectly cited republished versions instead of the originals.
The study pointed out that many AI systems showed confidence in their incorrect answers, rarely indicating uncertainty with phrases like "it appears" or "it’s possible."
Growing Concerns Among Users
Despite continuous improvements in AI systems, many users have expressed legitimate concerns about the accuracy of information provided. Google’s new "AI overviews" feature has been met with frustration over misleading or confusing answers. Recent examples include incorrect metrics presented in straightforward queries, where users noted discrepancies that could lead to misunderstanding.
Issues with AI Responses
- Lack of Clarity: Statements often lacked qualifiers, leading users to accept incorrect information as factual.
- Potential for Misinformation: Users are increasingly worried that inaccurate AI-generated information could extend beyond harmless errors and potentially lead to serious misunderstandings.
Google’s Responsibility in the AI Landscape
As AI systems evolve, there is growing pressure on companies like Google to ensure accurate and reliable information in their search results. Although Google claims its AI overviews are a favorite among users, many have voiced the desire to disable this feature due to its inconsistencies.
While Google is rolling out new AI capabilities, including an ‘AI Mode’ for improved search experiences, concerns remain regarding the accuracy of information delivered by these advanced tools. Reports suggest that over 90 million people in the United States may predominantly rely on AI for web browsing in the near future. Consequently, AI developers are expected to adopt more stringent standards to mitigate the spread of misinformation, as inaccuracies could lead to serious consequences in various contexts.
Developments in AI-powered search tools indicate a significant shift in how we access information, potentially reshaping our understanding of trust and reliability in digital contexts.