Mind Matters Natural and Artificial Intelligence News and Analysis
customers-using-smartphone-technology-chat-concept-ai-artifi-593424439-stockpack-adobestock
Customers using smartphone technology chat concept ai artificial intelligence Intelligent digital chatbot chat robot
Image Credit: VRVIRUS - Adobe Stock

Top accounting firm tries to get AI to do the writing and… splat!

Share
Facebook
Twitter/X
LinkedIn
Flipboard
Print
Email

It’s not clear why our new Masters of the Universe (in Tom Wolfe‘s words) are taking so long to learn a basic principle: machines do not think. As Associated Press’s Australia correspondent Rod McGuirk reports,

Deloitte Australia will partially refund the 440,000 Australian dollars ($290,000) paid by the Australian government for a report that was littered with apparent AI-generated errors, including a fabricated quote from a federal court judgment and references to nonexistent academic research papers.

The financial services firm’s report to the Department of Employment and Workplace Relations was originally published on the department’s website in July. A revised version was published Friday after Chris Rudge, a Sydney University researcher of health and welfare law, said he alerted the media that the report was “full of fabricated references.” “Deloitte to partially refund Australian government for report with apparent AI-generated errors,” October 7, 2025

McGuirk identifies the problem as hallucination, the tendency for AI chatbots (technically, large language models) to just make stuff up.

It can get pretty florid. As economics prof Gary Smith noted last year, for a while ChatGPT was claiming that the Soviets had sent bears into space.

The reason this happens is that the bot is designed in such a way that it will make up an answer rather than report no information. Now, here’s the difficulty: Obvious nonsense can be spotted easily and corrected by personnel at say, ChatGPT or Claude.

For example, in the Deloitte misfortune, if the federal court judgement was quoted as saying that every employee should eat spiders for lunch, it wouldn’t take long for the “spiders” to hit the fan. Everyone but the bot would know that that cannot be what a court ruled.

But much nonsense may not be obvious at all. A quotation from a court judgment that supports a policy perspective that sound reasonable may be fabricated. And no one who doesn’t check, as that health law researcher did, will ever know. After a while, the amount of completely erroneous information will start to become a more significant part of the overall information base.

Progress in reducing these errors has been, as computer engineering prof Robert J. Marks reports, slow and halting because of the underlying problems. His advice: “I suspect there will never be enough Band-Aids to stop all of the bleeding. So instead of ‘Trust but verify,’ I recommend ‘Verify, then trust.’”


Enjoying our content?
Support the Walter Bradley Center for Natural and Artificial Intelligence and ensure that we can continue to produce high-quality and informative content on the benefits as well as the challenges raised by artificial intelligence (AI) in light of the enduring truth of human exceptionalism.

Top accounting firm tries to get AI to do the writing and… splat!