Consulting firm quietly admitted to GPT-4o use after fake citations were found in August.
See full article...
See full article...
How many times do we have to see the fastest-to-the-bottom crowd make these mistakes before companies realize that AI isn't a magic wand?Deloitte and the DEWR buried that explanation in an updated version of the original report published Friday "to address a small number of corrections to references and footnotes," according to the DEWR website.
I mean, I guess just use chat-gpt instead of Deloitte, seems to be the result of their consultancy. The results will be similarly nonsense, but a lot cheaper.
It does worry me too that apparently nobody is sanity checking these reports before their general publication? Like if 10+ citations outright did not exist then Deloitte's analysis must have just been taken at face value with no serious review, and we are presumably using this to justify government policy?!This. I absolutely cannot understand why a consultancy, who's entire business model is "pay us large sums of money for our expert's advice" would rely on a LLM for even as much as grammar advice.
If your expert is ChatGPT, why do I pay you? I can write prompts myself. This is an incredibly fast way to sink your entire business model- if I were McKinsey or one of the others I'd be out there advertising "We know what we're doing, we don't need AI to do it poorly"
I'm not familiar with the Australian legal system, but they should actually have to pay a penalty multiplier for fraud, in other words pay more than they received for the fraudulent report. It also appears that they started with the conclusion that they wanted, and are claiming that their fraud does not alter the conclusion.Perhaps Lisa Crawford has a case for defamation or slander for having these false papers attributed to her. One way to stop the nonsense is to make it hurt. As it airs they are partially refunding the money but clearly all they did is engineer a few AI prompts to get the report. Make them refund it all, make them pay for defamation and send a message that this crap isn’t okay.
Same for the lawyers who submit briefs to the court with fake legal citations.
Genuine question - that makes me question the value of the summaries. How can we know the summaries are correct without reading the paper itself? Is there any research on not just lost nuance, but the hallucinations in AI summaries? I'd be interested in seeing it across approaches, such as NotebookLM and Kagi, which can pin to a set of sources, or requests to summarize a single paper across different models.I want to be fair about this, because I generally think ChatGPT is a useful tool for lit searches and summaries of papers (as with any summary, some nuance is lost). However, once I asked it for sources on a certain topic and it responded with hallucinated papers. My first clue that something wasn't quite right was when one of the papers (of which I was not previously aware) listed me as the first author...
"It is difficult to get a man to understand something, when his salary depends on his not understanding it." Attributed to Upton Sinclair, but unverified.Earlier this year, Deloitte declared it would start using generative AI for its reports as a way of enhancing the value provided to its clients. I don't remember if they said it in a specific report or not, but I recall seeing it.
The citation issue continues to trip people up across the spectrum, from lawyers to business analysts. It's striking how many supposedly smart people do not understand the limits of the tools they insist will deliver such amazing value.
4. Mistakes were made.How many times do we have to see the fastest-to-the-bottom crowd make these mistakes before companies realize that AI isn't a magic wand?
Anyone want to place a bet on :
- How long it takes to find a hallucination in the updated version.
- When another Deloitte client will discover hallucinations in the work product they received, or
- When Deloitte issues a statement blaming a low-level employee who's been disciplined for the errors?
Yeah, I was skeptical. I baselined this by using ChatGPT to summarize papers I'd already read, or on topics I'm already an expert in. Usually the summary was a re-wording of the abstract, with maybe some additional context from the paper. I've found these summaries to be pretty good, but you have to check.Genuine question - that makes me question the value of the summaries. How can we know the summaries are correct without reading the paper itself? Is there any research on not just lost nuance, but the hallucinations in AI summaries? I'd be interested in seeing it across approaches, such as NotebookLM and Kagi, which can pin to a set of sources, or requests to summarize a single paper across different models.
I occasionally use AI to summarize things, but I don't trust it past summarizing things where the ultimate goal is to point me to the actual authoritative source when I'm having a hard time finding it. Do you trust the summaries you get? And if so, why?
While I am in favor of sticking it to people who use AI like this in any way possible, I think you'd probably find it hard to make the case that it's "defaming" her. Unless the made up papers are about her doing experiments on humans or something of that magnitude. While it's wrong, if someone cited a fake paper I wrote on an inconclusive drug trial, it'd be hard to show that it had somehow damaged my reputation.Perhaps Lisa Crawford has a case for defamation or slander for having these false papers attributed to her. One way to stop the nonsense is to make it hurt. As it airs they are partially refunding the money but clearly all they did is engineer a few AI prompts to get the report. Make them refund it all, make them pay for defamation and send a message that this crap isn’t okay.
Same for the lawyers who submit briefs to the court with fake legal citations.
These big firms have been doing this for ages. It's just in the before times they'd have a room full of cheap college grads with limited experience do the work. They you have one person supervising multiple projects directing the work.This. I absolutely cannot understand why a consultancy, who's entire business model is "pay us large sums of money for our expert's advice" would rely on a LLM for even as much as grammar advice.
If your expert is ChatGPT, why do I pay you? I can write prompts myself. This is an incredibly fast way to sink your entire business model- if I were McKinsey or one of the others I'd be out there advertising "We know what we're doing, we don't need AI to do it poorly"
You're absolutely correct.Were there any em-dashes in the report?
Earlier this year, Deloitte declared it would start using generative AI for its reports as a way of enhancing the value provided to its clients. I don't remember if they said it in a specific report or not, but I recall seeing it.
The citation issue continues to trip people up across the spectrum, from lawyers to business analysts. It's striking how many supposedly smart people do not understand the limits of the tools they insist will deliver such amazing value.
4. Mistakes were made.
That was already a suggestion made.I mean, I guess just use chat-gpt instead of Deloitte, seems to be the result of their consultancy. The results will be similarly nonsense, but a lot cheaper.
There is no way I'm aware of to guarantee that an AI has summarized something perfectly without reading the documents yourself. ChatGPT openly acknowledges, if you ask it, that it is best at summarizing short, clear content. Since that's the content you least need summarized, it seems to be a use-case in search of a use. And because chatbots are not like humans and can't learn your personal preferences, there's no guarantee that a bot won't summarize 50 stories perfectly but fail on the 51st.Genuine question - that makes me question the value of the summaries. How can we know the summaries are correct without reading the paper itself? Is there any research on not just lost nuance, but the hallucinations in AI summaries? I'd be interested in seeing it across approaches, such as NotebookLM and Kagi, which can pin to a set of sources, or requests to summarize a single paper across different models.
I occasionally use AI to summarize things, but I don't trust it past summarizing things where the ultimate goal is to point me to the actual authoritative source when I'm having a hard time finding it, so I can verify the summary. Do you trust the summaries you get? And if so, why?
You're fooling yourself about the accuracy of the summaries.I want to be fair about this, because I generally think ChatGPT is a useful tool for lit searches and summaries of papers (as with any summary, some nuance is lost). However, once I asked it for sources on a certain topic and it responded with hallucinated papers. My first clue that something wasn't quite right was when one of the papers (of which I was not previously aware) listed me as the first author...