Consulting firm quietly admitted to GPT-4o use after fake citations were found in August.
See full article...
See full article...
Funnily enough, the Australian government studied this and found that AIs were worse than humans at summarizing!Genuine question - that makes me question the value of the summaries. How can we know the summaries are correct without reading the paper itself? Is there any research on not just lost nuance, but the hallucinations in AI summaries? I'd be interested in seeing it across approaches, such as NotebookLM and Kagi, which can pin to a set of sources, or requests to summarize a single paper across different models.
I occasionally use AI to summarize things, but I don't trust it past summarizing things where the ultimate goal is to point me to the actual authoritative source when I'm having a hard time finding it, so I can verify the summary. Do you trust the summaries you get? And if so, why?
This. I absolutely cannot understand why a consultancy, who's entire business model is "pay us large sums of money for our expert's advice" would rely on a LLM for even as much as grammar advice.
If your expert is ChatGPT, why do I pay you? I can write prompts myself. This is an incredibly fast way to sink your entire business model- if I were McKinsey or one of the others I'd be out there advertising "We know what we're doing, we don't need AI to do it poorly"
Agreed. A partial refund merely lowers the incentive to do it again. What's really required is a financially painful fine to act as deterrent. For these clowns and others.I'm not familiar with the Australian legal system, but they should actually have to pay a penalty multiplier for fraud, in other words pay more than they received for the fraudulent report. It also appears that they started with the conclusion that they wanted, and are claiming that their fraud does not alter the conclusion.
Anybody thinking about shorting Deloitte?
Yet governments and corporations keep using Deloitte. It's as if they are as stupid as the LLMs.5. Pants were lost.
Given that Deloitte's business model is to hire a whole bunch of kids right out of college, kinda supervise them a little bit, and charge boatloads of money for whatever flows out, this result is entirely unsurprising. And this is far from the first time that shoddy work has gotten them in trouble
Perhaps there's a claim under right of publicity or fraud.While I am in favor of sticking it to people who use AI like this in any way possible, I think you'd probably find it hard to make the case that it's "defaming" her. Unless the made up papers are about her doing experiments on humans or something of that magnitude. While it's wrong, if someone cited a fake paper I wrote on an inconclusive drug trial, it'd be hard to show that it had somehow damaged my reputation.
I developed a cancer-style staging system for the spread of GenAI adoption.the substance of the independent review is retained, and there are no changes to the recommendations.
The problem with counting on LLMs for lit searches is that they often fail on a very key point, such as "Does this paper support or reject a particular hypothesis?". It can write a nice paragraph summarizing the paper, but is far too easily confused on the key points because it can't really distinguish key points from secondary ones.I want to be fair about this, because I generally think ChatGPT is a useful tool for lit searches and summaries of papers (as with any summary, some nuance is lost). However, once I asked it for sources on a certain topic and it responded with hallucinated papers. My first clue that something wasn't quite right was when one of the papers (of which I was not previously aware) listed me as the first author...
Don't these management consulting companies run on fresh-out-of-college "consultants" who have zero or close to zero real world experience?
ETA ninja'ed
i usually see it attributed to mark twain. (also, speaking of upton, people might want to (re)read The Jungle again. that was his novel about the meat packing industry that aimed to help meat packing workers but instead got the president to start the FDA. seems timely, what with rfk, jr and all that."It is difficult to get a man to understand something, when his salary depends on his not understanding it." Attributed to Upton Sinclair, but unverified.
So you're saying that these reports are written to very rigorous professional standards? NoInterviewer: This report that Deloitte produced for Australia's Department of Employment and Workplace Relations this week...
Deloitte executive: The one with the false citations?
Interviewer: Yeah.
Deloitte executive: Yeah, that’s not very typical, I’d like to make that point.
Interviewer: Well, how was it un-typical?
Deloitte executive: Well there are a lot of these reports going around the world all the
time, and very seldom does anything like this happen. I just don’t want people
thinking that Deloitte's reports have false citations.
Interviewer: Did this report have false citations?
Deloitte executive: Well, I was thinking more about the other ones.
...
Interviewer: This report that Deloitte produced for Australia's Department of Employment and Workplace Relations this week...
Deloitte executive: The one with the false citations?
Interviewer: Yeah.
Deloitte executive: Yeah, that’s not very typical, I’d like to make that point.
Interviewer: Well, how was it un-typical?
Deloitte executive: Well there are a lot of these reports going around the world all the
time, and very seldom does anything like this happen. I just don’t want people
thinking that Deloitte's reports have false citations.
Interviewer: Did this report have false citations?
Deloitte executive: Well, I was thinking more about the other ones.
...
Sounds like Robodebt v2.0 is in preparation."Deloitte's "Targeted Compliance Framework Assurance Review"...focuses on the technical framework the government uses to automate penalties under the country's welfare system.".
I am not assured that the penalties will be compliant. Take them to court if you can.
Greed-coated lenses are 100% opaque.Earlier this year, Deloitte declared it would start using generative AI for its reports as a way of enhancing the value provided to its clients. I don't remember if they said it in a specific report or not, but I recall seeing it.
The citation issue continues to trip people up across the spectrum, from lawyers to business analysts. It's striking how many supposedly smart people do not understand the limits of the tools they insist will deliver such amazing value.
And the beautiful revolving door:It looks like AI might destroy the consultancy industry. Why pay millions to a fancy consultant when one can ask an LLM to crank out an equally worthless report? Management pays consultants to justify decisions they have already made and provide a way to deflect blame when they go awry. It sounds a lot cheaper to fire up an LLM, get the nonsense one wants and have a dumb computer to blame.
Genuine question - that makes me question the value of the summaries. How can we know the summaries are correct without reading the paper itself? Is there any research on not just lost nuance, but the hallucinations in AI summaries? I'd be interested in seeing it across approaches, such as NotebookLM and Kagi, which can pin to a set of sources, or requests to summarize a single paper across different models.
I occasionally use AI to summarize things, but I don't trust it past summarizing things where the ultimate goal is to point me to the actual authoritative source when I'm having a hard time finding it, so I can verify the summary. Do you trust the summaries you get? And if so, why?
The bigger joke is that people actually do read these reports.
Is the purpose of this framework to save taxpayer money by improving the penalizing of welfare recipients who either aren’t eligible or didn’t follow the rules when applying? If so then how many welfare recipients need to be penalized to make up $440.000 AUD? I’d really like to see that calculation.The report, which cost Australian taxpayers nearly $440,000 AUD (about $290,000 USD), focuses on the technical framework the government uses to automate penalties under the country's welfare system.
Yeah, I was skeptical. I baselined this by using ChatGPT to summarize papers I'd already read, or on topics I'm already an expert in. Usually the summary was a re-wording of the abstract, with maybe some additional context from the paper. I've found these summaries to be pretty good, but you have to check.
In the case I mentioned, the statement ChatGPT made about the topic is very probably correct, but the sources for the correct statement were made up.
This is a theme -- I've asked ChatGPT how to make calculations using a code that is popular in my field (and which I've used for about a decade), just as a curiosity. It is correct about what the code ought to able to do, and it is even correct about how the code would do it, but it totally makes up the keywords you'd need to set in the input file to make this happen.
Quite bold of Deloitte to use it without checking. It's a trivial thing to do. If I were a consulting firm in the age of generative AI, I would want to be very clear on how I add value to AI -- here it is deeply unclear how Deloitte did.
Yeah, law firms who allow this stupid shit have money and lazy attorneys. Fining them is just the cost of doing business.Perhaps Lisa Crawford has a case for defamation or slander for having these false papers attributed to her. One way to stop the nonsense is to make it hurt. As it airs they are partially refunding the money but clearly all they did is engineer a few AI prompts to get the report. Make them refund it all, make them pay for defamation and send a message that this crap isn’t okay.
Same for the lawyers who submit briefs to the court with fake legal citations.
You have some spelling errors, there.LLMs, the gift that keeps giving.
Not necessarily "people" in this case, but "a consultancy." And the answer is because it was cheaper for them to do it, if they got away with it.Why do people apparently hate thinking for themselves so much that they immediately abdicate that responsibility to an algorithm at the first opportunity?
Agent Smith was right
Deloitte does a good job with the UEFA Money League webpage detailing the revenue of the top football clubs in Europe.This has not been a good time for Deloitte; this mess with AI and the mess their former CEO is making in the WNBA. Anybody thinking about shorting Deloitte? For a professional services organization, using AI to do professional work, or the notorious work of its former CEO Cathy Engelbert sure isn't helping.
I would just say that “consulting” is such a broad term, that there is room for all kinds of outcomes. At one company I worked at, they brought in a consultant to make our undergraduate recruiting process better. The process they came up with worked great, but only for the first year after it was introduced. The problem was that the process required training and coordination among the people assigned to interview undergraduates each year. The company only bothered with the training that first year. Didn’t even bother providing the useful pre-printed forms the following year.View attachment 119666
As true as it's ever been.
Fun story involving non-Deloitte consultants: I was vehemently opposed to a project being forced through at my last company, not because I thought the goal was bad (in fact it was actually a good goal, if the process of implementation was done by competent people), but because the consulting company and integrator for some of the stuff was setting off all sorts of alarm bells in my talks with them. One of my last major acts of resistance was putting together a slideshow since execs love those, and giving a succinct list of issues that were going to be encountered if they took this approach, and the upfront and predicted long-term effects of those issues.
The astute reader will guess that yes, they did proceed with reckless abandon, shortly after I departed that company. One of my coworkers kept a copy of the presentation, and as they ran into each issue, annotated that with the date they did run into it, and some little screenshots from slack and emails of people freaking out, and then when it was all done, he realized it was 12 slides long, and printed it as a calendar and sent it to me![]()
that was used as part of the technical workstream to help "[assess] whether system code state can be mapped to business requirements and compliance needs
You found the older one, but there’s also this one from last month that concludes AI is bad at summarizing science.Funnily enough, the Australian government studied this and found that AIs were worse than humans at summarizing!
https://meincmagazine.com/ai/2024/09/...-ai-is-much-worse-than-humans-at-summarizing/
How much of that "amazing value" is offset by the negative PR and refunds to clients after the AI use and consequent hallucinations are made public?Earlier this year, Deloitte declared it would start using generative AI for its reports as a way of enhancing the value provided to its clients. I don't remember if they said it in a specific report or not, but I recall seeing it.
The citation issue continues to trip people up across the spectrum, from lawyers to business analysts. It's striking how many supposedly smart people do not understand the limits of the tools they insist will deliver such amazing value.