The cases spanned five counties, but they all had one thing in common: everyone had gone to the Brown County fair.
Dr Mole said:Some of the questions are easy enough to answer without a chatbot. A simple search on PubMed, a federal database of scientific literature, quickly pulls up examples of Salmonella being found in ice, for example.
I guess if you use it like a crappier wikipedia, it can be a good source of references if you ask for them and do a bunch of reading of said references yourself.This is the incredible danger of current LLM models. They use incredibly compelling language to assert confidence where the system itself literally IS NOT CAPABLE OF. Yes, the LLM said that ice was a "credible and likely" source, but ChatGPT isn't really able to do that, what it is doing instead is predicting that the words credible and likely are the most appropriate next words in a response!
Even if you know this is the major flaw of LLMs, it's really easy to fail to correct for that false assertion of confidence. Humans are creatures of language, and we're "programmed" to interpret confident language as evidence of knowledge and expertise. Even experts in the field (and TBH a health department should be an expert in public health outbreaks) can obviously be fooled to rely on LLM assertions because of this.
The tendency of LLMs to just make references up is pretty well known and has ended many a legal and consulting career already.if you ask for them and do a bunch of reading of said references yourself.
Thank you! Every time I hear about people wanting to use LLMs as a medical search engine I wonder if I've spent the past two decades hallucinating this tool that we already have!
I've had a few ongoing, very minor medical issues that I've mentioned to doctors with no success (Seborrheic dermatitis is one I've had for years and years). They usually shrugged their shoulders and said, "That’s weird," and didn't offer a helpful suggestion. I gave the symptoms to ChatGPT, and it diagnosed the problem right away and suggested an over-the-counter treatment which worked. It was honestly pretty amazing. I’m not saying this is a substitute for real doctors, and I’m sure a specialist would have diagnosed the same thing. But as a supplement to medical professionals, there’s value, I reckon.
Name checks out
That phrase is a sure butt-clencher.
I agree with your overall point but one clarification, it is not responding with “credible and likely” just because that is statistically a good response in general. Otherwise it would never respond in the negative, which it does (ask if it is possible if the salmonella is present due to spontaneous generation). It is generating the statistically likely response based on the input provided.This is the incredible danger of current LLM models. They use incredibly compelling language to assert confidence where the system itself literally IS NOT CAPABLE OF. Yes, the LLM said that ice was a "credible and likely" source, but ChatGPT isn't really able to do that, what it is doing instead is predicting that the words credible and likely are the most appropriate next words in a response!
Even if you know this is the major flaw of LLMs, it's really easy to fail to correct for that false assertion of confidence. Humans are creatures of language, and we're "programmed" to interpret confident language as evidence of knowledge and expertise. Even experts in the field (and TBH a health department should be an expert in public health outbreaks) can obviously be fooled to rely on LLM assertions because of this.
That's something I've been finding more than a little annoying about AI assistants. They feign cheer for helping and tell me everything is the greatest, most powerful, most superlative ever. They are the cheerful all-knowing assistant, Even when I instruct it to be objective, I still get that sense it is patronizing me. I would like it a lot more if it would just generate a flat response without trying to engage my enthusiasm.This is the incredible danger of current LLM models. They use incredibly compelling language to assert confidence where the system itself literally IS NOT CAPABLE OF. Yes, the LLM said that ice was a "credible and likely" source, but ChatGPT isn't really able to do that, what it is doing instead is predicting that the words credible and likely are the most appropriate next words in a response!
Even if you know this is the major flaw of LLMs, it's really easy to fail to correct for that false assertion of confidence. Humans are creatures of language, and we're "programmed" to interpret confident language as evidence of knowledge and expertise. Even experts in the field (and TBH a health department should be an expert in public health outbreaks) can obviously be fooled to rely on LLM assertions because of this.
And LLMs remain a solution in search of a problem.Some of the questions are easy enough to answer without a chatbot.
Unfortunately, it appears this tendency may have hit an Ars writer's career as well.The tendency of LLMs to just make references up is pretty well known and has ended many a legal and consulting career already.
That's something I've been finding more than a little annoying about AI assistants. They feign cheer for helping and tell me everything is the greatest, most powerful, most superlative ever. They are the cheerful all-knowing assistant, Even when I instruct it to be objective, I still get that sense it is patronizing me. I would like it a lot more if it would just generate a flat response without trying to engage my enthusiasm.
Although this technique did not follow a traditional surveillance protocol, AI was effective in this rural setting for rapid situational awareness and early case finding, especially because formal case reporting was delayed or limited.
And thirdly because elsewhere in the report it directly contradicts that first quote:In a small community, monitoring social media posts and photos, as well as personally contacting fair board members and persons who health department staff members had encountered at the fair, contributed to rapid situational awareness and early case finding but also contributed to reluctance to report, for fear of implicating a friend or neighbor as contributing to the outbreak.
AI was not used for case finding...
This is a worthwhile clarification, but shouldn't be confused for definitively indicating that the association of "credible and likely" is tied to the most important concepts in the prompt.I agree with your overall point but one clarification, it is not responding with “credible and likely” just because that is statistically a good response in general. Otherwise it would never respond in the negative, which it does (ask if it is possible if the salmonella is present due to spontaneous generation). It is generating the statistically likely response based on the input provided.
MMWR used to be an incredibly rigorous public health publication, but it's a part of the CDC, and the CDC staff has been cut to the bone so I don't think it's abnormal that publication standards have dropped quite a bit.Is it just me or is that weekly report something of a mess? I'm not convinced the bit about the effectiveness of AI is accurate at all. Firstly because it doesn't really make sense (how did ChatGPT help with situational awareness?), secondly because the key phrase is duplicated elsewhere in the report:
And thirdly because elsewhere in the report it directly contradicts that first quote:
I'd rather ask the furry community. Similar track record...I've had a few ongoing, very minor medical issues that I've mentioned to doctors with no success (Seborrheic dermatitis is one I've had for years and years). They usually shrugged their shoulders and said, "That’s weird," and didn't offer a helpful suggestion. I gave the symptoms to ChatGPT, and it diagnosed the problem right away and suggested an over-the-counter treatment which worked. It was honestly pretty amazing. I’m not saying this is a substitute for real doctors, and I’m sure a specialist would have diagnosed the same thing. But as a supplement to medical professionals, there’s value, I reckon.
In this case, the LLM got it right. But the answer here was seemingly so obvious it's a little mystifying why county health investigators needed to ask an LLM at all.This is a worthwhile clarification, but shouldn't be confused for definitively indicating that the association of "credible and likely" is tied to the most important concepts in the prompt.
Given the nature of LLMs, it's very, very likely a high probability that this is a sensible association with the prompt as a whole, but the weight of the various words in the prompt could also adjust how "correct" that response is (both toward a very "proper" weighting conceptually, or one that is leaning on the less important bits)
Sure, PubMed another tools may be available if we simply go use[...]
Well yeah, that's why I said you find and read the references. You don't just believe it because it 'has references'.The tendency of LLMs to just make references up is pretty well known and has ended many a legal and consulting career already.
And just a reminder here about how good LLMs are for medical issues:Thank you! Every time I hear about people wanting to use LLMs as a medical search engine I wonder if I've spent the past two decades hallucinating this tool that we already have!
And it matters. Infectious disease specialists, epidemiologists, and gastroenterologists are going to have a much better grasp of the essential context. A chatbot that pulls some answer out of the aether is useless because I need to know that source!
large makeshift cooler, described as being made of “a 10-ft length of non-food-grade corrugated black plastic farm drainage tile with four internal compartments.” It was reportedly hosed off at the start of the fair, but then never fully drained or cleaned again.
There are plenty of fallacies with using LLMs, but that really isn't one of them. The big breakthrough with the attention mechanism (from the Google paper, "Attention is all you need") that enabled LLMs to enter mainstream in the first place was precisely that it enabled distinguishing different meanings of the same, or similar, words and fragments depending on the context they appear in. It is the reason why modern LLM based translation software are much more likely to accurately preserve the meaning of texts when translating from one language to another that than any previous machine translation method.Because LLMs only preserve the connections between word-pieces, and not the actual meaning, you can have situations where an LLM considers "hyper" and "hypo" to be words that are both associated with a given suffix. And then it will take that suffix and determine the next token.
That's because an LLM has no motive ...