I totally agree on the hallucination front, that is a big problem.
But the privacy thing is less of a concern if we are only talking about "using" (not training) these medical models. You could upload your entire medical history to such a model within the context window, and then provide your latest symptoms to get a diagnostic from the LLM. The LLM has not remembered any information about you at this time (it isn't "within" the model itself), it is only in the context window.
Anyway, just pointing out that there is a world of difference between training these models and using these models. If you are only using, there is really very little fear of privacy assuming the whole interaction is discarded.
Sure, but have you seen the workload of most doctors? A huge portion of it is simply writing notes on patients.
Note writing is ripe for an LLm, even with hallucinations. Write out your quick bullet points, let the LLM expand that content, re-read and correct. Since a human remains in the loop, it doesn’t have to be perfect.
Have you tried GPT-4? Haven't seen it hallucinate once out of the 100+ questions I've asked it.
The one mistake I've seen it make was giving me the wrong pin number when asking it to read a datasheet via a plugin, but even then it was a real pin and I wasn't using core GPT-4.
> Mycin was never actually used in practice. This wasn't because of any weakness in its performance - in tests it outperformed members of the Stanford medical school. It was as much because of ethical and legal issues related to the use of computers in medicine - if it gives the wrong diagnosis, who do you sue?
A lot of medicine is about pattern recognition and in that sense this could help just as it does for software development by boosting newer doctors who are early in their careers and building their intuition. I've seen this gap first hand even with doctors from the best schools and residency programs.
Though, it's probably going to take a few more iterations before its ready for prime time.
I say this as a non medical person but one with three doctors in my immediate family
I think it could also help experienced doctors who may be a bit stuck in their way of thinking and any doctor trying to diagnose rarer diseases and disorders. so many medical issues seem to have overlapping or identical symptoms so it would certainly help to remind doctors of the more rare options they may need to test for.
It seems inevitable in the future I won't have a choice whether my info (however personal) gets sucked up into some corporate database via AI chatbots. Even if I were to refuse to use one personally, I have no control whether the nurse I talk to is just typing what I tell them into a chatbot on their laptop.
Don't be distracted by today's excitement over chatbots. Companies are out there consolidating this data right now through traditional methods. Oracle, for example, has been saying the quiet part out loud.
FWIW, healthcare has strong regulations around data. If your provider uses an approved chatbot (vs using public ChatGPT), there is more or less a guarantee that your data won't be sucked up by a corp.
I'd rather my doctor use a negotiated and approved model/tool than an unapproved one.
> there is more or less a guarantee that your data won't be sucked up by a corp
Not true. "The HIPAA Privacy Rule establishes the conditions under which protected health information may be used or disclosed by covered entities for research purposes" [1]. Your de-identified data can be sucked up by a corp for research purposes. You are protected because your identity will be removed before data is used for research.
Or even the doctor. They could consult with a chatbot rather than other doctors in refining their own opinion. I think that is more likely in developed countries, where this is more likely going to be a tool to be used by doctors directly, a more efficient way to sift through 10s if thousands if maybe relevant published studies and literature.
this is a great comment for a clear demonstration of how hard it is to make real boundaries for humans and their healthcare services. "antivax" is a policy statement, not a medical statement, yet the comment is likely in good faith. The blurry gray lines involve humans being a social animal, not just a single body; administration, rules of law, economics are all part of "policy"
There’s a push to privatize healthcare in Canada in the last few years. I worry the public system will crumble into an AI driven, grifter run trash heap and we’ll get the worst of everything where taxes stay high, because we have “free” healthcare, and real healthcare is private and only accessible to the rich.
>According to Google senior research director Greg Corrado, WSJ says, Med-PaLM 2 is still in its early stages. Corrado said that while he wouldn’t want it to be a part of his own family’s “healthcare journey,” he believes Med-PaLM 2 “takes the places in healthcare where AI can be beneficial and expands them by 10-fold.”
Google is suggesting this should be used in poor countries without enough doctors because they know it is (or essentially will be, after the tidal wave of malpractice suits happens) illegal in the US.
How is it illegal in the US? For one, it seems that it does conform to HIPPA. As per lawsuits, if the doctor is using it, they can't claim the chat bot gave them incorrect information - they should know better.
Devil's advocate: most health problems in developing nations are the basic stuff we solved 100 years ago in the west. An AI that can detect when to give someone a dewormer or some vitamines or something else without occupying a doctor could be very helpful
Not if the ai is hallucinating. Imagine if your doctor was extremely knowledgeable, but extremely high on acid every day and prone to confidently lying to you
Diagnosis is usually the easiest part of treating such conditions. The problems are more in lack of medical supplies, facilities, and personnel to administer treatment.
Double penalties. You are poor so you don't get good care and you also get to be a cobaye for a tech that can randomly (wrongly) guess your health problems.
And there it is. You aren’t worried it will make an error, you believe it will always be wrong. Since doctors misdiagnose all the time anyways, just claiming that it might misdiagnose wouldn’t be strong enough to disavow it (since it almost certainly will misdiagnose at least as much as a human doctor).
This idea that because a human is sometimes wrong, it's okay for the chatbot to be wrong is one I find to be dangerous and unhelpful. The fact that people are fallible shouldn't encourage us to choose solutions with a high or unpredictable error rate. Some failures have a greater cost (i.e. launching a rocket) than others (delivering a package).
Does the chatbot misdiagnose less than a doctor? When the chatbot does misdiagnose, is it in manner similar to a doctor? In the same way that AI generated images always had weird and creepy fingers, does this chatbot consistently misdiagnose the common cold as brain cancer?
IMHO, without some real data, we're simply experimenting on people. That is wholly immoral.
After some contradictory information on error rates and failure modes, a Google employee is quoted in the article:
'Corrado said that while he wouldn’t want it to be a part of his own family’s “healthcare journey,” he believes Med-PaLM 2 “takes the places in healthcare where AI can be beneficial and expands them by 10-fold.”'
Yikes! 10-fold expansion, excluding any interaction with people this person really cares about. Let us remember that this is a 10x times increase over the current "places where chatbot are beneficial in healthcare", that is to say zero.
> Tell that to someone when their kid dies because some chatbot hallucinated the wrong diagnosis or treatment.
MedPalm seems to be accessible only to healthcare facilities. If the chatbot hallucinated the wrong diagnosis or treatment, the doctor will still be liable.
Let's not do that (but still keep in mind, that it is true). An individuals understandable inability to deal with statistical truth in times of personal hardship should not stand in the way of the best medical care, statistically.
I barely even trust LLMs anymore to help me with orientating research, I keep finding out that half of what it says is a lie