ChatGPT generates cancer treatment plans that are full of errors — Study finds that ChatGPT provided false information when asked to design cancer treatment plans (www.businessinsider.com)
from L4s@lemmy.world to technology@lemmy.world on 26 Aug 2023 16:00
https://lemmy.world/post/3882690

ChatGPT generates cancer treatment plans that are full of errors — Study finds that ChatGPT provided false information when asked to design cancer treatment plans::Researchers at Brigham and Women’s Hospital found that cancer treatment plans generated by OpenAI’s revolutionary chatbot were full of errors.

#technology

threaded - newest

autotldr@lemmings.world on 26 Aug 2023 16:00 next collapse

This is the best summary I could come up with:


According to the study, which was published in the journal JAMA Oncology and initially reported by Bloomberg – when asked to generate treatment plans for a variety of cancer cases, one-third of the large language model’s responses contained incorrect information.

The chatbot sparked a rush to invest in AI companies and an intense debate over the long-term impact of artificial intelligence; Goldman Sachs research found it could affect 300 million jobs globally.

Famously, Google’s ChatGPT rival Bard wiped $120 billion off the company’s stock value when it gave an inaccurate answer to a question about the James Webb space telescope.

Earlier this month, a major study found that using AI to screen for breast cancer was safe, and suggested it could almost halve the workload of radiologists.

A computer scientist at Harvard recently found that GPT-4, the latest version of the model, could pass the US medical licensing exam with flying colors – and suggested it had better clinical judgment than some doctors.

The JAMA study found that 12.5% of ChatGPT’s responses were “hallucinated,” and that the chatbot was most likely to present incorrect information when asked about localized treatment for advanced diseases or immunotherapy.


The original article contains 523 words, the summary contains 195 words. Saved 63%. I’m a bot and I’m open source!

zeppo@lemmy.world on 26 Aug 2023 16:09 next collapse

I’m still confused that people don’t realize this. It’s not an oracle. It’s a program that generates sentences word by word based on statistical analysis, with no concept of fact checking. It’s even worse that someone actually did a study instead of simply acknowledging or realizing that ChatGPT is happy to just make stuff up.

fubo@lemmy.world on 26 Aug 2023 16:17 next collapse

It’s even worse that someone actually did a study instead of simply acknowledging or realizing that ChatGPT is happy to just make stuff up.

Sure, the world should just trust preconceptions instead of doing science to check our beliefs. That worked great for tens of thousands of years of prehistory.

zeppo@lemmy.world on 26 Aug 2023 16:21 next collapse

It’s not merely a preconception. It’s a rather obvious and well-known limitation of these systems. What I am decrying is that some people, from apparent ignorance, think things like “ChatGPT can give a reliable cancer treatment plan!” or “here, I’ll have it write a legal brief and not even check it for accuracy”. But sure, I agree with you, minus the needless sarcasm. It’s useful to prove or disprove even absurd hypotheses. And clearly people need to be definitely told that ChatGPT is not always factual, so hopefully this helps.

adeoxymus@lemmy.world on 26 Aug 2023 17:24 collapse

I’d say that a measurement always trumps arguments. At least you know how accurate they are, this statement cannot follow from reason:

The JAMA study found that 12.5% of ChatGPT’s responses were “hallucinated,” and that the chatbot was most likely to present incorrect information when asked about localized treatment for advanced diseases or immunotherapy.

zeppo@lemmy.world on 26 Aug 2023 17:38 collapse

That’s useful. It’s also good to note that the information the agent can relay depends heavily on the data used to train the model, so it could change.

PetDinosaurs@lemmy.world on 26 Aug 2023 17:21 next collapse

Why the hell are people down voting you?

This is absolutely correct. We need to do the science. Always. Doesn’t matter what the theory says. Doesn’t matter that our guess is probably correct.

Plus, all these studies tell us much more than just the conclusion.

yiliu@informis.land on 26 Aug 2023 17:26 next collapse

“After an extensive three-year study, I have discovered that touching a hot element with one’s bare hand does, in fact, hurt.”

“That seems like it was unnecessary…”

“Do U even science bro?!”

Not everything automatically deserves a study. Were there any non-rando people out there claiming that ChatGPT could totally generate legit cancer treatment plans that people could then follow?

Takumidesh@lemmy.world on 26 Aug 2023 17:51 next collapse

It’s not even a preconception, it’s willful ignorance, the website itself tells you multiple times that it is not accurate.

The bottom of every chat has this text: “Free Research Preview. ChatGPT may produce inaccurate information about people, places, or facts. ChatGPT August 3 Version”

And when you first use it, a modal pops up explaining the same thing.

Windex007@lemmy.world on 26 Aug 2023 20:47 collapse

ChatGPT isn’t some newly discovered sentient species.

It’s a machine designed and built by human engineers.

This is like suggesting that we study fortune cookies to see if they can accurately forecast the future. The manufacturer can simply tell you the limitation of their product… Being that they can not divine the future.

Zeth0s@lemmy.world on 26 Aug 2023 16:22 next collapse

Publish or perish, that’s why

agressivelyPassive@feddit.de on 26 Aug 2023 16:40 collapse

I’m trying really hard for the latter.

net00@lemm.ee on 26 Aug 2023 17:19 next collapse

Yeah this stuff was always marketed to automate simple and repetitive things we do daily. it’s mostly the media I guess who started misleading everyone into thinking this was AI like skynet. It’s still useful, not just as a all knowing AI god

inspxtr@lemmy.world on 26 Aug 2023 17:23 next collapse

while I agree it has become more of a common knowledge that they’re unreliable, this can add on to the myriad of examples for corporations, big organizations and government to abstain from using them, or at least be informed about these various cases with their nuances to know how to integrate them.

Why? I think partly because many of these organizations are racing to adopt them, for cost-cutting purposes, to chase the hype, or too slow to regulate them, … and there are/could still be very good uses that justify it in the first place.

I don’t think it’s good enough to have a blanket conception to not trust them completely. I think we need multiple examples of the good, the bad and the questionable in different domains to inform the people in charge, the people using them, and the people who might be affected by their use.

Kinda like the recent event at DefCon trying to exploit LLMs, it’s not enough we have some intuition about their harms, the people at the event aim to demonstrate the extremes of such harms AFAIK. These efforts can help inform developers/researchers to mitigate them, as well as showing concretely to anyone trying to adopt them how harmful they could be.

Regulators also need these examples in specific domains so they may be informed on how to create policies on them, sometimes building or modifying already existing policies of such domains.

zeppo@lemmy.world on 26 Aug 2023 17:45 next collapse

This is true and well-stated. Mainly what I wish people would understand is there are current appropriate uses, like ‘rewrite my marketing email’, but generating information that could result in great harm if inaccurate is an inappropriate use. It’s all about the specific model, though - if you had a ChatGPT system trained extensively on medical information, it would result in greater accuracy, but still the information would need expert human review before any decision were made. Mainly I wish the media had been more responsible and accurate in portraying these systems to the public.

jvisick@programming.dev on 26 Aug 2023 19:50 collapse

I don’t think it’s good enough to have a blanket conception to not trust them completely.

On the other hand, I actually think we should, as a rule, not trust the output of an LLM.

They’re great for generative purposes, but I don’t think there’s a single valid case where the accuracy of their response should be outright trusted. Any information you get from an AI model should be validated outright.

There are many cases where a simple once-over from a human is good enough, but any time it tells you something you didn’t already know you should not trust it and, if you want to rely on that information, you should validate that it’s accurate.

iforgotmyinstance@lemmy.world on 26 Aug 2023 18:41 next collapse

I know university professors struggling with this concept. They are so convinced using an LLM is plagiarism.

It can lead to plagiarism if you use it poorly, which is why you control the information you feed it. Then proofread and edit.

zeppo@lemmy.world on 26 Aug 2023 18:55 next collapse

Another related confusion in academia recently is the ‘AI detector’. It could easily be defeated with minor rewrites, if they were even accurate in the first place. My favorite misconception is there was a story of a professor who told students “I asked ChatGPT if it wrote this, and it said yes” which is just really not how it works.

ZodiacSF1969@sh.itjust.works on 27 Aug 2023 01:32 collapse

I can understand the plagiarism argument, though you have to extend the definition of it. If I am expected to write an essay, but I use ChatGPT instead, then I am fraudulently presenting the work as my own. Plagiarism might not be the right word, or maybe it’s a case where language is going to evolve so that plagiarism includes passing off AI generated work as your own. Either way it’s cheating unless I was specifically allowed to use AI.

iforgotmyinstance@lemmy.world on 27 Aug 2023 02:05 collapse

If the argument and the sources are incongruous, that isn’t the fault of the LLM/AI. That’s the authors fault for not proofreading and editing.

You assume an inherent morality of LLMs but they are amoral constructs. They are tools, and you limit yourself by not learning them.

ZodiacSF1969@sh.itjust.works on 28 Aug 2023 04:27 collapse

I didn’t say anything about the sources being incongruent? That’s a completely separate issue. We were talking about plagiarism.

I don’t understand the morality comment either, I didn’t ascribe any morality to AI, I was talking about whether using them fits the definition of plagiarism or not.

If you are expected to write it yourself, and you use an LLM to generate it, then that’s cheating in my opinion. Yes, of course we shoukd learn to use AI, but if you are told to do something and you get a person or LLM to do it for you, then you didn’t complete the task as you were told. And at university that can have consequences.

PreviouslyAmused@lemmy.ml on 26 Aug 2023 21:25 next collapse

But it’s supposed to be the future! I want the full talking spaceship like in Star Trek, not this … “initial learning steps” BS!

I was raised on SciFi and am now mad that I don’t have all the cool made up things from those shows/movies!

dual_sport_dork@lemmy.world on 26 Aug 2023 22:32 next collapse

This is why without some hitherto unknown or so far undeveloped capability of these sorts of LLM models, they’ll never actually be useful for performing any kind of mission critical work. The catch-22 is this: You can’t trust the AI to produce correct work without some kind of potentially dangerous, showstopping, or embarassing error. This isn’t a problem if you’re just, say, having it paint pictures. Or maybe even helping you twiddle the CSS on your web site. If there is a failure here, no one dies.

But what if your application is critical to life or safety? Like prescribing medical care, or designing a building that won’t fall down, or deciding which building the drone should bomb. Well, you have to get a trained or accredited professional in whatever field we’re talking about to check all of its work. And how much effort does that entail? As it turns out, pretty much exactly as much as having said trained or accredited professional do the work in the first place.

nfsu2@feddit.cl on 27 Aug 2023 17:09 collapse

true, I tried to explain this to my parents because they were scared of it and they seemed skeptical.

imperator3733@lemmy.world on 26 Aug 2023 16:16 next collapse

No duh - why would it have any ability to do that sort of task?

PeleSpirit@lemmy.world on 26 Aug 2023 16:39 next collapse

Because if it’s able to crawl all of the science pubs, then it would be able to try different combos until it works. Isn’t that how it could/is being used, to test stuff?

Ranessin@feddit.de on 26 Aug 2023 17:16 next collapse

It doesn’t check the stuff it generates other than on grammatical and orthographical errors. It’s not intelligent or has knowledge outside of how to create text. The text looks useful, but it doesn’t know what it contains in a way something intelligent would.

PeleSpirit@lemmy.world on 26 Aug 2023 17:18 next collapse

It seems like it could check for that though, which is what chatgpt doesn’t do but we all assumed would. I’m sure there are ai programs that could and do check for possibilities on only information we know to be true.

ZodiacSF1969@sh.itjust.works on 27 Aug 2023 01:55 collapse

People who understand the technology did not assume that, but yes the general public has a lot of misconceptions about it.

fsmacolyte@lemmy.world on 28 Aug 2023 06:16 collapse

Recent papers have shown that LLMs build internal world models but about a topic as niche and complicated as cancer treatment, a chatbot based on GPT-3.5 be woefully ill-equipped to do any kind of proper reasoning.

stephen01king@lemmy.zip on 26 Aug 2023 20:27 collapse

If you want an AI that can create cancer treatment, you need to train it on creating cancer treatment, and not just use one that is trained on general knowledge. Even if you train it on science publications, all it can now reliably do is mimic a science journal since it has not been trained on how to parse the knowledge in the journal itself.

amki@feddit.de on 26 Aug 2023 23:22 next collapse

Which is exactly the problem people think has been solved but isn’t anywhere near being solved. It cannot comprehend semantics, the meaning of things is completely beyond it and all other AIs.

Unfortunately saying I made a thing that creates vaguely human looking speech with little content isn’t astonishing to most people hence they are looking for something useful this breakthrough machine must be able to do and then they don’t find anything leading to these articles.

PeleSpirit@lemmy.world on 26 Aug 2023 23:23 collapse

Right, but can’t they tell it to also try thousands and thousands of combos that humans could never do? I think ChatGPT is both super amazing and as stupid as a rock at the same time. I thought the vaccine used an AI to do that. I’m obviously clueless, I’m seriously asking.

ZodiacSF1969@sh.itjust.works on 27 Aug 2023 01:54 collapse

I don’t know about AI, but there are already computer programs that try many different combinations of, for example, chemical structures with known pharmacological properties and then output new drugs that could possibly be used to treat something. Of course you have to verify with research and studies.

I’m sure there will be AI’s or machine learning programs, if not already, that can do this as well and perhaps improve upon the process. But they would need to be specifically trained for that purpose. ChatGPT is a LLM, it’s made to generate language that fits a given prompt, I would not expect it to be great at creating cancer treatments and I’m not sure why we needed a study to learn that. OpenAI tells you already that the results can be inaccurate or outright wrong.

PeleSpirit@lemmy.world on 27 Aug 2023 13:46 collapse

I’m in Seattle and surrounded by people who are techy while not being techy myself, so the innovations they talk about are mind blowing. I thought ChatGPT at first was like all the other tech I heard about. But when you think about it, they would never release that for free first of all, and it would be too powerful for evil people. I was just letting people know what a non-techy thought.

xkforce@lemmy.world on 26 Aug 2023 16:44 collapse

Part of the reason for studies like this is to debunk peoples’ expectations of AI’s capabilities. A lot of people are under the impression that cgatGPT can do ANYTHING and can think and reason when in reality it is a bullshitter that does nothing more than mimic what it thinks a suitable answer looks like. Just like a parrot.

TenderfootGungi@lemmy.world on 26 Aug 2023 16:38 next collapse

The computer science classroom in my high school had a poster stating: “Garbage in garbage out”

unreachable@lemmy.my.id on 26 Aug 2023 16:44 next collapse

chatgpt/bard is only the next iteration of MegaHAL

that’s why they called it “large language model”, not “artificial intelligent”

IceMan@lemmy.one on 26 Aug 2023 16:55 next collapse

Google is full of bullshit info too - what’s the big deal?

j4yt33@feddit.de on 26 Aug 2023 17:05 next collapse

Why would you ask it to do that in the first place??

dmonzel@lemmy.world on 26 Aug 2023 17:33 collapse

To prove to all of the tech bros that ChatGPT isn’t an actual AI, perhaps. At least that’s the feeling I get based on what the article says.

eager_eagle@lemmy.world on 26 Aug 2023 17:17 next collapse

<img alt="" src="https://lemmy.world/pictrs/image/fcbe76c5-ed9a-4d2f-8fd0-6715c1ad4e54.jpeg">

SirGolan@lemmy.sdf.org on 26 Aug 2023 20:53 collapse

I agree, but only because they used GPT 3.5 and not 4. Not that I think 4 would have been perfect or that you should follow medical advice from LLMs right now, but it would have been much more accurate.

eager_eagle@lemmy.world on 26 Aug 2023 21:14 collapse

I thought it was 4 before looking into it. It seems they ran the experiment in the same month GPT-4 was released, that’s unfortunate.

SirGolan@lemmy.sdf.org on 27 Aug 2023 23:43 collapse

Yeah what’s interesting is it was just published this week even though they did the tests in April. April is like 100 AI years ago.

eager_eagle@lemmy.world on 27 Aug 2023 23:50 collapse

That’s actually a short time for peer-reviewed work. Most things being published now, unless they’re in pre-print servers, are results from studies that likely started at least 6 months ago.

alienanimals@lemmy.world on 26 Aug 2023 17:22 next collapse

Clickbait written by an idiot who doesn’t understand technology. I guess they give out journalism degrees to anyone who can write a top 10 buzzfeed article.

elboyoloco@lemmy.world on 26 Aug 2023 17:45 next collapse

Scientist: Askes question to magic conch about cancer.

Conch: “Trying shoving bees up your ass.”

Scientists: 😡

PreviouslyAmused@lemmy.ml on 26 Aug 2023 21:23 collapse

Scientist to world: “magic conch is useless in trying to cure cancer.”

News media: “Magic Conch is completely useless!”

Sanctus@lemmy.world on 26 Aug 2023 20:13 next collapse

These studies are for the people out there who think ChatGPT thinks. Its a really good email assistant, and it can even get basic programming questions right if you are detailed with your prompt. Now everyone stop trying to make this thing like Finn’s mom in adventure time and just use it to helo you write a long email in a few seconds. Jfc.

lonke@feddit.nu on 26 Aug 2023 20:50 next collapse

I use ChatGPT primarily for programming, and it’s particularly well suited for programming.

“Even get basic programming questions right if you are detailed with your prompt”

is underselling its capabilities in that regard. Especially GPT-4 has been able to help me with everything from obscure adobe ExtendScript scripts to infrequently seen ‘unsafe’ C# OpenGL perspective matrix math. All with prompts of a sentence maximum.

Sanctus@lemmy.world on 26 Aug 2023 21:05 collapse

I’m specifically referring to ChatGPT. GPT-4 is a different beast that I’m sure is quite adept.

lonke@feddit.nu on 27 Aug 2023 01:07 collapse

ChatGPT is GPT 3.5 & GPT 4, as far as I’m aware.

3.5 is also very capable when it comes to programming, for any well known framework or language. It’s not as capable, but it is still very capable.

Sanctus@lemmy.world on 27 Aug 2023 02:09 collapse

I found it was okay with Unity libraries but really good with things like Excel.Interop and business libraries, as well as general programming concepts like linked lists.

For instance, I made a random dungeon generator using Unity’s visual scripting. It seemed to be unaware of the visual scripting library. But I’m automating excel processes right now and its on point with those.

lonke@feddit.nu on 27 Aug 2023 14:26 collapse

Well, the data it was trained on had a cutoff point in 2021 which would explain that.

I’ve used it (GPT 3) a fair amount for Unity, and I’m fairly pleased with the results, it’s saved me a fair amount of time. Implementing object pooling and editor window dialogues for scene translation management for example.

Of course, programming knowledge is required for it to be of consistent use, which, on second thought, may not be at all obvious.

Sanctus@lemmy.world on 27 Aug 2023 19:09 collapse

Maybe Visual Scripting is on the cusp of its knowledge. I thought it released in 2021. It has replaced my rubber ducky in corporate environments thats for sure. I plan on using it again for game development after this discussion. My visual scripting use case was off the beaten path which was probably why it had a hard time.

Corngood@lemmy.ml on 26 Aug 2023 20:59 next collapse

I’m going to need it to turn those emails back into the bullet points used to create them, so I don’t have to read the filler.

Sanctus@lemmy.world on 26 Aug 2023 21:05 collapse

It can leave the bullets if you tell it.

EssentialCoffee@midwest.social on 26 Aug 2023 23:04 collapse

I use it for D&D. It’s fantastic at coming up with adventures, NPCs, story hooks, taverns, etc.

All of those things are made up.

Sanctus@lemmy.world on 27 Aug 2023 19:09 collapse

Its fantastic at that. I had it help me with a Dark Heresy session. Its not bad at generating names, places, and even personalities for jobbers.

SirGolan@lemmy.sdf.org on 26 Aug 2023 20:52 next collapse

What’s with all the hit jobs on ChatGPT?

Prompts were input to the GPT-3.5-turbo-0301 model via the ChatGPT (OpenAI) interface.

This is the second paper I’ve seen recently to complain ChatGPT is crap and be using GPT3.5. There is a world of difference between 3.5 and 4. Unfortunately news sites aren’t savvy enough to pick up on that and just run with “ChatGPT sucks!” Also it’s not even ChatGPT if they’re using that model. The paper is wrong (or it’s old) because there’s no way to use that model in the ChatGPT interface. I don’t think there ever was either. It was probably ChatGPT 0301 or something which is (afaik) slightly different.

Anyway, tldr, paper is similar to “I tried running Diablo 4 on my Windows 95 computer and it didn’t work. Surprised Pikachu!”

eggymachus@sh.itjust.works on 26 Aug 2023 21:11 collapse

And this tech community is being weirdly luddite over it as well, saying stuff like “it’s only a bunch of statistics predicting what’s best to say next”. Guess what, so are you, sunshine.

PreviouslyAmused@lemmy.ml on 26 Aug 2023 21:19 next collapse

I mean, people are slightly more complicated than that. But sure, at their most basic, people simply communicate with statistical models.

eggymachus@sh.itjust.works on 27 Aug 2023 08:26 collapse

Ok, maybe slightly :) but it surprises me that the ability to emulate a basic human is dismissed as “just statistics”, since until a year ago it seemed like an impossible task…

markr@lemmy.world on 27 Aug 2023 11:03 collapse

The dismissal is coming from the class of people most threatened by these systems.

dukk@programming.dev on 26 Aug 2023 21:24 next collapse

IMO for AI to reach a useful point it needs to be able to learn. Now I’m no expert on neural networks, but if it can’t learn anything new once it’s been trained, it’s never really going to reach its true potential. It can imitate a human, but that’s about it. Once AI can really learn, it’ll become an order of magnitude more useful. Don’t get me wrong: all this AI work is a step in the right direction, but we’ll only be able to go so far with pre-trained models.

eggymachus@sh.itjust.works on 27 Aug 2023 08:22 collapse

Absolutely agree that this is a necessary next step!

SirGolan@lemmy.sdf.org on 26 Aug 2023 22:56 next collapse

Hah! That’s the response I always give! I’m not saying our brains work the exact same way because they don’t and there’s still a lot missing from current AI but I’ve definitely noticed that at least for myself, I do just predict the next word when I’m talking or writing (with some extra constraints). But even with LLMs there’s more going on then that since the attention mechanism allows it to consider parts of the prompt and what it’s already written as it’s trying to come up with the next word. On the other hand, I can go back and correct mistakes I make while writing and LLMs can’t do that…it’s just a linear stream.

eggymachus@sh.itjust.works on 27 Aug 2023 08:20 collapse

Agree, I have definitely fallen for the temptation to say what sounds better, rather than what’s exactly true… Less so in writing, possibly because it’s less of a linear stream.

amki@feddit.de on 26 Aug 2023 23:13 collapse

Might be true for you but most people do have a concept of true and false and don’t just dream up stuff to say.

Dultas@lemmy.world on 27 Aug 2023 02:26 next collapse

Do they? *Laughs nervously in American.

eggymachus@sh.itjust.works on 27 Aug 2023 08:16 next collapse

Yeah, I was probably a bit too caustic, and there’s more to (A)GI than an LLM can achieve on its own, but I do believe that some, and perhaps a large, part of human consciousness works in a similar manner.

I also think that LLMs can have models of concepts, otherwise they couldn’t do what they do. Probably also of truth and falsity, but perhaps with a lack of external grounding?

markr@lemmy.world on 27 Aug 2023 11:01 collapse

Actually we ‘dream up’ things to say quite a lot. As in our unconscious functions are far more important to our mental processes than we like to admit. Also we are basically not very good at evaluating the truth value of complex expressions.

Rexios@lemm.ee on 26 Aug 2023 22:57 next collapse

Okay and? GPT lies how is this news every other day? Lazy ass journalists.

mwguy@infosec.pub on 26 Aug 2023 23:16 next collapse

I asked a retard to spend a week looking at medical treatment plans and related information on the internet. Then asked him to guestimate a treatment plan for my actual cancer patient. How could they have got it wrong!

This is how I translate all these AI Language model says bullshit, bullshit.

CombatWombat1212@lemmy.world on 27 Aug 2023 01:46 next collapse

When did they ever claim that was able to

quadropiss@lemmy.world on 27 Aug 2023 11:41 collapse

Idk but the title sounds like it without generate a bunch of clicks🥰

drekly@lemmy.world on 27 Aug 2023 02:27 next collapse

It speeds things up for people who know what they’re talking about. The doctor asking for the plan could probably argue a few of the errors and GPT will say “oh you’re right, I’ll change that to something better” and then it’s good to go.

Yes you can’t just rely on it to be right all the time, but you can often use it to find the right answer with a small conversation, which would be quicker than just doing it alone.

I recently won a client with GPTs help in my industry.

I personally think I’m very knowledgeable in what I do, but to save time I asked what I should be looking out for, and it gave me a long list of areas to consider in a proposal. That list alone was a great starting block to get going. Some of the list wasn’t relevant to me or the client, so had to be ignored, but the majority of it was solid, and started me out an hour ahead, essentially tackling the planning stage for me.

To someone outside of my industry, if they used that list verbatim, they would have brought up a lot of irrelevant information and covered topics that would make no sense.

I feel it’s a tool or partner rather than a replacement for experts. It helps me get to where I need to go quicker, and it’s fantastic at brainstorming ideas or potential issues in plans. It takes some of the pressure off as I get things done.

Prethoryn@lemmy.world on 27 Aug 2023 02:43 next collapse

Look, I am all for seeing pros and cons. A.I. has a massive benefit to humanity and it has its issues but this article is just silly.

Why in the fuck are you using ChatGPT to set a cancer plan? When did ChatGPT claim to be a medical doctor.

Just go see a damn doctor.

clutch@lemmy.ml on 27 Aug 2023 11:16 next collapse

The issue is hospital administrators thinking that AI is the answer to boost profits

Kage520@lemmy.world on 27 Aug 2023 12:09 collapse

I have been getting surveys asking my opinion on ai as a healthcare practitioner (pharmacist). I feel like they are testing the waters.

AI is really dangerous for healthcare right now. I’m sure people are using it to ask regular questions they normally Google. I’m sure administrators are trying to see how they can use it to “take the pressure off” their employees (then fire some employees to “tighten the belt”).

If they can figure out how to fact check the AI results, maybe my opinion can change, but as long as AI can convincingly lie and not even know it’s lying, it’s a super dangerous tool.

Prethoryn@lemmy.world on 27 Aug 2023 15:30 collapse

For me the issue isn’t the tool. It’s people. The tool is used just as it is. A tool.

I always like to compare these things to other physical tools. If you take a philips screw driver to a flathead screw you don’t blame the tool you blame yourself for bringing the improper tool because as a human you can make mistakes. As a human you should have figured out prior, “do I need a flathead or philips?” There are tools capable of doing the job and doing it properly.

Same if you are an operator on a piece of machinery. If you take a forklift to destroy a house you probably aren’t going to get very far.

All of these tools were designed to make life easier and provide a positive to life when doing something but it is how you use the tool that matters.

The same with a gun. I am not a gun ownership kind of guy because of all the shit human beings that just can’t use one properly or claim to use it properly. Guns get more complicated and so do their use cases but the truth is a gun was designed to kill or defend from being killed (this is not a topic about gun rights just using it as an example.) However, in the hands of the wrong person a gun can kill unintentionally. That isn’t the guns fault after all its design was to kill.

ChatGPT wasn’t designed to kill, inherently. It wasn’t designed to do anything other than take databases of information and provide what it thinks is correct. If you as a person don’t know how to use it or what to do with it probably and you aren’t seeking actual medical attention or advice from a professional then I think that is the person’s fault.

ChatGPT can’t make a disclaimer for every little thing. A car on the other hand having a recall issue can. If you want to compare to a faulty part in a car then sure. Modify ChatGPT to just not provide medical advice.

See tools can be changed midway through. The tool isn’t the problem how the person uses the tool is the issue. Access to that tool and what that tool has access to can be an issue but the great thing about tools is laws can change and tools can change.

It isn’t the A.I.s fault if your legislature doesn’t care to enforce that change or law. The same legislature that half of Lemmy is opposed to literally all the time. Tools are only good in ways they can be used as well.

So let’s say for arguments sake the tool is dangerous and in your defense it absolutely can be used dangerously. Do you call upon the government to shut it down just like you would call upon the government to regulate or change gun laws?

Do you also ignore the positive impacts ChatGPT can have because it is doing something else terribly? Imagine a system that medical professionals do create and they modify a version that does provide good medical advice, accurate, and professional? What then? Is ChatGPT still bad? It’s not out of the realm of possibility. A.I isn’t the enemy because someone’s leadership decided to fire you. Leadership is the enemy. Tools are only as bad as the people using them.

Or for the sake of a recalled care that can kill they are as bad as the user manufacturing them. I don’t deny you can get a bad car, a bad screwdriver. My point is if you let the bad outweigh the good then you are missing the point. The bad should be handled by people who understand it better and can design laws and tools to enforce better usage to make something less bad. So again don’t blame the tool blame the people that aren’t protecting you with said tool.

SolNine@lemmy.ml on 27 Aug 2023 02:46 next collapse

GPT has been utter garbage lately. I feel as though it’s somehow become worse. I use it as a search engine alternative and it has RARELY been correct lately. I will respond to it, telling it that it is incorrect, and it will keep generating even more inaccurate answers… It’s to the point where it almost becomes entirely useless, where sometimes it used to find some of the correct information.

I don’t know what they did in 4.0 or whatever it is, but it’s just plain bad.

Elderos@lemmings.world on 27 Aug 2023 05:51 collapse

Yeah they basically killed it. I knew for a fact that it was way too good to be given like that to us common folks. I mean it still has some use cases, but its ability to “reason” about stuff or at least the illusion of it is basically gone.

Legendsofanus@lemmy.blahaj.zone on 27 Aug 2023 08:34 collapse

Can you give some recent use examples of this? I don’t use CHATGPT anymore but I used to get my code written by it and many of my classmates still use it

Elderos@lemmings.world on 27 Aug 2023 11:18 collapse

I think it is still pretty decent at “coding”, but it used that I could throw almost anything at it about virtually anything and most of the time there was no hallucinations if the prompt was good. It has always been prone to beong influenced by biased prompts so you had.to work around that.

Lately I feel like I really need to draw out answers out of it. You have to insist to get past the cookie-cutter reponses and all the newly added “I am not a specialist therefore I refuse to answer”. Historical questions are often refused too if deemed offensives. But really, its been unable to give me all sort of general facts like it used too by lying about the cutoff date of 2021, even after being corrected and apologizing. Some answers are purely 100% hallucinations. Its never been perfect but it used to be better than googling, well not anymore imo. It is still probably decent for non-historical questions, like coding.

SolNine@lemmy.ml on 27 Aug 2023 19:05 collapse

My experience mirrors this, though I don’t use it for much beyond a specific use case… I’m wondering why they neutered it so much?

obinice@lemmy.world on 27 Aug 2023 02:46 next collapse

Well, it’s a good thing absolutely no clinician is using it to figure out how to treat their patient’s cancer… then?

I imagine it also struggles when asked to go to the kitchen and make a cup of tea. Thankfully, nobody asks this, because it’s outside of the scope of the application.

clutch@lemmy.ml on 27 Aug 2023 11:15 collapse

The fear is that hospital administrators equipped with their MBA degrees will think about using it to replace expensive, experienced physicians and diagnosticians

whoisearth@lemmy.ca on 27 Aug 2023 11:35 next collapse

They’ve been trying this shit for decades already with established AI like Big Blue. This isn’t a new pattern. Those in charge need to keep driving costs down and profit up.

Race to the bottom.

obinice@lemmy.world on 29 Aug 2023 21:37 collapse

If that were legal, I’d absolutely be worried, you make a good point.

Even Doctor need special additional qualifications to do things like diagnose illnesses via radiographic imagery, etc. Specialised AI is making good progress in aiding these sorts of things, but a generalised and very poor AI like ChatGPT will never be legally certified to do this sort of thing.

Once we have a much more effective generalised AI, things will get more interesting. It’ll have to prove itself thoroughly though, before being certified, so it’ll still be a few years after it appears before we see it used in clinical applications.

quadropiss@lemmy.world on 27 Aug 2023 11:40 next collapse

😱😱😱😱😱😱😱😱😱 /j

LazyBane@lemmy.world on 27 Aug 2023 11:51 next collapse

People really need to get in their heads that AI can “hallucinate” random information and that any implementation on an AI needs a qualified human overseeing it.

grabyourmotherskeys@lemmy.world on 27 Aug 2023 18:40 collapse

Exactly, it’s stringing together information in a series of iterations, each time adding a new inference consistent with what came before. It has no way to know if that inference is correct.

sentient_loom@sh.itjust.works on 27 Aug 2023 12:14 next collapse

Why the fuck would anybody think a chat bot could create a cancer treatment plan?

5BC2E7@lemmy.world on 27 Aug 2023 13:17 next collapse

Because it’s been hyped. They had announced it could pass the medical licensing exam with good scores. The belief that it can replace a doctor has already been put forward

Touching_Grass@lemmy.world on 27 Aug 2023 13:51 collapse

It did pass it didn’t it. But who said it can replace doctors?

5BC2E7@lemmy.world on 28 Aug 2023 23:45 collapse

If my previous response didn’t answer your question. I’m not sure if it helps to tell you that no one said it can replace doctors.

solstice@lemmy.world on 28 Aug 2023 02:56 collapse

On two occasions I have been asked, ‘Pray, Mr. Babbage, if you put into the machine wrong figures, will the right answers come out?’ I am not able rightly to apprehend the kind of confusion of ideas that could provoke such a question.

Charles Babbage

Better tech, same stupid end users lmao

ours@lemmy.film on 28 Aug 2023 14:07 collapse

Good old “garbage in, garbage out”.

MrSlicer@lemmy.world on 27 Aug 2023 12:33 next collapse

So does my dog, this isn’t news.

Uncaged_Jay@lemmy.world on 27 Aug 2023 13:18 next collapse

“Hey, program that is basically just regurgitating information, how do we do this incredibly complex things that even we don’t understand yet?”

“Here ya go.”

“Wow, this is wrong.”

“No shit.”

JackbyDev@programming.dev on 27 Aug 2023 13:43 next collapse

“Be aware that ChatGPT may produce wrong or inaccurate results, what is your question?”

How beat cancer

wrong, inaccurate information

😱

lolcatnip@reddthat.com on 28 Aug 2023 15:28 collapse

It reminds me of some particularly badly written episodes of Star Trek where they use the holodeck to simulate some weird scenario involving exotic physics nobody understands, and it works perfectly.

sturmblast@lemmy.world on 27 Aug 2023 13:45 next collapse

Why is anyone surprised by this? It’s not meant to be your doctor.

Kodemystic@lemmy.kodemystic.dev on 27 Aug 2023 16:13 next collapse

Who tf is asking chatgpt for cancer treatments anyway?

playerwhoplayyes@lemmy.world on 27 Aug 2023 18:04 next collapse

Probably people who want to check AI accuracy or people who don’t want to search or go to the doctor and ask it to ChatGPT, even if I ask a cure, I will use other AI such as the bing AI, but still I go to the doctor, I will never ask an AI or search on the internet cures to cancer, never self-medicated.

solstice@lemmy.world on 28 Aug 2023 02:54 next collapse

It’s hilarious to me that people need to be told word for word that chat gpt is NOT literally the cure for cancer.

lightnsfw@reddthat.com on 28 Aug 2023 14:20 collapse

Real healthcare is expensive

Sanctus@lemmy.world on 27 Aug 2023 19:06 next collapse

I thought it released in 2021. Maybe it was on the cusp. I was basically using it to find what I couldn’t seem to find in the docs. Its definitely replaced my rubber ducky, but I still have to double check it after my Unity experience.

30mag@lemmy.world on 27 Aug 2023 20:02 next collapse

I am shocked.

NigelFrobisher@aussie.zone on 28 Aug 2023 02:23 next collapse

People really need to understand what LLMs are, and also what they are not. None of the messianic hype or even use of the term “AI” helps with this, and most of the ridiculous claims made in the space make me expect Peter Molyneux to be involved somehow.

dx1@lemmy.world on 28 Aug 2023 02:49 collapse

LLMs fit in the “weak AI” category. I’d be inclined to not call them “AI” at all, since there is no intelligence, just the illusion of intelligence (if I could just redefine the term “AI”). It’s possible to build intelligent AI, but probabilistic text construction isn’t even close.

fsmacolyte@lemmy.world on 28 Aug 2023 05:55 collapse

It’s possible to build intelligent AI

What does intelligent AI that we can currently build look like?

dx1@lemmy.world on 28 Aug 2023 16:31 collapse

There’s “can build” and “have built”. The basic idea is about continuously aggregating data and performing pattern analysis and basically cognitive schema assimilation/accommodation in the same way humans do. It’s absolutely doable, at least I think so.

fsmacolyte@lemmy.world on 28 Aug 2023 21:33 collapse

I haven’t heard of cognitive schema assimilation. That sounds interesting. It sounds like it might fall prey to challenges we’ve had with symbolic AI in the past though.

dx1@lemmy.world on 29 Aug 2023 01:10 collapse

It’s a concept from psychology. Instead of just a model of linguistic construction, the model has to actually be a comprehensive, data-forged model of reality as far as human observation goes/we care about. In poorly tuned, low-information scenarios, it would fall mostly into the same traps human do (e.g. falling for propaganda or pseudoscientific theories) but, if finely tuned, should emulate accurate theories and even predictive results with an expansive enough domain.

GBU_28@lemm.ee on 28 Aug 2023 02:34 next collapse

No one is building document traversal LLM in the healthcare space with off the shelf tools

Pyr_Pressure@lemmy.ca on 28 Aug 2023 03:07 next collapse

Chatgpt is a language / chatbot. Not a doctor. Has anyone claimed that it’s a doctor?

Agent641@lemmy.world on 28 Aug 2023 03:16 collapse

Chatgpt fails at basic math, and lies ablut the existence of technical documentation.

I mostly use it for recipe inspuration and discussing books Ive read recently. Just banter, you know? Nothing mission-critical.

IDontHavePantsOn@lemm.ee on 28 Aug 2023 05:39 collapse

Just a couple days ago it continually told me it was possible to re-tile part of my shower that is broken without cutting tiles, but none of the math added up. (18.5H x 21.5w area) “Place a 9” tile vertically. Place another 9“ tile vertically on top on the same side. Place another 9" tile on top vertically to cover the remainder of the area."

I told chatgpt it was wrong, which it admitted, and spit out another wrong answer. I tried specifying a few more times before I started a new chat and dumbed it down to just a simple math algorithm problem. The first part of the chat said it was possible, layed out the steps, and then said it wasn’t possible in the last sentence.

I surely wouldn’t trust chatgpt to advise my healthcare, but after seeing it spit out very wrong answers to a basic math question, I’m just wondering why anyone would try to have it advise anyone’s health are.

Brandon658@lemmy.world on 28 Aug 2023 16:04 collapse

People want to trust it as a source of quick knowledge. It is easier to be told 9 goes into 81 a total of 8 times trusting that the computer is always right, because it had access to everything, than to work out the answer given was wrong and is actually 9.

Think of WebMD. People love to self diagnose despite it commonly being known as a bad practice. But they do so because they can do it with less effort, faster, and cheaper than making an appointment to drive to an office so you can speak with a doctor that runs a few tests and gets back to you in a week saying they aren’t sure and need to do that process all over again.

raptir@lemm.ee on 28 Aug 2023 16:30 collapse

The healthcare issue is that I’m usually checking WebMD to see if what I’m experiencing is an actual issue that I need to go to the doctor for since it’s so expensive to go to a doctor.

UnbeatenDeployGoofy@lemmy.ml on 28 Aug 2023 03:45 next collapse

I suppose most sensible people already know that ChatGPT is not the answer for medical diagnosis.

Prompts were input to the GPT-3.5-turbo-0301 model via the ChatGPT (OpenAI) interface.

If the researcher wanted to investigate whether LLM is helpful, they should develop a model specifically using cancer treatment plans with GPT-4/3.5 before testing it thoroughly, in addition to entering prompts into the model that is available on OpenAI.

ours@lemmy.film on 28 Aug 2023 13:28 next collapse

Or they could feed the current model with a reputable source of medical information.

testo12@lemmynsfw.com on 28 Aug 2023 13:41 collapse

That wouldn’t guarantee correct answers.

It’s arguably more dangerous if ChatGPT gives mostly sane specific medical advice because it makes people put more trust in it than they should.

ours@lemmy.film on 28 Aug 2023 14:09 collapse

True but it would reduce the chances of it making stuff up entirely.

wewbull@feddit.uk on 28 Aug 2023 18:53 collapse

There have been a number of articles about how GPT has been out-diagnosing doctors in various domains. To me, that isn’t that surprising as diagnosis is a pattern matching problem, something a neuralnet will be very good at. Human doctors were seen to be discounting rare conditions just because they were rare and so “it was much more likely to be something else” even if the symptoms backed up the conclusion. A computer can be more objective about such things.

…but none of that needs AI/ML. We’ve had expert systems since the 60s.

It’s also very different from constructing a treatment plan, which is what we’re discussing here.

KIM_JONG_JUICEBOX@lemmy.ml on 28 Aug 2023 13:11 next collapse

Was this article summary written by chatgpt?

dangblingus@lemmy.world on 28 Aug 2023 13:18 next collapse

Considering ChatGPT’s learning was terminated in 2021, I would not ask it for current medical advice.

lolcatnip@reddthat.com on 28 Aug 2023 15:23 collapse

That’s not the main problem.

Quexotic@infosec.pub on 28 Aug 2023 14:31 next collapse

This is just stupid clickbait. Would you use a screwdriver as a hammer? No. Of course not. Anyone with even a little bit of sense understands that GPT is useful for some things and not others. Expecting it to write a cancer treatment plan, it’s just outlandish.

Even GPT says:I’m not a substitute for professional medical advice. Creating a cancer treatment plan requires specialized medical knowledge and the input of qualified healthcare professionals. It’s important to consult with oncologists and medical experts to develop an appropriate and effective treatment strategy for cancer patients. If you have questions about cancer treatment, I recommend reaching out to a medical professional.

Chunk@lemmy.world on 28 Aug 2023 14:46 collapse

Would you use a screwdriver as a hammer?

I use this argument all the time for Chat GPT :)

AngryCommieKender@lemmy.world on 28 Aug 2023 15:14 collapse

I unfortunately know someone that literally owns every tool he could need. He’s so damn impatient that when he needs a tool, whatever is in reach becomes that tool whether it’s the tool he needs or not. I do not allow him near my tools

shotgun_crab@lemmy.world on 28 Aug 2023 19:55 collapse

Wht would you even consider using chatGPT for this?