this post was submitted on 19 Nov 2024
9 points (80.0% liked)

Futurology

1852 readers
97 users here now

founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] Darrell_Winfield@lemmy.world 3 points 1 month ago* (last edited 1 month ago) (1 children)

Actual article here is worth a read because I find when newspapers write about these articles, they miss the point hard.

This actually has nothing to do with "memory", but instead reading text. They studied 52 doctors responses to standardized (read publicly available online) cases written in front of them. Half got access to LLM. Neither group was significantly different.

Then they ran 3 trials solely with LLM and find that these were significantly better.

My thoughts: 1: Terribly small sample size overall, but would like to see more LLM numbers. 2: The primary purpose of this study was to explore if doctors are better with LLMs helping them. We're not, and the authors discuss a very good point of "prompts matter". 3: As is always my gripe with these kinds of things, written text translates to real patients extremely poorly. A computerized text interface is better at handling and responding to text patients. Human doctors are still better at treating human patients.

[–] drspod@lemmy.ml 4 points 1 month ago (1 children)

They studied 52 doctors responses to standardized (read publicly available online) cases written in front of them.
...
Then they ran 3 trials solely with LLM and find that these were significantly better.

How do they know that the answers to the "standardized" publically available case studies were not in the training data of the LLM? Isn't it extremely likely that they were?

[–] Darrell_Winfield@lemmy.world 4 points 1 month ago

It's very likely that they were in the training data. I forgot to include that as a point. Unfortunately, though, that's a very difficult variable to control in the LLM research.