892
submitted 2 days ago* (last edited 2 days ago) by seahorse@midwest.social to c/technology@midwest.social
you are viewing a single comment's thread
view the rest of the comments
[-] GissaMittJobb@lemmy.ml 58 points 2 days ago

LLMs do not work that way. They are a bit less smart about it.

This is also why the first few generations of LLMs could never solve trivial math problems properly - it's because they don't actually do the math, so to speak.

[-] tyler@programming.dev 4 points 2 days ago

Overtraining has actually shown to result in emergent math behavior (in multiple independent studies), so that is no longer true. The studies were done where the input math samples are “poisoned” with incorrect answers to example math questions. Initially the LLM responds with incorrect answers, then when overtrained it finally “figures out” the underlying math and is able to solve the problems, even for the poisoned questions.

[-] Anticorp@lemmy.world 5 points 2 days ago

That's pretty interesting, and alarming.

Do you have these studies? I can't find much.

[-] tyler@programming.dev 2 points 1 day ago

I searched for like 20 minutes but was unable to find the article I was referencing. Not sure why. I read it less than a month ago and it referenced several studies done on the topic. I'll keep searching as I have time.

It's okay, man. If it really is improving, I'm sure it'll come up again at some point.

this post was submitted on 28 Jun 2024
892 points (98.9% liked)

Technology

1878 readers
93 users here now

Post articles or questions about technology

founded 2 years ago
MODERATORS