this post was submitted on 14 Jun 2024
45 points (92.5% liked)
Technology
60036 readers
2893 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
They absolutely do not learn and we absolutely do know how they work. It’s pretty simple.
https://jasonheppler.org/2024/05/23/i-made-this/
Yes, but the tokens are more than just a stream of letters, and aren't saved in the form of words. The information itself is organized into conceptual proximity to other concepts (and distinct from the text itself), and weighted in a way consistent with its training.
That's why these models can use analogies and metaphors in a persuasive way, in certain contexts. Mix concepts that the training data has never been shown before, and these LLMs can still output something consistent with those concepts.
Anthropic played around with their own model, emphasizing or deemphasizng particular concepts to observe some unexpected behavior.
And we'd have trouble saying whether a model "knows" something if we don't have a robust definition of when and whether a human brain "knows" something.