this post was submitted on 23 May 2024
227 points (100.0% liked)

TechTakes

1267 readers
166 users here now

Big brain tech dude got yet another clueless take over at HackerNews etc? Here's the place to vent. Orange site, VC foolishness, all welcome.

This is not debate club. Unless it’s amusing debate.

For actually-good tech, you want our NotAwfulTech community

founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] kogasa@programming.dev -3 points 3 months ago (9 children)

It must have some internal models of some things, or else it wouldn't be possible to consistently make coherent and mostly reasonable statements. But the fact that it has a reasonable model of things like grammar and conversation doesn't imply that it has a good model of literally anything else, which is unlike a human for whom a basic set of cognitive skills is presumably transferable. Still, the success of LLMs in their actual language-modeling objective is a promising indication that it's feasible for a ML model to learn complex abstractions.

[–] mawhrin@awful.systems 16 points 3 months ago (2 children)

it doesn't. that's why we're calling it “spicy autocompletion” .

load more comments (6 replies)