this post was submitted on 03 Jan 2025
76 points (100.0% liked)
Technology
37830 readers
390 users here now
A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.
Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.
Subcommunities on Beehaw:
This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Because everything we know about how the brain works says that it's not a statistical word predictor.
LLMs have no encoding of meaning or veracity.
There are some great philosophical exercises about this like the chinese room experiment.
There's also the fact that, empirically, human brains are bad at statistical inference but do not need to consume the entire internet and all written communication ever to have a conversation. Nor do they need to process a billion images of a bird to identify a bird.
Now of course because this exact argument has been had a billion times over the last few years your obvious comeback is "maybe it's a different kind of intelligence." Well fuck, maybe birds shit icecream. If you want to worship a chatbot made by a psycopath be my guest.
LLMs aren't just simple statistical predictors either. More generally, the universal approximation theorem is a thing - a neural network can be used to represent just about any function, so unless you think a human brain can't be represented by some function, it's possible to embed one in a neural network.
I'm not sure what you mean by this. The interpretability research I've seen suggests that modern LLMs do have a decent idea of whether their output is true, and in many cases lie knowingly because they have been accidentally taught, during RLHF, that making up an answer when you don't know one is a great way of getting more points. But it sounds like you're talking about something even more fundamental? Suffices to say, I think being good at text prediction does require figuring out which claims are truthful and which aren't.
The Chinese Room argument has been controversial since about the time it was first introduced. The general form of the most common argument against it is "just because any specific chip in your calculator is incapable of math doesn't mean your calculator as a system is", and that taken literally this experiment proves minds can't exist at all (indeed, Searle who invented this argument thought that human minds somehow stem directly from "physical–chemical properties of actual human brains", which sure is a wild idea). But also, the framing is rather misleading - quoting Scott Aaronson's "Quantum Computing Since Democritus":
I'm not sure what this proves - human brains can learn much faster because they already got most of their learning in the form of evolution optimizing their genetically-encoded brain structure over millions of years and billions of brains. A newborn human already has part of their brain structured in the right way to process vision, and hence needs only a bit of training to start doing it well. Artificial neural networks start out as randomly initialized and with a pretty generic structure, and hence need orders of magnitude more training.
Nah - personally, I don't actually care much about "self-awareness", because I don't think an intelligence needs to be "self-aware" (or "conscious", or a bunch of other words with underdefined meanings) to be dangerous; it just needs to have high enough capabilities. The reason why I noticed your comment is because it stood out to me as... epistemically unwise. You live in a world with inscrutable blackboxes who nobody really understands which can express wide ranges of human behavior including stuff like "writing poetry about the experience of self-awareness", and you're "absolutely sure" they're not self-aware? I don't think many of the history's philosophers of consciousness, say, would endorse a belief like that given such evidence.
It's possible to point out the exact silicon in the calculator that does the calculations, and also exactly how it does it. The fact that you don't understand it doesn't mean that nobody does. The way a calculator calculates is something that is very well understood by the people who designed it.
By the way, this brings us to the history of AI which is a history of 1) misunderstanding thought and 2) charlatans passing off impressive demos as something they're not. When George Boole invented boolean mathematics he thought he was building a mathematical model of human thought because he assumed that thought==logic and if he could represent logic such that he could do math on it, he could encode and manipulate thought mathematically.
The biggest clue that human brains are not logic machines is probably that we're bad at logic, but setting that aside when boolean computers were invented people tried to describe them as "electronic brains" and there was an assumption that they'd be thinking for us in no time. Turns out, those "thinking machines" were, in fact, highly mechanical and nobody would look at a univac today and suggest that it was ever capable of thought.
Arithmetic was something that we did with our brains and when we had machines that could do it that led us to think that we had created mechanical brains. It wasn't true then and it isn't true now.
Is it possible that someday we'll make machines that think? Perhaps. But I think we first need to really understand how the human brain works and what thought actually is.
There's this message pushed by the charlatans that we might create an emergent brain by feeding data into the right statistical training algorithm. They give mathematical structures misleading names like "neural networks" and let media hype and people's propensity to anthropomorphize take over from there.
I would say I do. It's not that high of a bar - one only needs some nandgame to understand how logic gates can be combined to do arithmetic. Understanding how doped silicon can be used to make a logic gate is harder but I've done a course on semiconductor physics and have an idea of how a field effect transistor works.
That's exactly my point, though. If you zoom in deeper, a calculator's microprocessor is itself composed of simpler and less capable components. There isn't specific a magical property of logic gates, nor of silicon (or doping) atoms, nor for that matter of elementary particles, that lets them do math - it's by building a certain device out of them that composes their elementary interactions that we can make a tool for this. Whereas Searle seems to just reject this idea entirely, and believes that humans being conscious implies you can zoom in to some purely physical or chemical property and claim that it produces the consciousness. Needless to say, I don't think that's true.
That was a common and reasonable position in, say, 2010, but the problem is: I think almost nobody in 2010 would have claimed that the space of things that you can make a program do without any extra understanding of thought included things like "write code" and "draw art" and "produce poetry". Now that it has happened, it may be tempting to goalpost-move and declare them as "not true thought", but the fact that nobody predicted it in advance ought to bring to mind the idea that maybe that entire line of thought was flawed, actually. I think that trying to cling to this idea would require to gradually discard all human activities as "not thought".
And that's us coming back around to the original line of argument - I don't at all agree that it's "easy to test" that even, say, modern LLMs "aren't thinking". Because the difference between the calculator example and an LLM is that in a calculator, we understand pretty much everything that happens and how arithmetic can be built out of the simpler parts, and so anyone suggesting that calculators need to be self-aware to do math would be wrong. But in a neural network, we have full understanding of the lowest layers of abstraction - how a single layer works, how activations are applied, how it can be trained to minimize a certain loss function via propagation - and no idea at all about how it works on a higher level. It's not even "only experts do", it's that nobody in the world understands how LLMs work under the hood, why they have the many and specific weird behaviors they do. That's concerning in many ways, but in particular I absolutely wouldn't assume with little evidence that there's no "self-awareness" going on. How would you know? It's an enormous blackbox.
There's certainly a lot of woo and scamming involved in modern AI (especially if one makes the mistake of reading Twitter), but I wouldn't say the term "neural network" is at all confusing? I agree on the anthropomorphization though, it gets very weird. That said, I can't help but notice that the way you phrased this message, it happens to be literally true. We know this because it already happened once. Evolution is just a particularly weird and long-running training algorithm and it eventually turned soup into humans, so clearly it's possible.
Every time there's an AI hype cycle the charlatans start accusing the naysayers of moving goalposts. Heck that exact same thing was happing constantly during the Watson hype. Remember that? Or before that the Alpha Go hype. Remember that?
I was editing my comment down to the core argument when you responded. But fundamentally you can't make a machine think without understanding thought. While I believe it is easy to test that Watson or ChatGPT are not thinking, because you can prove it through counterexample, the reality is that charlatans can always "but actually" those counterexamples aside by saying "it's a different kind of thought."
What we do know because this at least the 6th time this has happened is that the wow factor of the demo will wear off, most promised use cases won't materialize, everyone will realize it's still just an expensive stochastic parrot and, well, see you again for the next hype cycle a decade from now.
Not really. As far as I can see the goalpost moving is just objectively happening.
If "think" means anything coherent at all, then this is a factual claim. So what do you mean by it, then? Specifically: what event would have to happen for you to decide "oh shit, I was wrong, they sure did make a machine that could think"?