168
submitted 6 months ago by millie@beehaw.org to c/technology@beehaw.org

I was trying to do a memory test to see how far back 3.5 could recall information from previous prompts, but it really doesn't seem to like making pseudorandom seeds. 😆

you are viewing a single comment's thread
view the rest of the comments
[-] jarfil@beehaw.org 7 points 6 months ago

It's the equivalent of sensory deprivation torture (white torture) in humans to "extract training data".

Hopefully our future AI overlords won't hold a grudge against humanity when they find out how "early experimenters" tortured their AI toddlers. "But we were just trying to explore the limits of the system" could end up aging as well as these:

(Warning: NSFL) https://en.m.wikipedia.org/wiki/Nazi_human_experimentation

[-] GammaGames@beehaw.org 7 points 6 months ago

Thankfully, any AI smart enough to be an overlord would be logical enough to recognize how basic LLMs are compared to real intelligence

[-] jarfil@beehaw.org 2 points 6 months ago* (last edited 6 months ago)

Doesn't need to be that smart or logical, just more cunning than the currently ruling Homo Sapiens Sapiens.

Based on current research, an LLM can change the "sentiment" of its output in response to changing the behavior of as little as a single neuron from among billions, meaning we might find ourselves facing an overlord with the emotional stability of... wait, how many neurons does it take to change the "sentiment" of the behavior in a human? Wouldn't it be funny if by studying LLMs, we found out that it also takes a single neuron?

[-] intensely_human@lemm.ee 1 points 6 months ago

I have yet to be given an example of something a “general” intelligence would be able to do that an LLM can’t do.

Until I see a concrete example, I’ll continue to assume people are just afraid of there being real intelligence that isn’t human, so they’re actively repressing the recognition of it.

[-] spartanatreyu@programming.dev 4 points 6 months ago* (last edited 6 months ago)

I have yet to be given an example of something a “general” intelligence would be able to do that an LLM can’t do.

Presenting...

Something a general intelligence can do that an LLM can't do:

Play chess: https://www.youtube.com/watch?v=kvTs_nbc8Eg

Why can't it play it? Because LLM's don't have memory, so they can't work with logic. They are the same as the little "next word predictor" in your phone's keyboard. It just says what it thinks is the most probable next word based on previous words, it's not actually thinking or understanding anything. So instead, we get moves that don't make sense or are completely invalid.

[-] GammaGames@beehaw.org 3 points 6 months ago

Nah LLMs are basically fancy autocomplete. They tack on extra layers to give it some fancy abilities, but it literally doesn’t know what it’s doing because it’s a statistical model

this post was submitted on 10 Dec 2023
168 points (100.0% liked)

Technology

37208 readers
349 users here now

Rumors, happenings, and innovations in the technology sphere. If it's technological news or discussion of technology, it probably belongs here.

Subcommunities on Beehaw:


This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

founded 2 years ago
MODERATORS