356
submitted 6 months ago* (last edited 6 months ago) by tsonfeir@lemm.ee to c/technology@lemmy.world

A neuromorphic supercomputer called DeepSouth will be capable of 228 trillion synaptic operations per second, which is on par with the estimated number of operations in the human brain

Edit: updated link, no paywall

you are viewing a single comment's thread
view the rest of the comments
[-] ArbitraryValue@sh.itjust.works 340 points 6 months ago* (last edited 6 months ago)

A better title would be "Supercomputer that could conceivably simulate entire human brain, based on a rough estimate of what it would take to do that if we had any idea how to do that, will switch on in 2024".

[-] gibmiser@lemmy.world 95 points 6 months ago

For real. I'm reading the title all wondering how the fuck they mapped all the neuron connections and... nope, the real innovative part of the story is clickbait

[-] neuropean@kbin.social 46 points 6 months ago

That’s only counting connections. The brain learns by making new connections, through complex location and timing dependent inputs from other neurons. It’s way more complex than the number of connections, and if neuroscientists are still studying the building blocks we don’t have much hope of recreating it.

[-] IHeartBadCode@kbin.social 32 points 6 months ago

This also ignores that the brain is not wholly an electrical system. The are all kinds of chemical receptors within the brain that alter all kinds of neurological function. Kid of the reason why drugs are a thing. On small scales we have a pretty good idea how these work, at least for the receptors that we're aware of. On larger scales it's mostly guessing at this point. The brain has a knack of doing more than the sum of all parts on a pretty regular basis.

[-] 0ops@lemm.ee 1 points 6 months ago* (last edited 6 months ago)

Not to mention the scale and nature of the "dataset" that our brains were trained on. Millions of years of instinct encoded in DNA, plus a few years gathering data from dozens of senses 24/7 (including chemical receptors, like you said) and in turn manipulating our bodies, interacting with the environment, and observing the results. We've been doing all of this since embryo.

We can't just feed a model raw image and text data and expect it's intelligence to be comparable to ours. However you quantify intelligence/consciousness whatever, the text/image model's thought processes will be alien to ours, which makes sense because their "environment" is nothing like ours - just text and image input and output.

load more comments (1 replies)
load more comments (5 replies)
this post was submitted on 15 Dec 2023
356 points (83.8% liked)

Technology

55647 readers
2562 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS