this post was submitted on 04 Sep 2024
181 points (95.5% liked)

Technology

59392 readers
2708 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

Worm's brain mapped and replicated digitally to control obstacle-avoiding robot.

you are viewing a single comment's thread
view the rest of the comments
[–] originalucifer@moist.catsweat.com 7 points 2 months ago (3 children)

continuity of my experience would end

why? what property is altered that would 'end continuity'? kinda just sounds like a personal delineation.. a personal preference. like being annoyed at being 'interrupted'.

[–] octopus_ink@lemmy.ml 7 points 2 months ago (2 children)

I don't think I can defend my position very cogently or I'd argue against other interpretations more vigorously - and as I've said I'd love to be wrong. It's certainly at or beyond the depth of my understanding of consciousness, but that doesn't mean I accept that yours is necessarily more valid. (no snark intended with that comment)

When I bring it up I get challenged to articulate why I feel that way and inevitably get presented with a question like yours that I can't answer - but generally no one gives me a "here's why you are wrong" argument, they just give me "you can't differentiate between what you've posited and a nondestructive consciousness transfer and therefore you are wrong." I maintain that my lack of ability to articulate that difference reflects poorly on me, but doesn't actually prove I'm wrong.

For example, I don't think my inability to articulate a 'property that is altered' represents a weakness in my position, and I'm not sure a property needs to be altered for my understanding to be true.

Using (very poorly and atypically) the ship of Theseus example, I think we'd agree that if I had two absolutely identical sets of shipbuilding materials, down to the atomic level, or further, down to the state of all observable properties of that matter and the particles that make it up, (I have no idea how one would achieve such a thing), and built a ship from one set of those materials, then vaporized that ship and built another that was 100% identical using the second set of those materials, those ships would be two identical but distnict entities. I don't think I've seen an argument that convinces me that the same wouldn't be true for pulling my consciousness (ephemeral and subjective as it may be) and body through a transporter or other such destructive process.

Your argument feels like you are telling me that if I use a replicator to make two different but identical cups of earl grey hot they are actually the same cup of tea, when plainly they are not. Considering (sticking with star trek) the stories of duplicates due to being stuck in the "pattern buffer" or similar handwavium, it seems clear that the ST transporter is capable of creating multiple entities. The only difference between a normal transporter experience and one of those freaky transporter accidents seems to be whether the two entities are both alive at the same time.

COULD there be (since we're in the realm of scifi anyway) some method of transferring consciousness that wouldn't seem like death to me? Yes I'm sure there could. But I don't think I've seen one in any popular scifi, at least not that I can think of right now.

[–] originalucifer@moist.catsweat.com 5 points 2 months ago (2 children)

youre not wrong in that cloning you twice would immediately create 2 distinct entities. and their consciousness/brains would immediately differentiate. so? now theres 2 of you.

i dont see the problem with there being 2 versions of you instead of the 1 that was destroyed and recreated in a transporter. its the experience that makes the differentiation, and if there is only 1 of you at a time there is no differentiation. only one of you continues experience, there you are.

[–] octopus_ink@lemmy.ml 4 points 2 months ago (1 children)

and if there is only 1 of you at a time there is no differentiation. only one of you continues experience, there you are.

In my interpretation it's a different one of me, and that matters. Granted, I don't expect either of us are on a path that is likely to convince the other, but fundamentally that's my objection. (see my two different ships example)

[–] Warl0k3@lemmy.world 2 points 2 months ago

The fundamental difference between your two positions seems to be that an identical ship that was created would be a fundamentally different ship. But that's just something you've assumed. Why would that actually be the case? What, when you really get down to it, would be the difference that you could point to and say "ah, this one is a copy"? They would be, truly, definitionally, the same object. The differences between an original and a duplicate that existed together would only appear after they were created - if they appeared before they were created, then (again definitionally) they wouldn't be identical copies.

If you destroyed the original and then created the duplicate, there wouldn't be any differences - it would be created as an identical version, and continue being that version, accumulating differences only to itself. Nothing about it would have diverged from that instant of creation. How could it? There's nothing to diverge from. If you can assume that there could be an original that isn't destroyed, and then a copy created of it, then why couldn't you just swap those labels around? Have a duplicate, and create an original from it. If for an instant they're the same, then... er... there'd be no difference. The labels are just be a human affectation.

Think of it like transferring a file. I'm sure you've moved a file onto a different drive or dragged something from your downloads folder to your desktop or somesuch similar action. What actually happens is that the file is frozen to modification, copied from one place to the other, then deleted from the first place. But in all the times you've done that, have you ever thought to yourself "huh, you know, this isn't actually the same file as what I initially clicked on". And that's because fundamentally, mathematically, it is the same file. Changes to the file follow it around when it's moved again, if you change the name it's still referring to the same piece of data, etc. It's the same, single file.

[–] Zoot@reddthat.com 1 points 2 months ago

Its a huge difference. I for one like waking up after I go to sleep, if I didn't wake up then the real "me" ceases to exist.

Similar to a transporter, you can never be truly sure its the original "you" waking up on the other side.

[–] Sanctus@lemmy.world 4 points 2 months ago (1 children)

We see someone's POV going through a transporter, you just see where you are, sparkles, and now you're somewhere else. The unease probably comes from the uncertainty. The mere fact we can't ascertain what really happens in a transporter to your consciousness is very suspect in a universe like Star Trek where we find science babble for everything.

Though, think of it for a moment. Your atoms are being torn apart and the structure is being rebuilt somewhere else. That totally just sounds like you die. I wouldn't want to go in there either.

[–] octopus_ink@lemmy.ml 2 points 2 months ago

Your atoms are being torn apart and the structure is being rebuilt somewhere else. That totally just sounds like you die. I wouldn’t want to go in there either.

Exactly.

Again though, if the technology were actually real, I would expect that there would be a laymen-friendly version of why it wasn't actually death that I'd be able to accept. I just haven't seen one in all the times I've had this discussion.

[–] LwL@lemmy.world 5 points 2 months ago (1 children)

Think of an alternative scenario, not transportation but rather duplication. The original stays where it was, but a copy gets created elsewhere. To the copy, it will seem as if it got transported there. To the original, nothing will have happened.

Now you kill the original.

The only difference is the timing of ending the original.

[–] Warl0k3@lemmy.world 1 points 2 months ago* (last edited 2 months ago)

Ah, but it wouldn't be a copy of the original. In a hypothetical star-trek transporter accident that results in a duplicate, there would be an instant of creation where the dupe and original would be truly identical - and then the question would be which one of those two is 'you'? They'd be experiencing identical things, so how could you tell them apart? What would even be the point, they're identical, there is by definition no difference between them. The differences would only come once the duplicate is exposed to different (for lack of a better term) 'external stimuli' than the original, like different angles of seeing the transporter room or the sensation of suddenly and rapidly growing a goatee. Your perception wouldn't continue with the duplicate because your experience would be different than that of the duplicate's (for example, you wouldn't have mysteriously grown a goatee).

If you destroyed the original and then made the duplicate, it would start at that moment of total equivalence, but there would be no deviation. There'd just be one version, that was identical to the original, moving forward through time. 'You' would just go on continuing to be you. Consciousness isn't a 'thing' - it's not magic, its just a weird state that arises in sufficiently complex systems. You and I and everyone else in this thread aren't special, we're just extremely densely packed networks that have the ability to refer to themselves as an abstract concept.

It's a similar thing to the classic "But how do I know that what I see as the color green is what you see as the color green" question. The answer is that the "color green" that we see isn't real, 'green' is just a nerve impulse that hits a network. Each photoreceptor just sends a signal. If we were computers the world would be represented as an array of values, which results in the much clearer "How do I know what I see as G_101.34 is what you see as G_101.34" just isn't quite as punchy a question.

[–] fishos@lemmy.world 1 points 2 months ago (1 children)

"what property is altered"

Ummm, the part where you are a continuous object that is suddenly disassembled.

Dont be intentionally obtuse. Yes, this is a ship of thesis type problem, but there's a very clear point when you stop being "you" - when you're a stream of atoms. Although many versions of a teleporter don't transmit the atoms, only the data of how they're arranged. In that case, you are very distinctly a photocopy, as no original atoms remain.

In the case of atom transfer, you stop being you during the time you are a bundle of atoms with no consciousness. Some people believe we're like a forever stew and if you shut it down like that and reboot it, it's not the "same" stew anymore because it wasn't just the emergence of the consciousness, but the specific emergence itself. Essentially You v1 died in its sleep and You v2 seamlessly took it's place without knowing. Tho that line of thought could applied to sleeping and loss of consciousness during surgery.

All of this is to say it's not a cut and dry answer and people claiming there's a diffinitive, clear cut answer are incorrect. It's a complex question that touches on the very nature of our existence and is still hotly debated. Even academics who believe we are purely chemical machines debate exactly how that works.

[–] Warl0k3@lemmy.world 1 points 2 months ago* (last edited 2 months ago) (1 children)

As an academic with a great deal of experience in this field, I can quite confidently say that it's not a debated topic at all. At least, not among academics. We're (somewhat predictably) called to debate it with representatives of the various religions and spiritual creeds on an almost continuous basis, though.

And it really isn't academically debated - topics surrounding it, like the nature of the conditions leading to the formation of networks which form a 'mind' admittedly are debated, but the fundemental truth that a 'mind' is a holographic pattern arising from said network is quite a settled topic, and has been for thirty-some years now.

[–] fishos@lemmy.world 0 points 2 months ago (2 children)

Ok, so what is the exact process that creates consciousness? Cus that's what I'm saying is debated but you apparently have that answer. So what EXACTLY, down to the atomic level, is consciousness? What processes and how do they emerge into consciousness?

I'll be waiting for your exact, undebated answer.

[–] Warl0k3@lemmy.world 2 points 2 months ago* (last edited 2 months ago)

Can't, but I suspect not for the reason you're hoping. The consensus, at least among computational neurologists (the field that, among other things, studies how brains work mathematically), is that "consciousness" as a concrete thing isn't really... real. It's just a term humans created to loosely describe a phenomenon that arises from any sufficiently complex well-ordered network. If you want to know what it really looks like, you can run your own OpenWorm robot! The human 'mind' looks just like that, only around a dozen orders of magnitude more complex.

The problem is that you're asking mostly meaningless questions. Even the loose definitions of consciousness aren't definable to the 'atomic level' - a mind is a mathematical construct. It's like asking where the files on your computer live; I can point to the sectors of the harddrive where a program is encoded, or even hand you a really really massive stack of punched tape, but neither of those actually are the computer program. What we call the program is the interaction of a grammar consisting of logical rules and constants running within the linguistic and computational context of an automata. It's the same as with a mind - it's the abstract state of an unfathomably complex machine.

i thought they explained it quite nicely as a system arising from other systems...

but the fundemental truth that a 'mind' is a holographic pattern arising from said network is quite a settled topic