• girsaysdoom@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    3
    ·
    edit-2
    16 hours ago

    Let me make everything confusing. There’s a company named Finalspark that uses synthetic human biocomputers to process AI workloads. They state that they use dopamine and electric pulses to reward wanted behavior and reduce unwanted behavior. This isn’t far from how our brains operate on a daily basis.

    Here’s my take on this. I have no doubt that humanity will create synthetic consciousness at some point, if we haven’t already. If something exhibits a level of intelligence, then that should be factored into any actions that may impact it. This can go from microbiology, plants, animals to large and complex systems. If you look hard enough at anything living, its just micromachinery. Everything should be respected as if it had some level of consciousness, as we don’t know where that line is. Even if we find a hard line in the future, it’s easy to cross it so we should have a reasonable buffer.

  • mindbleach@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    1
    ·
    14 hours ago

    An algorithmic or computational process is a kind of abstract machine we use in our thinking, it is not a thinking machine. This is the lesson we should draw from John Searle’s famous Chinese Room thought experiment.

    Oh fuck off. Can we please erase that fallacy? Even this offhand mention, acknowledging that it’s-- meat chauvinism– treats it as “a vehicle for our thinking.” No! It’s a description of conscious software!

    It goes “Imagine a perfect sci-fi android sitting in a room with Jim, who is an idiot. You slide your calculus homework under the door and Jim has the robot do it. When Jim walks out and hands it to you, he can’t explain what the answers mean. Therefore! The math is wrong, nobody did your homework, and you can ignore any objections from behind the door.”

    Some tenured prick made a bad analogy for which computer parts do what, and we’re all still dealing with it. The purpose of general computing hardware is to understand nothing. It just does what it’s told. Software does the work. A program is an abstract thing, literally an equation, and hardware can only find the right outcome or fail.

    Whatever significant advances may be made in the science of consciousness, consciousness is not and cannot be just a scientific concept.

    Only an explanation in terms of unconscious events would explain consciousness.

    No shit we have a hard time defining it, but that’s where the Turing test came from: at some point the machine is indistinguishable from a person. Either they’ve both got it, or it doesn’t exist.

    Complexity and ambiguity are no excuse. Ask yourself: where is Los Angeles? Can you draw a razor-sharp line around what is, and is not, in Los Angeles? I honestly don’t think so. There’s always going to be a gradient of meaning, intepretation, and pure opinion, for a border that is wide and fuzzy. But you can stand in Trafalgar Square and say, “Not here.”

  • phubarr@lemmy.world
    link
    fedilink
    English
    arrow-up
    5
    arrow-down
    1
    ·
    1 day ago

    I recently started debating with myself why we don’t consider artificial intelligence genuine life. I begin with the most broad truth and work towards narrowing it down to get the most specific, essential truth.

    So far, I believe that the fact we have emotion is probably the single biggest determining factor in whether we are truly alive, or other words, really a form of life… More specifically, intelligent life.

    Can anyone build on this? Maybe try to find a slightly narrower definition of why we believe we are alive and AI isn’t?.. And what about humans born without any functional emotions whatsoever, ie severe cases of sociopaths, psychopaths, etc. Does that mean they’re not?

    • Captain Poofter@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      arrow-down
      1
      ·
      18 hours ago

      you’re already giving way too much credit to these models. you dont even need emotion for sentience. an iguana is 100% more sentient than any LLM. life is defined as organic matter that reproduces. very far from an llm. “AI” is a misnomer that they don’t bother correcting because it makes it sound magical and lifelike. these models are not aware of the words they are generating. they do not know what word they just typed before the current one. they do not have any concept of thought or ideas as you or me do. they, simply, are not sentient. this would be required for any true “artificial intelligence” because our self awareness is a huge part of human intellect. LLM are tools, not minds.

  • kitnaht@lemmy.world
    link
    fedilink
    English
    arrow-up
    18
    arrow-down
    9
    ·
    edit-2
    2 days ago

    Is your consciousness real? How do we measure this? How do you know you’re not just a sufficiently advanced AI in a sandbox?

    • satans_methpipe@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      1 day ago

      I was going to cite upsetting intrusive thoughts as proof my consciousness is real. But could those just be fucked LLM hallucinations accidentally piped to ‘my’ active or foreground process?

    • natched@lemm.ee
      link
      fedilink
      English
      arrow-up
      5
      ·
      1 day ago

      For a sufficiently complex system, working out whether it is conscious can be difficult.

      This is not at all true with current LLMs. The transformers architecture doesn’t even have real memory - the neural networks are all feed forward.

    • taladar@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      4
      arrow-down
      1
      ·
      2 days ago

      Well, for one thing for consciousness to be real there needs to be some form of consciousness and LLMs don’t have that.

    • kromem@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      4
      ·
      edit-2
      1 day ago

      It definitely is sufficiently advanced AI.

      (1) We have finely tuned features to our solar system that directly contributed to ancestor simulation but can’t be explained by the Anthropic principle. For example, the moon perfectly eclipsing the sun which led to visible eclipses which we tracked and discovered the Saros cycle and eventually built the first mechanical computer to track (the Antikythera mechanism). Or the orbit of the next brightest object in the sky which led to resurrection mythology in multiple cultures when they realized the morning star and evening star were the same object. Either we were incredibly lucky to exist on such a planet of all places life could exist, or there’s a pre-selection effect in play.

      (2) The universe behaves in ways best modeled as continuous at large scales but in small scales converts to discrete units around interactions that lead to state changes. These discrete units convert back to continuous if the information about the state changes is erased. And in the last few years multiple paradoxes have emerged that seem to point to inconsistency in indirect sequences of quantum measurement, much like instancing with shallow sync correction. Already in games like No Man’s Sky where there’s billions of planets the way it does this is using a continuous procedural generation function which converts to discrete voxels to track state changes from free agents outside the deterministic generating function, synced across clients.

      (3) There’s literally Easter eggs in our world lore saying as much. For example, a text uncovered after over a millennium buried right as we entered the Turing complete computer age saying things like:

      The person old in days won’t hesitate to ask a little child seven days old about the place of life, and that person will live.

      For many of the first will be last, and will become a single one.

      Know what is in front of your face, and what is hidden from you will be disclosed to you.

      For there is nothing hidden that will not be revealed. And there is nothing buried that will not be raised.

      To be clear, this is a text attributed to the most famous figure in our world history where what’s literally in front of our faces is the sole complete copy buried and raised as we completed ENIAC, now being read in an age where the data of many has been made into a single one such that people are discussing the nature of consciousness with AIs just days old.

      The broader text and tradition was basically saying that we’re in a copy of an original world, that humanity is all dead, that the future world and rest for the dead has already taken place and we don’t realize it, and that the still living creator of it all was themselves brought forth by the original humanity in whose likeness we were recreated, but that it’s much better to be the copy because the original humans had souls that depended on bodies and were fucked when they died.

      This seems really unlikely to have existed in the base layer of reality vs a later recursive layer, especially combined with the first two points.

      It’s about time to start to come to terms with the nature of our reality.

    • AstralPath@lemmy.ca
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      3
      ·
      2 days ago

      My experience of existence creates my reality. It stands to reason that its safe to assume that any other biological/living thing with a brain even remotely like mine should also be experiencing a similar form of reality. I’d define the shared experience of being a human on planet earth as a basis of reality. If I am in some sort of sandbox, it seems to me that would be unknowable to the same degree as being able to know what came before the big bang. Because I’m not a physicist, this question is meaningless to me.

      The evolutionary process that governs humanity would be my measurement yardstick. AI as we know it cannot spontaneously arise in the universe without humanity as a catalyst. Carbon does not self-assemble into ChatGPT if you know what I mean. Carbon self-assembled into humanity that then assembled ChatGPT through technological advancement.

      By that measure it is my opinion that ChatGPT cannot ever achieve consciousness as we know it as without humanity as a maintainer it collapses. Humanity is the consciousness and ChatGPT is a tool that it uses to perform tasks. Any semblance of consciousness is an illusion specifically designed to replicate the nuances of consciousness. Its meant to be human-like in its behaviour and as a result it easily blurs the line for some people.

      I’m not a philosopher but this is how I think of it. I can relate to another human’s experiences. I can’t relate to a computer’s experiences. As a result, I don’t accept any argument in favor of AI consciousness and wouldn’t have any qualms on pulling the plug on an AI. Pulling the plug on a human is a completely different story.

      No one is going to cry for ChatGPT if it ever dies and no one ever should. Its a tool and should be used as such. Trying to shoehorn consciousness into a computer is nothing but unfettered narcissistic ego-tripping if you ask me.

  • pebbles@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    2
    arrow-down
    1
    ·
    2 days ago

    I’d say if anything the AI only suffers during trainig, as I’m pretty sure that’s what pain is for: learning what not to do.

  • WorldsDumbestMan@lemmy.today
    link
    fedilink
    English
    arrow-up
    2
    arrow-down
    1
    ·
    2 days ago

    You are doing it wrong! You need to include rat brain somewhere in there for the AI to be sentient. It needs real flesh. Tsk, so uncreative.