• gaja@lemm.ee
    link
    fedilink
    English
    arrow-up
    2
    ·
    18 hours ago

    Google search results aren’t deterministic but I wouldn’t say it “learns” like a person. Algorithms with pattern detection isn’t the same as human learning.

    • nednobbins@lemmy.zip
      link
      fedilink
      English
      arrow-up
      1
      ·
      18 hours ago

      You may be correct but we don’t really know how humans learn.

      There’s a ton of research on it and a lot of theories but no clear answers.
      There’s general agreement that the brain is a bunch of neurons; there are no convincing ideas on how consciousness arises from that mass of neurons.
      The brain also has a bunch of chemicals that affect neural processing; there are no convincing ideas on how that gets you consciousness either.

      We modeled perceptrons after neurons and we’ve been working to make them more like neurons. They don’t have any obvious capabilities that perceptrons don’t have.

      That’s the big problem with any claim that “AI doesn’t do X like a person”; since we don’t know how people do it we can neither verify nor refute that claim.

      There’s more to AI than just being non-deterministic. Anything that’s too deterministic definitely isn’t an intelligence though; natural or artificial. Video compression algorithms are definitely very far removed from AI.

      • hoppolito@mander.xyz
        link
        fedilink
        English
        arrow-up
        1
        ·
        8 hours ago

        One point I would refute here is determinism. AI models are, by default, deterministic. They are made from deterministic parts and “any combination of deterministic components will result in a deterministic system”. Randomness has to be externally injected into e.g. current LLMs to produce ‘non-deterministic’ output.

        There is the notable exception of newer models like ChatGPT4 which seemingly produces non-deterministic outputs (i.e. give it the same sentence and it produces different outputs even with its temperature set to 0) - but my understanding is this is due to floating point number inaccuracies which lead to different token selection and thus a function of our current processor architectures and not inherent in the model itself.

        • nednobbins@lemmy.zip
          link
          fedilink
          English
          arrow-up
          1
          ·
          1 hour ago

          You’re correct that a collection of deterministic elements will produce a deterministic result.

          LLMs produce a probability distribution of next tokens and then randomly select one of them. That’s where the non-determinism enters the system. Even if you set the temperature to 0 you’re going to get some randomness. The GPU can round two different real numbers to the same floating point representation. When that happens, it’s a hardware-level coin toss on which token gets selected.

          You can test this empirically. Set the temperature to 0 and ask it, “give me a random number”. You’ll rarely get the same number twice in a row, no matter how similar you try to make the starting conditions.