• _stranger_@lemmy.world
      link
      fedilink
      arrow-up
      1
      ·
      23 days ago

      Because you’re not getting an answer to a question, you’re getting characters selected to appear like they statistically belong together given the context.

      • howrar@lemmy.ca
        link
        fedilink
        arrow-up
        0
        ·
        23 days ago

        A sentence saying she had her ovaries removed and that she is fertile don’t statistically belong together, so you’re not even getting that.

        • JcbAzPx@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          23 days ago

          You think that because you understand the meaning of words. LLM AI doesn’t. It uses math and math doesn’t care that it’s contradictory, it cares that the words individually usually came next in it’s training data.

            • Swedneck@discuss.tchncs.de
              link
              fedilink
              arrow-up
              1
              ·
              3 days ago

              and those tokens? just numbers, indexes. LLMs have no concept of language or words or anything, it’s literally just a statistical calculator where the numbers encode some combination of letter(s)

          • howrar@lemmy.ca
            link
            fedilink
            arrow-up
            0
            ·
            23 days ago

            It has nothing to do with the meaning. If your training set consists of a bunch of strings consisting of A’s and B’s together and another subset consisting of C’s and D’s together (i.e. [AB]+ and [CD]+ in regex) and the LLM outputs “ABBABBBDA”, then that’s statistically unlikely because D’s don’t appear with A’s and B’s. I have no idea what the meaning of these sequences are, nor do I need to know to see that it’s statistically unlikely.

            In the context of language and LLMs, “statistically likely” roughly means that some human somewhere out there is more likely to have written this than the alternatives because that’s where the training data comes from. The LLM doesn’t need to understand the meaning. It just needs to be able to compute probabilities, and the probability of this excerpt should be low because the probability that a human would’ve written this is low.

            • JcbAzPx@lemmy.world
              link
              fedilink
              English
              arrow-up
              1
              ·
              23 days ago

              Unless they grabbed discussion forums that happened to have examples of multiple people. It’s pretty common when talking about fertility, problems in that area will be brought up.

              People can use context and meaning to avoid that mistake, LLMs have to be forced not to through much slower QC by real people (something Google hates to do).