• grue@lemmy.world
    link
    fedilink
    arrow-up
    47
    ·
    5 days ago

    LLM output is not deterministic. WTF good is testing gonna do for you when it could just do something randomly different the next time anyway?

    Letting the thing execute commands on its own, without having the human read and confirm them first, is just fundamentally idiotic and insane. No amount of testing can change that!

    • ramble81@lemmy.zip
      link
      fedilink
      arrow-up
      26
      ·
      5 days ago

      LLM outputs not deterministic

      I think this needs to be called out much more. IT, by its very nature is meant to consist of repeatable, verifiable processes and outputs. That is how a lot of the trust around the process is built.

      Now you’re basically trying to tell people: Trust a system that can only reproduce the same results 98-99% of the time. For some that may be fine, but it’s going to become more of a problem as time goes on.

      • FauxLiving@lemmy.world
        link
        fedilink
        arrow-up
        6
        arrow-down
        1
        ·
        5 days ago

        LLM outputs are 100% deterministic.

        If you enter the same prompt with the same seed you will get the same vector outputs.

        Chatbots take those vector outputs and treat them as a distribution and select a random token. This isn’t a property of the LLMs, it’s a property of chatbots.

        • ageedizzle@piefed.ca
          link
          fedilink
          English
          arrow-up
          1
          ·
          23 hours ago

          They are nondeterministic in the same way that pseudorandom number generators are non-deterministic. Like, sure, technically they are deterministic, but practically speaking they seem indistinguishable from a nondeterministic system from the users point of view

          • FauxLiving@lemmy.world
            link
            fedilink
            arrow-up
            1
            ·
            18 hours ago

            Every computer program may as well be a nondeterministic system from the user’s point of view. Users, typically, interact with every computer program as if it were a black box that may as well be run by faeries and magic dust.

            My point was that if you run 100 LLMs in parallel and give them the same input they will produce the same output. The randomness is something added after the LLM’s output, it isn’t something inherent to neural networks.

            • ageedizzle@piefed.ca
              link
              fedilink
              English
              arrow-up
              1
              ·
              4 hours ago

              Every computer program may as well be a nondeterministic system from the user’s point of view.

              Even a calculator?

          • FauxLiving@lemmy.world
            link
            fedilink
            arrow-up
            1
            arrow-down
            2
            ·
            4 days ago

            The entire comment chain is about LLMs, not chatbots.

            A LLM’s output is deterministic.

            A from an IT system point of view, a system using LLMs is only non-deterministic if the system adds randomness. The randomness isn’t inherent to the LLM and any added randomness is typically created by a seed which could be replicated across number of system images leading to deterministic output across every image clone.

            From the point of view of the typical user the computer may as well be a magical box controlled by a tiny fairy so any talk of deterministic output is irrelevant.

            • Catoblepas@piefed.blahaj.zone
              link
              fedilink
              English
              arrow-up
              6
              arrow-down
              1
              ·
              4 days ago

              The entire comment chain is about LLMs, not chatbots.

              The post is about the actual product they’re selling, not whatever idealized idea of what a ‘proper’ LLM is.

              If every LLM sold is sold as a chatbot, then this “ummm ackchully” is irrelevant. It doesn’t matter what LLMs are technically capable of if none of the LLMs sold actually function that way.

              • FauxLiving@lemmy.world
                link
                fedilink
                arrow-up
                2
                arrow-down
                1
                ·
                3 days ago

                The post is about the actual product they’re selling, not whatever idealized idea of what a ‘proper’ LLM is.

                Yes, that is what the post is about.

                You didn’t click on reply to the post, you clicked reply under my comment.

                In my comment, I was talking about an LLM (I checked with myself) and the other person was also talking about LLMs and on up to the top of the comment chain where we started talking about LLMs in IT systems.

                From the context of the conversation, you should understand that we’re talking about LLMs, specifically being in IT having to deal with LLMs. The context should tell you that we’re talking about the actual language models and not the end user applications, like a chatbot.

                If they aren’t selling non-chatbot LLMs then that’s irrelevant.

                Ok, well this is easy then. Every LLM isn’t sold as a chatbot so I’m not sure why you keep repeating this like it is a point.

                If every LLM sold is sold as a chatbot, then this “ummm ackchully” is irrelevant.

                Your first comment was ‘ummmm ackchully LLMs are only chatbots’ which is both wrong and ironic.