• Zink@programming.dev
    link
    fedilink
    arrow-up
    109
    ·
    17 hours ago

    Wow, this is really impressive y’all!

    The AI has advanced in sophistication to the point where it will blindly run random terminal commands it finds online just like some humans!

    I wonder if it knows how to remove the french language package.

    • greybeard@feddit.online
      link
      fedilink
      English
      arrow-up
      3
      ·
      4 hours ago

      The problem (or safety) of LLMs is that they don’t learn from that mistake. The first time someone says “What’s this Windows folder doing taking up all this space?” and acts on it, they wont make that mistake again. LLM? It’ll keep making the same mistake over and over again.

      • skisnow@lemmy.ca
        link
        fedilink
        English
        arrow-up
        1
        ·
        4 hours ago

        I recently had an interaction where it made a really weird comment about a function that didn’t make sense, and when I asked it to explain what it meant, it said “let me have another look at the code to see what I meant”, and made up something even more nonsensical.

        It’s clear why it happened as well; when I asked it to explain itself, it had no access to its state of mind when it made the original statement; it has no memory of its own beyond the text the middleware feeds it each time. It was essentially being asked to explain what someone who wrote what it wrote, might have been thinking.

        • greybeard@feddit.online
          link
          fedilink
          English
          arrow-up
          2
          ·
          4 hours ago

          One of the fun things that self hosted LLMs let you do (the big tech ones might too), is that you can edit its answer. Then, ask it to justify that answer. It will try its best, because, as you said, it its entire state of mind is on the page.

          • skisnow@lemmy.ca
            link
            fedilink
            English
            arrow-up
            1
            ·
            edit-2
            3 hours ago

            One quirk of github copilot is that because it lets you choose which model to send a question to, you can gaslight Opus into apologising for something that gpt-4o told you.