Feel like we’ve got a lot of tech savvy people here seems like a good place to ask. Basically as a dumb guy that reads the news it seems like everyone that lost their mind (and savings) on crypto just pivoted to AI. In addition to that you’ve got all these people invested in AI companies running around with flashlights under their chins like “bro this is so scary how good we made this thing”. Seems like bullshit.

I’ve seen people generating bits of programming with it which seems useful but idk man. Coming from CNC I don’t think I’d just send it with some chatgpt code. Is it all hype? Is there something actually useful under there?

  • unknowing8343@discuss.tchncs.de
    link
    fedilink
    arrow-up
    1
    arrow-down
    1
    ·
    1 year ago

    You have not realised yet that… yes, it has all the right to be called AI. They are doing the same thing we do. Learn and then create thoughts based on those learnings.

    I even asked them to make up words that are not related to any language, and they create them, entirely new, never-used words, that are not even composites of others. These are creative machines. They might fail at answering some questions, but that is partially why we call it Artificial Intelligence. It’s not saying that it is a machine of truth. Just a machine that “learns” and “knows”. Sometimes correctly, sometimes wrong. Just like us.

    • nickwitha_k (he/him)@lemmy.sdf.org
      link
      fedilink
      arrow-up
      2
      ·
      1 year ago

      Incorrect. An LLM COULD be a part of a system that implements AI but, itself, possesses no intelligence. Claiming otherwise is akin to claiming that the Pythagorean theorem is an AI because it “understands” geometry. Neither actually understands the data that they are fed but, are good at producing results that make it seem that way.

      Human cognition does not work that way; it is much more complex and squishy. Association of current experiences with remembered experiences is only a fraction of what is going on in a brain related to cognition.

      • unknowing8343@discuss.tchncs.de
        link
        fedilink
        arrow-up
        0
        arrow-down
        1
        ·
        1 year ago

        I am not saying it works exactly like humans inside of the black box. I just say it works. It learns and then creates thoughts. And it works.

        You talk about how human cognition is more complex and squishy, but nobody really knows how it truly works inside.

        All I see is the same kind of blackbox. A kid trying many, many times to stand up, or to say “papa”, until it somehow works, and now the pathway is setup in the brain.

        Obviously ChatGPT is just dealing with text. But does it make it NOT intelligent? I think it makes it very text-intelligent. Just add together all the AI pieces we are building and you got yourself a general AI that will do anything we do.

        Yeah, maybe it does not work like our brain. But is a human brain structure the only possible structure for intelligence? I don’t think so.