• I still think calling any of these poorly trained algorithms “AI” is an insult to the very concept of intelligence, artificial or otherwise. It’s like training a very single-minded dog, but one that can’t do general learning and apply what it learns to dissimilar problem spaces. I’m very salty about everyone just accepting the empty marketing coopting of the terminology.

      • @The_Monocle_Debacle@lemmygrad.ml
        link
        fedilink
        8
        edit-2
        2 years ago

        most of these so-called “AI” implementations are just self-optimizing algorithms trained with incomplete or biased data for a very specific problem. A lot of them can’t even do something in the same problem space that wasn’t part of their training data correctly.

        • Amicese
          link
          fedilink
          3
          edit-2
          2 years ago

          Oh yeah I see what you mean. I struggle with discerning them though.

          I worry that the training data for deepfakes is suspiciously normative. (there seems to be no neurodiverse, queer, or (physically) disabled people in those training sets).

          • @southerntofu@lemmy.ml
            link
            fedilink
            32 years ago

            Well first deepfakes need to die. It’s a dangerous tech that should not exist at all and does not need any more research.

            To be fair, i haven’t dug into deepfake models, but i assume you would train them on the specific person you’re trying to deepfake: i mean for basic video stuff going with a pre-trained model may be ok but for audio there’s no way you can get away with it ;)

            • poVoq
              link
              fedilink
              0
              edit-2
              2 years ago

              There are also specific ML models for audio that sound pretty convincing in replicating a specific person’s voice.