• freakrho@lemmy.ml
    link
    fedilink
    arrow-up
    3
    ·
    3 years ago

    This algorithm creates missing data that fits the original image, but there are many different possibilities that can fit it. This can lead to false accusations if, for example, it is used to find a match in security footage. Here’s an example of the system creating a completely different face that fits the original image https://twitter.com/notlewistbh/status/1432936600745431041

    • poVoq@lemmy.mlOP
      link
      fedilink
      arrow-up
      2
      arrow-down
      1
      ·
      3 years ago

      No one is claiming that this should be used to enhance security camera footage though.

      • freakrho@lemmy.ml
        link
        fedilink
        arrow-up
        5
        ·
        3 years ago

        I think it’s important to consider it, because it totally will be used for cases like that if allowed

    • peppermint@lemmy.ml
      link
      fedilink
      arrow-up
      1
      ·
      3 years ago

      I’m starting to worry that the most popular perspective from which people not familiar with ML can tell loss functions and sampling algorithms apart is through the prism of BLM culture. Although, I agree with your sentiment that there is always a thin end of the wedge.