• msage@programming.dev
    link
    fedilink
    English
    arrow-up
    24
    arrow-down
    4
    ·
    6 months ago

    While I agree with your attitude, the whole ‘laissez-fair’ thing is probably a misunderstanding:

    There is nothing we can do to stop the AI.

    Nothing.

    The genie is out of the bottle, the Pandora’s box has been opened, everything is out and it won’t ever return. The world will never be the same, and it’s irrelevant what people think.

    That’s why we need to better understand the post-AI world we created, and figure out what do to now.

    Also, to hell with CP. (feels weird to use the word ‘fuck’ here)

    • retrospectology@lemmy.world
      link
      fedilink
      English
      arrow-up
      7
      arrow-down
      16
      ·
      edit-2
      6 months ago

      Thats not the question, the question is not “can we stop AI entirely” it’s about regulating its development and yes, we can make efforts to do that.

      This attitude of “it’s inevitable, can’t do anything about it” is eerily similar logic to what is used in climate denial and other right-wing efforts. It’s a really poor attitude to have, especially about something as consequential as AI.

      We have the best opportunity right now to create rules about its uses and development. The answer is not “do nothing” as if it’s some force of nature, as opposed toa tool created by humans.

      • msage@programming.dev
        link
        fedilink
        English
        arrow-up
        7
        arrow-down
        1
        ·
        6 months ago

        I hear you, and I don’t necessarily disagree with you, I just know that’s not how anything works.

        Regulations work for big companies, but there isn’t a big company behind this specific case. And those small-time users have run away and you can’t stop them.

        It’s like trying to regulate cameras to not store specific images. Like, I get the sentiment, but sorry, no. It’s not that I would not like that, it’s just not possible.

        • retrospectology@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          6
          ·
          6 months ago

          This argument could be applied to anything though. A lot of people get away with myrder, we should still try and do what we can to stop it from happening.

          You can’t sit in every car and force people to wear a seatbelt, we still have seatbelt laws and regulations for manufacturers.

          • msage@programming.dev
            link
            fedilink
            English
            arrow-up
            5
            arrow-down
            1
            ·
            6 months ago

            Physical things are much easier to regulate than software, much less serverless.

            We already regulate certain images, and it matters very little.

            The bigger payoff will be from educating the public and accepting that we can’t win every war.

            • retrospectology@lemmy.world
              link
              fedilink
              English
              arrow-up
              1
              arrow-down
              6
              ·
              edit-2
              6 months ago

              So accept defeat from the start, that’s really just a non-starter. AI models run on hardware, they are developed by specific people, their contents are distributed by specific individuals, code bases are hosted on hardware and on specific outlets.

              It really does sound like you’re just trying to make excuses to avoid regulation, not that you genuinely have a good reason to think it’s not possible to try.

      • L_Acacia@lemmy.one
        link
        fedilink
        English
        arrow-up
        3
        ·
        edit-2
        6 months ago

        The models used are not trained on CP. The models weight are distributed freely and anybody can train a LORA on his computer. Its already too late to ban open weight models.

      • GBU_28@lemm.ee
        link
        fedilink
        English
        arrow-up
        2
        ·
        edit-2
        6 months ago

        Dude the amount of open source, untrackable, distributed ai models is off the charts. This isn’t just about the models offered by subscription from the big players.

        • retrospectology@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          6 months ago

          This is still one of the weaker arguments. There is a lot of malware out there too, people are still prosecuted when they’re caught developing and distributing it, we don’t just throw up our hands and pretend there’s nothing that can be done.

          Like, yeah, some pedophile who also happens to be tech saavy might build his own AI model to make CP, that’s not some self-evident argument against attempting to stop them.

          • GBU_28@lemm.ee
            link
            fedilink
            English
            arrow-up
            1
            arrow-down
            1
            ·
            edit-2
            6 months ago

            No, like, the tools to do these things are common and readily available. It’s not malware, it’s generalized ai tools, completely embroiled with non image ai work.

            Pandora’s box is wide open. All of this work can be done trivially, completely offline with a basic PC. Anyone motivated can be offline and up and running in a weekend

            You’re asking to outlaw something like a spreadsheet.

            You download a general purpose image ai model, then train and prompt it completely offline