• Earthman_Jim@lemmy.zip
    link
    fedilink
    English
    arrow-up
    34
    arrow-down
    1
    ·
    edit-2
    2 months ago

    Why the fuck are people still using X unless they’re literally alt-right nazis and pedos?

    Americans: “tHeRe’S nOtHiNg wE cAn dO” Door dashes some 60 dollar chipotle while Xitting all over themselves.

    MAYBE STOP MAKING THESE FUCKERS RICHER EVERY FUCKING DAY?!?!?

    You don’t even have to have a general strike!! Just regain control of your fucking habits!!! Please!? Starve the actual beast.

    • shalafi@lemmy.world
      link
      fedilink
      English
      arrow-up
      13
      ·
      2 months ago

      I’m screaming this every day, and lemmy is especially ridiculous given our views on capitalism.

      In a thread about fast food prices last year, I was told I was privileged for suggesting that, “maybe stop buying their shit?”

      If every American had my wife and I’s spending habits, the economy would collapse in 3-4 months.

      • IronBird@lemmy.world
        link
        fedilink
        English
        arrow-up
        4
        ·
        2 months ago

        spending your $ consciously is literally the only way to fix this beast, the entire system is designed towards extracting it…therefor the only way to stress the system is to give your $ to (good) local/private businesses whenever possible.

      • Earthman_Jim@lemmy.zip
        link
        fedilink
        English
        arrow-up
        2
        ·
        edit-2
        2 months ago

        Exactly. Quiet quitting was a step in the right direction. What we need is a quiet strike. Take back our will from corporations, from food prep to social media/dating apps. Their hold is pervasive and destructive to the social fabric in almost every instance at this point. The key problem in the “free world” is that people have placed their faith in corporations and religious organizations and have learned to fear their neighbor by default, which is entirely backwards to a healthy society and hands all the power to the top.

    • BoycottTwitter@lemmy.zipOP
      link
      fedilink
      English
      arrow-up
      9
      arrow-down
      1
      ·
      2 months ago

      I totally agree with you. I literally made my username “BoycottTwitter” because it’s so important and so basic.

    • lechekaflan@lemmy.world
      link
      fedilink
      English
      arrow-up
      5
      arrow-down
      1
      ·
      edit-2
      2 months ago

      Why the fuck are people still using X

      For some they use it as a newsfeed without having to interact. For others, it’s utilized as a PR platform because partisans don’t limit themselves to Bluesky and Mastadon. Also, no need to pay the bastard for a blue checkmark.

    • architect@thelemmy.club
      link
      fedilink
      English
      arrow-up
      2
      ·
      2 months ago

      I had to sit at my in laws while they said straight to my face they were boycotting Coke products, Walmart, and Amazon. Right behind them was four dozen Coke cans and an Amazon box.

      Blast me to another fucking planet.

  • prac@lemmy.world
    link
    fedilink
    English
    arrow-up
    13
    ·
    2 months ago

    This is messed up tbh. Using AI to undress people—especially kids—shouldn’t even be technically possible, let alone.

    • Allero@lemmy.today
      link
      fedilink
      English
      arrow-up
      10
      arrow-down
      2
      ·
      2 months ago

      I feel like our relationship to it is also quite messed.

      AI doesn’t actually undress people, it just draws a naked body. It’s an artistic representation, not an X-ray. You’re not getting actual nudes in this process, and AI has no clue how the person looks like naked.

      Now, such images can be used to blackmail people, because again, our culture didn’t quite catch up with the fact that every nude image can absolutely be AI-generated fake. When it does, however, I fully expect creators of such things to be seen as odd creeps spreading their fantasies around and any nude imagery to be seen as fake by default.

      • nullroot@lemmy.world
        link
        fedilink
        English
        arrow-up
        8
        ·
        2 months ago

        It’s not an artistic representation, it’s worse. It’s algorithmic and to that extent it actually has a pretty good idea of what a person looks like naked based on their picture. That’s why it’s so disturbing.

        • bookmeat@lemmynsfw.com
          link
          fedilink
          English
          arrow-up
          1
          ·
          2 months ago

          Calling it an invasion of privacy is a stretch the way that copyright infringement is called theft.

        • aesthelete@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          1
          ·
          2 months ago

          Yeah they probably fed it a bunch of legitimate on/off content as well as stuff from people who used to do make “nudes” from celebrity photos with sheer / skimpy outfits as a creepy hobby.

            • Allero@lemmy.today
              link
              fedilink
              English
              arrow-up
              2
              arrow-down
              2
              ·
              2 months ago

              Honestly, I’d love to see more research on how AI CSAM consumption affects consumption of real CSAM and rates of sexual abuse.

              Because if it does reduce them, it might make sense to intentionally use datasets already involved in previous police investigations as training data. But only if there’s a clear reduction effect with AI materials.

              (Police has already used some materials, with victims’ consent, to crack down on CSAM sharing platforms in the past).

              • bookmeat@lemmynsfw.com
                link
                fedilink
                English
                arrow-up
                2
                ·
                2 months ago

                The idea is that to generate csam there was harm done to get the training data. This is why it’s bad.

                • Allero@lemmy.today
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  ·
                  2 months ago

                  That would be true if children were abused specifically to obtain the training data. But what I’m talking about is using the data that already exists, taken from police investigations and other sources. Of course, it also requires victim’s consent (as they grow old enough), as not everyone will agree to have materials of their abuse proliferate in any way.

                  Police has already used CSAM with victim’s consent to better impersonate CSAM platform admins in investigative operations, leading to arrests of more child abusers and those sharing the materials around. While controversial, this came as a net benefit as it allowed to reduce the amount of avenues for CSAM sharing and the amount of people able to do so.

                  The case with AI is milder, as it requires minimum human interaction, so no one will need to re-watch the materials as long as victims are already identified. It’s enough for the police to contact victims, get the agreement, and feed the data into AI without releasing the source. With enough data, AI could improve image and video generation, driving more watches away from real CSAM and reducing rates of abuse.

                  That is, if it works this way. There’s a glaring research hole in this area, and I believe it is paramount to figure out if it helps. Then, we could decide whether to include already produced CSAM into the data, or if adult data is sufficient to make it good enough for the intended audience to make a switch.

      • prac@lemmy.world
        link
        fedilink
        English
        arrow-up
        8
        ·
        2 months ago

        Idk, calling it ‘art’ feels like a reach. At the end of the day, it’s using someone’s real face for stuff they never agreed to. Fake or not, that’s still a massive violation of privacy.

    • andallthat@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      2 months ago

      Loophole. They didn’t cross their heart and hope to die. The only way is calling them out with Liar Liar Pants on Fire

  • ZILtoid1991@lemmy.world
    link
    fedilink
    English
    arrow-up
    7
    ·
    2 months ago

    Any government that does not ban Xitter after this mess are cucks to Musk and the Trump administration.

      • BreadstickNinja@lemmy.world
        link
        fedilink
        English
        arrow-up
        6
        ·
        2 months ago

        He sold a money-losing business with massive untapped potential to be a psyop to someone with infinite money and a desire for a psyop.

        Fucking great deal for both of them. Fucked in the ass for the rest of us.

    • KelvarCherry [They/Them]@lemmy.blahaj.zone
      link
      fedilink
      English
      arrow-up
      12
      ·
      2 months ago

      AI crawlers are scraping every site. Every site. Random public-but-unlisted hobby sites are getting scraped and spiking users’ data. There was a Lemmy post about someone who had that experience just yesterday.

      Think of how much Child Porn is stored on public sites that are shared in private groups. Also consider that FaceBook is the largest distributor of Child Sex Abuse material. These models are absolutely trained on Child Porn.

      • Earthman_Jim@lemmy.zip
        link
        fedilink
        English
        arrow-up
        4
        ·
        2 months ago

        “Also consider that FaceBook is the largest distributor of Child Sex Abuse material”

        Why is this not all over the news too? A rhetorical question. Sadly, I think we all know the answer by now.

    • Honytawk@feddit.nl
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      1
      ·
      2 months ago

      They don’t need to.

      Take pictures of normal dressed children, combine with pictures of naked adults. Now you have CP.

  • shirro@aussie.zone
    link
    fedilink
    English
    arrow-up
    3
    ·
    edit-2
    2 months ago

    The bad guys normalised racism, fraud, sexism and a lot of other evils and had people cheering for them as they flouted the law and protected each other from prosecution. It has been bad for ever regardless of who was in power with the two tier justice system that protected the elites while punishing everyone else. But we are in the end game now. The masks are off

    You too can be allowed to sample the vices of the wealthy elites, in moderation and under their control. In return all they ask is your loyalty. Seems kind of weird and out of touch to me. Regular people mostly just want good health, a place to raise a family, some free time and a decent income. No AI hallucination is going to substitute. They think so little of us.