There’s another round of CSAM attacks and it’s really disturbing to see those images. It was really bothering to see those and they weren’t taken down immediately. There was even a disgusting shithead in the comments who thought it was funny?? the fuck

It’s gone now but it was up for like an hour?? This really ruined my day and now I’m figuring out how to download tetris. It’s really sickening.

  • I Cast Fist
    link
    fedilink
    English
    191 year ago

    AFAIK, it all falls down on moderators’ shoulders. I don’t envy their jobs one bit :(

    • @[email protected]
      link
      fedilink
      English
      71 year ago

      How was it handled on Reddit? Did the moderators have to handle it there as well, or did Reddit filter it out beforehand?

        • @[email protected]
          link
          fedilink
          English
          21 year ago

          Are any of the examples that your provided libre/free and open-source? I wasn’t able to find any info for Google’s, and Cloudflare seems to only offer theirs for free if you are already using Cloudflare’s services. If not the examples that you provided, does there exist any tools that are libre/free and open-source?

          • @PM_Your_Nudes_Please
            link
            English
            14
            edit-2
            1 year ago

            It will also make it a battle of attrition. Because now we’re not only using AI to block CSAM; Trolls are using AI to generate CSAM.

            The issue is that these tools typically work by hashing the image (or a specific section of the image) and checking it against a database of known CSAM. That way you never actually need to view the file to compare it to the list. But with AI image generation, that list of known CSAM is essentially useless because trolls can just generate new images.

            • @fubo
              link
              English
              61 year ago

              Even without the issue of new AI-generated images, those hash-based scanning tools aren’t available to hobbyist projects like the typical Lemmy instance. If they were given to hobbyist projects, it would be really easy for an abuser to just tweak their image collection until it didn’t set off the filter.

              • snowe
                link
                fedilink
                English
                71 year ago

                You can use CloudFlare’s CSAM scanning tool completely for free. You can’t get access to the hashes, which would allow what you are talking about.

                • @fubo
                  link
                  English
                  51 year ago

                  Sure, for Lemmy instances who are Cloudflare customers. But I don’t think it can be integrated with the Lemmy code by default.

                  • snowe
                    link
                    fedilink
                    English
                    31 year ago

                    No it can’t, and it shouldn’t be. It’s better to stop the CSAM before it ever makes it to any server you control rather than wait and then need to deal with it.

              • @fubo
                link
                English
                41 year ago

                On the other hand, if the people who want those images can satisfy their urges using AI fakes, that could mean less spreading of images of actual abuse. It might even mean less abuse happening.

                However, because they’re terrible people, I have to suspect that’s not the case.

                • @[email protected]
                  link
                  fedilink
                  English
                  41 year ago

                  People who create the content and insane monsters, but a LOT of actual pedos (vs predators looking for a power play) are disgusted by their preference. I know a ton of them look to cartoons already for stimulation, so I think AI content could draw more people away from actual material. Hopefully if demand reduces there will be less creation of new real content as the potential profits fall more proportionate to the risk.

      • @ShittyRedditWasBetter
        link
        English
        -61 year ago

        Reddit pays people.

        Unless this place monetizes its done. But everyone in this place expects everything to be free.

        • @[email protected]
          link
          fedilink
          English
          21 year ago

          They don’t pay moderators, so that’s a moot point.

          But I do agree in general that there needs to be money flow to developers and admins, and potentially moderators as well. Perhaps that can be done with donations, or perhaps there needs to be a profit model, IDK, but I haven’t seen a long term solution here.

          My opinion is that the federated model is broken, and we should be looking into decentralized models where users share some of the burden. That way monetization wouldn’t be an issue because there isn’t a huge infrastructure cost.

          But I’ll stick around while it works.

          • @ShittyRedditWasBetter
            link
            English
            -21 year ago

            Reddit pays admins, I think we are mostly on the same page on the rest. Donations will never happen though. It’s going to be at least $100 a year per person+ unless you end up cutting corners on stability and I don’t see the user base accepting that really.

            • @[email protected]
              link
              fedilink
              English
              11 year ago

              Nah, it’s more like a few hundred/month/instance, so if an instance has 5k users, it’s $1-2/user/month. So about a quarter of what you suggested.

              But again, it’s unlikely to actually happen. Voluntary donations tend to suffer from the Bystander Effect.

    • hypelightfly
      link
      fedilink
      31 year ago

      Especially as Lemmy has even worse moderator tools than reddit (without custom tools) and the devs don’t give a shit.

      • SpaceBar
        link
        English
        11 year ago

        They really don’t care, do they?