I’m sure I’m not the first to think about this, but I haven’t seen it mentioned yet. I believe there is another problem with open registration instances besides just the threat of spamming.

If you go through the process of making a new post in ANY community and you attach an image and then cancel the post the image is STILL on the server at the given URL and is publically viewable if someone has that URL.

Theoretically, someone could upload illegal images this way and hotlink to them from another site. Because there are no posts on the instance with the bad image attached an instance admin would have no way of knowing the images were there unless they make a habit of browsing the pict-rs datastore regularly. There’s currently no easy way to moderate or delete images in the pict-rs datastore.

I don’t think I need to elaborate more on what kinds of images could be lurking on your very own server which could be hotlinked into VERY dark places on the web. Saying that you “didnt know” that they were there is not a defense. When the authorities are knocking on your door because you are hosting illegal images you will be sorry that you didn’t take a more active role in your user base.

I realize that even if you close or set manual registration that there’s still the danger of a bad user doing this very thing, but I think putting in some minor hurdles would greatly decrease the chances.

Regardless, I think there needs to be a better way to manage the pict-rs part of Lemmy and an easy way for admins and instance owners to be able to view EVERY attachment on their server to make sure there’s nothing there that could get them in trouble.

I run a small instance that just have people I personally know on it, so it’s not a worry for me. But the larger instances that are opening signups to strangers should be aware and take precautions.

  • @entropicshart
    link
    English
    18
    edit-2
    11 months ago

    Honestly this sounds like a bug that needs to be fixed to prevent image uploads before the post is committed.

    Furthermore, adding abilities to scan uploaded content for inappropriate images could help auto remove such content that makes it past moderation, or even before moderation

    • @[email protected]
      link
      fedilink
      English
      1
      edit-2
      11 months ago

      Yeah, definitely sounds like a bug. The fix doesn’t even have to prevent image uploads before post commit, it just needs a way to periodically cull images that don’t have any posts / comments referencing it. This seems like necessary functionality regardless to prevent server storage requirements from ballooning out of control and not to mention for CCPA and GDPR purposes.