- cross-posted to:
- [email protected]
- cross-posted to:
- [email protected]
In this paper we take a broad look at child sexual exploitation concerns on decentralized social media, present new findings on the nature and prevalence of child safety issues on the Fediverse, and offer several proposals to improve the ecosystem in a sustainable manner. We focus primarily on the Fediverse (i.e., the ecosystem supporting the ActivityPub protocol) and Mastodon, but several techniques could also be repurposed on decentralized networks such as Nostr, or semi-centralized networks such as Bluesky.
I am not sure how to feel about this. On the other hand it would mean that a node in the network cannot “outgrow” its own “moderation power”. But this would make the nodes vulnerable to a terrible form of attack or it would force the node to implement more rigorous sing-up procedures. All of these scenarios could end up pretty distopian!
I believe if users are empowered to participate in combating the problem, almost all users would. The question is, how can the nodes harvest this? What can ordinary users do to help?
Good point. If the main function would be implemented on all nodes and everyone would exchange information (hashes), which elements are harmful, we would maybe have a chance to skip central nodes from the equation. If the information is distributed throughout multiple maybe random nodes, but this would pose new problems - where to find that information.
I agree. Maybe the node can create some “unsafe” hashes and share them between the nodes - kind of like an antivirus creates hashes for malware. The user can always report to the admins - when there are more instances there are automatically more admins to handle the request themselves or with bots.