0
submitted 1 year ago by _Frog@feddit.ch to c/technology@feddit.ch

In this paper we take a broad look at child sexual exploitation concerns on decentralized social media, present new findings on the nature and prevalence of child safety issues on the Fediverse, and offer several proposals to improve the ecosystem in a sustainable manner. We focus primarily on the Fediverse (i.e., the ecosystem supporting the ActivityPub protocol) and Mastodon, but several techniques could also be repurposed on decentralized networks such as Nostr, or semi-centralized networks such as Bluesky.

you are viewing a single comment's thread
view the rest of the comments
[-] Maddox@feddit.ch 2 points 1 year ago

This contradicts somehow the use case of a federated network, but there would also be major benefits from something like this.

I am not sure how to feel about this. On the other hand it would mean that a node in the network cannot "outgrow" its own "moderation power". But this would make the nodes vulnerable to a terrible form of attack or it would force the node to implement more rigorous sing-up procedures. All of these scenarios could end up pretty distopian!

I believe if users are empowered to participate in combating the problem, almost all users would. The question is, how can the nodes harvest this? What can ordinary users do to help?

[-] _Frog@feddit.ch 2 points 1 year ago

Good point. If the main function would be implemented on all nodes and everyone would exchange information (hashes), which elements are harmful, we would maybe have a chance to skip central nodes from the equation. If the information is distributed throughout multiple maybe random nodes, but this would pose new problems - where to find that information.

I believe if users are empowered to participate in combating the problem, almost all users would. The question is, how can the nodes harvest this? What can ordinary users do to help?

I agree. Maybe the node can create some "unsafe" hashes and share them between the nodes - kind of like an antivirus creates hashes for malware. The user can always report to the admins - when there are more instances there are automatically more admins to handle the request themselves or with bots.

this post was submitted on 26 Jul 2023
0 points (NaN% liked)

TeCHnology

516 readers
1 users here now

Technology discussion for Switzerland. This community shall discuss various topics of technology in and around Switzerland.

founded 2 years ago
MODERATORS