Is there a way to automatically (and quickly) check texts and images for explicit content?
Will I take responsibility for this if I didn’t notice and delete it in time? If the app is going to scale, it’s almost impossible to check everything that users will create.
then, make sure you’re getting them to verify their email address before they can use the website. it’s too annoying to go through lots of email address just to be a douchbag on a new platform.
have a way people can report abusive content so you’ll get notifications in your email with a link to the content and you can review. if you’re app is scaling (which is a great thing!) - you’ll have more resources to invest into content moderation (but don’t worry about it too much right now) - just try to make it easy for people to report it (small icon in the bottom left or right of the content space that shows a group that has a workflow to report content.
That’s why we made these for you :
Detect negative sentiments form text:
Detect explicit content detection from images
Detect explicit content detection from videos
wow- super cool! thanks for sharing @redvivi
Thanks for sharing!
But it takes quite some time to check images (as it is in their example). Going to be tedious for users to wait that much.
Once executed once in Bubble’s backend from your app, the subsequent calls for different pictures are quick (around 1-2 seconds) thanks to Bubble’s intelligent caching.
Alternatively, you could also moderate the pictures at a later stage, not synchronised with the user upload.