Deepnude Censorship Nsfw Safesearch
you can turn on/off safesearch by changing ip to iran. is there better options, doing this programmatically?
nsfw createml imageclassifier with pretrained model
敏感词过滤器 also need a high speed censor engine using trie, also in diffferent language(like english or japanese)
english profanity:
https://github.com/snguyenthanh/better_profanity
https://github.com/zacanger/profane-words
https://github.com/MauriceButler/badwords
https://github.com/vzhou842/profanity-check
https://github.com/coffee-and-fun/google-profanity-words
chinese profanity:
https://github.com/gaohuifeng/sensitive-word-filter
https://github.com/lunzima/profanities.txt
https://github.com/nyx1987/forbiddenwords
https://github.com/insoxin/bannedwords
https://github.com/chason777777/mgck
https://github.com/k5h9999/keywordfilter
https://github.com/tomzhang/bannedwords
https://github.com/observerss/textfilter (需要回看历史 查看git历史)
https://github.com/aojiaotage/text-censor
deepnude nsfw nude picture detection:
https://github.com/yuanxiaosc/DeepNude-an-Image-to-Image-technology/blob/master/README-ZH.md
videoaudit 视频审核框架https://github.com/minitrill/VideoAudit
nsfw: not safe for work, inappropriate content, porn, offensive
https://github.com/rockyzhengwu/nsfw
https://github.com/alex000kim/nsfw_data_scraper
https://github.com/yahoo/open_nsfw
https://github.com/Rayraegah/nsfw_japan
https://github.com/fishsup/nsfw-image-classification
https://github.com/yangbisheng2009/nsfw-resnet
https://github.com/GantMan/nsfw_model
https://github.com/devzwy/open_nsfw_android
https://github.com/infinitered/nsfwjs
https://github.com/nsfw-filter/nsfw-filter
nudity, violence and drug
https://github.com/amshrbo/nsfw-detection
to train these networks, suitable datasets are required.
找训练集 找涉政 涉黄 暴力血腥图片训练集 找类似的文字训练集 百度aistudio可能有 github可能有 百度一下也可能有
同样的思路 情绪 情感打分也可以这样打分 根据不同的训练集进行打分
github violence detection:
https://github.com/topics/violence-detection
Bloody Image Classification with Global and Local Features
https://www.researchgate.net/publication/309365631_Bloody_Image_Classification_with_Global_and_Local_Features
Object content understanding in images and videos draws more and more attention nowadays. However, only few existing methods have addressed the problem of bloody scene detection in images. Along with the widespread popularity of the Internet, violent contents have affected our daily life. In this paper, we propose region-based techniques to identify a color image being bloody or not. Firstly, we have established a new dataset containing 25431 bloody images and 25431 non-bloody images. These annotated images are derived from the Violent Scenes Dataset, a public shared dataset for violent scenes detection in Hollywood movies and web videos. Secondly, we design a bloody image classification method with global visual features using Support Vector Machines. Thirdly, we also construct a novel bloody region identification approach using Convolutional Neural Networks. Finally, comparative experiments show that bloody image classification with local features is more effective.
search for nsfw filter providers on google/kaggle
kaggle nsfw image dataset
https://www.kaggle.com/datasets/laxmansingh/nsfw-images-data
https://www.kaggle.com/datasets/drakedtrex/my-nsfw-dataset/code
search for nsfw image/text on github:
https://github.com/alex000kim/nsfw_data_scraper
https://github.com/nsfw-filter/nsfw-filter
nsfwjs
https://github.com/arufian/Image-Censor-Lightning-Web-Component
https://github.com/enymuss/censorText
https://github.com/fmsky/resnet50_inappropriate_content_detect
https://github.com/CheranMahalingam/Image_Content_Moderation
文本审核框架:
https://github.com/minitrill/TextAudit
规避文本审查:有可能是加密了 但是人眼可以识别
https://github.com/kallydev/shutup
nudenet based inappropriate image censoring
ocr based word censoring
politics
涉及政治:领导人 人脸识别 图标识别
violence
血腥暴力:图像识别