Telegram bot that detects toxic comments based on Perspective API
-
Updated
Dec 9, 2021 - TypeScript
Telegram bot that detects toxic comments based on Perspective API
Comparing Toxic Texts with Transformers
This library can detect toxicity in the text or string or content and in return provide you the toxicity percentage in text or content with toxic words found in text.
Measure and mitigate gender bias in Danish toxicity classifiers and sentiment analysis models.
Toxicity detection in a conversation or phases.
In-game Toxic Language Detection: Shared Task and Attention Residuals
It is a trained Deep Learning model to predict different level of toxic comments. Toxicity like threats, obscenity, insults, and identity-based hate.
This is a application to analyse toxicity in social media using BERT and context analysis and aims to reduce toxicity
project for ms_system_design
This is a simple python program which uses a machine learning model to detect toxicity in tweets, GUI in Tkinter.
An anti-toxicity Discord bot to ease moderation.
This demo shows the functionality of the Voximplant instant messaging SDK, including silent supervision by a bot.
This work focuses on the development of machine learning models, in particular neural networks and SVM, where they can detect toxicity in comments. The topics we will be dealing with: a) Cost-sensitive learning, b) Class imbalance
An Explainable Toxicity detector for code review comments. Published in ESEM'2023
NLP deep learning model for multilingual toxicity detection in text 📚
This is a simple python program which uses a machine learning model to detect toxicity in tweets, developed in Flask.
A supervised learning based tool to identify toxic code review comments
AntiToxicBot is a bot that detects toxics in a chat using Data Science and Machine Learning technologies. The bot will warn admins about toxic users. Also, the admin can allow the bot to ban toxics.
Add a description, image, and links to the toxicity-detection topic page so that developers can more easily learn about it.
To associate your repository with the toxicity-detection topic, visit your repo's landing page and select "manage topics."