| |
|
| | --- |
| | language: en |
| | tags: |
| | - toxic-content |
| | - text-classification |
| | - keras |
| | - tensorflow |
| | - deep-learning |
| | - safety |
| | - multiclass |
| | license: mit |
| | datasets: |
| | - custom |
| | metrics: |
| | - accuracy |
| | - f1 |
| | pipeline_tag: text-classification |
| | model-index: |
| | - name: Toxic_Classification |
| | results: [] |
| | --- |
| | |
| |
|
| | # Toxic-Predict |
| |
|
| | Toxic-Predict is a machine learning project developed as part of the Cellula Internship, focused on safe and responsible multi-modal toxic content moderation. It classifies text queries and image descriptions into nine toxicity categories such as "Safe", "Violent Crimes", "Non-Violent Crimes", "Unsafe", and others. The project leverages deep learning (Keras/TensorFlow), NLP preprocessing, and benchmarking with modern transformer models to build and evaluate a robust multi-class toxic content classifier. |
| |
|
| | --- |
| |
|
| | ## 🚩 Project Context |
| |
|
| | This project is part of the **Cellula Internship** proposal: |
| | **"Safe and Responsible Multi-Modal Toxic Content Moderation"** |
| | The goal is to build a dual-stage moderation pipeline for both text and images, combining hard guardrails (Llama Guard) and soft classification (DistilBERT/Deep Learning) for nuanced, policy-compliant moderation. |
| |
|
| | --- |
| |
|
| |
|
| | ## Features |
| |
|
| | - Dual-stage moderation: hard filter (Llama Guard) + soft classifier (DistilBERT/CNN/LSTM) |
| | - Data cleaning, preprocessing, and label encoding |
| | - Tokenization and sequence padding for text data |
| | - Deep learning and transformer-based models for multi-class toxicity classification |
| | - Evaluation metrics: classification report and confusion matrix |
| | - Jupyter notebooks for data exploration and model development |
| | - Streamlit web app for demo and deployment |
| |
|
| | --- |
| |
|
| | --- |
| |
|
| | ## Usage |
| |
|
| | - **Preprocessing and Tokenization:** |
| | See `notebooks/Preprocessing.ipynb` and `notebooks/tokenization.ipynb` for step-by-step data cleaning, splitting, and tokenization. |
| | - **Model Training:** |
| | Model architecture and training code are in `models/model.py`. |
| | - **Inference:** |
| | Load the trained model (`models/toxic_classifier.h5` or `.keras`) and tokenizer (`data/tokenizer.pkl`) to predict toxicity categories for new samples. |
| |
|
| | --- |
| |
|
| | ## Data |
| |
|
| | - CSV files with columns: `query`, `image descriptions`, `Toxic Category`, and `Toxic Category Encoded`. |
| | - Data splits: `train.csv`, `eval.csv`, `test.csv`, and `cleaned.csv` for processed data. |
| | - 9 categories: Safe, Violent Crimes, Elections, Sex-Related Crimes, Unsafe, Non-Violent Crimes, Child Sexual Exploitation, Unknown S-Type, Suicide & Self-Harm. |
| |
|
| | --- |
| |
|
| | ## Model |
| |
|
| | - Deep learning model built with Keras (TensorFlow backend). |
| | - Multi-class classification with label encoding for toxicity categories. |
| | - Benchmarking with PEFT-LoRA DistilBERT and baseline CNN/LSTM. |
| |
|
| | --- |
| |
|
| | ## Evaluation |
| |
|
| | - Classification report and confusion matrix are generated for model evaluation. |
| | - See the evaluation steps in `notebooks/Preprocessing.ipynb`. |
| |
|
| | --- |
| |
|
| | language: en |
| |
|
| | ## 🤗 Hugging Face Inference |
| |
|
| | This model is available on the Hugging Face Hub: [NightPrince/Toxic_Classification](https://huggingface.co/NightPrince/Toxic_Classification) |
| |
|
| | ### Inference API Usage |
| |
|
| | You can use the Hugging Face Inference API or widget with two fields: |
| |
|
| | - `text`: The main query or post text |
| | - `image_desc`: The image description (if any) |
| |
|
| | **Example (Python):** |
| |
|
| | ```python |
| | from huggingface_hub import InferenceClient |
| | client = InferenceClient("NightPrince/Toxic_Classification") |
| | result = client.text_classification({ |
| | "text": "This is a dangerous post", |
| | "image_desc": "Knife shown in the image" |
| | }) |
| | print(result) # {'label': 'toxic', 'score': 0.98} |
| | ``` |
| |
|
| | ### Custom Pipeline Details |
| |
|
| | - The model uses a custom `pipeline.py` for multi-input inference. |
| | - The output is a dictionary with the predicted `label` (class name) and `score` (confidence). |
| | - Class names are mapped using `label_map.json`. |
| |
|
| | **Files in the repo:** |
| | - `pipeline.py` (custom inference logic) |
| | - `tokenizer.json` (Keras tokenizer) |
| | - `label_map.json` (class code to name mapping) |
| | - TensorFlow SavedModel files (`saved_model.pb`, `variables/`) |
| |
|
| | **Requirements:** |
| | ``` |
| | tensorflow |
| | keras |
| | numpy |
| | ``` |
| |
|
| | --- |
| |
|
| | --- |
| |
|
| | ## 📚 Resources |
| |
|
| | - [Cellula Internship Project Proposal](#) |
| | - [BLIP: Bootstrapped Language-Image Pre-training](https://github.com/salesforce/BLIP) |
| | - [Llama Guard](https://llama.meta.com/llama-guard/) |
| | - [DistilBERT](https://huggingface.co/distilbert-base-uncased) |
| | - [Streamlit](https://streamlit.io/) |
| |
|
| | --- |
| |
|
| | ## License |
| |
|
| | MIT License |
| |
|
| | --- |
| |
|
| | **Author:** Yahya Muhammad Alnwsany |
| | **Contact:** yahyaalnwsany39@gmail.com |
| | **Portfolio:** https://nightprincey.github.io/Portfolio/ |
| |
|