# DistilBERT Fine-Tuned for Sequence Classification ## Model Overview This is a fine-tuned version of the DistilBERT model designed for sequence classification tasks. It is inspired by the r/AmItheAsshole subreddit, where it has been trained on textual data to assess and classify user-submitted stories. - **Base Model**: [DistilBERT](https://huggingface.co/distilbert-base-uncased) - **Fine-Tuned For**: Sequence classification (e.g., sentiment analysis, AITA-type categorization) - **Dataset**: https://huggingface.co/datasets/MattBoraske/Reddit-AITA-2018-to-2022 - **Task**: Sequence classification with predefined labels. ## Model Details - **Architecture**: Transformer-based model (DistilBERT) - **Input Format**: Text sequences - **Output Format**: Classification labels with confidence scores - **Labels**: - `LABEL_0`: The Asshole - `LABEL_1`: Not the Asshole ## Intended Use This model is intended to provide insights and assessments for user-submitted textual scenarios. It works well for binary classification tasks. ### Example Usage ```python from transformers import pipeline classifier = pipeline( "text-classification", model="your-username/your-model-name" ) text = "I did not invite my friend for my wedding. AITA ?" result = classifier(text) print(result)