--- language: en tags: - emotion-classification - bert - lora license: mit --- # Emotion Classification Model This model is a fine-tuned version of `bert-base-uncased` on the "dair-ai/emotion" dataset, using LoRA (Low-Rank Adaptation) for efficient fine-tuning. label_list={"sadness", "joy", "love", "anger" ,"fear","surprise"} ## Model description [Describe your model, its architecture, and the task it performs] ## Intended uses & limitations [Describe what the model is intended for and any limitations] ## Training and evaluation data The model was trained on the "dair-ai/emotion" dataset. ## Training procedure [Describe your training procedure, hyperparameters, etc.] ## Eval results [Include your evaluation results] ## How to use Here's how you can use the model: ```python from transformers import AutoModelForSequenceClassification, AutoTokenizer model = AutoModelForSequenceClassification.from_pretrained("ahmetyaylalioglu/text-emotion-classifier") tokenizer = AutoTokenizer.from_pretrained("ahmetyaylalioglu/text-emotion-classifier") text = "I am feeling very happy today!" inputs = tokenizer(text, return_tensors="pt") outputs = model(**inputs) predictions = outputs.logits.argmax(-1) print(model.config.id2label[predictions.item()])