--- library_name: transformers tags: - sentiment-analysis - lora - roberta - fine-tuned - insurance --- # Model Card for RoBERTa LoRA Fine-Tuned for Insurance Review Rating This model is a fine-tuned version of RoBERTa (`roberta-large`) using LoRA adapters. It is specifically designed to classify English insurance reviews and assign a rating (on a scale of 1 to 5). ## Model Details ### Model Description This model uses RoBERTa (`roberta-large`) as its base architecture and was fine-tuned using Low-Rank Adaptation (LoRA) to adapt efficiently to the task of insurance review classification. The model predicts a rating from 1 to 5 based on the sentiment and context of a given review. LoRA fine-tuning reduces memory overhead and enables faster training compared to full fine-tuning. - **Developed by:** Lapujpuj - **Finetuned from model:** RoBERTa (`roberta-large`) - **Language(s) (NLP):** English - **License:** Apache-2.0 - **LoRA Configuration:** - Rank (r): 2 - LoRA Alpha: 16 - LoRA Dropout: 0.1 - **Task:** Sentiment-based rating prediction for insurance reviews ### Model Sources - **Repository:** [pujpuj/roberta-lora-token-classification](https://huggingface.co/pujpuj/roberta-lora-token-classification) - **Demo:** N/A --- ## Uses ### Direct Use This model can be directly used to assign a sentiment-based rating to insurance reviews. Input text is expected to be a sentence or paragraph in English. ### Downstream Use The model can be used as a building block for larger applications, such as customer feedback analysis, satisfaction prediction, or insurance service improvement. ### Out-of-Scope Use - The model is not designed for reviews in languages other than English. - It may not generalize well to domains outside of insurance-related reviews. - Avoid using the model for biased or malicious predictions. --- ## Bias, Risks, and Limitations ### Bias - The model is trained on a specific dataset of insurance reviews, which might include biases present in the training data (e.g., skewed ratings, linguistic or cultural biases). ### Risks - Predictions might not generalize well to other domains or review styles. - Inconsistent predictions may occur for ambiguous or mixed reviews. ### Recommendations - Always validate model outputs before making decisions. - Use the model in conjunction with other tools for a more comprehensive analysis. --- ## How to Get Started with the Model You can use the model with the following code snippet: ```python from transformers import AutoTokenizer from peft import PeftModel # Load tokenizer and model tokenizer = AutoTokenizer.from_pretrained("roberta-large") base_model = AutoModelForSequenceClassification.from_pretrained("roberta-large", num_labels=5) model = PeftModel.from_pretrained(base_model, "pujpuj/roberta-lora-token-classification") # Example prediction review = "The insurance service was quick and reliable." inputs = tokenizer(review, return_tensors="pt", truncation=True, padding=True) outputs = model(**inputs) rating = torch.argmax(outputs.logits, dim=1).item() + 1 print(f"Predicted rating: {rating}")