tashrifmahmud
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -12,32 +12,64 @@ metrics:
|
|
12 |
model-index:
|
13 |
- name: sentiment_analysis_model_v2
|
14 |
results: []
|
|
|
|
|
|
|
|
|
|
|
15 |
---
|
16 |
|
17 |
-
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
18 |
-
should probably proofread and complete it, then remove this comment. -->
|
19 |
-
|
20 |
# sentiment_analysis_model_v2
|
21 |
|
22 |
-
This model is a fine-tuned version of [tashrifmahmud/sentiment_analysis_model](https://huggingface.co/tashrifmahmud/sentiment_analysis_model) on
|
23 |
-
It achieves the following results on the evaluation set:
|
24 |
-
- Loss: 0.
|
25 |
-
- Accuracy: 0.
|
26 |
-
- Precision: 0.
|
27 |
-
- Recall: 0.
|
28 |
-
- F1: 0.
|
29 |
|
30 |
## Model description
|
31 |
|
32 |
-
|
|
|
|
|
|
|
|
|
|
|
33 |
|
34 |
## Intended uses & limitations
|
35 |
|
36 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
37 |
|
38 |
## Training and evaluation data
|
39 |
|
40 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
41 |
|
42 |
## Training procedure
|
43 |
|
@@ -66,4 +98,4 @@ The following hyperparameters were used during training:
|
|
66 |
- Transformers 4.46.2
|
67 |
- Pytorch 2.5.1+cu121
|
68 |
- Datasets 3.1.0
|
69 |
-
- Tokenizers 0.20.3
|
|
|
12 |
model-index:
|
13 |
- name: sentiment_analysis_model_v2
|
14 |
results: []
|
15 |
+
datasets:
|
16 |
+
- stanfordnlp/imdb
|
17 |
+
- cornell-movie-review-data/rotten_tomatoes
|
18 |
+
language:
|
19 |
+
- en
|
20 |
---
|
21 |
|
|
|
|
|
|
|
22 |
# sentiment_analysis_model_v2
|
23 |
|
24 |
+
This model is a fine-tuned version of [tashrifmahmud/sentiment_analysis_model](https://huggingface.co/tashrifmahmud/sentiment_analysis_model) on [IMDB](https://huggingface.co/datasets/stanfordnlp/imdb) and ["Rotten Tomatoes"](https://huggingface.co/datasets/cornell-movie-review-data/rotten_tomatoes) dataset.
|
25 |
+
It achieves the following results on the evaluation set (model has been updated at epoch 1):
|
26 |
+
- Loss: 0.3682
|
27 |
+
- Accuracy: 0.8396
|
28 |
+
- Precision: 0.8267
|
29 |
+
- Recall: 0.8593
|
30 |
+
- F1: 0.8427
|
31 |
|
32 |
## Model description
|
33 |
|
34 |
+
This model is a fine-tuned version of the DistilBERT transformer architecture for sentiment analysis. It was trained on the IMDB dataset for binary classification, distinguishing between positive and negative sentiment in movie reviews. The model has been further fine-tuned on the Rotten Tomatoes dataset to improve its generalization and performance on movie-related text.
|
35 |
+
|
36 |
+
**Architecture:** DistilBERT (a distilled version of BERT for faster inference).
|
37 |
+
**Task:** Sentiment Analysis (binary classification: positive or negative sentiment).
|
38 |
+
**Pre-training:** The model was pre-trained on a large corpus (BERT's original training).
|
39 |
+
**Fine-tuning:** Fine-tuned using both IMDB and Rotten Tomatoes datasets.
|
40 |
|
41 |
## Intended uses & limitations
|
42 |
|
43 |
+
**Intended uses:**
|
44 |
+
|
45 |
+
This model is suitable for classifying the sentiment of text, particularly movie reviews. It can be used in various applications such as:
|
46 |
+
Sentiment analysis for social media posts, customer reviews, or product feedback.
|
47 |
+
Analyzing movie reviews, comments, or related textual data.
|
48 |
+
As part of a sentiment-aware recommendation system, content moderation tool, or market research.
|
49 |
+
|
50 |
+
**Limitations:**
|
51 |
+
|
52 |
+
The model is specifically tuned for movie-related sentiment analysis. Its performance on non-movie-related text (e.g., general product reviews, news articles) may not be optimal.
|
53 |
+
The model may not perform well on texts with highly domain-specific terminology outside of movie-related contexts.
|
54 |
+
This model may struggle with sarcasm, irony, and nuanced expressions of sentiment, as is typical with many sentiment analysis models.
|
55 |
|
56 |
## Training and evaluation data
|
57 |
|
58 |
+
**Training data:**
|
59 |
+
|
60 |
+
**IMDB dataset:** The model was initially trained on the IMDB movie reviews dataset, which consists of 25,000 reviews labeled as positive or negative.
|
61 |
+
**Rotten Tomatoes dataset:** To improve the model's performance and generalization, it was further fine-tuned using the Rotten Tomatoes dataset, which contains movie reviews and ratings.
|
62 |
+
Evaluation data:
|
63 |
+
|
64 |
+
**Test data from Rotten Tomatoes:** The model's evaluation was performed using the test set of the Rotten Tomatoes dataset to assess its ability to generalize to unseen movie reviews.
|
65 |
+
|
66 |
+
**Improvement Metrics after fine-tuning on Rotten Tomatoes:**
|
67 |
+
- Accuracy increased from 82.1% to 84.3%.
|
68 |
+
- Precision improved from 81.61% to 84.81%.
|
69 |
+
- F1 Score saw a boost from 83.62% to 85.37%.
|
70 |
+
- Loss decreased from 0.4268 to 0.3621.
|
71 |
+
- Runtime was reduced from 17.7 seconds to 15.61 seconds.
|
72 |
+
- The model's throughput improved, with samples per second increasing from 56.49 to 64.05, and steps per second from 7.06 to 8.01.
|
73 |
|
74 |
## Training procedure
|
75 |
|
|
|
98 |
- Transformers 4.46.2
|
99 |
- Pytorch 2.5.1+cu121
|
100 |
- Datasets 3.1.0
|
101 |
+
- Tokenizers 0.20.3
|