Shivangsinha
commited on
Commit
•
d44df85
1
Parent(s):
f925544
Update README.md
Browse files
README.md
CHANGED
@@ -23,9 +23,8 @@ This model performs sentiment analysis with a specific focus on detecting sarcas
|
|
23 |
|
24 |
This model leverages the **Flan-T5-small** transformer architecture, fine-tuned on datasets including **IMDB** for sentiment analysis and **bharatiyabytes/sentimentWithSarcasm** for sarcasm detection. By combining these datasets, the model is better equipped to differentiate between sarcastic and genuine sentiment expressions, improving sentiment analysis accuracy in contexts where sarcasm is prevalent.
|
25 |
|
26 |
-
- **Developed by:**
|
27 |
-
- **Funded by [optional]:**
|
28 |
-
- **Shared by [optional]:** [Your Organization, if applicable]
|
29 |
- **Model type:** Text Classification
|
30 |
- **Language(s) (NLP):** English (en)
|
31 |
- **License:** Apache 2.0
|
@@ -33,7 +32,7 @@ This model leverages the **Flan-T5-small** transformer architecture, fine-tuned
|
|
33 |
|
34 |
### Model Sources [optional]
|
35 |
|
36 |
-
- **Repository:**
|
37 |
- **Paper [optional]:** [Link to paper if available]
|
38 |
- **Demo [optional]:** [Link to demo if available]
|
39 |
|
@@ -43,7 +42,7 @@ This model leverages the **Flan-T5-small** transformer architecture, fine-tuned
|
|
43 |
|
44 |
This model can be used directly for sentiment classification on texts where sarcasm may obscure intended sentiment. It provides accurate classifications by considering nuanced expressions, making it ideal for social media analysis, customer feedback processing, and sarcasm-rich content sources.
|
45 |
|
46 |
-
### Downstream Use
|
47 |
|
48 |
This model can serve as a base for further fine-tuning for domains that require sarcasm-aware sentiment analysis, such as customer service, public relations, and social media monitoring applications.
|
49 |
|
@@ -73,7 +72,7 @@ model = AutoModelForSequenceClassification.from_pretrained("your-username/sarcas
|
|
73 |
inputs = tokenizer("Your input text here", return_tensors="pt")
|
74 |
outputs = model(**inputs)
|
75 |
predictions = torch.argmax(outputs.logits, dim=-1)
|
76 |
-
|
77 |
## Training Details
|
78 |
|
79 |
### Training Data
|
@@ -82,7 +81,7 @@ The model was fine-tuned on a combination of IMDB (a general sentiment analysis
|
|
82 |
|
83 |
### Training Procedure
|
84 |
|
85 |
-
#### Preprocessing
|
86 |
|
87 |
Standard text preprocessing methods were applied, such as tokenization and lowercase transformation.
|
88 |
|
@@ -94,7 +93,7 @@ Epochs: 3
|
|
94 |
Batch Size: 16
|
95 |
Learning Rate: 3e-5
|
96 |
|
97 |
-
#### Speeds, Sizes, Times
|
98 |
|
99 |
Training took approximately 4 hours on an NVIDIA V100 GPU with a model size of 60M parameters.
|
100 |
|
@@ -125,7 +124,7 @@ The model achieved an accuracy of 88% on the sentiment classification task and a
|
|
125 |
|
126 |
The model shows strong performance in sarcasm-sensitive sentiment analysis, making it suitable for applications where nuanced sentiment interpretation is crucial.
|
127 |
|
128 |
-
## Model Examination
|
129 |
|
130 |
The model’s predictions have been examined to ensure that sarcastic content is accurately labeled, using interpretability tools such as SHAP to visualize model attention on sarcastic phrases.
|
131 |
|
@@ -141,7 +140,7 @@ Carbon emissions can be estimated using the [Machine Learning Impact calculator]
|
|
141 |
- **Compute Region:** USA
|
142 |
- **Carbon Emitted:** NA
|
143 |
|
144 |
-
## Technical Specifications
|
145 |
|
146 |
### Model Architecture and Objective
|
147 |
|
|
|
23 |
|
24 |
This model leverages the **Flan-T5-small** transformer architecture, fine-tuned on datasets including **IMDB** for sentiment analysis and **bharatiyabytes/sentimentWithSarcasm** for sarcasm detection. By combining these datasets, the model is better equipped to differentiate between sarcastic and genuine sentiment expressions, improving sentiment analysis accuracy in contexts where sarcasm is prevalent.
|
25 |
|
26 |
+
- **Developed by:** bharatiyabytes
|
27 |
+
- **Funded by [optional]:** Not yet funded
|
|
|
28 |
- **Model type:** Text Classification
|
29 |
- **Language(s) (NLP):** English (en)
|
30 |
- **License:** Apache 2.0
|
|
|
32 |
|
33 |
### Model Sources [optional]
|
34 |
|
35 |
+
- **Repository:** https://github.com/sohi-g/lets-talk-the-hype.git
|
36 |
- **Paper [optional]:** [Link to paper if available]
|
37 |
- **Demo [optional]:** [Link to demo if available]
|
38 |
|
|
|
42 |
|
43 |
This model can be used directly for sentiment classification on texts where sarcasm may obscure intended sentiment. It provides accurate classifications by considering nuanced expressions, making it ideal for social media analysis, customer feedback processing, and sarcasm-rich content sources.
|
44 |
|
45 |
+
### Downstream Use
|
46 |
|
47 |
This model can serve as a base for further fine-tuning for domains that require sarcasm-aware sentiment analysis, such as customer service, public relations, and social media monitoring applications.
|
48 |
|
|
|
72 |
inputs = tokenizer("Your input text here", return_tensors="pt")
|
73 |
outputs = model(**inputs)
|
74 |
predictions = torch.argmax(outputs.logits, dim=-1)
|
75 |
+
```
|
76 |
## Training Details
|
77 |
|
78 |
### Training Data
|
|
|
81 |
|
82 |
### Training Procedure
|
83 |
|
84 |
+
#### Preprocessing
|
85 |
|
86 |
Standard text preprocessing methods were applied, such as tokenization and lowercase transformation.
|
87 |
|
|
|
93 |
Batch Size: 16
|
94 |
Learning Rate: 3e-5
|
95 |
|
96 |
+
#### Speeds, Sizes, Times
|
97 |
|
98 |
Training took approximately 4 hours on an NVIDIA V100 GPU with a model size of 60M parameters.
|
99 |
|
|
|
124 |
|
125 |
The model shows strong performance in sarcasm-sensitive sentiment analysis, making it suitable for applications where nuanced sentiment interpretation is crucial.
|
126 |
|
127 |
+
## Model Examination
|
128 |
|
129 |
The model’s predictions have been examined to ensure that sarcastic content is accurately labeled, using interpretability tools such as SHAP to visualize model attention on sarcastic phrases.
|
130 |
|
|
|
140 |
- **Compute Region:** USA
|
141 |
- **Carbon Emitted:** NA
|
142 |
|
143 |
+
## Technical Specifications
|
144 |
|
145 |
### Model Architecture and Objective
|
146 |
|