Text Classification
Safetensors
English
t5
sarcasm
gen-ai
llms
Shivangsinha commited on
Commit
d44df85
1 Parent(s): f925544

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -10
README.md CHANGED
@@ -23,9 +23,8 @@ This model performs sentiment analysis with a specific focus on detecting sarcas
23
 
24
  This model leverages the **Flan-T5-small** transformer architecture, fine-tuned on datasets including **IMDB** for sentiment analysis and **bharatiyabytes/sentimentWithSarcasm** for sarcasm detection. By combining these datasets, the model is better equipped to differentiate between sarcastic and genuine sentiment expressions, improving sentiment analysis accuracy in contexts where sarcasm is prevalent.
25
 
26
- - **Developed by:** [Your Name/Organization]
27
- - **Funded by [optional]:** [Your Funder, if applicable]
28
- - **Shared by [optional]:** [Your Organization, if applicable]
29
  - **Model type:** Text Classification
30
  - **Language(s) (NLP):** English (en)
31
  - **License:** Apache 2.0
@@ -33,7 +32,7 @@ This model leverages the **Flan-T5-small** transformer architecture, fine-tuned
33
 
34
  ### Model Sources [optional]
35
 
36
- - **Repository:** [Repository link on Hugging Face]
37
  - **Paper [optional]:** [Link to paper if available]
38
  - **Demo [optional]:** [Link to demo if available]
39
 
@@ -43,7 +42,7 @@ This model leverages the **Flan-T5-small** transformer architecture, fine-tuned
43
 
44
  This model can be used directly for sentiment classification on texts where sarcasm may obscure intended sentiment. It provides accurate classifications by considering nuanced expressions, making it ideal for social media analysis, customer feedback processing, and sarcasm-rich content sources.
45
 
46
- ### Downstream Use [optional]
47
 
48
  This model can serve as a base for further fine-tuning for domains that require sarcasm-aware sentiment analysis, such as customer service, public relations, and social media monitoring applications.
49
 
@@ -73,7 +72,7 @@ model = AutoModelForSequenceClassification.from_pretrained("your-username/sarcas
73
  inputs = tokenizer("Your input text here", return_tensors="pt")
74
  outputs = model(**inputs)
75
  predictions = torch.argmax(outputs.logits, dim=-1)
76
-
77
  ## Training Details
78
 
79
  ### Training Data
@@ -82,7 +81,7 @@ The model was fine-tuned on a combination of IMDB (a general sentiment analysis
82
 
83
  ### Training Procedure
84
 
85
- #### Preprocessing [optional]
86
 
87
  Standard text preprocessing methods were applied, such as tokenization and lowercase transformation.
88
 
@@ -94,7 +93,7 @@ Epochs: 3
94
  Batch Size: 16
95
  Learning Rate: 3e-5
96
 
97
- #### Speeds, Sizes, Times [optional]
98
 
99
  Training took approximately 4 hours on an NVIDIA V100 GPU with a model size of 60M parameters.
100
 
@@ -125,7 +124,7 @@ The model achieved an accuracy of 88% on the sentiment classification task and a
125
 
126
  The model shows strong performance in sarcasm-sensitive sentiment analysis, making it suitable for applications where nuanced sentiment interpretation is crucial.
127
 
128
- ## Model Examination [optional]
129
 
130
  The model’s predictions have been examined to ensure that sarcastic content is accurately labeled, using interpretability tools such as SHAP to visualize model attention on sarcastic phrases.
131
 
@@ -141,7 +140,7 @@ Carbon emissions can be estimated using the [Machine Learning Impact calculator]
141
  - **Compute Region:** USA
142
  - **Carbon Emitted:** NA
143
 
144
- ## Technical Specifications [optional]
145
 
146
  ### Model Architecture and Objective
147
 
 
23
 
24
  This model leverages the **Flan-T5-small** transformer architecture, fine-tuned on datasets including **IMDB** for sentiment analysis and **bharatiyabytes/sentimentWithSarcasm** for sarcasm detection. By combining these datasets, the model is better equipped to differentiate between sarcastic and genuine sentiment expressions, improving sentiment analysis accuracy in contexts where sarcasm is prevalent.
25
 
26
+ - **Developed by:** bharatiyabytes
27
+ - **Funded by [optional]:** Not yet funded
 
28
  - **Model type:** Text Classification
29
  - **Language(s) (NLP):** English (en)
30
  - **License:** Apache 2.0
 
32
 
33
  ### Model Sources [optional]
34
 
35
+ - **Repository:** https://github.com/sohi-g/lets-talk-the-hype.git
36
  - **Paper [optional]:** [Link to paper if available]
37
  - **Demo [optional]:** [Link to demo if available]
38
 
 
42
 
43
  This model can be used directly for sentiment classification on texts where sarcasm may obscure intended sentiment. It provides accurate classifications by considering nuanced expressions, making it ideal for social media analysis, customer feedback processing, and sarcasm-rich content sources.
44
 
45
+ ### Downstream Use
46
 
47
  This model can serve as a base for further fine-tuning for domains that require sarcasm-aware sentiment analysis, such as customer service, public relations, and social media monitoring applications.
48
 
 
72
  inputs = tokenizer("Your input text here", return_tensors="pt")
73
  outputs = model(**inputs)
74
  predictions = torch.argmax(outputs.logits, dim=-1)
75
+ ```
76
  ## Training Details
77
 
78
  ### Training Data
 
81
 
82
  ### Training Procedure
83
 
84
+ #### Preprocessing
85
 
86
  Standard text preprocessing methods were applied, such as tokenization and lowercase transformation.
87
 
 
93
  Batch Size: 16
94
  Learning Rate: 3e-5
95
 
96
+ #### Speeds, Sizes, Times
97
 
98
  Training took approximately 4 hours on an NVIDIA V100 GPU with a model size of 60M parameters.
99
 
 
124
 
125
  The model shows strong performance in sarcasm-sensitive sentiment analysis, making it suitable for applications where nuanced sentiment interpretation is crucial.
126
 
127
+ ## Model Examination
128
 
129
  The model’s predictions have been examined to ensure that sarcastic content is accurately labeled, using interpretability tools such as SHAP to visualize model attention on sarcastic phrases.
130
 
 
140
  - **Compute Region:** USA
141
  - **Carbon Emitted:** NA
142
 
143
+ ## Technical Specifications
144
 
145
  ### Model Architecture and Objective
146