Mikimi commited on
Commit
06da107
·
verified ·
1 Parent(s): 78fb18a

bert-base-case-financial-news-twitter-sentiment

Browse files
Files changed (4) hide show
  1. README.md +5 -18
  2. config.json +13 -14
  3. model.safetensors +2 -2
  4. training_args.bin +1 -1
README.md CHANGED
@@ -1,12 +1,9 @@
1
  ---
2
  library_name: transformers
3
- license: mit
4
- base_model: roberta-large
5
  tags:
6
  - generated_from_trainer
7
- metrics:
8
- - accuracy
9
- - f1
10
  model-index:
11
  - name: twitter_trainer
12
  results: []
@@ -17,13 +14,7 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  # twitter_trainer
19
 
20
- This model is a fine-tuned version of [roberta-large](https://huggingface.co/roberta-large) on an unknown dataset.
21
- It achieves the following results on the evaluation set:
22
- - Loss: 0.6659
23
- - Accuracy: 72.8
24
- - P: 94.8793
25
- - R: 109.0833
26
- - F1: 101.4867
27
 
28
  ## Model description
29
 
@@ -51,17 +42,13 @@ The following hyperparameters were used during training:
51
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
52
  - lr_scheduler_type: linear
53
  - lr_scheduler_warmup_ratio: 0.1
54
- - num_epochs: 5
55
 
56
  ### Training results
57
 
58
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | P | R | F1 |
59
  |:-------------:|:-----:|:----:|:---------------:|:--------:|:-------:|:--------:|:--------:|
60
- | No log | 1.0 | 63 | 0.9032 | 64.3 | 89.9301 | 138.4284 | 109.0292 |
61
- | No log | 2.0 | 126 | 0.9033 | 64.3 | 89.9301 | 138.4284 | 109.0292 |
62
- | No log | 3.0 | 189 | 0.7329 | 71.5 | 91.8044 | 118.0691 | 103.2933 |
63
- | No log | 4.0 | 252 | 0.6866 | 72.5 | 95.0663 | 109.1293 | 101.6135 |
64
- | No log | 4.928 | 310 | 0.6659 | 72.8 | 94.8793 | 109.0833 | 101.4867 |
65
 
66
 
67
  ### Framework versions
 
1
  ---
2
  library_name: transformers
3
+ license: apache-2.0
4
+ base_model: bert-base-cased
5
  tags:
6
  - generated_from_trainer
 
 
 
7
  model-index:
8
  - name: twitter_trainer
9
  results: []
 
14
 
15
  # twitter_trainer
16
 
17
+ This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on an unknown dataset.
 
 
 
 
 
 
18
 
19
  ## Model description
20
 
 
42
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
43
  - lr_scheduler_type: linear
44
  - lr_scheduler_warmup_ratio: 0.1
45
+ - num_epochs: 1
46
 
47
  ### Training results
48
 
49
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | P | R | F1 |
50
  |:-------------:|:-----:|:----:|:---------------:|:--------:|:-------:|:--------:|:--------:|
51
+ | No log | 0.992 | 62 | 0.9867 | 66.5 | 90.1379 | 132.1537 | 107.1751 |
 
 
 
 
52
 
53
 
54
  ### Framework versions
config.json CHANGED
@@ -1,15 +1,14 @@
1
  {
2
- "_name_or_path": "roberta-large",
3
  "architectures": [
4
- "RobertaForSequenceClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
- "bos_token_id": 0,
8
  "classifier_dropout": null,
9
- "eos_token_id": 2,
10
  "hidden_act": "gelu",
11
  "hidden_dropout_prob": 0.1,
12
- "hidden_size": 1024,
13
  "id2label": {
14
  "0": "LABEL_0",
15
  "1": "LABEL_1",
@@ -18,7 +17,7 @@
18
  "4": "LABEL_4"
19
  },
20
  "initializer_range": 0.02,
21
- "intermediate_size": 4096,
22
  "label2id": {
23
  "LABEL_0": 0,
24
  "LABEL_1": 1,
@@ -26,17 +25,17 @@
26
  "LABEL_3": 3,
27
  "LABEL_4": 4
28
  },
29
- "layer_norm_eps": 1e-05,
30
- "max_position_embeddings": 514,
31
- "model_type": "roberta",
32
- "num_attention_heads": 16,
33
- "num_hidden_layers": 24,
34
- "pad_token_id": 1,
35
  "position_embedding_type": "absolute",
36
  "problem_type": "single_label_classification",
37
  "torch_dtype": "float32",
38
  "transformers_version": "4.47.1",
39
- "type_vocab_size": 1,
40
  "use_cache": true,
41
- "vocab_size": 50265
42
  }
 
1
  {
2
+ "_name_or_path": "bert-base-cased",
3
  "architectures": [
4
+ "BertForSequenceClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
 
7
  "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
  "hidden_act": "gelu",
10
  "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
  "id2label": {
13
  "0": "LABEL_0",
14
  "1": "LABEL_1",
 
17
  "4": "LABEL_4"
18
  },
19
  "initializer_range": 0.02,
20
+ "intermediate_size": 3072,
21
  "label2id": {
22
  "LABEL_0": 0,
23
  "LABEL_1": 1,
 
25
  "LABEL_3": 3,
26
  "LABEL_4": 4
27
  },
28
+ "layer_norm_eps": 1e-12,
29
+ "max_position_embeddings": 512,
30
+ "model_type": "bert",
31
+ "num_attention_heads": 12,
32
+ "num_hidden_layers": 12,
33
+ "pad_token_id": 0,
34
  "position_embedding_type": "absolute",
35
  "problem_type": "single_label_classification",
36
  "torch_dtype": "float32",
37
  "transformers_version": "4.47.1",
38
+ "type_vocab_size": 2,
39
  "use_cache": true,
40
+ "vocab_size": 28996
41
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6cc25227e6ed9fe79363dc88de349c0b44118587a9a2662839288e4774408872
3
- size 1421507716
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3da953ba079b14638c11b1aa95e4a1bba20744c7cc75dc13e2a09a8997c24d70
3
+ size 433279996
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ed71a0cf746d4ebd96eceb81361ed9b920afe1ae8eea8e2a79032bf2ce50aaf3
3
  size 5304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44962e8da4e7626306f98cdf9639ba3704847d2af1070260ff650b9f6504ad1e
3
  size 5304