Rasooli commited on
Commit
de7baa0
·
verified ·
1 Parent(s): 578aa64

Training in progress, epoch 1

Browse files
config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
- "_name_or_path": "pedramyazdipoor/persian_xlm_roberta_large",
3
  "architectures": [
4
- "XLMRobertaForSequenceClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
  "bos_token_id": 0,
@@ -10,27 +10,24 @@
10
  "gradient_checkpointing": false,
11
  "hidden_act": "gelu",
12
  "hidden_dropout_prob": 0.1,
13
- "hidden_size": 1024,
14
  "id2label": {
15
  "0": "Negative",
16
  "1": "Positive",
17
  "2": "Neutral"
18
  },
19
  "initializer_range": 0.02,
20
- "intermediate_size": 4096,
21
  "label2id": {
22
- "LABEL_0": 0,
23
- "LABEL_1": 1,
24
- "LABEL_2": 2
25
  },
26
- "language": "english",
27
- "layer_norm_eps": 1e-05,
28
  "max_position_embeddings": 514,
29
- "model_type": "xlm-roberta",
30
- "name": "XLMRoberta",
31
- "num_attention_heads": 16,
32
- "num_hidden_layers": 24,
33
- "output_past": true,
34
  "pad_token_id": 1,
35
  "position_embedding_type": "absolute",
36
  "problem_type": "single_label_classification",
@@ -38,5 +35,5 @@
38
  "transformers_version": "4.44.2",
39
  "type_vocab_size": 1,
40
  "use_cache": true,
41
- "vocab_size": 250002
42
  }
 
1
  {
2
+ "_name_or_path": "dadashzadeh/roberta-sentiment-persian",
3
  "architectures": [
4
+ "RobertaForSequenceClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
  "bos_token_id": 0,
 
10
  "gradient_checkpointing": false,
11
  "hidden_act": "gelu",
12
  "hidden_dropout_prob": 0.1,
13
+ "hidden_size": 768,
14
  "id2label": {
15
  "0": "Negative",
16
  "1": "Positive",
17
  "2": "Neutral"
18
  },
19
  "initializer_range": 0.02,
20
+ "intermediate_size": 3072,
21
  "label2id": {
22
+ "negative": 0,
23
+ "neutral": 2,
24
+ "positive": 1
25
  },
26
+ "layer_norm_eps": 1e-12,
 
27
  "max_position_embeddings": 514,
28
+ "model_type": "roberta",
29
+ "num_attention_heads": 12,
30
+ "num_hidden_layers": 12,
 
 
31
  "pad_token_id": 1,
32
  "position_embedding_type": "absolute",
33
  "problem_type": "single_label_classification",
 
35
  "transformers_version": "4.44.2",
36
  "type_vocab_size": 1,
37
  "use_cache": true,
38
+ "vocab_size": 42000
39
  }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6a6f504c89e40e2d97442bd92bd347079ac08915f380ecba88fbb51e5e0742df
3
- size 2239622772
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:102cb07e7094dd45e22920983ad13f88d3533756792d7d703e3dcd075330b134
3
+ size 473225820
runs/Sep15_04-12-29_1ea3fea1e80a/events.out.tfevents.1726373560.1ea3fea1e80a.1157.2 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:56865d14741d7df039d9f8e5841c50732ccc908fbe9a4dae17a1b68a3c18c1e2
3
- size 6875
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be4d4e3c3411997ddb6368e5a875e225ae2541ddb29f6b19b23938bfc32ae499
3
+ size 7666
runs/Sep15_04-37-54_1ea3fea1e80a/events.out.tfevents.1726375086.1ea3fea1e80a.8430.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21dc52df591f98d84e8a1e53673e79f0328b34792b727a74c1c55d56983d75ef
3
+ size 5217
runs/Sep15_04-39-05_1ea3fea1e80a/events.out.tfevents.1726375154.1ea3fea1e80a.8430.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35a6190c8a349602ee1b4aaa77a15ef7093992b8bbdabb740380924386bd2565
3
+ size 5215
runs/Sep15_04-39-28_1ea3fea1e80a/events.out.tfevents.1726375177.1ea3fea1e80a.8430.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a4aa6ac739040a655c5ab729f953e2b83f784389eb9e8391b06186f95030851
3
+ size 5795
special_tokens_map.json CHANGED
@@ -2,27 +2,27 @@
2
  "bos_token": {
3
  "content": "<s>",
4
  "lstrip": false,
5
- "normalized": false,
6
  "rstrip": false,
7
  "single_word": false
8
  },
9
  "cls_token": {
10
  "content": "<s>",
11
  "lstrip": false,
12
- "normalized": false,
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
  "eos_token": {
17
  "content": "</s>",
18
  "lstrip": false,
19
- "normalized": false,
20
  "rstrip": false,
21
  "single_word": false
22
  },
23
  "mask_token": {
24
  "content": "<mask>",
25
- "lstrip": false,
26
  "normalized": false,
27
  "rstrip": false,
28
  "single_word": false
@@ -30,21 +30,21 @@
30
  "pad_token": {
31
  "content": "<pad>",
32
  "lstrip": false,
33
- "normalized": false,
34
  "rstrip": false,
35
  "single_word": false
36
  },
37
  "sep_token": {
38
  "content": "</s>",
39
  "lstrip": false,
40
- "normalized": false,
41
  "rstrip": false,
42
  "single_word": false
43
  },
44
  "unk_token": {
45
  "content": "<unk>",
46
  "lstrip": false,
47
- "normalized": false,
48
  "rstrip": false,
49
  "single_word": false
50
  }
 
2
  "bos_token": {
3
  "content": "<s>",
4
  "lstrip": false,
5
+ "normalized": true,
6
  "rstrip": false,
7
  "single_word": false
8
  },
9
  "cls_token": {
10
  "content": "<s>",
11
  "lstrip": false,
12
+ "normalized": true,
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
  "eos_token": {
17
  "content": "</s>",
18
  "lstrip": false,
19
+ "normalized": true,
20
  "rstrip": false,
21
  "single_word": false
22
  },
23
  "mask_token": {
24
  "content": "<mask>",
25
+ "lstrip": true,
26
  "normalized": false,
27
  "rstrip": false,
28
  "single_word": false
 
30
  "pad_token": {
31
  "content": "<pad>",
32
  "lstrip": false,
33
+ "normalized": true,
34
  "rstrip": false,
35
  "single_word": false
36
  },
37
  "sep_token": {
38
  "content": "</s>",
39
  "lstrip": false,
40
+ "normalized": true,
41
  "rstrip": false,
42
  "single_word": false
43
  },
44
  "unk_token": {
45
  "content": "<unk>",
46
  "lstrip": false,
47
+ "normalized": true,
48
  "rstrip": false,
49
  "single_word": false
50
  }
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6247be075973430c70e5980bd5af624466882ca896e8994078f6707661e890ca
3
- size 17082898
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad6ff9d8d4dfc5111a9e2e3bf6cd868a97c6e54c84fab43ddba87ef05c541607
3
+ size 2749245
tokenizer_config.json CHANGED
@@ -1,9 +1,10 @@
1
  {
 
2
  "added_tokens_decoder": {
3
  "0": {
4
  "content": "<s>",
5
  "lstrip": false,
6
- "normalized": false,
7
  "rstrip": false,
8
  "single_word": false,
9
  "special": true
@@ -11,7 +12,7 @@
11
  "1": {
12
  "content": "<pad>",
13
  "lstrip": false,
14
- "normalized": false,
15
  "rstrip": false,
16
  "single_word": false,
17
  "special": true
@@ -19,7 +20,7 @@
19
  "2": {
20
  "content": "</s>",
21
  "lstrip": false,
22
- "normalized": false,
23
  "rstrip": false,
24
  "single_word": false,
25
  "special": true
@@ -27,38 +28,82 @@
27
  "3": {
28
  "content": "<unk>",
29
  "lstrip": false,
30
- "normalized": false,
31
  "rstrip": false,
32
  "single_word": false,
33
  "special": true
34
  },
35
- "250001": {
36
  "content": "<mask>",
37
- "lstrip": false,
38
  "normalized": false,
39
  "rstrip": false,
40
  "single_word": false,
41
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
42
  }
43
  },
44
  "bos_token": "<s>",
45
  "clean_up_tokenization_spaces": true,
46
  "cls_token": "<s>",
47
- "do_lower_case": false,
48
  "eos_token": "</s>",
49
- "full_tokenizer_file": null,
50
  "mask_token": "<mask>",
51
- "max_length": 32,
52
  "model_max_length": 512,
53
- "pad_to_multiple_of": null,
54
  "pad_token": "<pad>",
55
- "pad_token_type_id": 0,
56
- "padding_side": "right",
57
  "sep_token": "</s>",
58
- "sp_model_kwargs": {},
59
  "stride": 0,
60
- "tokenizer_class": "XLMRobertaTokenizer",
 
61
  "truncation_side": "right",
62
- "truncation_strategy": "only_first",
63
  "unk_token": "<unk>"
64
  }
 
1
  {
2
+ "add_prefix_space": false,
3
  "added_tokens_decoder": {
4
  "0": {
5
  "content": "<s>",
6
  "lstrip": false,
7
+ "normalized": true,
8
  "rstrip": false,
9
  "single_word": false,
10
  "special": true
 
12
  "1": {
13
  "content": "<pad>",
14
  "lstrip": false,
15
+ "normalized": true,
16
  "rstrip": false,
17
  "single_word": false,
18
  "special": true
 
20
  "2": {
21
  "content": "</s>",
22
  "lstrip": false,
23
+ "normalized": true,
24
  "rstrip": false,
25
  "single_word": false,
26
  "special": true
 
28
  "3": {
29
  "content": "<unk>",
30
  "lstrip": false,
31
+ "normalized": true,
32
  "rstrip": false,
33
  "single_word": false,
34
  "special": true
35
  },
36
+ "4": {
37
  "content": "<mask>",
38
+ "lstrip": true,
39
  "normalized": false,
40
  "rstrip": false,
41
  "single_word": false,
42
  "special": true
43
+ },
44
+ "5": {
45
+ "content": "<|endoftext|>",
46
+ "lstrip": false,
47
+ "normalized": true,
48
+ "rstrip": false,
49
+ "single_word": false,
50
+ "special": true
51
+ },
52
+ "6": {
53
+ "content": "<|startoftext|>",
54
+ "lstrip": false,
55
+ "normalized": true,
56
+ "rstrip": false,
57
+ "single_word": false,
58
+ "special": true
59
+ },
60
+ "7": {
61
+ "content": "<nl>",
62
+ "lstrip": false,
63
+ "normalized": true,
64
+ "rstrip": false,
65
+ "single_word": false,
66
+ "special": true
67
+ },
68
+ "8": {
69
+ "content": "<hs>",
70
+ "lstrip": false,
71
+ "normalized": true,
72
+ "rstrip": false,
73
+ "single_word": false,
74
+ "special": true
75
+ },
76
+ "9": {
77
+ "content": "<sep>",
78
+ "lstrip": false,
79
+ "normalized": true,
80
+ "rstrip": false,
81
+ "single_word": false,
82
+ "special": true
83
+ },
84
+ "10": {
85
+ "content": "<cls>",
86
+ "lstrip": false,
87
+ "normalized": true,
88
+ "rstrip": false,
89
+ "single_word": false,
90
+ "special": true
91
  }
92
  },
93
  "bos_token": "<s>",
94
  "clean_up_tokenization_spaces": true,
95
  "cls_token": "<s>",
 
96
  "eos_token": "</s>",
97
+ "errors": "replace",
98
  "mask_token": "<mask>",
99
+ "max_length": 512,
100
  "model_max_length": 512,
 
101
  "pad_token": "<pad>",
 
 
102
  "sep_token": "</s>",
 
103
  "stride": 0,
104
+ "tokenizer_class": "RobertaTokenizer",
105
+ "trim_offsets": true,
106
  "truncation_side": "right",
107
+ "truncation_strategy": "longest_first",
108
  "unk_token": "<unk>"
109
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aa8658cf6fac6cb7a55a3eed540a2b8ce79ee7b429656c925aae4aa317e284cd
3
  size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b684b44b2219b673134d95e5c6eef13f43a43426f0e413160c5be100398b6b93
3
  size 5176
vocab.json ADDED
The diff for this file is too large to render. See raw diff