Jonathancasjar commited on
Commit
04afbe1
1 Parent(s): 635b566

Training in progress, step 500

Browse files
.gitignore ADDED
@@ -0,0 +1 @@
 
 
1
+ checkpoint-*/
config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "distilroberta-base",
3
+ "architectures": [
4
+ "RobertaForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 768,
13
+ "id2label": {
14
+ "0": "not_equivalent",
15
+ "1": "equivalent"
16
+ },
17
+ "initializer_range": 0.02,
18
+ "intermediate_size": 3072,
19
+ "label2id": {
20
+ "equivalent": "1",
21
+ "not_equivalent": "0"
22
+ },
23
+ "layer_norm_eps": 1e-05,
24
+ "max_position_embeddings": 514,
25
+ "model_type": "roberta",
26
+ "num_attention_heads": 12,
27
+ "num_hidden_layers": 6,
28
+ "pad_token_id": 1,
29
+ "position_embedding_type": "absolute",
30
+ "problem_type": "single_label_classification",
31
+ "torch_dtype": "float32",
32
+ "transformers_version": "4.28.1",
33
+ "type_vocab_size": 1,
34
+ "use_cache": true,
35
+ "vocab_size": 50265
36
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9edf223993b81033d211904df681731835e28365a17cdc4154925b035c72e75c
3
+ size 328519925
runs/Apr21_13-06-29_cfa6e2d14f0d/1682082908.8732462/events.out.tfevents.1682082908.cfa6e2d14f0d.237.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c0b271f49d8041fc497ad59799c1c8ce5060edad6dbaf5e9d8bef454e30b454
3
+ size 6048
runs/Apr21_13-06-29_cfa6e2d14f0d/1682083077.3103273/events.out.tfevents.1682083077.cfa6e2d14f0d.237.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:131f399bd8089fc2cc84de44ebcbafe358eacf600a0fc4664ff09e74f717df17
3
+ size 6048
runs/Apr21_13-06-29_cfa6e2d14f0d/events.out.tfevents.1682082908.cfa6e2d14f0d.237.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f466d91598eb89919fc7cb72d4caf3a1f8dc64f0f7839d139332007ebdfce09b
3
+ size 4315
runs/Apr21_13-06-29_cfa6e2d14f0d/events.out.tfevents.1682083077.cfa6e2d14f0d.237.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8aed2187de9ccece7838cf46350037d7ae311dccfda42cab5846393a5e12b75e
3
+ size 4315
runs/Apr21_13-19-53_cfa6e2d14f0d/1682083211.5044575/events.out.tfevents.1682083211.cfa6e2d14f0d.237.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d07b9eaad75f83dd12ecffb02fc9daed53522f12060510a8a2ae1236a657adea
3
+ size 6048
runs/Apr21_13-19-53_cfa6e2d14f0d/events.out.tfevents.1682083211.cfa6e2d14f0d.237.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01169556f5f6399c2417b8f75a38c6c0b0575c88bebcbc85d86b25be5fcb4c34
3
+ size 4315
runs/Apr21_13-23-22_cfa6e2d14f0d/1682083434.0500302/events.out.tfevents.1682083434.cfa6e2d14f0d.237.7 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e409b00bbc6ec439fedba624230b2e1d683c580dc8eb17a8dac4264f37536c9d
3
+ size 6048
runs/Apr21_13-23-22_cfa6e2d14f0d/events.out.tfevents.1682083434.cfa6e2d14f0d.237.6 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0fb9014621ba914e416f7e8aeb84dcbf90d03370a2ff8cdf1d61f6b6015bfd56
3
+ size 4839
special_tokens_map.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<s>",
3
+ "cls_token": "<s>",
4
+ "eos_token": "</s>",
5
+ "mask_token": {
6
+ "content": "<mask>",
7
+ "lstrip": true,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "pad_token": "<pad>",
13
+ "sep_token": "</s>",
14
+ "unk_token": "<unk>"
15
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "bos_token": "<s>",
4
+ "clean_up_tokenization_spaces": true,
5
+ "cls_token": "<s>",
6
+ "eos_token": "</s>",
7
+ "errors": "replace",
8
+ "mask_token": "<mask>",
9
+ "model_max_length": 512,
10
+ "pad_token": "<pad>",
11
+ "sep_token": "</s>",
12
+ "tokenizer_class": "RobertaTokenizer",
13
+ "trim_offsets": true,
14
+ "unk_token": "<unk>"
15
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:11f18aac17d5a9d870a0218b2deeb79ddba91dbe170a58751b031ef8c9a51ffc
3
+ size 3707
vocab.json ADDED
The diff for this file is too large to render. See raw diff