Upload folder using huggingface_hub
Browse files- all_results.json +15 -0
- checkpoint-147264/config.json +39 -0
- checkpoint-147264/model.safetensors +3 -0
- checkpoint-147264/optimizer.pt +3 -0
- checkpoint-147264/rng_state.pth +3 -0
- checkpoint-147264/scheduler.pt +3 -0
- checkpoint-147264/special_tokens_map.json +30 -0
- checkpoint-147264/tokenizer.json +0 -0
- checkpoint-147264/tokenizer_config.json +67 -0
- checkpoint-147264/trainer_state.json +2127 -0
- checkpoint-147264/training_args.bin +3 -0
- config.json +39 -0
- eval_results.json +9 -0
- model.safetensors +3 -0
- predictions.txt +4909 -0
- special_tokens_map.json +30 -0
- tokenizer.json +0 -0
- tokenizer_config.json +67 -0
- train_results.json +9 -0
- trainer_state.json +3526 -0
- training_args.bin +3 -0
all_results.json
ADDED
@@ -0,0 +1,15 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"epoch": 5.0,
|
3 |
+
"eval_accuracy": 0.7337000966072083,
|
4 |
+
"eval_loss": 0.7322831153869629,
|
5 |
+
"eval_runtime": 7.5979,
|
6 |
+
"eval_samples": 4908,
|
7 |
+
"eval_samples_per_second": 645.964,
|
8 |
+
"eval_steps_per_second": 80.811,
|
9 |
+
"total_flos": 1.158449687808768e+17,
|
10 |
+
"train_loss": 0.5704954994082295,
|
11 |
+
"train_runtime": 14864.7666,
|
12 |
+
"train_samples": 392702,
|
13 |
+
"train_samples_per_second": 132.092,
|
14 |
+
"train_steps_per_second": 16.512
|
15 |
+
}
|
checkpoint-147264/config.json
ADDED
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "/home/ubuntu/utah/babylm-24/data/training/models/10M_babylm_ascii/SPM-Unigram_6144/DebertaV2-Base-10M_babylm-A",
|
3 |
+
"architectures": [
|
4 |
+
"DebertaV2ForSequenceClassification"
|
5 |
+
],
|
6 |
+
"attention_probs_dropout_prob": 0.1,
|
7 |
+
"hidden_act": "gelu",
|
8 |
+
"hidden_dropout_prob": 0.1,
|
9 |
+
"hidden_size": 768,
|
10 |
+
"id2label": {
|
11 |
+
"0": "LABEL_0",
|
12 |
+
"1": "LABEL_1",
|
13 |
+
"2": "LABEL_2"
|
14 |
+
},
|
15 |
+
"initializer_range": 0.02,
|
16 |
+
"intermediate_size": 1536,
|
17 |
+
"label2id": {
|
18 |
+
"0": 0,
|
19 |
+
"1": 1,
|
20 |
+
"2": 2
|
21 |
+
},
|
22 |
+
"layer_norm_eps": 1e-07,
|
23 |
+
"max_position_embeddings": 512,
|
24 |
+
"max_relative_positions": -1,
|
25 |
+
"model_type": "deberta-v2",
|
26 |
+
"num_attention_heads": 12,
|
27 |
+
"num_hidden_layers": 8,
|
28 |
+
"pad_token_id": 3,
|
29 |
+
"pooler_dropout": 0,
|
30 |
+
"pooler_hidden_act": "gelu",
|
31 |
+
"pooler_hidden_size": 768,
|
32 |
+
"pos_att_type": null,
|
33 |
+
"position_biased_input": true,
|
34 |
+
"relative_attention": false,
|
35 |
+
"torch_dtype": "float32",
|
36 |
+
"transformers_version": "4.44.2",
|
37 |
+
"type_vocab_size": 0,
|
38 |
+
"vocab_size": 6144
|
39 |
+
}
|
checkpoint-147264/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0575c4bc1a8943e345d617f7dcaac754a2941686ef65c5c0917ce3a4386ce1c5
|
3 |
+
size 174106588
|
checkpoint-147264/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2732f716a4ba778c841c21ad78f7e62a51d5b5ed6c5b31bbe101ff7ebd6db7c7
|
3 |
+
size 348294394
|
checkpoint-147264/rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:baa2cd28a67a69f42614108e9c5f40ed606f7281f0163c3b228eec0d153ec79b
|
3 |
+
size 14244
|
checkpoint-147264/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ffc1e9a9fb64f92c3fd07960e7506191a3e5ccfad21213c0c6044309bac08509
|
3 |
+
size 1064
|
checkpoint-147264/special_tokens_map.json
ADDED
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"cls_token": {
|
3 |
+
"content": "[CLS]",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": false,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"mask_token": {
|
10 |
+
"content": "[MASK]",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": false,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"pad_token": {
|
17 |
+
"content": "[PAD]",
|
18 |
+
"lstrip": false,
|
19 |
+
"normalized": false,
|
20 |
+
"rstrip": false,
|
21 |
+
"single_word": false
|
22 |
+
},
|
23 |
+
"sep_token": {
|
24 |
+
"content": "[SEP]",
|
25 |
+
"lstrip": false,
|
26 |
+
"normalized": false,
|
27 |
+
"rstrip": false,
|
28 |
+
"single_word": false
|
29 |
+
}
|
30 |
+
}
|
checkpoint-147264/tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-147264/tokenizer_config.json
ADDED
@@ -0,0 +1,67 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"added_tokens_decoder": {
|
3 |
+
"0": {
|
4 |
+
"content": "[UNK]",
|
5 |
+
"lstrip": false,
|
6 |
+
"normalized": false,
|
7 |
+
"rstrip": false,
|
8 |
+
"single_word": false,
|
9 |
+
"special": true
|
10 |
+
},
|
11 |
+
"1": {
|
12 |
+
"content": "[CLS]",
|
13 |
+
"lstrip": false,
|
14 |
+
"normalized": false,
|
15 |
+
"rstrip": false,
|
16 |
+
"single_word": false,
|
17 |
+
"special": true
|
18 |
+
},
|
19 |
+
"2": {
|
20 |
+
"content": "[SEP]",
|
21 |
+
"lstrip": false,
|
22 |
+
"normalized": false,
|
23 |
+
"rstrip": false,
|
24 |
+
"single_word": false,
|
25 |
+
"special": true
|
26 |
+
},
|
27 |
+
"3": {
|
28 |
+
"content": "[PAD]",
|
29 |
+
"lstrip": false,
|
30 |
+
"normalized": false,
|
31 |
+
"rstrip": false,
|
32 |
+
"single_word": false,
|
33 |
+
"special": true
|
34 |
+
},
|
35 |
+
"4": {
|
36 |
+
"content": "[MASK]",
|
37 |
+
"lstrip": false,
|
38 |
+
"normalized": false,
|
39 |
+
"rstrip": false,
|
40 |
+
"single_word": false,
|
41 |
+
"special": true
|
42 |
+
},
|
43 |
+
"5": {
|
44 |
+
"content": "[PAR]",
|
45 |
+
"lstrip": false,
|
46 |
+
"normalized": false,
|
47 |
+
"rstrip": false,
|
48 |
+
"single_word": false,
|
49 |
+
"special": true
|
50 |
+
},
|
51 |
+
"6": {
|
52 |
+
"content": "[TAB]",
|
53 |
+
"lstrip": false,
|
54 |
+
"normalized": false,
|
55 |
+
"rstrip": false,
|
56 |
+
"single_word": false,
|
57 |
+
"special": true
|
58 |
+
}
|
59 |
+
},
|
60 |
+
"clean_up_tokenization_spaces": false,
|
61 |
+
"cls_token": "[CLS]",
|
62 |
+
"mask_token": "[MASK]",
|
63 |
+
"model_max_length": 1000000000000000019884624838656,
|
64 |
+
"pad_token": "[PAD]",
|
65 |
+
"sep_token": "[SEP]",
|
66 |
+
"tokenizer_class": "PreTrainedTokenizerFast"
|
67 |
+
}
|
checkpoint-147264/trainer_state.json
ADDED
@@ -0,0 +1,2127 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"best_metric": 0.7337000966072083,
|
3 |
+
"best_model_checkpoint": "/home/ubuntu/utah/babylm-24/src/evaluation/results/finetune/DebertaV2-Base-10M_babylm-A/mnli/checkpoint-147264",
|
4 |
+
"epoch": 3.0,
|
5 |
+
"eval_steps": 500,
|
6 |
+
"global_step": 147264,
|
7 |
+
"is_hyper_param_search": false,
|
8 |
+
"is_local_process_zero": true,
|
9 |
+
"is_world_process_zero": true,
|
10 |
+
"log_history": [
|
11 |
+
{
|
12 |
+
"epoch": 0.010185788787483703,
|
13 |
+
"grad_norm": 3.972818613052368,
|
14 |
+
"learning_rate": 2.9938885267275102e-05,
|
15 |
+
"loss": 1.1058,
|
16 |
+
"step": 500
|
17 |
+
},
|
18 |
+
{
|
19 |
+
"epoch": 0.020371577574967405,
|
20 |
+
"grad_norm": 3.13022780418396,
|
21 |
+
"learning_rate": 2.9877770534550196e-05,
|
22 |
+
"loss": 1.0388,
|
23 |
+
"step": 1000
|
24 |
+
},
|
25 |
+
{
|
26 |
+
"epoch": 0.03055736636245111,
|
27 |
+
"grad_norm": 5.194879055023193,
|
28 |
+
"learning_rate": 2.9816655801825297e-05,
|
29 |
+
"loss": 0.999,
|
30 |
+
"step": 1500
|
31 |
+
},
|
32 |
+
{
|
33 |
+
"epoch": 0.04074315514993481,
|
34 |
+
"grad_norm": 6.858974933624268,
|
35 |
+
"learning_rate": 2.975554106910039e-05,
|
36 |
+
"loss": 0.9637,
|
37 |
+
"step": 2000
|
38 |
+
},
|
39 |
+
{
|
40 |
+
"epoch": 0.050928943937418515,
|
41 |
+
"grad_norm": 7.074331760406494,
|
42 |
+
"learning_rate": 2.9694426336375492e-05,
|
43 |
+
"loss": 0.9697,
|
44 |
+
"step": 2500
|
45 |
+
},
|
46 |
+
{
|
47 |
+
"epoch": 0.06111473272490222,
|
48 |
+
"grad_norm": 5.386518478393555,
|
49 |
+
"learning_rate": 2.9633311603650586e-05,
|
50 |
+
"loss": 0.9472,
|
51 |
+
"step": 3000
|
52 |
+
},
|
53 |
+
{
|
54 |
+
"epoch": 0.07130052151238592,
|
55 |
+
"grad_norm": 5.582085609436035,
|
56 |
+
"learning_rate": 2.9572196870925687e-05,
|
57 |
+
"loss": 0.9273,
|
58 |
+
"step": 3500
|
59 |
+
},
|
60 |
+
{
|
61 |
+
"epoch": 0.08148631029986962,
|
62 |
+
"grad_norm": 10.966410636901855,
|
63 |
+
"learning_rate": 2.951108213820078e-05,
|
64 |
+
"loss": 0.9316,
|
65 |
+
"step": 4000
|
66 |
+
},
|
67 |
+
{
|
68 |
+
"epoch": 0.09167209908735333,
|
69 |
+
"grad_norm": 7.840672492980957,
|
70 |
+
"learning_rate": 2.9449967405475882e-05,
|
71 |
+
"loss": 0.9237,
|
72 |
+
"step": 4500
|
73 |
+
},
|
74 |
+
{
|
75 |
+
"epoch": 0.10185788787483703,
|
76 |
+
"grad_norm": 7.442521095275879,
|
77 |
+
"learning_rate": 2.9388852672750976e-05,
|
78 |
+
"loss": 0.918,
|
79 |
+
"step": 5000
|
80 |
+
},
|
81 |
+
{
|
82 |
+
"epoch": 0.11204367666232073,
|
83 |
+
"grad_norm": 8.379261016845703,
|
84 |
+
"learning_rate": 2.9327737940026077e-05,
|
85 |
+
"loss": 0.8798,
|
86 |
+
"step": 5500
|
87 |
+
},
|
88 |
+
{
|
89 |
+
"epoch": 0.12222946544980444,
|
90 |
+
"grad_norm": 6.880769729614258,
|
91 |
+
"learning_rate": 2.926662320730117e-05,
|
92 |
+
"loss": 0.9045,
|
93 |
+
"step": 6000
|
94 |
+
},
|
95 |
+
{
|
96 |
+
"epoch": 0.13241525423728814,
|
97 |
+
"grad_norm": 5.965817451477051,
|
98 |
+
"learning_rate": 2.9205508474576272e-05,
|
99 |
+
"loss": 0.8941,
|
100 |
+
"step": 6500
|
101 |
+
},
|
102 |
+
{
|
103 |
+
"epoch": 0.14260104302477183,
|
104 |
+
"grad_norm": 9.67165756225586,
|
105 |
+
"learning_rate": 2.914439374185137e-05,
|
106 |
+
"loss": 0.878,
|
107 |
+
"step": 7000
|
108 |
+
},
|
109 |
+
{
|
110 |
+
"epoch": 0.15278683181225555,
|
111 |
+
"grad_norm": 6.4934892654418945,
|
112 |
+
"learning_rate": 2.9083279009126467e-05,
|
113 |
+
"loss": 0.8723,
|
114 |
+
"step": 7500
|
115 |
+
},
|
116 |
+
{
|
117 |
+
"epoch": 0.16297262059973924,
|
118 |
+
"grad_norm": 11.987072944641113,
|
119 |
+
"learning_rate": 2.9022164276401565e-05,
|
120 |
+
"loss": 0.8913,
|
121 |
+
"step": 8000
|
122 |
+
},
|
123 |
+
{
|
124 |
+
"epoch": 0.17315840938722293,
|
125 |
+
"grad_norm": 9.218280792236328,
|
126 |
+
"learning_rate": 2.8961049543676662e-05,
|
127 |
+
"loss": 0.868,
|
128 |
+
"step": 8500
|
129 |
+
},
|
130 |
+
{
|
131 |
+
"epoch": 0.18334419817470665,
|
132 |
+
"grad_norm": 6.697582244873047,
|
133 |
+
"learning_rate": 2.8899934810951763e-05,
|
134 |
+
"loss": 0.8618,
|
135 |
+
"step": 9000
|
136 |
+
},
|
137 |
+
{
|
138 |
+
"epoch": 0.19352998696219034,
|
139 |
+
"grad_norm": 8.035455703735352,
|
140 |
+
"learning_rate": 2.8838820078226858e-05,
|
141 |
+
"loss": 0.8641,
|
142 |
+
"step": 9500
|
143 |
+
},
|
144 |
+
{
|
145 |
+
"epoch": 0.20371577574967406,
|
146 |
+
"grad_norm": 6.890414237976074,
|
147 |
+
"learning_rate": 2.877770534550196e-05,
|
148 |
+
"loss": 0.8574,
|
149 |
+
"step": 10000
|
150 |
+
},
|
151 |
+
{
|
152 |
+
"epoch": 0.21390156453715775,
|
153 |
+
"grad_norm": 9.001678466796875,
|
154 |
+
"learning_rate": 2.8716590612777053e-05,
|
155 |
+
"loss": 0.8547,
|
156 |
+
"step": 10500
|
157 |
+
},
|
158 |
+
{
|
159 |
+
"epoch": 0.22408735332464147,
|
160 |
+
"grad_norm": 7.271191120147705,
|
161 |
+
"learning_rate": 2.8655475880052154e-05,
|
162 |
+
"loss": 0.8332,
|
163 |
+
"step": 11000
|
164 |
+
},
|
165 |
+
{
|
166 |
+
"epoch": 0.23427314211212516,
|
167 |
+
"grad_norm": 9.849822998046875,
|
168 |
+
"learning_rate": 2.8594361147327248e-05,
|
169 |
+
"loss": 0.8447,
|
170 |
+
"step": 11500
|
171 |
+
},
|
172 |
+
{
|
173 |
+
"epoch": 0.24445893089960888,
|
174 |
+
"grad_norm": 15.07829761505127,
|
175 |
+
"learning_rate": 2.853324641460235e-05,
|
176 |
+
"loss": 0.8251,
|
177 |
+
"step": 12000
|
178 |
+
},
|
179 |
+
{
|
180 |
+
"epoch": 0.25464471968709257,
|
181 |
+
"grad_norm": 6.886602878570557,
|
182 |
+
"learning_rate": 2.8472131681877446e-05,
|
183 |
+
"loss": 0.8444,
|
184 |
+
"step": 12500
|
185 |
+
},
|
186 |
+
{
|
187 |
+
"epoch": 0.2648305084745763,
|
188 |
+
"grad_norm": 9.94825267791748,
|
189 |
+
"learning_rate": 2.8411016949152544e-05,
|
190 |
+
"loss": 0.8504,
|
191 |
+
"step": 13000
|
192 |
+
},
|
193 |
+
{
|
194 |
+
"epoch": 0.27501629726205995,
|
195 |
+
"grad_norm": 8.0740966796875,
|
196 |
+
"learning_rate": 2.834990221642764e-05,
|
197 |
+
"loss": 0.8182,
|
198 |
+
"step": 13500
|
199 |
+
},
|
200 |
+
{
|
201 |
+
"epoch": 0.28520208604954367,
|
202 |
+
"grad_norm": 10.770129203796387,
|
203 |
+
"learning_rate": 2.828878748370274e-05,
|
204 |
+
"loss": 0.8373,
|
205 |
+
"step": 14000
|
206 |
+
},
|
207 |
+
{
|
208 |
+
"epoch": 0.2953878748370274,
|
209 |
+
"grad_norm": 9.214078903198242,
|
210 |
+
"learning_rate": 2.8227672750977836e-05,
|
211 |
+
"loss": 0.8089,
|
212 |
+
"step": 14500
|
213 |
+
},
|
214 |
+
{
|
215 |
+
"epoch": 0.3055736636245111,
|
216 |
+
"grad_norm": 14.231884956359863,
|
217 |
+
"learning_rate": 2.8166558018252934e-05,
|
218 |
+
"loss": 0.8337,
|
219 |
+
"step": 15000
|
220 |
+
},
|
221 |
+
{
|
222 |
+
"epoch": 0.31575945241199477,
|
223 |
+
"grad_norm": 9.379185676574707,
|
224 |
+
"learning_rate": 2.810544328552803e-05,
|
225 |
+
"loss": 0.8315,
|
226 |
+
"step": 15500
|
227 |
+
},
|
228 |
+
{
|
229 |
+
"epoch": 0.3259452411994785,
|
230 |
+
"grad_norm": 8.370843887329102,
|
231 |
+
"learning_rate": 2.804432855280313e-05,
|
232 |
+
"loss": 0.808,
|
233 |
+
"step": 16000
|
234 |
+
},
|
235 |
+
{
|
236 |
+
"epoch": 0.3361310299869622,
|
237 |
+
"grad_norm": 7.201444149017334,
|
238 |
+
"learning_rate": 2.7983213820078227e-05,
|
239 |
+
"loss": 0.808,
|
240 |
+
"step": 16500
|
241 |
+
},
|
242 |
+
{
|
243 |
+
"epoch": 0.34631681877444587,
|
244 |
+
"grad_norm": 10.696725845336914,
|
245 |
+
"learning_rate": 2.7922099087353324e-05,
|
246 |
+
"loss": 0.813,
|
247 |
+
"step": 17000
|
248 |
+
},
|
249 |
+
{
|
250 |
+
"epoch": 0.3565026075619296,
|
251 |
+
"grad_norm": 6.316204071044922,
|
252 |
+
"learning_rate": 2.7860984354628422e-05,
|
253 |
+
"loss": 0.8213,
|
254 |
+
"step": 17500
|
255 |
+
},
|
256 |
+
{
|
257 |
+
"epoch": 0.3666883963494133,
|
258 |
+
"grad_norm": 14.780086517333984,
|
259 |
+
"learning_rate": 2.779986962190352e-05,
|
260 |
+
"loss": 0.8165,
|
261 |
+
"step": 18000
|
262 |
+
},
|
263 |
+
{
|
264 |
+
"epoch": 0.376874185136897,
|
265 |
+
"grad_norm": 11.34945011138916,
|
266 |
+
"learning_rate": 2.773875488917862e-05,
|
267 |
+
"loss": 0.7943,
|
268 |
+
"step": 18500
|
269 |
+
},
|
270 |
+
{
|
271 |
+
"epoch": 0.3870599739243807,
|
272 |
+
"grad_norm": 7.136452674865723,
|
273 |
+
"learning_rate": 2.7677640156453718e-05,
|
274 |
+
"loss": 0.8017,
|
275 |
+
"step": 19000
|
276 |
+
},
|
277 |
+
{
|
278 |
+
"epoch": 0.3972457627118644,
|
279 |
+
"grad_norm": 6.209742069244385,
|
280 |
+
"learning_rate": 2.7616525423728815e-05,
|
281 |
+
"loss": 0.7941,
|
282 |
+
"step": 19500
|
283 |
+
},
|
284 |
+
{
|
285 |
+
"epoch": 0.4074315514993481,
|
286 |
+
"grad_norm": 8.923270225524902,
|
287 |
+
"learning_rate": 2.7555410691003913e-05,
|
288 |
+
"loss": 0.7826,
|
289 |
+
"step": 20000
|
290 |
+
},
|
291 |
+
{
|
292 |
+
"epoch": 0.41761734028683184,
|
293 |
+
"grad_norm": 6.371203899383545,
|
294 |
+
"learning_rate": 2.749429595827901e-05,
|
295 |
+
"loss": 0.7938,
|
296 |
+
"step": 20500
|
297 |
+
},
|
298 |
+
{
|
299 |
+
"epoch": 0.4278031290743155,
|
300 |
+
"grad_norm": 8.679354667663574,
|
301 |
+
"learning_rate": 2.7433181225554108e-05,
|
302 |
+
"loss": 0.7936,
|
303 |
+
"step": 21000
|
304 |
+
},
|
305 |
+
{
|
306 |
+
"epoch": 0.4379889178617992,
|
307 |
+
"grad_norm": 5.962934970855713,
|
308 |
+
"learning_rate": 2.7372066492829205e-05,
|
309 |
+
"loss": 0.7924,
|
310 |
+
"step": 21500
|
311 |
+
},
|
312 |
+
{
|
313 |
+
"epoch": 0.44817470664928294,
|
314 |
+
"grad_norm": 11.835200309753418,
|
315 |
+
"learning_rate": 2.7310951760104303e-05,
|
316 |
+
"loss": 0.7894,
|
317 |
+
"step": 22000
|
318 |
+
},
|
319 |
+
{
|
320 |
+
"epoch": 0.4583604954367666,
|
321 |
+
"grad_norm": 9.62806510925293,
|
322 |
+
"learning_rate": 2.72498370273794e-05,
|
323 |
+
"loss": 0.7963,
|
324 |
+
"step": 22500
|
325 |
+
},
|
326 |
+
{
|
327 |
+
"epoch": 0.4685462842242503,
|
328 |
+
"grad_norm": 7.6390509605407715,
|
329 |
+
"learning_rate": 2.7188722294654498e-05,
|
330 |
+
"loss": 0.783,
|
331 |
+
"step": 23000
|
332 |
+
},
|
333 |
+
{
|
334 |
+
"epoch": 0.47873207301173404,
|
335 |
+
"grad_norm": 5.826275825500488,
|
336 |
+
"learning_rate": 2.7127607561929596e-05,
|
337 |
+
"loss": 0.7859,
|
338 |
+
"step": 23500
|
339 |
+
},
|
340 |
+
{
|
341 |
+
"epoch": 0.48891786179921776,
|
342 |
+
"grad_norm": 5.199416637420654,
|
343 |
+
"learning_rate": 2.7066492829204693e-05,
|
344 |
+
"loss": 0.7989,
|
345 |
+
"step": 24000
|
346 |
+
},
|
347 |
+
{
|
348 |
+
"epoch": 0.4991036505867014,
|
349 |
+
"grad_norm": 12.38355541229248,
|
350 |
+
"learning_rate": 2.7005378096479794e-05,
|
351 |
+
"loss": 0.7864,
|
352 |
+
"step": 24500
|
353 |
+
},
|
354 |
+
{
|
355 |
+
"epoch": 0.5092894393741851,
|
356 |
+
"grad_norm": 7.700937271118164,
|
357 |
+
"learning_rate": 2.694426336375489e-05,
|
358 |
+
"loss": 0.7747,
|
359 |
+
"step": 25000
|
360 |
+
},
|
361 |
+
{
|
362 |
+
"epoch": 0.5194752281616688,
|
363 |
+
"grad_norm": 10.8720703125,
|
364 |
+
"learning_rate": 2.688314863102999e-05,
|
365 |
+
"loss": 0.7974,
|
366 |
+
"step": 25500
|
367 |
+
},
|
368 |
+
{
|
369 |
+
"epoch": 0.5296610169491526,
|
370 |
+
"grad_norm": 6.260867595672607,
|
371 |
+
"learning_rate": 2.6822033898305083e-05,
|
372 |
+
"loss": 0.7914,
|
373 |
+
"step": 26000
|
374 |
+
},
|
375 |
+
{
|
376 |
+
"epoch": 0.5398468057366362,
|
377 |
+
"grad_norm": 11.113471984863281,
|
378 |
+
"learning_rate": 2.6760919165580184e-05,
|
379 |
+
"loss": 0.782,
|
380 |
+
"step": 26500
|
381 |
+
},
|
382 |
+
{
|
383 |
+
"epoch": 0.5500325945241199,
|
384 |
+
"grad_norm": 7.497796058654785,
|
385 |
+
"learning_rate": 2.6699804432855282e-05,
|
386 |
+
"loss": 0.7642,
|
387 |
+
"step": 27000
|
388 |
+
},
|
389 |
+
{
|
390 |
+
"epoch": 0.5602183833116037,
|
391 |
+
"grad_norm": 6.1969170570373535,
|
392 |
+
"learning_rate": 2.663868970013038e-05,
|
393 |
+
"loss": 0.7693,
|
394 |
+
"step": 27500
|
395 |
+
},
|
396 |
+
{
|
397 |
+
"epoch": 0.5704041720990873,
|
398 |
+
"grad_norm": 9.965928077697754,
|
399 |
+
"learning_rate": 2.6577574967405477e-05,
|
400 |
+
"loss": 0.7681,
|
401 |
+
"step": 28000
|
402 |
+
},
|
403 |
+
{
|
404 |
+
"epoch": 0.5805899608865711,
|
405 |
+
"grad_norm": 5.052292346954346,
|
406 |
+
"learning_rate": 2.6516460234680575e-05,
|
407 |
+
"loss": 0.7664,
|
408 |
+
"step": 28500
|
409 |
+
},
|
410 |
+
{
|
411 |
+
"epoch": 0.5907757496740548,
|
412 |
+
"grad_norm": 6.7289276123046875,
|
413 |
+
"learning_rate": 2.6455345501955672e-05,
|
414 |
+
"loss": 0.7445,
|
415 |
+
"step": 29000
|
416 |
+
},
|
417 |
+
{
|
418 |
+
"epoch": 0.6009615384615384,
|
419 |
+
"grad_norm": 8.478078842163086,
|
420 |
+
"learning_rate": 2.639423076923077e-05,
|
421 |
+
"loss": 0.7859,
|
422 |
+
"step": 29500
|
423 |
+
},
|
424 |
+
{
|
425 |
+
"epoch": 0.6111473272490222,
|
426 |
+
"grad_norm": 6.2336106300354,
|
427 |
+
"learning_rate": 2.6333116036505867e-05,
|
428 |
+
"loss": 0.7598,
|
429 |
+
"step": 30000
|
430 |
+
},
|
431 |
+
{
|
432 |
+
"epoch": 0.6213331160365059,
|
433 |
+
"grad_norm": 11.82295036315918,
|
434 |
+
"learning_rate": 2.6272001303780965e-05,
|
435 |
+
"loss": 0.7507,
|
436 |
+
"step": 30500
|
437 |
+
},
|
438 |
+
{
|
439 |
+
"epoch": 0.6315189048239895,
|
440 |
+
"grad_norm": 8.545003890991211,
|
441 |
+
"learning_rate": 2.6210886571056066e-05,
|
442 |
+
"loss": 0.7769,
|
443 |
+
"step": 31000
|
444 |
+
},
|
445 |
+
{
|
446 |
+
"epoch": 0.6417046936114733,
|
447 |
+
"grad_norm": 9.849065780639648,
|
448 |
+
"learning_rate": 2.614977183833116e-05,
|
449 |
+
"loss": 0.7781,
|
450 |
+
"step": 31500
|
451 |
+
},
|
452 |
+
{
|
453 |
+
"epoch": 0.651890482398957,
|
454 |
+
"grad_norm": 7.641623020172119,
|
455 |
+
"learning_rate": 2.608865710560626e-05,
|
456 |
+
"loss": 0.7789,
|
457 |
+
"step": 32000
|
458 |
+
},
|
459 |
+
{
|
460 |
+
"epoch": 0.6620762711864406,
|
461 |
+
"grad_norm": 16.397785186767578,
|
462 |
+
"learning_rate": 2.6027542372881355e-05,
|
463 |
+
"loss": 0.7542,
|
464 |
+
"step": 32500
|
465 |
+
},
|
466 |
+
{
|
467 |
+
"epoch": 0.6722620599739244,
|
468 |
+
"grad_norm": 11.119671821594238,
|
469 |
+
"learning_rate": 2.5966427640156456e-05,
|
470 |
+
"loss": 0.7716,
|
471 |
+
"step": 33000
|
472 |
+
},
|
473 |
+
{
|
474 |
+
"epoch": 0.6824478487614081,
|
475 |
+
"grad_norm": 12.466675758361816,
|
476 |
+
"learning_rate": 2.590531290743155e-05,
|
477 |
+
"loss": 0.7743,
|
478 |
+
"step": 33500
|
479 |
+
},
|
480 |
+
{
|
481 |
+
"epoch": 0.6926336375488917,
|
482 |
+
"grad_norm": 8.151625633239746,
|
483 |
+
"learning_rate": 2.584419817470665e-05,
|
484 |
+
"loss": 0.7581,
|
485 |
+
"step": 34000
|
486 |
+
},
|
487 |
+
{
|
488 |
+
"epoch": 0.7028194263363755,
|
489 |
+
"grad_norm": 7.743143558502197,
|
490 |
+
"learning_rate": 2.5783083441981745e-05,
|
491 |
+
"loss": 0.7599,
|
492 |
+
"step": 34500
|
493 |
+
},
|
494 |
+
{
|
495 |
+
"epoch": 0.7130052151238592,
|
496 |
+
"grad_norm": 5.249680995941162,
|
497 |
+
"learning_rate": 2.5721968709256846e-05,
|
498 |
+
"loss": 0.7483,
|
499 |
+
"step": 35000
|
500 |
+
},
|
501 |
+
{
|
502 |
+
"epoch": 0.7231910039113429,
|
503 |
+
"grad_norm": 6.738178730010986,
|
504 |
+
"learning_rate": 2.566085397653194e-05,
|
505 |
+
"loss": 0.7781,
|
506 |
+
"step": 35500
|
507 |
+
},
|
508 |
+
{
|
509 |
+
"epoch": 0.7333767926988266,
|
510 |
+
"grad_norm": 6.7923102378845215,
|
511 |
+
"learning_rate": 2.559973924380704e-05,
|
512 |
+
"loss": 0.7464,
|
513 |
+
"step": 36000
|
514 |
+
},
|
515 |
+
{
|
516 |
+
"epoch": 0.7435625814863103,
|
517 |
+
"grad_norm": 15.575695991516113,
|
518 |
+
"learning_rate": 2.553862451108214e-05,
|
519 |
+
"loss": 0.7571,
|
520 |
+
"step": 36500
|
521 |
+
},
|
522 |
+
{
|
523 |
+
"epoch": 0.753748370273794,
|
524 |
+
"grad_norm": 13.128719329833984,
|
525 |
+
"learning_rate": 2.5477509778357236e-05,
|
526 |
+
"loss": 0.7625,
|
527 |
+
"step": 37000
|
528 |
+
},
|
529 |
+
{
|
530 |
+
"epoch": 0.7639341590612777,
|
531 |
+
"grad_norm": 4.938621997833252,
|
532 |
+
"learning_rate": 2.5416395045632337e-05,
|
533 |
+
"loss": 0.7477,
|
534 |
+
"step": 37500
|
535 |
+
},
|
536 |
+
{
|
537 |
+
"epoch": 0.7741199478487614,
|
538 |
+
"grad_norm": 8.46985912322998,
|
539 |
+
"learning_rate": 2.535528031290743e-05,
|
540 |
+
"loss": 0.7419,
|
541 |
+
"step": 38000
|
542 |
+
},
|
543 |
+
{
|
544 |
+
"epoch": 0.7843057366362451,
|
545 |
+
"grad_norm": 18.881208419799805,
|
546 |
+
"learning_rate": 2.5294165580182532e-05,
|
547 |
+
"loss": 0.7403,
|
548 |
+
"step": 38500
|
549 |
+
},
|
550 |
+
{
|
551 |
+
"epoch": 0.7944915254237288,
|
552 |
+
"grad_norm": 9.986156463623047,
|
553 |
+
"learning_rate": 2.5233050847457626e-05,
|
554 |
+
"loss": 0.7508,
|
555 |
+
"step": 39000
|
556 |
+
},
|
557 |
+
{
|
558 |
+
"epoch": 0.8046773142112125,
|
559 |
+
"grad_norm": 11.625401496887207,
|
560 |
+
"learning_rate": 2.5171936114732727e-05,
|
561 |
+
"loss": 0.7426,
|
562 |
+
"step": 39500
|
563 |
+
},
|
564 |
+
{
|
565 |
+
"epoch": 0.8148631029986962,
|
566 |
+
"grad_norm": 9.926522254943848,
|
567 |
+
"learning_rate": 2.511082138200782e-05,
|
568 |
+
"loss": 0.7365,
|
569 |
+
"step": 40000
|
570 |
+
},
|
571 |
+
{
|
572 |
+
"epoch": 0.8250488917861799,
|
573 |
+
"grad_norm": 8.104277610778809,
|
574 |
+
"learning_rate": 2.5049706649282922e-05,
|
575 |
+
"loss": 0.7532,
|
576 |
+
"step": 40500
|
577 |
+
},
|
578 |
+
{
|
579 |
+
"epoch": 0.8352346805736637,
|
580 |
+
"grad_norm": 5.898177623748779,
|
581 |
+
"learning_rate": 2.4988591916558017e-05,
|
582 |
+
"loss": 0.7657,
|
583 |
+
"step": 41000
|
584 |
+
},
|
585 |
+
{
|
586 |
+
"epoch": 0.8454204693611473,
|
587 |
+
"grad_norm": 6.856070041656494,
|
588 |
+
"learning_rate": 2.4927477183833118e-05,
|
589 |
+
"loss": 0.7438,
|
590 |
+
"step": 41500
|
591 |
+
},
|
592 |
+
{
|
593 |
+
"epoch": 0.855606258148631,
|
594 |
+
"grad_norm": 11.806628227233887,
|
595 |
+
"learning_rate": 2.4866362451108212e-05,
|
596 |
+
"loss": 0.7348,
|
597 |
+
"step": 42000
|
598 |
+
},
|
599 |
+
{
|
600 |
+
"epoch": 0.8657920469361148,
|
601 |
+
"grad_norm": 12.08478832244873,
|
602 |
+
"learning_rate": 2.4805247718383313e-05,
|
603 |
+
"loss": 0.7277,
|
604 |
+
"step": 42500
|
605 |
+
},
|
606 |
+
{
|
607 |
+
"epoch": 0.8759778357235984,
|
608 |
+
"grad_norm": 5.707648277282715,
|
609 |
+
"learning_rate": 2.474413298565841e-05,
|
610 |
+
"loss": 0.7458,
|
611 |
+
"step": 43000
|
612 |
+
},
|
613 |
+
{
|
614 |
+
"epoch": 0.8861636245110821,
|
615 |
+
"grad_norm": 5.36006498336792,
|
616 |
+
"learning_rate": 2.4683018252933508e-05,
|
617 |
+
"loss": 0.7446,
|
618 |
+
"step": 43500
|
619 |
+
},
|
620 |
+
{
|
621 |
+
"epoch": 0.8963494132985659,
|
622 |
+
"grad_norm": 6.603504180908203,
|
623 |
+
"learning_rate": 2.4621903520208605e-05,
|
624 |
+
"loss": 0.7278,
|
625 |
+
"step": 44000
|
626 |
+
},
|
627 |
+
{
|
628 |
+
"epoch": 0.9065352020860495,
|
629 |
+
"grad_norm": 7.789953231811523,
|
630 |
+
"learning_rate": 2.4560788787483703e-05,
|
631 |
+
"loss": 0.7375,
|
632 |
+
"step": 44500
|
633 |
+
},
|
634 |
+
{
|
635 |
+
"epoch": 0.9167209908735332,
|
636 |
+
"grad_norm": 11.424116134643555,
|
637 |
+
"learning_rate": 2.4499674054758804e-05,
|
638 |
+
"loss": 0.7297,
|
639 |
+
"step": 45000
|
640 |
+
},
|
641 |
+
{
|
642 |
+
"epoch": 0.926906779661017,
|
643 |
+
"grad_norm": 6.7079620361328125,
|
644 |
+
"learning_rate": 2.4438559322033898e-05,
|
645 |
+
"loss": 0.7411,
|
646 |
+
"step": 45500
|
647 |
+
},
|
648 |
+
{
|
649 |
+
"epoch": 0.9370925684485006,
|
650 |
+
"grad_norm": 11.931846618652344,
|
651 |
+
"learning_rate": 2.4377444589309e-05,
|
652 |
+
"loss": 0.749,
|
653 |
+
"step": 46000
|
654 |
+
},
|
655 |
+
{
|
656 |
+
"epoch": 0.9472783572359843,
|
657 |
+
"grad_norm": 8.87076473236084,
|
658 |
+
"learning_rate": 2.4316329856584093e-05,
|
659 |
+
"loss": 0.7312,
|
660 |
+
"step": 46500
|
661 |
+
},
|
662 |
+
{
|
663 |
+
"epoch": 0.9574641460234681,
|
664 |
+
"grad_norm": 12.787999153137207,
|
665 |
+
"learning_rate": 2.4255215123859194e-05,
|
666 |
+
"loss": 0.7244,
|
667 |
+
"step": 47000
|
668 |
+
},
|
669 |
+
{
|
670 |
+
"epoch": 0.9676499348109517,
|
671 |
+
"grad_norm": 6.917387962341309,
|
672 |
+
"learning_rate": 2.4194100391134288e-05,
|
673 |
+
"loss": 0.733,
|
674 |
+
"step": 47500
|
675 |
+
},
|
676 |
+
{
|
677 |
+
"epoch": 0.9778357235984355,
|
678 |
+
"grad_norm": 7.564793109893799,
|
679 |
+
"learning_rate": 2.413298565840939e-05,
|
680 |
+
"loss": 0.7253,
|
681 |
+
"step": 48000
|
682 |
+
},
|
683 |
+
{
|
684 |
+
"epoch": 0.9880215123859192,
|
685 |
+
"grad_norm": 7.665174961090088,
|
686 |
+
"learning_rate": 2.4071870925684483e-05,
|
687 |
+
"loss": 0.7191,
|
688 |
+
"step": 48500
|
689 |
+
},
|
690 |
+
{
|
691 |
+
"epoch": 0.9982073011734028,
|
692 |
+
"grad_norm": 9.235818862915039,
|
693 |
+
"learning_rate": 2.4010756192959584e-05,
|
694 |
+
"loss": 0.7288,
|
695 |
+
"step": 49000
|
696 |
+
},
|
697 |
+
{
|
698 |
+
"epoch": 1.0,
|
699 |
+
"eval_accuracy": 0.6841890811920166,
|
700 |
+
"eval_loss": 0.7356013059616089,
|
701 |
+
"eval_runtime": 8.0037,
|
702 |
+
"eval_samples_per_second": 613.213,
|
703 |
+
"eval_steps_per_second": 76.714,
|
704 |
+
"step": 49088
|
705 |
+
},
|
706 |
+
{
|
707 |
+
"epoch": 1.0083930899608866,
|
708 |
+
"grad_norm": 9.232441902160645,
|
709 |
+
"learning_rate": 2.3949641460234682e-05,
|
710 |
+
"loss": 0.666,
|
711 |
+
"step": 49500
|
712 |
+
},
|
713 |
+
{
|
714 |
+
"epoch": 1.0185788787483703,
|
715 |
+
"grad_norm": 5.098124980926514,
|
716 |
+
"learning_rate": 2.388852672750978e-05,
|
717 |
+
"loss": 0.6762,
|
718 |
+
"step": 50000
|
719 |
+
},
|
720 |
+
{
|
721 |
+
"epoch": 1.028764667535854,
|
722 |
+
"grad_norm": 6.507536888122559,
|
723 |
+
"learning_rate": 2.3827411994784877e-05,
|
724 |
+
"loss": 0.6509,
|
725 |
+
"step": 50500
|
726 |
+
},
|
727 |
+
{
|
728 |
+
"epoch": 1.0389504563233376,
|
729 |
+
"grad_norm": 12.516643524169922,
|
730 |
+
"learning_rate": 2.3766297262059974e-05,
|
731 |
+
"loss": 0.6569,
|
732 |
+
"step": 51000
|
733 |
+
},
|
734 |
+
{
|
735 |
+
"epoch": 1.0491362451108215,
|
736 |
+
"grad_norm": 11.644853591918945,
|
737 |
+
"learning_rate": 2.3705182529335072e-05,
|
738 |
+
"loss": 0.6498,
|
739 |
+
"step": 51500
|
740 |
+
},
|
741 |
+
{
|
742 |
+
"epoch": 1.0593220338983051,
|
743 |
+
"grad_norm": 15.303723335266113,
|
744 |
+
"learning_rate": 2.364406779661017e-05,
|
745 |
+
"loss": 0.6565,
|
746 |
+
"step": 52000
|
747 |
+
},
|
748 |
+
{
|
749 |
+
"epoch": 1.0695078226857888,
|
750 |
+
"grad_norm": 11.212811470031738,
|
751 |
+
"learning_rate": 2.3582953063885267e-05,
|
752 |
+
"loss": 0.6492,
|
753 |
+
"step": 52500
|
754 |
+
},
|
755 |
+
{
|
756 |
+
"epoch": 1.0796936114732725,
|
757 |
+
"grad_norm": 7.695071220397949,
|
758 |
+
"learning_rate": 2.3521838331160365e-05,
|
759 |
+
"loss": 0.6618,
|
760 |
+
"step": 53000
|
761 |
+
},
|
762 |
+
{
|
763 |
+
"epoch": 1.0898794002607561,
|
764 |
+
"grad_norm": 8.234328269958496,
|
765 |
+
"learning_rate": 2.3460723598435462e-05,
|
766 |
+
"loss": 0.6709,
|
767 |
+
"step": 53500
|
768 |
+
},
|
769 |
+
{
|
770 |
+
"epoch": 1.1000651890482398,
|
771 |
+
"grad_norm": 12.574545860290527,
|
772 |
+
"learning_rate": 2.339960886571056e-05,
|
773 |
+
"loss": 0.6518,
|
774 |
+
"step": 54000
|
775 |
+
},
|
776 |
+
{
|
777 |
+
"epoch": 1.1102509778357237,
|
778 |
+
"grad_norm": 5.914109706878662,
|
779 |
+
"learning_rate": 2.333849413298566e-05,
|
780 |
+
"loss": 0.6494,
|
781 |
+
"step": 54500
|
782 |
+
},
|
783 |
+
{
|
784 |
+
"epoch": 1.1204367666232073,
|
785 |
+
"grad_norm": 11.49263858795166,
|
786 |
+
"learning_rate": 2.3277379400260758e-05,
|
787 |
+
"loss": 0.6723,
|
788 |
+
"step": 55000
|
789 |
+
},
|
790 |
+
{
|
791 |
+
"epoch": 1.130622555410691,
|
792 |
+
"grad_norm": 11.228019714355469,
|
793 |
+
"learning_rate": 2.3216264667535856e-05,
|
794 |
+
"loss": 0.6328,
|
795 |
+
"step": 55500
|
796 |
+
},
|
797 |
+
{
|
798 |
+
"epoch": 1.1408083441981747,
|
799 |
+
"grad_norm": 14.478355407714844,
|
800 |
+
"learning_rate": 2.3155149934810953e-05,
|
801 |
+
"loss": 0.667,
|
802 |
+
"step": 56000
|
803 |
+
},
|
804 |
+
{
|
805 |
+
"epoch": 1.1509941329856583,
|
806 |
+
"grad_norm": 18.817468643188477,
|
807 |
+
"learning_rate": 2.309403520208605e-05,
|
808 |
+
"loss": 0.6674,
|
809 |
+
"step": 56500
|
810 |
+
},
|
811 |
+
{
|
812 |
+
"epoch": 1.161179921773142,
|
813 |
+
"grad_norm": 12.483678817749023,
|
814 |
+
"learning_rate": 2.303292046936115e-05,
|
815 |
+
"loss": 0.6677,
|
816 |
+
"step": 57000
|
817 |
+
},
|
818 |
+
{
|
819 |
+
"epoch": 1.1713657105606259,
|
820 |
+
"grad_norm": 7.133495807647705,
|
821 |
+
"learning_rate": 2.2971805736636246e-05,
|
822 |
+
"loss": 0.642,
|
823 |
+
"step": 57500
|
824 |
+
},
|
825 |
+
{
|
826 |
+
"epoch": 1.1815514993481095,
|
827 |
+
"grad_norm": 7.099282741546631,
|
828 |
+
"learning_rate": 2.2910691003911343e-05,
|
829 |
+
"loss": 0.664,
|
830 |
+
"step": 58000
|
831 |
+
},
|
832 |
+
{
|
833 |
+
"epoch": 1.1917372881355932,
|
834 |
+
"grad_norm": 9.188867568969727,
|
835 |
+
"learning_rate": 2.284957627118644e-05,
|
836 |
+
"loss": 0.6703,
|
837 |
+
"step": 58500
|
838 |
+
},
|
839 |
+
{
|
840 |
+
"epoch": 1.2019230769230769,
|
841 |
+
"grad_norm": 5.470687389373779,
|
842 |
+
"learning_rate": 2.278846153846154e-05,
|
843 |
+
"loss": 0.6575,
|
844 |
+
"step": 59000
|
845 |
+
},
|
846 |
+
{
|
847 |
+
"epoch": 1.2121088657105605,
|
848 |
+
"grad_norm": 8.013532638549805,
|
849 |
+
"learning_rate": 2.2727346805736636e-05,
|
850 |
+
"loss": 0.6465,
|
851 |
+
"step": 59500
|
852 |
+
},
|
853 |
+
{
|
854 |
+
"epoch": 1.2222946544980444,
|
855 |
+
"grad_norm": 8.846539497375488,
|
856 |
+
"learning_rate": 2.2666232073011734e-05,
|
857 |
+
"loss": 0.6682,
|
858 |
+
"step": 60000
|
859 |
+
},
|
860 |
+
{
|
861 |
+
"epoch": 1.232480443285528,
|
862 |
+
"grad_norm": 11.2225341796875,
|
863 |
+
"learning_rate": 2.260511734028683e-05,
|
864 |
+
"loss": 0.653,
|
865 |
+
"step": 60500
|
866 |
+
},
|
867 |
+
{
|
868 |
+
"epoch": 1.2426662320730117,
|
869 |
+
"grad_norm": 12.363032341003418,
|
870 |
+
"learning_rate": 2.254400260756193e-05,
|
871 |
+
"loss": 0.6807,
|
872 |
+
"step": 61000
|
873 |
+
},
|
874 |
+
{
|
875 |
+
"epoch": 1.2528520208604954,
|
876 |
+
"grad_norm": 8.01014232635498,
|
877 |
+
"learning_rate": 2.248288787483703e-05,
|
878 |
+
"loss": 0.6625,
|
879 |
+
"step": 61500
|
880 |
+
},
|
881 |
+
{
|
882 |
+
"epoch": 1.263037809647979,
|
883 |
+
"grad_norm": 20.960845947265625,
|
884 |
+
"learning_rate": 2.2421773142112124e-05,
|
885 |
+
"loss": 0.6524,
|
886 |
+
"step": 62000
|
887 |
+
},
|
888 |
+
{
|
889 |
+
"epoch": 1.2732235984354627,
|
890 |
+
"grad_norm": 13.620718002319336,
|
891 |
+
"learning_rate": 2.2360658409387225e-05,
|
892 |
+
"loss": 0.6625,
|
893 |
+
"step": 62500
|
894 |
+
},
|
895 |
+
{
|
896 |
+
"epoch": 1.2834093872229466,
|
897 |
+
"grad_norm": 18.55965232849121,
|
898 |
+
"learning_rate": 2.2299543676662322e-05,
|
899 |
+
"loss": 0.6434,
|
900 |
+
"step": 63000
|
901 |
+
},
|
902 |
+
{
|
903 |
+
"epoch": 1.2935951760104303,
|
904 |
+
"grad_norm": 12.919173240661621,
|
905 |
+
"learning_rate": 2.223842894393742e-05,
|
906 |
+
"loss": 0.652,
|
907 |
+
"step": 63500
|
908 |
+
},
|
909 |
+
{
|
910 |
+
"epoch": 1.303780964797914,
|
911 |
+
"grad_norm": 19.561386108398438,
|
912 |
+
"learning_rate": 2.2177314211212517e-05,
|
913 |
+
"loss": 0.6498,
|
914 |
+
"step": 64000
|
915 |
+
},
|
916 |
+
{
|
917 |
+
"epoch": 1.3139667535853976,
|
918 |
+
"grad_norm": 10.529995918273926,
|
919 |
+
"learning_rate": 2.2116199478487615e-05,
|
920 |
+
"loss": 0.666,
|
921 |
+
"step": 64500
|
922 |
+
},
|
923 |
+
{
|
924 |
+
"epoch": 1.3241525423728815,
|
925 |
+
"grad_norm": 13.268050193786621,
|
926 |
+
"learning_rate": 2.2055084745762713e-05,
|
927 |
+
"loss": 0.6499,
|
928 |
+
"step": 65000
|
929 |
+
},
|
930 |
+
{
|
931 |
+
"epoch": 1.3343383311603652,
|
932 |
+
"grad_norm": 6.3072285652160645,
|
933 |
+
"learning_rate": 2.199397001303781e-05,
|
934 |
+
"loss": 0.6798,
|
935 |
+
"step": 65500
|
936 |
+
},
|
937 |
+
{
|
938 |
+
"epoch": 1.3445241199478488,
|
939 |
+
"grad_norm": 5.304599761962891,
|
940 |
+
"learning_rate": 2.1932855280312908e-05,
|
941 |
+
"loss": 0.6559,
|
942 |
+
"step": 66000
|
943 |
+
},
|
944 |
+
{
|
945 |
+
"epoch": 1.3547099087353325,
|
946 |
+
"grad_norm": 13.698792457580566,
|
947 |
+
"learning_rate": 2.1871740547588005e-05,
|
948 |
+
"loss": 0.6441,
|
949 |
+
"step": 66500
|
950 |
+
},
|
951 |
+
{
|
952 |
+
"epoch": 1.3648956975228161,
|
953 |
+
"grad_norm": 8.951632499694824,
|
954 |
+
"learning_rate": 2.1810625814863103e-05,
|
955 |
+
"loss": 0.6564,
|
956 |
+
"step": 67000
|
957 |
+
},
|
958 |
+
{
|
959 |
+
"epoch": 1.3750814863102998,
|
960 |
+
"grad_norm": 5.84067440032959,
|
961 |
+
"learning_rate": 2.17495110821382e-05,
|
962 |
+
"loss": 0.6686,
|
963 |
+
"step": 67500
|
964 |
+
},
|
965 |
+
{
|
966 |
+
"epoch": 1.3852672750977835,
|
967 |
+
"grad_norm": 6.442080497741699,
|
968 |
+
"learning_rate": 2.16883963494133e-05,
|
969 |
+
"loss": 0.6517,
|
970 |
+
"step": 68000
|
971 |
+
},
|
972 |
+
{
|
973 |
+
"epoch": 1.3954530638852674,
|
974 |
+
"grad_norm": 5.514212608337402,
|
975 |
+
"learning_rate": 2.1627281616688395e-05,
|
976 |
+
"loss": 0.6664,
|
977 |
+
"step": 68500
|
978 |
+
},
|
979 |
+
{
|
980 |
+
"epoch": 1.405638852672751,
|
981 |
+
"grad_norm": 14.150158882141113,
|
982 |
+
"learning_rate": 2.1566166883963496e-05,
|
983 |
+
"loss": 0.6561,
|
984 |
+
"step": 69000
|
985 |
+
},
|
986 |
+
{
|
987 |
+
"epoch": 1.4158246414602347,
|
988 |
+
"grad_norm": 17.196884155273438,
|
989 |
+
"learning_rate": 2.150505215123859e-05,
|
990 |
+
"loss": 0.6513,
|
991 |
+
"step": 69500
|
992 |
+
},
|
993 |
+
{
|
994 |
+
"epoch": 1.4260104302477183,
|
995 |
+
"grad_norm": 6.181870937347412,
|
996 |
+
"learning_rate": 2.144393741851369e-05,
|
997 |
+
"loss": 0.65,
|
998 |
+
"step": 70000
|
999 |
+
},
|
1000 |
+
{
|
1001 |
+
"epoch": 1.436196219035202,
|
1002 |
+
"grad_norm": 14.865707397460938,
|
1003 |
+
"learning_rate": 2.1382822685788786e-05,
|
1004 |
+
"loss": 0.6596,
|
1005 |
+
"step": 70500
|
1006 |
+
},
|
1007 |
+
{
|
1008 |
+
"epoch": 1.4463820078226859,
|
1009 |
+
"grad_norm": 15.808574676513672,
|
1010 |
+
"learning_rate": 2.1321707953063886e-05,
|
1011 |
+
"loss": 0.6559,
|
1012 |
+
"step": 71000
|
1013 |
+
},
|
1014 |
+
{
|
1015 |
+
"epoch": 1.4565677966101696,
|
1016 |
+
"grad_norm": 16.76003074645996,
|
1017 |
+
"learning_rate": 2.1260593220338984e-05,
|
1018 |
+
"loss": 0.6467,
|
1019 |
+
"step": 71500
|
1020 |
+
},
|
1021 |
+
{
|
1022 |
+
"epoch": 1.4667535853976532,
|
1023 |
+
"grad_norm": 21.466825485229492,
|
1024 |
+
"learning_rate": 2.119947848761408e-05,
|
1025 |
+
"loss": 0.6579,
|
1026 |
+
"step": 72000
|
1027 |
+
},
|
1028 |
+
{
|
1029 |
+
"epoch": 1.4769393741851369,
|
1030 |
+
"grad_norm": 18.81052589416504,
|
1031 |
+
"learning_rate": 2.113836375488918e-05,
|
1032 |
+
"loss": 0.6699,
|
1033 |
+
"step": 72500
|
1034 |
+
},
|
1035 |
+
{
|
1036 |
+
"epoch": 1.4871251629726205,
|
1037 |
+
"grad_norm": 10.334957122802734,
|
1038 |
+
"learning_rate": 2.1077249022164277e-05,
|
1039 |
+
"loss": 0.655,
|
1040 |
+
"step": 73000
|
1041 |
+
},
|
1042 |
+
{
|
1043 |
+
"epoch": 1.4973109517601042,
|
1044 |
+
"grad_norm": 14.836852073669434,
|
1045 |
+
"learning_rate": 2.1016134289439378e-05,
|
1046 |
+
"loss": 0.66,
|
1047 |
+
"step": 73500
|
1048 |
+
},
|
1049 |
+
{
|
1050 |
+
"epoch": 1.5074967405475879,
|
1051 |
+
"grad_norm": 21.439233779907227,
|
1052 |
+
"learning_rate": 2.0955019556714472e-05,
|
1053 |
+
"loss": 0.6486,
|
1054 |
+
"step": 74000
|
1055 |
+
},
|
1056 |
+
{
|
1057 |
+
"epoch": 1.5176825293350718,
|
1058 |
+
"grad_norm": 21.09993553161621,
|
1059 |
+
"learning_rate": 2.0893904823989573e-05,
|
1060 |
+
"loss": 0.6432,
|
1061 |
+
"step": 74500
|
1062 |
+
},
|
1063 |
+
{
|
1064 |
+
"epoch": 1.5278683181225554,
|
1065 |
+
"grad_norm": 8.926955223083496,
|
1066 |
+
"learning_rate": 2.0832790091264667e-05,
|
1067 |
+
"loss": 0.6418,
|
1068 |
+
"step": 75000
|
1069 |
+
},
|
1070 |
+
{
|
1071 |
+
"epoch": 1.538054106910039,
|
1072 |
+
"grad_norm": 3.551163911819458,
|
1073 |
+
"learning_rate": 2.0771675358539768e-05,
|
1074 |
+
"loss": 0.6418,
|
1075 |
+
"step": 75500
|
1076 |
+
},
|
1077 |
+
{
|
1078 |
+
"epoch": 1.548239895697523,
|
1079 |
+
"grad_norm": 10.46932315826416,
|
1080 |
+
"learning_rate": 2.0710560625814862e-05,
|
1081 |
+
"loss": 0.657,
|
1082 |
+
"step": 76000
|
1083 |
+
},
|
1084 |
+
{
|
1085 |
+
"epoch": 1.5584256844850066,
|
1086 |
+
"grad_norm": 10.022995948791504,
|
1087 |
+
"learning_rate": 2.0649445893089963e-05,
|
1088 |
+
"loss": 0.6564,
|
1089 |
+
"step": 76500
|
1090 |
+
},
|
1091 |
+
{
|
1092 |
+
"epoch": 1.5686114732724903,
|
1093 |
+
"grad_norm": 15.299884796142578,
|
1094 |
+
"learning_rate": 2.0588331160365057e-05,
|
1095 |
+
"loss": 0.6535,
|
1096 |
+
"step": 77000
|
1097 |
+
},
|
1098 |
+
{
|
1099 |
+
"epoch": 1.578797262059974,
|
1100 |
+
"grad_norm": 11.149444580078125,
|
1101 |
+
"learning_rate": 2.0527216427640158e-05,
|
1102 |
+
"loss": 0.6409,
|
1103 |
+
"step": 77500
|
1104 |
+
},
|
1105 |
+
{
|
1106 |
+
"epoch": 1.5889830508474576,
|
1107 |
+
"grad_norm": 4.011179447174072,
|
1108 |
+
"learning_rate": 2.0466101694915252e-05,
|
1109 |
+
"loss": 0.6299,
|
1110 |
+
"step": 78000
|
1111 |
+
},
|
1112 |
+
{
|
1113 |
+
"epoch": 1.5991688396349413,
|
1114 |
+
"grad_norm": 13.744772911071777,
|
1115 |
+
"learning_rate": 2.0404986962190353e-05,
|
1116 |
+
"loss": 0.6511,
|
1117 |
+
"step": 78500
|
1118 |
+
},
|
1119 |
+
{
|
1120 |
+
"epoch": 1.609354628422425,
|
1121 |
+
"grad_norm": 11.129698753356934,
|
1122 |
+
"learning_rate": 2.0343872229465447e-05,
|
1123 |
+
"loss": 0.6349,
|
1124 |
+
"step": 79000
|
1125 |
+
},
|
1126 |
+
{
|
1127 |
+
"epoch": 1.6195404172099086,
|
1128 |
+
"grad_norm": 16.331953048706055,
|
1129 |
+
"learning_rate": 2.0282757496740548e-05,
|
1130 |
+
"loss": 0.6517,
|
1131 |
+
"step": 79500
|
1132 |
+
},
|
1133 |
+
{
|
1134 |
+
"epoch": 1.6297262059973925,
|
1135 |
+
"grad_norm": 12.816121101379395,
|
1136 |
+
"learning_rate": 2.0221642764015646e-05,
|
1137 |
+
"loss": 0.6409,
|
1138 |
+
"step": 80000
|
1139 |
+
},
|
1140 |
+
{
|
1141 |
+
"epoch": 1.6399119947848761,
|
1142 |
+
"grad_norm": 13.881726264953613,
|
1143 |
+
"learning_rate": 2.0160528031290743e-05,
|
1144 |
+
"loss": 0.6606,
|
1145 |
+
"step": 80500
|
1146 |
+
},
|
1147 |
+
{
|
1148 |
+
"epoch": 1.6500977835723598,
|
1149 |
+
"grad_norm": 14.215683937072754,
|
1150 |
+
"learning_rate": 2.0099413298565844e-05,
|
1151 |
+
"loss": 0.657,
|
1152 |
+
"step": 81000
|
1153 |
+
},
|
1154 |
+
{
|
1155 |
+
"epoch": 1.6602835723598437,
|
1156 |
+
"grad_norm": 19.557342529296875,
|
1157 |
+
"learning_rate": 2.003829856584094e-05,
|
1158 |
+
"loss": 0.6451,
|
1159 |
+
"step": 81500
|
1160 |
+
},
|
1161 |
+
{
|
1162 |
+
"epoch": 1.6704693611473274,
|
1163 |
+
"grad_norm": 17.83816909790039,
|
1164 |
+
"learning_rate": 1.997718383311604e-05,
|
1165 |
+
"loss": 0.6397,
|
1166 |
+
"step": 82000
|
1167 |
+
},
|
1168 |
+
{
|
1169 |
+
"epoch": 1.680655149934811,
|
1170 |
+
"grad_norm": 8.789198875427246,
|
1171 |
+
"learning_rate": 1.9916069100391133e-05,
|
1172 |
+
"loss": 0.6514,
|
1173 |
+
"step": 82500
|
1174 |
+
},
|
1175 |
+
{
|
1176 |
+
"epoch": 1.6908409387222947,
|
1177 |
+
"grad_norm": 11.361543655395508,
|
1178 |
+
"learning_rate": 1.9854954367666234e-05,
|
1179 |
+
"loss": 0.6466,
|
1180 |
+
"step": 83000
|
1181 |
+
},
|
1182 |
+
{
|
1183 |
+
"epoch": 1.7010267275097783,
|
1184 |
+
"grad_norm": 15.630517959594727,
|
1185 |
+
"learning_rate": 1.979383963494133e-05,
|
1186 |
+
"loss": 0.6316,
|
1187 |
+
"step": 83500
|
1188 |
+
},
|
1189 |
+
{
|
1190 |
+
"epoch": 1.711212516297262,
|
1191 |
+
"grad_norm": 10.55432415008545,
|
1192 |
+
"learning_rate": 1.973272490221643e-05,
|
1193 |
+
"loss": 0.6533,
|
1194 |
+
"step": 84000
|
1195 |
+
},
|
1196 |
+
{
|
1197 |
+
"epoch": 1.7213983050847457,
|
1198 |
+
"grad_norm": 8.428609848022461,
|
1199 |
+
"learning_rate": 1.9671610169491524e-05,
|
1200 |
+
"loss": 0.6478,
|
1201 |
+
"step": 84500
|
1202 |
+
},
|
1203 |
+
{
|
1204 |
+
"epoch": 1.7315840938722293,
|
1205 |
+
"grad_norm": 4.304576873779297,
|
1206 |
+
"learning_rate": 1.9610495436766625e-05,
|
1207 |
+
"loss": 0.6449,
|
1208 |
+
"step": 85000
|
1209 |
+
},
|
1210 |
+
{
|
1211 |
+
"epoch": 1.7417698826597132,
|
1212 |
+
"grad_norm": 21.92937469482422,
|
1213 |
+
"learning_rate": 1.9549380704041722e-05,
|
1214 |
+
"loss": 0.6425,
|
1215 |
+
"step": 85500
|
1216 |
+
},
|
1217 |
+
{
|
1218 |
+
"epoch": 1.7519556714471969,
|
1219 |
+
"grad_norm": 19.813392639160156,
|
1220 |
+
"learning_rate": 1.948826597131682e-05,
|
1221 |
+
"loss": 0.6238,
|
1222 |
+
"step": 86000
|
1223 |
+
},
|
1224 |
+
{
|
1225 |
+
"epoch": 1.7621414602346805,
|
1226 |
+
"grad_norm": 6.831646919250488,
|
1227 |
+
"learning_rate": 1.9427151238591917e-05,
|
1228 |
+
"loss": 0.6553,
|
1229 |
+
"step": 86500
|
1230 |
+
},
|
1231 |
+
{
|
1232 |
+
"epoch": 1.7723272490221644,
|
1233 |
+
"grad_norm": 11.058158874511719,
|
1234 |
+
"learning_rate": 1.9366036505867015e-05,
|
1235 |
+
"loss": 0.6528,
|
1236 |
+
"step": 87000
|
1237 |
+
},
|
1238 |
+
{
|
1239 |
+
"epoch": 1.782513037809648,
|
1240 |
+
"grad_norm": 7.01440954208374,
|
1241 |
+
"learning_rate": 1.9304921773142112e-05,
|
1242 |
+
"loss": 0.6506,
|
1243 |
+
"step": 87500
|
1244 |
+
},
|
1245 |
+
{
|
1246 |
+
"epoch": 1.7926988265971318,
|
1247 |
+
"grad_norm": 4.963765621185303,
|
1248 |
+
"learning_rate": 1.924380704041721e-05,
|
1249 |
+
"loss": 0.6412,
|
1250 |
+
"step": 88000
|
1251 |
+
},
|
1252 |
+
{
|
1253 |
+
"epoch": 1.8028846153846154,
|
1254 |
+
"grad_norm": 38.63767623901367,
|
1255 |
+
"learning_rate": 1.9182692307692307e-05,
|
1256 |
+
"loss": 0.6629,
|
1257 |
+
"step": 88500
|
1258 |
+
},
|
1259 |
+
{
|
1260 |
+
"epoch": 1.813070404172099,
|
1261 |
+
"grad_norm": 11.423843383789062,
|
1262 |
+
"learning_rate": 1.9121577574967405e-05,
|
1263 |
+
"loss": 0.673,
|
1264 |
+
"step": 89000
|
1265 |
+
},
|
1266 |
+
{
|
1267 |
+
"epoch": 1.8232561929595827,
|
1268 |
+
"grad_norm": 11.73025894165039,
|
1269 |
+
"learning_rate": 1.9060462842242506e-05,
|
1270 |
+
"loss": 0.625,
|
1271 |
+
"step": 89500
|
1272 |
+
},
|
1273 |
+
{
|
1274 |
+
"epoch": 1.8334419817470664,
|
1275 |
+
"grad_norm": 11.493837356567383,
|
1276 |
+
"learning_rate": 1.89993481095176e-05,
|
1277 |
+
"loss": 0.6658,
|
1278 |
+
"step": 90000
|
1279 |
+
},
|
1280 |
+
{
|
1281 |
+
"epoch": 1.84362777053455,
|
1282 |
+
"grad_norm": 12.196702003479004,
|
1283 |
+
"learning_rate": 1.89382333767927e-05,
|
1284 |
+
"loss": 0.6451,
|
1285 |
+
"step": 90500
|
1286 |
+
},
|
1287 |
+
{
|
1288 |
+
"epoch": 1.8538135593220337,
|
1289 |
+
"grad_norm": 9.690689086914062,
|
1290 |
+
"learning_rate": 1.8877118644067795e-05,
|
1291 |
+
"loss": 0.6282,
|
1292 |
+
"step": 91000
|
1293 |
+
},
|
1294 |
+
{
|
1295 |
+
"epoch": 1.8639993481095176,
|
1296 |
+
"grad_norm": 13.679101943969727,
|
1297 |
+
"learning_rate": 1.8816003911342896e-05,
|
1298 |
+
"loss": 0.6328,
|
1299 |
+
"step": 91500
|
1300 |
+
},
|
1301 |
+
{
|
1302 |
+
"epoch": 1.8741851368970013,
|
1303 |
+
"grad_norm": 14.264899253845215,
|
1304 |
+
"learning_rate": 1.8754889178617994e-05,
|
1305 |
+
"loss": 0.6355,
|
1306 |
+
"step": 92000
|
1307 |
+
},
|
1308 |
+
{
|
1309 |
+
"epoch": 1.8843709256844852,
|
1310 |
+
"grad_norm": 9.039682388305664,
|
1311 |
+
"learning_rate": 1.869377444589309e-05,
|
1312 |
+
"loss": 0.6471,
|
1313 |
+
"step": 92500
|
1314 |
+
},
|
1315 |
+
{
|
1316 |
+
"epoch": 1.8945567144719688,
|
1317 |
+
"grad_norm": 9.861899375915527,
|
1318 |
+
"learning_rate": 1.863265971316819e-05,
|
1319 |
+
"loss": 0.6549,
|
1320 |
+
"step": 93000
|
1321 |
+
},
|
1322 |
+
{
|
1323 |
+
"epoch": 1.9047425032594525,
|
1324 |
+
"grad_norm": 14.948457717895508,
|
1325 |
+
"learning_rate": 1.8571544980443286e-05,
|
1326 |
+
"loss": 0.6436,
|
1327 |
+
"step": 93500
|
1328 |
+
},
|
1329 |
+
{
|
1330 |
+
"epoch": 1.9149282920469362,
|
1331 |
+
"grad_norm": 15.613487243652344,
|
1332 |
+
"learning_rate": 1.8510430247718384e-05,
|
1333 |
+
"loss": 0.6355,
|
1334 |
+
"step": 94000
|
1335 |
+
},
|
1336 |
+
{
|
1337 |
+
"epoch": 1.9251140808344198,
|
1338 |
+
"grad_norm": 14.105792045593262,
|
1339 |
+
"learning_rate": 1.844931551499348e-05,
|
1340 |
+
"loss": 0.6345,
|
1341 |
+
"step": 94500
|
1342 |
+
},
|
1343 |
+
{
|
1344 |
+
"epoch": 1.9352998696219035,
|
1345 |
+
"grad_norm": 7.004807472229004,
|
1346 |
+
"learning_rate": 1.838820078226858e-05,
|
1347 |
+
"loss": 0.6456,
|
1348 |
+
"step": 95000
|
1349 |
+
},
|
1350 |
+
{
|
1351 |
+
"epoch": 1.9454856584093871,
|
1352 |
+
"grad_norm": 10.506879806518555,
|
1353 |
+
"learning_rate": 1.8327086049543677e-05,
|
1354 |
+
"loss": 0.6303,
|
1355 |
+
"step": 95500
|
1356 |
+
},
|
1357 |
+
{
|
1358 |
+
"epoch": 1.9556714471968708,
|
1359 |
+
"grad_norm": 11.595491409301758,
|
1360 |
+
"learning_rate": 1.8265971316818774e-05,
|
1361 |
+
"loss": 0.6509,
|
1362 |
+
"step": 96000
|
1363 |
+
},
|
1364 |
+
{
|
1365 |
+
"epoch": 1.9658572359843545,
|
1366 |
+
"grad_norm": 11.29542064666748,
|
1367 |
+
"learning_rate": 1.820485658409387e-05,
|
1368 |
+
"loss": 0.6273,
|
1369 |
+
"step": 96500
|
1370 |
+
},
|
1371 |
+
{
|
1372 |
+
"epoch": 1.9760430247718384,
|
1373 |
+
"grad_norm": 4.132607460021973,
|
1374 |
+
"learning_rate": 1.814374185136897e-05,
|
1375 |
+
"loss": 0.643,
|
1376 |
+
"step": 97000
|
1377 |
+
},
|
1378 |
+
{
|
1379 |
+
"epoch": 1.986228813559322,
|
1380 |
+
"grad_norm": 9.745738983154297,
|
1381 |
+
"learning_rate": 1.8082627118644067e-05,
|
1382 |
+
"loss": 0.6656,
|
1383 |
+
"step": 97500
|
1384 |
+
},
|
1385 |
+
{
|
1386 |
+
"epoch": 1.996414602346806,
|
1387 |
+
"grad_norm": 4.289452075958252,
|
1388 |
+
"learning_rate": 1.8021512385919164e-05,
|
1389 |
+
"loss": 0.6367,
|
1390 |
+
"step": 98000
|
1391 |
+
},
|
1392 |
+
{
|
1393 |
+
"epoch": 2.0,
|
1394 |
+
"eval_accuracy": 0.7218826413154602,
|
1395 |
+
"eval_loss": 0.6872764825820923,
|
1396 |
+
"eval_runtime": 7.6175,
|
1397 |
+
"eval_samples_per_second": 644.307,
|
1398 |
+
"eval_steps_per_second": 80.604,
|
1399 |
+
"step": 98176
|
1400 |
+
},
|
1401 |
+
{
|
1402 |
+
"epoch": 2.0066003911342896,
|
1403 |
+
"grad_norm": 19.260536193847656,
|
1404 |
+
"learning_rate": 1.7960397653194265e-05,
|
1405 |
+
"loss": 0.5568,
|
1406 |
+
"step": 98500
|
1407 |
+
},
|
1408 |
+
{
|
1409 |
+
"epoch": 2.0167861799217732,
|
1410 |
+
"grad_norm": 8.347018241882324,
|
1411 |
+
"learning_rate": 1.7899282920469363e-05,
|
1412 |
+
"loss": 0.5344,
|
1413 |
+
"step": 99000
|
1414 |
+
},
|
1415 |
+
{
|
1416 |
+
"epoch": 2.026971968709257,
|
1417 |
+
"grad_norm": 12.238682746887207,
|
1418 |
+
"learning_rate": 1.783816818774446e-05,
|
1419 |
+
"loss": 0.5685,
|
1420 |
+
"step": 99500
|
1421 |
+
},
|
1422 |
+
{
|
1423 |
+
"epoch": 2.0371577574967406,
|
1424 |
+
"grad_norm": 18.42786979675293,
|
1425 |
+
"learning_rate": 1.7777053455019558e-05,
|
1426 |
+
"loss": 0.5539,
|
1427 |
+
"step": 100000
|
1428 |
+
},
|
1429 |
+
{
|
1430 |
+
"epoch": 2.047343546284224,
|
1431 |
+
"grad_norm": 25.627717971801758,
|
1432 |
+
"learning_rate": 1.7715938722294655e-05,
|
1433 |
+
"loss": 0.5393,
|
1434 |
+
"step": 100500
|
1435 |
+
},
|
1436 |
+
{
|
1437 |
+
"epoch": 2.057529335071708,
|
1438 |
+
"grad_norm": 15.942806243896484,
|
1439 |
+
"learning_rate": 1.7654823989569753e-05,
|
1440 |
+
"loss": 0.5191,
|
1441 |
+
"step": 101000
|
1442 |
+
},
|
1443 |
+
{
|
1444 |
+
"epoch": 2.0677151238591915,
|
1445 |
+
"grad_norm": 7.591663360595703,
|
1446 |
+
"learning_rate": 1.759370925684485e-05,
|
1447 |
+
"loss": 0.5428,
|
1448 |
+
"step": 101500
|
1449 |
+
},
|
1450 |
+
{
|
1451 |
+
"epoch": 2.077900912646675,
|
1452 |
+
"grad_norm": 7.183356285095215,
|
1453 |
+
"learning_rate": 1.7532594524119948e-05,
|
1454 |
+
"loss": 0.5475,
|
1455 |
+
"step": 102000
|
1456 |
+
},
|
1457 |
+
{
|
1458 |
+
"epoch": 2.088086701434159,
|
1459 |
+
"grad_norm": 19.916288375854492,
|
1460 |
+
"learning_rate": 1.7471479791395046e-05,
|
1461 |
+
"loss": 0.5586,
|
1462 |
+
"step": 102500
|
1463 |
+
},
|
1464 |
+
{
|
1465 |
+
"epoch": 2.098272490221643,
|
1466 |
+
"grad_norm": 9.07480239868164,
|
1467 |
+
"learning_rate": 1.7410365058670143e-05,
|
1468 |
+
"loss": 0.5564,
|
1469 |
+
"step": 103000
|
1470 |
+
},
|
1471 |
+
{
|
1472 |
+
"epoch": 2.1084582790091266,
|
1473 |
+
"grad_norm": 7.647058010101318,
|
1474 |
+
"learning_rate": 1.734925032594524e-05,
|
1475 |
+
"loss": 0.543,
|
1476 |
+
"step": 103500
|
1477 |
+
},
|
1478 |
+
{
|
1479 |
+
"epoch": 2.1186440677966103,
|
1480 |
+
"grad_norm": 8.871644973754883,
|
1481 |
+
"learning_rate": 1.728813559322034e-05,
|
1482 |
+
"loss": 0.5623,
|
1483 |
+
"step": 104000
|
1484 |
+
},
|
1485 |
+
{
|
1486 |
+
"epoch": 2.128829856584094,
|
1487 |
+
"grad_norm": 9.516389846801758,
|
1488 |
+
"learning_rate": 1.7227020860495436e-05,
|
1489 |
+
"loss": 0.5464,
|
1490 |
+
"step": 104500
|
1491 |
+
},
|
1492 |
+
{
|
1493 |
+
"epoch": 2.1390156453715776,
|
1494 |
+
"grad_norm": 13.605746269226074,
|
1495 |
+
"learning_rate": 1.7165906127770537e-05,
|
1496 |
+
"loss": 0.5425,
|
1497 |
+
"step": 105000
|
1498 |
+
},
|
1499 |
+
{
|
1500 |
+
"epoch": 2.1492014341590613,
|
1501 |
+
"grad_norm": 14.016572952270508,
|
1502 |
+
"learning_rate": 1.710479139504563e-05,
|
1503 |
+
"loss": 0.5496,
|
1504 |
+
"step": 105500
|
1505 |
+
},
|
1506 |
+
{
|
1507 |
+
"epoch": 2.159387222946545,
|
1508 |
+
"grad_norm": 14.719120979309082,
|
1509 |
+
"learning_rate": 1.7043676662320732e-05,
|
1510 |
+
"loss": 0.5591,
|
1511 |
+
"step": 106000
|
1512 |
+
},
|
1513 |
+
{
|
1514 |
+
"epoch": 2.1695730117340286,
|
1515 |
+
"grad_norm": 14.922130584716797,
|
1516 |
+
"learning_rate": 1.6982561929595826e-05,
|
1517 |
+
"loss": 0.5462,
|
1518 |
+
"step": 106500
|
1519 |
+
},
|
1520 |
+
{
|
1521 |
+
"epoch": 2.1797588005215123,
|
1522 |
+
"grad_norm": 10.74787425994873,
|
1523 |
+
"learning_rate": 1.6921447196870927e-05,
|
1524 |
+
"loss": 0.5699,
|
1525 |
+
"step": 107000
|
1526 |
+
},
|
1527 |
+
{
|
1528 |
+
"epoch": 2.189944589308996,
|
1529 |
+
"grad_norm": 7.978420257568359,
|
1530 |
+
"learning_rate": 1.6860332464146024e-05,
|
1531 |
+
"loss": 0.5449,
|
1532 |
+
"step": 107500
|
1533 |
+
},
|
1534 |
+
{
|
1535 |
+
"epoch": 2.2001303780964796,
|
1536 |
+
"grad_norm": 15.361347198486328,
|
1537 |
+
"learning_rate": 1.6799217731421122e-05,
|
1538 |
+
"loss": 0.5443,
|
1539 |
+
"step": 108000
|
1540 |
+
},
|
1541 |
+
{
|
1542 |
+
"epoch": 2.2103161668839633,
|
1543 |
+
"grad_norm": 6.552661418914795,
|
1544 |
+
"learning_rate": 1.673810299869622e-05,
|
1545 |
+
"loss": 0.5663,
|
1546 |
+
"step": 108500
|
1547 |
+
},
|
1548 |
+
{
|
1549 |
+
"epoch": 2.2205019556714474,
|
1550 |
+
"grad_norm": 6.750521659851074,
|
1551 |
+
"learning_rate": 1.6676988265971317e-05,
|
1552 |
+
"loss": 0.5488,
|
1553 |
+
"step": 109000
|
1554 |
+
},
|
1555 |
+
{
|
1556 |
+
"epoch": 2.230687744458931,
|
1557 |
+
"grad_norm": 11.149799346923828,
|
1558 |
+
"learning_rate": 1.6615873533246415e-05,
|
1559 |
+
"loss": 0.5412,
|
1560 |
+
"step": 109500
|
1561 |
+
},
|
1562 |
+
{
|
1563 |
+
"epoch": 2.2408735332464147,
|
1564 |
+
"grad_norm": 19.362638473510742,
|
1565 |
+
"learning_rate": 1.6554758800521512e-05,
|
1566 |
+
"loss": 0.5505,
|
1567 |
+
"step": 110000
|
1568 |
+
},
|
1569 |
+
{
|
1570 |
+
"epoch": 2.2510593220338984,
|
1571 |
+
"grad_norm": 11.263615608215332,
|
1572 |
+
"learning_rate": 1.6493644067796613e-05,
|
1573 |
+
"loss": 0.5591,
|
1574 |
+
"step": 110500
|
1575 |
+
},
|
1576 |
+
{
|
1577 |
+
"epoch": 2.261245110821382,
|
1578 |
+
"grad_norm": 8.90456771850586,
|
1579 |
+
"learning_rate": 1.6432529335071707e-05,
|
1580 |
+
"loss": 0.5491,
|
1581 |
+
"step": 111000
|
1582 |
+
},
|
1583 |
+
{
|
1584 |
+
"epoch": 2.2714308996088657,
|
1585 |
+
"grad_norm": 11.962569236755371,
|
1586 |
+
"learning_rate": 1.6371414602346808e-05,
|
1587 |
+
"loss": 0.556,
|
1588 |
+
"step": 111500
|
1589 |
+
},
|
1590 |
+
{
|
1591 |
+
"epoch": 2.2816166883963493,
|
1592 |
+
"grad_norm": 8.397544860839844,
|
1593 |
+
"learning_rate": 1.6310299869621902e-05,
|
1594 |
+
"loss": 0.5561,
|
1595 |
+
"step": 112000
|
1596 |
+
},
|
1597 |
+
{
|
1598 |
+
"epoch": 2.291802477183833,
|
1599 |
+
"grad_norm": 16.376155853271484,
|
1600 |
+
"learning_rate": 1.6249185136897003e-05,
|
1601 |
+
"loss": 0.5435,
|
1602 |
+
"step": 112500
|
1603 |
+
},
|
1604 |
+
{
|
1605 |
+
"epoch": 2.3019882659713167,
|
1606 |
+
"grad_norm": 9.14609432220459,
|
1607 |
+
"learning_rate": 1.6188070404172097e-05,
|
1608 |
+
"loss": 0.5457,
|
1609 |
+
"step": 113000
|
1610 |
+
},
|
1611 |
+
{
|
1612 |
+
"epoch": 2.3121740547588003,
|
1613 |
+
"grad_norm": 26.53936767578125,
|
1614 |
+
"learning_rate": 1.61269556714472e-05,
|
1615 |
+
"loss": 0.5295,
|
1616 |
+
"step": 113500
|
1617 |
+
},
|
1618 |
+
{
|
1619 |
+
"epoch": 2.322359843546284,
|
1620 |
+
"grad_norm": 17.222721099853516,
|
1621 |
+
"learning_rate": 1.6065840938722293e-05,
|
1622 |
+
"loss": 0.5177,
|
1623 |
+
"step": 114000
|
1624 |
+
},
|
1625 |
+
{
|
1626 |
+
"epoch": 2.332545632333768,
|
1627 |
+
"grad_norm": 9.585100173950195,
|
1628 |
+
"learning_rate": 1.6004726205997394e-05,
|
1629 |
+
"loss": 0.5419,
|
1630 |
+
"step": 114500
|
1631 |
+
},
|
1632 |
+
{
|
1633 |
+
"epoch": 2.3427314211212518,
|
1634 |
+
"grad_norm": 9.392489433288574,
|
1635 |
+
"learning_rate": 1.5943611473272488e-05,
|
1636 |
+
"loss": 0.5296,
|
1637 |
+
"step": 115000
|
1638 |
+
},
|
1639 |
+
{
|
1640 |
+
"epoch": 2.3529172099087354,
|
1641 |
+
"grad_norm": 15.146162986755371,
|
1642 |
+
"learning_rate": 1.588249674054759e-05,
|
1643 |
+
"loss": 0.5518,
|
1644 |
+
"step": 115500
|
1645 |
+
},
|
1646 |
+
{
|
1647 |
+
"epoch": 2.363102998696219,
|
1648 |
+
"grad_norm": 15.631641387939453,
|
1649 |
+
"learning_rate": 1.582138200782269e-05,
|
1650 |
+
"loss": 0.5584,
|
1651 |
+
"step": 116000
|
1652 |
+
},
|
1653 |
+
{
|
1654 |
+
"epoch": 2.3732887874837028,
|
1655 |
+
"grad_norm": 17.875492095947266,
|
1656 |
+
"learning_rate": 1.5760267275097784e-05,
|
1657 |
+
"loss": 0.5601,
|
1658 |
+
"step": 116500
|
1659 |
+
},
|
1660 |
+
{
|
1661 |
+
"epoch": 2.3834745762711864,
|
1662 |
+
"grad_norm": 7.7471184730529785,
|
1663 |
+
"learning_rate": 1.5699152542372885e-05,
|
1664 |
+
"loss": 0.549,
|
1665 |
+
"step": 117000
|
1666 |
+
},
|
1667 |
+
{
|
1668 |
+
"epoch": 2.39366036505867,
|
1669 |
+
"grad_norm": 14.712841987609863,
|
1670 |
+
"learning_rate": 1.563803780964798e-05,
|
1671 |
+
"loss": 0.5293,
|
1672 |
+
"step": 117500
|
1673 |
+
},
|
1674 |
+
{
|
1675 |
+
"epoch": 2.4038461538461537,
|
1676 |
+
"grad_norm": 9.229011535644531,
|
1677 |
+
"learning_rate": 1.557692307692308e-05,
|
1678 |
+
"loss": 0.5483,
|
1679 |
+
"step": 118000
|
1680 |
+
},
|
1681 |
+
{
|
1682 |
+
"epoch": 2.4140319426336374,
|
1683 |
+
"grad_norm": 11.47548770904541,
|
1684 |
+
"learning_rate": 1.5515808344198174e-05,
|
1685 |
+
"loss": 0.5529,
|
1686 |
+
"step": 118500
|
1687 |
+
},
|
1688 |
+
{
|
1689 |
+
"epoch": 2.424217731421121,
|
1690 |
+
"grad_norm": 29.980873107910156,
|
1691 |
+
"learning_rate": 1.5454693611473275e-05,
|
1692 |
+
"loss": 0.5489,
|
1693 |
+
"step": 119000
|
1694 |
+
},
|
1695 |
+
{
|
1696 |
+
"epoch": 2.4344035202086047,
|
1697 |
+
"grad_norm": 10.478185653686523,
|
1698 |
+
"learning_rate": 1.539357887874837e-05,
|
1699 |
+
"loss": 0.5595,
|
1700 |
+
"step": 119500
|
1701 |
+
},
|
1702 |
+
{
|
1703 |
+
"epoch": 2.444589308996089,
|
1704 |
+
"grad_norm": 12.978096008300781,
|
1705 |
+
"learning_rate": 1.533246414602347e-05,
|
1706 |
+
"loss": 0.5612,
|
1707 |
+
"step": 120000
|
1708 |
+
},
|
1709 |
+
{
|
1710 |
+
"epoch": 2.4547750977835725,
|
1711 |
+
"grad_norm": 17.834806442260742,
|
1712 |
+
"learning_rate": 1.5271349413298564e-05,
|
1713 |
+
"loss": 0.5412,
|
1714 |
+
"step": 120500
|
1715 |
+
},
|
1716 |
+
{
|
1717 |
+
"epoch": 2.464960886571056,
|
1718 |
+
"grad_norm": 25.001754760742188,
|
1719 |
+
"learning_rate": 1.5210234680573665e-05,
|
1720 |
+
"loss": 0.5673,
|
1721 |
+
"step": 121000
|
1722 |
+
},
|
1723 |
+
{
|
1724 |
+
"epoch": 2.47514667535854,
|
1725 |
+
"grad_norm": 27.953767776489258,
|
1726 |
+
"learning_rate": 1.5149119947848761e-05,
|
1727 |
+
"loss": 0.5497,
|
1728 |
+
"step": 121500
|
1729 |
+
},
|
1730 |
+
{
|
1731 |
+
"epoch": 2.4853324641460235,
|
1732 |
+
"grad_norm": 9.370855331420898,
|
1733 |
+
"learning_rate": 1.508800521512386e-05,
|
1734 |
+
"loss": 0.5344,
|
1735 |
+
"step": 122000
|
1736 |
+
},
|
1737 |
+
{
|
1738 |
+
"epoch": 2.495518252933507,
|
1739 |
+
"grad_norm": 18.109821319580078,
|
1740 |
+
"learning_rate": 1.5026890482398956e-05,
|
1741 |
+
"loss": 0.5572,
|
1742 |
+
"step": 122500
|
1743 |
+
},
|
1744 |
+
{
|
1745 |
+
"epoch": 2.505704041720991,
|
1746 |
+
"grad_norm": 20.435270309448242,
|
1747 |
+
"learning_rate": 1.4965775749674055e-05,
|
1748 |
+
"loss": 0.5631,
|
1749 |
+
"step": 123000
|
1750 |
+
},
|
1751 |
+
{
|
1752 |
+
"epoch": 2.5158898305084745,
|
1753 |
+
"grad_norm": 10.656272888183594,
|
1754 |
+
"learning_rate": 1.4904661016949153e-05,
|
1755 |
+
"loss": 0.5344,
|
1756 |
+
"step": 123500
|
1757 |
+
},
|
1758 |
+
{
|
1759 |
+
"epoch": 2.526075619295958,
|
1760 |
+
"grad_norm": 18.796552658081055,
|
1761 |
+
"learning_rate": 1.484354628422425e-05,
|
1762 |
+
"loss": 0.5464,
|
1763 |
+
"step": 124000
|
1764 |
+
},
|
1765 |
+
{
|
1766 |
+
"epoch": 2.5362614080834422,
|
1767 |
+
"grad_norm": 9.630306243896484,
|
1768 |
+
"learning_rate": 1.4782431551499348e-05,
|
1769 |
+
"loss": 0.547,
|
1770 |
+
"step": 124500
|
1771 |
+
},
|
1772 |
+
{
|
1773 |
+
"epoch": 2.5464471968709255,
|
1774 |
+
"grad_norm": 6.3599467277526855,
|
1775 |
+
"learning_rate": 1.4721316818774445e-05,
|
1776 |
+
"loss": 0.5577,
|
1777 |
+
"step": 125000
|
1778 |
+
},
|
1779 |
+
{
|
1780 |
+
"epoch": 2.5566329856584096,
|
1781 |
+
"grad_norm": 8.967144966125488,
|
1782 |
+
"learning_rate": 1.4660202086049545e-05,
|
1783 |
+
"loss": 0.5524,
|
1784 |
+
"step": 125500
|
1785 |
+
},
|
1786 |
+
{
|
1787 |
+
"epoch": 2.5668187744458932,
|
1788 |
+
"grad_norm": 16.690263748168945,
|
1789 |
+
"learning_rate": 1.4599087353324642e-05,
|
1790 |
+
"loss": 0.5546,
|
1791 |
+
"step": 126000
|
1792 |
+
},
|
1793 |
+
{
|
1794 |
+
"epoch": 2.577004563233377,
|
1795 |
+
"grad_norm": 17.713790893554688,
|
1796 |
+
"learning_rate": 1.453797262059974e-05,
|
1797 |
+
"loss": 0.5686,
|
1798 |
+
"step": 126500
|
1799 |
+
},
|
1800 |
+
{
|
1801 |
+
"epoch": 2.5871903520208606,
|
1802 |
+
"grad_norm": 7.970760822296143,
|
1803 |
+
"learning_rate": 1.4476857887874837e-05,
|
1804 |
+
"loss": 0.5273,
|
1805 |
+
"step": 127000
|
1806 |
+
},
|
1807 |
+
{
|
1808 |
+
"epoch": 2.5973761408083442,
|
1809 |
+
"grad_norm": 8.869507789611816,
|
1810 |
+
"learning_rate": 1.4415743155149935e-05,
|
1811 |
+
"loss": 0.5472,
|
1812 |
+
"step": 127500
|
1813 |
+
},
|
1814 |
+
{
|
1815 |
+
"epoch": 2.607561929595828,
|
1816 |
+
"grad_norm": 17.247589111328125,
|
1817 |
+
"learning_rate": 1.4354628422425032e-05,
|
1818 |
+
"loss": 0.5444,
|
1819 |
+
"step": 128000
|
1820 |
+
},
|
1821 |
+
{
|
1822 |
+
"epoch": 2.6177477183833116,
|
1823 |
+
"grad_norm": 16.415138244628906,
|
1824 |
+
"learning_rate": 1.429351368970013e-05,
|
1825 |
+
"loss": 0.5445,
|
1826 |
+
"step": 128500
|
1827 |
+
},
|
1828 |
+
{
|
1829 |
+
"epoch": 2.627933507170795,
|
1830 |
+
"grad_norm": 10.408729553222656,
|
1831 |
+
"learning_rate": 1.4232398956975227e-05,
|
1832 |
+
"loss": 0.5667,
|
1833 |
+
"step": 129000
|
1834 |
+
},
|
1835 |
+
{
|
1836 |
+
"epoch": 2.638119295958279,
|
1837 |
+
"grad_norm": 53.77888488769531,
|
1838 |
+
"learning_rate": 1.4171284224250327e-05,
|
1839 |
+
"loss": 0.547,
|
1840 |
+
"step": 129500
|
1841 |
+
},
|
1842 |
+
{
|
1843 |
+
"epoch": 2.648305084745763,
|
1844 |
+
"grad_norm": 12.244620323181152,
|
1845 |
+
"learning_rate": 1.4110169491525424e-05,
|
1846 |
+
"loss": 0.5428,
|
1847 |
+
"step": 130000
|
1848 |
+
},
|
1849 |
+
{
|
1850 |
+
"epoch": 2.658490873533246,
|
1851 |
+
"grad_norm": 11.80124568939209,
|
1852 |
+
"learning_rate": 1.4049054758800522e-05,
|
1853 |
+
"loss": 0.5565,
|
1854 |
+
"step": 130500
|
1855 |
+
},
|
1856 |
+
{
|
1857 |
+
"epoch": 2.6686766623207303,
|
1858 |
+
"grad_norm": 13.250449180603027,
|
1859 |
+
"learning_rate": 1.398794002607562e-05,
|
1860 |
+
"loss": 0.5423,
|
1861 |
+
"step": 131000
|
1862 |
+
},
|
1863 |
+
{
|
1864 |
+
"epoch": 2.678862451108214,
|
1865 |
+
"grad_norm": 15.81285572052002,
|
1866 |
+
"learning_rate": 1.3926825293350719e-05,
|
1867 |
+
"loss": 0.5621,
|
1868 |
+
"step": 131500
|
1869 |
+
},
|
1870 |
+
{
|
1871 |
+
"epoch": 2.6890482398956976,
|
1872 |
+
"grad_norm": 22.584035873413086,
|
1873 |
+
"learning_rate": 1.3865710560625816e-05,
|
1874 |
+
"loss": 0.5521,
|
1875 |
+
"step": 132000
|
1876 |
+
},
|
1877 |
+
{
|
1878 |
+
"epoch": 2.6992340286831813,
|
1879 |
+
"grad_norm": 23.865680694580078,
|
1880 |
+
"learning_rate": 1.3804595827900914e-05,
|
1881 |
+
"loss": 0.551,
|
1882 |
+
"step": 132500
|
1883 |
+
},
|
1884 |
+
{
|
1885 |
+
"epoch": 2.709419817470665,
|
1886 |
+
"grad_norm": 18.28876304626465,
|
1887 |
+
"learning_rate": 1.3743481095176011e-05,
|
1888 |
+
"loss": 0.5451,
|
1889 |
+
"step": 133000
|
1890 |
+
},
|
1891 |
+
{
|
1892 |
+
"epoch": 2.7196056062581486,
|
1893 |
+
"grad_norm": 16.23969078063965,
|
1894 |
+
"learning_rate": 1.3682366362451109e-05,
|
1895 |
+
"loss": 0.5392,
|
1896 |
+
"step": 133500
|
1897 |
+
},
|
1898 |
+
{
|
1899 |
+
"epoch": 2.7297913950456323,
|
1900 |
+
"grad_norm": 14.673959732055664,
|
1901 |
+
"learning_rate": 1.3621251629726206e-05,
|
1902 |
+
"loss": 0.5451,
|
1903 |
+
"step": 134000
|
1904 |
+
},
|
1905 |
+
{
|
1906 |
+
"epoch": 2.739977183833116,
|
1907 |
+
"grad_norm": 8.019514083862305,
|
1908 |
+
"learning_rate": 1.3560136897001304e-05,
|
1909 |
+
"loss": 0.5568,
|
1910 |
+
"step": 134500
|
1911 |
+
},
|
1912 |
+
{
|
1913 |
+
"epoch": 2.7501629726205996,
|
1914 |
+
"grad_norm": 13.495898246765137,
|
1915 |
+
"learning_rate": 1.3499022164276401e-05,
|
1916 |
+
"loss": 0.5734,
|
1917 |
+
"step": 135000
|
1918 |
+
},
|
1919 |
+
{
|
1920 |
+
"epoch": 2.7603487614080837,
|
1921 |
+
"grad_norm": 7.548976421356201,
|
1922 |
+
"learning_rate": 1.3437907431551499e-05,
|
1923 |
+
"loss": 0.557,
|
1924 |
+
"step": 135500
|
1925 |
+
},
|
1926 |
+
{
|
1927 |
+
"epoch": 2.770534550195567,
|
1928 |
+
"grad_norm": 20.049760818481445,
|
1929 |
+
"learning_rate": 1.3376792698826597e-05,
|
1930 |
+
"loss": 0.5658,
|
1931 |
+
"step": 136000
|
1932 |
+
},
|
1933 |
+
{
|
1934 |
+
"epoch": 2.780720338983051,
|
1935 |
+
"grad_norm": 9.346122741699219,
|
1936 |
+
"learning_rate": 1.3315677966101694e-05,
|
1937 |
+
"loss": 0.535,
|
1938 |
+
"step": 136500
|
1939 |
+
},
|
1940 |
+
{
|
1941 |
+
"epoch": 2.7909061277705347,
|
1942 |
+
"grad_norm": 15.080660820007324,
|
1943 |
+
"learning_rate": 1.3254563233376792e-05,
|
1944 |
+
"loss": 0.5538,
|
1945 |
+
"step": 137000
|
1946 |
+
},
|
1947 |
+
{
|
1948 |
+
"epoch": 2.8010919165580184,
|
1949 |
+
"grad_norm": 11.485374450683594,
|
1950 |
+
"learning_rate": 1.3193448500651891e-05,
|
1951 |
+
"loss": 0.5419,
|
1952 |
+
"step": 137500
|
1953 |
+
},
|
1954 |
+
{
|
1955 |
+
"epoch": 2.811277705345502,
|
1956 |
+
"grad_norm": 12.089446067810059,
|
1957 |
+
"learning_rate": 1.3132333767926988e-05,
|
1958 |
+
"loss": 0.5651,
|
1959 |
+
"step": 138000
|
1960 |
+
},
|
1961 |
+
{
|
1962 |
+
"epoch": 2.8214634941329857,
|
1963 |
+
"grad_norm": 5.466490268707275,
|
1964 |
+
"learning_rate": 1.3071219035202088e-05,
|
1965 |
+
"loss": 0.5217,
|
1966 |
+
"step": 138500
|
1967 |
+
},
|
1968 |
+
{
|
1969 |
+
"epoch": 2.8316492829204694,
|
1970 |
+
"grad_norm": 12.89148235321045,
|
1971 |
+
"learning_rate": 1.3010104302477185e-05,
|
1972 |
+
"loss": 0.5477,
|
1973 |
+
"step": 139000
|
1974 |
+
},
|
1975 |
+
{
|
1976 |
+
"epoch": 2.841835071707953,
|
1977 |
+
"grad_norm": 5.610709190368652,
|
1978 |
+
"learning_rate": 1.2948989569752283e-05,
|
1979 |
+
"loss": 0.5377,
|
1980 |
+
"step": 139500
|
1981 |
+
},
|
1982 |
+
{
|
1983 |
+
"epoch": 2.8520208604954367,
|
1984 |
+
"grad_norm": 26.2186222076416,
|
1985 |
+
"learning_rate": 1.288787483702738e-05,
|
1986 |
+
"loss": 0.5281,
|
1987 |
+
"step": 140000
|
1988 |
+
},
|
1989 |
+
{
|
1990 |
+
"epoch": 2.8622066492829203,
|
1991 |
+
"grad_norm": 2.1066462993621826,
|
1992 |
+
"learning_rate": 1.2826760104302478e-05,
|
1993 |
+
"loss": 0.5387,
|
1994 |
+
"step": 140500
|
1995 |
+
},
|
1996 |
+
{
|
1997 |
+
"epoch": 2.872392438070404,
|
1998 |
+
"grad_norm": 13.6646728515625,
|
1999 |
+
"learning_rate": 1.2765645371577575e-05,
|
2000 |
+
"loss": 0.5416,
|
2001 |
+
"step": 141000
|
2002 |
+
},
|
2003 |
+
{
|
2004 |
+
"epoch": 2.8825782268578877,
|
2005 |
+
"grad_norm": 14.357284545898438,
|
2006 |
+
"learning_rate": 1.2704530638852673e-05,
|
2007 |
+
"loss": 0.5504,
|
2008 |
+
"step": 141500
|
2009 |
+
},
|
2010 |
+
{
|
2011 |
+
"epoch": 2.8927640156453718,
|
2012 |
+
"grad_norm": 14.08674144744873,
|
2013 |
+
"learning_rate": 1.264341590612777e-05,
|
2014 |
+
"loss": 0.5691,
|
2015 |
+
"step": 142000
|
2016 |
+
},
|
2017 |
+
{
|
2018 |
+
"epoch": 2.9029498044328554,
|
2019 |
+
"grad_norm": 20.460561752319336,
|
2020 |
+
"learning_rate": 1.2582301173402868e-05,
|
2021 |
+
"loss": 0.5624,
|
2022 |
+
"step": 142500
|
2023 |
+
},
|
2024 |
+
{
|
2025 |
+
"epoch": 2.913135593220339,
|
2026 |
+
"grad_norm": 12.731225967407227,
|
2027 |
+
"learning_rate": 1.2521186440677966e-05,
|
2028 |
+
"loss": 0.5439,
|
2029 |
+
"step": 143000
|
2030 |
+
},
|
2031 |
+
{
|
2032 |
+
"epoch": 2.9233213820078228,
|
2033 |
+
"grad_norm": 3.781522274017334,
|
2034 |
+
"learning_rate": 1.2460071707953065e-05,
|
2035 |
+
"loss": 0.5348,
|
2036 |
+
"step": 143500
|
2037 |
+
},
|
2038 |
+
{
|
2039 |
+
"epoch": 2.9335071707953064,
|
2040 |
+
"grad_norm": 14.624204635620117,
|
2041 |
+
"learning_rate": 1.2398956975228162e-05,
|
2042 |
+
"loss": 0.5589,
|
2043 |
+
"step": 144000
|
2044 |
+
},
|
2045 |
+
{
|
2046 |
+
"epoch": 2.94369295958279,
|
2047 |
+
"grad_norm": 18.551095962524414,
|
2048 |
+
"learning_rate": 1.233784224250326e-05,
|
2049 |
+
"loss": 0.532,
|
2050 |
+
"step": 144500
|
2051 |
+
},
|
2052 |
+
{
|
2053 |
+
"epoch": 2.9538787483702738,
|
2054 |
+
"grad_norm": 17.094831466674805,
|
2055 |
+
"learning_rate": 1.2276727509778357e-05,
|
2056 |
+
"loss": 0.558,
|
2057 |
+
"step": 145000
|
2058 |
+
},
|
2059 |
+
{
|
2060 |
+
"epoch": 2.9640645371577574,
|
2061 |
+
"grad_norm": 12.306907653808594,
|
2062 |
+
"learning_rate": 1.2215612777053455e-05,
|
2063 |
+
"loss": 0.542,
|
2064 |
+
"step": 145500
|
2065 |
+
},
|
2066 |
+
{
|
2067 |
+
"epoch": 2.974250325945241,
|
2068 |
+
"grad_norm": 12.134025573730469,
|
2069 |
+
"learning_rate": 1.2154498044328553e-05,
|
2070 |
+
"loss": 0.5429,
|
2071 |
+
"step": 146000
|
2072 |
+
},
|
2073 |
+
{
|
2074 |
+
"epoch": 2.9844361147327247,
|
2075 |
+
"grad_norm": 22.295795440673828,
|
2076 |
+
"learning_rate": 1.209338331160365e-05,
|
2077 |
+
"loss": 0.5632,
|
2078 |
+
"step": 146500
|
2079 |
+
},
|
2080 |
+
{
|
2081 |
+
"epoch": 2.9946219035202084,
|
2082 |
+
"grad_norm": 17.749858856201172,
|
2083 |
+
"learning_rate": 1.2032268578878748e-05,
|
2084 |
+
"loss": 0.5402,
|
2085 |
+
"step": 147000
|
2086 |
+
},
|
2087 |
+
{
|
2088 |
+
"epoch": 3.0,
|
2089 |
+
"eval_accuracy": 0.7337000966072083,
|
2090 |
+
"eval_loss": 0.7322831153869629,
|
2091 |
+
"eval_runtime": 8.3496,
|
2092 |
+
"eval_samples_per_second": 587.813,
|
2093 |
+
"eval_steps_per_second": 73.536,
|
2094 |
+
"step": 147264
|
2095 |
+
}
|
2096 |
+
],
|
2097 |
+
"logging_steps": 500,
|
2098 |
+
"max_steps": 245440,
|
2099 |
+
"num_input_tokens_seen": 0,
|
2100 |
+
"num_train_epochs": 5,
|
2101 |
+
"save_steps": 500,
|
2102 |
+
"stateful_callbacks": {
|
2103 |
+
"EarlyStoppingCallback": {
|
2104 |
+
"args": {
|
2105 |
+
"early_stopping_patience": 3,
|
2106 |
+
"early_stopping_threshold": 0.001
|
2107 |
+
},
|
2108 |
+
"attributes": {
|
2109 |
+
"early_stopping_patience_counter": 0
|
2110 |
+
}
|
2111 |
+
},
|
2112 |
+
"TrainerControl": {
|
2113 |
+
"args": {
|
2114 |
+
"should_epoch_stop": false,
|
2115 |
+
"should_evaluate": false,
|
2116 |
+
"should_log": false,
|
2117 |
+
"should_save": true,
|
2118 |
+
"should_training_stop": false
|
2119 |
+
},
|
2120 |
+
"attributes": {}
|
2121 |
+
}
|
2122 |
+
},
|
2123 |
+
"total_flos": 6.950698126852608e+16,
|
2124 |
+
"train_batch_size": 8,
|
2125 |
+
"trial_name": null,
|
2126 |
+
"trial_params": null
|
2127 |
+
}
|
checkpoint-147264/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9d044fe66ccd6291067c67bf5a07a2d7b47163784c08e151e57205d231bd75e9
|
3 |
+
size 5368
|
config.json
ADDED
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "/home/ubuntu/utah/babylm-24/data/training/models/10M_babylm_ascii/SPM-Unigram_6144/DebertaV2-Base-10M_babylm-A",
|
3 |
+
"architectures": [
|
4 |
+
"DebertaV2ForSequenceClassification"
|
5 |
+
],
|
6 |
+
"attention_probs_dropout_prob": 0.1,
|
7 |
+
"hidden_act": "gelu",
|
8 |
+
"hidden_dropout_prob": 0.1,
|
9 |
+
"hidden_size": 768,
|
10 |
+
"id2label": {
|
11 |
+
"0": "LABEL_0",
|
12 |
+
"1": "LABEL_1",
|
13 |
+
"2": "LABEL_2"
|
14 |
+
},
|
15 |
+
"initializer_range": 0.02,
|
16 |
+
"intermediate_size": 1536,
|
17 |
+
"label2id": {
|
18 |
+
"0": 0,
|
19 |
+
"1": 1,
|
20 |
+
"2": 2
|
21 |
+
},
|
22 |
+
"layer_norm_eps": 1e-07,
|
23 |
+
"max_position_embeddings": 512,
|
24 |
+
"max_relative_positions": -1,
|
25 |
+
"model_type": "deberta-v2",
|
26 |
+
"num_attention_heads": 12,
|
27 |
+
"num_hidden_layers": 8,
|
28 |
+
"pad_token_id": 3,
|
29 |
+
"pooler_dropout": 0,
|
30 |
+
"pooler_hidden_act": "gelu",
|
31 |
+
"pooler_hidden_size": 768,
|
32 |
+
"pos_att_type": null,
|
33 |
+
"position_biased_input": true,
|
34 |
+
"relative_attention": false,
|
35 |
+
"torch_dtype": "float32",
|
36 |
+
"transformers_version": "4.44.2",
|
37 |
+
"type_vocab_size": 0,
|
38 |
+
"vocab_size": 6144
|
39 |
+
}
|
eval_results.json
ADDED
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"epoch": 5.0,
|
3 |
+
"eval_accuracy": 0.7337000966072083,
|
4 |
+
"eval_loss": 0.7322831153869629,
|
5 |
+
"eval_runtime": 7.5979,
|
6 |
+
"eval_samples": 4908,
|
7 |
+
"eval_samples_per_second": 645.964,
|
8 |
+
"eval_steps_per_second": 80.811
|
9 |
+
}
|
model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0575c4bc1a8943e345d617f7dcaac754a2941686ef65c5c0917ce3a4386ce1c5
|
3 |
+
size 174106588
|
predictions.txt
ADDED
@@ -0,0 +1,4909 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
index prediction
|
2 |
+
0 2
|
3 |
+
1 2
|
4 |
+
2 2
|
5 |
+
3 0
|
6 |
+
4 2
|
7 |
+
5 1
|
8 |
+
6 2
|
9 |
+
7 1
|
10 |
+
8 1
|
11 |
+
9 2
|
12 |
+
10 1
|
13 |
+
11 1
|
14 |
+
12 1
|
15 |
+
13 2
|
16 |
+
14 0
|
17 |
+
15 1
|
18 |
+
16 0
|
19 |
+
17 0
|
20 |
+
18 1
|
21 |
+
19 2
|
22 |
+
20 0
|
23 |
+
21 0
|
24 |
+
22 1
|
25 |
+
23 0
|
26 |
+
24 2
|
27 |
+
25 0
|
28 |
+
26 1
|
29 |
+
27 1
|
30 |
+
28 1
|
31 |
+
29 0
|
32 |
+
30 2
|
33 |
+
31 1
|
34 |
+
32 0
|
35 |
+
33 1
|
36 |
+
34 1
|
37 |
+
35 1
|
38 |
+
36 0
|
39 |
+
37 0
|
40 |
+
38 1
|
41 |
+
39 0
|
42 |
+
40 0
|
43 |
+
41 1
|
44 |
+
42 1
|
45 |
+
43 0
|
46 |
+
44 2
|
47 |
+
45 2
|
48 |
+
46 0
|
49 |
+
47 0
|
50 |
+
48 2
|
51 |
+
49 2
|
52 |
+
50 2
|
53 |
+
51 0
|
54 |
+
52 2
|
55 |
+
53 0
|
56 |
+
54 2
|
57 |
+
55 2
|
58 |
+
56 1
|
59 |
+
57 1
|
60 |
+
58 0
|
61 |
+
59 1
|
62 |
+
60 2
|
63 |
+
61 1
|
64 |
+
62 0
|
65 |
+
63 1
|
66 |
+
64 1
|
67 |
+
65 1
|
68 |
+
66 0
|
69 |
+
67 1
|
70 |
+
68 0
|
71 |
+
69 1
|
72 |
+
70 0
|
73 |
+
71 0
|
74 |
+
72 1
|
75 |
+
73 0
|
76 |
+
74 2
|
77 |
+
75 0
|
78 |
+
76 1
|
79 |
+
77 1
|
80 |
+
78 0
|
81 |
+
79 0
|
82 |
+
80 2
|
83 |
+
81 1
|
84 |
+
82 2
|
85 |
+
83 2
|
86 |
+
84 0
|
87 |
+
85 2
|
88 |
+
86 0
|
89 |
+
87 1
|
90 |
+
88 1
|
91 |
+
89 0
|
92 |
+
90 0
|
93 |
+
91 2
|
94 |
+
92 1
|
95 |
+
93 2
|
96 |
+
94 1
|
97 |
+
95 0
|
98 |
+
96 1
|
99 |
+
97 0
|
100 |
+
98 2
|
101 |
+
99 1
|
102 |
+
100 1
|
103 |
+
101 0
|
104 |
+
102 1
|
105 |
+
103 1
|
106 |
+
104 1
|
107 |
+
105 0
|
108 |
+
106 0
|
109 |
+
107 1
|
110 |
+
108 2
|
111 |
+
109 2
|
112 |
+
110 0
|
113 |
+
111 0
|
114 |
+
112 1
|
115 |
+
113 1
|
116 |
+
114 0
|
117 |
+
115 0
|
118 |
+
116 1
|
119 |
+
117 1
|
120 |
+
118 2
|
121 |
+
119 2
|
122 |
+
120 2
|
123 |
+
121 1
|
124 |
+
122 2
|
125 |
+
123 2
|
126 |
+
124 0
|
127 |
+
125 1
|
128 |
+
126 0
|
129 |
+
127 1
|
130 |
+
128 1
|
131 |
+
129 0
|
132 |
+
130 0
|
133 |
+
131 0
|
134 |
+
132 2
|
135 |
+
133 0
|
136 |
+
134 0
|
137 |
+
135 2
|
138 |
+
136 1
|
139 |
+
137 2
|
140 |
+
138 2
|
141 |
+
139 2
|
142 |
+
140 2
|
143 |
+
141 2
|
144 |
+
142 1
|
145 |
+
143 2
|
146 |
+
144 0
|
147 |
+
145 0
|
148 |
+
146 1
|
149 |
+
147 1
|
150 |
+
148 1
|
151 |
+
149 2
|
152 |
+
150 1
|
153 |
+
151 1
|
154 |
+
152 1
|
155 |
+
153 1
|
156 |
+
154 1
|
157 |
+
155 0
|
158 |
+
156 0
|
159 |
+
157 0
|
160 |
+
158 1
|
161 |
+
159 2
|
162 |
+
160 1
|
163 |
+
161 1
|
164 |
+
162 0
|
165 |
+
163 2
|
166 |
+
164 2
|
167 |
+
165 0
|
168 |
+
166 0
|
169 |
+
167 0
|
170 |
+
168 1
|
171 |
+
169 2
|
172 |
+
170 2
|
173 |
+
171 0
|
174 |
+
172 1
|
175 |
+
173 2
|
176 |
+
174 0
|
177 |
+
175 1
|
178 |
+
176 1
|
179 |
+
177 1
|
180 |
+
178 0
|
181 |
+
179 1
|
182 |
+
180 1
|
183 |
+
181 1
|
184 |
+
182 2
|
185 |
+
183 2
|
186 |
+
184 2
|
187 |
+
185 0
|
188 |
+
186 0
|
189 |
+
187 1
|
190 |
+
188 2
|
191 |
+
189 0
|
192 |
+
190 1
|
193 |
+
191 1
|
194 |
+
192 0
|
195 |
+
193 2
|
196 |
+
194 1
|
197 |
+
195 2
|
198 |
+
196 1
|
199 |
+
197 1
|
200 |
+
198 2
|
201 |
+
199 2
|
202 |
+
200 0
|
203 |
+
201 0
|
204 |
+
202 2
|
205 |
+
203 0
|
206 |
+
204 1
|
207 |
+
205 0
|
208 |
+
206 1
|
209 |
+
207 0
|
210 |
+
208 0
|
211 |
+
209 0
|
212 |
+
210 1
|
213 |
+
211 0
|
214 |
+
212 1
|
215 |
+
213 0
|
216 |
+
214 2
|
217 |
+
215 1
|
218 |
+
216 0
|
219 |
+
217 0
|
220 |
+
218 2
|
221 |
+
219 2
|
222 |
+
220 1
|
223 |
+
221 2
|
224 |
+
222 0
|
225 |
+
223 2
|
226 |
+
224 1
|
227 |
+
225 0
|
228 |
+
226 0
|
229 |
+
227 1
|
230 |
+
228 2
|
231 |
+
229 1
|
232 |
+
230 2
|
233 |
+
231 0
|
234 |
+
232 1
|
235 |
+
233 0
|
236 |
+
234 0
|
237 |
+
235 2
|
238 |
+
236 0
|
239 |
+
237 2
|
240 |
+
238 2
|
241 |
+
239 0
|
242 |
+
240 0
|
243 |
+
241 1
|
244 |
+
242 1
|
245 |
+
243 0
|
246 |
+
244 1
|
247 |
+
245 1
|
248 |
+
246 1
|
249 |
+
247 2
|
250 |
+
248 2
|
251 |
+
249 1
|
252 |
+
250 1
|
253 |
+
251 0
|
254 |
+
252 2
|
255 |
+
253 0
|
256 |
+
254 2
|
257 |
+
255 2
|
258 |
+
256 0
|
259 |
+
257 1
|
260 |
+
258 0
|
261 |
+
259 0
|
262 |
+
260 2
|
263 |
+
261 0
|
264 |
+
262 0
|
265 |
+
263 1
|
266 |
+
264 2
|
267 |
+
265 2
|
268 |
+
266 1
|
269 |
+
267 1
|
270 |
+
268 2
|
271 |
+
269 1
|
272 |
+
270 2
|
273 |
+
271 1
|
274 |
+
272 2
|
275 |
+
273 2
|
276 |
+
274 0
|
277 |
+
275 1
|
278 |
+
276 1
|
279 |
+
277 1
|
280 |
+
278 0
|
281 |
+
279 0
|
282 |
+
280 0
|
283 |
+
281 0
|
284 |
+
282 1
|
285 |
+
283 0
|
286 |
+
284 1
|
287 |
+
285 0
|
288 |
+
286 1
|
289 |
+
287 0
|
290 |
+
288 0
|
291 |
+
289 1
|
292 |
+
290 2
|
293 |
+
291 0
|
294 |
+
292 1
|
295 |
+
293 2
|
296 |
+
294 1
|
297 |
+
295 1
|
298 |
+
296 1
|
299 |
+
297 1
|
300 |
+
298 1
|
301 |
+
299 0
|
302 |
+
300 1
|
303 |
+
301 0
|
304 |
+
302 2
|
305 |
+
303 2
|
306 |
+
304 1
|
307 |
+
305 0
|
308 |
+
306 0
|
309 |
+
307 1
|
310 |
+
308 0
|
311 |
+
309 1
|
312 |
+
310 2
|
313 |
+
311 0
|
314 |
+
312 0
|
315 |
+
313 1
|
316 |
+
314 1
|
317 |
+
315 1
|
318 |
+
316 2
|
319 |
+
317 0
|
320 |
+
318 1
|
321 |
+
319 2
|
322 |
+
320 1
|
323 |
+
321 1
|
324 |
+
322 2
|
325 |
+
323 2
|
326 |
+
324 0
|
327 |
+
325 2
|
328 |
+
326 1
|
329 |
+
327 0
|
330 |
+
328 1
|
331 |
+
329 2
|
332 |
+
330 1
|
333 |
+
331 0
|
334 |
+
332 0
|
335 |
+
333 0
|
336 |
+
334 1
|
337 |
+
335 2
|
338 |
+
336 0
|
339 |
+
337 1
|
340 |
+
338 0
|
341 |
+
339 2
|
342 |
+
340 2
|
343 |
+
341 0
|
344 |
+
342 1
|
345 |
+
343 2
|
346 |
+
344 0
|
347 |
+
345 0
|
348 |
+
346 0
|
349 |
+
347 0
|
350 |
+
348 0
|
351 |
+
349 0
|
352 |
+
350 1
|
353 |
+
351 1
|
354 |
+
352 1
|
355 |
+
353 1
|
356 |
+
354 1
|
357 |
+
355 1
|
358 |
+
356 1
|
359 |
+
357 1
|
360 |
+
358 1
|
361 |
+
359 1
|
362 |
+
360 1
|
363 |
+
361 0
|
364 |
+
362 2
|
365 |
+
363 1
|
366 |
+
364 1
|
367 |
+
365 0
|
368 |
+
366 1
|
369 |
+
367 0
|
370 |
+
368 2
|
371 |
+
369 0
|
372 |
+
370 2
|
373 |
+
371 2
|
374 |
+
372 1
|
375 |
+
373 0
|
376 |
+
374 0
|
377 |
+
375 0
|
378 |
+
376 2
|
379 |
+
377 1
|
380 |
+
378 0
|
381 |
+
379 2
|
382 |
+
380 2
|
383 |
+
381 0
|
384 |
+
382 1
|
385 |
+
383 0
|
386 |
+
384 2
|
387 |
+
385 1
|
388 |
+
386 0
|
389 |
+
387 0
|
390 |
+
388 0
|
391 |
+
389 2
|
392 |
+
390 0
|
393 |
+
391 1
|
394 |
+
392 0
|
395 |
+
393 0
|
396 |
+
394 1
|
397 |
+
395 0
|
398 |
+
396 0
|
399 |
+
397 1
|
400 |
+
398 2
|
401 |
+
399 0
|
402 |
+
400 1
|
403 |
+
401 0
|
404 |
+
402 0
|
405 |
+
403 2
|
406 |
+
404 2
|
407 |
+
405 0
|
408 |
+
406 1
|
409 |
+
407 2
|
410 |
+
408 1
|
411 |
+
409 2
|
412 |
+
410 2
|
413 |
+
411 1
|
414 |
+
412 1
|
415 |
+
413 1
|
416 |
+
414 2
|
417 |
+
415 0
|
418 |
+
416 1
|
419 |
+
417 2
|
420 |
+
418 0
|
421 |
+
419 0
|
422 |
+
420 0
|
423 |
+
421 2
|
424 |
+
422 1
|
425 |
+
423 0
|
426 |
+
424 0
|
427 |
+
425 0
|
428 |
+
426 2
|
429 |
+
427 0
|
430 |
+
428 1
|
431 |
+
429 2
|
432 |
+
430 0
|
433 |
+
431 1
|
434 |
+
432 1
|
435 |
+
433 0
|
436 |
+
434 1
|
437 |
+
435 2
|
438 |
+
436 2
|
439 |
+
437 0
|
440 |
+
438 0
|
441 |
+
439 1
|
442 |
+
440 1
|
443 |
+
441 1
|
444 |
+
442 1
|
445 |
+
443 1
|
446 |
+
444 1
|
447 |
+
445 2
|
448 |
+
446 2
|
449 |
+
447 0
|
450 |
+
448 2
|
451 |
+
449 1
|
452 |
+
450 1
|
453 |
+
451 1
|
454 |
+
452 1
|
455 |
+
453 0
|
456 |
+
454 1
|
457 |
+
455 1
|
458 |
+
456 0
|
459 |
+
457 1
|
460 |
+
458 1
|
461 |
+
459 1
|
462 |
+
460 2
|
463 |
+
461 2
|
464 |
+
462 1
|
465 |
+
463 1
|
466 |
+
464 2
|
467 |
+
465 1
|
468 |
+
466 0
|
469 |
+
467 0
|
470 |
+
468 1
|
471 |
+
469 0
|
472 |
+
470 0
|
473 |
+
471 2
|
474 |
+
472 0
|
475 |
+
473 1
|
476 |
+
474 2
|
477 |
+
475 0
|
478 |
+
476 0
|
479 |
+
477 2
|
480 |
+
478 1
|
481 |
+
479 2
|
482 |
+
480 1
|
483 |
+
481 0
|
484 |
+
482 2
|
485 |
+
483 0
|
486 |
+
484 0
|
487 |
+
485 1
|
488 |
+
486 1
|
489 |
+
487 1
|
490 |
+
488 0
|
491 |
+
489 0
|
492 |
+
490 1
|
493 |
+
491 2
|
494 |
+
492 1
|
495 |
+
493 0
|
496 |
+
494 2
|
497 |
+
495 1
|
498 |
+
496 2
|
499 |
+
497 2
|
500 |
+
498 2
|
501 |
+
499 1
|
502 |
+
500 0
|
503 |
+
501 0
|
504 |
+
502 1
|
505 |
+
503 1
|
506 |
+
504 1
|
507 |
+
505 0
|
508 |
+
506 0
|
509 |
+
507 1
|
510 |
+
508 0
|
511 |
+
509 1
|
512 |
+
510 1
|
513 |
+
511 2
|
514 |
+
512 1
|
515 |
+
513 2
|
516 |
+
514 1
|
517 |
+
515 1
|
518 |
+
516 2
|
519 |
+
517 2
|
520 |
+
518 2
|
521 |
+
519 1
|
522 |
+
520 2
|
523 |
+
521 1
|
524 |
+
522 2
|
525 |
+
523 0
|
526 |
+
524 1
|
527 |
+
525 1
|
528 |
+
526 2
|
529 |
+
527 1
|
530 |
+
528 0
|
531 |
+
529 1
|
532 |
+
530 1
|
533 |
+
531 1
|
534 |
+
532 0
|
535 |
+
533 1
|
536 |
+
534 1
|
537 |
+
535 1
|
538 |
+
536 2
|
539 |
+
537 0
|
540 |
+
538 2
|
541 |
+
539 2
|
542 |
+
540 1
|
543 |
+
541 0
|
544 |
+
542 2
|
545 |
+
543 2
|
546 |
+
544 2
|
547 |
+
545 1
|
548 |
+
546 0
|
549 |
+
547 1
|
550 |
+
548 2
|
551 |
+
549 1
|
552 |
+
550 0
|
553 |
+
551 0
|
554 |
+
552 0
|
555 |
+
553 1
|
556 |
+
554 2
|
557 |
+
555 0
|
558 |
+
556 1
|
559 |
+
557 0
|
560 |
+
558 1
|
561 |
+
559 1
|
562 |
+
560 0
|
563 |
+
561 1
|
564 |
+
562 1
|
565 |
+
563 1
|
566 |
+
564 2
|
567 |
+
565 1
|
568 |
+
566 2
|
569 |
+
567 2
|
570 |
+
568 2
|
571 |
+
569 1
|
572 |
+
570 0
|
573 |
+
571 1
|
574 |
+
572 2
|
575 |
+
573 2
|
576 |
+
574 2
|
577 |
+
575 1
|
578 |
+
576 1
|
579 |
+
577 1
|
580 |
+
578 1
|
581 |
+
579 1
|
582 |
+
580 0
|
583 |
+
581 0
|
584 |
+
582 0
|
585 |
+
583 1
|
586 |
+
584 0
|
587 |
+
585 0
|
588 |
+
586 0
|
589 |
+
587 1
|
590 |
+
588 2
|
591 |
+
589 0
|
592 |
+
590 0
|
593 |
+
591 0
|
594 |
+
592 0
|
595 |
+
593 0
|
596 |
+
594 0
|
597 |
+
595 0
|
598 |
+
596 1
|
599 |
+
597 2
|
600 |
+
598 1
|
601 |
+
599 1
|
602 |
+
600 2
|
603 |
+
601 2
|
604 |
+
602 1
|
605 |
+
603 1
|
606 |
+
604 0
|
607 |
+
605 0
|
608 |
+
606 2
|
609 |
+
607 2
|
610 |
+
608 1
|
611 |
+
609 1
|
612 |
+
610 2
|
613 |
+
611 2
|
614 |
+
612 0
|
615 |
+
613 1
|
616 |
+
614 2
|
617 |
+
615 0
|
618 |
+
616 0
|
619 |
+
617 0
|
620 |
+
618 0
|
621 |
+
619 2
|
622 |
+
620 0
|
623 |
+
621 1
|
624 |
+
622 1
|
625 |
+
623 2
|
626 |
+
624 1
|
627 |
+
625 2
|
628 |
+
626 0
|
629 |
+
627 2
|
630 |
+
628 2
|
631 |
+
629 1
|
632 |
+
630 2
|
633 |
+
631 2
|
634 |
+
632 1
|
635 |
+
633 1
|
636 |
+
634 1
|
637 |
+
635 0
|
638 |
+
636 0
|
639 |
+
637 1
|
640 |
+
638 2
|
641 |
+
639 1
|
642 |
+
640 0
|
643 |
+
641 0
|
644 |
+
642 0
|
645 |
+
643 1
|
646 |
+
644 0
|
647 |
+
645 0
|
648 |
+
646 2
|
649 |
+
647 2
|
650 |
+
648 2
|
651 |
+
649 0
|
652 |
+
650 2
|
653 |
+
651 0
|
654 |
+
652 1
|
655 |
+
653 2
|
656 |
+
654 2
|
657 |
+
655 0
|
658 |
+
656 2
|
659 |
+
657 1
|
660 |
+
658 0
|
661 |
+
659 0
|
662 |
+
660 2
|
663 |
+
661 1
|
664 |
+
662 0
|
665 |
+
663 0
|
666 |
+
664 0
|
667 |
+
665 0
|
668 |
+
666 1
|
669 |
+
667 1
|
670 |
+
668 2
|
671 |
+
669 1
|
672 |
+
670 2
|
673 |
+
671 1
|
674 |
+
672 0
|
675 |
+
673 1
|
676 |
+
674 0
|
677 |
+
675 1
|
678 |
+
676 2
|
679 |
+
677 0
|
680 |
+
678 2
|
681 |
+
679 1
|
682 |
+
680 0
|
683 |
+
681 2
|
684 |
+
682 2
|
685 |
+
683 1
|
686 |
+
684 0
|
687 |
+
685 0
|
688 |
+
686 1
|
689 |
+
687 2
|
690 |
+
688 0
|
691 |
+
689 1
|
692 |
+
690 2
|
693 |
+
691 1
|
694 |
+
692 1
|
695 |
+
693 0
|
696 |
+
694 1
|
697 |
+
695 2
|
698 |
+
696 1
|
699 |
+
697 2
|
700 |
+
698 0
|
701 |
+
699 1
|
702 |
+
700 1
|
703 |
+
701 0
|
704 |
+
702 0
|
705 |
+
703 0
|
706 |
+
704 2
|
707 |
+
705 0
|
708 |
+
706 1
|
709 |
+
707 0
|
710 |
+
708 0
|
711 |
+
709 1
|
712 |
+
710 0
|
713 |
+
711 0
|
714 |
+
712 1
|
715 |
+
713 0
|
716 |
+
714 2
|
717 |
+
715 1
|
718 |
+
716 1
|
719 |
+
717 1
|
720 |
+
718 0
|
721 |
+
719 2
|
722 |
+
720 0
|
723 |
+
721 0
|
724 |
+
722 2
|
725 |
+
723 2
|
726 |
+
724 0
|
727 |
+
725 2
|
728 |
+
726 1
|
729 |
+
727 0
|
730 |
+
728 1
|
731 |
+
729 0
|
732 |
+
730 2
|
733 |
+
731 0
|
734 |
+
732 2
|
735 |
+
733 0
|
736 |
+
734 0
|
737 |
+
735 1
|
738 |
+
736 0
|
739 |
+
737 2
|
740 |
+
738 0
|
741 |
+
739 2
|
742 |
+
740 0
|
743 |
+
741 0
|
744 |
+
742 1
|
745 |
+
743 1
|
746 |
+
744 2
|
747 |
+
745 1
|
748 |
+
746 1
|
749 |
+
747 0
|
750 |
+
748 1
|
751 |
+
749 2
|
752 |
+
750 2
|
753 |
+
751 0
|
754 |
+
752 2
|
755 |
+
753 1
|
756 |
+
754 2
|
757 |
+
755 1
|
758 |
+
756 0
|
759 |
+
757 0
|
760 |
+
758 0
|
761 |
+
759 0
|
762 |
+
760 1
|
763 |
+
761 1
|
764 |
+
762 1
|
765 |
+
763 1
|
766 |
+
764 1
|
767 |
+
765 1
|
768 |
+
766 1
|
769 |
+
767 0
|
770 |
+
768 1
|
771 |
+
769 2
|
772 |
+
770 0
|
773 |
+
771 1
|
774 |
+
772 1
|
775 |
+
773 0
|
776 |
+
774 0
|
777 |
+
775 0
|
778 |
+
776 2
|
779 |
+
777 0
|
780 |
+
778 0
|
781 |
+
779 1
|
782 |
+
780 1
|
783 |
+
781 2
|
784 |
+
782 2
|
785 |
+
783 2
|
786 |
+
784 1
|
787 |
+
785 1
|
788 |
+
786 0
|
789 |
+
787 0
|
790 |
+
788 0
|
791 |
+
789 2
|
792 |
+
790 2
|
793 |
+
791 0
|
794 |
+
792 2
|
795 |
+
793 2
|
796 |
+
794 2
|
797 |
+
795 1
|
798 |
+
796 1
|
799 |
+
797 0
|
800 |
+
798 0
|
801 |
+
799 2
|
802 |
+
800 2
|
803 |
+
801 2
|
804 |
+
802 2
|
805 |
+
803 1
|
806 |
+
804 0
|
807 |
+
805 2
|
808 |
+
806 2
|
809 |
+
807 2
|
810 |
+
808 2
|
811 |
+
809 0
|
812 |
+
810 0
|
813 |
+
811 2
|
814 |
+
812 2
|
815 |
+
813 2
|
816 |
+
814 0
|
817 |
+
815 2
|
818 |
+
816 2
|
819 |
+
817 1
|
820 |
+
818 1
|
821 |
+
819 0
|
822 |
+
820 1
|
823 |
+
821 0
|
824 |
+
822 1
|
825 |
+
823 2
|
826 |
+
824 2
|
827 |
+
825 1
|
828 |
+
826 1
|
829 |
+
827 0
|
830 |
+
828 0
|
831 |
+
829 1
|
832 |
+
830 0
|
833 |
+
831 1
|
834 |
+
832 1
|
835 |
+
833 2
|
836 |
+
834 1
|
837 |
+
835 1
|
838 |
+
836 1
|
839 |
+
837 2
|
840 |
+
838 2
|
841 |
+
839 0
|
842 |
+
840 0
|
843 |
+
841 0
|
844 |
+
842 1
|
845 |
+
843 0
|
846 |
+
844 2
|
847 |
+
845 0
|
848 |
+
846 1
|
849 |
+
847 2
|
850 |
+
848 1
|
851 |
+
849 0
|
852 |
+
850 2
|
853 |
+
851 0
|
854 |
+
852 1
|
855 |
+
853 1
|
856 |
+
854 2
|
857 |
+
855 2
|
858 |
+
856 2
|
859 |
+
857 1
|
860 |
+
858 0
|
861 |
+
859 1
|
862 |
+
860 0
|
863 |
+
861 1
|
864 |
+
862 1
|
865 |
+
863 0
|
866 |
+
864 1
|
867 |
+
865 1
|
868 |
+
866 2
|
869 |
+
867 2
|
870 |
+
868 0
|
871 |
+
869 1
|
872 |
+
870 2
|
873 |
+
871 2
|
874 |
+
872 2
|
875 |
+
873 0
|
876 |
+
874 1
|
877 |
+
875 1
|
878 |
+
876 2
|
879 |
+
877 2
|
880 |
+
878 0
|
881 |
+
879 1
|
882 |
+
880 0
|
883 |
+
881 1
|
884 |
+
882 0
|
885 |
+
883 2
|
886 |
+
884 2
|
887 |
+
885 0
|
888 |
+
886 0
|
889 |
+
887 2
|
890 |
+
888 0
|
891 |
+
889 0
|
892 |
+
890 0
|
893 |
+
891 2
|
894 |
+
892 1
|
895 |
+
893 2
|
896 |
+
894 0
|
897 |
+
895 0
|
898 |
+
896 0
|
899 |
+
897 0
|
900 |
+
898 0
|
901 |
+
899 2
|
902 |
+
900 1
|
903 |
+
901 1
|
904 |
+
902 0
|
905 |
+
903 2
|
906 |
+
904 1
|
907 |
+
905 1
|
908 |
+
906 2
|
909 |
+
907 0
|
910 |
+
908 1
|
911 |
+
909 1
|
912 |
+
910 2
|
913 |
+
911 1
|
914 |
+
912 2
|
915 |
+
913 1
|
916 |
+
914 0
|
917 |
+
915 1
|
918 |
+
916 2
|
919 |
+
917 1
|
920 |
+
918 0
|
921 |
+
919 0
|
922 |
+
920 2
|
923 |
+
921 1
|
924 |
+
922 0
|
925 |
+
923 0
|
926 |
+
924 2
|
927 |
+
925 2
|
928 |
+
926 1
|
929 |
+
927 2
|
930 |
+
928 1
|
931 |
+
929 1
|
932 |
+
930 2
|
933 |
+
931 2
|
934 |
+
932 0
|
935 |
+
933 0
|
936 |
+
934 1
|
937 |
+
935 0
|
938 |
+
936 0
|
939 |
+
937 1
|
940 |
+
938 0
|
941 |
+
939 0
|
942 |
+
940 0
|
943 |
+
941 1
|
944 |
+
942 2
|
945 |
+
943 1
|
946 |
+
944 2
|
947 |
+
945 2
|
948 |
+
946 0
|
949 |
+
947 0
|
950 |
+
948 2
|
951 |
+
949 0
|
952 |
+
950 0
|
953 |
+
951 2
|
954 |
+
952 0
|
955 |
+
953 2
|
956 |
+
954 2
|
957 |
+
955 1
|
958 |
+
956 0
|
959 |
+
957 1
|
960 |
+
958 0
|
961 |
+
959 1
|
962 |
+
960 2
|
963 |
+
961 1
|
964 |
+
962 2
|
965 |
+
963 0
|
966 |
+
964 1
|
967 |
+
965 0
|
968 |
+
966 1
|
969 |
+
967 1
|
970 |
+
968 2
|
971 |
+
969 0
|
972 |
+
970 2
|
973 |
+
971 1
|
974 |
+
972 2
|
975 |
+
973 0
|
976 |
+
974 1
|
977 |
+
975 0
|
978 |
+
976 2
|
979 |
+
977 2
|
980 |
+
978 2
|
981 |
+
979 0
|
982 |
+
980 1
|
983 |
+
981 0
|
984 |
+
982 0
|
985 |
+
983 1
|
986 |
+
984 0
|
987 |
+
985 0
|
988 |
+
986 2
|
989 |
+
987 2
|
990 |
+
988 1
|
991 |
+
989 1
|
992 |
+
990 1
|
993 |
+
991 2
|
994 |
+
992 0
|
995 |
+
993 0
|
996 |
+
994 0
|
997 |
+
995 0
|
998 |
+
996 0
|
999 |
+
997 1
|
1000 |
+
998 0
|
1001 |
+
999 1
|
1002 |
+
1000 2
|
1003 |
+
1001 2
|
1004 |
+
1002 2
|
1005 |
+
1003 1
|
1006 |
+
1004 2
|
1007 |
+
1005 2
|
1008 |
+
1006 2
|
1009 |
+
1007 1
|
1010 |
+
1008 1
|
1011 |
+
1009 1
|
1012 |
+
1010 1
|
1013 |
+
1011 0
|
1014 |
+
1012 0
|
1015 |
+
1013 2
|
1016 |
+
1014 1
|
1017 |
+
1015 1
|
1018 |
+
1016 0
|
1019 |
+
1017 2
|
1020 |
+
1018 2
|
1021 |
+
1019 2
|
1022 |
+
1020 2
|
1023 |
+
1021 1
|
1024 |
+
1022 0
|
1025 |
+
1023 0
|
1026 |
+
1024 0
|
1027 |
+
1025 1
|
1028 |
+
1026 1
|
1029 |
+
1027 2
|
1030 |
+
1028 1
|
1031 |
+
1029 2
|
1032 |
+
1030 2
|
1033 |
+
1031 2
|
1034 |
+
1032 2
|
1035 |
+
1033 1
|
1036 |
+
1034 2
|
1037 |
+
1035 0
|
1038 |
+
1036 0
|
1039 |
+
1037 2
|
1040 |
+
1038 0
|
1041 |
+
1039 1
|
1042 |
+
1040 2
|
1043 |
+
1041 1
|
1044 |
+
1042 1
|
1045 |
+
1043 1
|
1046 |
+
1044 0
|
1047 |
+
1045 2
|
1048 |
+
1046 0
|
1049 |
+
1047 0
|
1050 |
+
1048 2
|
1051 |
+
1049 2
|
1052 |
+
1050 1
|
1053 |
+
1051 0
|
1054 |
+
1052 1
|
1055 |
+
1053 1
|
1056 |
+
1054 1
|
1057 |
+
1055 1
|
1058 |
+
1056 2
|
1059 |
+
1057 1
|
1060 |
+
1058 2
|
1061 |
+
1059 0
|
1062 |
+
1060 0
|
1063 |
+
1061 2
|
1064 |
+
1062 0
|
1065 |
+
1063 1
|
1066 |
+
1064 1
|
1067 |
+
1065 1
|
1068 |
+
1066 0
|
1069 |
+
1067 2
|
1070 |
+
1068 1
|
1071 |
+
1069 1
|
1072 |
+
1070 1
|
1073 |
+
1071 0
|
1074 |
+
1072 1
|
1075 |
+
1073 2
|
1076 |
+
1074 2
|
1077 |
+
1075 2
|
1078 |
+
1076 2
|
1079 |
+
1077 1
|
1080 |
+
1078 1
|
1081 |
+
1079 0
|
1082 |
+
1080 0
|
1083 |
+
1081 1
|
1084 |
+
1082 1
|
1085 |
+
1083 2
|
1086 |
+
1084 1
|
1087 |
+
1085 1
|
1088 |
+
1086 1
|
1089 |
+
1087 0
|
1090 |
+
1088 1
|
1091 |
+
1089 2
|
1092 |
+
1090 2
|
1093 |
+
1091 1
|
1094 |
+
1092 0
|
1095 |
+
1093 1
|
1096 |
+
1094 2
|
1097 |
+
1095 2
|
1098 |
+
1096 1
|
1099 |
+
1097 0
|
1100 |
+
1098 1
|
1101 |
+
1099 2
|
1102 |
+
1100 2
|
1103 |
+
1101 0
|
1104 |
+
1102 1
|
1105 |
+
1103 0
|
1106 |
+
1104 0
|
1107 |
+
1105 1
|
1108 |
+
1106 2
|
1109 |
+
1107 1
|
1110 |
+
1108 0
|
1111 |
+
1109 0
|
1112 |
+
1110 1
|
1113 |
+
1111 1
|
1114 |
+
1112 0
|
1115 |
+
1113 2
|
1116 |
+
1114 0
|
1117 |
+
1115 0
|
1118 |
+
1116 0
|
1119 |
+
1117 1
|
1120 |
+
1118 0
|
1121 |
+
1119 0
|
1122 |
+
1120 2
|
1123 |
+
1121 2
|
1124 |
+
1122 1
|
1125 |
+
1123 1
|
1126 |
+
1124 0
|
1127 |
+
1125 1
|
1128 |
+
1126 2
|
1129 |
+
1127 2
|
1130 |
+
1128 2
|
1131 |
+
1129 1
|
1132 |
+
1130 0
|
1133 |
+
1131 0
|
1134 |
+
1132 1
|
1135 |
+
1133 1
|
1136 |
+
1134 0
|
1137 |
+
1135 1
|
1138 |
+
1136 1
|
1139 |
+
1137 2
|
1140 |
+
1138 2
|
1141 |
+
1139 0
|
1142 |
+
1140 0
|
1143 |
+
1141 2
|
1144 |
+
1142 2
|
1145 |
+
1143 0
|
1146 |
+
1144 1
|
1147 |
+
1145 1
|
1148 |
+
1146 1
|
1149 |
+
1147 2
|
1150 |
+
1148 1
|
1151 |
+
1149 0
|
1152 |
+
1150 0
|
1153 |
+
1151 1
|
1154 |
+
1152 1
|
1155 |
+
1153 1
|
1156 |
+
1154 1
|
1157 |
+
1155 1
|
1158 |
+
1156 0
|
1159 |
+
1157 1
|
1160 |
+
1158 0
|
1161 |
+
1159 2
|
1162 |
+
1160 0
|
1163 |
+
1161 1
|
1164 |
+
1162 2
|
1165 |
+
1163 1
|
1166 |
+
1164 1
|
1167 |
+
1165 2
|
1168 |
+
1166 2
|
1169 |
+
1167 1
|
1170 |
+
1168 1
|
1171 |
+
1169 1
|
1172 |
+
1170 1
|
1173 |
+
1171 1
|
1174 |
+
1172 2
|
1175 |
+
1173 1
|
1176 |
+
1174 0
|
1177 |
+
1175 0
|
1178 |
+
1176 0
|
1179 |
+
1177 2
|
1180 |
+
1178 0
|
1181 |
+
1179 2
|
1182 |
+
1180 0
|
1183 |
+
1181 0
|
1184 |
+
1182 2
|
1185 |
+
1183 2
|
1186 |
+
1184 1
|
1187 |
+
1185 0
|
1188 |
+
1186 0
|
1189 |
+
1187 0
|
1190 |
+
1188 2
|
1191 |
+
1189 1
|
1192 |
+
1190 0
|
1193 |
+
1191 0
|
1194 |
+
1192 0
|
1195 |
+
1193 1
|
1196 |
+
1194 2
|
1197 |
+
1195 0
|
1198 |
+
1196 0
|
1199 |
+
1197 1
|
1200 |
+
1198 1
|
1201 |
+
1199 2
|
1202 |
+
1200 1
|
1203 |
+
1201 2
|
1204 |
+
1202 0
|
1205 |
+
1203 2
|
1206 |
+
1204 0
|
1207 |
+
1205 2
|
1208 |
+
1206 0
|
1209 |
+
1207 1
|
1210 |
+
1208 0
|
1211 |
+
1209 1
|
1212 |
+
1210 0
|
1213 |
+
1211 2
|
1214 |
+
1212 2
|
1215 |
+
1213 1
|
1216 |
+
1214 1
|
1217 |
+
1215 2
|
1218 |
+
1216 0
|
1219 |
+
1217 1
|
1220 |
+
1218 1
|
1221 |
+
1219 2
|
1222 |
+
1220 1
|
1223 |
+
1221 1
|
1224 |
+
1222 1
|
1225 |
+
1223 2
|
1226 |
+
1224 0
|
1227 |
+
1225 2
|
1228 |
+
1226 1
|
1229 |
+
1227 2
|
1230 |
+
1228 1
|
1231 |
+
1229 0
|
1232 |
+
1230 2
|
1233 |
+
1231 1
|
1234 |
+
1232 0
|
1235 |
+
1233 1
|
1236 |
+
1234 0
|
1237 |
+
1235 1
|
1238 |
+
1236 2
|
1239 |
+
1237 1
|
1240 |
+
1238 1
|
1241 |
+
1239 1
|
1242 |
+
1240 1
|
1243 |
+
1241 1
|
1244 |
+
1242 2
|
1245 |
+
1243 1
|
1246 |
+
1244 1
|
1247 |
+
1245 1
|
1248 |
+
1246 2
|
1249 |
+
1247 0
|
1250 |
+
1248 0
|
1251 |
+
1249 2
|
1252 |
+
1250 0
|
1253 |
+
1251 0
|
1254 |
+
1252 0
|
1255 |
+
1253 0
|
1256 |
+
1254 2
|
1257 |
+
1255 2
|
1258 |
+
1256 0
|
1259 |
+
1257 1
|
1260 |
+
1258 1
|
1261 |
+
1259 1
|
1262 |
+
1260 2
|
1263 |
+
1261 1
|
1264 |
+
1262 2
|
1265 |
+
1263 1
|
1266 |
+
1264 2
|
1267 |
+
1265 0
|
1268 |
+
1266 1
|
1269 |
+
1267 0
|
1270 |
+
1268 1
|
1271 |
+
1269 0
|
1272 |
+
1270 0
|
1273 |
+
1271 0
|
1274 |
+
1272 0
|
1275 |
+
1273 1
|
1276 |
+
1274 1
|
1277 |
+
1275 0
|
1278 |
+
1276 0
|
1279 |
+
1277 0
|
1280 |
+
1278 0
|
1281 |
+
1279 0
|
1282 |
+
1280 0
|
1283 |
+
1281 2
|
1284 |
+
1282 0
|
1285 |
+
1283 1
|
1286 |
+
1284 1
|
1287 |
+
1285 1
|
1288 |
+
1286 1
|
1289 |
+
1287 0
|
1290 |
+
1288 0
|
1291 |
+
1289 0
|
1292 |
+
1290 1
|
1293 |
+
1291 0
|
1294 |
+
1292 0
|
1295 |
+
1293 0
|
1296 |
+
1294 0
|
1297 |
+
1295 0
|
1298 |
+
1296 2
|
1299 |
+
1297 1
|
1300 |
+
1298 0
|
1301 |
+
1299 0
|
1302 |
+
1300 1
|
1303 |
+
1301 1
|
1304 |
+
1302 1
|
1305 |
+
1303 1
|
1306 |
+
1304 1
|
1307 |
+
1305 2
|
1308 |
+
1306 2
|
1309 |
+
1307 1
|
1310 |
+
1308 0
|
1311 |
+
1309 2
|
1312 |
+
1310 1
|
1313 |
+
1311 0
|
1314 |
+
1312 0
|
1315 |
+
1313 1
|
1316 |
+
1314 1
|
1317 |
+
1315 1
|
1318 |
+
1316 1
|
1319 |
+
1317 1
|
1320 |
+
1318 1
|
1321 |
+
1319 1
|
1322 |
+
1320 0
|
1323 |
+
1321 2
|
1324 |
+
1322 0
|
1325 |
+
1323 1
|
1326 |
+
1324 0
|
1327 |
+
1325 2
|
1328 |
+
1326 2
|
1329 |
+
1327 1
|
1330 |
+
1328 2
|
1331 |
+
1329 2
|
1332 |
+
1330 0
|
1333 |
+
1331 0
|
1334 |
+
1332 0
|
1335 |
+
1333 2
|
1336 |
+
1334 1
|
1337 |
+
1335 0
|
1338 |
+
1336 0
|
1339 |
+
1337 1
|
1340 |
+
1338 0
|
1341 |
+
1339 1
|
1342 |
+
1340 1
|
1343 |
+
1341 0
|
1344 |
+
1342 1
|
1345 |
+
1343 1
|
1346 |
+
1344 1
|
1347 |
+
1345 1
|
1348 |
+
1346 2
|
1349 |
+
1347 2
|
1350 |
+
1348 2
|
1351 |
+
1349 1
|
1352 |
+
1350 1
|
1353 |
+
1351 0
|
1354 |
+
1352 1
|
1355 |
+
1353 1
|
1356 |
+
1354 2
|
1357 |
+
1355 1
|
1358 |
+
1356 0
|
1359 |
+
1357 0
|
1360 |
+
1358 1
|
1361 |
+
1359 1
|
1362 |
+
1360 1
|
1363 |
+
1361 0
|
1364 |
+
1362 1
|
1365 |
+
1363 1
|
1366 |
+
1364 1
|
1367 |
+
1365 2
|
1368 |
+
1366 0
|
1369 |
+
1367 0
|
1370 |
+
1368 2
|
1371 |
+
1369 0
|
1372 |
+
1370 2
|
1373 |
+
1371 1
|
1374 |
+
1372 2
|
1375 |
+
1373 0
|
1376 |
+
1374 0
|
1377 |
+
1375 1
|
1378 |
+
1376 1
|
1379 |
+
1377 0
|
1380 |
+
1378 2
|
1381 |
+
1379 0
|
1382 |
+
1380 2
|
1383 |
+
1381 0
|
1384 |
+
1382 2
|
1385 |
+
1383 2
|
1386 |
+
1384 0
|
1387 |
+
1385 0
|
1388 |
+
1386 0
|
1389 |
+
1387 1
|
1390 |
+
1388 2
|
1391 |
+
1389 2
|
1392 |
+
1390 0
|
1393 |
+
1391 1
|
1394 |
+
1392 0
|
1395 |
+
1393 1
|
1396 |
+
1394 1
|
1397 |
+
1395 1
|
1398 |
+
1396 1
|
1399 |
+
1397 0
|
1400 |
+
1398 0
|
1401 |
+
1399 0
|
1402 |
+
1400 0
|
1403 |
+
1401 2
|
1404 |
+
1402 0
|
1405 |
+
1403 2
|
1406 |
+
1404 2
|
1407 |
+
1405 2
|
1408 |
+
1406 1
|
1409 |
+
1407 1
|
1410 |
+
1408 1
|
1411 |
+
1409 0
|
1412 |
+
1410 2
|
1413 |
+
1411 2
|
1414 |
+
1412 1
|
1415 |
+
1413 0
|
1416 |
+
1414 1
|
1417 |
+
1415 0
|
1418 |
+
1416 0
|
1419 |
+
1417 1
|
1420 |
+
1418 1
|
1421 |
+
1419 0
|
1422 |
+
1420 1
|
1423 |
+
1421 2
|
1424 |
+
1422 1
|
1425 |
+
1423 1
|
1426 |
+
1424 2
|
1427 |
+
1425 0
|
1428 |
+
1426 0
|
1429 |
+
1427 0
|
1430 |
+
1428 2
|
1431 |
+
1429 2
|
1432 |
+
1430 0
|
1433 |
+
1431 1
|
1434 |
+
1432 0
|
1435 |
+
1433 2
|
1436 |
+
1434 0
|
1437 |
+
1435 2
|
1438 |
+
1436 2
|
1439 |
+
1437 0
|
1440 |
+
1438 0
|
1441 |
+
1439 1
|
1442 |
+
1440 1
|
1443 |
+
1441 2
|
1444 |
+
1442 1
|
1445 |
+
1443 2
|
1446 |
+
1444 1
|
1447 |
+
1445 2
|
1448 |
+
1446 1
|
1449 |
+
1447 2
|
1450 |
+
1448 0
|
1451 |
+
1449 1
|
1452 |
+
1450 1
|
1453 |
+
1451 2
|
1454 |
+
1452 1
|
1455 |
+
1453 0
|
1456 |
+
1454 2
|
1457 |
+
1455 0
|
1458 |
+
1456 2
|
1459 |
+
1457 1
|
1460 |
+
1458 1
|
1461 |
+
1459 1
|
1462 |
+
1460 0
|
1463 |
+
1461 1
|
1464 |
+
1462 2
|
1465 |
+
1463 0
|
1466 |
+
1464 0
|
1467 |
+
1465 2
|
1468 |
+
1466 0
|
1469 |
+
1467 1
|
1470 |
+
1468 1
|
1471 |
+
1469 1
|
1472 |
+
1470 0
|
1473 |
+
1471 0
|
1474 |
+
1472 1
|
1475 |
+
1473 1
|
1476 |
+
1474 0
|
1477 |
+
1475 2
|
1478 |
+
1476 2
|
1479 |
+
1477 0
|
1480 |
+
1478 1
|
1481 |
+
1479 0
|
1482 |
+
1480 0
|
1483 |
+
1481 1
|
1484 |
+
1482 0
|
1485 |
+
1483 0
|
1486 |
+
1484 1
|
1487 |
+
1485 0
|
1488 |
+
1486 2
|
1489 |
+
1487 2
|
1490 |
+
1488 1
|
1491 |
+
1489 0
|
1492 |
+
1490 1
|
1493 |
+
1491 1
|
1494 |
+
1492 1
|
1495 |
+
1493 1
|
1496 |
+
1494 0
|
1497 |
+
1495 0
|
1498 |
+
1496 2
|
1499 |
+
1497 2
|
1500 |
+
1498 0
|
1501 |
+
1499 1
|
1502 |
+
1500 1
|
1503 |
+
1501 0
|
1504 |
+
1502 0
|
1505 |
+
1503 0
|
1506 |
+
1504 0
|
1507 |
+
1505 2
|
1508 |
+
1506 0
|
1509 |
+
1507 0
|
1510 |
+
1508 0
|
1511 |
+
1509 0
|
1512 |
+
1510 1
|
1513 |
+
1511 2
|
1514 |
+
1512 1
|
1515 |
+
1513 1
|
1516 |
+
1514 2
|
1517 |
+
1515 0
|
1518 |
+
1516 0
|
1519 |
+
1517 2
|
1520 |
+
1518 1
|
1521 |
+
1519 0
|
1522 |
+
1520 1
|
1523 |
+
1521 2
|
1524 |
+
1522 1
|
1525 |
+
1523 2
|
1526 |
+
1524 1
|
1527 |
+
1525 0
|
1528 |
+
1526 2
|
1529 |
+
1527 2
|
1530 |
+
1528 0
|
1531 |
+
1529 1
|
1532 |
+
1530 1
|
1533 |
+
1531 0
|
1534 |
+
1532 2
|
1535 |
+
1533 0
|
1536 |
+
1534 2
|
1537 |
+
1535 0
|
1538 |
+
1536 1
|
1539 |
+
1537 0
|
1540 |
+
1538 1
|
1541 |
+
1539 1
|
1542 |
+
1540 1
|
1543 |
+
1541 0
|
1544 |
+
1542 0
|
1545 |
+
1543 0
|
1546 |
+
1544 0
|
1547 |
+
1545 1
|
1548 |
+
1546 2
|
1549 |
+
1547 0
|
1550 |
+
1548 2
|
1551 |
+
1549 2
|
1552 |
+
1550 1
|
1553 |
+
1551 1
|
1554 |
+
1552 2
|
1555 |
+
1553 1
|
1556 |
+
1554 2
|
1557 |
+
1555 1
|
1558 |
+
1556 1
|
1559 |
+
1557 1
|
1560 |
+
1558 2
|
1561 |
+
1559 1
|
1562 |
+
1560 1
|
1563 |
+
1561 0
|
1564 |
+
1562 0
|
1565 |
+
1563 0
|
1566 |
+
1564 1
|
1567 |
+
1565 2
|
1568 |
+
1566 1
|
1569 |
+
1567 2
|
1570 |
+
1568 1
|
1571 |
+
1569 1
|
1572 |
+
1570 2
|
1573 |
+
1571 1
|
1574 |
+
1572 1
|
1575 |
+
1573 1
|
1576 |
+
1574 2
|
1577 |
+
1575 2
|
1578 |
+
1576 2
|
1579 |
+
1577 1
|
1580 |
+
1578 0
|
1581 |
+
1579 0
|
1582 |
+
1580 2
|
1583 |
+
1581 1
|
1584 |
+
1582 1
|
1585 |
+
1583 0
|
1586 |
+
1584 1
|
1587 |
+
1585 1
|
1588 |
+
1586 1
|
1589 |
+
1587 1
|
1590 |
+
1588 0
|
1591 |
+
1589 1
|
1592 |
+
1590 1
|
1593 |
+
1591 0
|
1594 |
+
1592 1
|
1595 |
+
1593 0
|
1596 |
+
1594 1
|
1597 |
+
1595 1
|
1598 |
+
1596 2
|
1599 |
+
1597 0
|
1600 |
+
1598 0
|
1601 |
+
1599 1
|
1602 |
+
1600 0
|
1603 |
+
1601 0
|
1604 |
+
1602 1
|
1605 |
+
1603 2
|
1606 |
+
1604 2
|
1607 |
+
1605 1
|
1608 |
+
1606 0
|
1609 |
+
1607 1
|
1610 |
+
1608 0
|
1611 |
+
1609 1
|
1612 |
+
1610 0
|
1613 |
+
1611 1
|
1614 |
+
1612 1
|
1615 |
+
1613 1
|
1616 |
+
1614 2
|
1617 |
+
1615 0
|
1618 |
+
1616 2
|
1619 |
+
1617 0
|
1620 |
+
1618 0
|
1621 |
+
1619 1
|
1622 |
+
1620 0
|
1623 |
+
1621 1
|
1624 |
+
1622 0
|
1625 |
+
1623 0
|
1626 |
+
1624 0
|
1627 |
+
1625 0
|
1628 |
+
1626 0
|
1629 |
+
1627 2
|
1630 |
+
1628 0
|
1631 |
+
1629 1
|
1632 |
+
1630 1
|
1633 |
+
1631 1
|
1634 |
+
1632 2
|
1635 |
+
1633 0
|
1636 |
+
1634 2
|
1637 |
+
1635 1
|
1638 |
+
1636 0
|
1639 |
+
1637 1
|
1640 |
+
1638 0
|
1641 |
+
1639 0
|
1642 |
+
1640 0
|
1643 |
+
1641 1
|
1644 |
+
1642 0
|
1645 |
+
1643 0
|
1646 |
+
1644 1
|
1647 |
+
1645 0
|
1648 |
+
1646 0
|
1649 |
+
1647 0
|
1650 |
+
1648 1
|
1651 |
+
1649 1
|
1652 |
+
1650 0
|
1653 |
+
1651 0
|
1654 |
+
1652 1
|
1655 |
+
1653 1
|
1656 |
+
1654 1
|
1657 |
+
1655 2
|
1658 |
+
1656 2
|
1659 |
+
1657 2
|
1660 |
+
1658 1
|
1661 |
+
1659 1
|
1662 |
+
1660 0
|
1663 |
+
1661 1
|
1664 |
+
1662 2
|
1665 |
+
1663 2
|
1666 |
+
1664 2
|
1667 |
+
1665 1
|
1668 |
+
1666 0
|
1669 |
+
1667 1
|
1670 |
+
1668 1
|
1671 |
+
1669 2
|
1672 |
+
1670 2
|
1673 |
+
1671 1
|
1674 |
+
1672 0
|
1675 |
+
1673 1
|
1676 |
+
1674 0
|
1677 |
+
1675 0
|
1678 |
+
1676 1
|
1679 |
+
1677 2
|
1680 |
+
1678 2
|
1681 |
+
1679 0
|
1682 |
+
1680 0
|
1683 |
+
1681 0
|
1684 |
+
1682 1
|
1685 |
+
1683 0
|
1686 |
+
1684 2
|
1687 |
+
1685 0
|
1688 |
+
1686 1
|
1689 |
+
1687 0
|
1690 |
+
1688 1
|
1691 |
+
1689 0
|
1692 |
+
1690 0
|
1693 |
+
1691 2
|
1694 |
+
1692 0
|
1695 |
+
1693 1
|
1696 |
+
1694 1
|
1697 |
+
1695 0
|
1698 |
+
1696 1
|
1699 |
+
1697 1
|
1700 |
+
1698 1
|
1701 |
+
1699 1
|
1702 |
+
1700 1
|
1703 |
+
1701 1
|
1704 |
+
1702 2
|
1705 |
+
1703 0
|
1706 |
+
1704 1
|
1707 |
+
1705 2
|
1708 |
+
1706 0
|
1709 |
+
1707 1
|
1710 |
+
1708 1
|
1711 |
+
1709 2
|
1712 |
+
1710 1
|
1713 |
+
1711 1
|
1714 |
+
1712 1
|
1715 |
+
1713 2
|
1716 |
+
1714 0
|
1717 |
+
1715 1
|
1718 |
+
1716 0
|
1719 |
+
1717 0
|
1720 |
+
1718 2
|
1721 |
+
1719 2
|
1722 |
+
1720 2
|
1723 |
+
1721 2
|
1724 |
+
1722 2
|
1725 |
+
1723 1
|
1726 |
+
1724 2
|
1727 |
+
1725 0
|
1728 |
+
1726 0
|
1729 |
+
1727 2
|
1730 |
+
1728 1
|
1731 |
+
1729 0
|
1732 |
+
1730 1
|
1733 |
+
1731 0
|
1734 |
+
1732 1
|
1735 |
+
1733 2
|
1736 |
+
1734 0
|
1737 |
+
1735 2
|
1738 |
+
1736 2
|
1739 |
+
1737 2
|
1740 |
+
1738 0
|
1741 |
+
1739 1
|
1742 |
+
1740 1
|
1743 |
+
1741 1
|
1744 |
+
1742 1
|
1745 |
+
1743 1
|
1746 |
+
1744 0
|
1747 |
+
1745 1
|
1748 |
+
1746 1
|
1749 |
+
1747 2
|
1750 |
+
1748 1
|
1751 |
+
1749 0
|
1752 |
+
1750 1
|
1753 |
+
1751 2
|
1754 |
+
1752 2
|
1755 |
+
1753 1
|
1756 |
+
1754 0
|
1757 |
+
1755 2
|
1758 |
+
1756 1
|
1759 |
+
1757 2
|
1760 |
+
1758 2
|
1761 |
+
1759 2
|
1762 |
+
1760 0
|
1763 |
+
1761 1
|
1764 |
+
1762 1
|
1765 |
+
1763 2
|
1766 |
+
1764 0
|
1767 |
+
1765 0
|
1768 |
+
1766 1
|
1769 |
+
1767 0
|
1770 |
+
1768 1
|
1771 |
+
1769 2
|
1772 |
+
1770 2
|
1773 |
+
1771 2
|
1774 |
+
1772 1
|
1775 |
+
1773 2
|
1776 |
+
1774 0
|
1777 |
+
1775 2
|
1778 |
+
1776 1
|
1779 |
+
1777 1
|
1780 |
+
1778 0
|
1781 |
+
1779 1
|
1782 |
+
1780 2
|
1783 |
+
1781 0
|
1784 |
+
1782 2
|
1785 |
+
1783 1
|
1786 |
+
1784 1
|
1787 |
+
1785 0
|
1788 |
+
1786 1
|
1789 |
+
1787 1
|
1790 |
+
1788 1
|
1791 |
+
1789 2
|
1792 |
+
1790 1
|
1793 |
+
1791 1
|
1794 |
+
1792 2
|
1795 |
+
1793 0
|
1796 |
+
1794 2
|
1797 |
+
1795 2
|
1798 |
+
1796 1
|
1799 |
+
1797 0
|
1800 |
+
1798 0
|
1801 |
+
1799 1
|
1802 |
+
1800 0
|
1803 |
+
1801 2
|
1804 |
+
1802 1
|
1805 |
+
1803 2
|
1806 |
+
1804 2
|
1807 |
+
1805 1
|
1808 |
+
1806 0
|
1809 |
+
1807 0
|
1810 |
+
1808 1
|
1811 |
+
1809 1
|
1812 |
+
1810 0
|
1813 |
+
1811 2
|
1814 |
+
1812 0
|
1815 |
+
1813 2
|
1816 |
+
1814 2
|
1817 |
+
1815 0
|
1818 |
+
1816 0
|
1819 |
+
1817 2
|
1820 |
+
1818 1
|
1821 |
+
1819 0
|
1822 |
+
1820 0
|
1823 |
+
1821 2
|
1824 |
+
1822 1
|
1825 |
+
1823 0
|
1826 |
+
1824 1
|
1827 |
+
1825 1
|
1828 |
+
1826 1
|
1829 |
+
1827 1
|
1830 |
+
1828 1
|
1831 |
+
1829 0
|
1832 |
+
1830 1
|
1833 |
+
1831 1
|
1834 |
+
1832 1
|
1835 |
+
1833 0
|
1836 |
+
1834 2
|
1837 |
+
1835 1
|
1838 |
+
1836 0
|
1839 |
+
1837 1
|
1840 |
+
1838 1
|
1841 |
+
1839 0
|
1842 |
+
1840 0
|
1843 |
+
1841 1
|
1844 |
+
1842 1
|
1845 |
+
1843 1
|
1846 |
+
1844 0
|
1847 |
+
1845 0
|
1848 |
+
1846 2
|
1849 |
+
1847 2
|
1850 |
+
1848 1
|
1851 |
+
1849 2
|
1852 |
+
1850 0
|
1853 |
+
1851 2
|
1854 |
+
1852 1
|
1855 |
+
1853 1
|
1856 |
+
1854 2
|
1857 |
+
1855 2
|
1858 |
+
1856 0
|
1859 |
+
1857 2
|
1860 |
+
1858 1
|
1861 |
+
1859 1
|
1862 |
+
1860 0
|
1863 |
+
1861 2
|
1864 |
+
1862 1
|
1865 |
+
1863 2
|
1866 |
+
1864 1
|
1867 |
+
1865 0
|
1868 |
+
1866 0
|
1869 |
+
1867 0
|
1870 |
+
1868 2
|
1871 |
+
1869 0
|
1872 |
+
1870 2
|
1873 |
+
1871 2
|
1874 |
+
1872 2
|
1875 |
+
1873 0
|
1876 |
+
1874 1
|
1877 |
+
1875 0
|
1878 |
+
1876 1
|
1879 |
+
1877 0
|
1880 |
+
1878 2
|
1881 |
+
1879 0
|
1882 |
+
1880 2
|
1883 |
+
1881 2
|
1884 |
+
1882 2
|
1885 |
+
1883 1
|
1886 |
+
1884 2
|
1887 |
+
1885 1
|
1888 |
+
1886 0
|
1889 |
+
1887 0
|
1890 |
+
1888 2
|
1891 |
+
1889 1
|
1892 |
+
1890 2
|
1893 |
+
1891 0
|
1894 |
+
1892 1
|
1895 |
+
1893 0
|
1896 |
+
1894 1
|
1897 |
+
1895 0
|
1898 |
+
1896 1
|
1899 |
+
1897 0
|
1900 |
+
1898 1
|
1901 |
+
1899 1
|
1902 |
+
1900 1
|
1903 |
+
1901 2
|
1904 |
+
1902 1
|
1905 |
+
1903 0
|
1906 |
+
1904 0
|
1907 |
+
1905 0
|
1908 |
+
1906 0
|
1909 |
+
1907 0
|
1910 |
+
1908 1
|
1911 |
+
1909 0
|
1912 |
+
1910 2
|
1913 |
+
1911 0
|
1914 |
+
1912 0
|
1915 |
+
1913 0
|
1916 |
+
1914 2
|
1917 |
+
1915 1
|
1918 |
+
1916 0
|
1919 |
+
1917 1
|
1920 |
+
1918 2
|
1921 |
+
1919 2
|
1922 |
+
1920 2
|
1923 |
+
1921 0
|
1924 |
+
1922 1
|
1925 |
+
1923 2
|
1926 |
+
1924 1
|
1927 |
+
1925 2
|
1928 |
+
1926 0
|
1929 |
+
1927 0
|
1930 |
+
1928 0
|
1931 |
+
1929 1
|
1932 |
+
1930 1
|
1933 |
+
1931 2
|
1934 |
+
1932 0
|
1935 |
+
1933 1
|
1936 |
+
1934 2
|
1937 |
+
1935 0
|
1938 |
+
1936 0
|
1939 |
+
1937 1
|
1940 |
+
1938 1
|
1941 |
+
1939 0
|
1942 |
+
1940 1
|
1943 |
+
1941 1
|
1944 |
+
1942 1
|
1945 |
+
1943 2
|
1946 |
+
1944 0
|
1947 |
+
1945 1
|
1948 |
+
1946 0
|
1949 |
+
1947 0
|
1950 |
+
1948 0
|
1951 |
+
1949 2
|
1952 |
+
1950 2
|
1953 |
+
1951 0
|
1954 |
+
1952 0
|
1955 |
+
1953 1
|
1956 |
+
1954 1
|
1957 |
+
1955 1
|
1958 |
+
1956 2
|
1959 |
+
1957 2
|
1960 |
+
1958 2
|
1961 |
+
1959 1
|
1962 |
+
1960 1
|
1963 |
+
1961 1
|
1964 |
+
1962 1
|
1965 |
+
1963 2
|
1966 |
+
1964 2
|
1967 |
+
1965 2
|
1968 |
+
1966 1
|
1969 |
+
1967 2
|
1970 |
+
1968 2
|
1971 |
+
1969 1
|
1972 |
+
1970 1
|
1973 |
+
1971 1
|
1974 |
+
1972 2
|
1975 |
+
1973 0
|
1976 |
+
1974 2
|
1977 |
+
1975 1
|
1978 |
+
1976 2
|
1979 |
+
1977 0
|
1980 |
+
1978 2
|
1981 |
+
1979 1
|
1982 |
+
1980 1
|
1983 |
+
1981 1
|
1984 |
+
1982 1
|
1985 |
+
1983 1
|
1986 |
+
1984 1
|
1987 |
+
1985 2
|
1988 |
+
1986 1
|
1989 |
+
1987 1
|
1990 |
+
1988 0
|
1991 |
+
1989 0
|
1992 |
+
1990 2
|
1993 |
+
1991 2
|
1994 |
+
1992 0
|
1995 |
+
1993 1
|
1996 |
+
1994 1
|
1997 |
+
1995 1
|
1998 |
+
1996 0
|
1999 |
+
1997 0
|
2000 |
+
1998 1
|
2001 |
+
1999 2
|
2002 |
+
2000 0
|
2003 |
+
2001 0
|
2004 |
+
2002 2
|
2005 |
+
2003 0
|
2006 |
+
2004 2
|
2007 |
+
2005 1
|
2008 |
+
2006 0
|
2009 |
+
2007 1
|
2010 |
+
2008 1
|
2011 |
+
2009 1
|
2012 |
+
2010 1
|
2013 |
+
2011 1
|
2014 |
+
2012 2
|
2015 |
+
2013 0
|
2016 |
+
2014 2
|
2017 |
+
2015 2
|
2018 |
+
2016 1
|
2019 |
+
2017 0
|
2020 |
+
2018 1
|
2021 |
+
2019 0
|
2022 |
+
2020 1
|
2023 |
+
2021 1
|
2024 |
+
2022 1
|
2025 |
+
2023 2
|
2026 |
+
2024 1
|
2027 |
+
2025 0
|
2028 |
+
2026 0
|
2029 |
+
2027 2
|
2030 |
+
2028 0
|
2031 |
+
2029 1
|
2032 |
+
2030 1
|
2033 |
+
2031 1
|
2034 |
+
2032 1
|
2035 |
+
2033 0
|
2036 |
+
2034 0
|
2037 |
+
2035 0
|
2038 |
+
2036 1
|
2039 |
+
2037 2
|
2040 |
+
2038 1
|
2041 |
+
2039 0
|
2042 |
+
2040 1
|
2043 |
+
2041 0
|
2044 |
+
2042 1
|
2045 |
+
2043 2
|
2046 |
+
2044 1
|
2047 |
+
2045 1
|
2048 |
+
2046 1
|
2049 |
+
2047 1
|
2050 |
+
2048 1
|
2051 |
+
2049 1
|
2052 |
+
2050 0
|
2053 |
+
2051 2
|
2054 |
+
2052 1
|
2055 |
+
2053 1
|
2056 |
+
2054 1
|
2057 |
+
2055 0
|
2058 |
+
2056 2
|
2059 |
+
2057 2
|
2060 |
+
2058 1
|
2061 |
+
2059 0
|
2062 |
+
2060 1
|
2063 |
+
2061 0
|
2064 |
+
2062 0
|
2065 |
+
2063 1
|
2066 |
+
2064 2
|
2067 |
+
2065 2
|
2068 |
+
2066 0
|
2069 |
+
2067 1
|
2070 |
+
2068 0
|
2071 |
+
2069 0
|
2072 |
+
2070 2
|
2073 |
+
2071 2
|
2074 |
+
2072 0
|
2075 |
+
2073 1
|
2076 |
+
2074 1
|
2077 |
+
2075 0
|
2078 |
+
2076 0
|
2079 |
+
2077 1
|
2080 |
+
2078 1
|
2081 |
+
2079 2
|
2082 |
+
2080 0
|
2083 |
+
2081 2
|
2084 |
+
2082 1
|
2085 |
+
2083 2
|
2086 |
+
2084 1
|
2087 |
+
2085 2
|
2088 |
+
2086 0
|
2089 |
+
2087 1
|
2090 |
+
2088 1
|
2091 |
+
2089 2
|
2092 |
+
2090 2
|
2093 |
+
2091 0
|
2094 |
+
2092 0
|
2095 |
+
2093 2
|
2096 |
+
2094 0
|
2097 |
+
2095 1
|
2098 |
+
2096 0
|
2099 |
+
2097 0
|
2100 |
+
2098 2
|
2101 |
+
2099 1
|
2102 |
+
2100 2
|
2103 |
+
2101 0
|
2104 |
+
2102 0
|
2105 |
+
2103 0
|
2106 |
+
2104 0
|
2107 |
+
2105 0
|
2108 |
+
2106 1
|
2109 |
+
2107 1
|
2110 |
+
2108 1
|
2111 |
+
2109 1
|
2112 |
+
2110 2
|
2113 |
+
2111 0
|
2114 |
+
2112 1
|
2115 |
+
2113 1
|
2116 |
+
2114 2
|
2117 |
+
2115 0
|
2118 |
+
2116 2
|
2119 |
+
2117 0
|
2120 |
+
2118 0
|
2121 |
+
2119 0
|
2122 |
+
2120 1
|
2123 |
+
2121 0
|
2124 |
+
2122 2
|
2125 |
+
2123 2
|
2126 |
+
2124 1
|
2127 |
+
2125 2
|
2128 |
+
2126 0
|
2129 |
+
2127 2
|
2130 |
+
2128 1
|
2131 |
+
2129 1
|
2132 |
+
2130 0
|
2133 |
+
2131 2
|
2134 |
+
2132 2
|
2135 |
+
2133 1
|
2136 |
+
2134 1
|
2137 |
+
2135 0
|
2138 |
+
2136 0
|
2139 |
+
2137 0
|
2140 |
+
2138 0
|
2141 |
+
2139 1
|
2142 |
+
2140 2
|
2143 |
+
2141 1
|
2144 |
+
2142 1
|
2145 |
+
2143 1
|
2146 |
+
2144 0
|
2147 |
+
2145 2
|
2148 |
+
2146 1
|
2149 |
+
2147 0
|
2150 |
+
2148 1
|
2151 |
+
2149 2
|
2152 |
+
2150 1
|
2153 |
+
2151 0
|
2154 |
+
2152 1
|
2155 |
+
2153 1
|
2156 |
+
2154 0
|
2157 |
+
2155 2
|
2158 |
+
2156 1
|
2159 |
+
2157 1
|
2160 |
+
2158 0
|
2161 |
+
2159 1
|
2162 |
+
2160 2
|
2163 |
+
2161 2
|
2164 |
+
2162 1
|
2165 |
+
2163 0
|
2166 |
+
2164 0
|
2167 |
+
2165 2
|
2168 |
+
2166 0
|
2169 |
+
2167 1
|
2170 |
+
2168 1
|
2171 |
+
2169 1
|
2172 |
+
2170 2
|
2173 |
+
2171 1
|
2174 |
+
2172 1
|
2175 |
+
2173 1
|
2176 |
+
2174 1
|
2177 |
+
2175 1
|
2178 |
+
2176 0
|
2179 |
+
2177 1
|
2180 |
+
2178 0
|
2181 |
+
2179 0
|
2182 |
+
2180 1
|
2183 |
+
2181 1
|
2184 |
+
2182 1
|
2185 |
+
2183 0
|
2186 |
+
2184 0
|
2187 |
+
2185 1
|
2188 |
+
2186 1
|
2189 |
+
2187 1
|
2190 |
+
2188 0
|
2191 |
+
2189 1
|
2192 |
+
2190 0
|
2193 |
+
2191 1
|
2194 |
+
2192 1
|
2195 |
+
2193 0
|
2196 |
+
2194 1
|
2197 |
+
2195 0
|
2198 |
+
2196 1
|
2199 |
+
2197 1
|
2200 |
+
2198 1
|
2201 |
+
2199 0
|
2202 |
+
2200 0
|
2203 |
+
2201 1
|
2204 |
+
2202 0
|
2205 |
+
2203 1
|
2206 |
+
2204 1
|
2207 |
+
2205 2
|
2208 |
+
2206 2
|
2209 |
+
2207 2
|
2210 |
+
2208 0
|
2211 |
+
2209 0
|
2212 |
+
2210 1
|
2213 |
+
2211 1
|
2214 |
+
2212 0
|
2215 |
+
2213 1
|
2216 |
+
2214 1
|
2217 |
+
2215 0
|
2218 |
+
2216 1
|
2219 |
+
2217 1
|
2220 |
+
2218 0
|
2221 |
+
2219 0
|
2222 |
+
2220 2
|
2223 |
+
2221 2
|
2224 |
+
2222 1
|
2225 |
+
2223 1
|
2226 |
+
2224 0
|
2227 |
+
2225 1
|
2228 |
+
2226 1
|
2229 |
+
2227 0
|
2230 |
+
2228 1
|
2231 |
+
2229 1
|
2232 |
+
2230 0
|
2233 |
+
2231 0
|
2234 |
+
2232 1
|
2235 |
+
2233 0
|
2236 |
+
2234 2
|
2237 |
+
2235 0
|
2238 |
+
2236 2
|
2239 |
+
2237 0
|
2240 |
+
2238 2
|
2241 |
+
2239 0
|
2242 |
+
2240 1
|
2243 |
+
2241 1
|
2244 |
+
2242 0
|
2245 |
+
2243 0
|
2246 |
+
2244 0
|
2247 |
+
2245 0
|
2248 |
+
2246 2
|
2249 |
+
2247 2
|
2250 |
+
2248 0
|
2251 |
+
2249 1
|
2252 |
+
2250 0
|
2253 |
+
2251 2
|
2254 |
+
2252 0
|
2255 |
+
2253 1
|
2256 |
+
2254 0
|
2257 |
+
2255 1
|
2258 |
+
2256 0
|
2259 |
+
2257 0
|
2260 |
+
2258 2
|
2261 |
+
2259 0
|
2262 |
+
2260 2
|
2263 |
+
2261 2
|
2264 |
+
2262 1
|
2265 |
+
2263 1
|
2266 |
+
2264 0
|
2267 |
+
2265 2
|
2268 |
+
2266 2
|
2269 |
+
2267 1
|
2270 |
+
2268 1
|
2271 |
+
2269 0
|
2272 |
+
2270 2
|
2273 |
+
2271 0
|
2274 |
+
2272 0
|
2275 |
+
2273 2
|
2276 |
+
2274 2
|
2277 |
+
2275 1
|
2278 |
+
2276 0
|
2279 |
+
2277 1
|
2280 |
+
2278 1
|
2281 |
+
2279 2
|
2282 |
+
2280 2
|
2283 |
+
2281 0
|
2284 |
+
2282 2
|
2285 |
+
2283 2
|
2286 |
+
2284 0
|
2287 |
+
2285 2
|
2288 |
+
2286 2
|
2289 |
+
2287 1
|
2290 |
+
2288 0
|
2291 |
+
2289 1
|
2292 |
+
2290 0
|
2293 |
+
2291 2
|
2294 |
+
2292 0
|
2295 |
+
2293 2
|
2296 |
+
2294 0
|
2297 |
+
2295 2
|
2298 |
+
2296 1
|
2299 |
+
2297 0
|
2300 |
+
2298 2
|
2301 |
+
2299 1
|
2302 |
+
2300 1
|
2303 |
+
2301 2
|
2304 |
+
2302 0
|
2305 |
+
2303 2
|
2306 |
+
2304 1
|
2307 |
+
2305 2
|
2308 |
+
2306 2
|
2309 |
+
2307 2
|
2310 |
+
2308 2
|
2311 |
+
2309 0
|
2312 |
+
2310 0
|
2313 |
+
2311 1
|
2314 |
+
2312 1
|
2315 |
+
2313 0
|
2316 |
+
2314 0
|
2317 |
+
2315 1
|
2318 |
+
2316 1
|
2319 |
+
2317 0
|
2320 |
+
2318 1
|
2321 |
+
2319 1
|
2322 |
+
2320 1
|
2323 |
+
2321 0
|
2324 |
+
2322 0
|
2325 |
+
2323 1
|
2326 |
+
2324 2
|
2327 |
+
2325 1
|
2328 |
+
2326 0
|
2329 |
+
2327 0
|
2330 |
+
2328 2
|
2331 |
+
2329 1
|
2332 |
+
2330 0
|
2333 |
+
2331 0
|
2334 |
+
2332 0
|
2335 |
+
2333 0
|
2336 |
+
2334 1
|
2337 |
+
2335 1
|
2338 |
+
2336 1
|
2339 |
+
2337 2
|
2340 |
+
2338 1
|
2341 |
+
2339 1
|
2342 |
+
2340 2
|
2343 |
+
2341 1
|
2344 |
+
2342 0
|
2345 |
+
2343 2
|
2346 |
+
2344 1
|
2347 |
+
2345 2
|
2348 |
+
2346 1
|
2349 |
+
2347 0
|
2350 |
+
2348 2
|
2351 |
+
2349 1
|
2352 |
+
2350 0
|
2353 |
+
2351 1
|
2354 |
+
2352 1
|
2355 |
+
2353 1
|
2356 |
+
2354 2
|
2357 |
+
2355 1
|
2358 |
+
2356 0
|
2359 |
+
2357 1
|
2360 |
+
2358 0
|
2361 |
+
2359 0
|
2362 |
+
2360 0
|
2363 |
+
2361 1
|
2364 |
+
2362 0
|
2365 |
+
2363 0
|
2366 |
+
2364 0
|
2367 |
+
2365 0
|
2368 |
+
2366 0
|
2369 |
+
2367 1
|
2370 |
+
2368 2
|
2371 |
+
2369 1
|
2372 |
+
2370 0
|
2373 |
+
2371 0
|
2374 |
+
2372 2
|
2375 |
+
2373 0
|
2376 |
+
2374 0
|
2377 |
+
2375 0
|
2378 |
+
2376 1
|
2379 |
+
2377 2
|
2380 |
+
2378 0
|
2381 |
+
2379 0
|
2382 |
+
2380 2
|
2383 |
+
2381 1
|
2384 |
+
2382 1
|
2385 |
+
2383 0
|
2386 |
+
2384 0
|
2387 |
+
2385 2
|
2388 |
+
2386 2
|
2389 |
+
2387 0
|
2390 |
+
2388 1
|
2391 |
+
2389 0
|
2392 |
+
2390 1
|
2393 |
+
2391 0
|
2394 |
+
2392 0
|
2395 |
+
2393 1
|
2396 |
+
2394 1
|
2397 |
+
2395 2
|
2398 |
+
2396 0
|
2399 |
+
2397 1
|
2400 |
+
2398 0
|
2401 |
+
2399 1
|
2402 |
+
2400 0
|
2403 |
+
2401 2
|
2404 |
+
2402 1
|
2405 |
+
2403 1
|
2406 |
+
2404 2
|
2407 |
+
2405 0
|
2408 |
+
2406 1
|
2409 |
+
2407 2
|
2410 |
+
2408 1
|
2411 |
+
2409 2
|
2412 |
+
2410 1
|
2413 |
+
2411 1
|
2414 |
+
2412 1
|
2415 |
+
2413 0
|
2416 |
+
2414 2
|
2417 |
+
2415 0
|
2418 |
+
2416 2
|
2419 |
+
2417 1
|
2420 |
+
2418 2
|
2421 |
+
2419 0
|
2422 |
+
2420 0
|
2423 |
+
2421 2
|
2424 |
+
2422 2
|
2425 |
+
2423 0
|
2426 |
+
2424 2
|
2427 |
+
2425 1
|
2428 |
+
2426 0
|
2429 |
+
2427 1
|
2430 |
+
2428 0
|
2431 |
+
2429 1
|
2432 |
+
2430 1
|
2433 |
+
2431 1
|
2434 |
+
2432 1
|
2435 |
+
2433 2
|
2436 |
+
2434 0
|
2437 |
+
2435 2
|
2438 |
+
2436 1
|
2439 |
+
2437 2
|
2440 |
+
2438 1
|
2441 |
+
2439 1
|
2442 |
+
2440 0
|
2443 |
+
2441 2
|
2444 |
+
2442 0
|
2445 |
+
2443 2
|
2446 |
+
2444 1
|
2447 |
+
2445 0
|
2448 |
+
2446 1
|
2449 |
+
2447 0
|
2450 |
+
2448 1
|
2451 |
+
2449 1
|
2452 |
+
2450 0
|
2453 |
+
2451 1
|
2454 |
+
2452 1
|
2455 |
+
2453 1
|
2456 |
+
2454 1
|
2457 |
+
2455 1
|
2458 |
+
2456 1
|
2459 |
+
2457 1
|
2460 |
+
2458 1
|
2461 |
+
2459 0
|
2462 |
+
2460 0
|
2463 |
+
2461 1
|
2464 |
+
2462 0
|
2465 |
+
2463 2
|
2466 |
+
2464 2
|
2467 |
+
2465 0
|
2468 |
+
2466 0
|
2469 |
+
2467 0
|
2470 |
+
2468 2
|
2471 |
+
2469 0
|
2472 |
+
2470 0
|
2473 |
+
2471 0
|
2474 |
+
2472 2
|
2475 |
+
2473 2
|
2476 |
+
2474 2
|
2477 |
+
2475 1
|
2478 |
+
2476 2
|
2479 |
+
2477 0
|
2480 |
+
2478 2
|
2481 |
+
2479 1
|
2482 |
+
2480 2
|
2483 |
+
2481 2
|
2484 |
+
2482 1
|
2485 |
+
2483 2
|
2486 |
+
2484 1
|
2487 |
+
2485 0
|
2488 |
+
2486 1
|
2489 |
+
2487 0
|
2490 |
+
2488 0
|
2491 |
+
2489 0
|
2492 |
+
2490 1
|
2493 |
+
2491 0
|
2494 |
+
2492 1
|
2495 |
+
2493 1
|
2496 |
+
2494 2
|
2497 |
+
2495 2
|
2498 |
+
2496 0
|
2499 |
+
2497 2
|
2500 |
+
2498 0
|
2501 |
+
2499 2
|
2502 |
+
2500 0
|
2503 |
+
2501 2
|
2504 |
+
2502 0
|
2505 |
+
2503 2
|
2506 |
+
2504 0
|
2507 |
+
2505 0
|
2508 |
+
2506 1
|
2509 |
+
2507 1
|
2510 |
+
2508 2
|
2511 |
+
2509 1
|
2512 |
+
2510 1
|
2513 |
+
2511 1
|
2514 |
+
2512 0
|
2515 |
+
2513 1
|
2516 |
+
2514 0
|
2517 |
+
2515 1
|
2518 |
+
2516 2
|
2519 |
+
2517 2
|
2520 |
+
2518 1
|
2521 |
+
2519 2
|
2522 |
+
2520 0
|
2523 |
+
2521 0
|
2524 |
+
2522 2
|
2525 |
+
2523 1
|
2526 |
+
2524 2
|
2527 |
+
2525 1
|
2528 |
+
2526 1
|
2529 |
+
2527 0
|
2530 |
+
2528 2
|
2531 |
+
2529 1
|
2532 |
+
2530 0
|
2533 |
+
2531 0
|
2534 |
+
2532 2
|
2535 |
+
2533 1
|
2536 |
+
2534 0
|
2537 |
+
2535 1
|
2538 |
+
2536 2
|
2539 |
+
2537 1
|
2540 |
+
2538 0
|
2541 |
+
2539 2
|
2542 |
+
2540 2
|
2543 |
+
2541 1
|
2544 |
+
2542 0
|
2545 |
+
2543 2
|
2546 |
+
2544 1
|
2547 |
+
2545 0
|
2548 |
+
2546 2
|
2549 |
+
2547 2
|
2550 |
+
2548 2
|
2551 |
+
2549 2
|
2552 |
+
2550 2
|
2553 |
+
2551 1
|
2554 |
+
2552 2
|
2555 |
+
2553 2
|
2556 |
+
2554 0
|
2557 |
+
2555 0
|
2558 |
+
2556 2
|
2559 |
+
2557 1
|
2560 |
+
2558 1
|
2561 |
+
2559 2
|
2562 |
+
2560 2
|
2563 |
+
2561 0
|
2564 |
+
2562 0
|
2565 |
+
2563 1
|
2566 |
+
2564 2
|
2567 |
+
2565 1
|
2568 |
+
2566 2
|
2569 |
+
2567 0
|
2570 |
+
2568 1
|
2571 |
+
2569 0
|
2572 |
+
2570 0
|
2573 |
+
2571 1
|
2574 |
+
2572 2
|
2575 |
+
2573 1
|
2576 |
+
2574 2
|
2577 |
+
2575 1
|
2578 |
+
2576 2
|
2579 |
+
2577 0
|
2580 |
+
2578 0
|
2581 |
+
2579 2
|
2582 |
+
2580 0
|
2583 |
+
2581 0
|
2584 |
+
2582 1
|
2585 |
+
2583 0
|
2586 |
+
2584 0
|
2587 |
+
2585 1
|
2588 |
+
2586 2
|
2589 |
+
2587 1
|
2590 |
+
2588 0
|
2591 |
+
2589 2
|
2592 |
+
2590 0
|
2593 |
+
2591 1
|
2594 |
+
2592 0
|
2595 |
+
2593 2
|
2596 |
+
2594 0
|
2597 |
+
2595 1
|
2598 |
+
2596 0
|
2599 |
+
2597 1
|
2600 |
+
2598 1
|
2601 |
+
2599 2
|
2602 |
+
2600 2
|
2603 |
+
2601 0
|
2604 |
+
2602 1
|
2605 |
+
2603 1
|
2606 |
+
2604 1
|
2607 |
+
2605 0
|
2608 |
+
2606 1
|
2609 |
+
2607 1
|
2610 |
+
2608 2
|
2611 |
+
2609 0
|
2612 |
+
2610 0
|
2613 |
+
2611 1
|
2614 |
+
2612 1
|
2615 |
+
2613 1
|
2616 |
+
2614 0
|
2617 |
+
2615 0
|
2618 |
+
2616 0
|
2619 |
+
2617 2
|
2620 |
+
2618 1
|
2621 |
+
2619 2
|
2622 |
+
2620 0
|
2623 |
+
2621 0
|
2624 |
+
2622 0
|
2625 |
+
2623 0
|
2626 |
+
2624 1
|
2627 |
+
2625 1
|
2628 |
+
2626 2
|
2629 |
+
2627 2
|
2630 |
+
2628 2
|
2631 |
+
2629 1
|
2632 |
+
2630 1
|
2633 |
+
2631 1
|
2634 |
+
2632 0
|
2635 |
+
2633 1
|
2636 |
+
2634 1
|
2637 |
+
2635 0
|
2638 |
+
2636 1
|
2639 |
+
2637 0
|
2640 |
+
2638 0
|
2641 |
+
2639 1
|
2642 |
+
2640 2
|
2643 |
+
2641 2
|
2644 |
+
2642 2
|
2645 |
+
2643 1
|
2646 |
+
2644 0
|
2647 |
+
2645 1
|
2648 |
+
2646 0
|
2649 |
+
2647 2
|
2650 |
+
2648 1
|
2651 |
+
2649 0
|
2652 |
+
2650 1
|
2653 |
+
2651 1
|
2654 |
+
2652 2
|
2655 |
+
2653 0
|
2656 |
+
2654 1
|
2657 |
+
2655 0
|
2658 |
+
2656 2
|
2659 |
+
2657 0
|
2660 |
+
2658 2
|
2661 |
+
2659 0
|
2662 |
+
2660 1
|
2663 |
+
2661 1
|
2664 |
+
2662 1
|
2665 |
+
2663 1
|
2666 |
+
2664 2
|
2667 |
+
2665 2
|
2668 |
+
2666 1
|
2669 |
+
2667 0
|
2670 |
+
2668 0
|
2671 |
+
2669 2
|
2672 |
+
2670 0
|
2673 |
+
2671 0
|
2674 |
+
2672 2
|
2675 |
+
2673 1
|
2676 |
+
2674 0
|
2677 |
+
2675 0
|
2678 |
+
2676 1
|
2679 |
+
2677 0
|
2680 |
+
2678 1
|
2681 |
+
2679 0
|
2682 |
+
2680 2
|
2683 |
+
2681 1
|
2684 |
+
2682 1
|
2685 |
+
2683 1
|
2686 |
+
2684 2
|
2687 |
+
2685 1
|
2688 |
+
2686 1
|
2689 |
+
2687 2
|
2690 |
+
2688 0
|
2691 |
+
2689 2
|
2692 |
+
2690 0
|
2693 |
+
2691 1
|
2694 |
+
2692 1
|
2695 |
+
2693 2
|
2696 |
+
2694 1
|
2697 |
+
2695 0
|
2698 |
+
2696 2
|
2699 |
+
2697 1
|
2700 |
+
2698 1
|
2701 |
+
2699 2
|
2702 |
+
2700 0
|
2703 |
+
2701 1
|
2704 |
+
2702 2
|
2705 |
+
2703 2
|
2706 |
+
2704 0
|
2707 |
+
2705 1
|
2708 |
+
2706 1
|
2709 |
+
2707 2
|
2710 |
+
2708 1
|
2711 |
+
2709 1
|
2712 |
+
2710 1
|
2713 |
+
2711 0
|
2714 |
+
2712 0
|
2715 |
+
2713 1
|
2716 |
+
2714 1
|
2717 |
+
2715 2
|
2718 |
+
2716 1
|
2719 |
+
2717 1
|
2720 |
+
2718 1
|
2721 |
+
2719 1
|
2722 |
+
2720 1
|
2723 |
+
2721 2
|
2724 |
+
2722 1
|
2725 |
+
2723 1
|
2726 |
+
2724 0
|
2727 |
+
2725 2
|
2728 |
+
2726 2
|
2729 |
+
2727 2
|
2730 |
+
2728 0
|
2731 |
+
2729 2
|
2732 |
+
2730 0
|
2733 |
+
2731 1
|
2734 |
+
2732 0
|
2735 |
+
2733 1
|
2736 |
+
2734 2
|
2737 |
+
2735 1
|
2738 |
+
2736 1
|
2739 |
+
2737 2
|
2740 |
+
2738 0
|
2741 |
+
2739 2
|
2742 |
+
2740 2
|
2743 |
+
2741 1
|
2744 |
+
2742 1
|
2745 |
+
2743 1
|
2746 |
+
2744 2
|
2747 |
+
2745 0
|
2748 |
+
2746 1
|
2749 |
+
2747 1
|
2750 |
+
2748 1
|
2751 |
+
2749 1
|
2752 |
+
2750 1
|
2753 |
+
2751 1
|
2754 |
+
2752 0
|
2755 |
+
2753 1
|
2756 |
+
2754 0
|
2757 |
+
2755 0
|
2758 |
+
2756 2
|
2759 |
+
2757 1
|
2760 |
+
2758 1
|
2761 |
+
2759 0
|
2762 |
+
2760 1
|
2763 |
+
2761 2
|
2764 |
+
2762 0
|
2765 |
+
2763 0
|
2766 |
+
2764 1
|
2767 |
+
2765 2
|
2768 |
+
2766 1
|
2769 |
+
2767 1
|
2770 |
+
2768 1
|
2771 |
+
2769 1
|
2772 |
+
2770 0
|
2773 |
+
2771 0
|
2774 |
+
2772 0
|
2775 |
+
2773 0
|
2776 |
+
2774 1
|
2777 |
+
2775 1
|
2778 |
+
2776 1
|
2779 |
+
2777 0
|
2780 |
+
2778 1
|
2781 |
+
2779 2
|
2782 |
+
2780 0
|
2783 |
+
2781 1
|
2784 |
+
2782 2
|
2785 |
+
2783 1
|
2786 |
+
2784 0
|
2787 |
+
2785 2
|
2788 |
+
2786 0
|
2789 |
+
2787 0
|
2790 |
+
2788 1
|
2791 |
+
2789 1
|
2792 |
+
2790 0
|
2793 |
+
2791 0
|
2794 |
+
2792 1
|
2795 |
+
2793 1
|
2796 |
+
2794 0
|
2797 |
+
2795 0
|
2798 |
+
2796 1
|
2799 |
+
2797 2
|
2800 |
+
2798 0
|
2801 |
+
2799 2
|
2802 |
+
2800 2
|
2803 |
+
2801 2
|
2804 |
+
2802 0
|
2805 |
+
2803 0
|
2806 |
+
2804 1
|
2807 |
+
2805 0
|
2808 |
+
2806 0
|
2809 |
+
2807 1
|
2810 |
+
2808 1
|
2811 |
+
2809 0
|
2812 |
+
2810 2
|
2813 |
+
2811 1
|
2814 |
+
2812 0
|
2815 |
+
2813 2
|
2816 |
+
2814 1
|
2817 |
+
2815 2
|
2818 |
+
2816 1
|
2819 |
+
2817 0
|
2820 |
+
2818 0
|
2821 |
+
2819 0
|
2822 |
+
2820 1
|
2823 |
+
2821 2
|
2824 |
+
2822 1
|
2825 |
+
2823 2
|
2826 |
+
2824 0
|
2827 |
+
2825 2
|
2828 |
+
2826 2
|
2829 |
+
2827 0
|
2830 |
+
2828 1
|
2831 |
+
2829 2
|
2832 |
+
2830 1
|
2833 |
+
2831 0
|
2834 |
+
2832 2
|
2835 |
+
2833 2
|
2836 |
+
2834 0
|
2837 |
+
2835 2
|
2838 |
+
2836 1
|
2839 |
+
2837 1
|
2840 |
+
2838 2
|
2841 |
+
2839 0
|
2842 |
+
2840 2
|
2843 |
+
2841 2
|
2844 |
+
2842 1
|
2845 |
+
2843 0
|
2846 |
+
2844 1
|
2847 |
+
2845 2
|
2848 |
+
2846 0
|
2849 |
+
2847 2
|
2850 |
+
2848 0
|
2851 |
+
2849 0
|
2852 |
+
2850 0
|
2853 |
+
2851 0
|
2854 |
+
2852 2
|
2855 |
+
2853 0
|
2856 |
+
2854 1
|
2857 |
+
2855 1
|
2858 |
+
2856 0
|
2859 |
+
2857 0
|
2860 |
+
2858 2
|
2861 |
+
2859 0
|
2862 |
+
2860 0
|
2863 |
+
2861 1
|
2864 |
+
2862 1
|
2865 |
+
2863 2
|
2866 |
+
2864 2
|
2867 |
+
2865 2
|
2868 |
+
2866 0
|
2869 |
+
2867 1
|
2870 |
+
2868 1
|
2871 |
+
2869 2
|
2872 |
+
2870 1
|
2873 |
+
2871 2
|
2874 |
+
2872 2
|
2875 |
+
2873 2
|
2876 |
+
2874 1
|
2877 |
+
2875 1
|
2878 |
+
2876 0
|
2879 |
+
2877 1
|
2880 |
+
2878 0
|
2881 |
+
2879 2
|
2882 |
+
2880 1
|
2883 |
+
2881 0
|
2884 |
+
2882 2
|
2885 |
+
2883 1
|
2886 |
+
2884 2
|
2887 |
+
2885 1
|
2888 |
+
2886 2
|
2889 |
+
2887 0
|
2890 |
+
2888 0
|
2891 |
+
2889 1
|
2892 |
+
2890 0
|
2893 |
+
2891 1
|
2894 |
+
2892 1
|
2895 |
+
2893 0
|
2896 |
+
2894 1
|
2897 |
+
2895 2
|
2898 |
+
2896 0
|
2899 |
+
2897 0
|
2900 |
+
2898 1
|
2901 |
+
2899 0
|
2902 |
+
2900 1
|
2903 |
+
2901 1
|
2904 |
+
2902 2
|
2905 |
+
2903 0
|
2906 |
+
2904 0
|
2907 |
+
2905 1
|
2908 |
+
2906 1
|
2909 |
+
2907 2
|
2910 |
+
2908 0
|
2911 |
+
2909 1
|
2912 |
+
2910 2
|
2913 |
+
2911 0
|
2914 |
+
2912 0
|
2915 |
+
2913 1
|
2916 |
+
2914 1
|
2917 |
+
2915 1
|
2918 |
+
2916 0
|
2919 |
+
2917 0
|
2920 |
+
2918 0
|
2921 |
+
2919 0
|
2922 |
+
2920 1
|
2923 |
+
2921 2
|
2924 |
+
2922 1
|
2925 |
+
2923 0
|
2926 |
+
2924 1
|
2927 |
+
2925 0
|
2928 |
+
2926 0
|
2929 |
+
2927 1
|
2930 |
+
2928 1
|
2931 |
+
2929 0
|
2932 |
+
2930 0
|
2933 |
+
2931 0
|
2934 |
+
2932 1
|
2935 |
+
2933 0
|
2936 |
+
2934 2
|
2937 |
+
2935 0
|
2938 |
+
2936 2
|
2939 |
+
2937 0
|
2940 |
+
2938 1
|
2941 |
+
2939 2
|
2942 |
+
2940 2
|
2943 |
+
2941 1
|
2944 |
+
2942 0
|
2945 |
+
2943 2
|
2946 |
+
2944 2
|
2947 |
+
2945 1
|
2948 |
+
2946 0
|
2949 |
+
2947 2
|
2950 |
+
2948 1
|
2951 |
+
2949 1
|
2952 |
+
2950 1
|
2953 |
+
2951 0
|
2954 |
+
2952 2
|
2955 |
+
2953 2
|
2956 |
+
2954 1
|
2957 |
+
2955 1
|
2958 |
+
2956 1
|
2959 |
+
2957 1
|
2960 |
+
2958 1
|
2961 |
+
2959 2
|
2962 |
+
2960 1
|
2963 |
+
2961 2
|
2964 |
+
2962 0
|
2965 |
+
2963 0
|
2966 |
+
2964 1
|
2967 |
+
2965 2
|
2968 |
+
2966 0
|
2969 |
+
2967 1
|
2970 |
+
2968 0
|
2971 |
+
2969 1
|
2972 |
+
2970 2
|
2973 |
+
2971 0
|
2974 |
+
2972 2
|
2975 |
+
2973 2
|
2976 |
+
2974 1
|
2977 |
+
2975 0
|
2978 |
+
2976 0
|
2979 |
+
2977 0
|
2980 |
+
2978 1
|
2981 |
+
2979 1
|
2982 |
+
2980 2
|
2983 |
+
2981 1
|
2984 |
+
2982 0
|
2985 |
+
2983 2
|
2986 |
+
2984 1
|
2987 |
+
2985 1
|
2988 |
+
2986 2
|
2989 |
+
2987 2
|
2990 |
+
2988 0
|
2991 |
+
2989 1
|
2992 |
+
2990 2
|
2993 |
+
2991 2
|
2994 |
+
2992 1
|
2995 |
+
2993 0
|
2996 |
+
2994 2
|
2997 |
+
2995 2
|
2998 |
+
2996 1
|
2999 |
+
2997 2
|
3000 |
+
2998 1
|
3001 |
+
2999 2
|
3002 |
+
3000 0
|
3003 |
+
3001 0
|
3004 |
+
3002 2
|
3005 |
+
3003 0
|
3006 |
+
3004 1
|
3007 |
+
3005 1
|
3008 |
+
3006 2
|
3009 |
+
3007 1
|
3010 |
+
3008 2
|
3011 |
+
3009 1
|
3012 |
+
3010 1
|
3013 |
+
3011 2
|
3014 |
+
3012 1
|
3015 |
+
3013 2
|
3016 |
+
3014 2
|
3017 |
+
3015 2
|
3018 |
+
3016 2
|
3019 |
+
3017 0
|
3020 |
+
3018 0
|
3021 |
+
3019 2
|
3022 |
+
3020 0
|
3023 |
+
3021 0
|
3024 |
+
3022 1
|
3025 |
+
3023 1
|
3026 |
+
3024 1
|
3027 |
+
3025 0
|
3028 |
+
3026 0
|
3029 |
+
3027 2
|
3030 |
+
3028 0
|
3031 |
+
3029 1
|
3032 |
+
3030 0
|
3033 |
+
3031 1
|
3034 |
+
3032 1
|
3035 |
+
3033 0
|
3036 |
+
3034 0
|
3037 |
+
3035 0
|
3038 |
+
3036 1
|
3039 |
+
3037 1
|
3040 |
+
3038 0
|
3041 |
+
3039 1
|
3042 |
+
3040 0
|
3043 |
+
3041 2
|
3044 |
+
3042 0
|
3045 |
+
3043 1
|
3046 |
+
3044 0
|
3047 |
+
3045 0
|
3048 |
+
3046 1
|
3049 |
+
3047 2
|
3050 |
+
3048 0
|
3051 |
+
3049 2
|
3052 |
+
3050 1
|
3053 |
+
3051 1
|
3054 |
+
3052 2
|
3055 |
+
3053 1
|
3056 |
+
3054 1
|
3057 |
+
3055 1
|
3058 |
+
3056 1
|
3059 |
+
3057 2
|
3060 |
+
3058 1
|
3061 |
+
3059 1
|
3062 |
+
3060 2
|
3063 |
+
3061 0
|
3064 |
+
3062 1
|
3065 |
+
3063 0
|
3066 |
+
3064 0
|
3067 |
+
3065 2
|
3068 |
+
3066 0
|
3069 |
+
3067 1
|
3070 |
+
3068 1
|
3071 |
+
3069 0
|
3072 |
+
3070 1
|
3073 |
+
3071 1
|
3074 |
+
3072 1
|
3075 |
+
3073 2
|
3076 |
+
3074 2
|
3077 |
+
3075 0
|
3078 |
+
3076 1
|
3079 |
+
3077 1
|
3080 |
+
3078 2
|
3081 |
+
3079 0
|
3082 |
+
3080 0
|
3083 |
+
3081 1
|
3084 |
+
3082 1
|
3085 |
+
3083 0
|
3086 |
+
3084 0
|
3087 |
+
3085 2
|
3088 |
+
3086 0
|
3089 |
+
3087 2
|
3090 |
+
3088 2
|
3091 |
+
3089 2
|
3092 |
+
3090 0
|
3093 |
+
3091 2
|
3094 |
+
3092 0
|
3095 |
+
3093 0
|
3096 |
+
3094 0
|
3097 |
+
3095 0
|
3098 |
+
3096 2
|
3099 |
+
3097 1
|
3100 |
+
3098 0
|
3101 |
+
3099 2
|
3102 |
+
3100 0
|
3103 |
+
3101 1
|
3104 |
+
3102 1
|
3105 |
+
3103 0
|
3106 |
+
3104 1
|
3107 |
+
3105 1
|
3108 |
+
3106 2
|
3109 |
+
3107 2
|
3110 |
+
3108 1
|
3111 |
+
3109 0
|
3112 |
+
3110 0
|
3113 |
+
3111 0
|
3114 |
+
3112 1
|
3115 |
+
3113 2
|
3116 |
+
3114 0
|
3117 |
+
3115 2
|
3118 |
+
3116 1
|
3119 |
+
3117 1
|
3120 |
+
3118 2
|
3121 |
+
3119 0
|
3122 |
+
3120 1
|
3123 |
+
3121 1
|
3124 |
+
3122 1
|
3125 |
+
3123 1
|
3126 |
+
3124 2
|
3127 |
+
3125 2
|
3128 |
+
3126 2
|
3129 |
+
3127 2
|
3130 |
+
3128 0
|
3131 |
+
3129 1
|
3132 |
+
3130 0
|
3133 |
+
3131 1
|
3134 |
+
3132 1
|
3135 |
+
3133 0
|
3136 |
+
3134 1
|
3137 |
+
3135 1
|
3138 |
+
3136 1
|
3139 |
+
3137 1
|
3140 |
+
3138 0
|
3141 |
+
3139 2
|
3142 |
+
3140 2
|
3143 |
+
3141 1
|
3144 |
+
3142 2
|
3145 |
+
3143 0
|
3146 |
+
3144 1
|
3147 |
+
3145 0
|
3148 |
+
3146 2
|
3149 |
+
3147 2
|
3150 |
+
3148 1
|
3151 |
+
3149 0
|
3152 |
+
3150 1
|
3153 |
+
3151 0
|
3154 |
+
3152 2
|
3155 |
+
3153 0
|
3156 |
+
3154 2
|
3157 |
+
3155 1
|
3158 |
+
3156 2
|
3159 |
+
3157 2
|
3160 |
+
3158 1
|
3161 |
+
3159 1
|
3162 |
+
3160 1
|
3163 |
+
3161 2
|
3164 |
+
3162 2
|
3165 |
+
3163 1
|
3166 |
+
3164 0
|
3167 |
+
3165 0
|
3168 |
+
3166 2
|
3169 |
+
3167 2
|
3170 |
+
3168 2
|
3171 |
+
3169 2
|
3172 |
+
3170 0
|
3173 |
+
3171 0
|
3174 |
+
3172 0
|
3175 |
+
3173 1
|
3176 |
+
3174 1
|
3177 |
+
3175 1
|
3178 |
+
3176 0
|
3179 |
+
3177 2
|
3180 |
+
3178 0
|
3181 |
+
3179 1
|
3182 |
+
3180 1
|
3183 |
+
3181 2
|
3184 |
+
3182 0
|
3185 |
+
3183 0
|
3186 |
+
3184 0
|
3187 |
+
3185 0
|
3188 |
+
3186 1
|
3189 |
+
3187 2
|
3190 |
+
3188 1
|
3191 |
+
3189 2
|
3192 |
+
3190 0
|
3193 |
+
3191 1
|
3194 |
+
3192 1
|
3195 |
+
3193 2
|
3196 |
+
3194 1
|
3197 |
+
3195 1
|
3198 |
+
3196 2
|
3199 |
+
3197 0
|
3200 |
+
3198 1
|
3201 |
+
3199 1
|
3202 |
+
3200 1
|
3203 |
+
3201 0
|
3204 |
+
3202 1
|
3205 |
+
3203 1
|
3206 |
+
3204 1
|
3207 |
+
3205 1
|
3208 |
+
3206 1
|
3209 |
+
3207 2
|
3210 |
+
3208 1
|
3211 |
+
3209 2
|
3212 |
+
3210 1
|
3213 |
+
3211 1
|
3214 |
+
3212 0
|
3215 |
+
3213 1
|
3216 |
+
3214 1
|
3217 |
+
3215 1
|
3218 |
+
3216 1
|
3219 |
+
3217 2
|
3220 |
+
3218 2
|
3221 |
+
3219 1
|
3222 |
+
3220 2
|
3223 |
+
3221 0
|
3224 |
+
3222 1
|
3225 |
+
3223 0
|
3226 |
+
3224 2
|
3227 |
+
3225 1
|
3228 |
+
3226 2
|
3229 |
+
3227 2
|
3230 |
+
3228 2
|
3231 |
+
3229 0
|
3232 |
+
3230 0
|
3233 |
+
3231 1
|
3234 |
+
3232 0
|
3235 |
+
3233 0
|
3236 |
+
3234 1
|
3237 |
+
3235 2
|
3238 |
+
3236 0
|
3239 |
+
3237 2
|
3240 |
+
3238 1
|
3241 |
+
3239 2
|
3242 |
+
3240 2
|
3243 |
+
3241 2
|
3244 |
+
3242 1
|
3245 |
+
3243 0
|
3246 |
+
3244 1
|
3247 |
+
3245 2
|
3248 |
+
3246 0
|
3249 |
+
3247 1
|
3250 |
+
3248 2
|
3251 |
+
3249 1
|
3252 |
+
3250 1
|
3253 |
+
3251 2
|
3254 |
+
3252 0
|
3255 |
+
3253 1
|
3256 |
+
3254 0
|
3257 |
+
3255 1
|
3258 |
+
3256 1
|
3259 |
+
3257 1
|
3260 |
+
3258 1
|
3261 |
+
3259 2
|
3262 |
+
3260 2
|
3263 |
+
3261 0
|
3264 |
+
3262 2
|
3265 |
+
3263 2
|
3266 |
+
3264 1
|
3267 |
+
3265 1
|
3268 |
+
3266 1
|
3269 |
+
3267 1
|
3270 |
+
3268 1
|
3271 |
+
3269 2
|
3272 |
+
3270 1
|
3273 |
+
3271 1
|
3274 |
+
3272 1
|
3275 |
+
3273 1
|
3276 |
+
3274 2
|
3277 |
+
3275 0
|
3278 |
+
3276 2
|
3279 |
+
3277 0
|
3280 |
+
3278 0
|
3281 |
+
3279 0
|
3282 |
+
3280 1
|
3283 |
+
3281 0
|
3284 |
+
3282 1
|
3285 |
+
3283 2
|
3286 |
+
3284 1
|
3287 |
+
3285 0
|
3288 |
+
3286 0
|
3289 |
+
3287 2
|
3290 |
+
3288 1
|
3291 |
+
3289 1
|
3292 |
+
3290 1
|
3293 |
+
3291 0
|
3294 |
+
3292 0
|
3295 |
+
3293 0
|
3296 |
+
3294 0
|
3297 |
+
3295 2
|
3298 |
+
3296 0
|
3299 |
+
3297 0
|
3300 |
+
3298 1
|
3301 |
+
3299 0
|
3302 |
+
3300 1
|
3303 |
+
3301 0
|
3304 |
+
3302 2
|
3305 |
+
3303 0
|
3306 |
+
3304 1
|
3307 |
+
3305 0
|
3308 |
+
3306 2
|
3309 |
+
3307 1
|
3310 |
+
3308 1
|
3311 |
+
3309 1
|
3312 |
+
3310 1
|
3313 |
+
3311 2
|
3314 |
+
3312 1
|
3315 |
+
3313 2
|
3316 |
+
3314 0
|
3317 |
+
3315 2
|
3318 |
+
3316 0
|
3319 |
+
3317 1
|
3320 |
+
3318 1
|
3321 |
+
3319 0
|
3322 |
+
3320 1
|
3323 |
+
3321 1
|
3324 |
+
3322 0
|
3325 |
+
3323 2
|
3326 |
+
3324 2
|
3327 |
+
3325 1
|
3328 |
+
3326 2
|
3329 |
+
3327 1
|
3330 |
+
3328 2
|
3331 |
+
3329 0
|
3332 |
+
3330 2
|
3333 |
+
3331 2
|
3334 |
+
3332 0
|
3335 |
+
3333 2
|
3336 |
+
3334 1
|
3337 |
+
3335 1
|
3338 |
+
3336 2
|
3339 |
+
3337 1
|
3340 |
+
3338 0
|
3341 |
+
3339 1
|
3342 |
+
3340 1
|
3343 |
+
3341 0
|
3344 |
+
3342 0
|
3345 |
+
3343 0
|
3346 |
+
3344 0
|
3347 |
+
3345 2
|
3348 |
+
3346 0
|
3349 |
+
3347 0
|
3350 |
+
3348 1
|
3351 |
+
3349 1
|
3352 |
+
3350 2
|
3353 |
+
3351 1
|
3354 |
+
3352 1
|
3355 |
+
3353 0
|
3356 |
+
3354 1
|
3357 |
+
3355 0
|
3358 |
+
3356 0
|
3359 |
+
3357 0
|
3360 |
+
3358 0
|
3361 |
+
3359 2
|
3362 |
+
3360 1
|
3363 |
+
3361 1
|
3364 |
+
3362 1
|
3365 |
+
3363 2
|
3366 |
+
3364 1
|
3367 |
+
3365 2
|
3368 |
+
3366 1
|
3369 |
+
3367 1
|
3370 |
+
3368 0
|
3371 |
+
3369 0
|
3372 |
+
3370 1
|
3373 |
+
3371 0
|
3374 |
+
3372 1
|
3375 |
+
3373 2
|
3376 |
+
3374 2
|
3377 |
+
3375 1
|
3378 |
+
3376 2
|
3379 |
+
3377 1
|
3380 |
+
3378 0
|
3381 |
+
3379 1
|
3382 |
+
3380 2
|
3383 |
+
3381 1
|
3384 |
+
3382 1
|
3385 |
+
3383 2
|
3386 |
+
3384 1
|
3387 |
+
3385 1
|
3388 |
+
3386 1
|
3389 |
+
3387 0
|
3390 |
+
3388 1
|
3391 |
+
3389 0
|
3392 |
+
3390 0
|
3393 |
+
3391 1
|
3394 |
+
3392 1
|
3395 |
+
3393 2
|
3396 |
+
3394 1
|
3397 |
+
3395 1
|
3398 |
+
3396 2
|
3399 |
+
3397 0
|
3400 |
+
3398 0
|
3401 |
+
3399 2
|
3402 |
+
3400 1
|
3403 |
+
3401 1
|
3404 |
+
3402 2
|
3405 |
+
3403 0
|
3406 |
+
3404 1
|
3407 |
+
3405 2
|
3408 |
+
3406 2
|
3409 |
+
3407 2
|
3410 |
+
3408 0
|
3411 |
+
3409 2
|
3412 |
+
3410 2
|
3413 |
+
3411 0
|
3414 |
+
3412 0
|
3415 |
+
3413 1
|
3416 |
+
3414 2
|
3417 |
+
3415 1
|
3418 |
+
3416 0
|
3419 |
+
3417 0
|
3420 |
+
3418 0
|
3421 |
+
3419 1
|
3422 |
+
3420 1
|
3423 |
+
3421 0
|
3424 |
+
3422 0
|
3425 |
+
3423 1
|
3426 |
+
3424 0
|
3427 |
+
3425 1
|
3428 |
+
3426 2
|
3429 |
+
3427 0
|
3430 |
+
3428 0
|
3431 |
+
3429 0
|
3432 |
+
3430 0
|
3433 |
+
3431 1
|
3434 |
+
3432 0
|
3435 |
+
3433 1
|
3436 |
+
3434 1
|
3437 |
+
3435 1
|
3438 |
+
3436 1
|
3439 |
+
3437 1
|
3440 |
+
3438 1
|
3441 |
+
3439 2
|
3442 |
+
3440 1
|
3443 |
+
3441 1
|
3444 |
+
3442 0
|
3445 |
+
3443 1
|
3446 |
+
3444 1
|
3447 |
+
3445 0
|
3448 |
+
3446 2
|
3449 |
+
3447 2
|
3450 |
+
3448 2
|
3451 |
+
3449 0
|
3452 |
+
3450 1
|
3453 |
+
3451 2
|
3454 |
+
3452 1
|
3455 |
+
3453 2
|
3456 |
+
3454 2
|
3457 |
+
3455 0
|
3458 |
+
3456 1
|
3459 |
+
3457 2
|
3460 |
+
3458 1
|
3461 |
+
3459 0
|
3462 |
+
3460 2
|
3463 |
+
3461 2
|
3464 |
+
3462 1
|
3465 |
+
3463 2
|
3466 |
+
3464 2
|
3467 |
+
3465 1
|
3468 |
+
3466 2
|
3469 |
+
3467 0
|
3470 |
+
3468 2
|
3471 |
+
3469 0
|
3472 |
+
3470 1
|
3473 |
+
3471 1
|
3474 |
+
3472 1
|
3475 |
+
3473 0
|
3476 |
+
3474 0
|
3477 |
+
3475 1
|
3478 |
+
3476 1
|
3479 |
+
3477 1
|
3480 |
+
3478 0
|
3481 |
+
3479 0
|
3482 |
+
3480 0
|
3483 |
+
3481 1
|
3484 |
+
3482 1
|
3485 |
+
3483 1
|
3486 |
+
3484 1
|
3487 |
+
3485 0
|
3488 |
+
3486 2
|
3489 |
+
3487 1
|
3490 |
+
3488 1
|
3491 |
+
3489 1
|
3492 |
+
3490 2
|
3493 |
+
3491 1
|
3494 |
+
3492 2
|
3495 |
+
3493 0
|
3496 |
+
3494 1
|
3497 |
+
3495 1
|
3498 |
+
3496 0
|
3499 |
+
3497 0
|
3500 |
+
3498 1
|
3501 |
+
3499 1
|
3502 |
+
3500 2
|
3503 |
+
3501 1
|
3504 |
+
3502 2
|
3505 |
+
3503 1
|
3506 |
+
3504 2
|
3507 |
+
3505 1
|
3508 |
+
3506 0
|
3509 |
+
3507 2
|
3510 |
+
3508 2
|
3511 |
+
3509 0
|
3512 |
+
3510 2
|
3513 |
+
3511 2
|
3514 |
+
3512 2
|
3515 |
+
3513 0
|
3516 |
+
3514 2
|
3517 |
+
3515 2
|
3518 |
+
3516 1
|
3519 |
+
3517 0
|
3520 |
+
3518 2
|
3521 |
+
3519 0
|
3522 |
+
3520 1
|
3523 |
+
3521 1
|
3524 |
+
3522 0
|
3525 |
+
3523 1
|
3526 |
+
3524 1
|
3527 |
+
3525 0
|
3528 |
+
3526 1
|
3529 |
+
3527 0
|
3530 |
+
3528 2
|
3531 |
+
3529 2
|
3532 |
+
3530 2
|
3533 |
+
3531 0
|
3534 |
+
3532 1
|
3535 |
+
3533 0
|
3536 |
+
3534 0
|
3537 |
+
3535 0
|
3538 |
+
3536 1
|
3539 |
+
3537 0
|
3540 |
+
3538 1
|
3541 |
+
3539 0
|
3542 |
+
3540 2
|
3543 |
+
3541 2
|
3544 |
+
3542 0
|
3545 |
+
3543 1
|
3546 |
+
3544 2
|
3547 |
+
3545 1
|
3548 |
+
3546 0
|
3549 |
+
3547 0
|
3550 |
+
3548 1
|
3551 |
+
3549 0
|
3552 |
+
3550 0
|
3553 |
+
3551 1
|
3554 |
+
3552 1
|
3555 |
+
3553 0
|
3556 |
+
3554 1
|
3557 |
+
3555 2
|
3558 |
+
3556 0
|
3559 |
+
3557 1
|
3560 |
+
3558 0
|
3561 |
+
3559 0
|
3562 |
+
3560 1
|
3563 |
+
3561 0
|
3564 |
+
3562 2
|
3565 |
+
3563 2
|
3566 |
+
3564 0
|
3567 |
+
3565 1
|
3568 |
+
3566 2
|
3569 |
+
3567 2
|
3570 |
+
3568 0
|
3571 |
+
3569 2
|
3572 |
+
3570 0
|
3573 |
+
3571 2
|
3574 |
+
3572 2
|
3575 |
+
3573 1
|
3576 |
+
3574 0
|
3577 |
+
3575 2
|
3578 |
+
3576 1
|
3579 |
+
3577 0
|
3580 |
+
3578 2
|
3581 |
+
3579 1
|
3582 |
+
3580 1
|
3583 |
+
3581 2
|
3584 |
+
3582 2
|
3585 |
+
3583 2
|
3586 |
+
3584 0
|
3587 |
+
3585 0
|
3588 |
+
3586 1
|
3589 |
+
3587 1
|
3590 |
+
3588 1
|
3591 |
+
3589 0
|
3592 |
+
3590 2
|
3593 |
+
3591 2
|
3594 |
+
3592 2
|
3595 |
+
3593 0
|
3596 |
+
3594 1
|
3597 |
+
3595 2
|
3598 |
+
3596 0
|
3599 |
+
3597 0
|
3600 |
+
3598 0
|
3601 |
+
3599 0
|
3602 |
+
3600 1
|
3603 |
+
3601 2
|
3604 |
+
3602 0
|
3605 |
+
3603 1
|
3606 |
+
3604 1
|
3607 |
+
3605 0
|
3608 |
+
3606 0
|
3609 |
+
3607 2
|
3610 |
+
3608 0
|
3611 |
+
3609 1
|
3612 |
+
3610 2
|
3613 |
+
3611 1
|
3614 |
+
3612 0
|
3615 |
+
3613 2
|
3616 |
+
3614 1
|
3617 |
+
3615 1
|
3618 |
+
3616 1
|
3619 |
+
3617 2
|
3620 |
+
3618 1
|
3621 |
+
3619 1
|
3622 |
+
3620 0
|
3623 |
+
3621 0
|
3624 |
+
3622 2
|
3625 |
+
3623 0
|
3626 |
+
3624 2
|
3627 |
+
3625 0
|
3628 |
+
3626 1
|
3629 |
+
3627 0
|
3630 |
+
3628 2
|
3631 |
+
3629 2
|
3632 |
+
3630 1
|
3633 |
+
3631 1
|
3634 |
+
3632 0
|
3635 |
+
3633 1
|
3636 |
+
3634 2
|
3637 |
+
3635 1
|
3638 |
+
3636 2
|
3639 |
+
3637 1
|
3640 |
+
3638 2
|
3641 |
+
3639 1
|
3642 |
+
3640 0
|
3643 |
+
3641 1
|
3644 |
+
3642 0
|
3645 |
+
3643 2
|
3646 |
+
3644 1
|
3647 |
+
3645 0
|
3648 |
+
3646 1
|
3649 |
+
3647 2
|
3650 |
+
3648 1
|
3651 |
+
3649 0
|
3652 |
+
3650 0
|
3653 |
+
3651 0
|
3654 |
+
3652 0
|
3655 |
+
3653 1
|
3656 |
+
3654 0
|
3657 |
+
3655 0
|
3658 |
+
3656 2
|
3659 |
+
3657 2
|
3660 |
+
3658 1
|
3661 |
+
3659 1
|
3662 |
+
3660 2
|
3663 |
+
3661 0
|
3664 |
+
3662 0
|
3665 |
+
3663 2
|
3666 |
+
3664 0
|
3667 |
+
3665 0
|
3668 |
+
3666 2
|
3669 |
+
3667 0
|
3670 |
+
3668 0
|
3671 |
+
3669 0
|
3672 |
+
3670 1
|
3673 |
+
3671 0
|
3674 |
+
3672 1
|
3675 |
+
3673 0
|
3676 |
+
3674 0
|
3677 |
+
3675 1
|
3678 |
+
3676 2
|
3679 |
+
3677 2
|
3680 |
+
3678 1
|
3681 |
+
3679 1
|
3682 |
+
3680 1
|
3683 |
+
3681 0
|
3684 |
+
3682 0
|
3685 |
+
3683 0
|
3686 |
+
3684 2
|
3687 |
+
3685 2
|
3688 |
+
3686 0
|
3689 |
+
3687 2
|
3690 |
+
3688 0
|
3691 |
+
3689 1
|
3692 |
+
3690 1
|
3693 |
+
3691 0
|
3694 |
+
3692 0
|
3695 |
+
3693 2
|
3696 |
+
3694 1
|
3697 |
+
3695 2
|
3698 |
+
3696 1
|
3699 |
+
3697 2
|
3700 |
+
3698 0
|
3701 |
+
3699 0
|
3702 |
+
3700 1
|
3703 |
+
3701 2
|
3704 |
+
3702 1
|
3705 |
+
3703 0
|
3706 |
+
3704 0
|
3707 |
+
3705 0
|
3708 |
+
3706 1
|
3709 |
+
3707 0
|
3710 |
+
3708 1
|
3711 |
+
3709 0
|
3712 |
+
3710 2
|
3713 |
+
3711 0
|
3714 |
+
3712 1
|
3715 |
+
3713 1
|
3716 |
+
3714 2
|
3717 |
+
3715 1
|
3718 |
+
3716 1
|
3719 |
+
3717 0
|
3720 |
+
3718 0
|
3721 |
+
3719 2
|
3722 |
+
3720 2
|
3723 |
+
3721 1
|
3724 |
+
3722 1
|
3725 |
+
3723 1
|
3726 |
+
3724 1
|
3727 |
+
3725 1
|
3728 |
+
3726 1
|
3729 |
+
3727 2
|
3730 |
+
3728 1
|
3731 |
+
3729 0
|
3732 |
+
3730 1
|
3733 |
+
3731 2
|
3734 |
+
3732 2
|
3735 |
+
3733 0
|
3736 |
+
3734 0
|
3737 |
+
3735 1
|
3738 |
+
3736 1
|
3739 |
+
3737 0
|
3740 |
+
3738 0
|
3741 |
+
3739 0
|
3742 |
+
3740 0
|
3743 |
+
3741 0
|
3744 |
+
3742 0
|
3745 |
+
3743 1
|
3746 |
+
3744 2
|
3747 |
+
3745 1
|
3748 |
+
3746 1
|
3749 |
+
3747 0
|
3750 |
+
3748 0
|
3751 |
+
3749 1
|
3752 |
+
3750 2
|
3753 |
+
3751 0
|
3754 |
+
3752 1
|
3755 |
+
3753 1
|
3756 |
+
3754 1
|
3757 |
+
3755 0
|
3758 |
+
3756 2
|
3759 |
+
3757 0
|
3760 |
+
3758 0
|
3761 |
+
3759 2
|
3762 |
+
3760 2
|
3763 |
+
3761 1
|
3764 |
+
3762 2
|
3765 |
+
3763 1
|
3766 |
+
3764 1
|
3767 |
+
3765 2
|
3768 |
+
3766 2
|
3769 |
+
3767 0
|
3770 |
+
3768 0
|
3771 |
+
3769 1
|
3772 |
+
3770 0
|
3773 |
+
3771 0
|
3774 |
+
3772 1
|
3775 |
+
3773 0
|
3776 |
+
3774 0
|
3777 |
+
3775 0
|
3778 |
+
3776 0
|
3779 |
+
3777 2
|
3780 |
+
3778 2
|
3781 |
+
3779 0
|
3782 |
+
3780 0
|
3783 |
+
3781 1
|
3784 |
+
3782 2
|
3785 |
+
3783 0
|
3786 |
+
3784 1
|
3787 |
+
3785 0
|
3788 |
+
3786 0
|
3789 |
+
3787 2
|
3790 |
+
3788 1
|
3791 |
+
3789 1
|
3792 |
+
3790 1
|
3793 |
+
3791 2
|
3794 |
+
3792 2
|
3795 |
+
3793 2
|
3796 |
+
3794 2
|
3797 |
+
3795 0
|
3798 |
+
3796 1
|
3799 |
+
3797 1
|
3800 |
+
3798 2
|
3801 |
+
3799 1
|
3802 |
+
3800 2
|
3803 |
+
3801 1
|
3804 |
+
3802 2
|
3805 |
+
3803 2
|
3806 |
+
3804 0
|
3807 |
+
3805 0
|
3808 |
+
3806 1
|
3809 |
+
3807 1
|
3810 |
+
3808 2
|
3811 |
+
3809 1
|
3812 |
+
3810 0
|
3813 |
+
3811 1
|
3814 |
+
3812 1
|
3815 |
+
3813 1
|
3816 |
+
3814 2
|
3817 |
+
3815 1
|
3818 |
+
3816 2
|
3819 |
+
3817 1
|
3820 |
+
3818 2
|
3821 |
+
3819 2
|
3822 |
+
3820 0
|
3823 |
+
3821 0
|
3824 |
+
3822 1
|
3825 |
+
3823 1
|
3826 |
+
3824 2
|
3827 |
+
3825 2
|
3828 |
+
3826 2
|
3829 |
+
3827 2
|
3830 |
+
3828 2
|
3831 |
+
3829 2
|
3832 |
+
3830 2
|
3833 |
+
3831 0
|
3834 |
+
3832 1
|
3835 |
+
3833 2
|
3836 |
+
3834 1
|
3837 |
+
3835 1
|
3838 |
+
3836 1
|
3839 |
+
3837 2
|
3840 |
+
3838 2
|
3841 |
+
3839 0
|
3842 |
+
3840 1
|
3843 |
+
3841 1
|
3844 |
+
3842 0
|
3845 |
+
3843 2
|
3846 |
+
3844 1
|
3847 |
+
3845 0
|
3848 |
+
3846 2
|
3849 |
+
3847 1
|
3850 |
+
3848 0
|
3851 |
+
3849 0
|
3852 |
+
3850 0
|
3853 |
+
3851 0
|
3854 |
+
3852 2
|
3855 |
+
3853 1
|
3856 |
+
3854 2
|
3857 |
+
3855 1
|
3858 |
+
3856 0
|
3859 |
+
3857 1
|
3860 |
+
3858 0
|
3861 |
+
3859 0
|
3862 |
+
3860 0
|
3863 |
+
3861 0
|
3864 |
+
3862 2
|
3865 |
+
3863 0
|
3866 |
+
3864 0
|
3867 |
+
3865 0
|
3868 |
+
3866 2
|
3869 |
+
3867 0
|
3870 |
+
3868 0
|
3871 |
+
3869 0
|
3872 |
+
3870 1
|
3873 |
+
3871 0
|
3874 |
+
3872 0
|
3875 |
+
3873 1
|
3876 |
+
3874 0
|
3877 |
+
3875 1
|
3878 |
+
3876 0
|
3879 |
+
3877 2
|
3880 |
+
3878 1
|
3881 |
+
3879 0
|
3882 |
+
3880 1
|
3883 |
+
3881 2
|
3884 |
+
3882 2
|
3885 |
+
3883 1
|
3886 |
+
3884 2
|
3887 |
+
3885 0
|
3888 |
+
3886 1
|
3889 |
+
3887 1
|
3890 |
+
3888 0
|
3891 |
+
3889 0
|
3892 |
+
3890 0
|
3893 |
+
3891 0
|
3894 |
+
3892 0
|
3895 |
+
3893 1
|
3896 |
+
3894 0
|
3897 |
+
3895 0
|
3898 |
+
3896 2
|
3899 |
+
3897 0
|
3900 |
+
3898 0
|
3901 |
+
3899 1
|
3902 |
+
3900 0
|
3903 |
+
3901 2
|
3904 |
+
3902 2
|
3905 |
+
3903 2
|
3906 |
+
3904 2
|
3907 |
+
3905 0
|
3908 |
+
3906 2
|
3909 |
+
3907 1
|
3910 |
+
3908 0
|
3911 |
+
3909 2
|
3912 |
+
3910 1
|
3913 |
+
3911 0
|
3914 |
+
3912 2
|
3915 |
+
3913 2
|
3916 |
+
3914 0
|
3917 |
+
3915 1
|
3918 |
+
3916 0
|
3919 |
+
3917 1
|
3920 |
+
3918 2
|
3921 |
+
3919 1
|
3922 |
+
3920 0
|
3923 |
+
3921 2
|
3924 |
+
3922 2
|
3925 |
+
3923 1
|
3926 |
+
3924 1
|
3927 |
+
3925 1
|
3928 |
+
3926 1
|
3929 |
+
3927 1
|
3930 |
+
3928 2
|
3931 |
+
3929 1
|
3932 |
+
3930 1
|
3933 |
+
3931 0
|
3934 |
+
3932 1
|
3935 |
+
3933 0
|
3936 |
+
3934 0
|
3937 |
+
3935 2
|
3938 |
+
3936 2
|
3939 |
+
3937 1
|
3940 |
+
3938 1
|
3941 |
+
3939 2
|
3942 |
+
3940 1
|
3943 |
+
3941 0
|
3944 |
+
3942 0
|
3945 |
+
3943 0
|
3946 |
+
3944 0
|
3947 |
+
3945 2
|
3948 |
+
3946 0
|
3949 |
+
3947 2
|
3950 |
+
3948 1
|
3951 |
+
3949 0
|
3952 |
+
3950 0
|
3953 |
+
3951 0
|
3954 |
+
3952 0
|
3955 |
+
3953 0
|
3956 |
+
3954 1
|
3957 |
+
3955 0
|
3958 |
+
3956 1
|
3959 |
+
3957 1
|
3960 |
+
3958 1
|
3961 |
+
3959 0
|
3962 |
+
3960 2
|
3963 |
+
3961 0
|
3964 |
+
3962 0
|
3965 |
+
3963 0
|
3966 |
+
3964 1
|
3967 |
+
3965 1
|
3968 |
+
3966 0
|
3969 |
+
3967 2
|
3970 |
+
3968 0
|
3971 |
+
3969 1
|
3972 |
+
3970 2
|
3973 |
+
3971 0
|
3974 |
+
3972 2
|
3975 |
+
3973 1
|
3976 |
+
3974 2
|
3977 |
+
3975 1
|
3978 |
+
3976 2
|
3979 |
+
3977 1
|
3980 |
+
3978 2
|
3981 |
+
3979 2
|
3982 |
+
3980 0
|
3983 |
+
3981 2
|
3984 |
+
3982 0
|
3985 |
+
3983 2
|
3986 |
+
3984 1
|
3987 |
+
3985 1
|
3988 |
+
3986 2
|
3989 |
+
3987 0
|
3990 |
+
3988 0
|
3991 |
+
3989 2
|
3992 |
+
3990 2
|
3993 |
+
3991 1
|
3994 |
+
3992 0
|
3995 |
+
3993 2
|
3996 |
+
3994 1
|
3997 |
+
3995 2
|
3998 |
+
3996 1
|
3999 |
+
3997 2
|
4000 |
+
3998 0
|
4001 |
+
3999 0
|
4002 |
+
4000 0
|
4003 |
+
4001 2
|
4004 |
+
4002 2
|
4005 |
+
4003 0
|
4006 |
+
4004 2
|
4007 |
+
4005 2
|
4008 |
+
4006 1
|
4009 |
+
4007 2
|
4010 |
+
4008 2
|
4011 |
+
4009 1
|
4012 |
+
4010 2
|
4013 |
+
4011 1
|
4014 |
+
4012 2
|
4015 |
+
4013 1
|
4016 |
+
4014 0
|
4017 |
+
4015 1
|
4018 |
+
4016 0
|
4019 |
+
4017 0
|
4020 |
+
4018 0
|
4021 |
+
4019 2
|
4022 |
+
4020 1
|
4023 |
+
4021 0
|
4024 |
+
4022 0
|
4025 |
+
4023 0
|
4026 |
+
4024 0
|
4027 |
+
4025 1
|
4028 |
+
4026 0
|
4029 |
+
4027 2
|
4030 |
+
4028 2
|
4031 |
+
4029 2
|
4032 |
+
4030 0
|
4033 |
+
4031 2
|
4034 |
+
4032 1
|
4035 |
+
4033 2
|
4036 |
+
4034 0
|
4037 |
+
4035 2
|
4038 |
+
4036 2
|
4039 |
+
4037 1
|
4040 |
+
4038 1
|
4041 |
+
4039 2
|
4042 |
+
4040 1
|
4043 |
+
4041 0
|
4044 |
+
4042 0
|
4045 |
+
4043 0
|
4046 |
+
4044 1
|
4047 |
+
4045 2
|
4048 |
+
4046 0
|
4049 |
+
4047 1
|
4050 |
+
4048 2
|
4051 |
+
4049 0
|
4052 |
+
4050 1
|
4053 |
+
4051 0
|
4054 |
+
4052 0
|
4055 |
+
4053 1
|
4056 |
+
4054 0
|
4057 |
+
4055 0
|
4058 |
+
4056 1
|
4059 |
+
4057 0
|
4060 |
+
4058 0
|
4061 |
+
4059 1
|
4062 |
+
4060 2
|
4063 |
+
4061 0
|
4064 |
+
4062 0
|
4065 |
+
4063 2
|
4066 |
+
4064 1
|
4067 |
+
4065 0
|
4068 |
+
4066 2
|
4069 |
+
4067 1
|
4070 |
+
4068 2
|
4071 |
+
4069 2
|
4072 |
+
4070 2
|
4073 |
+
4071 2
|
4074 |
+
4072 0
|
4075 |
+
4073 2
|
4076 |
+
4074 1
|
4077 |
+
4075 1
|
4078 |
+
4076 0
|
4079 |
+
4077 1
|
4080 |
+
4078 0
|
4081 |
+
4079 1
|
4082 |
+
4080 1
|
4083 |
+
4081 2
|
4084 |
+
4082 2
|
4085 |
+
4083 1
|
4086 |
+
4084 0
|
4087 |
+
4085 0
|
4088 |
+
4086 1
|
4089 |
+
4087 1
|
4090 |
+
4088 0
|
4091 |
+
4089 2
|
4092 |
+
4090 0
|
4093 |
+
4091 0
|
4094 |
+
4092 2
|
4095 |
+
4093 0
|
4096 |
+
4094 0
|
4097 |
+
4095 1
|
4098 |
+
4096 2
|
4099 |
+
4097 2
|
4100 |
+
4098 2
|
4101 |
+
4099 1
|
4102 |
+
4100 1
|
4103 |
+
4101 1
|
4104 |
+
4102 1
|
4105 |
+
4103 1
|
4106 |
+
4104 2
|
4107 |
+
4105 0
|
4108 |
+
4106 0
|
4109 |
+
4107 2
|
4110 |
+
4108 0
|
4111 |
+
4109 1
|
4112 |
+
4110 0
|
4113 |
+
4111 1
|
4114 |
+
4112 0
|
4115 |
+
4113 1
|
4116 |
+
4114 0
|
4117 |
+
4115 2
|
4118 |
+
4116 2
|
4119 |
+
4117 1
|
4120 |
+
4118 1
|
4121 |
+
4119 2
|
4122 |
+
4120 0
|
4123 |
+
4121 1
|
4124 |
+
4122 1
|
4125 |
+
4123 0
|
4126 |
+
4124 0
|
4127 |
+
4125 2
|
4128 |
+
4126 0
|
4129 |
+
4127 2
|
4130 |
+
4128 1
|
4131 |
+
4129 2
|
4132 |
+
4130 1
|
4133 |
+
4131 1
|
4134 |
+
4132 0
|
4135 |
+
4133 0
|
4136 |
+
4134 0
|
4137 |
+
4135 1
|
4138 |
+
4136 0
|
4139 |
+
4137 0
|
4140 |
+
4138 0
|
4141 |
+
4139 2
|
4142 |
+
4140 0
|
4143 |
+
4141 0
|
4144 |
+
4142 2
|
4145 |
+
4143 0
|
4146 |
+
4144 1
|
4147 |
+
4145 0
|
4148 |
+
4146 2
|
4149 |
+
4147 2
|
4150 |
+
4148 0
|
4151 |
+
4149 2
|
4152 |
+
4150 1
|
4153 |
+
4151 0
|
4154 |
+
4152 1
|
4155 |
+
4153 1
|
4156 |
+
4154 2
|
4157 |
+
4155 1
|
4158 |
+
4156 0
|
4159 |
+
4157 1
|
4160 |
+
4158 0
|
4161 |
+
4159 1
|
4162 |
+
4160 1
|
4163 |
+
4161 1
|
4164 |
+
4162 2
|
4165 |
+
4163 0
|
4166 |
+
4164 0
|
4167 |
+
4165 1
|
4168 |
+
4166 2
|
4169 |
+
4167 1
|
4170 |
+
4168 1
|
4171 |
+
4169 0
|
4172 |
+
4170 2
|
4173 |
+
4171 1
|
4174 |
+
4172 0
|
4175 |
+
4173 2
|
4176 |
+
4174 0
|
4177 |
+
4175 2
|
4178 |
+
4176 0
|
4179 |
+
4177 1
|
4180 |
+
4178 1
|
4181 |
+
4179 1
|
4182 |
+
4180 1
|
4183 |
+
4181 0
|
4184 |
+
4182 2
|
4185 |
+
4183 0
|
4186 |
+
4184 1
|
4187 |
+
4185 2
|
4188 |
+
4186 1
|
4189 |
+
4187 2
|
4190 |
+
4188 0
|
4191 |
+
4189 0
|
4192 |
+
4190 1
|
4193 |
+
4191 2
|
4194 |
+
4192 1
|
4195 |
+
4193 1
|
4196 |
+
4194 2
|
4197 |
+
4195 0
|
4198 |
+
4196 2
|
4199 |
+
4197 1
|
4200 |
+
4198 0
|
4201 |
+
4199 0
|
4202 |
+
4200 1
|
4203 |
+
4201 0
|
4204 |
+
4202 1
|
4205 |
+
4203 0
|
4206 |
+
4204 2
|
4207 |
+
4205 2
|
4208 |
+
4206 0
|
4209 |
+
4207 2
|
4210 |
+
4208 1
|
4211 |
+
4209 2
|
4212 |
+
4210 0
|
4213 |
+
4211 2
|
4214 |
+
4212 0
|
4215 |
+
4213 1
|
4216 |
+
4214 2
|
4217 |
+
4215 1
|
4218 |
+
4216 0
|
4219 |
+
4217 0
|
4220 |
+
4218 2
|
4221 |
+
4219 2
|
4222 |
+
4220 0
|
4223 |
+
4221 2
|
4224 |
+
4222 0
|
4225 |
+
4223 1
|
4226 |
+
4224 1
|
4227 |
+
4225 0
|
4228 |
+
4226 1
|
4229 |
+
4227 1
|
4230 |
+
4228 0
|
4231 |
+
4229 1
|
4232 |
+
4230 1
|
4233 |
+
4231 2
|
4234 |
+
4232 2
|
4235 |
+
4233 0
|
4236 |
+
4234 2
|
4237 |
+
4235 1
|
4238 |
+
4236 2
|
4239 |
+
4237 2
|
4240 |
+
4238 0
|
4241 |
+
4239 0
|
4242 |
+
4240 0
|
4243 |
+
4241 0
|
4244 |
+
4242 0
|
4245 |
+
4243 1
|
4246 |
+
4244 1
|
4247 |
+
4245 2
|
4248 |
+
4246 2
|
4249 |
+
4247 1
|
4250 |
+
4248 0
|
4251 |
+
4249 0
|
4252 |
+
4250 2
|
4253 |
+
4251 0
|
4254 |
+
4252 1
|
4255 |
+
4253 0
|
4256 |
+
4254 2
|
4257 |
+
4255 1
|
4258 |
+
4256 0
|
4259 |
+
4257 0
|
4260 |
+
4258 2
|
4261 |
+
4259 2
|
4262 |
+
4260 1
|
4263 |
+
4261 2
|
4264 |
+
4262 0
|
4265 |
+
4263 0
|
4266 |
+
4264 2
|
4267 |
+
4265 0
|
4268 |
+
4266 2
|
4269 |
+
4267 1
|
4270 |
+
4268 2
|
4271 |
+
4269 0
|
4272 |
+
4270 0
|
4273 |
+
4271 0
|
4274 |
+
4272 1
|
4275 |
+
4273 0
|
4276 |
+
4274 2
|
4277 |
+
4275 0
|
4278 |
+
4276 2
|
4279 |
+
4277 1
|
4280 |
+
4278 2
|
4281 |
+
4279 1
|
4282 |
+
4280 1
|
4283 |
+
4281 2
|
4284 |
+
4282 0
|
4285 |
+
4283 0
|
4286 |
+
4284 1
|
4287 |
+
4285 2
|
4288 |
+
4286 0
|
4289 |
+
4287 0
|
4290 |
+
4288 0
|
4291 |
+
4289 1
|
4292 |
+
4290 0
|
4293 |
+
4291 1
|
4294 |
+
4292 1
|
4295 |
+
4293 0
|
4296 |
+
4294 0
|
4297 |
+
4295 2
|
4298 |
+
4296 0
|
4299 |
+
4297 2
|
4300 |
+
4298 2
|
4301 |
+
4299 2
|
4302 |
+
4300 2
|
4303 |
+
4301 2
|
4304 |
+
4302 0
|
4305 |
+
4303 1
|
4306 |
+
4304 1
|
4307 |
+
4305 1
|
4308 |
+
4306 2
|
4309 |
+
4307 2
|
4310 |
+
4308 2
|
4311 |
+
4309 1
|
4312 |
+
4310 2
|
4313 |
+
4311 1
|
4314 |
+
4312 0
|
4315 |
+
4313 1
|
4316 |
+
4314 0
|
4317 |
+
4315 1
|
4318 |
+
4316 2
|
4319 |
+
4317 1
|
4320 |
+
4318 0
|
4321 |
+
4319 2
|
4322 |
+
4320 2
|
4323 |
+
4321 0
|
4324 |
+
4322 1
|
4325 |
+
4323 0
|
4326 |
+
4324 1
|
4327 |
+
4325 2
|
4328 |
+
4326 1
|
4329 |
+
4327 1
|
4330 |
+
4328 2
|
4331 |
+
4329 2
|
4332 |
+
4330 0
|
4333 |
+
4331 2
|
4334 |
+
4332 2
|
4335 |
+
4333 2
|
4336 |
+
4334 2
|
4337 |
+
4335 2
|
4338 |
+
4336 0
|
4339 |
+
4337 0
|
4340 |
+
4338 0
|
4341 |
+
4339 0
|
4342 |
+
4340 0
|
4343 |
+
4341 2
|
4344 |
+
4342 0
|
4345 |
+
4343 0
|
4346 |
+
4344 0
|
4347 |
+
4345 2
|
4348 |
+
4346 1
|
4349 |
+
4347 0
|
4350 |
+
4348 0
|
4351 |
+
4349 2
|
4352 |
+
4350 1
|
4353 |
+
4351 1
|
4354 |
+
4352 0
|
4355 |
+
4353 1
|
4356 |
+
4354 1
|
4357 |
+
4355 0
|
4358 |
+
4356 1
|
4359 |
+
4357 2
|
4360 |
+
4358 1
|
4361 |
+
4359 0
|
4362 |
+
4360 2
|
4363 |
+
4361 2
|
4364 |
+
4362 0
|
4365 |
+
4363 1
|
4366 |
+
4364 0
|
4367 |
+
4365 1
|
4368 |
+
4366 0
|
4369 |
+
4367 1
|
4370 |
+
4368 0
|
4371 |
+
4369 1
|
4372 |
+
4370 0
|
4373 |
+
4371 1
|
4374 |
+
4372 1
|
4375 |
+
4373 1
|
4376 |
+
4374 1
|
4377 |
+
4375 0
|
4378 |
+
4376 1
|
4379 |
+
4377 1
|
4380 |
+
4378 2
|
4381 |
+
4379 2
|
4382 |
+
4380 0
|
4383 |
+
4381 0
|
4384 |
+
4382 2
|
4385 |
+
4383 1
|
4386 |
+
4384 2
|
4387 |
+
4385 0
|
4388 |
+
4386 1
|
4389 |
+
4387 1
|
4390 |
+
4388 2
|
4391 |
+
4389 0
|
4392 |
+
4390 2
|
4393 |
+
4391 1
|
4394 |
+
4392 1
|
4395 |
+
4393 1
|
4396 |
+
4394 2
|
4397 |
+
4395 1
|
4398 |
+
4396 1
|
4399 |
+
4397 1
|
4400 |
+
4398 1
|
4401 |
+
4399 1
|
4402 |
+
4400 1
|
4403 |
+
4401 2
|
4404 |
+
4402 2
|
4405 |
+
4403 2
|
4406 |
+
4404 0
|
4407 |
+
4405 1
|
4408 |
+
4406 2
|
4409 |
+
4407 0
|
4410 |
+
4408 2
|
4411 |
+
4409 2
|
4412 |
+
4410 1
|
4413 |
+
4411 1
|
4414 |
+
4412 1
|
4415 |
+
4413 0
|
4416 |
+
4414 1
|
4417 |
+
4415 2
|
4418 |
+
4416 0
|
4419 |
+
4417 2
|
4420 |
+
4418 2
|
4421 |
+
4419 0
|
4422 |
+
4420 2
|
4423 |
+
4421 1
|
4424 |
+
4422 1
|
4425 |
+
4423 1
|
4426 |
+
4424 1
|
4427 |
+
4425 2
|
4428 |
+
4426 0
|
4429 |
+
4427 1
|
4430 |
+
4428 0
|
4431 |
+
4429 2
|
4432 |
+
4430 0
|
4433 |
+
4431 0
|
4434 |
+
4432 0
|
4435 |
+
4433 0
|
4436 |
+
4434 2
|
4437 |
+
4435 1
|
4438 |
+
4436 0
|
4439 |
+
4437 2
|
4440 |
+
4438 1
|
4441 |
+
4439 1
|
4442 |
+
4440 1
|
4443 |
+
4441 1
|
4444 |
+
4442 1
|
4445 |
+
4443 0
|
4446 |
+
4444 2
|
4447 |
+
4445 0
|
4448 |
+
4446 1
|
4449 |
+
4447 1
|
4450 |
+
4448 2
|
4451 |
+
4449 1
|
4452 |
+
4450 0
|
4453 |
+
4451 1
|
4454 |
+
4452 0
|
4455 |
+
4453 1
|
4456 |
+
4454 0
|
4457 |
+
4455 2
|
4458 |
+
4456 2
|
4459 |
+
4457 2
|
4460 |
+
4458 0
|
4461 |
+
4459 1
|
4462 |
+
4460 2
|
4463 |
+
4461 1
|
4464 |
+
4462 2
|
4465 |
+
4463 2
|
4466 |
+
4464 0
|
4467 |
+
4465 2
|
4468 |
+
4466 0
|
4469 |
+
4467 1
|
4470 |
+
4468 0
|
4471 |
+
4469 2
|
4472 |
+
4470 2
|
4473 |
+
4471 0
|
4474 |
+
4472 1
|
4475 |
+
4473 2
|
4476 |
+
4474 0
|
4477 |
+
4475 2
|
4478 |
+
4476 0
|
4479 |
+
4477 1
|
4480 |
+
4478 1
|
4481 |
+
4479 1
|
4482 |
+
4480 2
|
4483 |
+
4481 1
|
4484 |
+
4482 0
|
4485 |
+
4483 1
|
4486 |
+
4484 1
|
4487 |
+
4485 1
|
4488 |
+
4486 1
|
4489 |
+
4487 1
|
4490 |
+
4488 1
|
4491 |
+
4489 0
|
4492 |
+
4490 2
|
4493 |
+
4491 2
|
4494 |
+
4492 2
|
4495 |
+
4493 1
|
4496 |
+
4494 1
|
4497 |
+
4495 1
|
4498 |
+
4496 1
|
4499 |
+
4497 1
|
4500 |
+
4498 0
|
4501 |
+
4499 2
|
4502 |
+
4500 1
|
4503 |
+
4501 0
|
4504 |
+
4502 0
|
4505 |
+
4503 1
|
4506 |
+
4504 2
|
4507 |
+
4505 2
|
4508 |
+
4506 2
|
4509 |
+
4507 2
|
4510 |
+
4508 0
|
4511 |
+
4509 2
|
4512 |
+
4510 0
|
4513 |
+
4511 0
|
4514 |
+
4512 2
|
4515 |
+
4513 2
|
4516 |
+
4514 2
|
4517 |
+
4515 1
|
4518 |
+
4516 1
|
4519 |
+
4517 0
|
4520 |
+
4518 1
|
4521 |
+
4519 1
|
4522 |
+
4520 0
|
4523 |
+
4521 0
|
4524 |
+
4522 0
|
4525 |
+
4523 2
|
4526 |
+
4524 2
|
4527 |
+
4525 0
|
4528 |
+
4526 0
|
4529 |
+
4527 1
|
4530 |
+
4528 0
|
4531 |
+
4529 1
|
4532 |
+
4530 1
|
4533 |
+
4531 0
|
4534 |
+
4532 0
|
4535 |
+
4533 1
|
4536 |
+
4534 0
|
4537 |
+
4535 1
|
4538 |
+
4536 1
|
4539 |
+
4537 0
|
4540 |
+
4538 1
|
4541 |
+
4539 1
|
4542 |
+
4540 1
|
4543 |
+
4541 0
|
4544 |
+
4542 1
|
4545 |
+
4543 1
|
4546 |
+
4544 0
|
4547 |
+
4545 2
|
4548 |
+
4546 0
|
4549 |
+
4547 1
|
4550 |
+
4548 0
|
4551 |
+
4549 0
|
4552 |
+
4550 0
|
4553 |
+
4551 0
|
4554 |
+
4552 0
|
4555 |
+
4553 2
|
4556 |
+
4554 0
|
4557 |
+
4555 2
|
4558 |
+
4556 2
|
4559 |
+
4557 2
|
4560 |
+
4558 2
|
4561 |
+
4559 2
|
4562 |
+
4560 2
|
4563 |
+
4561 0
|
4564 |
+
4562 1
|
4565 |
+
4563 1
|
4566 |
+
4564 1
|
4567 |
+
4565 2
|
4568 |
+
4566 1
|
4569 |
+
4567 0
|
4570 |
+
4568 2
|
4571 |
+
4569 1
|
4572 |
+
4570 0
|
4573 |
+
4571 2
|
4574 |
+
4572 2
|
4575 |
+
4573 1
|
4576 |
+
4574 1
|
4577 |
+
4575 1
|
4578 |
+
4576 0
|
4579 |
+
4577 2
|
4580 |
+
4578 1
|
4581 |
+
4579 0
|
4582 |
+
4580 0
|
4583 |
+
4581 0
|
4584 |
+
4582 2
|
4585 |
+
4583 1
|
4586 |
+
4584 0
|
4587 |
+
4585 0
|
4588 |
+
4586 2
|
4589 |
+
4587 0
|
4590 |
+
4588 2
|
4591 |
+
4589 2
|
4592 |
+
4590 1
|
4593 |
+
4591 0
|
4594 |
+
4592 0
|
4595 |
+
4593 1
|
4596 |
+
4594 1
|
4597 |
+
4595 0
|
4598 |
+
4596 2
|
4599 |
+
4597 1
|
4600 |
+
4598 2
|
4601 |
+
4599 0
|
4602 |
+
4600 1
|
4603 |
+
4601 0
|
4604 |
+
4602 0
|
4605 |
+
4603 0
|
4606 |
+
4604 1
|
4607 |
+
4605 1
|
4608 |
+
4606 2
|
4609 |
+
4607 1
|
4610 |
+
4608 1
|
4611 |
+
4609 0
|
4612 |
+
4610 1
|
4613 |
+
4611 0
|
4614 |
+
4612 0
|
4615 |
+
4613 0
|
4616 |
+
4614 0
|
4617 |
+
4615 2
|
4618 |
+
4616 2
|
4619 |
+
4617 0
|
4620 |
+
4618 0
|
4621 |
+
4619 2
|
4622 |
+
4620 1
|
4623 |
+
4621 2
|
4624 |
+
4622 1
|
4625 |
+
4623 1
|
4626 |
+
4624 2
|
4627 |
+
4625 1
|
4628 |
+
4626 2
|
4629 |
+
4627 2
|
4630 |
+
4628 2
|
4631 |
+
4629 1
|
4632 |
+
4630 1
|
4633 |
+
4631 2
|
4634 |
+
4632 1
|
4635 |
+
4633 1
|
4636 |
+
4634 1
|
4637 |
+
4635 2
|
4638 |
+
4636 0
|
4639 |
+
4637 0
|
4640 |
+
4638 1
|
4641 |
+
4639 1
|
4642 |
+
4640 1
|
4643 |
+
4641 0
|
4644 |
+
4642 0
|
4645 |
+
4643 0
|
4646 |
+
4644 2
|
4647 |
+
4645 1
|
4648 |
+
4646 1
|
4649 |
+
4647 2
|
4650 |
+
4648 2
|
4651 |
+
4649 0
|
4652 |
+
4650 2
|
4653 |
+
4651 1
|
4654 |
+
4652 1
|
4655 |
+
4653 0
|
4656 |
+
4654 2
|
4657 |
+
4655 2
|
4658 |
+
4656 1
|
4659 |
+
4657 2
|
4660 |
+
4658 1
|
4661 |
+
4659 1
|
4662 |
+
4660 2
|
4663 |
+
4661 0
|
4664 |
+
4662 1
|
4665 |
+
4663 0
|
4666 |
+
4664 0
|
4667 |
+
4665 1
|
4668 |
+
4666 2
|
4669 |
+
4667 1
|
4670 |
+
4668 0
|
4671 |
+
4669 0
|
4672 |
+
4670 2
|
4673 |
+
4671 1
|
4674 |
+
4672 2
|
4675 |
+
4673 1
|
4676 |
+
4674 0
|
4677 |
+
4675 2
|
4678 |
+
4676 0
|
4679 |
+
4677 1
|
4680 |
+
4678 0
|
4681 |
+
4679 1
|
4682 |
+
4680 0
|
4683 |
+
4681 2
|
4684 |
+
4682 2
|
4685 |
+
4683 2
|
4686 |
+
4684 2
|
4687 |
+
4685 0
|
4688 |
+
4686 1
|
4689 |
+
4687 2
|
4690 |
+
4688 2
|
4691 |
+
4689 1
|
4692 |
+
4690 2
|
4693 |
+
4691 1
|
4694 |
+
4692 0
|
4695 |
+
4693 2
|
4696 |
+
4694 1
|
4697 |
+
4695 1
|
4698 |
+
4696 2
|
4699 |
+
4697 2
|
4700 |
+
4698 1
|
4701 |
+
4699 1
|
4702 |
+
4700 0
|
4703 |
+
4701 0
|
4704 |
+
4702 1
|
4705 |
+
4703 1
|
4706 |
+
4704 1
|
4707 |
+
4705 2
|
4708 |
+
4706 0
|
4709 |
+
4707 0
|
4710 |
+
4708 1
|
4711 |
+
4709 1
|
4712 |
+
4710 1
|
4713 |
+
4711 2
|
4714 |
+
4712 0
|
4715 |
+
4713 0
|
4716 |
+
4714 1
|
4717 |
+
4715 0
|
4718 |
+
4716 0
|
4719 |
+
4717 0
|
4720 |
+
4718 1
|
4721 |
+
4719 2
|
4722 |
+
4720 1
|
4723 |
+
4721 2
|
4724 |
+
4722 2
|
4725 |
+
4723 0
|
4726 |
+
4724 0
|
4727 |
+
4725 0
|
4728 |
+
4726 1
|
4729 |
+
4727 1
|
4730 |
+
4728 1
|
4731 |
+
4729 0
|
4732 |
+
4730 2
|
4733 |
+
4731 2
|
4734 |
+
4732 0
|
4735 |
+
4733 2
|
4736 |
+
4734 2
|
4737 |
+
4735 2
|
4738 |
+
4736 1
|
4739 |
+
4737 0
|
4740 |
+
4738 1
|
4741 |
+
4739 1
|
4742 |
+
4740 0
|
4743 |
+
4741 2
|
4744 |
+
4742 1
|
4745 |
+
4743 2
|
4746 |
+
4744 2
|
4747 |
+
4745 1
|
4748 |
+
4746 1
|
4749 |
+
4747 0
|
4750 |
+
4748 2
|
4751 |
+
4749 1
|
4752 |
+
4750 2
|
4753 |
+
4751 2
|
4754 |
+
4752 0
|
4755 |
+
4753 1
|
4756 |
+
4754 0
|
4757 |
+
4755 2
|
4758 |
+
4756 1
|
4759 |
+
4757 1
|
4760 |
+
4758 0
|
4761 |
+
4759 2
|
4762 |
+
4760 0
|
4763 |
+
4761 1
|
4764 |
+
4762 0
|
4765 |
+
4763 1
|
4766 |
+
4764 1
|
4767 |
+
4765 1
|
4768 |
+
4766 2
|
4769 |
+
4767 1
|
4770 |
+
4768 1
|
4771 |
+
4769 0
|
4772 |
+
4770 1
|
4773 |
+
4771 0
|
4774 |
+
4772 2
|
4775 |
+
4773 1
|
4776 |
+
4774 2
|
4777 |
+
4775 2
|
4778 |
+
4776 1
|
4779 |
+
4777 1
|
4780 |
+
4778 0
|
4781 |
+
4779 1
|
4782 |
+
4780 1
|
4783 |
+
4781 2
|
4784 |
+
4782 1
|
4785 |
+
4783 0
|
4786 |
+
4784 1
|
4787 |
+
4785 1
|
4788 |
+
4786 2
|
4789 |
+
4787 0
|
4790 |
+
4788 2
|
4791 |
+
4789 0
|
4792 |
+
4790 1
|
4793 |
+
4791 0
|
4794 |
+
4792 1
|
4795 |
+
4793 1
|
4796 |
+
4794 0
|
4797 |
+
4795 1
|
4798 |
+
4796 2
|
4799 |
+
4797 0
|
4800 |
+
4798 1
|
4801 |
+
4799 1
|
4802 |
+
4800 1
|
4803 |
+
4801 1
|
4804 |
+
4802 0
|
4805 |
+
4803 1
|
4806 |
+
4804 1
|
4807 |
+
4805 1
|
4808 |
+
4806 1
|
4809 |
+
4807 2
|
4810 |
+
4808 0
|
4811 |
+
4809 0
|
4812 |
+
4810 1
|
4813 |
+
4811 0
|
4814 |
+
4812 0
|
4815 |
+
4813 0
|
4816 |
+
4814 0
|
4817 |
+
4815 1
|
4818 |
+
4816 0
|
4819 |
+
4817 2
|
4820 |
+
4818 1
|
4821 |
+
4819 1
|
4822 |
+
4820 2
|
4823 |
+
4821 0
|
4824 |
+
4822 2
|
4825 |
+
4823 1
|
4826 |
+
4824 0
|
4827 |
+
4825 2
|
4828 |
+
4826 0
|
4829 |
+
4827 2
|
4830 |
+
4828 1
|
4831 |
+
4829 0
|
4832 |
+
4830 2
|
4833 |
+
4831 0
|
4834 |
+
4832 2
|
4835 |
+
4833 2
|
4836 |
+
4834 2
|
4837 |
+
4835 2
|
4838 |
+
4836 1
|
4839 |
+
4837 0
|
4840 |
+
4838 0
|
4841 |
+
4839 0
|
4842 |
+
4840 1
|
4843 |
+
4841 1
|
4844 |
+
4842 1
|
4845 |
+
4843 1
|
4846 |
+
4844 2
|
4847 |
+
4845 0
|
4848 |
+
4846 0
|
4849 |
+
4847 0
|
4850 |
+
4848 0
|
4851 |
+
4849 1
|
4852 |
+
4850 0
|
4853 |
+
4851 1
|
4854 |
+
4852 1
|
4855 |
+
4853 0
|
4856 |
+
4854 0
|
4857 |
+
4855 1
|
4858 |
+
4856 1
|
4859 |
+
4857 2
|
4860 |
+
4858 1
|
4861 |
+
4859 2
|
4862 |
+
4860 0
|
4863 |
+
4861 0
|
4864 |
+
4862 0
|
4865 |
+
4863 1
|
4866 |
+
4864 2
|
4867 |
+
4865 1
|
4868 |
+
4866 1
|
4869 |
+
4867 2
|
4870 |
+
4868 2
|
4871 |
+
4869 1
|
4872 |
+
4870 0
|
4873 |
+
4871 0
|
4874 |
+
4872 2
|
4875 |
+
4873 2
|
4876 |
+
4874 0
|
4877 |
+
4875 0
|
4878 |
+
4876 0
|
4879 |
+
4877 2
|
4880 |
+
4878 1
|
4881 |
+
4879 0
|
4882 |
+
4880 2
|
4883 |
+
4881 0
|
4884 |
+
4882 1
|
4885 |
+
4883 2
|
4886 |
+
4884 1
|
4887 |
+
4885 2
|
4888 |
+
4886 0
|
4889 |
+
4887 2
|
4890 |
+
4888 2
|
4891 |
+
4889 0
|
4892 |
+
4890 0
|
4893 |
+
4891 0
|
4894 |
+
4892 2
|
4895 |
+
4893 1
|
4896 |
+
4894 2
|
4897 |
+
4895 1
|
4898 |
+
4896 1
|
4899 |
+
4897 1
|
4900 |
+
4898 1
|
4901 |
+
4899 0
|
4902 |
+
4900 0
|
4903 |
+
4901 0
|
4904 |
+
4902 2
|
4905 |
+
4903 0
|
4906 |
+
4904 0
|
4907 |
+
4905 1
|
4908 |
+
4906 0
|
4909 |
+
4907 1
|
special_tokens_map.json
ADDED
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"cls_token": {
|
3 |
+
"content": "[CLS]",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": false,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"mask_token": {
|
10 |
+
"content": "[MASK]",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": false,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"pad_token": {
|
17 |
+
"content": "[PAD]",
|
18 |
+
"lstrip": false,
|
19 |
+
"normalized": false,
|
20 |
+
"rstrip": false,
|
21 |
+
"single_word": false
|
22 |
+
},
|
23 |
+
"sep_token": {
|
24 |
+
"content": "[SEP]",
|
25 |
+
"lstrip": false,
|
26 |
+
"normalized": false,
|
27 |
+
"rstrip": false,
|
28 |
+
"single_word": false
|
29 |
+
}
|
30 |
+
}
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
ADDED
@@ -0,0 +1,67 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"added_tokens_decoder": {
|
3 |
+
"0": {
|
4 |
+
"content": "[UNK]",
|
5 |
+
"lstrip": false,
|
6 |
+
"normalized": false,
|
7 |
+
"rstrip": false,
|
8 |
+
"single_word": false,
|
9 |
+
"special": true
|
10 |
+
},
|
11 |
+
"1": {
|
12 |
+
"content": "[CLS]",
|
13 |
+
"lstrip": false,
|
14 |
+
"normalized": false,
|
15 |
+
"rstrip": false,
|
16 |
+
"single_word": false,
|
17 |
+
"special": true
|
18 |
+
},
|
19 |
+
"2": {
|
20 |
+
"content": "[SEP]",
|
21 |
+
"lstrip": false,
|
22 |
+
"normalized": false,
|
23 |
+
"rstrip": false,
|
24 |
+
"single_word": false,
|
25 |
+
"special": true
|
26 |
+
},
|
27 |
+
"3": {
|
28 |
+
"content": "[PAD]",
|
29 |
+
"lstrip": false,
|
30 |
+
"normalized": false,
|
31 |
+
"rstrip": false,
|
32 |
+
"single_word": false,
|
33 |
+
"special": true
|
34 |
+
},
|
35 |
+
"4": {
|
36 |
+
"content": "[MASK]",
|
37 |
+
"lstrip": false,
|
38 |
+
"normalized": false,
|
39 |
+
"rstrip": false,
|
40 |
+
"single_word": false,
|
41 |
+
"special": true
|
42 |
+
},
|
43 |
+
"5": {
|
44 |
+
"content": "[PAR]",
|
45 |
+
"lstrip": false,
|
46 |
+
"normalized": false,
|
47 |
+
"rstrip": false,
|
48 |
+
"single_word": false,
|
49 |
+
"special": true
|
50 |
+
},
|
51 |
+
"6": {
|
52 |
+
"content": "[TAB]",
|
53 |
+
"lstrip": false,
|
54 |
+
"normalized": false,
|
55 |
+
"rstrip": false,
|
56 |
+
"single_word": false,
|
57 |
+
"special": true
|
58 |
+
}
|
59 |
+
},
|
60 |
+
"clean_up_tokenization_spaces": false,
|
61 |
+
"cls_token": "[CLS]",
|
62 |
+
"mask_token": "[MASK]",
|
63 |
+
"model_max_length": 1000000000000000019884624838656,
|
64 |
+
"pad_token": "[PAD]",
|
65 |
+
"sep_token": "[SEP]",
|
66 |
+
"tokenizer_class": "PreTrainedTokenizerFast"
|
67 |
+
}
|
train_results.json
ADDED
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"epoch": 5.0,
|
3 |
+
"total_flos": 1.158449687808768e+17,
|
4 |
+
"train_loss": 0.5704954994082295,
|
5 |
+
"train_runtime": 14864.7666,
|
6 |
+
"train_samples": 392702,
|
7 |
+
"train_samples_per_second": 132.092,
|
8 |
+
"train_steps_per_second": 16.512
|
9 |
+
}
|
trainer_state.json
ADDED
@@ -0,0 +1,3526 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"best_metric": 0.7337000966072083,
|
3 |
+
"best_model_checkpoint": "/home/ubuntu/utah/babylm-24/src/evaluation/results/finetune/DebertaV2-Base-10M_babylm-A/mnli/checkpoint-147264",
|
4 |
+
"epoch": 5.0,
|
5 |
+
"eval_steps": 500,
|
6 |
+
"global_step": 245440,
|
7 |
+
"is_hyper_param_search": false,
|
8 |
+
"is_local_process_zero": true,
|
9 |
+
"is_world_process_zero": true,
|
10 |
+
"log_history": [
|
11 |
+
{
|
12 |
+
"epoch": 0.010185788787483703,
|
13 |
+
"grad_norm": 3.972818613052368,
|
14 |
+
"learning_rate": 2.9938885267275102e-05,
|
15 |
+
"loss": 1.1058,
|
16 |
+
"step": 500
|
17 |
+
},
|
18 |
+
{
|
19 |
+
"epoch": 0.020371577574967405,
|
20 |
+
"grad_norm": 3.13022780418396,
|
21 |
+
"learning_rate": 2.9877770534550196e-05,
|
22 |
+
"loss": 1.0388,
|
23 |
+
"step": 1000
|
24 |
+
},
|
25 |
+
{
|
26 |
+
"epoch": 0.03055736636245111,
|
27 |
+
"grad_norm": 5.194879055023193,
|
28 |
+
"learning_rate": 2.9816655801825297e-05,
|
29 |
+
"loss": 0.999,
|
30 |
+
"step": 1500
|
31 |
+
},
|
32 |
+
{
|
33 |
+
"epoch": 0.04074315514993481,
|
34 |
+
"grad_norm": 6.858974933624268,
|
35 |
+
"learning_rate": 2.975554106910039e-05,
|
36 |
+
"loss": 0.9637,
|
37 |
+
"step": 2000
|
38 |
+
},
|
39 |
+
{
|
40 |
+
"epoch": 0.050928943937418515,
|
41 |
+
"grad_norm": 7.074331760406494,
|
42 |
+
"learning_rate": 2.9694426336375492e-05,
|
43 |
+
"loss": 0.9697,
|
44 |
+
"step": 2500
|
45 |
+
},
|
46 |
+
{
|
47 |
+
"epoch": 0.06111473272490222,
|
48 |
+
"grad_norm": 5.386518478393555,
|
49 |
+
"learning_rate": 2.9633311603650586e-05,
|
50 |
+
"loss": 0.9472,
|
51 |
+
"step": 3000
|
52 |
+
},
|
53 |
+
{
|
54 |
+
"epoch": 0.07130052151238592,
|
55 |
+
"grad_norm": 5.582085609436035,
|
56 |
+
"learning_rate": 2.9572196870925687e-05,
|
57 |
+
"loss": 0.9273,
|
58 |
+
"step": 3500
|
59 |
+
},
|
60 |
+
{
|
61 |
+
"epoch": 0.08148631029986962,
|
62 |
+
"grad_norm": 10.966410636901855,
|
63 |
+
"learning_rate": 2.951108213820078e-05,
|
64 |
+
"loss": 0.9316,
|
65 |
+
"step": 4000
|
66 |
+
},
|
67 |
+
{
|
68 |
+
"epoch": 0.09167209908735333,
|
69 |
+
"grad_norm": 7.840672492980957,
|
70 |
+
"learning_rate": 2.9449967405475882e-05,
|
71 |
+
"loss": 0.9237,
|
72 |
+
"step": 4500
|
73 |
+
},
|
74 |
+
{
|
75 |
+
"epoch": 0.10185788787483703,
|
76 |
+
"grad_norm": 7.442521095275879,
|
77 |
+
"learning_rate": 2.9388852672750976e-05,
|
78 |
+
"loss": 0.918,
|
79 |
+
"step": 5000
|
80 |
+
},
|
81 |
+
{
|
82 |
+
"epoch": 0.11204367666232073,
|
83 |
+
"grad_norm": 8.379261016845703,
|
84 |
+
"learning_rate": 2.9327737940026077e-05,
|
85 |
+
"loss": 0.8798,
|
86 |
+
"step": 5500
|
87 |
+
},
|
88 |
+
{
|
89 |
+
"epoch": 0.12222946544980444,
|
90 |
+
"grad_norm": 6.880769729614258,
|
91 |
+
"learning_rate": 2.926662320730117e-05,
|
92 |
+
"loss": 0.9045,
|
93 |
+
"step": 6000
|
94 |
+
},
|
95 |
+
{
|
96 |
+
"epoch": 0.13241525423728814,
|
97 |
+
"grad_norm": 5.965817451477051,
|
98 |
+
"learning_rate": 2.9205508474576272e-05,
|
99 |
+
"loss": 0.8941,
|
100 |
+
"step": 6500
|
101 |
+
},
|
102 |
+
{
|
103 |
+
"epoch": 0.14260104302477183,
|
104 |
+
"grad_norm": 9.67165756225586,
|
105 |
+
"learning_rate": 2.914439374185137e-05,
|
106 |
+
"loss": 0.878,
|
107 |
+
"step": 7000
|
108 |
+
},
|
109 |
+
{
|
110 |
+
"epoch": 0.15278683181225555,
|
111 |
+
"grad_norm": 6.4934892654418945,
|
112 |
+
"learning_rate": 2.9083279009126467e-05,
|
113 |
+
"loss": 0.8723,
|
114 |
+
"step": 7500
|
115 |
+
},
|
116 |
+
{
|
117 |
+
"epoch": 0.16297262059973924,
|
118 |
+
"grad_norm": 11.987072944641113,
|
119 |
+
"learning_rate": 2.9022164276401565e-05,
|
120 |
+
"loss": 0.8913,
|
121 |
+
"step": 8000
|
122 |
+
},
|
123 |
+
{
|
124 |
+
"epoch": 0.17315840938722293,
|
125 |
+
"grad_norm": 9.218280792236328,
|
126 |
+
"learning_rate": 2.8961049543676662e-05,
|
127 |
+
"loss": 0.868,
|
128 |
+
"step": 8500
|
129 |
+
},
|
130 |
+
{
|
131 |
+
"epoch": 0.18334419817470665,
|
132 |
+
"grad_norm": 6.697582244873047,
|
133 |
+
"learning_rate": 2.8899934810951763e-05,
|
134 |
+
"loss": 0.8618,
|
135 |
+
"step": 9000
|
136 |
+
},
|
137 |
+
{
|
138 |
+
"epoch": 0.19352998696219034,
|
139 |
+
"grad_norm": 8.035455703735352,
|
140 |
+
"learning_rate": 2.8838820078226858e-05,
|
141 |
+
"loss": 0.8641,
|
142 |
+
"step": 9500
|
143 |
+
},
|
144 |
+
{
|
145 |
+
"epoch": 0.20371577574967406,
|
146 |
+
"grad_norm": 6.890414237976074,
|
147 |
+
"learning_rate": 2.877770534550196e-05,
|
148 |
+
"loss": 0.8574,
|
149 |
+
"step": 10000
|
150 |
+
},
|
151 |
+
{
|
152 |
+
"epoch": 0.21390156453715775,
|
153 |
+
"grad_norm": 9.001678466796875,
|
154 |
+
"learning_rate": 2.8716590612777053e-05,
|
155 |
+
"loss": 0.8547,
|
156 |
+
"step": 10500
|
157 |
+
},
|
158 |
+
{
|
159 |
+
"epoch": 0.22408735332464147,
|
160 |
+
"grad_norm": 7.271191120147705,
|
161 |
+
"learning_rate": 2.8655475880052154e-05,
|
162 |
+
"loss": 0.8332,
|
163 |
+
"step": 11000
|
164 |
+
},
|
165 |
+
{
|
166 |
+
"epoch": 0.23427314211212516,
|
167 |
+
"grad_norm": 9.849822998046875,
|
168 |
+
"learning_rate": 2.8594361147327248e-05,
|
169 |
+
"loss": 0.8447,
|
170 |
+
"step": 11500
|
171 |
+
},
|
172 |
+
{
|
173 |
+
"epoch": 0.24445893089960888,
|
174 |
+
"grad_norm": 15.07829761505127,
|
175 |
+
"learning_rate": 2.853324641460235e-05,
|
176 |
+
"loss": 0.8251,
|
177 |
+
"step": 12000
|
178 |
+
},
|
179 |
+
{
|
180 |
+
"epoch": 0.25464471968709257,
|
181 |
+
"grad_norm": 6.886602878570557,
|
182 |
+
"learning_rate": 2.8472131681877446e-05,
|
183 |
+
"loss": 0.8444,
|
184 |
+
"step": 12500
|
185 |
+
},
|
186 |
+
{
|
187 |
+
"epoch": 0.2648305084745763,
|
188 |
+
"grad_norm": 9.94825267791748,
|
189 |
+
"learning_rate": 2.8411016949152544e-05,
|
190 |
+
"loss": 0.8504,
|
191 |
+
"step": 13000
|
192 |
+
},
|
193 |
+
{
|
194 |
+
"epoch": 0.27501629726205995,
|
195 |
+
"grad_norm": 8.0740966796875,
|
196 |
+
"learning_rate": 2.834990221642764e-05,
|
197 |
+
"loss": 0.8182,
|
198 |
+
"step": 13500
|
199 |
+
},
|
200 |
+
{
|
201 |
+
"epoch": 0.28520208604954367,
|
202 |
+
"grad_norm": 10.770129203796387,
|
203 |
+
"learning_rate": 2.828878748370274e-05,
|
204 |
+
"loss": 0.8373,
|
205 |
+
"step": 14000
|
206 |
+
},
|
207 |
+
{
|
208 |
+
"epoch": 0.2953878748370274,
|
209 |
+
"grad_norm": 9.214078903198242,
|
210 |
+
"learning_rate": 2.8227672750977836e-05,
|
211 |
+
"loss": 0.8089,
|
212 |
+
"step": 14500
|
213 |
+
},
|
214 |
+
{
|
215 |
+
"epoch": 0.3055736636245111,
|
216 |
+
"grad_norm": 14.231884956359863,
|
217 |
+
"learning_rate": 2.8166558018252934e-05,
|
218 |
+
"loss": 0.8337,
|
219 |
+
"step": 15000
|
220 |
+
},
|
221 |
+
{
|
222 |
+
"epoch": 0.31575945241199477,
|
223 |
+
"grad_norm": 9.379185676574707,
|
224 |
+
"learning_rate": 2.810544328552803e-05,
|
225 |
+
"loss": 0.8315,
|
226 |
+
"step": 15500
|
227 |
+
},
|
228 |
+
{
|
229 |
+
"epoch": 0.3259452411994785,
|
230 |
+
"grad_norm": 8.370843887329102,
|
231 |
+
"learning_rate": 2.804432855280313e-05,
|
232 |
+
"loss": 0.808,
|
233 |
+
"step": 16000
|
234 |
+
},
|
235 |
+
{
|
236 |
+
"epoch": 0.3361310299869622,
|
237 |
+
"grad_norm": 7.201444149017334,
|
238 |
+
"learning_rate": 2.7983213820078227e-05,
|
239 |
+
"loss": 0.808,
|
240 |
+
"step": 16500
|
241 |
+
},
|
242 |
+
{
|
243 |
+
"epoch": 0.34631681877444587,
|
244 |
+
"grad_norm": 10.696725845336914,
|
245 |
+
"learning_rate": 2.7922099087353324e-05,
|
246 |
+
"loss": 0.813,
|
247 |
+
"step": 17000
|
248 |
+
},
|
249 |
+
{
|
250 |
+
"epoch": 0.3565026075619296,
|
251 |
+
"grad_norm": 6.316204071044922,
|
252 |
+
"learning_rate": 2.7860984354628422e-05,
|
253 |
+
"loss": 0.8213,
|
254 |
+
"step": 17500
|
255 |
+
},
|
256 |
+
{
|
257 |
+
"epoch": 0.3666883963494133,
|
258 |
+
"grad_norm": 14.780086517333984,
|
259 |
+
"learning_rate": 2.779986962190352e-05,
|
260 |
+
"loss": 0.8165,
|
261 |
+
"step": 18000
|
262 |
+
},
|
263 |
+
{
|
264 |
+
"epoch": 0.376874185136897,
|
265 |
+
"grad_norm": 11.34945011138916,
|
266 |
+
"learning_rate": 2.773875488917862e-05,
|
267 |
+
"loss": 0.7943,
|
268 |
+
"step": 18500
|
269 |
+
},
|
270 |
+
{
|
271 |
+
"epoch": 0.3870599739243807,
|
272 |
+
"grad_norm": 7.136452674865723,
|
273 |
+
"learning_rate": 2.7677640156453718e-05,
|
274 |
+
"loss": 0.8017,
|
275 |
+
"step": 19000
|
276 |
+
},
|
277 |
+
{
|
278 |
+
"epoch": 0.3972457627118644,
|
279 |
+
"grad_norm": 6.209742069244385,
|
280 |
+
"learning_rate": 2.7616525423728815e-05,
|
281 |
+
"loss": 0.7941,
|
282 |
+
"step": 19500
|
283 |
+
},
|
284 |
+
{
|
285 |
+
"epoch": 0.4074315514993481,
|
286 |
+
"grad_norm": 8.923270225524902,
|
287 |
+
"learning_rate": 2.7555410691003913e-05,
|
288 |
+
"loss": 0.7826,
|
289 |
+
"step": 20000
|
290 |
+
},
|
291 |
+
{
|
292 |
+
"epoch": 0.41761734028683184,
|
293 |
+
"grad_norm": 6.371203899383545,
|
294 |
+
"learning_rate": 2.749429595827901e-05,
|
295 |
+
"loss": 0.7938,
|
296 |
+
"step": 20500
|
297 |
+
},
|
298 |
+
{
|
299 |
+
"epoch": 0.4278031290743155,
|
300 |
+
"grad_norm": 8.679354667663574,
|
301 |
+
"learning_rate": 2.7433181225554108e-05,
|
302 |
+
"loss": 0.7936,
|
303 |
+
"step": 21000
|
304 |
+
},
|
305 |
+
{
|
306 |
+
"epoch": 0.4379889178617992,
|
307 |
+
"grad_norm": 5.962934970855713,
|
308 |
+
"learning_rate": 2.7372066492829205e-05,
|
309 |
+
"loss": 0.7924,
|
310 |
+
"step": 21500
|
311 |
+
},
|
312 |
+
{
|
313 |
+
"epoch": 0.44817470664928294,
|
314 |
+
"grad_norm": 11.835200309753418,
|
315 |
+
"learning_rate": 2.7310951760104303e-05,
|
316 |
+
"loss": 0.7894,
|
317 |
+
"step": 22000
|
318 |
+
},
|
319 |
+
{
|
320 |
+
"epoch": 0.4583604954367666,
|
321 |
+
"grad_norm": 9.62806510925293,
|
322 |
+
"learning_rate": 2.72498370273794e-05,
|
323 |
+
"loss": 0.7963,
|
324 |
+
"step": 22500
|
325 |
+
},
|
326 |
+
{
|
327 |
+
"epoch": 0.4685462842242503,
|
328 |
+
"grad_norm": 7.6390509605407715,
|
329 |
+
"learning_rate": 2.7188722294654498e-05,
|
330 |
+
"loss": 0.783,
|
331 |
+
"step": 23000
|
332 |
+
},
|
333 |
+
{
|
334 |
+
"epoch": 0.47873207301173404,
|
335 |
+
"grad_norm": 5.826275825500488,
|
336 |
+
"learning_rate": 2.7127607561929596e-05,
|
337 |
+
"loss": 0.7859,
|
338 |
+
"step": 23500
|
339 |
+
},
|
340 |
+
{
|
341 |
+
"epoch": 0.48891786179921776,
|
342 |
+
"grad_norm": 5.199416637420654,
|
343 |
+
"learning_rate": 2.7066492829204693e-05,
|
344 |
+
"loss": 0.7989,
|
345 |
+
"step": 24000
|
346 |
+
},
|
347 |
+
{
|
348 |
+
"epoch": 0.4991036505867014,
|
349 |
+
"grad_norm": 12.38355541229248,
|
350 |
+
"learning_rate": 2.7005378096479794e-05,
|
351 |
+
"loss": 0.7864,
|
352 |
+
"step": 24500
|
353 |
+
},
|
354 |
+
{
|
355 |
+
"epoch": 0.5092894393741851,
|
356 |
+
"grad_norm": 7.700937271118164,
|
357 |
+
"learning_rate": 2.694426336375489e-05,
|
358 |
+
"loss": 0.7747,
|
359 |
+
"step": 25000
|
360 |
+
},
|
361 |
+
{
|
362 |
+
"epoch": 0.5194752281616688,
|
363 |
+
"grad_norm": 10.8720703125,
|
364 |
+
"learning_rate": 2.688314863102999e-05,
|
365 |
+
"loss": 0.7974,
|
366 |
+
"step": 25500
|
367 |
+
},
|
368 |
+
{
|
369 |
+
"epoch": 0.5296610169491526,
|
370 |
+
"grad_norm": 6.260867595672607,
|
371 |
+
"learning_rate": 2.6822033898305083e-05,
|
372 |
+
"loss": 0.7914,
|
373 |
+
"step": 26000
|
374 |
+
},
|
375 |
+
{
|
376 |
+
"epoch": 0.5398468057366362,
|
377 |
+
"grad_norm": 11.113471984863281,
|
378 |
+
"learning_rate": 2.6760919165580184e-05,
|
379 |
+
"loss": 0.782,
|
380 |
+
"step": 26500
|
381 |
+
},
|
382 |
+
{
|
383 |
+
"epoch": 0.5500325945241199,
|
384 |
+
"grad_norm": 7.497796058654785,
|
385 |
+
"learning_rate": 2.6699804432855282e-05,
|
386 |
+
"loss": 0.7642,
|
387 |
+
"step": 27000
|
388 |
+
},
|
389 |
+
{
|
390 |
+
"epoch": 0.5602183833116037,
|
391 |
+
"grad_norm": 6.1969170570373535,
|
392 |
+
"learning_rate": 2.663868970013038e-05,
|
393 |
+
"loss": 0.7693,
|
394 |
+
"step": 27500
|
395 |
+
},
|
396 |
+
{
|
397 |
+
"epoch": 0.5704041720990873,
|
398 |
+
"grad_norm": 9.965928077697754,
|
399 |
+
"learning_rate": 2.6577574967405477e-05,
|
400 |
+
"loss": 0.7681,
|
401 |
+
"step": 28000
|
402 |
+
},
|
403 |
+
{
|
404 |
+
"epoch": 0.5805899608865711,
|
405 |
+
"grad_norm": 5.052292346954346,
|
406 |
+
"learning_rate": 2.6516460234680575e-05,
|
407 |
+
"loss": 0.7664,
|
408 |
+
"step": 28500
|
409 |
+
},
|
410 |
+
{
|
411 |
+
"epoch": 0.5907757496740548,
|
412 |
+
"grad_norm": 6.7289276123046875,
|
413 |
+
"learning_rate": 2.6455345501955672e-05,
|
414 |
+
"loss": 0.7445,
|
415 |
+
"step": 29000
|
416 |
+
},
|
417 |
+
{
|
418 |
+
"epoch": 0.6009615384615384,
|
419 |
+
"grad_norm": 8.478078842163086,
|
420 |
+
"learning_rate": 2.639423076923077e-05,
|
421 |
+
"loss": 0.7859,
|
422 |
+
"step": 29500
|
423 |
+
},
|
424 |
+
{
|
425 |
+
"epoch": 0.6111473272490222,
|
426 |
+
"grad_norm": 6.2336106300354,
|
427 |
+
"learning_rate": 2.6333116036505867e-05,
|
428 |
+
"loss": 0.7598,
|
429 |
+
"step": 30000
|
430 |
+
},
|
431 |
+
{
|
432 |
+
"epoch": 0.6213331160365059,
|
433 |
+
"grad_norm": 11.82295036315918,
|
434 |
+
"learning_rate": 2.6272001303780965e-05,
|
435 |
+
"loss": 0.7507,
|
436 |
+
"step": 30500
|
437 |
+
},
|
438 |
+
{
|
439 |
+
"epoch": 0.6315189048239895,
|
440 |
+
"grad_norm": 8.545003890991211,
|
441 |
+
"learning_rate": 2.6210886571056066e-05,
|
442 |
+
"loss": 0.7769,
|
443 |
+
"step": 31000
|
444 |
+
},
|
445 |
+
{
|
446 |
+
"epoch": 0.6417046936114733,
|
447 |
+
"grad_norm": 9.849065780639648,
|
448 |
+
"learning_rate": 2.614977183833116e-05,
|
449 |
+
"loss": 0.7781,
|
450 |
+
"step": 31500
|
451 |
+
},
|
452 |
+
{
|
453 |
+
"epoch": 0.651890482398957,
|
454 |
+
"grad_norm": 7.641623020172119,
|
455 |
+
"learning_rate": 2.608865710560626e-05,
|
456 |
+
"loss": 0.7789,
|
457 |
+
"step": 32000
|
458 |
+
},
|
459 |
+
{
|
460 |
+
"epoch": 0.6620762711864406,
|
461 |
+
"grad_norm": 16.397785186767578,
|
462 |
+
"learning_rate": 2.6027542372881355e-05,
|
463 |
+
"loss": 0.7542,
|
464 |
+
"step": 32500
|
465 |
+
},
|
466 |
+
{
|
467 |
+
"epoch": 0.6722620599739244,
|
468 |
+
"grad_norm": 11.119671821594238,
|
469 |
+
"learning_rate": 2.5966427640156456e-05,
|
470 |
+
"loss": 0.7716,
|
471 |
+
"step": 33000
|
472 |
+
},
|
473 |
+
{
|
474 |
+
"epoch": 0.6824478487614081,
|
475 |
+
"grad_norm": 12.466675758361816,
|
476 |
+
"learning_rate": 2.590531290743155e-05,
|
477 |
+
"loss": 0.7743,
|
478 |
+
"step": 33500
|
479 |
+
},
|
480 |
+
{
|
481 |
+
"epoch": 0.6926336375488917,
|
482 |
+
"grad_norm": 8.151625633239746,
|
483 |
+
"learning_rate": 2.584419817470665e-05,
|
484 |
+
"loss": 0.7581,
|
485 |
+
"step": 34000
|
486 |
+
},
|
487 |
+
{
|
488 |
+
"epoch": 0.7028194263363755,
|
489 |
+
"grad_norm": 7.743143558502197,
|
490 |
+
"learning_rate": 2.5783083441981745e-05,
|
491 |
+
"loss": 0.7599,
|
492 |
+
"step": 34500
|
493 |
+
},
|
494 |
+
{
|
495 |
+
"epoch": 0.7130052151238592,
|
496 |
+
"grad_norm": 5.249680995941162,
|
497 |
+
"learning_rate": 2.5721968709256846e-05,
|
498 |
+
"loss": 0.7483,
|
499 |
+
"step": 35000
|
500 |
+
},
|
501 |
+
{
|
502 |
+
"epoch": 0.7231910039113429,
|
503 |
+
"grad_norm": 6.738178730010986,
|
504 |
+
"learning_rate": 2.566085397653194e-05,
|
505 |
+
"loss": 0.7781,
|
506 |
+
"step": 35500
|
507 |
+
},
|
508 |
+
{
|
509 |
+
"epoch": 0.7333767926988266,
|
510 |
+
"grad_norm": 6.7923102378845215,
|
511 |
+
"learning_rate": 2.559973924380704e-05,
|
512 |
+
"loss": 0.7464,
|
513 |
+
"step": 36000
|
514 |
+
},
|
515 |
+
{
|
516 |
+
"epoch": 0.7435625814863103,
|
517 |
+
"grad_norm": 15.575695991516113,
|
518 |
+
"learning_rate": 2.553862451108214e-05,
|
519 |
+
"loss": 0.7571,
|
520 |
+
"step": 36500
|
521 |
+
},
|
522 |
+
{
|
523 |
+
"epoch": 0.753748370273794,
|
524 |
+
"grad_norm": 13.128719329833984,
|
525 |
+
"learning_rate": 2.5477509778357236e-05,
|
526 |
+
"loss": 0.7625,
|
527 |
+
"step": 37000
|
528 |
+
},
|
529 |
+
{
|
530 |
+
"epoch": 0.7639341590612777,
|
531 |
+
"grad_norm": 4.938621997833252,
|
532 |
+
"learning_rate": 2.5416395045632337e-05,
|
533 |
+
"loss": 0.7477,
|
534 |
+
"step": 37500
|
535 |
+
},
|
536 |
+
{
|
537 |
+
"epoch": 0.7741199478487614,
|
538 |
+
"grad_norm": 8.46985912322998,
|
539 |
+
"learning_rate": 2.535528031290743e-05,
|
540 |
+
"loss": 0.7419,
|
541 |
+
"step": 38000
|
542 |
+
},
|
543 |
+
{
|
544 |
+
"epoch": 0.7843057366362451,
|
545 |
+
"grad_norm": 18.881208419799805,
|
546 |
+
"learning_rate": 2.5294165580182532e-05,
|
547 |
+
"loss": 0.7403,
|
548 |
+
"step": 38500
|
549 |
+
},
|
550 |
+
{
|
551 |
+
"epoch": 0.7944915254237288,
|
552 |
+
"grad_norm": 9.986156463623047,
|
553 |
+
"learning_rate": 2.5233050847457626e-05,
|
554 |
+
"loss": 0.7508,
|
555 |
+
"step": 39000
|
556 |
+
},
|
557 |
+
{
|
558 |
+
"epoch": 0.8046773142112125,
|
559 |
+
"grad_norm": 11.625401496887207,
|
560 |
+
"learning_rate": 2.5171936114732727e-05,
|
561 |
+
"loss": 0.7426,
|
562 |
+
"step": 39500
|
563 |
+
},
|
564 |
+
{
|
565 |
+
"epoch": 0.8148631029986962,
|
566 |
+
"grad_norm": 9.926522254943848,
|
567 |
+
"learning_rate": 2.511082138200782e-05,
|
568 |
+
"loss": 0.7365,
|
569 |
+
"step": 40000
|
570 |
+
},
|
571 |
+
{
|
572 |
+
"epoch": 0.8250488917861799,
|
573 |
+
"grad_norm": 8.104277610778809,
|
574 |
+
"learning_rate": 2.5049706649282922e-05,
|
575 |
+
"loss": 0.7532,
|
576 |
+
"step": 40500
|
577 |
+
},
|
578 |
+
{
|
579 |
+
"epoch": 0.8352346805736637,
|
580 |
+
"grad_norm": 5.898177623748779,
|
581 |
+
"learning_rate": 2.4988591916558017e-05,
|
582 |
+
"loss": 0.7657,
|
583 |
+
"step": 41000
|
584 |
+
},
|
585 |
+
{
|
586 |
+
"epoch": 0.8454204693611473,
|
587 |
+
"grad_norm": 6.856070041656494,
|
588 |
+
"learning_rate": 2.4927477183833118e-05,
|
589 |
+
"loss": 0.7438,
|
590 |
+
"step": 41500
|
591 |
+
},
|
592 |
+
{
|
593 |
+
"epoch": 0.855606258148631,
|
594 |
+
"grad_norm": 11.806628227233887,
|
595 |
+
"learning_rate": 2.4866362451108212e-05,
|
596 |
+
"loss": 0.7348,
|
597 |
+
"step": 42000
|
598 |
+
},
|
599 |
+
{
|
600 |
+
"epoch": 0.8657920469361148,
|
601 |
+
"grad_norm": 12.08478832244873,
|
602 |
+
"learning_rate": 2.4805247718383313e-05,
|
603 |
+
"loss": 0.7277,
|
604 |
+
"step": 42500
|
605 |
+
},
|
606 |
+
{
|
607 |
+
"epoch": 0.8759778357235984,
|
608 |
+
"grad_norm": 5.707648277282715,
|
609 |
+
"learning_rate": 2.474413298565841e-05,
|
610 |
+
"loss": 0.7458,
|
611 |
+
"step": 43000
|
612 |
+
},
|
613 |
+
{
|
614 |
+
"epoch": 0.8861636245110821,
|
615 |
+
"grad_norm": 5.36006498336792,
|
616 |
+
"learning_rate": 2.4683018252933508e-05,
|
617 |
+
"loss": 0.7446,
|
618 |
+
"step": 43500
|
619 |
+
},
|
620 |
+
{
|
621 |
+
"epoch": 0.8963494132985659,
|
622 |
+
"grad_norm": 6.603504180908203,
|
623 |
+
"learning_rate": 2.4621903520208605e-05,
|
624 |
+
"loss": 0.7278,
|
625 |
+
"step": 44000
|
626 |
+
},
|
627 |
+
{
|
628 |
+
"epoch": 0.9065352020860495,
|
629 |
+
"grad_norm": 7.789953231811523,
|
630 |
+
"learning_rate": 2.4560788787483703e-05,
|
631 |
+
"loss": 0.7375,
|
632 |
+
"step": 44500
|
633 |
+
},
|
634 |
+
{
|
635 |
+
"epoch": 0.9167209908735332,
|
636 |
+
"grad_norm": 11.424116134643555,
|
637 |
+
"learning_rate": 2.4499674054758804e-05,
|
638 |
+
"loss": 0.7297,
|
639 |
+
"step": 45000
|
640 |
+
},
|
641 |
+
{
|
642 |
+
"epoch": 0.926906779661017,
|
643 |
+
"grad_norm": 6.7079620361328125,
|
644 |
+
"learning_rate": 2.4438559322033898e-05,
|
645 |
+
"loss": 0.7411,
|
646 |
+
"step": 45500
|
647 |
+
},
|
648 |
+
{
|
649 |
+
"epoch": 0.9370925684485006,
|
650 |
+
"grad_norm": 11.931846618652344,
|
651 |
+
"learning_rate": 2.4377444589309e-05,
|
652 |
+
"loss": 0.749,
|
653 |
+
"step": 46000
|
654 |
+
},
|
655 |
+
{
|
656 |
+
"epoch": 0.9472783572359843,
|
657 |
+
"grad_norm": 8.87076473236084,
|
658 |
+
"learning_rate": 2.4316329856584093e-05,
|
659 |
+
"loss": 0.7312,
|
660 |
+
"step": 46500
|
661 |
+
},
|
662 |
+
{
|
663 |
+
"epoch": 0.9574641460234681,
|
664 |
+
"grad_norm": 12.787999153137207,
|
665 |
+
"learning_rate": 2.4255215123859194e-05,
|
666 |
+
"loss": 0.7244,
|
667 |
+
"step": 47000
|
668 |
+
},
|
669 |
+
{
|
670 |
+
"epoch": 0.9676499348109517,
|
671 |
+
"grad_norm": 6.917387962341309,
|
672 |
+
"learning_rate": 2.4194100391134288e-05,
|
673 |
+
"loss": 0.733,
|
674 |
+
"step": 47500
|
675 |
+
},
|
676 |
+
{
|
677 |
+
"epoch": 0.9778357235984355,
|
678 |
+
"grad_norm": 7.564793109893799,
|
679 |
+
"learning_rate": 2.413298565840939e-05,
|
680 |
+
"loss": 0.7253,
|
681 |
+
"step": 48000
|
682 |
+
},
|
683 |
+
{
|
684 |
+
"epoch": 0.9880215123859192,
|
685 |
+
"grad_norm": 7.665174961090088,
|
686 |
+
"learning_rate": 2.4071870925684483e-05,
|
687 |
+
"loss": 0.7191,
|
688 |
+
"step": 48500
|
689 |
+
},
|
690 |
+
{
|
691 |
+
"epoch": 0.9982073011734028,
|
692 |
+
"grad_norm": 9.235818862915039,
|
693 |
+
"learning_rate": 2.4010756192959584e-05,
|
694 |
+
"loss": 0.7288,
|
695 |
+
"step": 49000
|
696 |
+
},
|
697 |
+
{
|
698 |
+
"epoch": 1.0,
|
699 |
+
"eval_accuracy": 0.6841890811920166,
|
700 |
+
"eval_loss": 0.7356013059616089,
|
701 |
+
"eval_runtime": 8.0037,
|
702 |
+
"eval_samples_per_second": 613.213,
|
703 |
+
"eval_steps_per_second": 76.714,
|
704 |
+
"step": 49088
|
705 |
+
},
|
706 |
+
{
|
707 |
+
"epoch": 1.0083930899608866,
|
708 |
+
"grad_norm": 9.232441902160645,
|
709 |
+
"learning_rate": 2.3949641460234682e-05,
|
710 |
+
"loss": 0.666,
|
711 |
+
"step": 49500
|
712 |
+
},
|
713 |
+
{
|
714 |
+
"epoch": 1.0185788787483703,
|
715 |
+
"grad_norm": 5.098124980926514,
|
716 |
+
"learning_rate": 2.388852672750978e-05,
|
717 |
+
"loss": 0.6762,
|
718 |
+
"step": 50000
|
719 |
+
},
|
720 |
+
{
|
721 |
+
"epoch": 1.028764667535854,
|
722 |
+
"grad_norm": 6.507536888122559,
|
723 |
+
"learning_rate": 2.3827411994784877e-05,
|
724 |
+
"loss": 0.6509,
|
725 |
+
"step": 50500
|
726 |
+
},
|
727 |
+
{
|
728 |
+
"epoch": 1.0389504563233376,
|
729 |
+
"grad_norm": 12.516643524169922,
|
730 |
+
"learning_rate": 2.3766297262059974e-05,
|
731 |
+
"loss": 0.6569,
|
732 |
+
"step": 51000
|
733 |
+
},
|
734 |
+
{
|
735 |
+
"epoch": 1.0491362451108215,
|
736 |
+
"grad_norm": 11.644853591918945,
|
737 |
+
"learning_rate": 2.3705182529335072e-05,
|
738 |
+
"loss": 0.6498,
|
739 |
+
"step": 51500
|
740 |
+
},
|
741 |
+
{
|
742 |
+
"epoch": 1.0593220338983051,
|
743 |
+
"grad_norm": 15.303723335266113,
|
744 |
+
"learning_rate": 2.364406779661017e-05,
|
745 |
+
"loss": 0.6565,
|
746 |
+
"step": 52000
|
747 |
+
},
|
748 |
+
{
|
749 |
+
"epoch": 1.0695078226857888,
|
750 |
+
"grad_norm": 11.212811470031738,
|
751 |
+
"learning_rate": 2.3582953063885267e-05,
|
752 |
+
"loss": 0.6492,
|
753 |
+
"step": 52500
|
754 |
+
},
|
755 |
+
{
|
756 |
+
"epoch": 1.0796936114732725,
|
757 |
+
"grad_norm": 7.695071220397949,
|
758 |
+
"learning_rate": 2.3521838331160365e-05,
|
759 |
+
"loss": 0.6618,
|
760 |
+
"step": 53000
|
761 |
+
},
|
762 |
+
{
|
763 |
+
"epoch": 1.0898794002607561,
|
764 |
+
"grad_norm": 8.234328269958496,
|
765 |
+
"learning_rate": 2.3460723598435462e-05,
|
766 |
+
"loss": 0.6709,
|
767 |
+
"step": 53500
|
768 |
+
},
|
769 |
+
{
|
770 |
+
"epoch": 1.1000651890482398,
|
771 |
+
"grad_norm": 12.574545860290527,
|
772 |
+
"learning_rate": 2.339960886571056e-05,
|
773 |
+
"loss": 0.6518,
|
774 |
+
"step": 54000
|
775 |
+
},
|
776 |
+
{
|
777 |
+
"epoch": 1.1102509778357237,
|
778 |
+
"grad_norm": 5.914109706878662,
|
779 |
+
"learning_rate": 2.333849413298566e-05,
|
780 |
+
"loss": 0.6494,
|
781 |
+
"step": 54500
|
782 |
+
},
|
783 |
+
{
|
784 |
+
"epoch": 1.1204367666232073,
|
785 |
+
"grad_norm": 11.49263858795166,
|
786 |
+
"learning_rate": 2.3277379400260758e-05,
|
787 |
+
"loss": 0.6723,
|
788 |
+
"step": 55000
|
789 |
+
},
|
790 |
+
{
|
791 |
+
"epoch": 1.130622555410691,
|
792 |
+
"grad_norm": 11.228019714355469,
|
793 |
+
"learning_rate": 2.3216264667535856e-05,
|
794 |
+
"loss": 0.6328,
|
795 |
+
"step": 55500
|
796 |
+
},
|
797 |
+
{
|
798 |
+
"epoch": 1.1408083441981747,
|
799 |
+
"grad_norm": 14.478355407714844,
|
800 |
+
"learning_rate": 2.3155149934810953e-05,
|
801 |
+
"loss": 0.667,
|
802 |
+
"step": 56000
|
803 |
+
},
|
804 |
+
{
|
805 |
+
"epoch": 1.1509941329856583,
|
806 |
+
"grad_norm": 18.817468643188477,
|
807 |
+
"learning_rate": 2.309403520208605e-05,
|
808 |
+
"loss": 0.6674,
|
809 |
+
"step": 56500
|
810 |
+
},
|
811 |
+
{
|
812 |
+
"epoch": 1.161179921773142,
|
813 |
+
"grad_norm": 12.483678817749023,
|
814 |
+
"learning_rate": 2.303292046936115e-05,
|
815 |
+
"loss": 0.6677,
|
816 |
+
"step": 57000
|
817 |
+
},
|
818 |
+
{
|
819 |
+
"epoch": 1.1713657105606259,
|
820 |
+
"grad_norm": 7.133495807647705,
|
821 |
+
"learning_rate": 2.2971805736636246e-05,
|
822 |
+
"loss": 0.642,
|
823 |
+
"step": 57500
|
824 |
+
},
|
825 |
+
{
|
826 |
+
"epoch": 1.1815514993481095,
|
827 |
+
"grad_norm": 7.099282741546631,
|
828 |
+
"learning_rate": 2.2910691003911343e-05,
|
829 |
+
"loss": 0.664,
|
830 |
+
"step": 58000
|
831 |
+
},
|
832 |
+
{
|
833 |
+
"epoch": 1.1917372881355932,
|
834 |
+
"grad_norm": 9.188867568969727,
|
835 |
+
"learning_rate": 2.284957627118644e-05,
|
836 |
+
"loss": 0.6703,
|
837 |
+
"step": 58500
|
838 |
+
},
|
839 |
+
{
|
840 |
+
"epoch": 1.2019230769230769,
|
841 |
+
"grad_norm": 5.470687389373779,
|
842 |
+
"learning_rate": 2.278846153846154e-05,
|
843 |
+
"loss": 0.6575,
|
844 |
+
"step": 59000
|
845 |
+
},
|
846 |
+
{
|
847 |
+
"epoch": 1.2121088657105605,
|
848 |
+
"grad_norm": 8.013532638549805,
|
849 |
+
"learning_rate": 2.2727346805736636e-05,
|
850 |
+
"loss": 0.6465,
|
851 |
+
"step": 59500
|
852 |
+
},
|
853 |
+
{
|
854 |
+
"epoch": 1.2222946544980444,
|
855 |
+
"grad_norm": 8.846539497375488,
|
856 |
+
"learning_rate": 2.2666232073011734e-05,
|
857 |
+
"loss": 0.6682,
|
858 |
+
"step": 60000
|
859 |
+
},
|
860 |
+
{
|
861 |
+
"epoch": 1.232480443285528,
|
862 |
+
"grad_norm": 11.2225341796875,
|
863 |
+
"learning_rate": 2.260511734028683e-05,
|
864 |
+
"loss": 0.653,
|
865 |
+
"step": 60500
|
866 |
+
},
|
867 |
+
{
|
868 |
+
"epoch": 1.2426662320730117,
|
869 |
+
"grad_norm": 12.363032341003418,
|
870 |
+
"learning_rate": 2.254400260756193e-05,
|
871 |
+
"loss": 0.6807,
|
872 |
+
"step": 61000
|
873 |
+
},
|
874 |
+
{
|
875 |
+
"epoch": 1.2528520208604954,
|
876 |
+
"grad_norm": 8.01014232635498,
|
877 |
+
"learning_rate": 2.248288787483703e-05,
|
878 |
+
"loss": 0.6625,
|
879 |
+
"step": 61500
|
880 |
+
},
|
881 |
+
{
|
882 |
+
"epoch": 1.263037809647979,
|
883 |
+
"grad_norm": 20.960845947265625,
|
884 |
+
"learning_rate": 2.2421773142112124e-05,
|
885 |
+
"loss": 0.6524,
|
886 |
+
"step": 62000
|
887 |
+
},
|
888 |
+
{
|
889 |
+
"epoch": 1.2732235984354627,
|
890 |
+
"grad_norm": 13.620718002319336,
|
891 |
+
"learning_rate": 2.2360658409387225e-05,
|
892 |
+
"loss": 0.6625,
|
893 |
+
"step": 62500
|
894 |
+
},
|
895 |
+
{
|
896 |
+
"epoch": 1.2834093872229466,
|
897 |
+
"grad_norm": 18.55965232849121,
|
898 |
+
"learning_rate": 2.2299543676662322e-05,
|
899 |
+
"loss": 0.6434,
|
900 |
+
"step": 63000
|
901 |
+
},
|
902 |
+
{
|
903 |
+
"epoch": 1.2935951760104303,
|
904 |
+
"grad_norm": 12.919173240661621,
|
905 |
+
"learning_rate": 2.223842894393742e-05,
|
906 |
+
"loss": 0.652,
|
907 |
+
"step": 63500
|
908 |
+
},
|
909 |
+
{
|
910 |
+
"epoch": 1.303780964797914,
|
911 |
+
"grad_norm": 19.561386108398438,
|
912 |
+
"learning_rate": 2.2177314211212517e-05,
|
913 |
+
"loss": 0.6498,
|
914 |
+
"step": 64000
|
915 |
+
},
|
916 |
+
{
|
917 |
+
"epoch": 1.3139667535853976,
|
918 |
+
"grad_norm": 10.529995918273926,
|
919 |
+
"learning_rate": 2.2116199478487615e-05,
|
920 |
+
"loss": 0.666,
|
921 |
+
"step": 64500
|
922 |
+
},
|
923 |
+
{
|
924 |
+
"epoch": 1.3241525423728815,
|
925 |
+
"grad_norm": 13.268050193786621,
|
926 |
+
"learning_rate": 2.2055084745762713e-05,
|
927 |
+
"loss": 0.6499,
|
928 |
+
"step": 65000
|
929 |
+
},
|
930 |
+
{
|
931 |
+
"epoch": 1.3343383311603652,
|
932 |
+
"grad_norm": 6.3072285652160645,
|
933 |
+
"learning_rate": 2.199397001303781e-05,
|
934 |
+
"loss": 0.6798,
|
935 |
+
"step": 65500
|
936 |
+
},
|
937 |
+
{
|
938 |
+
"epoch": 1.3445241199478488,
|
939 |
+
"grad_norm": 5.304599761962891,
|
940 |
+
"learning_rate": 2.1932855280312908e-05,
|
941 |
+
"loss": 0.6559,
|
942 |
+
"step": 66000
|
943 |
+
},
|
944 |
+
{
|
945 |
+
"epoch": 1.3547099087353325,
|
946 |
+
"grad_norm": 13.698792457580566,
|
947 |
+
"learning_rate": 2.1871740547588005e-05,
|
948 |
+
"loss": 0.6441,
|
949 |
+
"step": 66500
|
950 |
+
},
|
951 |
+
{
|
952 |
+
"epoch": 1.3648956975228161,
|
953 |
+
"grad_norm": 8.951632499694824,
|
954 |
+
"learning_rate": 2.1810625814863103e-05,
|
955 |
+
"loss": 0.6564,
|
956 |
+
"step": 67000
|
957 |
+
},
|
958 |
+
{
|
959 |
+
"epoch": 1.3750814863102998,
|
960 |
+
"grad_norm": 5.84067440032959,
|
961 |
+
"learning_rate": 2.17495110821382e-05,
|
962 |
+
"loss": 0.6686,
|
963 |
+
"step": 67500
|
964 |
+
},
|
965 |
+
{
|
966 |
+
"epoch": 1.3852672750977835,
|
967 |
+
"grad_norm": 6.442080497741699,
|
968 |
+
"learning_rate": 2.16883963494133e-05,
|
969 |
+
"loss": 0.6517,
|
970 |
+
"step": 68000
|
971 |
+
},
|
972 |
+
{
|
973 |
+
"epoch": 1.3954530638852674,
|
974 |
+
"grad_norm": 5.514212608337402,
|
975 |
+
"learning_rate": 2.1627281616688395e-05,
|
976 |
+
"loss": 0.6664,
|
977 |
+
"step": 68500
|
978 |
+
},
|
979 |
+
{
|
980 |
+
"epoch": 1.405638852672751,
|
981 |
+
"grad_norm": 14.150158882141113,
|
982 |
+
"learning_rate": 2.1566166883963496e-05,
|
983 |
+
"loss": 0.6561,
|
984 |
+
"step": 69000
|
985 |
+
},
|
986 |
+
{
|
987 |
+
"epoch": 1.4158246414602347,
|
988 |
+
"grad_norm": 17.196884155273438,
|
989 |
+
"learning_rate": 2.150505215123859e-05,
|
990 |
+
"loss": 0.6513,
|
991 |
+
"step": 69500
|
992 |
+
},
|
993 |
+
{
|
994 |
+
"epoch": 1.4260104302477183,
|
995 |
+
"grad_norm": 6.181870937347412,
|
996 |
+
"learning_rate": 2.144393741851369e-05,
|
997 |
+
"loss": 0.65,
|
998 |
+
"step": 70000
|
999 |
+
},
|
1000 |
+
{
|
1001 |
+
"epoch": 1.436196219035202,
|
1002 |
+
"grad_norm": 14.865707397460938,
|
1003 |
+
"learning_rate": 2.1382822685788786e-05,
|
1004 |
+
"loss": 0.6596,
|
1005 |
+
"step": 70500
|
1006 |
+
},
|
1007 |
+
{
|
1008 |
+
"epoch": 1.4463820078226859,
|
1009 |
+
"grad_norm": 15.808574676513672,
|
1010 |
+
"learning_rate": 2.1321707953063886e-05,
|
1011 |
+
"loss": 0.6559,
|
1012 |
+
"step": 71000
|
1013 |
+
},
|
1014 |
+
{
|
1015 |
+
"epoch": 1.4565677966101696,
|
1016 |
+
"grad_norm": 16.76003074645996,
|
1017 |
+
"learning_rate": 2.1260593220338984e-05,
|
1018 |
+
"loss": 0.6467,
|
1019 |
+
"step": 71500
|
1020 |
+
},
|
1021 |
+
{
|
1022 |
+
"epoch": 1.4667535853976532,
|
1023 |
+
"grad_norm": 21.466825485229492,
|
1024 |
+
"learning_rate": 2.119947848761408e-05,
|
1025 |
+
"loss": 0.6579,
|
1026 |
+
"step": 72000
|
1027 |
+
},
|
1028 |
+
{
|
1029 |
+
"epoch": 1.4769393741851369,
|
1030 |
+
"grad_norm": 18.81052589416504,
|
1031 |
+
"learning_rate": 2.113836375488918e-05,
|
1032 |
+
"loss": 0.6699,
|
1033 |
+
"step": 72500
|
1034 |
+
},
|
1035 |
+
{
|
1036 |
+
"epoch": 1.4871251629726205,
|
1037 |
+
"grad_norm": 10.334957122802734,
|
1038 |
+
"learning_rate": 2.1077249022164277e-05,
|
1039 |
+
"loss": 0.655,
|
1040 |
+
"step": 73000
|
1041 |
+
},
|
1042 |
+
{
|
1043 |
+
"epoch": 1.4973109517601042,
|
1044 |
+
"grad_norm": 14.836852073669434,
|
1045 |
+
"learning_rate": 2.1016134289439378e-05,
|
1046 |
+
"loss": 0.66,
|
1047 |
+
"step": 73500
|
1048 |
+
},
|
1049 |
+
{
|
1050 |
+
"epoch": 1.5074967405475879,
|
1051 |
+
"grad_norm": 21.439233779907227,
|
1052 |
+
"learning_rate": 2.0955019556714472e-05,
|
1053 |
+
"loss": 0.6486,
|
1054 |
+
"step": 74000
|
1055 |
+
},
|
1056 |
+
{
|
1057 |
+
"epoch": 1.5176825293350718,
|
1058 |
+
"grad_norm": 21.09993553161621,
|
1059 |
+
"learning_rate": 2.0893904823989573e-05,
|
1060 |
+
"loss": 0.6432,
|
1061 |
+
"step": 74500
|
1062 |
+
},
|
1063 |
+
{
|
1064 |
+
"epoch": 1.5278683181225554,
|
1065 |
+
"grad_norm": 8.926955223083496,
|
1066 |
+
"learning_rate": 2.0832790091264667e-05,
|
1067 |
+
"loss": 0.6418,
|
1068 |
+
"step": 75000
|
1069 |
+
},
|
1070 |
+
{
|
1071 |
+
"epoch": 1.538054106910039,
|
1072 |
+
"grad_norm": 3.551163911819458,
|
1073 |
+
"learning_rate": 2.0771675358539768e-05,
|
1074 |
+
"loss": 0.6418,
|
1075 |
+
"step": 75500
|
1076 |
+
},
|
1077 |
+
{
|
1078 |
+
"epoch": 1.548239895697523,
|
1079 |
+
"grad_norm": 10.46932315826416,
|
1080 |
+
"learning_rate": 2.0710560625814862e-05,
|
1081 |
+
"loss": 0.657,
|
1082 |
+
"step": 76000
|
1083 |
+
},
|
1084 |
+
{
|
1085 |
+
"epoch": 1.5584256844850066,
|
1086 |
+
"grad_norm": 10.022995948791504,
|
1087 |
+
"learning_rate": 2.0649445893089963e-05,
|
1088 |
+
"loss": 0.6564,
|
1089 |
+
"step": 76500
|
1090 |
+
},
|
1091 |
+
{
|
1092 |
+
"epoch": 1.5686114732724903,
|
1093 |
+
"grad_norm": 15.299884796142578,
|
1094 |
+
"learning_rate": 2.0588331160365057e-05,
|
1095 |
+
"loss": 0.6535,
|
1096 |
+
"step": 77000
|
1097 |
+
},
|
1098 |
+
{
|
1099 |
+
"epoch": 1.578797262059974,
|
1100 |
+
"grad_norm": 11.149444580078125,
|
1101 |
+
"learning_rate": 2.0527216427640158e-05,
|
1102 |
+
"loss": 0.6409,
|
1103 |
+
"step": 77500
|
1104 |
+
},
|
1105 |
+
{
|
1106 |
+
"epoch": 1.5889830508474576,
|
1107 |
+
"grad_norm": 4.011179447174072,
|
1108 |
+
"learning_rate": 2.0466101694915252e-05,
|
1109 |
+
"loss": 0.6299,
|
1110 |
+
"step": 78000
|
1111 |
+
},
|
1112 |
+
{
|
1113 |
+
"epoch": 1.5991688396349413,
|
1114 |
+
"grad_norm": 13.744772911071777,
|
1115 |
+
"learning_rate": 2.0404986962190353e-05,
|
1116 |
+
"loss": 0.6511,
|
1117 |
+
"step": 78500
|
1118 |
+
},
|
1119 |
+
{
|
1120 |
+
"epoch": 1.609354628422425,
|
1121 |
+
"grad_norm": 11.129698753356934,
|
1122 |
+
"learning_rate": 2.0343872229465447e-05,
|
1123 |
+
"loss": 0.6349,
|
1124 |
+
"step": 79000
|
1125 |
+
},
|
1126 |
+
{
|
1127 |
+
"epoch": 1.6195404172099086,
|
1128 |
+
"grad_norm": 16.331953048706055,
|
1129 |
+
"learning_rate": 2.0282757496740548e-05,
|
1130 |
+
"loss": 0.6517,
|
1131 |
+
"step": 79500
|
1132 |
+
},
|
1133 |
+
{
|
1134 |
+
"epoch": 1.6297262059973925,
|
1135 |
+
"grad_norm": 12.816121101379395,
|
1136 |
+
"learning_rate": 2.0221642764015646e-05,
|
1137 |
+
"loss": 0.6409,
|
1138 |
+
"step": 80000
|
1139 |
+
},
|
1140 |
+
{
|
1141 |
+
"epoch": 1.6399119947848761,
|
1142 |
+
"grad_norm": 13.881726264953613,
|
1143 |
+
"learning_rate": 2.0160528031290743e-05,
|
1144 |
+
"loss": 0.6606,
|
1145 |
+
"step": 80500
|
1146 |
+
},
|
1147 |
+
{
|
1148 |
+
"epoch": 1.6500977835723598,
|
1149 |
+
"grad_norm": 14.215683937072754,
|
1150 |
+
"learning_rate": 2.0099413298565844e-05,
|
1151 |
+
"loss": 0.657,
|
1152 |
+
"step": 81000
|
1153 |
+
},
|
1154 |
+
{
|
1155 |
+
"epoch": 1.6602835723598437,
|
1156 |
+
"grad_norm": 19.557342529296875,
|
1157 |
+
"learning_rate": 2.003829856584094e-05,
|
1158 |
+
"loss": 0.6451,
|
1159 |
+
"step": 81500
|
1160 |
+
},
|
1161 |
+
{
|
1162 |
+
"epoch": 1.6704693611473274,
|
1163 |
+
"grad_norm": 17.83816909790039,
|
1164 |
+
"learning_rate": 1.997718383311604e-05,
|
1165 |
+
"loss": 0.6397,
|
1166 |
+
"step": 82000
|
1167 |
+
},
|
1168 |
+
{
|
1169 |
+
"epoch": 1.680655149934811,
|
1170 |
+
"grad_norm": 8.789198875427246,
|
1171 |
+
"learning_rate": 1.9916069100391133e-05,
|
1172 |
+
"loss": 0.6514,
|
1173 |
+
"step": 82500
|
1174 |
+
},
|
1175 |
+
{
|
1176 |
+
"epoch": 1.6908409387222947,
|
1177 |
+
"grad_norm": 11.361543655395508,
|
1178 |
+
"learning_rate": 1.9854954367666234e-05,
|
1179 |
+
"loss": 0.6466,
|
1180 |
+
"step": 83000
|
1181 |
+
},
|
1182 |
+
{
|
1183 |
+
"epoch": 1.7010267275097783,
|
1184 |
+
"grad_norm": 15.630517959594727,
|
1185 |
+
"learning_rate": 1.979383963494133e-05,
|
1186 |
+
"loss": 0.6316,
|
1187 |
+
"step": 83500
|
1188 |
+
},
|
1189 |
+
{
|
1190 |
+
"epoch": 1.711212516297262,
|
1191 |
+
"grad_norm": 10.55432415008545,
|
1192 |
+
"learning_rate": 1.973272490221643e-05,
|
1193 |
+
"loss": 0.6533,
|
1194 |
+
"step": 84000
|
1195 |
+
},
|
1196 |
+
{
|
1197 |
+
"epoch": 1.7213983050847457,
|
1198 |
+
"grad_norm": 8.428609848022461,
|
1199 |
+
"learning_rate": 1.9671610169491524e-05,
|
1200 |
+
"loss": 0.6478,
|
1201 |
+
"step": 84500
|
1202 |
+
},
|
1203 |
+
{
|
1204 |
+
"epoch": 1.7315840938722293,
|
1205 |
+
"grad_norm": 4.304576873779297,
|
1206 |
+
"learning_rate": 1.9610495436766625e-05,
|
1207 |
+
"loss": 0.6449,
|
1208 |
+
"step": 85000
|
1209 |
+
},
|
1210 |
+
{
|
1211 |
+
"epoch": 1.7417698826597132,
|
1212 |
+
"grad_norm": 21.92937469482422,
|
1213 |
+
"learning_rate": 1.9549380704041722e-05,
|
1214 |
+
"loss": 0.6425,
|
1215 |
+
"step": 85500
|
1216 |
+
},
|
1217 |
+
{
|
1218 |
+
"epoch": 1.7519556714471969,
|
1219 |
+
"grad_norm": 19.813392639160156,
|
1220 |
+
"learning_rate": 1.948826597131682e-05,
|
1221 |
+
"loss": 0.6238,
|
1222 |
+
"step": 86000
|
1223 |
+
},
|
1224 |
+
{
|
1225 |
+
"epoch": 1.7621414602346805,
|
1226 |
+
"grad_norm": 6.831646919250488,
|
1227 |
+
"learning_rate": 1.9427151238591917e-05,
|
1228 |
+
"loss": 0.6553,
|
1229 |
+
"step": 86500
|
1230 |
+
},
|
1231 |
+
{
|
1232 |
+
"epoch": 1.7723272490221644,
|
1233 |
+
"grad_norm": 11.058158874511719,
|
1234 |
+
"learning_rate": 1.9366036505867015e-05,
|
1235 |
+
"loss": 0.6528,
|
1236 |
+
"step": 87000
|
1237 |
+
},
|
1238 |
+
{
|
1239 |
+
"epoch": 1.782513037809648,
|
1240 |
+
"grad_norm": 7.01440954208374,
|
1241 |
+
"learning_rate": 1.9304921773142112e-05,
|
1242 |
+
"loss": 0.6506,
|
1243 |
+
"step": 87500
|
1244 |
+
},
|
1245 |
+
{
|
1246 |
+
"epoch": 1.7926988265971318,
|
1247 |
+
"grad_norm": 4.963765621185303,
|
1248 |
+
"learning_rate": 1.924380704041721e-05,
|
1249 |
+
"loss": 0.6412,
|
1250 |
+
"step": 88000
|
1251 |
+
},
|
1252 |
+
{
|
1253 |
+
"epoch": 1.8028846153846154,
|
1254 |
+
"grad_norm": 38.63767623901367,
|
1255 |
+
"learning_rate": 1.9182692307692307e-05,
|
1256 |
+
"loss": 0.6629,
|
1257 |
+
"step": 88500
|
1258 |
+
},
|
1259 |
+
{
|
1260 |
+
"epoch": 1.813070404172099,
|
1261 |
+
"grad_norm": 11.423843383789062,
|
1262 |
+
"learning_rate": 1.9121577574967405e-05,
|
1263 |
+
"loss": 0.673,
|
1264 |
+
"step": 89000
|
1265 |
+
},
|
1266 |
+
{
|
1267 |
+
"epoch": 1.8232561929595827,
|
1268 |
+
"grad_norm": 11.73025894165039,
|
1269 |
+
"learning_rate": 1.9060462842242506e-05,
|
1270 |
+
"loss": 0.625,
|
1271 |
+
"step": 89500
|
1272 |
+
},
|
1273 |
+
{
|
1274 |
+
"epoch": 1.8334419817470664,
|
1275 |
+
"grad_norm": 11.493837356567383,
|
1276 |
+
"learning_rate": 1.89993481095176e-05,
|
1277 |
+
"loss": 0.6658,
|
1278 |
+
"step": 90000
|
1279 |
+
},
|
1280 |
+
{
|
1281 |
+
"epoch": 1.84362777053455,
|
1282 |
+
"grad_norm": 12.196702003479004,
|
1283 |
+
"learning_rate": 1.89382333767927e-05,
|
1284 |
+
"loss": 0.6451,
|
1285 |
+
"step": 90500
|
1286 |
+
},
|
1287 |
+
{
|
1288 |
+
"epoch": 1.8538135593220337,
|
1289 |
+
"grad_norm": 9.690689086914062,
|
1290 |
+
"learning_rate": 1.8877118644067795e-05,
|
1291 |
+
"loss": 0.6282,
|
1292 |
+
"step": 91000
|
1293 |
+
},
|
1294 |
+
{
|
1295 |
+
"epoch": 1.8639993481095176,
|
1296 |
+
"grad_norm": 13.679101943969727,
|
1297 |
+
"learning_rate": 1.8816003911342896e-05,
|
1298 |
+
"loss": 0.6328,
|
1299 |
+
"step": 91500
|
1300 |
+
},
|
1301 |
+
{
|
1302 |
+
"epoch": 1.8741851368970013,
|
1303 |
+
"grad_norm": 14.264899253845215,
|
1304 |
+
"learning_rate": 1.8754889178617994e-05,
|
1305 |
+
"loss": 0.6355,
|
1306 |
+
"step": 92000
|
1307 |
+
},
|
1308 |
+
{
|
1309 |
+
"epoch": 1.8843709256844852,
|
1310 |
+
"grad_norm": 9.039682388305664,
|
1311 |
+
"learning_rate": 1.869377444589309e-05,
|
1312 |
+
"loss": 0.6471,
|
1313 |
+
"step": 92500
|
1314 |
+
},
|
1315 |
+
{
|
1316 |
+
"epoch": 1.8945567144719688,
|
1317 |
+
"grad_norm": 9.861899375915527,
|
1318 |
+
"learning_rate": 1.863265971316819e-05,
|
1319 |
+
"loss": 0.6549,
|
1320 |
+
"step": 93000
|
1321 |
+
},
|
1322 |
+
{
|
1323 |
+
"epoch": 1.9047425032594525,
|
1324 |
+
"grad_norm": 14.948457717895508,
|
1325 |
+
"learning_rate": 1.8571544980443286e-05,
|
1326 |
+
"loss": 0.6436,
|
1327 |
+
"step": 93500
|
1328 |
+
},
|
1329 |
+
{
|
1330 |
+
"epoch": 1.9149282920469362,
|
1331 |
+
"grad_norm": 15.613487243652344,
|
1332 |
+
"learning_rate": 1.8510430247718384e-05,
|
1333 |
+
"loss": 0.6355,
|
1334 |
+
"step": 94000
|
1335 |
+
},
|
1336 |
+
{
|
1337 |
+
"epoch": 1.9251140808344198,
|
1338 |
+
"grad_norm": 14.105792045593262,
|
1339 |
+
"learning_rate": 1.844931551499348e-05,
|
1340 |
+
"loss": 0.6345,
|
1341 |
+
"step": 94500
|
1342 |
+
},
|
1343 |
+
{
|
1344 |
+
"epoch": 1.9352998696219035,
|
1345 |
+
"grad_norm": 7.004807472229004,
|
1346 |
+
"learning_rate": 1.838820078226858e-05,
|
1347 |
+
"loss": 0.6456,
|
1348 |
+
"step": 95000
|
1349 |
+
},
|
1350 |
+
{
|
1351 |
+
"epoch": 1.9454856584093871,
|
1352 |
+
"grad_norm": 10.506879806518555,
|
1353 |
+
"learning_rate": 1.8327086049543677e-05,
|
1354 |
+
"loss": 0.6303,
|
1355 |
+
"step": 95500
|
1356 |
+
},
|
1357 |
+
{
|
1358 |
+
"epoch": 1.9556714471968708,
|
1359 |
+
"grad_norm": 11.595491409301758,
|
1360 |
+
"learning_rate": 1.8265971316818774e-05,
|
1361 |
+
"loss": 0.6509,
|
1362 |
+
"step": 96000
|
1363 |
+
},
|
1364 |
+
{
|
1365 |
+
"epoch": 1.9658572359843545,
|
1366 |
+
"grad_norm": 11.29542064666748,
|
1367 |
+
"learning_rate": 1.820485658409387e-05,
|
1368 |
+
"loss": 0.6273,
|
1369 |
+
"step": 96500
|
1370 |
+
},
|
1371 |
+
{
|
1372 |
+
"epoch": 1.9760430247718384,
|
1373 |
+
"grad_norm": 4.132607460021973,
|
1374 |
+
"learning_rate": 1.814374185136897e-05,
|
1375 |
+
"loss": 0.643,
|
1376 |
+
"step": 97000
|
1377 |
+
},
|
1378 |
+
{
|
1379 |
+
"epoch": 1.986228813559322,
|
1380 |
+
"grad_norm": 9.745738983154297,
|
1381 |
+
"learning_rate": 1.8082627118644067e-05,
|
1382 |
+
"loss": 0.6656,
|
1383 |
+
"step": 97500
|
1384 |
+
},
|
1385 |
+
{
|
1386 |
+
"epoch": 1.996414602346806,
|
1387 |
+
"grad_norm": 4.289452075958252,
|
1388 |
+
"learning_rate": 1.8021512385919164e-05,
|
1389 |
+
"loss": 0.6367,
|
1390 |
+
"step": 98000
|
1391 |
+
},
|
1392 |
+
{
|
1393 |
+
"epoch": 2.0,
|
1394 |
+
"eval_accuracy": 0.7218826413154602,
|
1395 |
+
"eval_loss": 0.6872764825820923,
|
1396 |
+
"eval_runtime": 7.6175,
|
1397 |
+
"eval_samples_per_second": 644.307,
|
1398 |
+
"eval_steps_per_second": 80.604,
|
1399 |
+
"step": 98176
|
1400 |
+
},
|
1401 |
+
{
|
1402 |
+
"epoch": 2.0066003911342896,
|
1403 |
+
"grad_norm": 19.260536193847656,
|
1404 |
+
"learning_rate": 1.7960397653194265e-05,
|
1405 |
+
"loss": 0.5568,
|
1406 |
+
"step": 98500
|
1407 |
+
},
|
1408 |
+
{
|
1409 |
+
"epoch": 2.0167861799217732,
|
1410 |
+
"grad_norm": 8.347018241882324,
|
1411 |
+
"learning_rate": 1.7899282920469363e-05,
|
1412 |
+
"loss": 0.5344,
|
1413 |
+
"step": 99000
|
1414 |
+
},
|
1415 |
+
{
|
1416 |
+
"epoch": 2.026971968709257,
|
1417 |
+
"grad_norm": 12.238682746887207,
|
1418 |
+
"learning_rate": 1.783816818774446e-05,
|
1419 |
+
"loss": 0.5685,
|
1420 |
+
"step": 99500
|
1421 |
+
},
|
1422 |
+
{
|
1423 |
+
"epoch": 2.0371577574967406,
|
1424 |
+
"grad_norm": 18.42786979675293,
|
1425 |
+
"learning_rate": 1.7777053455019558e-05,
|
1426 |
+
"loss": 0.5539,
|
1427 |
+
"step": 100000
|
1428 |
+
},
|
1429 |
+
{
|
1430 |
+
"epoch": 2.047343546284224,
|
1431 |
+
"grad_norm": 25.627717971801758,
|
1432 |
+
"learning_rate": 1.7715938722294655e-05,
|
1433 |
+
"loss": 0.5393,
|
1434 |
+
"step": 100500
|
1435 |
+
},
|
1436 |
+
{
|
1437 |
+
"epoch": 2.057529335071708,
|
1438 |
+
"grad_norm": 15.942806243896484,
|
1439 |
+
"learning_rate": 1.7654823989569753e-05,
|
1440 |
+
"loss": 0.5191,
|
1441 |
+
"step": 101000
|
1442 |
+
},
|
1443 |
+
{
|
1444 |
+
"epoch": 2.0677151238591915,
|
1445 |
+
"grad_norm": 7.591663360595703,
|
1446 |
+
"learning_rate": 1.759370925684485e-05,
|
1447 |
+
"loss": 0.5428,
|
1448 |
+
"step": 101500
|
1449 |
+
},
|
1450 |
+
{
|
1451 |
+
"epoch": 2.077900912646675,
|
1452 |
+
"grad_norm": 7.183356285095215,
|
1453 |
+
"learning_rate": 1.7532594524119948e-05,
|
1454 |
+
"loss": 0.5475,
|
1455 |
+
"step": 102000
|
1456 |
+
},
|
1457 |
+
{
|
1458 |
+
"epoch": 2.088086701434159,
|
1459 |
+
"grad_norm": 19.916288375854492,
|
1460 |
+
"learning_rate": 1.7471479791395046e-05,
|
1461 |
+
"loss": 0.5586,
|
1462 |
+
"step": 102500
|
1463 |
+
},
|
1464 |
+
{
|
1465 |
+
"epoch": 2.098272490221643,
|
1466 |
+
"grad_norm": 9.07480239868164,
|
1467 |
+
"learning_rate": 1.7410365058670143e-05,
|
1468 |
+
"loss": 0.5564,
|
1469 |
+
"step": 103000
|
1470 |
+
},
|
1471 |
+
{
|
1472 |
+
"epoch": 2.1084582790091266,
|
1473 |
+
"grad_norm": 7.647058010101318,
|
1474 |
+
"learning_rate": 1.734925032594524e-05,
|
1475 |
+
"loss": 0.543,
|
1476 |
+
"step": 103500
|
1477 |
+
},
|
1478 |
+
{
|
1479 |
+
"epoch": 2.1186440677966103,
|
1480 |
+
"grad_norm": 8.871644973754883,
|
1481 |
+
"learning_rate": 1.728813559322034e-05,
|
1482 |
+
"loss": 0.5623,
|
1483 |
+
"step": 104000
|
1484 |
+
},
|
1485 |
+
{
|
1486 |
+
"epoch": 2.128829856584094,
|
1487 |
+
"grad_norm": 9.516389846801758,
|
1488 |
+
"learning_rate": 1.7227020860495436e-05,
|
1489 |
+
"loss": 0.5464,
|
1490 |
+
"step": 104500
|
1491 |
+
},
|
1492 |
+
{
|
1493 |
+
"epoch": 2.1390156453715776,
|
1494 |
+
"grad_norm": 13.605746269226074,
|
1495 |
+
"learning_rate": 1.7165906127770537e-05,
|
1496 |
+
"loss": 0.5425,
|
1497 |
+
"step": 105000
|
1498 |
+
},
|
1499 |
+
{
|
1500 |
+
"epoch": 2.1492014341590613,
|
1501 |
+
"grad_norm": 14.016572952270508,
|
1502 |
+
"learning_rate": 1.710479139504563e-05,
|
1503 |
+
"loss": 0.5496,
|
1504 |
+
"step": 105500
|
1505 |
+
},
|
1506 |
+
{
|
1507 |
+
"epoch": 2.159387222946545,
|
1508 |
+
"grad_norm": 14.719120979309082,
|
1509 |
+
"learning_rate": 1.7043676662320732e-05,
|
1510 |
+
"loss": 0.5591,
|
1511 |
+
"step": 106000
|
1512 |
+
},
|
1513 |
+
{
|
1514 |
+
"epoch": 2.1695730117340286,
|
1515 |
+
"grad_norm": 14.922130584716797,
|
1516 |
+
"learning_rate": 1.6982561929595826e-05,
|
1517 |
+
"loss": 0.5462,
|
1518 |
+
"step": 106500
|
1519 |
+
},
|
1520 |
+
{
|
1521 |
+
"epoch": 2.1797588005215123,
|
1522 |
+
"grad_norm": 10.74787425994873,
|
1523 |
+
"learning_rate": 1.6921447196870927e-05,
|
1524 |
+
"loss": 0.5699,
|
1525 |
+
"step": 107000
|
1526 |
+
},
|
1527 |
+
{
|
1528 |
+
"epoch": 2.189944589308996,
|
1529 |
+
"grad_norm": 7.978420257568359,
|
1530 |
+
"learning_rate": 1.6860332464146024e-05,
|
1531 |
+
"loss": 0.5449,
|
1532 |
+
"step": 107500
|
1533 |
+
},
|
1534 |
+
{
|
1535 |
+
"epoch": 2.2001303780964796,
|
1536 |
+
"grad_norm": 15.361347198486328,
|
1537 |
+
"learning_rate": 1.6799217731421122e-05,
|
1538 |
+
"loss": 0.5443,
|
1539 |
+
"step": 108000
|
1540 |
+
},
|
1541 |
+
{
|
1542 |
+
"epoch": 2.2103161668839633,
|
1543 |
+
"grad_norm": 6.552661418914795,
|
1544 |
+
"learning_rate": 1.673810299869622e-05,
|
1545 |
+
"loss": 0.5663,
|
1546 |
+
"step": 108500
|
1547 |
+
},
|
1548 |
+
{
|
1549 |
+
"epoch": 2.2205019556714474,
|
1550 |
+
"grad_norm": 6.750521659851074,
|
1551 |
+
"learning_rate": 1.6676988265971317e-05,
|
1552 |
+
"loss": 0.5488,
|
1553 |
+
"step": 109000
|
1554 |
+
},
|
1555 |
+
{
|
1556 |
+
"epoch": 2.230687744458931,
|
1557 |
+
"grad_norm": 11.149799346923828,
|
1558 |
+
"learning_rate": 1.6615873533246415e-05,
|
1559 |
+
"loss": 0.5412,
|
1560 |
+
"step": 109500
|
1561 |
+
},
|
1562 |
+
{
|
1563 |
+
"epoch": 2.2408735332464147,
|
1564 |
+
"grad_norm": 19.362638473510742,
|
1565 |
+
"learning_rate": 1.6554758800521512e-05,
|
1566 |
+
"loss": 0.5505,
|
1567 |
+
"step": 110000
|
1568 |
+
},
|
1569 |
+
{
|
1570 |
+
"epoch": 2.2510593220338984,
|
1571 |
+
"grad_norm": 11.263615608215332,
|
1572 |
+
"learning_rate": 1.6493644067796613e-05,
|
1573 |
+
"loss": 0.5591,
|
1574 |
+
"step": 110500
|
1575 |
+
},
|
1576 |
+
{
|
1577 |
+
"epoch": 2.261245110821382,
|
1578 |
+
"grad_norm": 8.90456771850586,
|
1579 |
+
"learning_rate": 1.6432529335071707e-05,
|
1580 |
+
"loss": 0.5491,
|
1581 |
+
"step": 111000
|
1582 |
+
},
|
1583 |
+
{
|
1584 |
+
"epoch": 2.2714308996088657,
|
1585 |
+
"grad_norm": 11.962569236755371,
|
1586 |
+
"learning_rate": 1.6371414602346808e-05,
|
1587 |
+
"loss": 0.556,
|
1588 |
+
"step": 111500
|
1589 |
+
},
|
1590 |
+
{
|
1591 |
+
"epoch": 2.2816166883963493,
|
1592 |
+
"grad_norm": 8.397544860839844,
|
1593 |
+
"learning_rate": 1.6310299869621902e-05,
|
1594 |
+
"loss": 0.5561,
|
1595 |
+
"step": 112000
|
1596 |
+
},
|
1597 |
+
{
|
1598 |
+
"epoch": 2.291802477183833,
|
1599 |
+
"grad_norm": 16.376155853271484,
|
1600 |
+
"learning_rate": 1.6249185136897003e-05,
|
1601 |
+
"loss": 0.5435,
|
1602 |
+
"step": 112500
|
1603 |
+
},
|
1604 |
+
{
|
1605 |
+
"epoch": 2.3019882659713167,
|
1606 |
+
"grad_norm": 9.14609432220459,
|
1607 |
+
"learning_rate": 1.6188070404172097e-05,
|
1608 |
+
"loss": 0.5457,
|
1609 |
+
"step": 113000
|
1610 |
+
},
|
1611 |
+
{
|
1612 |
+
"epoch": 2.3121740547588003,
|
1613 |
+
"grad_norm": 26.53936767578125,
|
1614 |
+
"learning_rate": 1.61269556714472e-05,
|
1615 |
+
"loss": 0.5295,
|
1616 |
+
"step": 113500
|
1617 |
+
},
|
1618 |
+
{
|
1619 |
+
"epoch": 2.322359843546284,
|
1620 |
+
"grad_norm": 17.222721099853516,
|
1621 |
+
"learning_rate": 1.6065840938722293e-05,
|
1622 |
+
"loss": 0.5177,
|
1623 |
+
"step": 114000
|
1624 |
+
},
|
1625 |
+
{
|
1626 |
+
"epoch": 2.332545632333768,
|
1627 |
+
"grad_norm": 9.585100173950195,
|
1628 |
+
"learning_rate": 1.6004726205997394e-05,
|
1629 |
+
"loss": 0.5419,
|
1630 |
+
"step": 114500
|
1631 |
+
},
|
1632 |
+
{
|
1633 |
+
"epoch": 2.3427314211212518,
|
1634 |
+
"grad_norm": 9.392489433288574,
|
1635 |
+
"learning_rate": 1.5943611473272488e-05,
|
1636 |
+
"loss": 0.5296,
|
1637 |
+
"step": 115000
|
1638 |
+
},
|
1639 |
+
{
|
1640 |
+
"epoch": 2.3529172099087354,
|
1641 |
+
"grad_norm": 15.146162986755371,
|
1642 |
+
"learning_rate": 1.588249674054759e-05,
|
1643 |
+
"loss": 0.5518,
|
1644 |
+
"step": 115500
|
1645 |
+
},
|
1646 |
+
{
|
1647 |
+
"epoch": 2.363102998696219,
|
1648 |
+
"grad_norm": 15.631641387939453,
|
1649 |
+
"learning_rate": 1.582138200782269e-05,
|
1650 |
+
"loss": 0.5584,
|
1651 |
+
"step": 116000
|
1652 |
+
},
|
1653 |
+
{
|
1654 |
+
"epoch": 2.3732887874837028,
|
1655 |
+
"grad_norm": 17.875492095947266,
|
1656 |
+
"learning_rate": 1.5760267275097784e-05,
|
1657 |
+
"loss": 0.5601,
|
1658 |
+
"step": 116500
|
1659 |
+
},
|
1660 |
+
{
|
1661 |
+
"epoch": 2.3834745762711864,
|
1662 |
+
"grad_norm": 7.7471184730529785,
|
1663 |
+
"learning_rate": 1.5699152542372885e-05,
|
1664 |
+
"loss": 0.549,
|
1665 |
+
"step": 117000
|
1666 |
+
},
|
1667 |
+
{
|
1668 |
+
"epoch": 2.39366036505867,
|
1669 |
+
"grad_norm": 14.712841987609863,
|
1670 |
+
"learning_rate": 1.563803780964798e-05,
|
1671 |
+
"loss": 0.5293,
|
1672 |
+
"step": 117500
|
1673 |
+
},
|
1674 |
+
{
|
1675 |
+
"epoch": 2.4038461538461537,
|
1676 |
+
"grad_norm": 9.229011535644531,
|
1677 |
+
"learning_rate": 1.557692307692308e-05,
|
1678 |
+
"loss": 0.5483,
|
1679 |
+
"step": 118000
|
1680 |
+
},
|
1681 |
+
{
|
1682 |
+
"epoch": 2.4140319426336374,
|
1683 |
+
"grad_norm": 11.47548770904541,
|
1684 |
+
"learning_rate": 1.5515808344198174e-05,
|
1685 |
+
"loss": 0.5529,
|
1686 |
+
"step": 118500
|
1687 |
+
},
|
1688 |
+
{
|
1689 |
+
"epoch": 2.424217731421121,
|
1690 |
+
"grad_norm": 29.980873107910156,
|
1691 |
+
"learning_rate": 1.5454693611473275e-05,
|
1692 |
+
"loss": 0.5489,
|
1693 |
+
"step": 119000
|
1694 |
+
},
|
1695 |
+
{
|
1696 |
+
"epoch": 2.4344035202086047,
|
1697 |
+
"grad_norm": 10.478185653686523,
|
1698 |
+
"learning_rate": 1.539357887874837e-05,
|
1699 |
+
"loss": 0.5595,
|
1700 |
+
"step": 119500
|
1701 |
+
},
|
1702 |
+
{
|
1703 |
+
"epoch": 2.444589308996089,
|
1704 |
+
"grad_norm": 12.978096008300781,
|
1705 |
+
"learning_rate": 1.533246414602347e-05,
|
1706 |
+
"loss": 0.5612,
|
1707 |
+
"step": 120000
|
1708 |
+
},
|
1709 |
+
{
|
1710 |
+
"epoch": 2.4547750977835725,
|
1711 |
+
"grad_norm": 17.834806442260742,
|
1712 |
+
"learning_rate": 1.5271349413298564e-05,
|
1713 |
+
"loss": 0.5412,
|
1714 |
+
"step": 120500
|
1715 |
+
},
|
1716 |
+
{
|
1717 |
+
"epoch": 2.464960886571056,
|
1718 |
+
"grad_norm": 25.001754760742188,
|
1719 |
+
"learning_rate": 1.5210234680573665e-05,
|
1720 |
+
"loss": 0.5673,
|
1721 |
+
"step": 121000
|
1722 |
+
},
|
1723 |
+
{
|
1724 |
+
"epoch": 2.47514667535854,
|
1725 |
+
"grad_norm": 27.953767776489258,
|
1726 |
+
"learning_rate": 1.5149119947848761e-05,
|
1727 |
+
"loss": 0.5497,
|
1728 |
+
"step": 121500
|
1729 |
+
},
|
1730 |
+
{
|
1731 |
+
"epoch": 2.4853324641460235,
|
1732 |
+
"grad_norm": 9.370855331420898,
|
1733 |
+
"learning_rate": 1.508800521512386e-05,
|
1734 |
+
"loss": 0.5344,
|
1735 |
+
"step": 122000
|
1736 |
+
},
|
1737 |
+
{
|
1738 |
+
"epoch": 2.495518252933507,
|
1739 |
+
"grad_norm": 18.109821319580078,
|
1740 |
+
"learning_rate": 1.5026890482398956e-05,
|
1741 |
+
"loss": 0.5572,
|
1742 |
+
"step": 122500
|
1743 |
+
},
|
1744 |
+
{
|
1745 |
+
"epoch": 2.505704041720991,
|
1746 |
+
"grad_norm": 20.435270309448242,
|
1747 |
+
"learning_rate": 1.4965775749674055e-05,
|
1748 |
+
"loss": 0.5631,
|
1749 |
+
"step": 123000
|
1750 |
+
},
|
1751 |
+
{
|
1752 |
+
"epoch": 2.5158898305084745,
|
1753 |
+
"grad_norm": 10.656272888183594,
|
1754 |
+
"learning_rate": 1.4904661016949153e-05,
|
1755 |
+
"loss": 0.5344,
|
1756 |
+
"step": 123500
|
1757 |
+
},
|
1758 |
+
{
|
1759 |
+
"epoch": 2.526075619295958,
|
1760 |
+
"grad_norm": 18.796552658081055,
|
1761 |
+
"learning_rate": 1.484354628422425e-05,
|
1762 |
+
"loss": 0.5464,
|
1763 |
+
"step": 124000
|
1764 |
+
},
|
1765 |
+
{
|
1766 |
+
"epoch": 2.5362614080834422,
|
1767 |
+
"grad_norm": 9.630306243896484,
|
1768 |
+
"learning_rate": 1.4782431551499348e-05,
|
1769 |
+
"loss": 0.547,
|
1770 |
+
"step": 124500
|
1771 |
+
},
|
1772 |
+
{
|
1773 |
+
"epoch": 2.5464471968709255,
|
1774 |
+
"grad_norm": 6.3599467277526855,
|
1775 |
+
"learning_rate": 1.4721316818774445e-05,
|
1776 |
+
"loss": 0.5577,
|
1777 |
+
"step": 125000
|
1778 |
+
},
|
1779 |
+
{
|
1780 |
+
"epoch": 2.5566329856584096,
|
1781 |
+
"grad_norm": 8.967144966125488,
|
1782 |
+
"learning_rate": 1.4660202086049545e-05,
|
1783 |
+
"loss": 0.5524,
|
1784 |
+
"step": 125500
|
1785 |
+
},
|
1786 |
+
{
|
1787 |
+
"epoch": 2.5668187744458932,
|
1788 |
+
"grad_norm": 16.690263748168945,
|
1789 |
+
"learning_rate": 1.4599087353324642e-05,
|
1790 |
+
"loss": 0.5546,
|
1791 |
+
"step": 126000
|
1792 |
+
},
|
1793 |
+
{
|
1794 |
+
"epoch": 2.577004563233377,
|
1795 |
+
"grad_norm": 17.713790893554688,
|
1796 |
+
"learning_rate": 1.453797262059974e-05,
|
1797 |
+
"loss": 0.5686,
|
1798 |
+
"step": 126500
|
1799 |
+
},
|
1800 |
+
{
|
1801 |
+
"epoch": 2.5871903520208606,
|
1802 |
+
"grad_norm": 7.970760822296143,
|
1803 |
+
"learning_rate": 1.4476857887874837e-05,
|
1804 |
+
"loss": 0.5273,
|
1805 |
+
"step": 127000
|
1806 |
+
},
|
1807 |
+
{
|
1808 |
+
"epoch": 2.5973761408083442,
|
1809 |
+
"grad_norm": 8.869507789611816,
|
1810 |
+
"learning_rate": 1.4415743155149935e-05,
|
1811 |
+
"loss": 0.5472,
|
1812 |
+
"step": 127500
|
1813 |
+
},
|
1814 |
+
{
|
1815 |
+
"epoch": 2.607561929595828,
|
1816 |
+
"grad_norm": 17.247589111328125,
|
1817 |
+
"learning_rate": 1.4354628422425032e-05,
|
1818 |
+
"loss": 0.5444,
|
1819 |
+
"step": 128000
|
1820 |
+
},
|
1821 |
+
{
|
1822 |
+
"epoch": 2.6177477183833116,
|
1823 |
+
"grad_norm": 16.415138244628906,
|
1824 |
+
"learning_rate": 1.429351368970013e-05,
|
1825 |
+
"loss": 0.5445,
|
1826 |
+
"step": 128500
|
1827 |
+
},
|
1828 |
+
{
|
1829 |
+
"epoch": 2.627933507170795,
|
1830 |
+
"grad_norm": 10.408729553222656,
|
1831 |
+
"learning_rate": 1.4232398956975227e-05,
|
1832 |
+
"loss": 0.5667,
|
1833 |
+
"step": 129000
|
1834 |
+
},
|
1835 |
+
{
|
1836 |
+
"epoch": 2.638119295958279,
|
1837 |
+
"grad_norm": 53.77888488769531,
|
1838 |
+
"learning_rate": 1.4171284224250327e-05,
|
1839 |
+
"loss": 0.547,
|
1840 |
+
"step": 129500
|
1841 |
+
},
|
1842 |
+
{
|
1843 |
+
"epoch": 2.648305084745763,
|
1844 |
+
"grad_norm": 12.244620323181152,
|
1845 |
+
"learning_rate": 1.4110169491525424e-05,
|
1846 |
+
"loss": 0.5428,
|
1847 |
+
"step": 130000
|
1848 |
+
},
|
1849 |
+
{
|
1850 |
+
"epoch": 2.658490873533246,
|
1851 |
+
"grad_norm": 11.80124568939209,
|
1852 |
+
"learning_rate": 1.4049054758800522e-05,
|
1853 |
+
"loss": 0.5565,
|
1854 |
+
"step": 130500
|
1855 |
+
},
|
1856 |
+
{
|
1857 |
+
"epoch": 2.6686766623207303,
|
1858 |
+
"grad_norm": 13.250449180603027,
|
1859 |
+
"learning_rate": 1.398794002607562e-05,
|
1860 |
+
"loss": 0.5423,
|
1861 |
+
"step": 131000
|
1862 |
+
},
|
1863 |
+
{
|
1864 |
+
"epoch": 2.678862451108214,
|
1865 |
+
"grad_norm": 15.81285572052002,
|
1866 |
+
"learning_rate": 1.3926825293350719e-05,
|
1867 |
+
"loss": 0.5621,
|
1868 |
+
"step": 131500
|
1869 |
+
},
|
1870 |
+
{
|
1871 |
+
"epoch": 2.6890482398956976,
|
1872 |
+
"grad_norm": 22.584035873413086,
|
1873 |
+
"learning_rate": 1.3865710560625816e-05,
|
1874 |
+
"loss": 0.5521,
|
1875 |
+
"step": 132000
|
1876 |
+
},
|
1877 |
+
{
|
1878 |
+
"epoch": 2.6992340286831813,
|
1879 |
+
"grad_norm": 23.865680694580078,
|
1880 |
+
"learning_rate": 1.3804595827900914e-05,
|
1881 |
+
"loss": 0.551,
|
1882 |
+
"step": 132500
|
1883 |
+
},
|
1884 |
+
{
|
1885 |
+
"epoch": 2.709419817470665,
|
1886 |
+
"grad_norm": 18.28876304626465,
|
1887 |
+
"learning_rate": 1.3743481095176011e-05,
|
1888 |
+
"loss": 0.5451,
|
1889 |
+
"step": 133000
|
1890 |
+
},
|
1891 |
+
{
|
1892 |
+
"epoch": 2.7196056062581486,
|
1893 |
+
"grad_norm": 16.23969078063965,
|
1894 |
+
"learning_rate": 1.3682366362451109e-05,
|
1895 |
+
"loss": 0.5392,
|
1896 |
+
"step": 133500
|
1897 |
+
},
|
1898 |
+
{
|
1899 |
+
"epoch": 2.7297913950456323,
|
1900 |
+
"grad_norm": 14.673959732055664,
|
1901 |
+
"learning_rate": 1.3621251629726206e-05,
|
1902 |
+
"loss": 0.5451,
|
1903 |
+
"step": 134000
|
1904 |
+
},
|
1905 |
+
{
|
1906 |
+
"epoch": 2.739977183833116,
|
1907 |
+
"grad_norm": 8.019514083862305,
|
1908 |
+
"learning_rate": 1.3560136897001304e-05,
|
1909 |
+
"loss": 0.5568,
|
1910 |
+
"step": 134500
|
1911 |
+
},
|
1912 |
+
{
|
1913 |
+
"epoch": 2.7501629726205996,
|
1914 |
+
"grad_norm": 13.495898246765137,
|
1915 |
+
"learning_rate": 1.3499022164276401e-05,
|
1916 |
+
"loss": 0.5734,
|
1917 |
+
"step": 135000
|
1918 |
+
},
|
1919 |
+
{
|
1920 |
+
"epoch": 2.7603487614080837,
|
1921 |
+
"grad_norm": 7.548976421356201,
|
1922 |
+
"learning_rate": 1.3437907431551499e-05,
|
1923 |
+
"loss": 0.557,
|
1924 |
+
"step": 135500
|
1925 |
+
},
|
1926 |
+
{
|
1927 |
+
"epoch": 2.770534550195567,
|
1928 |
+
"grad_norm": 20.049760818481445,
|
1929 |
+
"learning_rate": 1.3376792698826597e-05,
|
1930 |
+
"loss": 0.5658,
|
1931 |
+
"step": 136000
|
1932 |
+
},
|
1933 |
+
{
|
1934 |
+
"epoch": 2.780720338983051,
|
1935 |
+
"grad_norm": 9.346122741699219,
|
1936 |
+
"learning_rate": 1.3315677966101694e-05,
|
1937 |
+
"loss": 0.535,
|
1938 |
+
"step": 136500
|
1939 |
+
},
|
1940 |
+
{
|
1941 |
+
"epoch": 2.7909061277705347,
|
1942 |
+
"grad_norm": 15.080660820007324,
|
1943 |
+
"learning_rate": 1.3254563233376792e-05,
|
1944 |
+
"loss": 0.5538,
|
1945 |
+
"step": 137000
|
1946 |
+
},
|
1947 |
+
{
|
1948 |
+
"epoch": 2.8010919165580184,
|
1949 |
+
"grad_norm": 11.485374450683594,
|
1950 |
+
"learning_rate": 1.3193448500651891e-05,
|
1951 |
+
"loss": 0.5419,
|
1952 |
+
"step": 137500
|
1953 |
+
},
|
1954 |
+
{
|
1955 |
+
"epoch": 2.811277705345502,
|
1956 |
+
"grad_norm": 12.089446067810059,
|
1957 |
+
"learning_rate": 1.3132333767926988e-05,
|
1958 |
+
"loss": 0.5651,
|
1959 |
+
"step": 138000
|
1960 |
+
},
|
1961 |
+
{
|
1962 |
+
"epoch": 2.8214634941329857,
|
1963 |
+
"grad_norm": 5.466490268707275,
|
1964 |
+
"learning_rate": 1.3071219035202088e-05,
|
1965 |
+
"loss": 0.5217,
|
1966 |
+
"step": 138500
|
1967 |
+
},
|
1968 |
+
{
|
1969 |
+
"epoch": 2.8316492829204694,
|
1970 |
+
"grad_norm": 12.89148235321045,
|
1971 |
+
"learning_rate": 1.3010104302477185e-05,
|
1972 |
+
"loss": 0.5477,
|
1973 |
+
"step": 139000
|
1974 |
+
},
|
1975 |
+
{
|
1976 |
+
"epoch": 2.841835071707953,
|
1977 |
+
"grad_norm": 5.610709190368652,
|
1978 |
+
"learning_rate": 1.2948989569752283e-05,
|
1979 |
+
"loss": 0.5377,
|
1980 |
+
"step": 139500
|
1981 |
+
},
|
1982 |
+
{
|
1983 |
+
"epoch": 2.8520208604954367,
|
1984 |
+
"grad_norm": 26.2186222076416,
|
1985 |
+
"learning_rate": 1.288787483702738e-05,
|
1986 |
+
"loss": 0.5281,
|
1987 |
+
"step": 140000
|
1988 |
+
},
|
1989 |
+
{
|
1990 |
+
"epoch": 2.8622066492829203,
|
1991 |
+
"grad_norm": 2.1066462993621826,
|
1992 |
+
"learning_rate": 1.2826760104302478e-05,
|
1993 |
+
"loss": 0.5387,
|
1994 |
+
"step": 140500
|
1995 |
+
},
|
1996 |
+
{
|
1997 |
+
"epoch": 2.872392438070404,
|
1998 |
+
"grad_norm": 13.6646728515625,
|
1999 |
+
"learning_rate": 1.2765645371577575e-05,
|
2000 |
+
"loss": 0.5416,
|
2001 |
+
"step": 141000
|
2002 |
+
},
|
2003 |
+
{
|
2004 |
+
"epoch": 2.8825782268578877,
|
2005 |
+
"grad_norm": 14.357284545898438,
|
2006 |
+
"learning_rate": 1.2704530638852673e-05,
|
2007 |
+
"loss": 0.5504,
|
2008 |
+
"step": 141500
|
2009 |
+
},
|
2010 |
+
{
|
2011 |
+
"epoch": 2.8927640156453718,
|
2012 |
+
"grad_norm": 14.08674144744873,
|
2013 |
+
"learning_rate": 1.264341590612777e-05,
|
2014 |
+
"loss": 0.5691,
|
2015 |
+
"step": 142000
|
2016 |
+
},
|
2017 |
+
{
|
2018 |
+
"epoch": 2.9029498044328554,
|
2019 |
+
"grad_norm": 20.460561752319336,
|
2020 |
+
"learning_rate": 1.2582301173402868e-05,
|
2021 |
+
"loss": 0.5624,
|
2022 |
+
"step": 142500
|
2023 |
+
},
|
2024 |
+
{
|
2025 |
+
"epoch": 2.913135593220339,
|
2026 |
+
"grad_norm": 12.731225967407227,
|
2027 |
+
"learning_rate": 1.2521186440677966e-05,
|
2028 |
+
"loss": 0.5439,
|
2029 |
+
"step": 143000
|
2030 |
+
},
|
2031 |
+
{
|
2032 |
+
"epoch": 2.9233213820078228,
|
2033 |
+
"grad_norm": 3.781522274017334,
|
2034 |
+
"learning_rate": 1.2460071707953065e-05,
|
2035 |
+
"loss": 0.5348,
|
2036 |
+
"step": 143500
|
2037 |
+
},
|
2038 |
+
{
|
2039 |
+
"epoch": 2.9335071707953064,
|
2040 |
+
"grad_norm": 14.624204635620117,
|
2041 |
+
"learning_rate": 1.2398956975228162e-05,
|
2042 |
+
"loss": 0.5589,
|
2043 |
+
"step": 144000
|
2044 |
+
},
|
2045 |
+
{
|
2046 |
+
"epoch": 2.94369295958279,
|
2047 |
+
"grad_norm": 18.551095962524414,
|
2048 |
+
"learning_rate": 1.233784224250326e-05,
|
2049 |
+
"loss": 0.532,
|
2050 |
+
"step": 144500
|
2051 |
+
},
|
2052 |
+
{
|
2053 |
+
"epoch": 2.9538787483702738,
|
2054 |
+
"grad_norm": 17.094831466674805,
|
2055 |
+
"learning_rate": 1.2276727509778357e-05,
|
2056 |
+
"loss": 0.558,
|
2057 |
+
"step": 145000
|
2058 |
+
},
|
2059 |
+
{
|
2060 |
+
"epoch": 2.9640645371577574,
|
2061 |
+
"grad_norm": 12.306907653808594,
|
2062 |
+
"learning_rate": 1.2215612777053455e-05,
|
2063 |
+
"loss": 0.542,
|
2064 |
+
"step": 145500
|
2065 |
+
},
|
2066 |
+
{
|
2067 |
+
"epoch": 2.974250325945241,
|
2068 |
+
"grad_norm": 12.134025573730469,
|
2069 |
+
"learning_rate": 1.2154498044328553e-05,
|
2070 |
+
"loss": 0.5429,
|
2071 |
+
"step": 146000
|
2072 |
+
},
|
2073 |
+
{
|
2074 |
+
"epoch": 2.9844361147327247,
|
2075 |
+
"grad_norm": 22.295795440673828,
|
2076 |
+
"learning_rate": 1.209338331160365e-05,
|
2077 |
+
"loss": 0.5632,
|
2078 |
+
"step": 146500
|
2079 |
+
},
|
2080 |
+
{
|
2081 |
+
"epoch": 2.9946219035202084,
|
2082 |
+
"grad_norm": 17.749858856201172,
|
2083 |
+
"learning_rate": 1.2032268578878748e-05,
|
2084 |
+
"loss": 0.5402,
|
2085 |
+
"step": 147000
|
2086 |
+
},
|
2087 |
+
{
|
2088 |
+
"epoch": 3.0,
|
2089 |
+
"eval_accuracy": 0.7337000966072083,
|
2090 |
+
"eval_loss": 0.7322831153869629,
|
2091 |
+
"eval_runtime": 8.3496,
|
2092 |
+
"eval_samples_per_second": 587.813,
|
2093 |
+
"eval_steps_per_second": 73.536,
|
2094 |
+
"step": 147264
|
2095 |
+
},
|
2096 |
+
{
|
2097 |
+
"epoch": 3.0048076923076925,
|
2098 |
+
"grad_norm": 22.746931076049805,
|
2099 |
+
"learning_rate": 1.1971153846153847e-05,
|
2100 |
+
"loss": 0.5024,
|
2101 |
+
"step": 147500
|
2102 |
+
},
|
2103 |
+
{
|
2104 |
+
"epoch": 3.014993481095176,
|
2105 |
+
"grad_norm": 19.812463760375977,
|
2106 |
+
"learning_rate": 1.1910039113428944e-05,
|
2107 |
+
"loss": 0.455,
|
2108 |
+
"step": 148000
|
2109 |
+
},
|
2110 |
+
{
|
2111 |
+
"epoch": 3.02517926988266,
|
2112 |
+
"grad_norm": 8.608039855957031,
|
2113 |
+
"learning_rate": 1.1848924380704042e-05,
|
2114 |
+
"loss": 0.4529,
|
2115 |
+
"step": 148500
|
2116 |
+
},
|
2117 |
+
{
|
2118 |
+
"epoch": 3.0353650586701435,
|
2119 |
+
"grad_norm": 21.67936897277832,
|
2120 |
+
"learning_rate": 1.178780964797914e-05,
|
2121 |
+
"loss": 0.467,
|
2122 |
+
"step": 149000
|
2123 |
+
},
|
2124 |
+
{
|
2125 |
+
"epoch": 3.045550847457627,
|
2126 |
+
"grad_norm": 23.777921676635742,
|
2127 |
+
"learning_rate": 1.1726694915254239e-05,
|
2128 |
+
"loss": 0.4399,
|
2129 |
+
"step": 149500
|
2130 |
+
},
|
2131 |
+
{
|
2132 |
+
"epoch": 3.055736636245111,
|
2133 |
+
"grad_norm": 5.8297600746154785,
|
2134 |
+
"learning_rate": 1.1665580182529336e-05,
|
2135 |
+
"loss": 0.4605,
|
2136 |
+
"step": 150000
|
2137 |
+
},
|
2138 |
+
{
|
2139 |
+
"epoch": 3.0659224250325945,
|
2140 |
+
"grad_norm": 1.6593589782714844,
|
2141 |
+
"learning_rate": 1.1604465449804434e-05,
|
2142 |
+
"loss": 0.4668,
|
2143 |
+
"step": 150500
|
2144 |
+
},
|
2145 |
+
{
|
2146 |
+
"epoch": 3.076108213820078,
|
2147 |
+
"grad_norm": 41.140750885009766,
|
2148 |
+
"learning_rate": 1.1543350717079531e-05,
|
2149 |
+
"loss": 0.4558,
|
2150 |
+
"step": 151000
|
2151 |
+
},
|
2152 |
+
{
|
2153 |
+
"epoch": 3.086294002607562,
|
2154 |
+
"grad_norm": 14.093750953674316,
|
2155 |
+
"learning_rate": 1.1482235984354629e-05,
|
2156 |
+
"loss": 0.4645,
|
2157 |
+
"step": 151500
|
2158 |
+
},
|
2159 |
+
{
|
2160 |
+
"epoch": 3.0964797913950455,
|
2161 |
+
"grad_norm": 25.050161361694336,
|
2162 |
+
"learning_rate": 1.1421121251629727e-05,
|
2163 |
+
"loss": 0.4549,
|
2164 |
+
"step": 152000
|
2165 |
+
},
|
2166 |
+
{
|
2167 |
+
"epoch": 3.106665580182529,
|
2168 |
+
"grad_norm": 12.942317962646484,
|
2169 |
+
"learning_rate": 1.1360006518904824e-05,
|
2170 |
+
"loss": 0.4675,
|
2171 |
+
"step": 152500
|
2172 |
+
},
|
2173 |
+
{
|
2174 |
+
"epoch": 3.1168513689700132,
|
2175 |
+
"grad_norm": 26.615070343017578,
|
2176 |
+
"learning_rate": 1.1298891786179922e-05,
|
2177 |
+
"loss": 0.4606,
|
2178 |
+
"step": 153000
|
2179 |
+
},
|
2180 |
+
{
|
2181 |
+
"epoch": 3.127037157757497,
|
2182 |
+
"grad_norm": 35.78953552246094,
|
2183 |
+
"learning_rate": 1.123777705345502e-05,
|
2184 |
+
"loss": 0.4754,
|
2185 |
+
"step": 153500
|
2186 |
+
},
|
2187 |
+
{
|
2188 |
+
"epoch": 3.1372229465449806,
|
2189 |
+
"grad_norm": 17.106658935546875,
|
2190 |
+
"learning_rate": 1.1176662320730117e-05,
|
2191 |
+
"loss": 0.46,
|
2192 |
+
"step": 154000
|
2193 |
+
},
|
2194 |
+
{
|
2195 |
+
"epoch": 3.1474087353324642,
|
2196 |
+
"grad_norm": 5.81476354598999,
|
2197 |
+
"learning_rate": 1.1115547588005214e-05,
|
2198 |
+
"loss": 0.455,
|
2199 |
+
"step": 154500
|
2200 |
+
},
|
2201 |
+
{
|
2202 |
+
"epoch": 3.157594524119948,
|
2203 |
+
"grad_norm": 13.205123901367188,
|
2204 |
+
"learning_rate": 1.1054432855280312e-05,
|
2205 |
+
"loss": 0.4443,
|
2206 |
+
"step": 155000
|
2207 |
+
},
|
2208 |
+
{
|
2209 |
+
"epoch": 3.1677803129074316,
|
2210 |
+
"grad_norm": 44.934471130371094,
|
2211 |
+
"learning_rate": 1.0993318122555411e-05,
|
2212 |
+
"loss": 0.447,
|
2213 |
+
"step": 155500
|
2214 |
+
},
|
2215 |
+
{
|
2216 |
+
"epoch": 3.1779661016949152,
|
2217 |
+
"grad_norm": 8.879769325256348,
|
2218 |
+
"learning_rate": 1.0932203389830509e-05,
|
2219 |
+
"loss": 0.4536,
|
2220 |
+
"step": 156000
|
2221 |
+
},
|
2222 |
+
{
|
2223 |
+
"epoch": 3.188151890482399,
|
2224 |
+
"grad_norm": 22.281354904174805,
|
2225 |
+
"learning_rate": 1.0871088657105608e-05,
|
2226 |
+
"loss": 0.4541,
|
2227 |
+
"step": 156500
|
2228 |
+
},
|
2229 |
+
{
|
2230 |
+
"epoch": 3.1983376792698825,
|
2231 |
+
"grad_norm": 46.504112243652344,
|
2232 |
+
"learning_rate": 1.0809973924380705e-05,
|
2233 |
+
"loss": 0.4467,
|
2234 |
+
"step": 157000
|
2235 |
+
},
|
2236 |
+
{
|
2237 |
+
"epoch": 3.208523468057366,
|
2238 |
+
"grad_norm": 1.4721815586090088,
|
2239 |
+
"learning_rate": 1.0748859191655803e-05,
|
2240 |
+
"loss": 0.4474,
|
2241 |
+
"step": 157500
|
2242 |
+
},
|
2243 |
+
{
|
2244 |
+
"epoch": 3.21870925684485,
|
2245 |
+
"grad_norm": 18.72897720336914,
|
2246 |
+
"learning_rate": 1.06877444589309e-05,
|
2247 |
+
"loss": 0.4465,
|
2248 |
+
"step": 158000
|
2249 |
+
},
|
2250 |
+
{
|
2251 |
+
"epoch": 3.228895045632334,
|
2252 |
+
"grad_norm": 13.687529563903809,
|
2253 |
+
"learning_rate": 1.0626629726205998e-05,
|
2254 |
+
"loss": 0.4674,
|
2255 |
+
"step": 158500
|
2256 |
+
},
|
2257 |
+
{
|
2258 |
+
"epoch": 3.2390808344198176,
|
2259 |
+
"grad_norm": 2.0822532176971436,
|
2260 |
+
"learning_rate": 1.0565514993481096e-05,
|
2261 |
+
"loss": 0.4516,
|
2262 |
+
"step": 159000
|
2263 |
+
},
|
2264 |
+
{
|
2265 |
+
"epoch": 3.2492666232073013,
|
2266 |
+
"grad_norm": 15.965363502502441,
|
2267 |
+
"learning_rate": 1.0504400260756193e-05,
|
2268 |
+
"loss": 0.4582,
|
2269 |
+
"step": 159500
|
2270 |
+
},
|
2271 |
+
{
|
2272 |
+
"epoch": 3.259452411994785,
|
2273 |
+
"grad_norm": 19.683805465698242,
|
2274 |
+
"learning_rate": 1.044328552803129e-05,
|
2275 |
+
"loss": 0.452,
|
2276 |
+
"step": 160000
|
2277 |
+
},
|
2278 |
+
{
|
2279 |
+
"epoch": 3.2696382007822686,
|
2280 |
+
"grad_norm": 30.53873062133789,
|
2281 |
+
"learning_rate": 1.0382170795306388e-05,
|
2282 |
+
"loss": 0.4532,
|
2283 |
+
"step": 160500
|
2284 |
+
},
|
2285 |
+
{
|
2286 |
+
"epoch": 3.2798239895697523,
|
2287 |
+
"grad_norm": 14.192473411560059,
|
2288 |
+
"learning_rate": 1.0321056062581486e-05,
|
2289 |
+
"loss": 0.4576,
|
2290 |
+
"step": 161000
|
2291 |
+
},
|
2292 |
+
{
|
2293 |
+
"epoch": 3.290009778357236,
|
2294 |
+
"grad_norm": 19.519214630126953,
|
2295 |
+
"learning_rate": 1.0259941329856583e-05,
|
2296 |
+
"loss": 0.4531,
|
2297 |
+
"step": 161500
|
2298 |
+
},
|
2299 |
+
{
|
2300 |
+
"epoch": 3.3001955671447196,
|
2301 |
+
"grad_norm": 11.308916091918945,
|
2302 |
+
"learning_rate": 1.0198826597131683e-05,
|
2303 |
+
"loss": 0.4374,
|
2304 |
+
"step": 162000
|
2305 |
+
},
|
2306 |
+
{
|
2307 |
+
"epoch": 3.3103813559322033,
|
2308 |
+
"grad_norm": 22.54949378967285,
|
2309 |
+
"learning_rate": 1.013771186440678e-05,
|
2310 |
+
"loss": 0.4641,
|
2311 |
+
"step": 162500
|
2312 |
+
},
|
2313 |
+
{
|
2314 |
+
"epoch": 3.320567144719687,
|
2315 |
+
"grad_norm": 27.339025497436523,
|
2316 |
+
"learning_rate": 1.0076597131681878e-05,
|
2317 |
+
"loss": 0.4603,
|
2318 |
+
"step": 163000
|
2319 |
+
},
|
2320 |
+
{
|
2321 |
+
"epoch": 3.3307529335071706,
|
2322 |
+
"grad_norm": 10.614120483398438,
|
2323 |
+
"learning_rate": 1.0015482398956975e-05,
|
2324 |
+
"loss": 0.4482,
|
2325 |
+
"step": 163500
|
2326 |
+
},
|
2327 |
+
{
|
2328 |
+
"epoch": 3.3409387222946547,
|
2329 |
+
"grad_norm": 19.92940902709961,
|
2330 |
+
"learning_rate": 9.954367666232073e-06,
|
2331 |
+
"loss": 0.4503,
|
2332 |
+
"step": 164000
|
2333 |
+
},
|
2334 |
+
{
|
2335 |
+
"epoch": 3.3511245110821384,
|
2336 |
+
"grad_norm": 21.356348037719727,
|
2337 |
+
"learning_rate": 9.89325293350717e-06,
|
2338 |
+
"loss": 0.4542,
|
2339 |
+
"step": 164500
|
2340 |
+
},
|
2341 |
+
{
|
2342 |
+
"epoch": 3.361310299869622,
|
2343 |
+
"grad_norm": 19.11351203918457,
|
2344 |
+
"learning_rate": 9.83213820078227e-06,
|
2345 |
+
"loss": 0.4653,
|
2346 |
+
"step": 165000
|
2347 |
+
},
|
2348 |
+
{
|
2349 |
+
"epoch": 3.3714960886571057,
|
2350 |
+
"grad_norm": 15.949396133422852,
|
2351 |
+
"learning_rate": 9.771023468057367e-06,
|
2352 |
+
"loss": 0.4583,
|
2353 |
+
"step": 165500
|
2354 |
+
},
|
2355 |
+
{
|
2356 |
+
"epoch": 3.3816818774445894,
|
2357 |
+
"grad_norm": 10.171916961669922,
|
2358 |
+
"learning_rate": 9.709908735332465e-06,
|
2359 |
+
"loss": 0.4747,
|
2360 |
+
"step": 166000
|
2361 |
+
},
|
2362 |
+
{
|
2363 |
+
"epoch": 3.391867666232073,
|
2364 |
+
"grad_norm": 17.548404693603516,
|
2365 |
+
"learning_rate": 9.648794002607562e-06,
|
2366 |
+
"loss": 0.4252,
|
2367 |
+
"step": 166500
|
2368 |
+
},
|
2369 |
+
{
|
2370 |
+
"epoch": 3.4020534550195567,
|
2371 |
+
"grad_norm": 26.116491317749023,
|
2372 |
+
"learning_rate": 9.58767926988266e-06,
|
2373 |
+
"loss": 0.4429,
|
2374 |
+
"step": 167000
|
2375 |
+
},
|
2376 |
+
{
|
2377 |
+
"epoch": 3.4122392438070404,
|
2378 |
+
"grad_norm": 8.717327117919922,
|
2379 |
+
"learning_rate": 9.526564537157757e-06,
|
2380 |
+
"loss": 0.4752,
|
2381 |
+
"step": 167500
|
2382 |
+
},
|
2383 |
+
{
|
2384 |
+
"epoch": 3.422425032594524,
|
2385 |
+
"grad_norm": 5.2715535163879395,
|
2386 |
+
"learning_rate": 9.465449804432857e-06,
|
2387 |
+
"loss": 0.4689,
|
2388 |
+
"step": 168000
|
2389 |
+
},
|
2390 |
+
{
|
2391 |
+
"epoch": 3.4326108213820077,
|
2392 |
+
"grad_norm": 15.369921684265137,
|
2393 |
+
"learning_rate": 9.404335071707954e-06,
|
2394 |
+
"loss": 0.474,
|
2395 |
+
"step": 168500
|
2396 |
+
},
|
2397 |
+
{
|
2398 |
+
"epoch": 3.4427966101694913,
|
2399 |
+
"grad_norm": 23.541234970092773,
|
2400 |
+
"learning_rate": 9.343220338983052e-06,
|
2401 |
+
"loss": 0.4717,
|
2402 |
+
"step": 169000
|
2403 |
+
},
|
2404 |
+
{
|
2405 |
+
"epoch": 3.4529823989569755,
|
2406 |
+
"grad_norm": 25.710681915283203,
|
2407 |
+
"learning_rate": 9.28210560625815e-06,
|
2408 |
+
"loss": 0.459,
|
2409 |
+
"step": 169500
|
2410 |
+
},
|
2411 |
+
{
|
2412 |
+
"epoch": 3.463168187744459,
|
2413 |
+
"grad_norm": 36.22392272949219,
|
2414 |
+
"learning_rate": 9.220990873533247e-06,
|
2415 |
+
"loss": 0.4374,
|
2416 |
+
"step": 170000
|
2417 |
+
},
|
2418 |
+
{
|
2419 |
+
"epoch": 3.4733539765319428,
|
2420 |
+
"grad_norm": 17.486045837402344,
|
2421 |
+
"learning_rate": 9.159876140808344e-06,
|
2422 |
+
"loss": 0.4682,
|
2423 |
+
"step": 170500
|
2424 |
+
},
|
2425 |
+
{
|
2426 |
+
"epoch": 3.4835397653194264,
|
2427 |
+
"grad_norm": 21.079090118408203,
|
2428 |
+
"learning_rate": 9.098761408083442e-06,
|
2429 |
+
"loss": 0.461,
|
2430 |
+
"step": 171000
|
2431 |
+
},
|
2432 |
+
{
|
2433 |
+
"epoch": 3.49372555410691,
|
2434 |
+
"grad_norm": 47.44207763671875,
|
2435 |
+
"learning_rate": 9.03764667535854e-06,
|
2436 |
+
"loss": 0.428,
|
2437 |
+
"step": 171500
|
2438 |
+
},
|
2439 |
+
{
|
2440 |
+
"epoch": 3.5039113428943938,
|
2441 |
+
"grad_norm": 10.158509254455566,
|
2442 |
+
"learning_rate": 8.976531942633637e-06,
|
2443 |
+
"loss": 0.4705,
|
2444 |
+
"step": 172000
|
2445 |
+
},
|
2446 |
+
{
|
2447 |
+
"epoch": 3.5140971316818774,
|
2448 |
+
"grad_norm": 2.3025388717651367,
|
2449 |
+
"learning_rate": 8.915417209908735e-06,
|
2450 |
+
"loss": 0.4794,
|
2451 |
+
"step": 172500
|
2452 |
+
},
|
2453 |
+
{
|
2454 |
+
"epoch": 3.524282920469361,
|
2455 |
+
"grad_norm": 7.106923580169678,
|
2456 |
+
"learning_rate": 8.854302477183832e-06,
|
2457 |
+
"loss": 0.4596,
|
2458 |
+
"step": 173000
|
2459 |
+
},
|
2460 |
+
{
|
2461 |
+
"epoch": 3.5344687092568448,
|
2462 |
+
"grad_norm": 11.846363067626953,
|
2463 |
+
"learning_rate": 8.79318774445893e-06,
|
2464 |
+
"loss": 0.4548,
|
2465 |
+
"step": 173500
|
2466 |
+
},
|
2467 |
+
{
|
2468 |
+
"epoch": 3.5446544980443284,
|
2469 |
+
"grad_norm": 23.35795783996582,
|
2470 |
+
"learning_rate": 8.73207301173403e-06,
|
2471 |
+
"loss": 0.4627,
|
2472 |
+
"step": 174000
|
2473 |
+
},
|
2474 |
+
{
|
2475 |
+
"epoch": 3.554840286831812,
|
2476 |
+
"grad_norm": 51.494873046875,
|
2477 |
+
"learning_rate": 8.670958279009128e-06,
|
2478 |
+
"loss": 0.443,
|
2479 |
+
"step": 174500
|
2480 |
+
},
|
2481 |
+
{
|
2482 |
+
"epoch": 3.565026075619296,
|
2483 |
+
"grad_norm": 17.30136489868164,
|
2484 |
+
"learning_rate": 8.609843546284226e-06,
|
2485 |
+
"loss": 0.4431,
|
2486 |
+
"step": 175000
|
2487 |
+
},
|
2488 |
+
{
|
2489 |
+
"epoch": 3.5752118644067794,
|
2490 |
+
"grad_norm": 6.228327751159668,
|
2491 |
+
"learning_rate": 8.548728813559323e-06,
|
2492 |
+
"loss": 0.4494,
|
2493 |
+
"step": 175500
|
2494 |
+
},
|
2495 |
+
{
|
2496 |
+
"epoch": 3.5853976531942635,
|
2497 |
+
"grad_norm": 7.742433071136475,
|
2498 |
+
"learning_rate": 8.48761408083442e-06,
|
2499 |
+
"loss": 0.4591,
|
2500 |
+
"step": 176000
|
2501 |
+
},
|
2502 |
+
{
|
2503 |
+
"epoch": 3.595583441981747,
|
2504 |
+
"grad_norm": 17.392818450927734,
|
2505 |
+
"learning_rate": 8.426499348109518e-06,
|
2506 |
+
"loss": 0.4572,
|
2507 |
+
"step": 176500
|
2508 |
+
},
|
2509 |
+
{
|
2510 |
+
"epoch": 3.605769230769231,
|
2511 |
+
"grad_norm": 9.273465156555176,
|
2512 |
+
"learning_rate": 8.365384615384616e-06,
|
2513 |
+
"loss": 0.4592,
|
2514 |
+
"step": 177000
|
2515 |
+
},
|
2516 |
+
{
|
2517 |
+
"epoch": 3.6159550195567145,
|
2518 |
+
"grad_norm": 10.859349250793457,
|
2519 |
+
"learning_rate": 8.304269882659713e-06,
|
2520 |
+
"loss": 0.4507,
|
2521 |
+
"step": 177500
|
2522 |
+
},
|
2523 |
+
{
|
2524 |
+
"epoch": 3.626140808344198,
|
2525 |
+
"grad_norm": 16.180192947387695,
|
2526 |
+
"learning_rate": 8.243155149934811e-06,
|
2527 |
+
"loss": 0.4615,
|
2528 |
+
"step": 178000
|
2529 |
+
},
|
2530 |
+
{
|
2531 |
+
"epoch": 3.636326597131682,
|
2532 |
+
"grad_norm": 7.720341682434082,
|
2533 |
+
"learning_rate": 8.182040417209908e-06,
|
2534 |
+
"loss": 0.4614,
|
2535 |
+
"step": 178500
|
2536 |
+
},
|
2537 |
+
{
|
2538 |
+
"epoch": 3.6465123859191655,
|
2539 |
+
"grad_norm": 10.102255821228027,
|
2540 |
+
"learning_rate": 8.120925684485006e-06,
|
2541 |
+
"loss": 0.4504,
|
2542 |
+
"step": 179000
|
2543 |
+
},
|
2544 |
+
{
|
2545 |
+
"epoch": 3.656698174706649,
|
2546 |
+
"grad_norm": 9.698884010314941,
|
2547 |
+
"learning_rate": 8.059810951760104e-06,
|
2548 |
+
"loss": 0.4791,
|
2549 |
+
"step": 179500
|
2550 |
+
},
|
2551 |
+
{
|
2552 |
+
"epoch": 3.666883963494133,
|
2553 |
+
"grad_norm": 13.645587921142578,
|
2554 |
+
"learning_rate": 7.998696219035203e-06,
|
2555 |
+
"loss": 0.4655,
|
2556 |
+
"step": 180000
|
2557 |
+
},
|
2558 |
+
{
|
2559 |
+
"epoch": 3.677069752281617,
|
2560 |
+
"grad_norm": 6.0963358879089355,
|
2561 |
+
"learning_rate": 7.9375814863103e-06,
|
2562 |
+
"loss": 0.4522,
|
2563 |
+
"step": 180500
|
2564 |
+
},
|
2565 |
+
{
|
2566 |
+
"epoch": 3.6872555410691,
|
2567 |
+
"grad_norm": 19.082395553588867,
|
2568 |
+
"learning_rate": 7.876466753585398e-06,
|
2569 |
+
"loss": 0.4606,
|
2570 |
+
"step": 181000
|
2571 |
+
},
|
2572 |
+
{
|
2573 |
+
"epoch": 3.6974413298565842,
|
2574 |
+
"grad_norm": 21.68135643005371,
|
2575 |
+
"learning_rate": 7.815352020860495e-06,
|
2576 |
+
"loss": 0.475,
|
2577 |
+
"step": 181500
|
2578 |
+
},
|
2579 |
+
{
|
2580 |
+
"epoch": 3.707627118644068,
|
2581 |
+
"grad_norm": 22.7216796875,
|
2582 |
+
"learning_rate": 7.754237288135593e-06,
|
2583 |
+
"loss": 0.4561,
|
2584 |
+
"step": 182000
|
2585 |
+
},
|
2586 |
+
{
|
2587 |
+
"epoch": 3.7178129074315516,
|
2588 |
+
"grad_norm": 5.14304780960083,
|
2589 |
+
"learning_rate": 7.69312255541069e-06,
|
2590 |
+
"loss": 0.4561,
|
2591 |
+
"step": 182500
|
2592 |
+
},
|
2593 |
+
{
|
2594 |
+
"epoch": 3.7279986962190352,
|
2595 |
+
"grad_norm": 12.879047393798828,
|
2596 |
+
"learning_rate": 7.63200782268579e-06,
|
2597 |
+
"loss": 0.4619,
|
2598 |
+
"step": 183000
|
2599 |
+
},
|
2600 |
+
{
|
2601 |
+
"epoch": 3.738184485006519,
|
2602 |
+
"grad_norm": 20.963592529296875,
|
2603 |
+
"learning_rate": 7.570893089960887e-06,
|
2604 |
+
"loss": 0.4488,
|
2605 |
+
"step": 183500
|
2606 |
+
},
|
2607 |
+
{
|
2608 |
+
"epoch": 3.7483702737940026,
|
2609 |
+
"grad_norm": 15.685653686523438,
|
2610 |
+
"learning_rate": 7.509778357235985e-06,
|
2611 |
+
"loss": 0.4515,
|
2612 |
+
"step": 184000
|
2613 |
+
},
|
2614 |
+
{
|
2615 |
+
"epoch": 3.7585560625814862,
|
2616 |
+
"grad_norm": 14.296714782714844,
|
2617 |
+
"learning_rate": 7.4486636245110824e-06,
|
2618 |
+
"loss": 0.4461,
|
2619 |
+
"step": 184500
|
2620 |
+
},
|
2621 |
+
{
|
2622 |
+
"epoch": 3.76874185136897,
|
2623 |
+
"grad_norm": 8.592365264892578,
|
2624 |
+
"learning_rate": 7.38754889178618e-06,
|
2625 |
+
"loss": 0.4571,
|
2626 |
+
"step": 185000
|
2627 |
+
},
|
2628 |
+
{
|
2629 |
+
"epoch": 3.7789276401564535,
|
2630 |
+
"grad_norm": 32.55515670776367,
|
2631 |
+
"learning_rate": 7.326434159061278e-06,
|
2632 |
+
"loss": 0.4601,
|
2633 |
+
"step": 185500
|
2634 |
+
},
|
2635 |
+
{
|
2636 |
+
"epoch": 3.7891134289439377,
|
2637 |
+
"grad_norm": 19.179519653320312,
|
2638 |
+
"learning_rate": 7.265319426336376e-06,
|
2639 |
+
"loss": 0.4591,
|
2640 |
+
"step": 186000
|
2641 |
+
},
|
2642 |
+
{
|
2643 |
+
"epoch": 3.799299217731421,
|
2644 |
+
"grad_norm": 11.681869506835938,
|
2645 |
+
"learning_rate": 7.2042046936114735e-06,
|
2646 |
+
"loss": 0.4638,
|
2647 |
+
"step": 186500
|
2648 |
+
},
|
2649 |
+
{
|
2650 |
+
"epoch": 3.809485006518905,
|
2651 |
+
"grad_norm": 17.548564910888672,
|
2652 |
+
"learning_rate": 7.143089960886571e-06,
|
2653 |
+
"loss": 0.4595,
|
2654 |
+
"step": 187000
|
2655 |
+
},
|
2656 |
+
{
|
2657 |
+
"epoch": 3.8196707953063886,
|
2658 |
+
"grad_norm": 34.2673225402832,
|
2659 |
+
"learning_rate": 7.0819752281616686e-06,
|
2660 |
+
"loss": 0.4622,
|
2661 |
+
"step": 187500
|
2662 |
+
},
|
2663 |
+
{
|
2664 |
+
"epoch": 3.8298565840938723,
|
2665 |
+
"grad_norm": 25.60137939453125,
|
2666 |
+
"learning_rate": 7.020860495436767e-06,
|
2667 |
+
"loss": 0.4653,
|
2668 |
+
"step": 188000
|
2669 |
+
},
|
2670 |
+
{
|
2671 |
+
"epoch": 3.840042372881356,
|
2672 |
+
"grad_norm": 14.652885437011719,
|
2673 |
+
"learning_rate": 6.9597457627118645e-06,
|
2674 |
+
"loss": 0.4484,
|
2675 |
+
"step": 188500
|
2676 |
+
},
|
2677 |
+
{
|
2678 |
+
"epoch": 3.8502281616688396,
|
2679 |
+
"grad_norm": 20.71872901916504,
|
2680 |
+
"learning_rate": 6.898631029986962e-06,
|
2681 |
+
"loss": 0.4717,
|
2682 |
+
"step": 189000
|
2683 |
+
},
|
2684 |
+
{
|
2685 |
+
"epoch": 3.8604139504563233,
|
2686 |
+
"grad_norm": 6.4770989418029785,
|
2687 |
+
"learning_rate": 6.83751629726206e-06,
|
2688 |
+
"loss": 0.462,
|
2689 |
+
"step": 189500
|
2690 |
+
},
|
2691 |
+
{
|
2692 |
+
"epoch": 3.870599739243807,
|
2693 |
+
"grad_norm": 21.029338836669922,
|
2694 |
+
"learning_rate": 6.776401564537158e-06,
|
2695 |
+
"loss": 0.4312,
|
2696 |
+
"step": 190000
|
2697 |
+
},
|
2698 |
+
{
|
2699 |
+
"epoch": 3.8807855280312906,
|
2700 |
+
"grad_norm": 27.267345428466797,
|
2701 |
+
"learning_rate": 6.7152868318122556e-06,
|
2702 |
+
"loss": 0.4623,
|
2703 |
+
"step": 190500
|
2704 |
+
},
|
2705 |
+
{
|
2706 |
+
"epoch": 3.8909713168187743,
|
2707 |
+
"grad_norm": 16.270328521728516,
|
2708 |
+
"learning_rate": 6.654172099087354e-06,
|
2709 |
+
"loss": 0.4594,
|
2710 |
+
"step": 191000
|
2711 |
+
},
|
2712 |
+
{
|
2713 |
+
"epoch": 3.9011571056062584,
|
2714 |
+
"grad_norm": 15.952796936035156,
|
2715 |
+
"learning_rate": 6.5930573663624515e-06,
|
2716 |
+
"loss": 0.4478,
|
2717 |
+
"step": 191500
|
2718 |
+
},
|
2719 |
+
{
|
2720 |
+
"epoch": 3.9113428943937416,
|
2721 |
+
"grad_norm": 24.082805633544922,
|
2722 |
+
"learning_rate": 6.531942633637549e-06,
|
2723 |
+
"loss": 0.4572,
|
2724 |
+
"step": 192000
|
2725 |
+
},
|
2726 |
+
{
|
2727 |
+
"epoch": 3.9215286831812257,
|
2728 |
+
"grad_norm": 17.784955978393555,
|
2729 |
+
"learning_rate": 6.470827900912647e-06,
|
2730 |
+
"loss": 0.453,
|
2731 |
+
"step": 192500
|
2732 |
+
},
|
2733 |
+
{
|
2734 |
+
"epoch": 3.9317144719687094,
|
2735 |
+
"grad_norm": 15.306330680847168,
|
2736 |
+
"learning_rate": 6.409713168187744e-06,
|
2737 |
+
"loss": 0.4516,
|
2738 |
+
"step": 193000
|
2739 |
+
},
|
2740 |
+
{
|
2741 |
+
"epoch": 3.941900260756193,
|
2742 |
+
"grad_norm": 17.07221031188965,
|
2743 |
+
"learning_rate": 6.348598435462842e-06,
|
2744 |
+
"loss": 0.4513,
|
2745 |
+
"step": 193500
|
2746 |
+
},
|
2747 |
+
{
|
2748 |
+
"epoch": 3.9520860495436767,
|
2749 |
+
"grad_norm": 1.9630801677703857,
|
2750 |
+
"learning_rate": 6.287483702737941e-06,
|
2751 |
+
"loss": 0.4339,
|
2752 |
+
"step": 194000
|
2753 |
+
},
|
2754 |
+
{
|
2755 |
+
"epoch": 3.9622718383311604,
|
2756 |
+
"grad_norm": 23.103532791137695,
|
2757 |
+
"learning_rate": 6.2263689700130385e-06,
|
2758 |
+
"loss": 0.475,
|
2759 |
+
"step": 194500
|
2760 |
+
},
|
2761 |
+
{
|
2762 |
+
"epoch": 3.972457627118644,
|
2763 |
+
"grad_norm": 9.09752082824707,
|
2764 |
+
"learning_rate": 6.165254237288136e-06,
|
2765 |
+
"loss": 0.4422,
|
2766 |
+
"step": 195000
|
2767 |
+
},
|
2768 |
+
{
|
2769 |
+
"epoch": 3.9826434159061277,
|
2770 |
+
"grad_norm": 23.36123275756836,
|
2771 |
+
"learning_rate": 6.104139504563234e-06,
|
2772 |
+
"loss": 0.4578,
|
2773 |
+
"step": 195500
|
2774 |
+
},
|
2775 |
+
{
|
2776 |
+
"epoch": 3.9928292046936114,
|
2777 |
+
"grad_norm": 15.44927978515625,
|
2778 |
+
"learning_rate": 6.043024771838331e-06,
|
2779 |
+
"loss": 0.4815,
|
2780 |
+
"step": 196000
|
2781 |
+
},
|
2782 |
+
{
|
2783 |
+
"epoch": 4.0,
|
2784 |
+
"eval_accuracy": 0.7300326228141785,
|
2785 |
+
"eval_loss": 0.8482908010482788,
|
2786 |
+
"eval_runtime": 7.2526,
|
2787 |
+
"eval_samples_per_second": 676.722,
|
2788 |
+
"eval_steps_per_second": 84.659,
|
2789 |
+
"step": 196352
|
2790 |
+
},
|
2791 |
+
{
|
2792 |
+
"epoch": 4.003014993481095,
|
2793 |
+
"grad_norm": 18.638896942138672,
|
2794 |
+
"learning_rate": 5.981910039113429e-06,
|
2795 |
+
"loss": 0.4162,
|
2796 |
+
"step": 196500
|
2797 |
+
},
|
2798 |
+
{
|
2799 |
+
"epoch": 4.013200782268579,
|
2800 |
+
"grad_norm": 23.482778549194336,
|
2801 |
+
"learning_rate": 5.920795306388527e-06,
|
2802 |
+
"loss": 0.3857,
|
2803 |
+
"step": 197000
|
2804 |
+
},
|
2805 |
+
{
|
2806 |
+
"epoch": 4.023386571056062,
|
2807 |
+
"grad_norm": 19.941246032714844,
|
2808 |
+
"learning_rate": 5.859680573663625e-06,
|
2809 |
+
"loss": 0.4055,
|
2810 |
+
"step": 197500
|
2811 |
+
},
|
2812 |
+
{
|
2813 |
+
"epoch": 4.0335723598435465,
|
2814 |
+
"grad_norm": 55.201602935791016,
|
2815 |
+
"learning_rate": 5.798565840938722e-06,
|
2816 |
+
"loss": 0.4054,
|
2817 |
+
"step": 198000
|
2818 |
+
},
|
2819 |
+
{
|
2820 |
+
"epoch": 4.04375814863103,
|
2821 |
+
"grad_norm": 9.683618545532227,
|
2822 |
+
"learning_rate": 5.7374511082138206e-06,
|
2823 |
+
"loss": 0.3772,
|
2824 |
+
"step": 198500
|
2825 |
+
},
|
2826 |
+
{
|
2827 |
+
"epoch": 4.053943937418514,
|
2828 |
+
"grad_norm": 14.284856796264648,
|
2829 |
+
"learning_rate": 5.676336375488918e-06,
|
2830 |
+
"loss": 0.3761,
|
2831 |
+
"step": 199000
|
2832 |
+
},
|
2833 |
+
{
|
2834 |
+
"epoch": 4.064129726205997,
|
2835 |
+
"grad_norm": 33.600894927978516,
|
2836 |
+
"learning_rate": 5.615221642764016e-06,
|
2837 |
+
"loss": 0.394,
|
2838 |
+
"step": 199500
|
2839 |
+
},
|
2840 |
+
{
|
2841 |
+
"epoch": 4.074315514993481,
|
2842 |
+
"grad_norm": 4.879152774810791,
|
2843 |
+
"learning_rate": 5.554106910039114e-06,
|
2844 |
+
"loss": 0.3754,
|
2845 |
+
"step": 200000
|
2846 |
+
},
|
2847 |
+
{
|
2848 |
+
"epoch": 4.084501303780965,
|
2849 |
+
"grad_norm": 4.289336681365967,
|
2850 |
+
"learning_rate": 5.492992177314212e-06,
|
2851 |
+
"loss": 0.3822,
|
2852 |
+
"step": 200500
|
2853 |
+
},
|
2854 |
+
{
|
2855 |
+
"epoch": 4.094687092568448,
|
2856 |
+
"grad_norm": 13.225927352905273,
|
2857 |
+
"learning_rate": 5.431877444589309e-06,
|
2858 |
+
"loss": 0.385,
|
2859 |
+
"step": 201000
|
2860 |
+
},
|
2861 |
+
{
|
2862 |
+
"epoch": 4.1048728813559325,
|
2863 |
+
"grad_norm": 30.11479949951172,
|
2864 |
+
"learning_rate": 5.370762711864407e-06,
|
2865 |
+
"loss": 0.4043,
|
2866 |
+
"step": 201500
|
2867 |
+
},
|
2868 |
+
{
|
2869 |
+
"epoch": 4.115058670143416,
|
2870 |
+
"grad_norm": 30.226152420043945,
|
2871 |
+
"learning_rate": 5.309647979139504e-06,
|
2872 |
+
"loss": 0.3785,
|
2873 |
+
"step": 202000
|
2874 |
+
},
|
2875 |
+
{
|
2876 |
+
"epoch": 4.1252444589309,
|
2877 |
+
"grad_norm": 57.69206619262695,
|
2878 |
+
"learning_rate": 5.248533246414602e-06,
|
2879 |
+
"loss": 0.3783,
|
2880 |
+
"step": 202500
|
2881 |
+
},
|
2882 |
+
{
|
2883 |
+
"epoch": 4.135430247718383,
|
2884 |
+
"grad_norm": 0.5498570203781128,
|
2885 |
+
"learning_rate": 5.187418513689701e-06,
|
2886 |
+
"loss": 0.387,
|
2887 |
+
"step": 203000
|
2888 |
+
},
|
2889 |
+
{
|
2890 |
+
"epoch": 4.145616036505867,
|
2891 |
+
"grad_norm": 1.2860363721847534,
|
2892 |
+
"learning_rate": 5.126303780964799e-06,
|
2893 |
+
"loss": 0.3788,
|
2894 |
+
"step": 203500
|
2895 |
+
},
|
2896 |
+
{
|
2897 |
+
"epoch": 4.15580182529335,
|
2898 |
+
"grad_norm": 24.312036514282227,
|
2899 |
+
"learning_rate": 5.065189048239896e-06,
|
2900 |
+
"loss": 0.3958,
|
2901 |
+
"step": 204000
|
2902 |
+
},
|
2903 |
+
{
|
2904 |
+
"epoch": 4.1659876140808345,
|
2905 |
+
"grad_norm": 31.0595645904541,
|
2906 |
+
"learning_rate": 5.004074315514994e-06,
|
2907 |
+
"loss": 0.368,
|
2908 |
+
"step": 204500
|
2909 |
+
},
|
2910 |
+
{
|
2911 |
+
"epoch": 4.176173402868318,
|
2912 |
+
"grad_norm": 30.00829315185547,
|
2913 |
+
"learning_rate": 4.942959582790091e-06,
|
2914 |
+
"loss": 0.387,
|
2915 |
+
"step": 205000
|
2916 |
+
},
|
2917 |
+
{
|
2918 |
+
"epoch": 4.186359191655802,
|
2919 |
+
"grad_norm": 5.656859874725342,
|
2920 |
+
"learning_rate": 4.881844850065189e-06,
|
2921 |
+
"loss": 0.3876,
|
2922 |
+
"step": 205500
|
2923 |
+
},
|
2924 |
+
{
|
2925 |
+
"epoch": 4.196544980443286,
|
2926 |
+
"grad_norm": 28.364885330200195,
|
2927 |
+
"learning_rate": 4.820730117340287e-06,
|
2928 |
+
"loss": 0.3974,
|
2929 |
+
"step": 206000
|
2930 |
+
},
|
2931 |
+
{
|
2932 |
+
"epoch": 4.206730769230769,
|
2933 |
+
"grad_norm": 21.450817108154297,
|
2934 |
+
"learning_rate": 4.759615384615385e-06,
|
2935 |
+
"loss": 0.4304,
|
2936 |
+
"step": 206500
|
2937 |
+
},
|
2938 |
+
{
|
2939 |
+
"epoch": 4.216916558018253,
|
2940 |
+
"grad_norm": 26.55447769165039,
|
2941 |
+
"learning_rate": 4.698500651890482e-06,
|
2942 |
+
"loss": 0.3598,
|
2943 |
+
"step": 207000
|
2944 |
+
},
|
2945 |
+
{
|
2946 |
+
"epoch": 4.2271023468057365,
|
2947 |
+
"grad_norm": 63.503299713134766,
|
2948 |
+
"learning_rate": 4.637385919165581e-06,
|
2949 |
+
"loss": 0.3972,
|
2950 |
+
"step": 207500
|
2951 |
+
},
|
2952 |
+
{
|
2953 |
+
"epoch": 4.237288135593221,
|
2954 |
+
"grad_norm": 6.910649299621582,
|
2955 |
+
"learning_rate": 4.576271186440678e-06,
|
2956 |
+
"loss": 0.3788,
|
2957 |
+
"step": 208000
|
2958 |
+
},
|
2959 |
+
{
|
2960 |
+
"epoch": 4.247473924380704,
|
2961 |
+
"grad_norm": 6.5545172691345215,
|
2962 |
+
"learning_rate": 4.515156453715776e-06,
|
2963 |
+
"loss": 0.3978,
|
2964 |
+
"step": 208500
|
2965 |
+
},
|
2966 |
+
{
|
2967 |
+
"epoch": 4.257659713168188,
|
2968 |
+
"grad_norm": 17.815195083618164,
|
2969 |
+
"learning_rate": 4.454041720990874e-06,
|
2970 |
+
"loss": 0.4014,
|
2971 |
+
"step": 209000
|
2972 |
+
},
|
2973 |
+
{
|
2974 |
+
"epoch": 4.267845501955671,
|
2975 |
+
"grad_norm": 33.210113525390625,
|
2976 |
+
"learning_rate": 4.392926988265972e-06,
|
2977 |
+
"loss": 0.3891,
|
2978 |
+
"step": 209500
|
2979 |
+
},
|
2980 |
+
{
|
2981 |
+
"epoch": 4.278031290743155,
|
2982 |
+
"grad_norm": 33.96702194213867,
|
2983 |
+
"learning_rate": 4.331812255541069e-06,
|
2984 |
+
"loss": 0.3891,
|
2985 |
+
"step": 210000
|
2986 |
+
},
|
2987 |
+
{
|
2988 |
+
"epoch": 4.2882170795306385,
|
2989 |
+
"grad_norm": 13.674240112304688,
|
2990 |
+
"learning_rate": 4.270697522816167e-06,
|
2991 |
+
"loss": 0.3743,
|
2992 |
+
"step": 210500
|
2993 |
+
},
|
2994 |
+
{
|
2995 |
+
"epoch": 4.298402868318123,
|
2996 |
+
"grad_norm": 1.694655179977417,
|
2997 |
+
"learning_rate": 4.209582790091264e-06,
|
2998 |
+
"loss": 0.4072,
|
2999 |
+
"step": 211000
|
3000 |
+
},
|
3001 |
+
{
|
3002 |
+
"epoch": 4.308588657105606,
|
3003 |
+
"grad_norm": 10.407938957214355,
|
3004 |
+
"learning_rate": 4.148468057366362e-06,
|
3005 |
+
"loss": 0.3788,
|
3006 |
+
"step": 211500
|
3007 |
+
},
|
3008 |
+
{
|
3009 |
+
"epoch": 4.31877444589309,
|
3010 |
+
"grad_norm": 28.451736450195312,
|
3011 |
+
"learning_rate": 4.087353324641461e-06,
|
3012 |
+
"loss": 0.44,
|
3013 |
+
"step": 212000
|
3014 |
+
},
|
3015 |
+
{
|
3016 |
+
"epoch": 4.328960234680574,
|
3017 |
+
"grad_norm": 40.655616760253906,
|
3018 |
+
"learning_rate": 4.026238591916559e-06,
|
3019 |
+
"loss": 0.3846,
|
3020 |
+
"step": 212500
|
3021 |
+
},
|
3022 |
+
{
|
3023 |
+
"epoch": 4.339146023468057,
|
3024 |
+
"grad_norm": 33.640682220458984,
|
3025 |
+
"learning_rate": 3.965123859191656e-06,
|
3026 |
+
"loss": 0.3967,
|
3027 |
+
"step": 213000
|
3028 |
+
},
|
3029 |
+
{
|
3030 |
+
"epoch": 4.349331812255541,
|
3031 |
+
"grad_norm": 77.51473236083984,
|
3032 |
+
"learning_rate": 3.904009126466754e-06,
|
3033 |
+
"loss": 0.384,
|
3034 |
+
"step": 213500
|
3035 |
+
},
|
3036 |
+
{
|
3037 |
+
"epoch": 4.3595176010430245,
|
3038 |
+
"grad_norm": 15.64946174621582,
|
3039 |
+
"learning_rate": 3.842894393741851e-06,
|
3040 |
+
"loss": 0.3935,
|
3041 |
+
"step": 214000
|
3042 |
+
},
|
3043 |
+
{
|
3044 |
+
"epoch": 4.369703389830509,
|
3045 |
+
"grad_norm": 38.03608703613281,
|
3046 |
+
"learning_rate": 3.7817796610169493e-06,
|
3047 |
+
"loss": 0.3933,
|
3048 |
+
"step": 214500
|
3049 |
+
},
|
3050 |
+
{
|
3051 |
+
"epoch": 4.379889178617992,
|
3052 |
+
"grad_norm": 24.385404586791992,
|
3053 |
+
"learning_rate": 3.720664928292047e-06,
|
3054 |
+
"loss": 0.3879,
|
3055 |
+
"step": 215000
|
3056 |
+
},
|
3057 |
+
{
|
3058 |
+
"epoch": 4.390074967405476,
|
3059 |
+
"grad_norm": 19.359683990478516,
|
3060 |
+
"learning_rate": 3.659550195567145e-06,
|
3061 |
+
"loss": 0.4027,
|
3062 |
+
"step": 215500
|
3063 |
+
},
|
3064 |
+
{
|
3065 |
+
"epoch": 4.400260756192959,
|
3066 |
+
"grad_norm": 20.058656692504883,
|
3067 |
+
"learning_rate": 3.598435462842243e-06,
|
3068 |
+
"loss": 0.3947,
|
3069 |
+
"step": 216000
|
3070 |
+
},
|
3071 |
+
{
|
3072 |
+
"epoch": 4.410446544980443,
|
3073 |
+
"grad_norm": 22.407522201538086,
|
3074 |
+
"learning_rate": 3.5373207301173403e-06,
|
3075 |
+
"loss": 0.3692,
|
3076 |
+
"step": 216500
|
3077 |
+
},
|
3078 |
+
{
|
3079 |
+
"epoch": 4.4206323337679265,
|
3080 |
+
"grad_norm": 0.43725308775901794,
|
3081 |
+
"learning_rate": 3.476205997392438e-06,
|
3082 |
+
"loss": 0.3876,
|
3083 |
+
"step": 217000
|
3084 |
+
},
|
3085 |
+
{
|
3086 |
+
"epoch": 4.430818122555411,
|
3087 |
+
"grad_norm": 6.114492416381836,
|
3088 |
+
"learning_rate": 3.4150912646675363e-06,
|
3089 |
+
"loss": 0.3812,
|
3090 |
+
"step": 217500
|
3091 |
+
},
|
3092 |
+
{
|
3093 |
+
"epoch": 4.441003911342895,
|
3094 |
+
"grad_norm": 8.737237930297852,
|
3095 |
+
"learning_rate": 3.353976531942634e-06,
|
3096 |
+
"loss": 0.4218,
|
3097 |
+
"step": 218000
|
3098 |
+
},
|
3099 |
+
{
|
3100 |
+
"epoch": 4.451189700130378,
|
3101 |
+
"grad_norm": 21.46657371520996,
|
3102 |
+
"learning_rate": 3.2928617992177314e-06,
|
3103 |
+
"loss": 0.3878,
|
3104 |
+
"step": 218500
|
3105 |
+
},
|
3106 |
+
{
|
3107 |
+
"epoch": 4.461375488917862,
|
3108 |
+
"grad_norm": 28.781583786010742,
|
3109 |
+
"learning_rate": 3.2317470664928294e-06,
|
3110 |
+
"loss": 0.4014,
|
3111 |
+
"step": 219000
|
3112 |
+
},
|
3113 |
+
{
|
3114 |
+
"epoch": 4.471561277705345,
|
3115 |
+
"grad_norm": 65.7254867553711,
|
3116 |
+
"learning_rate": 3.170632333767927e-06,
|
3117 |
+
"loss": 0.3846,
|
3118 |
+
"step": 219500
|
3119 |
+
},
|
3120 |
+
{
|
3121 |
+
"epoch": 4.481747066492829,
|
3122 |
+
"grad_norm": 42.237693786621094,
|
3123 |
+
"learning_rate": 3.109517601043025e-06,
|
3124 |
+
"loss": 0.3945,
|
3125 |
+
"step": 220000
|
3126 |
+
},
|
3127 |
+
{
|
3128 |
+
"epoch": 4.491932855280313,
|
3129 |
+
"grad_norm": 22.627384185791016,
|
3130 |
+
"learning_rate": 3.048402868318123e-06,
|
3131 |
+
"loss": 0.3846,
|
3132 |
+
"step": 220500
|
3133 |
+
},
|
3134 |
+
{
|
3135 |
+
"epoch": 4.502118644067797,
|
3136 |
+
"grad_norm": 5.28376579284668,
|
3137 |
+
"learning_rate": 2.9872881355932204e-06,
|
3138 |
+
"loss": 0.3951,
|
3139 |
+
"step": 221000
|
3140 |
+
},
|
3141 |
+
{
|
3142 |
+
"epoch": 4.51230443285528,
|
3143 |
+
"grad_norm": 29.31270980834961,
|
3144 |
+
"learning_rate": 2.926173402868318e-06,
|
3145 |
+
"loss": 0.387,
|
3146 |
+
"step": 221500
|
3147 |
+
},
|
3148 |
+
{
|
3149 |
+
"epoch": 4.522490221642764,
|
3150 |
+
"grad_norm": 28.06836700439453,
|
3151 |
+
"learning_rate": 2.8650586701434163e-06,
|
3152 |
+
"loss": 0.3758,
|
3153 |
+
"step": 222000
|
3154 |
+
},
|
3155 |
+
{
|
3156 |
+
"epoch": 4.532676010430247,
|
3157 |
+
"grad_norm": 19.28433609008789,
|
3158 |
+
"learning_rate": 2.803943937418514e-06,
|
3159 |
+
"loss": 0.4011,
|
3160 |
+
"step": 222500
|
3161 |
+
},
|
3162 |
+
{
|
3163 |
+
"epoch": 4.542861799217731,
|
3164 |
+
"grad_norm": 28.01453399658203,
|
3165 |
+
"learning_rate": 2.7428292046936114e-06,
|
3166 |
+
"loss": 0.4123,
|
3167 |
+
"step": 223000
|
3168 |
+
},
|
3169 |
+
{
|
3170 |
+
"epoch": 4.5530475880052155,
|
3171 |
+
"grad_norm": 22.563831329345703,
|
3172 |
+
"learning_rate": 2.6817144719687094e-06,
|
3173 |
+
"loss": 0.3901,
|
3174 |
+
"step": 223500
|
3175 |
+
},
|
3176 |
+
{
|
3177 |
+
"epoch": 4.563233376792699,
|
3178 |
+
"grad_norm": 6.881279945373535,
|
3179 |
+
"learning_rate": 2.620599739243807e-06,
|
3180 |
+
"loss": 0.3641,
|
3181 |
+
"step": 224000
|
3182 |
+
},
|
3183 |
+
{
|
3184 |
+
"epoch": 4.573419165580183,
|
3185 |
+
"grad_norm": 10.799131393432617,
|
3186 |
+
"learning_rate": 2.559485006518905e-06,
|
3187 |
+
"loss": 0.388,
|
3188 |
+
"step": 224500
|
3189 |
+
},
|
3190 |
+
{
|
3191 |
+
"epoch": 4.583604954367666,
|
3192 |
+
"grad_norm": 8.748978614807129,
|
3193 |
+
"learning_rate": 2.498370273794003e-06,
|
3194 |
+
"loss": 0.3855,
|
3195 |
+
"step": 225000
|
3196 |
+
},
|
3197 |
+
{
|
3198 |
+
"epoch": 4.59379074315515,
|
3199 |
+
"grad_norm": 30.88077163696289,
|
3200 |
+
"learning_rate": 2.4372555410691004e-06,
|
3201 |
+
"loss": 0.3846,
|
3202 |
+
"step": 225500
|
3203 |
+
},
|
3204 |
+
{
|
3205 |
+
"epoch": 4.603976531942633,
|
3206 |
+
"grad_norm": 9.857905387878418,
|
3207 |
+
"learning_rate": 2.376140808344198e-06,
|
3208 |
+
"loss": 0.3887,
|
3209 |
+
"step": 226000
|
3210 |
+
},
|
3211 |
+
{
|
3212 |
+
"epoch": 4.6141623207301175,
|
3213 |
+
"grad_norm": 18.84724235534668,
|
3214 |
+
"learning_rate": 2.3150260756192964e-06,
|
3215 |
+
"loss": 0.3981,
|
3216 |
+
"step": 226500
|
3217 |
+
},
|
3218 |
+
{
|
3219 |
+
"epoch": 4.624348109517601,
|
3220 |
+
"grad_norm": 1.5743709802627563,
|
3221 |
+
"learning_rate": 2.253911342894394e-06,
|
3222 |
+
"loss": 0.4198,
|
3223 |
+
"step": 227000
|
3224 |
+
},
|
3225 |
+
{
|
3226 |
+
"epoch": 4.634533898305085,
|
3227 |
+
"grad_norm": 1.2036515474319458,
|
3228 |
+
"learning_rate": 2.1927966101694915e-06,
|
3229 |
+
"loss": 0.3883,
|
3230 |
+
"step": 227500
|
3231 |
+
},
|
3232 |
+
{
|
3233 |
+
"epoch": 4.644719687092568,
|
3234 |
+
"grad_norm": 25.730968475341797,
|
3235 |
+
"learning_rate": 2.1316818774445895e-06,
|
3236 |
+
"loss": 0.3791,
|
3237 |
+
"step": 228000
|
3238 |
+
},
|
3239 |
+
{
|
3240 |
+
"epoch": 4.654905475880052,
|
3241 |
+
"grad_norm": 32.65802764892578,
|
3242 |
+
"learning_rate": 2.070567144719687e-06,
|
3243 |
+
"loss": 0.3772,
|
3244 |
+
"step": 228500
|
3245 |
+
},
|
3246 |
+
{
|
3247 |
+
"epoch": 4.665091264667536,
|
3248 |
+
"grad_norm": 3.2796003818511963,
|
3249 |
+
"learning_rate": 2.009452411994785e-06,
|
3250 |
+
"loss": 0.3994,
|
3251 |
+
"step": 229000
|
3252 |
+
},
|
3253 |
+
{
|
3254 |
+
"epoch": 4.675277053455019,
|
3255 |
+
"grad_norm": 12.903874397277832,
|
3256 |
+
"learning_rate": 1.948337679269883e-06,
|
3257 |
+
"loss": 0.3725,
|
3258 |
+
"step": 229500
|
3259 |
+
},
|
3260 |
+
{
|
3261 |
+
"epoch": 4.6854628422425035,
|
3262 |
+
"grad_norm": 1.0021297931671143,
|
3263 |
+
"learning_rate": 1.8872229465449805e-06,
|
3264 |
+
"loss": 0.393,
|
3265 |
+
"step": 230000
|
3266 |
+
},
|
3267 |
+
{
|
3268 |
+
"epoch": 4.695648631029987,
|
3269 |
+
"grad_norm": 5.547977447509766,
|
3270 |
+
"learning_rate": 1.8261082138200783e-06,
|
3271 |
+
"loss": 0.3907,
|
3272 |
+
"step": 230500
|
3273 |
+
},
|
3274 |
+
{
|
3275 |
+
"epoch": 4.705834419817471,
|
3276 |
+
"grad_norm": 5.375828742980957,
|
3277 |
+
"learning_rate": 1.764993481095176e-06,
|
3278 |
+
"loss": 0.4023,
|
3279 |
+
"step": 231000
|
3280 |
+
},
|
3281 |
+
{
|
3282 |
+
"epoch": 4.716020208604954,
|
3283 |
+
"grad_norm": 20.087383270263672,
|
3284 |
+
"learning_rate": 1.703878748370274e-06,
|
3285 |
+
"loss": 0.3652,
|
3286 |
+
"step": 231500
|
3287 |
+
},
|
3288 |
+
{
|
3289 |
+
"epoch": 4.726205997392438,
|
3290 |
+
"grad_norm": 19.617931365966797,
|
3291 |
+
"learning_rate": 1.6427640156453715e-06,
|
3292 |
+
"loss": 0.3842,
|
3293 |
+
"step": 232000
|
3294 |
+
},
|
3295 |
+
{
|
3296 |
+
"epoch": 4.736391786179921,
|
3297 |
+
"grad_norm": 22.911376953125,
|
3298 |
+
"learning_rate": 1.5816492829204695e-06,
|
3299 |
+
"loss": 0.3871,
|
3300 |
+
"step": 232500
|
3301 |
+
},
|
3302 |
+
{
|
3303 |
+
"epoch": 4.7465775749674055,
|
3304 |
+
"grad_norm": 31.00206184387207,
|
3305 |
+
"learning_rate": 1.5205345501955673e-06,
|
3306 |
+
"loss": 0.3918,
|
3307 |
+
"step": 233000
|
3308 |
+
},
|
3309 |
+
{
|
3310 |
+
"epoch": 4.756763363754889,
|
3311 |
+
"grad_norm": 39.066619873046875,
|
3312 |
+
"learning_rate": 1.4594198174706648e-06,
|
3313 |
+
"loss": 0.3743,
|
3314 |
+
"step": 233500
|
3315 |
+
},
|
3316 |
+
{
|
3317 |
+
"epoch": 4.766949152542373,
|
3318 |
+
"grad_norm": 9.460115432739258,
|
3319 |
+
"learning_rate": 1.3983050847457628e-06,
|
3320 |
+
"loss": 0.388,
|
3321 |
+
"step": 234000
|
3322 |
+
},
|
3323 |
+
{
|
3324 |
+
"epoch": 4.777134941329857,
|
3325 |
+
"grad_norm": 10.770241737365723,
|
3326 |
+
"learning_rate": 1.3371903520208605e-06,
|
3327 |
+
"loss": 0.3986,
|
3328 |
+
"step": 234500
|
3329 |
+
},
|
3330 |
+
{
|
3331 |
+
"epoch": 4.78732073011734,
|
3332 |
+
"grad_norm": 29.36595916748047,
|
3333 |
+
"learning_rate": 1.2760756192959583e-06,
|
3334 |
+
"loss": 0.3836,
|
3335 |
+
"step": 235000
|
3336 |
+
},
|
3337 |
+
{
|
3338 |
+
"epoch": 4.797506518904824,
|
3339 |
+
"grad_norm": 21.498857498168945,
|
3340 |
+
"learning_rate": 1.214960886571056e-06,
|
3341 |
+
"loss": 0.3839,
|
3342 |
+
"step": 235500
|
3343 |
+
},
|
3344 |
+
{
|
3345 |
+
"epoch": 4.8076923076923075,
|
3346 |
+
"grad_norm": 2.3236119747161865,
|
3347 |
+
"learning_rate": 1.153846153846154e-06,
|
3348 |
+
"loss": 0.3983,
|
3349 |
+
"step": 236000
|
3350 |
+
},
|
3351 |
+
{
|
3352 |
+
"epoch": 4.817878096479792,
|
3353 |
+
"grad_norm": 10.451170921325684,
|
3354 |
+
"learning_rate": 1.0927314211212516e-06,
|
3355 |
+
"loss": 0.3853,
|
3356 |
+
"step": 236500
|
3357 |
+
},
|
3358 |
+
{
|
3359 |
+
"epoch": 4.828063885267275,
|
3360 |
+
"grad_norm": 18.35541343688965,
|
3361 |
+
"learning_rate": 1.0316166883963496e-06,
|
3362 |
+
"loss": 0.3779,
|
3363 |
+
"step": 237000
|
3364 |
+
},
|
3365 |
+
{
|
3366 |
+
"epoch": 4.838249674054759,
|
3367 |
+
"grad_norm": 56.08652877807617,
|
3368 |
+
"learning_rate": 9.705019556714473e-07,
|
3369 |
+
"loss": 0.3608,
|
3370 |
+
"step": 237500
|
3371 |
+
},
|
3372 |
+
{
|
3373 |
+
"epoch": 4.848435462842242,
|
3374 |
+
"grad_norm": 30.668987274169922,
|
3375 |
+
"learning_rate": 9.09387222946545e-07,
|
3376 |
+
"loss": 0.3901,
|
3377 |
+
"step": 238000
|
3378 |
+
},
|
3379 |
+
{
|
3380 |
+
"epoch": 4.858621251629726,
|
3381 |
+
"grad_norm": 44.03152084350586,
|
3382 |
+
"learning_rate": 8.482724902216428e-07,
|
3383 |
+
"loss": 0.3742,
|
3384 |
+
"step": 238500
|
3385 |
+
},
|
3386 |
+
{
|
3387 |
+
"epoch": 4.8688070404172095,
|
3388 |
+
"grad_norm": 46.421485900878906,
|
3389 |
+
"learning_rate": 7.871577574967406e-07,
|
3390 |
+
"loss": 0.4039,
|
3391 |
+
"step": 239000
|
3392 |
+
},
|
3393 |
+
{
|
3394 |
+
"epoch": 4.878992829204694,
|
3395 |
+
"grad_norm": 41.14004898071289,
|
3396 |
+
"learning_rate": 7.260430247718384e-07,
|
3397 |
+
"loss": 0.4113,
|
3398 |
+
"step": 239500
|
3399 |
+
},
|
3400 |
+
{
|
3401 |
+
"epoch": 4.889178617992178,
|
3402 |
+
"grad_norm": 3.9679319858551025,
|
3403 |
+
"learning_rate": 6.649282920469362e-07,
|
3404 |
+
"loss": 0.3856,
|
3405 |
+
"step": 240000
|
3406 |
+
},
|
3407 |
+
{
|
3408 |
+
"epoch": 4.899364406779661,
|
3409 |
+
"grad_norm": 30.101577758789062,
|
3410 |
+
"learning_rate": 6.038135593220339e-07,
|
3411 |
+
"loss": 0.3742,
|
3412 |
+
"step": 240500
|
3413 |
+
},
|
3414 |
+
{
|
3415 |
+
"epoch": 4.909550195567145,
|
3416 |
+
"grad_norm": 12.514345169067383,
|
3417 |
+
"learning_rate": 5.426988265971316e-07,
|
3418 |
+
"loss": 0.3795,
|
3419 |
+
"step": 241000
|
3420 |
+
},
|
3421 |
+
{
|
3422 |
+
"epoch": 4.919735984354628,
|
3423 |
+
"grad_norm": 33.0556526184082,
|
3424 |
+
"learning_rate": 4.815840938722295e-07,
|
3425 |
+
"loss": 0.3865,
|
3426 |
+
"step": 241500
|
3427 |
+
},
|
3428 |
+
{
|
3429 |
+
"epoch": 4.929921773142112,
|
3430 |
+
"grad_norm": 36.99669647216797,
|
3431 |
+
"learning_rate": 4.2046936114732726e-07,
|
3432 |
+
"loss": 0.4271,
|
3433 |
+
"step": 242000
|
3434 |
+
},
|
3435 |
+
{
|
3436 |
+
"epoch": 4.9401075619295955,
|
3437 |
+
"grad_norm": 2.8075144290924072,
|
3438 |
+
"learning_rate": 3.59354628422425e-07,
|
3439 |
+
"loss": 0.3867,
|
3440 |
+
"step": 242500
|
3441 |
+
},
|
3442 |
+
{
|
3443 |
+
"epoch": 4.95029335071708,
|
3444 |
+
"grad_norm": 0.9676657319068909,
|
3445 |
+
"learning_rate": 2.9823989569752284e-07,
|
3446 |
+
"loss": 0.3994,
|
3447 |
+
"step": 243000
|
3448 |
+
},
|
3449 |
+
{
|
3450 |
+
"epoch": 4.960479139504563,
|
3451 |
+
"grad_norm": 26.419252395629883,
|
3452 |
+
"learning_rate": 2.3712516297262062e-07,
|
3453 |
+
"loss": 0.3956,
|
3454 |
+
"step": 243500
|
3455 |
+
},
|
3456 |
+
{
|
3457 |
+
"epoch": 4.970664928292047,
|
3458 |
+
"grad_norm": 37.89094543457031,
|
3459 |
+
"learning_rate": 1.7601043024771838e-07,
|
3460 |
+
"loss": 0.3856,
|
3461 |
+
"step": 244000
|
3462 |
+
},
|
3463 |
+
{
|
3464 |
+
"epoch": 4.98085071707953,
|
3465 |
+
"grad_norm": 20.763473510742188,
|
3466 |
+
"learning_rate": 1.1489569752281617e-07,
|
3467 |
+
"loss": 0.4161,
|
3468 |
+
"step": 244500
|
3469 |
+
},
|
3470 |
+
{
|
3471 |
+
"epoch": 4.991036505867014,
|
3472 |
+
"grad_norm": 10.952351570129395,
|
3473 |
+
"learning_rate": 5.378096479791395e-08,
|
3474 |
+
"loss": 0.4074,
|
3475 |
+
"step": 245000
|
3476 |
+
},
|
3477 |
+
{
|
3478 |
+
"epoch": 5.0,
|
3479 |
+
"eval_accuracy": 0.7310513257980347,
|
3480 |
+
"eval_loss": 1.0102216005325317,
|
3481 |
+
"eval_runtime": 7.6909,
|
3482 |
+
"eval_samples_per_second": 638.153,
|
3483 |
+
"eval_steps_per_second": 79.834,
|
3484 |
+
"step": 245440
|
3485 |
+
},
|
3486 |
+
{
|
3487 |
+
"epoch": 5.0,
|
3488 |
+
"step": 245440,
|
3489 |
+
"total_flos": 1.158449687808768e+17,
|
3490 |
+
"train_loss": 0.5704954994082295,
|
3491 |
+
"train_runtime": 14864.7666,
|
3492 |
+
"train_samples_per_second": 132.092,
|
3493 |
+
"train_steps_per_second": 16.512
|
3494 |
+
}
|
3495 |
+
],
|
3496 |
+
"logging_steps": 500,
|
3497 |
+
"max_steps": 245440,
|
3498 |
+
"num_input_tokens_seen": 0,
|
3499 |
+
"num_train_epochs": 5,
|
3500 |
+
"save_steps": 500,
|
3501 |
+
"stateful_callbacks": {
|
3502 |
+
"EarlyStoppingCallback": {
|
3503 |
+
"args": {
|
3504 |
+
"early_stopping_patience": 3,
|
3505 |
+
"early_stopping_threshold": 0.001
|
3506 |
+
},
|
3507 |
+
"attributes": {
|
3508 |
+
"early_stopping_patience_counter": 0
|
3509 |
+
}
|
3510 |
+
},
|
3511 |
+
"TrainerControl": {
|
3512 |
+
"args": {
|
3513 |
+
"should_epoch_stop": false,
|
3514 |
+
"should_evaluate": false,
|
3515 |
+
"should_log": false,
|
3516 |
+
"should_save": true,
|
3517 |
+
"should_training_stop": true
|
3518 |
+
},
|
3519 |
+
"attributes": {}
|
3520 |
+
}
|
3521 |
+
},
|
3522 |
+
"total_flos": 1.158449687808768e+17,
|
3523 |
+
"train_batch_size": 8,
|
3524 |
+
"trial_name": null,
|
3525 |
+
"trial_params": null
|
3526 |
+
}
|
training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9d044fe66ccd6291067c67bf5a07a2d7b47163784c08e151e57205d231bd75e9
|
3 |
+
size 5368
|