rabby33 commited on
Commit
d0954d9
1 Parent(s): cfb9aa3

Training in progress, step 500

Browse files
=0.20.1 ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ usage: accelerate <command> [<args>]
2
+
3
+ positional arguments:
4
+ {config,estimate-memory,env,launch,tpu-config,test}
5
+ accelerate command helpers
6
+
7
+ options:
8
+ -h, --help show this help message and exit
added_tokens.json ADDED
@@ -0,0 +1,102 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "<extra_id_0>": 32099,
3
+ "<extra_id_10>": 32089,
4
+ "<extra_id_11>": 32088,
5
+ "<extra_id_12>": 32087,
6
+ "<extra_id_13>": 32086,
7
+ "<extra_id_14>": 32085,
8
+ "<extra_id_15>": 32084,
9
+ "<extra_id_16>": 32083,
10
+ "<extra_id_17>": 32082,
11
+ "<extra_id_18>": 32081,
12
+ "<extra_id_19>": 32080,
13
+ "<extra_id_1>": 32098,
14
+ "<extra_id_20>": 32079,
15
+ "<extra_id_21>": 32078,
16
+ "<extra_id_22>": 32077,
17
+ "<extra_id_23>": 32076,
18
+ "<extra_id_24>": 32075,
19
+ "<extra_id_25>": 32074,
20
+ "<extra_id_26>": 32073,
21
+ "<extra_id_27>": 32072,
22
+ "<extra_id_28>": 32071,
23
+ "<extra_id_29>": 32070,
24
+ "<extra_id_2>": 32097,
25
+ "<extra_id_30>": 32069,
26
+ "<extra_id_31>": 32068,
27
+ "<extra_id_32>": 32067,
28
+ "<extra_id_33>": 32066,
29
+ "<extra_id_34>": 32065,
30
+ "<extra_id_35>": 32064,
31
+ "<extra_id_36>": 32063,
32
+ "<extra_id_37>": 32062,
33
+ "<extra_id_38>": 32061,
34
+ "<extra_id_39>": 32060,
35
+ "<extra_id_3>": 32096,
36
+ "<extra_id_40>": 32059,
37
+ "<extra_id_41>": 32058,
38
+ "<extra_id_42>": 32057,
39
+ "<extra_id_43>": 32056,
40
+ "<extra_id_44>": 32055,
41
+ "<extra_id_45>": 32054,
42
+ "<extra_id_46>": 32053,
43
+ "<extra_id_47>": 32052,
44
+ "<extra_id_48>": 32051,
45
+ "<extra_id_49>": 32050,
46
+ "<extra_id_4>": 32095,
47
+ "<extra_id_50>": 32049,
48
+ "<extra_id_51>": 32048,
49
+ "<extra_id_52>": 32047,
50
+ "<extra_id_53>": 32046,
51
+ "<extra_id_54>": 32045,
52
+ "<extra_id_55>": 32044,
53
+ "<extra_id_56>": 32043,
54
+ "<extra_id_57>": 32042,
55
+ "<extra_id_58>": 32041,
56
+ "<extra_id_59>": 32040,
57
+ "<extra_id_5>": 32094,
58
+ "<extra_id_60>": 32039,
59
+ "<extra_id_61>": 32038,
60
+ "<extra_id_62>": 32037,
61
+ "<extra_id_63>": 32036,
62
+ "<extra_id_64>": 32035,
63
+ "<extra_id_65>": 32034,
64
+ "<extra_id_66>": 32033,
65
+ "<extra_id_67>": 32032,
66
+ "<extra_id_68>": 32031,
67
+ "<extra_id_69>": 32030,
68
+ "<extra_id_6>": 32093,
69
+ "<extra_id_70>": 32029,
70
+ "<extra_id_71>": 32028,
71
+ "<extra_id_72>": 32027,
72
+ "<extra_id_73>": 32026,
73
+ "<extra_id_74>": 32025,
74
+ "<extra_id_75>": 32024,
75
+ "<extra_id_76>": 32023,
76
+ "<extra_id_77>": 32022,
77
+ "<extra_id_78>": 32021,
78
+ "<extra_id_79>": 32020,
79
+ "<extra_id_7>": 32092,
80
+ "<extra_id_80>": 32019,
81
+ "<extra_id_81>": 32018,
82
+ "<extra_id_82>": 32017,
83
+ "<extra_id_83>": 32016,
84
+ "<extra_id_84>": 32015,
85
+ "<extra_id_85>": 32014,
86
+ "<extra_id_86>": 32013,
87
+ "<extra_id_87>": 32012,
88
+ "<extra_id_88>": 32011,
89
+ "<extra_id_89>": 32010,
90
+ "<extra_id_8>": 32091,
91
+ "<extra_id_90>": 32009,
92
+ "<extra_id_91>": 32008,
93
+ "<extra_id_92>": 32007,
94
+ "<extra_id_93>": 32006,
95
+ "<extra_id_94>": 32005,
96
+ "<extra_id_95>": 32004,
97
+ "<extra_id_96>": 32003,
98
+ "<extra_id_97>": 32002,
99
+ "<extra_id_98>": 32001,
100
+ "<extra_id_99>": 32000,
101
+ "<extra_id_9>": 32090
102
+ }
config.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "csebuetnlp/banglat5",
3
+ "architectures": [
4
+ "T5ForConditionalGeneration"
5
+ ],
6
+ "classifier_dropout": 0.0,
7
+ "d_ff": 2048,
8
+ "d_kv": 64,
9
+ "d_model": 768,
10
+ "decoder_start_token_id": 0,
11
+ "dense_act_fn": "gelu_new",
12
+ "dropout_rate": 0.1,
13
+ "eos_token_id": 1,
14
+ "feed_forward_proj": "gated-gelu",
15
+ "gradient_checkpointing": false,
16
+ "initializer_factor": 1.0,
17
+ "is_encoder_decoder": true,
18
+ "is_gated_act": true,
19
+ "layer_norm_epsilon": 1e-06,
20
+ "model_type": "t5",
21
+ "num_decoder_layers": 12,
22
+ "num_heads": 12,
23
+ "num_layers": 12,
24
+ "output_past": true,
25
+ "pad_token_id": 0,
26
+ "relative_attention_max_distance": 128,
27
+ "relative_attention_num_buckets": 32,
28
+ "tie_word_embeddings": false,
29
+ "torch_dtype": "float32",
30
+ "transformers_version": "4.37.0",
31
+ "use_cache": true,
32
+ "vocab_size": 32128
33
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22c2faa69222b71c5d706aa63199a52f0d45efa531170ceb094649734de12a73
3
+ size 990345064
runs/Feb26_09-20-49_d79522c38279/events.out.tfevents.1708939250.d79522c38279.34.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3a6a36f2ce3056a54342ec9c3943343682009ef7cd120bfb54e642434f64fd5
3
+ size 4700
runs/Feb26_09-21-58_d79522c38279/events.out.tfevents.1708939318.d79522c38279.34.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f2de636ae024586f465299657732496ec1f3656725651de0f587af4b6a2e792c
3
+ size 4700
runs/Feb26_09-22-08_d79522c38279/events.out.tfevents.1708939329.d79522c38279.34.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a341825a27ceefb30f5b97fa089df4032a35b566bd6189a6276ab308535e8116
3
+ size 4700
runs/Feb26_09-22-18_d79522c38279/events.out.tfevents.1708939339.d79522c38279.34.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4b8138b9a3f346796de326ee6e1f12db9f995650ba97366f36b566750b2b296
3
+ size 4700
runs/Feb26_09-22-37_d79522c38279/events.out.tfevents.1708939358.d79522c38279.34.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0954db96b63c3244fafb9fefb7188a5e95bcf68ba50c65ca656798aa783ab12d
3
+ size 4697
runs/Feb26_09-23-32_d79522c38279/events.out.tfevents.1708939413.d79522c38279.34.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:247acf5ecc438e86378de5ca0998c630004fa55de1517ae8f3e8f76b1abf742a
3
+ size 5226
special_tokens_map.json ADDED
@@ -0,0 +1,125 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<extra_id_0>",
4
+ "<extra_id_1>",
5
+ "<extra_id_2>",
6
+ "<extra_id_3>",
7
+ "<extra_id_4>",
8
+ "<extra_id_5>",
9
+ "<extra_id_6>",
10
+ "<extra_id_7>",
11
+ "<extra_id_8>",
12
+ "<extra_id_9>",
13
+ "<extra_id_10>",
14
+ "<extra_id_11>",
15
+ "<extra_id_12>",
16
+ "<extra_id_13>",
17
+ "<extra_id_14>",
18
+ "<extra_id_15>",
19
+ "<extra_id_16>",
20
+ "<extra_id_17>",
21
+ "<extra_id_18>",
22
+ "<extra_id_19>",
23
+ "<extra_id_20>",
24
+ "<extra_id_21>",
25
+ "<extra_id_22>",
26
+ "<extra_id_23>",
27
+ "<extra_id_24>",
28
+ "<extra_id_25>",
29
+ "<extra_id_26>",
30
+ "<extra_id_27>",
31
+ "<extra_id_28>",
32
+ "<extra_id_29>",
33
+ "<extra_id_30>",
34
+ "<extra_id_31>",
35
+ "<extra_id_32>",
36
+ "<extra_id_33>",
37
+ "<extra_id_34>",
38
+ "<extra_id_35>",
39
+ "<extra_id_36>",
40
+ "<extra_id_37>",
41
+ "<extra_id_38>",
42
+ "<extra_id_39>",
43
+ "<extra_id_40>",
44
+ "<extra_id_41>",
45
+ "<extra_id_42>",
46
+ "<extra_id_43>",
47
+ "<extra_id_44>",
48
+ "<extra_id_45>",
49
+ "<extra_id_46>",
50
+ "<extra_id_47>",
51
+ "<extra_id_48>",
52
+ "<extra_id_49>",
53
+ "<extra_id_50>",
54
+ "<extra_id_51>",
55
+ "<extra_id_52>",
56
+ "<extra_id_53>",
57
+ "<extra_id_54>",
58
+ "<extra_id_55>",
59
+ "<extra_id_56>",
60
+ "<extra_id_57>",
61
+ "<extra_id_58>",
62
+ "<extra_id_59>",
63
+ "<extra_id_60>",
64
+ "<extra_id_61>",
65
+ "<extra_id_62>",
66
+ "<extra_id_63>",
67
+ "<extra_id_64>",
68
+ "<extra_id_65>",
69
+ "<extra_id_66>",
70
+ "<extra_id_67>",
71
+ "<extra_id_68>",
72
+ "<extra_id_69>",
73
+ "<extra_id_70>",
74
+ "<extra_id_71>",
75
+ "<extra_id_72>",
76
+ "<extra_id_73>",
77
+ "<extra_id_74>",
78
+ "<extra_id_75>",
79
+ "<extra_id_76>",
80
+ "<extra_id_77>",
81
+ "<extra_id_78>",
82
+ "<extra_id_79>",
83
+ "<extra_id_80>",
84
+ "<extra_id_81>",
85
+ "<extra_id_82>",
86
+ "<extra_id_83>",
87
+ "<extra_id_84>",
88
+ "<extra_id_85>",
89
+ "<extra_id_86>",
90
+ "<extra_id_87>",
91
+ "<extra_id_88>",
92
+ "<extra_id_89>",
93
+ "<extra_id_90>",
94
+ "<extra_id_91>",
95
+ "<extra_id_92>",
96
+ "<extra_id_93>",
97
+ "<extra_id_94>",
98
+ "<extra_id_95>",
99
+ "<extra_id_96>",
100
+ "<extra_id_97>",
101
+ "<extra_id_98>",
102
+ "<extra_id_99>"
103
+ ],
104
+ "eos_token": {
105
+ "content": "</s>",
106
+ "lstrip": false,
107
+ "normalized": false,
108
+ "rstrip": false,
109
+ "single_word": false
110
+ },
111
+ "pad_token": {
112
+ "content": "<pad>",
113
+ "lstrip": false,
114
+ "normalized": false,
115
+ "rstrip": false,
116
+ "single_word": false
117
+ },
118
+ "unk_token": {
119
+ "content": "<unk>",
120
+ "lstrip": false,
121
+ "normalized": false,
122
+ "rstrip": false,
123
+ "single_word": false
124
+ }
125
+ }
spiece.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7dcab96935a2a51b1461c84e44c952ea8a3640c8bc3e2c6ae7a21d855454ae27
3
+ size 1111492
tokenizer_config.json ADDED
@@ -0,0 +1,939 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<pad>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "</s>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "<unk>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "32000": {
28
+ "content": "<extra_id_99>",
29
+ "lstrip": true,
30
+ "normalized": false,
31
+ "rstrip": true,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "32001": {
36
+ "content": "<extra_id_98>",
37
+ "lstrip": true,
38
+ "normalized": false,
39
+ "rstrip": true,
40
+ "single_word": false,
41
+ "special": true
42
+ },
43
+ "32002": {
44
+ "content": "<extra_id_97>",
45
+ "lstrip": true,
46
+ "normalized": false,
47
+ "rstrip": true,
48
+ "single_word": false,
49
+ "special": true
50
+ },
51
+ "32003": {
52
+ "content": "<extra_id_96>",
53
+ "lstrip": true,
54
+ "normalized": false,
55
+ "rstrip": true,
56
+ "single_word": false,
57
+ "special": true
58
+ },
59
+ "32004": {
60
+ "content": "<extra_id_95>",
61
+ "lstrip": true,
62
+ "normalized": false,
63
+ "rstrip": true,
64
+ "single_word": false,
65
+ "special": true
66
+ },
67
+ "32005": {
68
+ "content": "<extra_id_94>",
69
+ "lstrip": true,
70
+ "normalized": false,
71
+ "rstrip": true,
72
+ "single_word": false,
73
+ "special": true
74
+ },
75
+ "32006": {
76
+ "content": "<extra_id_93>",
77
+ "lstrip": true,
78
+ "normalized": false,
79
+ "rstrip": true,
80
+ "single_word": false,
81
+ "special": true
82
+ },
83
+ "32007": {
84
+ "content": "<extra_id_92>",
85
+ "lstrip": true,
86
+ "normalized": false,
87
+ "rstrip": true,
88
+ "single_word": false,
89
+ "special": true
90
+ },
91
+ "32008": {
92
+ "content": "<extra_id_91>",
93
+ "lstrip": true,
94
+ "normalized": false,
95
+ "rstrip": true,
96
+ "single_word": false,
97
+ "special": true
98
+ },
99
+ "32009": {
100
+ "content": "<extra_id_90>",
101
+ "lstrip": true,
102
+ "normalized": false,
103
+ "rstrip": true,
104
+ "single_word": false,
105
+ "special": true
106
+ },
107
+ "32010": {
108
+ "content": "<extra_id_89>",
109
+ "lstrip": true,
110
+ "normalized": false,
111
+ "rstrip": true,
112
+ "single_word": false,
113
+ "special": true
114
+ },
115
+ "32011": {
116
+ "content": "<extra_id_88>",
117
+ "lstrip": true,
118
+ "normalized": false,
119
+ "rstrip": true,
120
+ "single_word": false,
121
+ "special": true
122
+ },
123
+ "32012": {
124
+ "content": "<extra_id_87>",
125
+ "lstrip": true,
126
+ "normalized": false,
127
+ "rstrip": true,
128
+ "single_word": false,
129
+ "special": true
130
+ },
131
+ "32013": {
132
+ "content": "<extra_id_86>",
133
+ "lstrip": true,
134
+ "normalized": false,
135
+ "rstrip": true,
136
+ "single_word": false,
137
+ "special": true
138
+ },
139
+ "32014": {
140
+ "content": "<extra_id_85>",
141
+ "lstrip": true,
142
+ "normalized": false,
143
+ "rstrip": true,
144
+ "single_word": false,
145
+ "special": true
146
+ },
147
+ "32015": {
148
+ "content": "<extra_id_84>",
149
+ "lstrip": true,
150
+ "normalized": false,
151
+ "rstrip": true,
152
+ "single_word": false,
153
+ "special": true
154
+ },
155
+ "32016": {
156
+ "content": "<extra_id_83>",
157
+ "lstrip": true,
158
+ "normalized": false,
159
+ "rstrip": true,
160
+ "single_word": false,
161
+ "special": true
162
+ },
163
+ "32017": {
164
+ "content": "<extra_id_82>",
165
+ "lstrip": true,
166
+ "normalized": false,
167
+ "rstrip": true,
168
+ "single_word": false,
169
+ "special": true
170
+ },
171
+ "32018": {
172
+ "content": "<extra_id_81>",
173
+ "lstrip": true,
174
+ "normalized": false,
175
+ "rstrip": true,
176
+ "single_word": false,
177
+ "special": true
178
+ },
179
+ "32019": {
180
+ "content": "<extra_id_80>",
181
+ "lstrip": true,
182
+ "normalized": false,
183
+ "rstrip": true,
184
+ "single_word": false,
185
+ "special": true
186
+ },
187
+ "32020": {
188
+ "content": "<extra_id_79>",
189
+ "lstrip": true,
190
+ "normalized": false,
191
+ "rstrip": true,
192
+ "single_word": false,
193
+ "special": true
194
+ },
195
+ "32021": {
196
+ "content": "<extra_id_78>",
197
+ "lstrip": true,
198
+ "normalized": false,
199
+ "rstrip": true,
200
+ "single_word": false,
201
+ "special": true
202
+ },
203
+ "32022": {
204
+ "content": "<extra_id_77>",
205
+ "lstrip": true,
206
+ "normalized": false,
207
+ "rstrip": true,
208
+ "single_word": false,
209
+ "special": true
210
+ },
211
+ "32023": {
212
+ "content": "<extra_id_76>",
213
+ "lstrip": true,
214
+ "normalized": false,
215
+ "rstrip": true,
216
+ "single_word": false,
217
+ "special": true
218
+ },
219
+ "32024": {
220
+ "content": "<extra_id_75>",
221
+ "lstrip": true,
222
+ "normalized": false,
223
+ "rstrip": true,
224
+ "single_word": false,
225
+ "special": true
226
+ },
227
+ "32025": {
228
+ "content": "<extra_id_74>",
229
+ "lstrip": true,
230
+ "normalized": false,
231
+ "rstrip": true,
232
+ "single_word": false,
233
+ "special": true
234
+ },
235
+ "32026": {
236
+ "content": "<extra_id_73>",
237
+ "lstrip": true,
238
+ "normalized": false,
239
+ "rstrip": true,
240
+ "single_word": false,
241
+ "special": true
242
+ },
243
+ "32027": {
244
+ "content": "<extra_id_72>",
245
+ "lstrip": true,
246
+ "normalized": false,
247
+ "rstrip": true,
248
+ "single_word": false,
249
+ "special": true
250
+ },
251
+ "32028": {
252
+ "content": "<extra_id_71>",
253
+ "lstrip": true,
254
+ "normalized": false,
255
+ "rstrip": true,
256
+ "single_word": false,
257
+ "special": true
258
+ },
259
+ "32029": {
260
+ "content": "<extra_id_70>",
261
+ "lstrip": true,
262
+ "normalized": false,
263
+ "rstrip": true,
264
+ "single_word": false,
265
+ "special": true
266
+ },
267
+ "32030": {
268
+ "content": "<extra_id_69>",
269
+ "lstrip": true,
270
+ "normalized": false,
271
+ "rstrip": true,
272
+ "single_word": false,
273
+ "special": true
274
+ },
275
+ "32031": {
276
+ "content": "<extra_id_68>",
277
+ "lstrip": true,
278
+ "normalized": false,
279
+ "rstrip": true,
280
+ "single_word": false,
281
+ "special": true
282
+ },
283
+ "32032": {
284
+ "content": "<extra_id_67>",
285
+ "lstrip": true,
286
+ "normalized": false,
287
+ "rstrip": true,
288
+ "single_word": false,
289
+ "special": true
290
+ },
291
+ "32033": {
292
+ "content": "<extra_id_66>",
293
+ "lstrip": true,
294
+ "normalized": false,
295
+ "rstrip": true,
296
+ "single_word": false,
297
+ "special": true
298
+ },
299
+ "32034": {
300
+ "content": "<extra_id_65>",
301
+ "lstrip": true,
302
+ "normalized": false,
303
+ "rstrip": true,
304
+ "single_word": false,
305
+ "special": true
306
+ },
307
+ "32035": {
308
+ "content": "<extra_id_64>",
309
+ "lstrip": true,
310
+ "normalized": false,
311
+ "rstrip": true,
312
+ "single_word": false,
313
+ "special": true
314
+ },
315
+ "32036": {
316
+ "content": "<extra_id_63>",
317
+ "lstrip": true,
318
+ "normalized": false,
319
+ "rstrip": true,
320
+ "single_word": false,
321
+ "special": true
322
+ },
323
+ "32037": {
324
+ "content": "<extra_id_62>",
325
+ "lstrip": true,
326
+ "normalized": false,
327
+ "rstrip": true,
328
+ "single_word": false,
329
+ "special": true
330
+ },
331
+ "32038": {
332
+ "content": "<extra_id_61>",
333
+ "lstrip": true,
334
+ "normalized": false,
335
+ "rstrip": true,
336
+ "single_word": false,
337
+ "special": true
338
+ },
339
+ "32039": {
340
+ "content": "<extra_id_60>",
341
+ "lstrip": true,
342
+ "normalized": false,
343
+ "rstrip": true,
344
+ "single_word": false,
345
+ "special": true
346
+ },
347
+ "32040": {
348
+ "content": "<extra_id_59>",
349
+ "lstrip": true,
350
+ "normalized": false,
351
+ "rstrip": true,
352
+ "single_word": false,
353
+ "special": true
354
+ },
355
+ "32041": {
356
+ "content": "<extra_id_58>",
357
+ "lstrip": true,
358
+ "normalized": false,
359
+ "rstrip": true,
360
+ "single_word": false,
361
+ "special": true
362
+ },
363
+ "32042": {
364
+ "content": "<extra_id_57>",
365
+ "lstrip": true,
366
+ "normalized": false,
367
+ "rstrip": true,
368
+ "single_word": false,
369
+ "special": true
370
+ },
371
+ "32043": {
372
+ "content": "<extra_id_56>",
373
+ "lstrip": true,
374
+ "normalized": false,
375
+ "rstrip": true,
376
+ "single_word": false,
377
+ "special": true
378
+ },
379
+ "32044": {
380
+ "content": "<extra_id_55>",
381
+ "lstrip": true,
382
+ "normalized": false,
383
+ "rstrip": true,
384
+ "single_word": false,
385
+ "special": true
386
+ },
387
+ "32045": {
388
+ "content": "<extra_id_54>",
389
+ "lstrip": true,
390
+ "normalized": false,
391
+ "rstrip": true,
392
+ "single_word": false,
393
+ "special": true
394
+ },
395
+ "32046": {
396
+ "content": "<extra_id_53>",
397
+ "lstrip": true,
398
+ "normalized": false,
399
+ "rstrip": true,
400
+ "single_word": false,
401
+ "special": true
402
+ },
403
+ "32047": {
404
+ "content": "<extra_id_52>",
405
+ "lstrip": true,
406
+ "normalized": false,
407
+ "rstrip": true,
408
+ "single_word": false,
409
+ "special": true
410
+ },
411
+ "32048": {
412
+ "content": "<extra_id_51>",
413
+ "lstrip": true,
414
+ "normalized": false,
415
+ "rstrip": true,
416
+ "single_word": false,
417
+ "special": true
418
+ },
419
+ "32049": {
420
+ "content": "<extra_id_50>",
421
+ "lstrip": true,
422
+ "normalized": false,
423
+ "rstrip": true,
424
+ "single_word": false,
425
+ "special": true
426
+ },
427
+ "32050": {
428
+ "content": "<extra_id_49>",
429
+ "lstrip": true,
430
+ "normalized": false,
431
+ "rstrip": true,
432
+ "single_word": false,
433
+ "special": true
434
+ },
435
+ "32051": {
436
+ "content": "<extra_id_48>",
437
+ "lstrip": true,
438
+ "normalized": false,
439
+ "rstrip": true,
440
+ "single_word": false,
441
+ "special": true
442
+ },
443
+ "32052": {
444
+ "content": "<extra_id_47>",
445
+ "lstrip": true,
446
+ "normalized": false,
447
+ "rstrip": true,
448
+ "single_word": false,
449
+ "special": true
450
+ },
451
+ "32053": {
452
+ "content": "<extra_id_46>",
453
+ "lstrip": true,
454
+ "normalized": false,
455
+ "rstrip": true,
456
+ "single_word": false,
457
+ "special": true
458
+ },
459
+ "32054": {
460
+ "content": "<extra_id_45>",
461
+ "lstrip": true,
462
+ "normalized": false,
463
+ "rstrip": true,
464
+ "single_word": false,
465
+ "special": true
466
+ },
467
+ "32055": {
468
+ "content": "<extra_id_44>",
469
+ "lstrip": true,
470
+ "normalized": false,
471
+ "rstrip": true,
472
+ "single_word": false,
473
+ "special": true
474
+ },
475
+ "32056": {
476
+ "content": "<extra_id_43>",
477
+ "lstrip": true,
478
+ "normalized": false,
479
+ "rstrip": true,
480
+ "single_word": false,
481
+ "special": true
482
+ },
483
+ "32057": {
484
+ "content": "<extra_id_42>",
485
+ "lstrip": true,
486
+ "normalized": false,
487
+ "rstrip": true,
488
+ "single_word": false,
489
+ "special": true
490
+ },
491
+ "32058": {
492
+ "content": "<extra_id_41>",
493
+ "lstrip": true,
494
+ "normalized": false,
495
+ "rstrip": true,
496
+ "single_word": false,
497
+ "special": true
498
+ },
499
+ "32059": {
500
+ "content": "<extra_id_40>",
501
+ "lstrip": true,
502
+ "normalized": false,
503
+ "rstrip": true,
504
+ "single_word": false,
505
+ "special": true
506
+ },
507
+ "32060": {
508
+ "content": "<extra_id_39>",
509
+ "lstrip": true,
510
+ "normalized": false,
511
+ "rstrip": true,
512
+ "single_word": false,
513
+ "special": true
514
+ },
515
+ "32061": {
516
+ "content": "<extra_id_38>",
517
+ "lstrip": true,
518
+ "normalized": false,
519
+ "rstrip": true,
520
+ "single_word": false,
521
+ "special": true
522
+ },
523
+ "32062": {
524
+ "content": "<extra_id_37>",
525
+ "lstrip": true,
526
+ "normalized": false,
527
+ "rstrip": true,
528
+ "single_word": false,
529
+ "special": true
530
+ },
531
+ "32063": {
532
+ "content": "<extra_id_36>",
533
+ "lstrip": true,
534
+ "normalized": false,
535
+ "rstrip": true,
536
+ "single_word": false,
537
+ "special": true
538
+ },
539
+ "32064": {
540
+ "content": "<extra_id_35>",
541
+ "lstrip": true,
542
+ "normalized": false,
543
+ "rstrip": true,
544
+ "single_word": false,
545
+ "special": true
546
+ },
547
+ "32065": {
548
+ "content": "<extra_id_34>",
549
+ "lstrip": true,
550
+ "normalized": false,
551
+ "rstrip": true,
552
+ "single_word": false,
553
+ "special": true
554
+ },
555
+ "32066": {
556
+ "content": "<extra_id_33>",
557
+ "lstrip": true,
558
+ "normalized": false,
559
+ "rstrip": true,
560
+ "single_word": false,
561
+ "special": true
562
+ },
563
+ "32067": {
564
+ "content": "<extra_id_32>",
565
+ "lstrip": true,
566
+ "normalized": false,
567
+ "rstrip": true,
568
+ "single_word": false,
569
+ "special": true
570
+ },
571
+ "32068": {
572
+ "content": "<extra_id_31>",
573
+ "lstrip": true,
574
+ "normalized": false,
575
+ "rstrip": true,
576
+ "single_word": false,
577
+ "special": true
578
+ },
579
+ "32069": {
580
+ "content": "<extra_id_30>",
581
+ "lstrip": true,
582
+ "normalized": false,
583
+ "rstrip": true,
584
+ "single_word": false,
585
+ "special": true
586
+ },
587
+ "32070": {
588
+ "content": "<extra_id_29>",
589
+ "lstrip": true,
590
+ "normalized": false,
591
+ "rstrip": true,
592
+ "single_word": false,
593
+ "special": true
594
+ },
595
+ "32071": {
596
+ "content": "<extra_id_28>",
597
+ "lstrip": true,
598
+ "normalized": false,
599
+ "rstrip": true,
600
+ "single_word": false,
601
+ "special": true
602
+ },
603
+ "32072": {
604
+ "content": "<extra_id_27>",
605
+ "lstrip": true,
606
+ "normalized": false,
607
+ "rstrip": true,
608
+ "single_word": false,
609
+ "special": true
610
+ },
611
+ "32073": {
612
+ "content": "<extra_id_26>",
613
+ "lstrip": true,
614
+ "normalized": false,
615
+ "rstrip": true,
616
+ "single_word": false,
617
+ "special": true
618
+ },
619
+ "32074": {
620
+ "content": "<extra_id_25>",
621
+ "lstrip": true,
622
+ "normalized": false,
623
+ "rstrip": true,
624
+ "single_word": false,
625
+ "special": true
626
+ },
627
+ "32075": {
628
+ "content": "<extra_id_24>",
629
+ "lstrip": true,
630
+ "normalized": false,
631
+ "rstrip": true,
632
+ "single_word": false,
633
+ "special": true
634
+ },
635
+ "32076": {
636
+ "content": "<extra_id_23>",
637
+ "lstrip": true,
638
+ "normalized": false,
639
+ "rstrip": true,
640
+ "single_word": false,
641
+ "special": true
642
+ },
643
+ "32077": {
644
+ "content": "<extra_id_22>",
645
+ "lstrip": true,
646
+ "normalized": false,
647
+ "rstrip": true,
648
+ "single_word": false,
649
+ "special": true
650
+ },
651
+ "32078": {
652
+ "content": "<extra_id_21>",
653
+ "lstrip": true,
654
+ "normalized": false,
655
+ "rstrip": true,
656
+ "single_word": false,
657
+ "special": true
658
+ },
659
+ "32079": {
660
+ "content": "<extra_id_20>",
661
+ "lstrip": true,
662
+ "normalized": false,
663
+ "rstrip": true,
664
+ "single_word": false,
665
+ "special": true
666
+ },
667
+ "32080": {
668
+ "content": "<extra_id_19>",
669
+ "lstrip": true,
670
+ "normalized": false,
671
+ "rstrip": true,
672
+ "single_word": false,
673
+ "special": true
674
+ },
675
+ "32081": {
676
+ "content": "<extra_id_18>",
677
+ "lstrip": true,
678
+ "normalized": false,
679
+ "rstrip": true,
680
+ "single_word": false,
681
+ "special": true
682
+ },
683
+ "32082": {
684
+ "content": "<extra_id_17>",
685
+ "lstrip": true,
686
+ "normalized": false,
687
+ "rstrip": true,
688
+ "single_word": false,
689
+ "special": true
690
+ },
691
+ "32083": {
692
+ "content": "<extra_id_16>",
693
+ "lstrip": true,
694
+ "normalized": false,
695
+ "rstrip": true,
696
+ "single_word": false,
697
+ "special": true
698
+ },
699
+ "32084": {
700
+ "content": "<extra_id_15>",
701
+ "lstrip": true,
702
+ "normalized": false,
703
+ "rstrip": true,
704
+ "single_word": false,
705
+ "special": true
706
+ },
707
+ "32085": {
708
+ "content": "<extra_id_14>",
709
+ "lstrip": true,
710
+ "normalized": false,
711
+ "rstrip": true,
712
+ "single_word": false,
713
+ "special": true
714
+ },
715
+ "32086": {
716
+ "content": "<extra_id_13>",
717
+ "lstrip": true,
718
+ "normalized": false,
719
+ "rstrip": true,
720
+ "single_word": false,
721
+ "special": true
722
+ },
723
+ "32087": {
724
+ "content": "<extra_id_12>",
725
+ "lstrip": true,
726
+ "normalized": false,
727
+ "rstrip": true,
728
+ "single_word": false,
729
+ "special": true
730
+ },
731
+ "32088": {
732
+ "content": "<extra_id_11>",
733
+ "lstrip": true,
734
+ "normalized": false,
735
+ "rstrip": true,
736
+ "single_word": false,
737
+ "special": true
738
+ },
739
+ "32089": {
740
+ "content": "<extra_id_10>",
741
+ "lstrip": true,
742
+ "normalized": false,
743
+ "rstrip": true,
744
+ "single_word": false,
745
+ "special": true
746
+ },
747
+ "32090": {
748
+ "content": "<extra_id_9>",
749
+ "lstrip": true,
750
+ "normalized": false,
751
+ "rstrip": true,
752
+ "single_word": false,
753
+ "special": true
754
+ },
755
+ "32091": {
756
+ "content": "<extra_id_8>",
757
+ "lstrip": true,
758
+ "normalized": false,
759
+ "rstrip": true,
760
+ "single_word": false,
761
+ "special": true
762
+ },
763
+ "32092": {
764
+ "content": "<extra_id_7>",
765
+ "lstrip": true,
766
+ "normalized": false,
767
+ "rstrip": true,
768
+ "single_word": false,
769
+ "special": true
770
+ },
771
+ "32093": {
772
+ "content": "<extra_id_6>",
773
+ "lstrip": true,
774
+ "normalized": false,
775
+ "rstrip": true,
776
+ "single_word": false,
777
+ "special": true
778
+ },
779
+ "32094": {
780
+ "content": "<extra_id_5>",
781
+ "lstrip": true,
782
+ "normalized": false,
783
+ "rstrip": true,
784
+ "single_word": false,
785
+ "special": true
786
+ },
787
+ "32095": {
788
+ "content": "<extra_id_4>",
789
+ "lstrip": true,
790
+ "normalized": false,
791
+ "rstrip": true,
792
+ "single_word": false,
793
+ "special": true
794
+ },
795
+ "32096": {
796
+ "content": "<extra_id_3>",
797
+ "lstrip": true,
798
+ "normalized": false,
799
+ "rstrip": true,
800
+ "single_word": false,
801
+ "special": true
802
+ },
803
+ "32097": {
804
+ "content": "<extra_id_2>",
805
+ "lstrip": true,
806
+ "normalized": false,
807
+ "rstrip": true,
808
+ "single_word": false,
809
+ "special": true
810
+ },
811
+ "32098": {
812
+ "content": "<extra_id_1>",
813
+ "lstrip": true,
814
+ "normalized": false,
815
+ "rstrip": true,
816
+ "single_word": false,
817
+ "special": true
818
+ },
819
+ "32099": {
820
+ "content": "<extra_id_0>",
821
+ "lstrip": true,
822
+ "normalized": false,
823
+ "rstrip": true,
824
+ "single_word": false,
825
+ "special": true
826
+ }
827
+ },
828
+ "additional_special_tokens": [
829
+ "<extra_id_0>",
830
+ "<extra_id_1>",
831
+ "<extra_id_2>",
832
+ "<extra_id_3>",
833
+ "<extra_id_4>",
834
+ "<extra_id_5>",
835
+ "<extra_id_6>",
836
+ "<extra_id_7>",
837
+ "<extra_id_8>",
838
+ "<extra_id_9>",
839
+ "<extra_id_10>",
840
+ "<extra_id_11>",
841
+ "<extra_id_12>",
842
+ "<extra_id_13>",
843
+ "<extra_id_14>",
844
+ "<extra_id_15>",
845
+ "<extra_id_16>",
846
+ "<extra_id_17>",
847
+ "<extra_id_18>",
848
+ "<extra_id_19>",
849
+ "<extra_id_20>",
850
+ "<extra_id_21>",
851
+ "<extra_id_22>",
852
+ "<extra_id_23>",
853
+ "<extra_id_24>",
854
+ "<extra_id_25>",
855
+ "<extra_id_26>",
856
+ "<extra_id_27>",
857
+ "<extra_id_28>",
858
+ "<extra_id_29>",
859
+ "<extra_id_30>",
860
+ "<extra_id_31>",
861
+ "<extra_id_32>",
862
+ "<extra_id_33>",
863
+ "<extra_id_34>",
864
+ "<extra_id_35>",
865
+ "<extra_id_36>",
866
+ "<extra_id_37>",
867
+ "<extra_id_38>",
868
+ "<extra_id_39>",
869
+ "<extra_id_40>",
870
+ "<extra_id_41>",
871
+ "<extra_id_42>",
872
+ "<extra_id_43>",
873
+ "<extra_id_44>",
874
+ "<extra_id_45>",
875
+ "<extra_id_46>",
876
+ "<extra_id_47>",
877
+ "<extra_id_48>",
878
+ "<extra_id_49>",
879
+ "<extra_id_50>",
880
+ "<extra_id_51>",
881
+ "<extra_id_52>",
882
+ "<extra_id_53>",
883
+ "<extra_id_54>",
884
+ "<extra_id_55>",
885
+ "<extra_id_56>",
886
+ "<extra_id_57>",
887
+ "<extra_id_58>",
888
+ "<extra_id_59>",
889
+ "<extra_id_60>",
890
+ "<extra_id_61>",
891
+ "<extra_id_62>",
892
+ "<extra_id_63>",
893
+ "<extra_id_64>",
894
+ "<extra_id_65>",
895
+ "<extra_id_66>",
896
+ "<extra_id_67>",
897
+ "<extra_id_68>",
898
+ "<extra_id_69>",
899
+ "<extra_id_70>",
900
+ "<extra_id_71>",
901
+ "<extra_id_72>",
902
+ "<extra_id_73>",
903
+ "<extra_id_74>",
904
+ "<extra_id_75>",
905
+ "<extra_id_76>",
906
+ "<extra_id_77>",
907
+ "<extra_id_78>",
908
+ "<extra_id_79>",
909
+ "<extra_id_80>",
910
+ "<extra_id_81>",
911
+ "<extra_id_82>",
912
+ "<extra_id_83>",
913
+ "<extra_id_84>",
914
+ "<extra_id_85>",
915
+ "<extra_id_86>",
916
+ "<extra_id_87>",
917
+ "<extra_id_88>",
918
+ "<extra_id_89>",
919
+ "<extra_id_90>",
920
+ "<extra_id_91>",
921
+ "<extra_id_92>",
922
+ "<extra_id_93>",
923
+ "<extra_id_94>",
924
+ "<extra_id_95>",
925
+ "<extra_id_96>",
926
+ "<extra_id_97>",
927
+ "<extra_id_98>",
928
+ "<extra_id_99>"
929
+ ],
930
+ "clean_up_tokenization_spaces": true,
931
+ "eos_token": "</s>",
932
+ "extra_ids": 100,
933
+ "legacy": true,
934
+ "model_max_length": 512,
935
+ "pad_token": "<pad>",
936
+ "sp_model_kwargs": {},
937
+ "tokenizer_class": "T5Tokenizer",
938
+ "unk_token": "<unk>"
939
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e8a8b8d3a4e565f3964577c0900c811dc4081b3e725d6f2210d0f1c988da9c0
3
+ size 4411
wandb/debug-internal.log ADDED
@@ -0,0 +1,327 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-02-26 09:21:05,531 INFO StreamThr :196 [internal.py:wandb_internal():86] W&B internal server running at pid: 196, started at: 2024-02-26 09:21:05.531199
2
+ 2024-02-26 09:21:05,533 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status
3
+ 2024-02-26 09:21:06,208 INFO WriterThread:196 [datastore.py:open_for_write():87] open: /kaggle/working/wandb/run-20240226_092105-91uhs91a/run-91uhs91a.wandb
4
+ 2024-02-26 09:21:06,208 DEBUG SenderThread:196 [sender.py:send():382] send: header
5
+ 2024-02-26 09:21:06,211 DEBUG SenderThread:196 [sender.py:send():382] send: run
6
+ 2024-02-26 09:21:06,389 INFO SenderThread:196 [dir_watcher.py:__init__():211] watching files in: /kaggle/working/wandb/run-20240226_092105-91uhs91a/files
7
+ 2024-02-26 09:21:06,389 INFO SenderThread:196 [sender.py:_start_run_threads():1136] run started: 91uhs91a with start time 1708939265.532258
8
+ 2024-02-26 09:21:06,397 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: check_version
9
+ 2024-02-26 09:21:06,397 DEBUG SenderThread:196 [sender.py:send_request():409] send_request: check_version
10
+ 2024-02-26 09:21:06,464 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: run_start
11
+ 2024-02-26 09:21:06,475 DEBUG HandlerThread:196 [system_info.py:__init__():27] System info init
12
+ 2024-02-26 09:21:06,475 DEBUG HandlerThread:196 [system_info.py:__init__():42] System info init done
13
+ 2024-02-26 09:21:06,475 INFO HandlerThread:196 [system_monitor.py:start():194] Starting system monitor
14
+ 2024-02-26 09:21:06,475 INFO SystemMonitor:196 [system_monitor.py:_start():158] Starting system asset monitoring threads
15
+ 2024-02-26 09:21:06,475 INFO HandlerThread:196 [system_monitor.py:probe():214] Collecting system info
16
+ 2024-02-26 09:21:06,476 INFO SystemMonitor:196 [interfaces.py:start():190] Started cpu monitoring
17
+ 2024-02-26 09:21:06,476 INFO SystemMonitor:196 [interfaces.py:start():190] Started disk monitoring
18
+ 2024-02-26 09:21:06,477 INFO SystemMonitor:196 [interfaces.py:start():190] Started gpu monitoring
19
+ 2024-02-26 09:21:06,477 INFO SystemMonitor:196 [interfaces.py:start():190] Started memory monitoring
20
+ 2024-02-26 09:21:06,478 INFO SystemMonitor:196 [interfaces.py:start():190] Started network monitoring
21
+ 2024-02-26 09:21:06,484 DEBUG HandlerThread:196 [system_info.py:probe():151] Probing system
22
+ 2024-02-26 09:21:06,486 DEBUG HandlerThread:196 [gitlib.py:_init_repo():56] git repository is invalid
23
+ 2024-02-26 09:21:06,486 DEBUG HandlerThread:196 [system_info.py:probe():199] Probing system done
24
+ 2024-02-26 09:21:06,487 DEBUG HandlerThread:196 [system_monitor.py:probe():223] {'os': 'Linux-5.15.133+-x86_64-with-glibc2.31', 'python': '3.10.13', 'heartbeatAt': '2024-02-26T09:21:06.484258', 'startedAt': '2024-02-26T09:21:05.525040', 'docker': None, 'cuda': None, 'args': (), 'state': 'running', 'program': 'kaggle.ipynb', 'codePathLocal': None, 'root': '/kaggle/working', 'host': 'd79522c38279', 'username': 'root', 'executable': '/opt/conda/bin/python3.10', 'cpu_count': 2, 'cpu_count_logical': 4, 'cpu_freq': {'current': 2000.186, 'min': 0.0, 'max': 0.0}, 'cpu_freq_per_core': [{'current': 2000.186, 'min': 0.0, 'max': 0.0}, {'current': 2000.186, 'min': 0.0, 'max': 0.0}, {'current': 2000.186, 'min': 0.0, 'max': 0.0}, {'current': 2000.186, 'min': 0.0, 'max': 0.0}], 'disk': {'/': {'total': 8062.387607574463, 'used': 5437.818241119385}}, 'gpu': 'Tesla P100-PCIE-16GB', 'gpu_count': 1, 'gpu_devices': [{'name': 'Tesla P100-PCIE-16GB', 'memory_total': 17179869184}], 'memory': {'total': 31.357559204101562}}
25
+ 2024-02-26 09:21:06,487 INFO HandlerThread:196 [system_monitor.py:probe():224] Finished collecting system info
26
+ 2024-02-26 09:21:06,487 INFO HandlerThread:196 [system_monitor.py:probe():227] Publishing system info
27
+ 2024-02-26 09:21:06,487 DEBUG HandlerThread:196 [system_info.py:_save_conda():208] Saving list of conda packages installed into the current environment
28
+ 2024-02-26 09:21:07,391 INFO Thread-12 :196 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240226_092105-91uhs91a/files/conda-environment.yaml
29
+ 2024-02-26 09:21:58,406 INFO Thread-12 :196 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240226_092105-91uhs91a/files/conda-environment.yaml
30
+ 2024-02-26 09:21:58,864 DEBUG HandlerThread:196 [system_info.py:_save_conda():220] Saving conda packages done
31
+ 2024-02-26 09:21:58,865 INFO HandlerThread:196 [system_monitor.py:probe():229] Finished publishing system info
32
+ 2024-02-26 09:21:58,872 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
33
+ 2024-02-26 09:21:58,872 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
34
+ 2024-02-26 09:21:58,873 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
35
+ 2024-02-26 09:21:58,873 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
36
+ 2024-02-26 09:21:58,873 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
37
+ 2024-02-26 09:21:58,873 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
38
+ 2024-02-26 09:21:58,873 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
39
+ 2024-02-26 09:21:58,873 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
40
+ 2024-02-26 09:21:58,873 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
41
+ 2024-02-26 09:21:58,873 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
42
+ 2024-02-26 09:21:58,873 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
43
+ 2024-02-26 09:21:58,873 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: python_packages
44
+ 2024-02-26 09:21:58,873 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: internal_messages
45
+ 2024-02-26 09:21:58,873 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: stop_status
46
+ 2024-02-26 09:21:58,874 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: pause
47
+ 2024-02-26 09:21:58,874 INFO HandlerThread:196 [handler.py:handle_request_pause():708] stopping system metrics thread
48
+ 2024-02-26 09:21:58,874 INFO HandlerThread:196 [system_monitor.py:finish():203] Stopping system monitor
49
+ 2024-02-26 09:21:58,875 DEBUG SystemMonitor:196 [system_monitor.py:_start():172] Starting system metrics aggregation loop
50
+ 2024-02-26 09:21:58,875 DEBUG SystemMonitor:196 [system_monitor.py:_start():179] Finished system metrics aggregation loop
51
+ 2024-02-26 09:21:58,875 DEBUG SystemMonitor:196 [system_monitor.py:_start():183] Publishing last batch of metrics
52
+ 2024-02-26 09:21:58,876 DEBUG SenderThread:196 [sender.py:send_request():409] send_request: python_packages
53
+ 2024-02-26 09:21:58,878 INFO HandlerThread:196 [interfaces.py:finish():202] Joined cpu monitor
54
+ 2024-02-26 09:21:58,879 INFO HandlerThread:196 [interfaces.py:finish():202] Joined disk monitor
55
+ 2024-02-26 09:21:58,879 DEBUG SenderThread:196 [sender.py:send():382] send: telemetry
56
+ 2024-02-26 09:21:58,879 DEBUG SenderThread:196 [sender.py:send():382] send: config
57
+ 2024-02-26 09:21:58,881 DEBUG SenderThread:196 [sender.py:send():382] send: metric
58
+ 2024-02-26 09:21:58,881 DEBUG SenderThread:196 [sender.py:send():382] send: telemetry
59
+ 2024-02-26 09:21:58,881 DEBUG SenderThread:196 [sender.py:send():382] send: metric
60
+ 2024-02-26 09:21:58,881 WARNING SenderThread:196 [sender.py:send_metric():1354] Seen metric with glob (shouldn't happen)
61
+ 2024-02-26 09:21:58,881 DEBUG SenderThread:196 [sender.py:send():382] send: telemetry
62
+ 2024-02-26 09:21:58,882 DEBUG SenderThread:196 [sender.py:send_request():409] send_request: stop_status
63
+ 2024-02-26 09:21:58,891 INFO HandlerThread:196 [interfaces.py:finish():202] Joined gpu monitor
64
+ 2024-02-26 09:21:58,892 INFO HandlerThread:196 [interfaces.py:finish():202] Joined memory monitor
65
+ 2024-02-26 09:21:58,892 INFO HandlerThread:196 [interfaces.py:finish():202] Joined network monitor
66
+ 2024-02-26 09:21:58,892 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
67
+ 2024-02-26 09:21:58,892 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
68
+ 2024-02-26 09:21:58,892 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
69
+ 2024-02-26 09:21:58,892 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
70
+ 2024-02-26 09:21:58,892 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
71
+ 2024-02-26 09:21:58,892 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
72
+ 2024-02-26 09:21:58,893 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
73
+ 2024-02-26 09:21:58,893 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
74
+ 2024-02-26 09:21:58,893 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: resume
75
+ 2024-02-26 09:21:58,893 INFO HandlerThread:196 [handler.py:handle_request_resume():699] starting system metrics thread
76
+ 2024-02-26 09:21:58,893 INFO HandlerThread:196 [system_monitor.py:start():194] Starting system monitor
77
+ 2024-02-26 09:21:58,893 INFO SystemMonitor:196 [system_monitor.py:_start():158] Starting system asset monitoring threads
78
+ 2024-02-26 09:21:58,893 INFO SystemMonitor:196 [interfaces.py:start():190] Started cpu monitoring
79
+ 2024-02-26 09:21:58,894 INFO SystemMonitor:196 [interfaces.py:start():190] Started disk monitoring
80
+ 2024-02-26 09:21:58,895 INFO SystemMonitor:196 [interfaces.py:start():190] Started gpu monitoring
81
+ 2024-02-26 09:21:58,896 INFO SystemMonitor:196 [interfaces.py:start():190] Started memory monitoring
82
+ 2024-02-26 09:21:58,897 INFO SystemMonitor:196 [interfaces.py:start():190] Started network monitoring
83
+ 2024-02-26 09:21:59,035 DEBUG SenderThread:196 [sender.py:send():382] send: config
84
+ 2024-02-26 09:21:59,037 DEBUG SenderThread:196 [sender.py:send():382] send: metric
85
+ 2024-02-26 09:21:59,037 DEBUG SenderThread:196 [sender.py:send():382] send: metric
86
+ 2024-02-26 09:21:59,037 WARNING SenderThread:196 [sender.py:send_metric():1354] Seen metric with glob (shouldn't happen)
87
+ 2024-02-26 09:21:59,037 DEBUG SenderThread:196 [sender.py:send():382] send: files
88
+ 2024-02-26 09:21:59,038 INFO SenderThread:196 [sender.py:_save_file():1403] saving file wandb-metadata.json with policy now
89
+ 2024-02-26 09:21:59,038 DEBUG SenderThread:196 [sender.py:send():382] send: stats
90
+ 2024-02-26 09:21:59,335 INFO wandb-upload_0:196 [upload_job.py:push():131] Uploaded file /tmp/tmp5wprt0czwandb/qmdsgk7x-wandb-metadata.json
91
+ 2024-02-26 09:21:59,406 INFO Thread-12 :196 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240226_092105-91uhs91a/files/wandb-metadata.json
92
+ 2024-02-26 09:21:59,406 INFO Thread-12 :196 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240226_092105-91uhs91a/files/requirements.txt
93
+ 2024-02-26 09:21:59,602 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: pause
94
+ 2024-02-26 09:21:59,602 INFO HandlerThread:196 [handler.py:handle_request_pause():708] stopping system metrics thread
95
+ 2024-02-26 09:21:59,602 INFO HandlerThread:196 [system_monitor.py:finish():203] Stopping system monitor
96
+ 2024-02-26 09:21:59,603 INFO HandlerThread:196 [interfaces.py:finish():202] Joined cpu monitor
97
+ 2024-02-26 09:21:59,603 INFO HandlerThread:196 [interfaces.py:finish():202] Joined disk monitor
98
+ 2024-02-26 09:21:59,603 DEBUG SystemMonitor:196 [system_monitor.py:_start():172] Starting system metrics aggregation loop
99
+ 2024-02-26 09:21:59,604 DEBUG SystemMonitor:196 [system_monitor.py:_start():179] Finished system metrics aggregation loop
100
+ 2024-02-26 09:21:59,604 DEBUG SystemMonitor:196 [system_monitor.py:_start():183] Publishing last batch of metrics
101
+ 2024-02-26 09:21:59,611 INFO HandlerThread:196 [interfaces.py:finish():202] Joined gpu monitor
102
+ 2024-02-26 09:21:59,611 INFO HandlerThread:196 [interfaces.py:finish():202] Joined memory monitor
103
+ 2024-02-26 09:21:59,611 INFO HandlerThread:196 [interfaces.py:finish():202] Joined network monitor
104
+ 2024-02-26 09:21:59,611 DEBUG SenderThread:196 [sender.py:send():382] send: stats
105
+ 2024-02-26 09:22:01,613 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
106
+ 2024-02-26 09:22:01,749 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
107
+ 2024-02-26 09:22:06,619 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
108
+ 2024-02-26 09:22:06,755 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
109
+ 2024-02-26 09:22:07,409 INFO Thread-12 :196 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240226_092105-91uhs91a/files/config.yaml
110
+ 2024-02-26 09:22:08,331 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: resume
111
+ 2024-02-26 09:22:08,331 INFO HandlerThread:196 [handler.py:handle_request_resume():699] starting system metrics thread
112
+ 2024-02-26 09:22:08,331 INFO HandlerThread:196 [system_monitor.py:start():194] Starting system monitor
113
+ 2024-02-26 09:22:08,331 INFO SystemMonitor:196 [system_monitor.py:_start():158] Starting system asset monitoring threads
114
+ 2024-02-26 09:22:08,332 INFO SystemMonitor:196 [interfaces.py:start():190] Started cpu monitoring
115
+ 2024-02-26 09:22:08,332 INFO SystemMonitor:196 [interfaces.py:start():190] Started disk monitoring
116
+ 2024-02-26 09:22:08,335 INFO SystemMonitor:196 [interfaces.py:start():190] Started gpu monitoring
117
+ 2024-02-26 09:22:08,335 INFO SystemMonitor:196 [interfaces.py:start():190] Started memory monitoring
118
+ 2024-02-26 09:22:08,336 INFO SystemMonitor:196 [interfaces.py:start():190] Started network monitoring
119
+ 2024-02-26 09:22:09,404 DEBUG SenderThread:196 [sender.py:send():382] send: config
120
+ 2024-02-26 09:22:09,406 DEBUG SenderThread:196 [sender.py:send():382] send: metric
121
+ 2024-02-26 09:22:09,406 DEBUG SenderThread:196 [sender.py:send():382] send: metric
122
+ 2024-02-26 09:22:09,406 WARNING SenderThread:196 [sender.py:send_metric():1354] Seen metric with glob (shouldn't happen)
123
+ 2024-02-26 09:22:09,927 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: pause
124
+ 2024-02-26 09:22:09,927 INFO HandlerThread:196 [handler.py:handle_request_pause():708] stopping system metrics thread
125
+ 2024-02-26 09:22:09,927 INFO HandlerThread:196 [system_monitor.py:finish():203] Stopping system monitor
126
+ 2024-02-26 09:22:09,927 INFO HandlerThread:196 [interfaces.py:finish():202] Joined cpu monitor
127
+ 2024-02-26 09:22:09,928 INFO HandlerThread:196 [interfaces.py:finish():202] Joined disk monitor
128
+ 2024-02-26 09:22:09,928 DEBUG SystemMonitor:196 [system_monitor.py:_start():172] Starting system metrics aggregation loop
129
+ 2024-02-26 09:22:09,928 DEBUG SystemMonitor:196 [system_monitor.py:_start():179] Finished system metrics aggregation loop
130
+ 2024-02-26 09:22:09,928 DEBUG SystemMonitor:196 [system_monitor.py:_start():183] Publishing last batch of metrics
131
+ 2024-02-26 09:22:09,934 INFO HandlerThread:196 [interfaces.py:finish():202] Joined gpu monitor
132
+ 2024-02-26 09:22:09,934 INFO HandlerThread:196 [interfaces.py:finish():202] Joined memory monitor
133
+ 2024-02-26 09:22:09,934 INFO HandlerThread:196 [interfaces.py:finish():202] Joined network monitor
134
+ 2024-02-26 09:22:09,935 DEBUG SenderThread:196 [sender.py:send():382] send: stats
135
+ 2024-02-26 09:22:11,810 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
136
+ 2024-02-26 09:22:11,936 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
137
+ 2024-02-26 09:22:16,811 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
138
+ 2024-02-26 09:22:16,936 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
139
+ 2024-02-26 09:22:18,738 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: resume
140
+ 2024-02-26 09:22:18,738 INFO HandlerThread:196 [handler.py:handle_request_resume():699] starting system metrics thread
141
+ 2024-02-26 09:22:18,738 INFO HandlerThread:196 [system_monitor.py:start():194] Starting system monitor
142
+ 2024-02-26 09:22:18,738 INFO SystemMonitor:196 [system_monitor.py:_start():158] Starting system asset monitoring threads
143
+ 2024-02-26 09:22:18,739 INFO SystemMonitor:196 [interfaces.py:start():190] Started cpu monitoring
144
+ 2024-02-26 09:22:18,740 INFO SystemMonitor:196 [interfaces.py:start():190] Started disk monitoring
145
+ 2024-02-26 09:22:18,740 INFO SystemMonitor:196 [interfaces.py:start():190] Started gpu monitoring
146
+ 2024-02-26 09:22:18,742 INFO SystemMonitor:196 [interfaces.py:start():190] Started memory monitoring
147
+ 2024-02-26 09:22:18,743 INFO SystemMonitor:196 [interfaces.py:start():190] Started network monitoring
148
+ 2024-02-26 09:22:19,271 DEBUG SenderThread:196 [sender.py:send():382] send: config
149
+ 2024-02-26 09:22:19,272 DEBUG SenderThread:196 [sender.py:send():382] send: metric
150
+ 2024-02-26 09:22:19,273 DEBUG SenderThread:196 [sender.py:send():382] send: metric
151
+ 2024-02-26 09:22:19,273 WARNING SenderThread:196 [sender.py:send_metric():1354] Seen metric with glob (shouldn't happen)
152
+ 2024-02-26 09:22:19,668 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: pause
153
+ 2024-02-26 09:22:19,669 INFO HandlerThread:196 [handler.py:handle_request_pause():708] stopping system metrics thread
154
+ 2024-02-26 09:22:19,669 INFO HandlerThread:196 [system_monitor.py:finish():203] Stopping system monitor
155
+ 2024-02-26 09:22:19,669 DEBUG SystemMonitor:196 [system_monitor.py:_start():172] Starting system metrics aggregation loop
156
+ 2024-02-26 09:22:19,669 DEBUG SystemMonitor:196 [system_monitor.py:_start():179] Finished system metrics aggregation loop
157
+ 2024-02-26 09:22:19,669 DEBUG SystemMonitor:196 [system_monitor.py:_start():183] Publishing last batch of metrics
158
+ 2024-02-26 09:22:19,670 INFO HandlerThread:196 [interfaces.py:finish():202] Joined cpu monitor
159
+ 2024-02-26 09:22:19,670 INFO HandlerThread:196 [interfaces.py:finish():202] Joined disk monitor
160
+ 2024-02-26 09:22:19,675 INFO HandlerThread:196 [interfaces.py:finish():202] Joined gpu monitor
161
+ 2024-02-26 09:22:19,676 INFO HandlerThread:196 [interfaces.py:finish():202] Joined memory monitor
162
+ 2024-02-26 09:22:19,676 INFO HandlerThread:196 [interfaces.py:finish():202] Joined network monitor
163
+ 2024-02-26 09:22:19,676 DEBUG SenderThread:196 [sender.py:send():382] send: stats
164
+ 2024-02-26 09:22:21,813 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
165
+ 2024-02-26 09:22:22,677 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
166
+ 2024-02-26 09:22:26,814 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
167
+ 2024-02-26 09:22:27,678 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
168
+ 2024-02-26 09:22:31,815 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
169
+ 2024-02-26 09:22:32,679 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
170
+ 2024-02-26 09:22:36,816 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
171
+ 2024-02-26 09:22:37,685 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
172
+ 2024-02-26 09:22:37,695 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: resume
173
+ 2024-02-26 09:22:37,696 INFO HandlerThread:196 [handler.py:handle_request_resume():699] starting system metrics thread
174
+ 2024-02-26 09:22:37,696 INFO HandlerThread:196 [system_monitor.py:start():194] Starting system monitor
175
+ 2024-02-26 09:22:37,696 INFO SystemMonitor:196 [system_monitor.py:_start():158] Starting system asset monitoring threads
176
+ 2024-02-26 09:22:37,696 INFO SystemMonitor:196 [interfaces.py:start():190] Started cpu monitoring
177
+ 2024-02-26 09:22:37,697 INFO SystemMonitor:196 [interfaces.py:start():190] Started disk monitoring
178
+ 2024-02-26 09:22:37,699 INFO SystemMonitor:196 [interfaces.py:start():190] Started gpu monitoring
179
+ 2024-02-26 09:22:37,701 INFO SystemMonitor:196 [interfaces.py:start():190] Started memory monitoring
180
+ 2024-02-26 09:22:37,702 INFO SystemMonitor:196 [interfaces.py:start():190] Started network monitoring
181
+ 2024-02-26 09:22:38,419 INFO Thread-12 :196 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240226_092105-91uhs91a/files/config.yaml
182
+ 2024-02-26 09:22:38,766 DEBUG SenderThread:196 [sender.py:send():382] send: config
183
+ 2024-02-26 09:22:38,768 DEBUG SenderThread:196 [sender.py:send():382] send: metric
184
+ 2024-02-26 09:22:38,768 DEBUG SenderThread:196 [sender.py:send():382] send: metric
185
+ 2024-02-26 09:22:38,768 WARNING SenderThread:196 [sender.py:send_metric():1354] Seen metric with glob (shouldn't happen)
186
+ 2024-02-26 09:22:42,174 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
187
+ 2024-02-26 09:22:43,769 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
188
+ 2024-02-26 09:22:47,175 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
189
+ 2024-02-26 09:22:48,770 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
190
+ 2024-02-26 09:22:52,181 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
191
+ 2024-02-26 09:22:53,770 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
192
+ 2024-02-26 09:22:57,184 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
193
+ 2024-02-26 09:22:58,771 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
194
+ 2024-02-26 09:23:02,185 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
195
+ 2024-02-26 09:23:03,772 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
196
+ 2024-02-26 09:23:07,189 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
197
+ 2024-02-26 09:23:08,778 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
198
+ 2024-02-26 09:23:09,430 INFO Thread-12 :196 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240226_092105-91uhs91a/files/config.yaml
199
+ 2024-02-26 09:23:12,191 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
200
+ 2024-02-26 09:23:13,902 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
201
+ 2024-02-26 09:23:17,192 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
202
+ 2024-02-26 09:23:18,902 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
203
+ 2024-02-26 09:23:19,821 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: pause
204
+ 2024-02-26 09:23:19,821 INFO HandlerThread:196 [handler.py:handle_request_pause():708] stopping system metrics thread
205
+ 2024-02-26 09:23:19,821 INFO HandlerThread:196 [system_monitor.py:finish():203] Stopping system monitor
206
+ 2024-02-26 09:23:19,821 DEBUG SystemMonitor:196 [system_monitor.py:_start():172] Starting system metrics aggregation loop
207
+ 2024-02-26 09:23:19,821 DEBUG SystemMonitor:196 [system_monitor.py:_start():179] Finished system metrics aggregation loop
208
+ 2024-02-26 09:23:19,821 DEBUG SystemMonitor:196 [system_monitor.py:_start():183] Publishing last batch of metrics
209
+ 2024-02-26 09:23:19,822 INFO HandlerThread:196 [interfaces.py:finish():202] Joined cpu monitor
210
+ 2024-02-26 09:23:19,822 INFO HandlerThread:196 [interfaces.py:finish():202] Joined disk monitor
211
+ 2024-02-26 09:23:19,830 INFO HandlerThread:196 [interfaces.py:finish():202] Joined gpu monitor
212
+ 2024-02-26 09:23:19,830 INFO HandlerThread:196 [interfaces.py:finish():202] Joined memory monitor
213
+ 2024-02-26 09:23:19,830 INFO HandlerThread:196 [interfaces.py:finish():202] Joined network monitor
214
+ 2024-02-26 09:23:19,830 DEBUG SenderThread:196 [sender.py:send():382] send: stats
215
+ 2024-02-26 09:23:22,193 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
216
+ 2024-02-26 09:23:24,832 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
217
+ 2024-02-26 09:23:27,194 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
218
+ 2024-02-26 09:23:29,832 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
219
+ 2024-02-26 09:23:32,195 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
220
+ 2024-02-26 09:23:32,481 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: resume
221
+ 2024-02-26 09:23:32,481 INFO HandlerThread:196 [handler.py:handle_request_resume():699] starting system metrics thread
222
+ 2024-02-26 09:23:32,481 INFO HandlerThread:196 [system_monitor.py:start():194] Starting system monitor
223
+ 2024-02-26 09:23:32,482 INFO SystemMonitor:196 [system_monitor.py:_start():158] Starting system asset monitoring threads
224
+ 2024-02-26 09:23:32,482 INFO SystemMonitor:196 [interfaces.py:start():190] Started cpu monitoring
225
+ 2024-02-26 09:23:32,483 INFO SystemMonitor:196 [interfaces.py:start():190] Started disk monitoring
226
+ 2024-02-26 09:23:32,484 INFO SystemMonitor:196 [interfaces.py:start():190] Started gpu monitoring
227
+ 2024-02-26 09:23:32,485 INFO SystemMonitor:196 [interfaces.py:start():190] Started memory monitoring
228
+ 2024-02-26 09:23:32,486 INFO SystemMonitor:196 [interfaces.py:start():190] Started network monitoring
229
+ 2024-02-26 09:23:33,088 DEBUG SenderThread:196 [sender.py:send():382] send: config
230
+ 2024-02-26 09:23:33,090 DEBUG SenderThread:196 [sender.py:send():382] send: metric
231
+ 2024-02-26 09:23:33,090 DEBUG SenderThread:196 [sender.py:send():382] send: metric
232
+ 2024-02-26 09:23:33,090 WARNING SenderThread:196 [sender.py:send_metric():1354] Seen metric with glob (shouldn't happen)
233
+ 2024-02-26 09:23:35,091 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
234
+ 2024-02-26 09:23:37,198 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
235
+ 2024-02-26 09:23:40,097 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
236
+ 2024-02-26 09:23:40,440 INFO Thread-12 :196 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240226_092105-91uhs91a/files/config.yaml
237
+ 2024-02-26 09:23:42,199 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
238
+ 2024-02-26 09:23:45,217 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
239
+ 2024-02-26 09:23:47,200 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
240
+ 2024-02-26 09:23:50,218 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
241
+ 2024-02-26 09:23:52,201 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
242
+ 2024-02-26 09:23:55,219 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
243
+ 2024-02-26 09:23:57,202 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
244
+ 2024-02-26 09:24:00,220 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
245
+ 2024-02-26 09:24:02,203 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
246
+ 2024-02-26 09:24:05,221 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
247
+ 2024-02-26 09:24:07,204 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
248
+ 2024-02-26 09:24:10,221 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
249
+ 2024-02-26 09:24:12,209 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
250
+ 2024-02-26 09:24:15,222 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
251
+ 2024-02-26 09:24:17,210 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
252
+ 2024-02-26 09:24:20,223 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
253
+ 2024-02-26 09:24:22,211 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
254
+ 2024-02-26 09:24:25,224 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
255
+ 2024-02-26 09:24:27,212 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
256
+ 2024-02-26 09:24:30,225 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
257
+ 2024-02-26 09:24:32,213 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
258
+ 2024-02-26 09:24:32,486 DEBUG SystemMonitor:196 [system_monitor.py:_start():172] Starting system metrics aggregation loop
259
+ 2024-02-26 09:24:32,488 DEBUG SenderThread:196 [sender.py:send():382] send: stats
260
+ 2024-02-26 09:24:35,489 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
261
+ 2024-02-26 09:24:37,214 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
262
+ 2024-02-26 09:24:40,490 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
263
+ 2024-02-26 09:24:42,215 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
264
+ 2024-02-26 09:24:45,491 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
265
+ 2024-02-26 09:24:47,219 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
266
+ 2024-02-26 09:24:50,491 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
267
+ 2024-02-26 09:24:52,220 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
268
+ 2024-02-26 09:24:55,492 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
269
+ 2024-02-26 09:24:57,221 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
270
+ 2024-02-26 09:25:00,493 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
271
+ 2024-02-26 09:25:02,226 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
272
+ 2024-02-26 09:25:02,488 DEBUG SenderThread:196 [sender.py:send():382] send: stats
273
+ 2024-02-26 09:25:06,490 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
274
+ 2024-02-26 09:25:07,227 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
275
+ 2024-02-26 09:25:07,470 INFO Thread-12 :196 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240226_092105-91uhs91a/files/output.log
276
+ 2024-02-26 09:25:09,470 INFO Thread-12 :196 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240226_092105-91uhs91a/files/output.log
277
+ 2024-02-26 09:25:11,847 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
278
+ 2024-02-26 09:25:12,228 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
279
+ 2024-02-26 09:25:16,848 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
280
+ 2024-02-26 09:25:17,229 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
281
+ 2024-02-26 09:25:21,849 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
282
+ 2024-02-26 09:25:22,230 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
283
+ 2024-02-26 09:25:26,850 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
284
+ 2024-02-26 09:25:27,247 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
285
+ 2024-02-26 09:25:31,851 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
286
+ 2024-02-26 09:25:32,284 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
287
+ 2024-02-26 09:25:32,489 DEBUG SenderThread:196 [sender.py:send():382] send: stats
288
+ 2024-02-26 09:25:34,173 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: partial_history
289
+ 2024-02-26 09:25:34,175 DEBUG SenderThread:196 [sender.py:send():382] send: metric
290
+ 2024-02-26 09:25:34,176 DEBUG SenderThread:196 [sender.py:send():382] send: metric
291
+ 2024-02-26 09:25:34,176 DEBUG SenderThread:196 [sender.py:send():382] send: metric
292
+ 2024-02-26 09:25:34,176 DEBUG SenderThread:196 [sender.py:send():382] send: metric
293
+ 2024-02-26 09:25:34,176 DEBUG SenderThread:196 [sender.py:send():382] send: metric
294
+ 2024-02-26 09:25:34,176 DEBUG SenderThread:196 [sender.py:send():382] send: metric
295
+ 2024-02-26 09:25:34,176 DEBUG SenderThread:196 [sender.py:send():382] send: metric
296
+ 2024-02-26 09:25:34,176 DEBUG SenderThread:196 [sender.py:send():382] send: history
297
+ 2024-02-26 09:25:34,177 DEBUG SenderThread:196 [sender.py:send_request():409] send_request: summary_record
298
+ 2024-02-26 09:25:34,177 INFO SenderThread:196 [sender.py:_save_file():1403] saving file wandb-summary.json with policy end
299
+ 2024-02-26 09:25:34,480 INFO Thread-12 :196 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240226_092105-91uhs91a/files/wandb-summary.json
300
+ 2024-02-26 09:25:37,178 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
301
+ 2024-02-26 09:25:37,295 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
302
+ 2024-02-26 09:25:42,215 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
303
+ 2024-02-26 09:25:42,302 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
304
+ 2024-02-26 09:25:42,484 INFO Thread-12 :196 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240226_092105-91uhs91a/files/config.yaml
305
+ 2024-02-26 09:25:44,419 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: partial_history
306
+ 2024-02-26 09:25:44,420 DEBUG SenderThread:196 [sender.py:send():382] send: metric
307
+ 2024-02-26 09:25:44,421 DEBUG SenderThread:196 [sender.py:send():382] send: metric
308
+ 2024-02-26 09:25:44,421 DEBUG SenderThread:196 [sender.py:send():382] send: history
309
+ 2024-02-26 09:25:44,421 DEBUG SenderThread:196 [sender.py:send_request():409] send_request: summary_record
310
+ 2024-02-26 09:25:44,421 INFO SenderThread:196 [sender.py:_save_file():1403] saving file wandb-summary.json with policy end
311
+ 2024-02-26 09:25:44,485 INFO Thread-12 :196 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240226_092105-91uhs91a/files/wandb-summary.json
312
+ 2024-02-26 09:25:47,401 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
313
+ 2024-02-26 09:25:47,422 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
314
+ 2024-02-26 09:25:52,424 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
315
+ 2024-02-26 09:25:52,494 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
316
+ 2024-02-26 09:25:57,424 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
317
+ 2024-02-26 09:25:57,510 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
318
+ 2024-02-26 09:26:02,425 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
319
+ 2024-02-26 09:26:02,490 DEBUG SenderThread:196 [sender.py:send():382] send: stats
320
+ 2024-02-26 09:26:02,510 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
321
+ 2024-02-26 09:26:07,491 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
322
+ 2024-02-26 09:26:07,512 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
323
+ 2024-02-26 09:26:12,498 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
324
+ 2024-02-26 09:26:12,518 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
325
+ 2024-02-26 09:26:13,499 INFO Thread-12 :196 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240226_092105-91uhs91a/files/config.yaml
326
+ 2024-02-26 09:26:17,520 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
327
+ 2024-02-26 09:26:17,670 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
wandb/debug.log ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-02-26 09:21:05,526 INFO MainThread:34 [wandb_setup.py:_flush():76] Current SDK version is 0.16.2
2
+ 2024-02-26 09:21:05,526 INFO MainThread:34 [wandb_setup.py:_flush():76] Configure stats pid to 34
3
+ 2024-02-26 09:21:05,526 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-02-26 09:21:05,526 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-02-26 09:21:05,526 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-02-26 09:21:05,527 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-02-26 09:21:05,527 INFO MainThread:34 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-02-26 09:21:05,527 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
9
+ 2024-02-26 09:21:05,527 INFO MainThread:34 [wandb_init.py:_log_setup():526] Logging user logs to /kaggle/working/wandb/run-20240226_092105-91uhs91a/logs/debug.log
10
+ 2024-02-26 09:21:05,527 INFO MainThread:34 [wandb_init.py:_log_setup():527] Logging internal logs to /kaggle/working/wandb/run-20240226_092105-91uhs91a/logs/debug-internal.log
11
+ 2024-02-26 09:21:05,527 INFO MainThread:34 [wandb_init.py:_jupyter_setup():472] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x77fcc2873700>
12
+ 2024-02-26 09:21:05,527 INFO MainThread:34 [wandb_init.py:init():566] calling init triggers
13
+ 2024-02-26 09:21:05,527 INFO MainThread:34 [wandb_init.py:init():573] wandb.init called with sweep_config: {}
14
+ config: {}
15
+ 2024-02-26 09:21:05,527 INFO MainThread:34 [wandb_init.py:init():616] starting backend
16
+ 2024-02-26 09:21:05,528 INFO MainThread:34 [wandb_init.py:init():620] setting up manager
17
+ 2024-02-26 09:21:05,529 INFO MainThread:34 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-02-26 09:21:05,532 INFO MainThread:34 [wandb_init.py:init():628] backend started and connected
19
+ 2024-02-26 09:21:05,543 INFO MainThread:34 [wandb_run.py:_label_probe_notebook():1294] probe notebook
20
+ 2024-02-26 09:21:06,207 INFO MainThread:34 [wandb_init.py:init():720] updated telemetry
21
+ 2024-02-26 09:21:06,210 INFO MainThread:34 [wandb_init.py:init():753] communicating run to backend with 90.0 second timeout
22
+ 2024-02-26 09:21:06,396 INFO MainThread:34 [wandb_run.py:_on_init():2254] communicating current version
23
+ 2024-02-26 09:21:06,458 INFO MainThread:34 [wandb_run.py:_on_init():2263] got version response upgrade_message: "wandb version 0.16.3 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
24
+
25
+ 2024-02-26 09:21:06,458 INFO MainThread:34 [wandb_init.py:init():804] starting run threads in backend
26
+ 2024-02-26 09:21:36,490 INFO MainThread:34 [wandb_run.py:_console_start():2233] atexit reg
27
+ 2024-02-26 09:21:36,490 INFO MainThread:34 [wandb_run.py:_redirect():2088] redirect: wrap_raw
28
+ 2024-02-26 09:21:36,491 INFO MainThread:34 [wandb_run.py:_redirect():2153] Wrapping output streams.
29
+ 2024-02-26 09:21:36,491 INFO MainThread:34 [wandb_run.py:_redirect():2178] Redirects installed.
30
+ 2024-02-26 09:21:36,492 INFO MainThread:34 [wandb_init.py:init():847] run started, returning control to user process
31
+ 2024-02-26 09:21:36,500 INFO MainThread:34 [wandb_run.py:_config_callback():1342] config_cb None None {'vocab_size': 32128, 'd_model': 768, 'd_kv': 64, 'd_ff': 2048, 'num_layers': 12, 'num_decoder_layers': 12, 'num_heads': 12, 'relative_attention_num_buckets': 32, 'relative_attention_max_distance': 128, 'dropout_rate': 0.1, 'classifier_dropout': 0.0, 'layer_norm_epsilon': 1e-06, 'initializer_factor': 1.0, 'feed_forward_proj': 'gated-gelu', 'use_cache': True, 'dense_act_fn': 'gelu_new', 'is_gated_act': True, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['T5ForConditionalGeneration'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': 1, 'sep_token_id': None, 'decoder_start_token_id': 0, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'csebuetnlp/banglat5', 'transformers_version': '4.37.0', 'gradient_checkpointing': False, 'model_type': 't5', 'output_past': True, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 300, 'per_device_eval_batch_size': 32, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0005, 'weight_decay': 0.1, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 5, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Feb26_09-20-49_d79522c38279', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 3, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': False, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': None, 'generation_num_beams': None, 'generation_config': None}
32
+ 2024-02-26 09:21:37,880 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
33
+ 2024-02-26 09:21:37,880 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
34
+ 2024-02-26 09:21:58,243 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
35
+ 2024-02-26 09:21:58,790 INFO MainThread:34 [wandb_run.py:_config_callback():1342] config_cb None None {'vocab_size': 32128, 'd_model': 768, 'd_kv': 64, 'd_ff': 2048, 'num_layers': 12, 'num_decoder_layers': 12, 'num_heads': 12, 'relative_attention_num_buckets': 32, 'relative_attention_max_distance': 128, 'dropout_rate': 0.1, 'classifier_dropout': 0.0, 'layer_norm_epsilon': 1e-06, 'initializer_factor': 1.0, 'feed_forward_proj': 'gated-gelu', 'use_cache': True, 'dense_act_fn': 'gelu_new', 'is_gated_act': True, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['T5ForConditionalGeneration'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': 1, 'sep_token_id': None, 'decoder_start_token_id': 0, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'csebuetnlp/banglat5', 'transformers_version': '4.37.0', 'gradient_checkpointing': False, 'model_type': 't5', 'output_past': True, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 300, 'per_device_eval_batch_size': 16, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0005, 'weight_decay': 0.1, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 5, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Feb26_09-21-58_d79522c38279', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 3, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': False, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': None, 'generation_num_beams': None, 'generation_config': None}
36
+ 2024-02-26 09:21:59,601 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
37
+ 2024-02-26 09:21:59,601 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
38
+ 2024-02-26 09:22:08,330 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
39
+ 2024-02-26 09:22:09,400 INFO MainThread:34 [wandb_run.py:_config_callback():1342] config_cb None None {'vocab_size': 32128, 'd_model': 768, 'd_kv': 64, 'd_ff': 2048, 'num_layers': 12, 'num_decoder_layers': 12, 'num_heads': 12, 'relative_attention_num_buckets': 32, 'relative_attention_max_distance': 128, 'dropout_rate': 0.1, 'classifier_dropout': 0.0, 'layer_norm_epsilon': 1e-06, 'initializer_factor': 1.0, 'feed_forward_proj': 'gated-gelu', 'use_cache': True, 'dense_act_fn': 'gelu_new', 'is_gated_act': True, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['T5ForConditionalGeneration'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': 1, 'sep_token_id': None, 'decoder_start_token_id': 0, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'csebuetnlp/banglat5', 'transformers_version': '4.37.0', 'gradient_checkpointing': False, 'model_type': 't5', 'output_past': True, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 300, 'per_device_eval_batch_size': 16, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0005, 'weight_decay': 0.1, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 5, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Feb26_09-22-08_d79522c38279', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 3, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': False, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': None, 'generation_num_beams': None, 'generation_config': None}
40
+ 2024-02-26 09:22:09,926 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
41
+ 2024-02-26 09:22:09,926 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
42
+ 2024-02-26 09:22:18,737 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
43
+ 2024-02-26 09:22:19,266 INFO MainThread:34 [wandb_run.py:_config_callback():1342] config_cb None None {'vocab_size': 32128, 'd_model': 768, 'd_kv': 64, 'd_ff': 2048, 'num_layers': 12, 'num_decoder_layers': 12, 'num_heads': 12, 'relative_attention_num_buckets': 32, 'relative_attention_max_distance': 128, 'dropout_rate': 0.1, 'classifier_dropout': 0.0, 'layer_norm_epsilon': 1e-06, 'initializer_factor': 1.0, 'feed_forward_proj': 'gated-gelu', 'use_cache': True, 'dense_act_fn': 'gelu_new', 'is_gated_act': True, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['T5ForConditionalGeneration'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': 1, 'sep_token_id': None, 'decoder_start_token_id': 0, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'csebuetnlp/banglat5', 'transformers_version': '4.37.0', 'gradient_checkpointing': False, 'model_type': 't5', 'output_past': True, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 300, 'per_device_eval_batch_size': 10, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0005, 'weight_decay': 0.1, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 5, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Feb26_09-22-18_d79522c38279', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 3, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': False, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': None, 'generation_num_beams': None, 'generation_config': None}
44
+ 2024-02-26 09:22:19,668 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
45
+ 2024-02-26 09:22:19,668 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
46
+ 2024-02-26 09:22:37,695 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
47
+ 2024-02-26 09:22:38,762 INFO MainThread:34 [wandb_run.py:_config_callback():1342] config_cb None None {'vocab_size': 32128, 'd_model': 768, 'd_kv': 64, 'd_ff': 2048, 'num_layers': 12, 'num_decoder_layers': 12, 'num_heads': 12, 'relative_attention_num_buckets': 32, 'relative_attention_max_distance': 128, 'dropout_rate': 0.1, 'classifier_dropout': 0.0, 'layer_norm_epsilon': 1e-06, 'initializer_factor': 1.0, 'feed_forward_proj': 'gated-gelu', 'use_cache': True, 'dense_act_fn': 'gelu_new', 'is_gated_act': True, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['T5ForConditionalGeneration'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': 1, 'sep_token_id': None, 'decoder_start_token_id': 0, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'csebuetnlp/banglat5', 'transformers_version': '4.37.0', 'gradient_checkpointing': False, 'model_type': 't5', 'output_past': True, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0005, 'weight_decay': 0.1, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 5, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Feb26_09-22-37_d79522c38279', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 3, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': False, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': None, 'generation_num_beams': None, 'generation_config': None}
48
+ 2024-02-26 09:23:19,820 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
49
+ 2024-02-26 09:23:19,820 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
50
+ 2024-02-26 09:23:32,480 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
51
+ 2024-02-26 09:23:33,079 INFO MainThread:34 [wandb_run.py:_config_callback():1342] config_cb None None {'vocab_size': 32128, 'd_model': 768, 'd_kv': 64, 'd_ff': 2048, 'num_layers': 12, 'num_decoder_layers': 12, 'num_heads': 12, 'relative_attention_num_buckets': 32, 'relative_attention_max_distance': 128, 'dropout_rate': 0.1, 'classifier_dropout': 0.0, 'layer_norm_epsilon': 1e-06, 'initializer_factor': 1.0, 'feed_forward_proj': 'gated-gelu', 'use_cache': True, 'dense_act_fn': 'gelu_new', 'is_gated_act': True, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['T5ForConditionalGeneration'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': 1, 'sep_token_id': None, 'decoder_start_token_id': 0, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'csebuetnlp/banglat5', 'transformers_version': '4.37.0', 'gradient_checkpointing': False, 'model_type': 't5', 'output_past': True, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 16, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0005, 'weight_decay': 0.1, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 5, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Feb26_09-23-32_d79522c38279', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 3, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': False, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': None, 'generation_num_beams': None, 'generation_config': None}
wandb/run-20240226_092105-91uhs91a/files/conda-environment.yaml ADDED
@@ -0,0 +1,1079 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: base
2
+ channels:
3
+ - pytorch
4
+ - file:///tmp/conda
5
+ - rapidsai
6
+ - nvidia
7
+ - conda-forge
8
+ - defaults
9
+ dependencies:
10
+ - _libgcc_mutex=0.1=conda_forge
11
+ - _openmp_mutex=4.5=2_gnu
12
+ - aiohttp=3.9.1=py310h2372a71_0
13
+ - aiosignal=1.3.1=pyhd8ed1ab_0
14
+ - annotated-types=0.6.0=pyhd8ed1ab_0
15
+ - anyio=4.2.0=pyhd8ed1ab_0
16
+ - archspec=0.2.2=pyhd8ed1ab_0
17
+ - argon2-cffi=23.1.0=pyhd8ed1ab_0
18
+ - argon2-cffi-bindings=21.2.0=py310h2372a71_4
19
+ - arrow=1.3.0=pyhd8ed1ab_0
20
+ - arrow-cpp=11.0.0=ha770c72_9_cpu
21
+ - asttokens=2.4.1=pyhd8ed1ab_0
22
+ - async-timeout=4.0.3=pyhd8ed1ab_0
23
+ - atk-1.0=2.38.0=hd4edc92_1
24
+ - attrs=23.2.0=pyh71513ae_0
25
+ - aws-c-auth=0.6.26=h987a71b_2
26
+ - aws-c-cal=0.5.21=h48707d8_2
27
+ - aws-c-common=0.8.14=h0b41bf4_0
28
+ - aws-c-compression=0.2.16=h03acc5a_5
29
+ - aws-c-event-stream=0.2.20=h00877a2_4
30
+ - aws-c-http=0.7.6=hf342b9f_0
31
+ - aws-c-io=0.13.19=h5b20300_3
32
+ - aws-c-mqtt=0.8.6=hc4349f7_12
33
+ - aws-c-s3=0.2.7=h909e904_1
34
+ - aws-c-sdkutils=0.1.9=h03acc5a_0
35
+ - aws-checksums=0.1.14=h03acc5a_5
36
+ - aws-crt-cpp=0.19.8=hf7fbfca_12
37
+ - aws-sdk-cpp=1.10.57=h17c43bd_8
38
+ - beautifulsoup4=4.12.2=pyha770c72_0
39
+ - bleach=6.1.0=pyhd8ed1ab_0
40
+ - bokeh=3.3.3=pyhd8ed1ab_0
41
+ - boltons=23.1.1=pyhd8ed1ab_0
42
+ - brotli=1.0.9=h166bdaf_9
43
+ - brotli-bin=1.0.9=h166bdaf_9
44
+ - brotli-python=1.0.9=py310hd8f1fbe_9
45
+ - brotlipy=0.7.0=py310h7f8727e_1002
46
+ - bzip2=1.0.8=h7b6447c_0
47
+ - c-ares=1.25.0=hd590300_0
48
+ - ca-certificates=2023.11.17=hbcca054_0
49
+ - cached-property=1.5.2=hd8ed1ab_1
50
+ - cached_property=1.5.2=pyha770c72_1
51
+ - cairo=1.18.0=h3faef2a_0
52
+ - cartopy=0.22.0=py310hcc13569_1
53
+ - catalogue=2.0.10=py310hff52083_0
54
+ - certifi=2023.11.17=pyhd8ed1ab_0
55
+ - cffi=1.16.0=py310h2fee648_0
56
+ - charset-normalizer=3.3.2=pyhd8ed1ab_0
57
+ - click=8.1.7=unix_pyh707e725_0
58
+ - cloudpathlib=0.16.0=pyhd8ed1ab_0
59
+ - colorama=0.4.6=pyhd8ed1ab_0
60
+ - comm=0.2.1=pyhd8ed1ab_0
61
+ - conda=23.7.4=py310hff52083_0
62
+ - conda-libmamba-solver=23.7.0=pyhd8ed1ab_0
63
+ - conda-package-handling=2.2.0=pyh38be061_0
64
+ - conda-package-streaming=0.9.0=pyhd8ed1ab_0
65
+ - confection=0.1.4=py310h17c5347_0
66
+ - contourpy=1.2.0=py310hd41b1e2_0
67
+ - cryptography=41.0.7=py310hb8475ec_1
68
+ - cuda-cccl=12.3.101=0
69
+ - cuda-cudart=12.3.101=0
70
+ - cuda-cudart-dev=12.3.101=0
71
+ - cuda-nvcc-dev_linux-64=12.1.105=ha770c72_0
72
+ - cuda-nvcc-impl=12.1.105=hd3aeb46_0
73
+ - cuda-nvcc-tools=12.1.105=hd3aeb46_0
74
+ - cuda-nvrtc=12.1.105=hd3aeb46_0
75
+ - cuda-profiler-api=12.3.101=0
76
+ - cuda-python=12.3.0=py310h7ca8cf1_0
77
+ - cuda-version=12.1=h1d6eff3_2
78
+ - cudf=23.08.00=cuda12_py310_230809_g8150d38e08_0
79
+ - cuml=23.08.00=cuda12_py310_230809_gd7162cdea_0
80
+ - cupy=13.0.0=py310h7aad9d2_0
81
+ - cupy-core=13.0.0=py310had4011e_0
82
+ - curl=8.5.0=hca28451_0
83
+ - cycler=0.12.1=pyhd8ed1ab_0
84
+ - cymem=2.0.8=py310hc6cd4ac_1
85
+ - cython-blis=0.7.10=py310h1f7b6fc_2
86
+ - cytoolz=0.12.2=py310h2372a71_1
87
+ - dask-cuda=23.08.00=py310_230809_gefbd6ca_0
88
+ - dask-cudf=23.08.00=cuda12_py310_230809_g8150d38e08_0
89
+ - debugpy=1.8.0=py310hc6cd4ac_1
90
+ - decorator=5.1.1=pyhd8ed1ab_0
91
+ - defusedxml=0.7.1=pyhd8ed1ab_0
92
+ - distro=1.9.0=pyhd8ed1ab_0
93
+ - dlenv-tf-2-15-gpu=1.0.20240111=py310ha20f8e0_0
94
+ - dlpack=0.5=h9c3ff4c_0
95
+ - entrypoints=0.4=pyhd8ed1ab_0
96
+ - exceptiongroup=1.2.0=pyhd8ed1ab_2
97
+ - executing=2.0.1=pyhd8ed1ab_0
98
+ - expat=2.5.0=hcb278e6_1
99
+ - fastrlock=0.8.2=py310hc6cd4ac_2
100
+ - fftw=3.3.10=nompi_hc118613_108
101
+ - fmt=9.1.0=h924138e_0
102
+ - font-ttf-dejavu-sans-mono=2.37=hab24e00_0
103
+ - font-ttf-inconsolata=3.000=h77eed37_0
104
+ - font-ttf-source-code-pro=2.038=h77eed37_0
105
+ - font-ttf-ubuntu=0.83=h77eed37_1
106
+ - fontconfig=2.14.2=h14ed4e7_0
107
+ - fonts-conda-ecosystem=1=0
108
+ - fonts-conda-forge=1=0
109
+ - fqdn=1.5.1=pyhd8ed1ab_0
110
+ - freetype=2.12.1=h267a509_2
111
+ - fribidi=1.0.10=h36c2ea0_0
112
+ - frozenlist=1.4.1=py310h2372a71_0
113
+ - fsspec=2023.12.2=pyhca7485f_0
114
+ - gdk-pixbuf=2.42.10=h829c605_4
115
+ - geos=3.11.1=h27087fc_0
116
+ - gettext=0.21.1=h27087fc_0
117
+ - gflags=2.2.2=he1b5a44_1004
118
+ - ghostscript=10.02.1=h59595ed_0
119
+ - giflib=5.2.1=h0b41bf4_3
120
+ - glog=0.6.0=h6f12383_0
121
+ - gmock=1.14.0=ha770c72_1
122
+ - gmp=6.3.0=h59595ed_0
123
+ - google-api-core-grpc=2.11.1=hd8ed1ab_0
124
+ - google-auth=2.26.1=pyhca7485f_0
125
+ - google-cloud-core=2.4.1=pyhd8ed1ab_0
126
+ - google-cloud-datastore=2.19.0=pyhd8ed1ab_0
127
+ - googleapis-common-protos=1.62.0=pyhd8ed1ab_0
128
+ - graphite2=1.3.13=h58526e2_1001
129
+ - graphviz=9.0.0=h78e8752_1
130
+ - grpc-cpp=1.51.1=h27aab58_3
131
+ - gtest=1.14.0=h00ab1b0_1
132
+ - gtk2=2.24.33=h7f000aa_3
133
+ - gts=0.7.6=h977cf35_4
134
+ - harfbuzz=8.3.0=h3d44ed6_0
135
+ - icu=73.2=h59595ed_0
136
+ - idna=3.6=pyhd8ed1ab_0
137
+ - imagemagick=7.1.1_26=pl5321h3fa1221_0
138
+ - importlib_metadata=7.0.1=hd8ed1ab_0
139
+ - importlib_resources=6.1.1=pyhd8ed1ab_0
140
+ - intel-openmp=2023.1.0=hdb19cb5_46306
141
+ - ipykernel=6.28.0=pyhd33586a_0
142
+ - ipython=8.20.0=pyh707e725_0
143
+ - ipython_genutils=0.2.0=py_1
144
+ - isoduration=20.11.0=pyhd8ed1ab_0
145
+ - jbig=2.1=h7f98852_2003
146
+ - jedi=0.19.1=pyhd8ed1ab_0
147
+ - jinja2=3.1.2=pyhd8ed1ab_1
148
+ - joblib=1.3.2=pyhd8ed1ab_0
149
+ - jsonpatch=1.33=pyhd8ed1ab_0
150
+ - jsonpointer=2.4=py310hff52083_3
151
+ - jsonschema=4.20.0=pyhd8ed1ab_0
152
+ - jsonschema-specifications=2023.12.1=pyhd8ed1ab_0
153
+ - jsonschema-with-format-nongpl=4.20.0=pyhd8ed1ab_0
154
+ - jupyter_client=8.6.0=pyhd8ed1ab_0
155
+ - jupyter_core=5.7.1=py310hff52083_0
156
+ - jupyter_events=0.9.0=pyhd8ed1ab_0
157
+ - jupyter_server_terminals=0.5.1=pyhd8ed1ab_0
158
+ - jupyterlab_pygments=0.3.0=pyhd8ed1ab_0
159
+ - keyutils=1.6.1=h166bdaf_0
160
+ - kiwisolver=1.4.5=py310hd41b1e2_1
161
+ - krb5=1.21.2=h659d440_0
162
+ - langcodes=3.3.0=pyhd8ed1ab_0
163
+ - lcms2=2.16=hb7c19ff_0
164
+ - ld_impl_linux-64=2.40=h41732ed_0
165
+ - lerc=4.0.0=h27087fc_0
166
+ - libabseil=20230125.0=cxx17_hcb278e6_1
167
+ - libarchive=3.6.2=h039dbb9_1
168
+ - libarrow=11.0.0=h33598ff_9_cpu
169
+ - libblas=3.9.0=21_linux64_openblas
170
+ - libbrotlicommon=1.0.9=h166bdaf_9
171
+ - libbrotlidec=1.0.9=h166bdaf_9
172
+ - libbrotlienc=1.0.9=h166bdaf_9
173
+ - libcblas=3.9.0=21_linux64_openblas
174
+ - libcrc32c=1.1.2=h9c3ff4c_0
175
+ - libcublas=12.1.3.1=hd3aeb46_0
176
+ - libcublas-dev=12.1.3.1=0
177
+ - libcudf=23.08.00=cuda12_230809_g8150d38e08_0
178
+ - libcufft=11.0.2.54=hd3aeb46_0
179
+ - libcufile=1.8.1.2=0
180
+ - libcufile-dev=1.8.1.2=0
181
+ - libcuml=23.08.00=cuda12_230809_gd7162cdea_0
182
+ - libcumlprims=23.08.00=cuda12_230809_g71c0a86_0
183
+ - libcurand=10.3.2.106=hd3aeb46_0
184
+ - libcurand-dev=10.3.2.106=0
185
+ - libcurl=8.5.0=hca28451_0
186
+ - libcusolver=11.4.5.107=hd3aeb46_0
187
+ - libcusolver-dev=11.4.5.107=0
188
+ - libcusparse=12.1.0.106=hd3aeb46_0
189
+ - libcusparse-dev=12.1.0.106=0
190
+ - libdeflate=1.19=hd590300_0
191
+ - libedit=3.1.20191231=he28a2e2_2
192
+ - libev=4.33=hd590300_2
193
+ - libevent=2.1.10=h28343ad_4
194
+ - libexpat=2.5.0=hcb278e6_1
195
+ - libffi=3.4.2=h7f98852_5
196
+ - libgcc-ng=13.2.0=h807b86a_3
197
+ - libgd=2.3.3=h119a65a_9
198
+ - libgfortran-ng=13.2.0=h69a702a_3
199
+ - libgfortran5=13.2.0=ha4646dd_3
200
+ - libglib=2.78.3=h783c2da_0
201
+ - libgomp=13.2.0=h807b86a_3
202
+ - libgoogle-cloud=2.8.0=h3c06191_0
203
+ - libgrpc=1.51.1=hcf146ea_3
204
+ - libhwloc=2.9.3=default_h554bfaf_1009
205
+ - libiconv=1.17=hd590300_2
206
+ - libjpeg-turbo=3.0.0=hd590300_1
207
+ - libkvikio=23.08.00=cuda12_230809_g51a9036_0
208
+ - liblapack=3.9.0=21_linux64_openblas
209
+ - libllvm14=14.0.6=hcd5def8_4
210
+ - libmamba=1.5.0=h658169a_0
211
+ - libmambapy=1.5.0=py310h8aae740_0
212
+ - libnghttp2=1.58.0=h47da74e_1
213
+ - libnsl=2.0.1=hd590300_0
214
+ - libnuma=2.0.16=h0b41bf4_1
215
+ - libnvjitlink=12.1.105=hd3aeb46_0
216
+ - libopenblas=0.3.26=pthreads_h413a1c8_0
217
+ - libpng=1.6.39=h753d276_0
218
+ - libprotobuf=3.21.12=hfc55251_2
219
+ - libraft=23.08.00=cuda12_230809_ge588d7b5_0
220
+ - libraft-headers=23.08.00=cuda12_230809_ge588d7b5_0
221
+ - libraft-headers-only=23.08.00=cuda12_230809_ge588d7b5_0
222
+ - librmm=23.08.00=cuda12_230809_gf3af0e8d_0
223
+ - librsvg=2.56.3=he3f83f7_1
224
+ - libsodium=1.0.18=h36c2ea0_1
225
+ - libsolv=0.7.27=hfc55251_0
226
+ - libsqlite=3.44.2=h2797004_0
227
+ - libssh2=1.11.0=h0841786_0
228
+ - libstdcxx-ng=13.2.0=h7e041cc_3
229
+ - libthrift=0.18.0=h5e4af38_0
230
+ - libtiff=4.6.0=ha9c0a0a_2
231
+ - libutf8proc=2.8.0=h166bdaf_0
232
+ - libuuid=2.38.1=h0b41bf4_0
233
+ - libuv=1.46.0=hd590300_0
234
+ - libwebp=1.3.2=h658648e_1
235
+ - libwebp-base=1.3.2=hd590300_0
236
+ - libxcb=1.15=h0b41bf4_0
237
+ - libxcrypt=4.4.36=hd590300_1
238
+ - libxml2=2.12.3=h232c23b_0
239
+ - libzlib=1.2.13=hd590300_5
240
+ - llvm-openmp=8.0.1=hc9558a2_0
241
+ - locket=1.0.0=pyhd8ed1ab_0
242
+ - lz4=4.3.3=py310h350c4a5_0
243
+ - lz4-c=1.9.4=hcb278e6_0
244
+ - lzo=2.10=h516909a_1000
245
+ - magma-cuda121=2.6.1=1
246
+ - mamba=1.5.0=py310h51d5547_0
247
+ - markdown-it-py=3.0.0=pyhd8ed1ab_0
248
+ - matplotlib-base=3.8.2=py310h62c0568_0
249
+ - matplotlib-inline=0.1.6=pyhd8ed1ab_0
250
+ - mdurl=0.1.2=pyhd8ed1ab_0
251
+ - menuinst=2.0.1=py310hff52083_0
252
+ - mkl=2023.1.0=h213fc3f_46344
253
+ - msgpack-python=1.0.7=py310hd41b1e2_0
254
+ - multidict=6.0.4=py310h2372a71_1
255
+ - munkres=1.1.4=pyh9f0ad1d_0
256
+ - murmurhash=1.0.10=py310hc6cd4ac_1
257
+ - nb_conda=2.2.1=unix_7
258
+ - nb_conda_kernels=2.3.1=pyhd8ed1ab_3
259
+ - nbclassic=1.0.0=pyhb4ecaf3_1
260
+ - nbconvert-pandoc=7.14.0=pyhd8ed1ab_0
261
+ - nbformat=5.9.2=pyhd8ed1ab_0
262
+ - nccl=2.19.4.1=h3a97aeb_0
263
+ - ncurses=6.4=h59595ed_2
264
+ - nest-asyncio=1.5.8=pyhd8ed1ab_0
265
+ - nodejs=20.9.0=hb753e55_0
266
+ - notebook-shim=0.2.3=pyhd8ed1ab_0
267
+ - numba=0.58.1=py310h7dc5dd1_0
268
+ - nvcomp=2.6.1=h10b603f_3
269
+ - nvtx=0.2.8=py310h2372a71_1
270
+ - openjpeg=2.5.0=h488ebb8_3
271
+ - openmp=8.0.1=0
272
+ - openssl=3.2.0=hd590300_1
273
+ - orc=1.8.2=hfdbbad2_2
274
+ - overrides=7.4.0=pyhd8ed1ab_0
275
+ - pandoc=3.1.3=h32600fe_0
276
+ - pandocfilters=1.5.0=pyhd8ed1ab_0
277
+ - pango=1.50.14=ha41ecd1_2
278
+ - parquet-cpp=1.5.1=2
279
+ - parso=0.8.3=pyhd8ed1ab_0
280
+ - partd=1.4.1=pyhd8ed1ab_0
281
+ - pathy=0.10.3=py310h06a4308_0
282
+ - pcre2=10.42=hcad00b1_0
283
+ - perl=5.32.1=7_hd590300_perl5
284
+ - pickleshare=0.7.5=py_1003
285
+ - pip=23.3.2=pyhd8ed1ab_0
286
+ - pixman=0.43.0=h59595ed_0
287
+ - pkg-config=0.29.2=h36c2ea0_1008
288
+ - pkgutil-resolve-name=1.3.10=pyhd8ed1ab_1
289
+ - platformdirs=4.1.0=pyhd8ed1ab_0
290
+ - pluggy=1.3.0=pyhd8ed1ab_0
291
+ - preshed=3.0.9=py310hc6cd4ac_1
292
+ - proj=9.3.1=h1d62c97_0
293
+ - prometheus_client=0.19.0=pyhd8ed1ab_0
294
+ - proto-plus=1.23.0=pyhd8ed1ab_0
295
+ - pthread-stubs=0.4=h36c2ea0_1001
296
+ - ptyprocess=0.7.0=pyhd3deb0d_0
297
+ - pure_eval=0.2.2=pyhd8ed1ab_0
298
+ - pyarrow=11.0.0=py310h633f555_9_cpu
299
+ - pyasn1=0.5.1=pyhd8ed1ab_0
300
+ - pyasn1-modules=0.3.0=pyhd8ed1ab_0
301
+ - pybind11-abi=4=hd8ed1ab_3
302
+ - pycosat=0.6.6=py310h2372a71_0
303
+ - pydantic=2.5.3=pyhd8ed1ab_0
304
+ - pydantic-core=2.14.6=py310hcb5633a_1
305
+ - pygments=2.17.2=pyhd8ed1ab_0
306
+ - pylibraft=23.08.00=cuda12_py310_230809_ge588d7b5_0
307
+ - pynvml=11.4.1=pyhd8ed1ab_0
308
+ - pyopenssl=23.3.0=pyhd8ed1ab_0
309
+ - pyparsing=3.1.1=pyhd8ed1ab_0
310
+ - pyproj=3.6.1=py310hd5c30f3_5
311
+ - pyshp=2.3.1=pyhd8ed1ab_0
312
+ - pysocks=1.7.1=py310h06a4308_0
313
+ - python=3.10.13=hd12c33a_1_cpython
314
+ - python-dateutil=2.8.2=pyhd8ed1ab_0
315
+ - python-fastjsonschema=2.19.1=pyhd8ed1ab_0
316
+ - python-json-logger=2.0.7=pyhd8ed1ab_0
317
+ - python_abi=3.10=4_cp310
318
+ - pytz=2023.3.post1=pyhd8ed1ab_0
319
+ - pyu2f=0.1.5=pyhd8ed1ab_0
320
+ - pyyaml=6.0.1=py310h2372a71_1
321
+ - raft-dask=23.08.00=cuda12_py310_230809_ge588d7b5_0
322
+ - rdma-core=28.9=h59595ed_1
323
+ - re2=2023.02.02=hcb278e6_0
324
+ - readline=8.2=h8228510_1
325
+ - referencing=0.32.1=pyhd8ed1ab_0
326
+ - reproc=14.2.4.post0=hd590300_1
327
+ - reproc-cpp=14.2.4.post0=h59595ed_1
328
+ - requests=2.31.0=pyhd8ed1ab_0
329
+ - rfc3339-validator=0.1.4=pyhd8ed1ab_0
330
+ - rfc3986-validator=0.1.1=pyh9f0ad1d_0
331
+ - rich=13.7.0=pyhd8ed1ab_0
332
+ - rmm=23.08.00=cuda12_py310_230809_gf3af0e8d_0
333
+ - rpds-py=0.16.2=py310hcb5633a_0
334
+ - rsa=4.9=pyhd8ed1ab_0
335
+ - ruamel.yaml=0.17.40=py310h2372a71_0
336
+ - ruamel.yaml.clib=0.2.7=py310h2372a71_2
337
+ - ruamel_yaml=0.15.100=py310h7f8727e_0
338
+ - s2n=1.3.41=h3358134_0
339
+ - send2trash=1.8.2=pyh41d4057_0
340
+ - setuptools=69.0.3=pyhd8ed1ab_0
341
+ - shapely=1.8.5=py310h5b266fc_2
342
+ - shellingham=1.5.4=pyhd8ed1ab_0
343
+ - smart_open=6.4.0=pyhd8ed1ab_0
344
+ - snappy=1.1.10=h9fff704_0
345
+ - sniffio=1.3.0=pyhd8ed1ab_0
346
+ - sortedcontainers=2.4.0=pyhd8ed1ab_0
347
+ - soupsieve=2.5=pyhd8ed1ab_1
348
+ - spacy=3.7.2=py310hcb52e73_0
349
+ - spacy-legacy=3.0.12=pyhd8ed1ab_0
350
+ - spacy-loggers=1.0.5=pyhd8ed1ab_0
351
+ - spdlog=1.11.0=h9b3ece8_1
352
+ - sqlite=3.38.2=hc218d9a_0
353
+ - srsly=2.4.8=py310hc6cd4ac_1
354
+ - stack_data=0.6.2=pyhd8ed1ab_0
355
+ - tblib=3.0.0=pyhd8ed1ab_0
356
+ - terminado=0.18.0=pyh0d859eb_0
357
+ - thinc=8.2.2=py310hcb52e73_0
358
+ - tinycss2=1.2.1=pyhd8ed1ab_0
359
+ - tk=8.6.13=noxft_h4845f30_101
360
+ - toolz=0.12.0=pyhd8ed1ab_0
361
+ - tornado=6.3.3=py310h2372a71_1
362
+ - tqdm=4.66.1=pyhd8ed1ab_0
363
+ - traitlets=5.9.0=pyhd8ed1ab_0
364
+ - treelite=3.2.0=py310h1be96d9_0
365
+ - truststore=0.8.0=pyhd8ed1ab_0
366
+ - typer=0.9.0=pyhd8ed1ab_0
367
+ - types-python-dateutil=2.8.19.20240106=pyhd8ed1ab_0
368
+ - typing-extensions=4.9.0=hd8ed1ab_0
369
+ - typing_extensions=4.9.0=pyha770c72_0
370
+ - typing_utils=0.1.0=pyhd8ed1ab_0
371
+ - ucx=1.14.1=h195a15c_5
372
+ - ucx-proc=1.0.0=gpu
373
+ - ucx-py=0.33.00=py310_230809_gea1eb8f_0
374
+ - unicodedata2=15.1.0=py310h2372a71_0
375
+ - uri-template=1.3.0=pyhd8ed1ab_0
376
+ - wasabi=1.1.2=py310hff52083_0
377
+ - wcwidth=0.2.13=pyhd8ed1ab_0
378
+ - weasel=0.3.4=pyhd8ed1ab_0
379
+ - webcolors=1.13=pyhd8ed1ab_0
380
+ - webencodings=0.5.1=pyhd8ed1ab_2
381
+ - websocket-client=1.7.0=pyhd8ed1ab_0
382
+ - wheel=0.42.0=pyhd8ed1ab_0
383
+ - xorg-kbproto=1.0.7=h7f98852_1002
384
+ - xorg-libice=1.1.1=hd590300_0
385
+ - xorg-libsm=1.2.4=h7391055_0
386
+ - xorg-libx11=1.8.7=h8ee46fc_0
387
+ - xorg-libxau=1.0.11=hd590300_0
388
+ - xorg-libxdmcp=1.1.3=h7f98852_0
389
+ - xorg-libxext=1.3.4=h0b41bf4_2
390
+ - xorg-libxrender=0.9.11=hd590300_0
391
+ - xorg-libxt=1.3.0=hd590300_1
392
+ - xorg-renderproto=0.11.1=h7f98852_1002
393
+ - xorg-xextproto=7.3.0=h0b41bf4_1003
394
+ - xorg-xproto=7.0.31=h7f98852_1007
395
+ - xyzservices=2023.10.1=pyhd8ed1ab_0
396
+ - xz=5.2.6=h166bdaf_0
397
+ - yaml=0.2.5=h7b6447c_0
398
+ - yaml-cpp=0.7.0=h59595ed_3
399
+ - zeromq=4.3.5=h59595ed_0
400
+ - zict=3.0.0=pyhd8ed1ab_0
401
+ - zipp=3.17.0=pyhd8ed1ab_0
402
+ - zlib=1.2.13=hd590300_5
403
+ - zstandard=0.22.0=py310h1275a96_0
404
+ - zstd=1.5.5=hfc55251_0
405
+ - pip:
406
+ - absl-py==1.4.0
407
+ - accelerate==0.27.2
408
+ - access==1.1.9
409
+ - affine==2.4.0
410
+ - aiobotocore==2.11.0
411
+ - aiofiles==22.1.0
412
+ - aiohttp-cors==0.7.0
413
+ - aioitertools==0.11.0
414
+ - aiorwlock==1.3.0
415
+ - aiosqlite==0.19.0
416
+ - albumentations==1.3.1
417
+ - alembic==1.13.1
418
+ - altair==5.2.0
419
+ - annoy==1.17.3
420
+ - apache-beam==2.46.0
421
+ - aplus==0.11.0
422
+ - appdirs==1.4.4
423
+ - array-record==0.5.0
424
+ - arviz==0.17.0
425
+ - astroid==3.0.2
426
+ - astropy==6.0.0
427
+ - astropy-iers-data==0.2024.1.22.0.30.30
428
+ - astunparse==1.6.3
429
+ - async-lru==2.0.4
430
+ - audioread==3.0.1
431
+ - autopep8==2.0.4
432
+ - babel==2.14.0
433
+ - backoff==2.2.1
434
+ - bayesian-optimization==1.4.3
435
+ - bayespy==0.5.26
436
+ - beatrix-jupyterlab==2023.128.151533
437
+ - bidict==0.22.1
438
+ - biopython==1.83
439
+ - blake3==0.2.1
440
+ - blessed==1.20.0
441
+ - blinker==1.7.0
442
+ - blosc2==2.4.0
443
+ - boruta==0.3
444
+ - boto3==1.26.100
445
+ - botocore==1.34.22
446
+ - bqplot==0.12.42
447
+ - branca==0.7.0
448
+ - brewer2mpl==1.4.1
449
+ - cachetools==4.2.4
450
+ - catalyst==22.4
451
+ - catboost==1.2.2
452
+ - category-encoders==2.6.3
453
+ - cesium==0.12.1
454
+ - chex==0.1.85
455
+ - cleverhans==4.0.0
456
+ - click-plugins==1.1.1
457
+ - cligj==0.7.2
458
+ - cloud-tpu-client==0.10
459
+ - cloud-tpu-profiler==2.4.0
460
+ - cloudpickle==2.2.1
461
+ - cmdstanpy==1.2.0
462
+ - cmudict==1.0.16
463
+ - colorcet==3.0.1
464
+ - colorful==0.5.6
465
+ - colorlog==6.8.0
466
+ - colorlover==0.3.0
467
+ - contextily==1.5.0
468
+ - convertdate==2.4.0
469
+ - crcmod==1.7
470
+ - cufflinks==0.17.3
471
+ - cvxcanon==0.1.2
472
+ - cysignals==1.11.4
473
+ - cython==3.0.8
474
+ - daal==2024.1.0
475
+ - daal4py==2024.1.0
476
+ - dacite==1.8.1
477
+ - dask==2024.1.0
478
+ - dataclasses-json==0.6.3
479
+ - dataproc-jupyter-plugin==0.1.66
480
+ - datasets==2.1.0
481
+ - datashader==0.16.0
482
+ - datatile==1.0.3
483
+ - db-dtypes==1.2.0
484
+ - deap==1.4.1
485
+ - deepdiff==6.7.1
486
+ - delorean==1.0.0
487
+ - deprecated==1.2.14
488
+ - deprecation==2.1.0
489
+ - descartes==1.1.0
490
+ - dill==0.3.7
491
+ - dipy==1.8.0
492
+ - distlib==0.3.8
493
+ - distributed==2024.1.0
494
+ - dm-tree==0.1.8
495
+ - docker==7.0.0
496
+ - docker-pycreds==0.4.0
497
+ - docopt==0.6.2
498
+ - docstring-parser==0.15
499
+ - docstring-to-markdown==0.13
500
+ - docutils==0.20.1
501
+ - earthengine-api==0.1.386
502
+ - easydict==1.11
503
+ - easyocr==1.7.1
504
+ - ecos==2.0.12
505
+ - eli5==0.13.0
506
+ - emoji==1.4.2
507
+ - en-core-web-lg==3.7.1
508
+ - en-core-web-sm==3.7.1
509
+ - ephem==4.1.5
510
+ - esda==2.5.1
511
+ - essentia==2.1b6.dev1110
512
+ - et-xmlfile==1.1.0
513
+ - etils==1.6.0
514
+ - evaluate==0.4.1
515
+ - explainable-ai-sdk==1.3.3
516
+ - farama-notifications==0.0.4
517
+ - fastai==2.7.13
518
+ - fastapi==0.108.0
519
+ - fastavro==1.9.3
520
+ - fastcore==1.5.29
521
+ - fastdownload==0.0.7
522
+ - fasteners==0.19
523
+ - fastprogress==1.0.3
524
+ - fasttext==0.9.2
525
+ - fbpca==1.0
526
+ - feather-format==0.4.1
527
+ - featuretools==1.28.0
528
+ - filelock==3.13.1
529
+ - fiona==1.9.5
530
+ - fitter==1.7.0
531
+ - flake8==7.0.0
532
+ - flashtext==2.7
533
+ - flask==3.0.1
534
+ - flatbuffers==23.5.26
535
+ - flax==0.7.5
536
+ - folium==0.15.1
537
+ - fonttools==4.47.0
538
+ - frozendict==2.4.0
539
+ - ftfy==6.0.3
540
+ - funcy==2.0
541
+ - fury==0.9.0
542
+ - future==0.18.3
543
+ - fuzzywuzzy==0.18.0
544
+ - gast==0.5.4
545
+ - gatspy==0.3
546
+ - gcsfs==2023.12.2.post1
547
+ - gensim==4.3.2
548
+ - geographiclib==2.0
549
+ - geohash==1.0
550
+ - geojson==3.1.0
551
+ - geopandas==0.14.2
552
+ - geoplot==0.5.1
553
+ - geopy==2.4.1
554
+ - geoviews==1.11.0
555
+ - ggplot==0.11.5
556
+ - giddy==2.3.5
557
+ - gitdb==4.0.11
558
+ - gitpython==3.1.41
559
+ - google-ai-generativelanguage==0.4.0
560
+ - google-api-core==2.15.0
561
+ - google-api-python-client==2.115.0
562
+ - google-apitools==0.5.31
563
+ - google-auth-httplib2==0.1.1
564
+ - google-auth-oauthlib==1.2.0
565
+ - google-cloud-aiplatform==0.6.0a1
566
+ - google-cloud-artifact-registry==1.10.0
567
+ - google-cloud-automl==1.0.1
568
+ - google-cloud-bigquery==2.34.4
569
+ - google-cloud-bigtable==1.7.3
570
+ - google-cloud-dlp==3.14.0
571
+ - google-cloud-jupyter-config==0.0.5
572
+ - google-cloud-language==2.12.0
573
+ - google-cloud-monitoring==2.18.0
574
+ - google-cloud-pubsub==2.19.0
575
+ - google-cloud-pubsublite==1.9.0
576
+ - google-cloud-recommendations-ai==0.7.1
577
+ - google-cloud-resource-manager==1.11.0
578
+ - google-cloud-spanner==3.40.1
579
+ - google-cloud-storage==1.44.0
580
+ - google-cloud-translate==3.12.1
581
+ - google-cloud-videointelligence==2.12.0
582
+ - google-cloud-vision==2.8.0
583
+ - google-crc32c==1.5.0
584
+ - google-generativeai==0.3.2
585
+ - google-pasta==0.2.0
586
+ - google-resumable-media==2.7.0
587
+ - gplearn==0.4.2
588
+ - gpustat==1.0.0
589
+ - gpxpy==1.6.2
590
+ - greenlet==3.0.3
591
+ - grpc-google-iam-v1==0.12.7
592
+ - grpcio==1.60.0
593
+ - grpcio-status==1.48.2
594
+ - gviz-api==1.10.0
595
+ - gym==0.26.2
596
+ - gym-notices==0.0.8
597
+ - gymnasium==0.29.0
598
+ - h11==0.14.0
599
+ - h2o==3.44.0.3
600
+ - h5netcdf==1.3.0
601
+ - h5py==3.10.0
602
+ - haversine==2.8.1
603
+ - hdfs==2.7.3
604
+ - hep-ml==0.7.2
605
+ - hijri-converter==2.3.1
606
+ - hmmlearn==0.3.0
607
+ - holidays==0.24
608
+ - holoviews==1.18.1
609
+ - hpsklearn==0.1.0
610
+ - html5lib==1.1
611
+ - htmlmin==0.1.12
612
+ - httplib2==0.21.0
613
+ - httptools==0.6.1
614
+ - huggingface-hub==0.20.3
615
+ - humanize==4.9.0
616
+ - hunspell==0.5.5
617
+ - husl==4.0.3
618
+ - hydra-slayer==0.5.0
619
+ - hyperopt==0.2.7
620
+ - hypertools==0.8.0
621
+ - igraph==0.11.3
622
+ - imagecodecs==2024.1.1
623
+ - imagehash==4.3.1
624
+ - imageio==2.33.1
625
+ - imbalanced-learn==0.11.0
626
+ - imgaug==0.4.0
627
+ - importlib-metadata==6.11.0
628
+ - inequality==1.0.1
629
+ - iniconfig==2.0.0
630
+ - ipydatawidgets==4.3.5
631
+ - ipyleaflet==0.18.2
632
+ - ipympl==0.7.0
633
+ - ipython-genutils==0.2.0
634
+ - ipython-sql==0.5.0
635
+ - ipyvolume==0.6.3
636
+ - ipyvue==1.10.1
637
+ - ipyvuetify==1.8.10
638
+ - ipywebrtc==0.6.0
639
+ - ipywidgets==7.7.1
640
+ - isort==5.13.2
641
+ - isoweek==1.3.3
642
+ - itsdangerous==2.1.2
643
+ - janome==0.5.0
644
+ - jaraco-classes==3.3.0
645
+ - jax==0.4.23
646
+ - jax-jumpy==1.0.0
647
+ - jaxlib==0.4.23.dev20240116
648
+ - jeepney==0.8.0
649
+ - jieba==0.42.1
650
+ - jmespath==1.0.1
651
+ - json5==0.9.14
652
+ - jupyter-client==7.4.9
653
+ - jupyter-console==6.6.3
654
+ - jupyter-http-over-ws==0.0.8
655
+ - jupyter-lsp==1.5.1
656
+ - jupyter-server==2.12.5
657
+ - jupyter-server-fileid==0.9.1
658
+ - jupyter-server-mathjax==0.2.6
659
+ - jupyter-server-proxy==4.1.0
660
+ - jupyter-server-ydoc==0.8.0
661
+ - jupyter-ydoc==0.2.5
662
+ - jupyterlab==4.0.11
663
+ - jupyterlab-git==0.44.0
664
+ - jupyterlab-lsp==5.0.2
665
+ - jupyterlab-server==2.25.2
666
+ - jupyterlab-widgets==3.0.9
667
+ - jupytext==1.16.0
668
+ - kaggle==1.6.3
669
+ - kaggle-environments==1.14.3
670
+ - kagglehub==0.1.6
671
+ - keras==2.15.0
672
+ - keras-core==0.1.7
673
+ - keras-cv==0.8.1
674
+ - keras-nlp==0.7.0
675
+ - keras-tuner==1.4.6
676
+ - kernels-mixer==0.0.7
677
+ - keyring==24.3.0
678
+ - keyrings-google-artifactregistry-auth==1.1.2
679
+ - kfp==2.5.0
680
+ - kfp-pipeline-spec==0.2.2
681
+ - kfp-server-api==2.0.5
682
+ - kmapper==2.0.1
683
+ - kmodes==0.12.2
684
+ - korean-lunar-calendar==0.3.1
685
+ - kornia==0.7.1
686
+ - kt-legacy==1.0.5
687
+ - kubernetes==26.1.0
688
+ - langid==1.1.6
689
+ - lazy-loader==0.3
690
+ - learntools==0.3.4
691
+ - leven==1.0.4
692
+ - levenshtein==0.23.0
693
+ - libclang==16.0.6
694
+ - libpysal==4.9.2
695
+ - librosa==0.10.1
696
+ - lightgbm==4.2.0
697
+ - lightning-utilities==0.10.1
698
+ - lime==0.2.0.1
699
+ - line-profiler==4.1.2
700
+ - linkify-it-py==2.0.2
701
+ - llvmlite==0.41.1
702
+ - lml==0.1.0
703
+ - loguru==0.7.2
704
+ - lunarcalendar==0.0.9
705
+ - lxml==5.1.0
706
+ - mako==1.3.1
707
+ - mapclassify==2.6.1
708
+ - marisa-trie==1.1.0
709
+ - markdown==3.5.2
710
+ - markovify==0.9.4
711
+ - markupsafe==2.1.4
712
+ - marshmallow==3.20.2
713
+ - matplotlib==3.7.4
714
+ - matplotlib-venn==0.11.9
715
+ - mccabe==0.7.0
716
+ - mdit-py-plugins==0.4.0
717
+ - memory-profiler==0.61.0
718
+ - mercantile==1.2.1
719
+ - mgwr==2.2.1
720
+ - missingno==0.5.2
721
+ - mistune==0.8.4
722
+ - mizani==0.9.3
723
+ - ml-dtypes==0.2.0
724
+ - mlcrate==0.2.0
725
+ - mlens==0.2.3
726
+ - mlxtend==0.23.1
727
+ - mmh3==4.1.0
728
+ - mne==1.6.1
729
+ - mnist==0.2.2
730
+ - mock==5.1.0
731
+ - momepy==0.7.0
732
+ - more-itertools==10.2.0
733
+ - mpld3==0.5.10
734
+ - mpmath==1.3.0
735
+ - msgpack-numpy==0.4.8
736
+ - multimethod==1.10
737
+ - multipledispatch==1.0.0
738
+ - multiprocess==0.70.15
739
+ - mypy-extensions==1.0.0
740
+ - namex==0.0.7
741
+ - nbclient==0.5.13
742
+ - nbconvert==6.4.5
743
+ - nbdime==3.2.0
744
+ - ndindex==1.7
745
+ - networkx==3.2.1
746
+ - nibabel==5.2.0
747
+ - nilearn==0.10.2
748
+ - ninja==1.11.1.1
749
+ - nltk==3.2.4
750
+ - normalizer==0.0.1
751
+ - nose==1.3.7
752
+ - notebook==6.5.6
753
+ - notebook-executor==0.2
754
+ - numexpr==2.8.8
755
+ - numpy==1.24.4
756
+ - nvidia-cublas-cu11==11.10.3.66
757
+ - nvidia-cuda-nvrtc-cu11==11.7.99
758
+ - nvidia-cuda-runtime-cu11==11.7.99
759
+ - nvidia-cudnn-cu11==8.5.0.96
760
+ - nvidia-ml-py==11.495.46
761
+ - oauth2client==4.1.3
762
+ - oauthlib==3.2.2
763
+ - objsize==0.6.1
764
+ - odfpy==1.4.1
765
+ - olefile==0.47
766
+ - onnx==1.15.0
767
+ - opencensus==0.11.4
768
+ - opencensus-context==0.1.3
769
+ - opencv-contrib-python==4.9.0.80
770
+ - opencv-python==4.9.0.80
771
+ - opencv-python-headless==4.9.0.80
772
+ - openpyxl==3.1.2
773
+ - openslide-python==1.3.1
774
+ - opentelemetry-api==1.22.0
775
+ - opentelemetry-exporter-otlp==1.22.0
776
+ - opentelemetry-exporter-otlp-proto-common==1.22.0
777
+ - opentelemetry-exporter-otlp-proto-grpc==1.22.0
778
+ - opentelemetry-exporter-otlp-proto-http==1.22.0
779
+ - opentelemetry-proto==1.22.0
780
+ - opentelemetry-sdk==1.22.0
781
+ - opentelemetry-semantic-conventions==0.43b0
782
+ - opt-einsum==3.3.0
783
+ - optax==0.1.8
784
+ - optuna==3.5.0
785
+ - orbax-checkpoint==0.4.4
786
+ - ordered-set==4.1.0
787
+ - orderedmultidict==1.0.1
788
+ - orjson==3.9.10
789
+ - ortools==9.4.1874
790
+ - osmnx==1.8.1
791
+ - packaging==21.3
792
+ - pandas==2.2.0
793
+ - pandas-datareader==0.10.0
794
+ - pandas-profiling==3.6.6
795
+ - pandas-summary==0.2.0
796
+ - pandasql==0.7.3
797
+ - panel==1.3.7
798
+ - papermill==2.5.0
799
+ - param==2.0.2
800
+ - path==16.9.0
801
+ - path-py==12.5.0
802
+ - pathos==0.3.1
803
+ - patsy==0.5.6
804
+ - pdf2image==1.17.0
805
+ - pettingzoo==1.24.0
806
+ - pexpect==4.9.0
807
+ - phik==0.12.4
808
+ - pillow==9.5.0
809
+ - plotly==5.18.0
810
+ - plotly-express==0.4.1
811
+ - plotnine==0.12.4
812
+ - pointpats==2.4.0
813
+ - polars==0.20.5
814
+ - polyglot==16.7.4
815
+ - pooch==1.8.0
816
+ - portalocker==2.8.2
817
+ - pox==0.3.3
818
+ - ppca==0.0.4
819
+ - ppft==1.7.6.7
820
+ - preprocessing==0.1.13
821
+ - prettytable==3.9.0
822
+ - progressbar2==4.3.2
823
+ - promise==2.3
824
+ - prompt-toolkit==3.0.43
825
+ - pronouncing==0.2.0
826
+ - prophet==1.1.1
827
+ - protobuf==3.20.3
828
+ - psutil==5.9.3
829
+ - pudb==2024.1
830
+ - pulp==2.8.0
831
+ - py-cpuinfo==9.0.0
832
+ - py-spy==0.3.14
833
+ - py4j==0.10.9.7
834
+ - pyaml==23.12.0
835
+ - pyarabic==0.6.15
836
+ - pyastronomy==0.20.0
837
+ - pybind11==2.11.1
838
+ - pyclipper==1.3.0.post5
839
+ - pycodestyle==2.11.1
840
+ - pycparser==2.21
841
+ - pycryptodome==3.20.0
842
+ - pyct==0.5.0
843
+ - pycuda==2024.1
844
+ - pydegensac==0.1.2
845
+ - pydicom==2.4.4
846
+ - pydocstyle==6.3.0
847
+ - pydot==1.4.2
848
+ - pydub==0.25.1
849
+ - pyemd==1.0.0
850
+ - pyerfa==2.0.1.1
851
+ - pyexcel-io==0.6.6
852
+ - pyexcel-ods==0.6.0
853
+ - pyfasttext==0.4.6
854
+ - pyflakes==3.2.0
855
+ - pygltflib==1.16.1
856
+ - pyjwt==2.8.0
857
+ - pykalman==0.9.5
858
+ - pyldavis==3.4.1
859
+ - pylint==3.0.3
860
+ - pymc3==3.11.4
861
+ - pymeeus==0.5.12
862
+ - pymongo==3.13.0
863
+ - pympler==1.0.1
864
+ - pynndescent==0.5.11
865
+ - pynvrtc==9.2
866
+ - pyocr==0.8.5
867
+ - pypdf==4.0.0
868
+ - pysal==23.7
869
+ - pytesseract==0.3.10
870
+ - pytest==7.4.4
871
+ - python-bidi==0.4.2
872
+ - python-dotenv==1.0.0
873
+ - python-graphviz==0.20.1
874
+ - python-levenshtein==0.23.0
875
+ - python-louvain==0.16
876
+ - python-lsp-jsonrpc==1.1.2
877
+ - python-lsp-server==1.10.0
878
+ - python-slugify==8.0.1
879
+ - python-utils==3.8.1
880
+ - pythreejs==2.4.2
881
+ - pytoolconfig==1.3.1
882
+ - pytools==2023.1.1
883
+ - pytorch-ignite==0.4.13
884
+ - pytorch-lightning==2.1.3
885
+ - pyupset==0.1.1.post7
886
+ - pyviz-comms==3.0.1
887
+ - pywavelets==1.5.0
888
+ - pyzmq==24.0.1
889
+ - qgrid==1.3.1
890
+ - qtconsole==5.5.1
891
+ - qtpy==2.4.1
892
+ - quantecon==0.7.1
893
+ - quantities==0.15.0
894
+ - qudida==0.0.4
895
+ - rapidfuzz==3.6.1
896
+ - rasterio==1.3.9
897
+ - rasterstats==0.19.0
898
+ - ray==2.9.0
899
+ - ray-cpp==2.9.0
900
+ - regex==2023.12.25
901
+ - requests-oauthlib==1.3.1
902
+ - requests-toolbelt==0.10.1
903
+ - responses==0.18.0
904
+ - retrying==1.3.4
905
+ - rgf-python==3.12.0
906
+ - rich-click==1.7.3
907
+ - rope==1.12.0
908
+ - rtree==1.2.0
909
+ - s2sphere==0.2.5
910
+ - s3fs==2023.12.2
911
+ - s3transfer==0.6.2
912
+ - sacrebleu==2.4.0
913
+ - safetensors==0.4.2
914
+ - scattertext==0.1.19
915
+ - scikit-image==0.22.0
916
+ - scikit-learn==1.2.2
917
+ - scikit-learn-intelex==2024.1.0
918
+ - scikit-multilearn==0.2.0
919
+ - scikit-optimize==0.9.0
920
+ - scikit-plot==0.3.7
921
+ - scikit-surprise==1.1.3
922
+ - scipy==1.11.4
923
+ - seaborn==0.12.2
924
+ - secretstorage==3.3.3
925
+ - segment-anything==1.0
926
+ - segregation==2.5
927
+ - semver==3.0.2
928
+ - sentencepiece==0.1.99
929
+ - sentry-sdk==1.39.2
930
+ - setproctitle==1.3.3
931
+ - setuptools-git==1.2
932
+ - setuptools-scm==8.0.4
933
+ - shap==0.44.0
934
+ - shimmy==1.3.0
935
+ - simpervisor==1.0.0
936
+ - simpleitk==2.3.1
937
+ - simplejson==3.19.2
938
+ - six==1.16.0
939
+ - sklearn-pandas==2.2.0
940
+ - slicer==0.0.7
941
+ - smhasher==0.150.1
942
+ - smmap==5.0.1
943
+ - snowballstemmer==2.2.0
944
+ - snuggs==1.4.7
945
+ - soundfile==0.12.1
946
+ - soxr==0.3.7
947
+ - spaghetti==1.7.4
948
+ - spectral==0.23.1
949
+ - spglm==1.1.0
950
+ - sphinx-rtd-theme==0.2.4
951
+ - spint==1.0.7
952
+ - splot==1.1.5.post1
953
+ - spopt==0.6.0
954
+ - spreg==1.4.2
955
+ - spvcm==0.3.0
956
+ - sqlalchemy==2.0.25
957
+ - sqlparse==0.4.4
958
+ - squarify==0.4.3
959
+ - stable-baselines3==2.1.0
960
+ - stack-data==0.6.3
961
+ - stanio==0.3.0
962
+ - starlette==0.32.0.post1
963
+ - statsmodels==0.14.1
964
+ - stemming==1.0.1
965
+ - stop-words==2018.7.23
966
+ - stopit==1.1.2
967
+ - stumpy==1.12.0
968
+ - sympy==1.12
969
+ - tables==3.9.2
970
+ - tabulate==0.9.0
971
+ - tangled-up-in-unicode==0.2.0
972
+ - tbb==2021.11.0
973
+ - tenacity==8.2.3
974
+ - tensorboard==2.15.1
975
+ - tensorboard-data-server==0.7.2
976
+ - tensorboard-plugin-profile==2.15.0
977
+ - tensorboardx==2.6.2.2
978
+ - tensorflow==2.15.0
979
+ - tensorflow-cloud==0.1.16
980
+ - tensorflow-datasets==4.9.4
981
+ - tensorflow-decision-forests==1.8.1
982
+ - tensorflow-estimator==2.15.0
983
+ - tensorflow-hub==0.15.0
984
+ - tensorflow-io==0.35.0
985
+ - tensorflow-io-gcs-filesystem==0.35.0
986
+ - tensorflow-metadata==0.14.0
987
+ - tensorflow-probability==0.23.0
988
+ - tensorflow-serving-api==2.14.1
989
+ - tensorflow-text==2.15.0
990
+ - tensorflow-transform==0.14.0
991
+ - tensorflowjs==4.16.0
992
+ - tensorpack==0.11
993
+ - tensorstore==0.1.45
994
+ - termcolor==2.4.0
995
+ - testpath==0.6.0
996
+ - text-unidecode==1.3
997
+ - textblob==0.17.1
998
+ - texttable==1.7.0
999
+ - theano==1.0.5
1000
+ - theano-pymc==1.1.2
1001
+ - threadpoolctl==3.2.0
1002
+ - tifffile==2023.12.9
1003
+ - timm==0.9.12
1004
+ - tobler==0.11.2
1005
+ - tokenizers==0.15.1
1006
+ - toml==0.10.2
1007
+ - tomli==2.0.1
1008
+ - tomlkit==0.12.3
1009
+ - torch==1.13.1
1010
+ - torchaudio==2.1.2
1011
+ - torchdata==0.7.1
1012
+ - torchinfo==1.8.0
1013
+ - torchmetrics==1.3.0.post0
1014
+ - torchtext==0.16.2
1015
+ - torchvision==0.16.2
1016
+ - tpot==0.12.1
1017
+ - traceml==1.0.8
1018
+ - traittypes==0.2.1
1019
+ - transformers==4.37.0
1020
+ - treelite-runtime==3.2.0
1021
+ - trueskill==0.4.5
1022
+ - trx-python==0.2.9
1023
+ - tsfresh==0.20.1
1024
+ - typeguard==4.1.5
1025
+ - typing-inspect==0.9.0
1026
+ - tzdata==2023.4
1027
+ - tzlocal==5.2
1028
+ - uc-micro-py==1.0.2
1029
+ - ujson==5.9.0
1030
+ - umap-learn==0.5.5
1031
+ - unidecode==1.3.8
1032
+ - update-checker==0.18.0
1033
+ - uritemplate==3.0.1
1034
+ - urllib3==1.26.18
1035
+ - urwid==2.4.6
1036
+ - urwid-readline==0.13
1037
+ - uvicorn==0.25.0
1038
+ - uvloop==0.19.0
1039
+ - vaex==4.17.0
1040
+ - vaex-astro==0.9.3
1041
+ - vaex-core==4.17.1
1042
+ - vaex-hdf5==0.14.1
1043
+ - vaex-jupyter==0.8.2
1044
+ - vaex-ml==0.18.3
1045
+ - vaex-server==0.9.0
1046
+ - vaex-viz==0.5.4
1047
+ - vec-noise==1.1.4
1048
+ - vecstack==0.4.0
1049
+ - virtualenv==20.21.0
1050
+ - visions==0.7.5
1051
+ - vowpalwabbit==9.9.0
1052
+ - vtk==9.3.0
1053
+ - wand==0.6.13
1054
+ - wandb==0.16.2
1055
+ - watchfiles==0.21.0
1056
+ - wavio==0.0.8
1057
+ - websockets==12.0
1058
+ - werkzeug==3.0.1
1059
+ - wfdb==4.1.2
1060
+ - whatthepatch==1.0.5
1061
+ - widgetsnbextension==3.6.6
1062
+ - witwidget==1.8.1
1063
+ - woodwork==0.27.0
1064
+ - wordcloud==1.9.3
1065
+ - wordsegment==1.3.1
1066
+ - wrapt==1.14.1
1067
+ - wurlitzer==3.0.3
1068
+ - xarray==2024.1.0
1069
+ - xarray-einstats==0.7.0
1070
+ - xgboost==2.0.3
1071
+ - xvfbwrapper==0.2.9
1072
+ - xxhash==3.4.1
1073
+ - y-py==0.6.2
1074
+ - yapf==0.40.2
1075
+ - yarl==1.9.4
1076
+ - ydata-profiling==4.6.4
1077
+ - yellowbrick==1.5
1078
+ - ypy-websocket==0.8.4
1079
+ prefix: /opt/conda
wandb/run-20240226_092105-91uhs91a/files/config.yaml ADDED
@@ -0,0 +1,695 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.10.13
7
+ cli_version: 0.16.2
8
+ framework: huggingface
9
+ huggingface_version: 4.37.0
10
+ is_jupyter_run: true
11
+ is_kaggle_kernel: true
12
+ start_time: 1708939265.532258
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 2
17
+ - 3
18
+ - 5
19
+ - 11
20
+ - 12
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 100
27
+ - 105
28
+ 2:
29
+ - 1
30
+ - 2
31
+ - 3
32
+ - 5
33
+ - 11
34
+ - 12
35
+ - 49
36
+ - 51
37
+ - 53
38
+ - 55
39
+ - 71
40
+ - 100
41
+ - 105
42
+ 3:
43
+ - 7
44
+ - 23
45
+ 4: 3.10.13
46
+ 5: 0.16.2
47
+ 6: 4.37.0
48
+ 8:
49
+ - 1
50
+ - 2
51
+ - 5
52
+ 9:
53
+ 1: transformers_trainer
54
+ 13: linux-x86_64
55
+ m:
56
+ - 1: train/global_step
57
+ 6:
58
+ - 3
59
+ - 1: eval/loss
60
+ 5: 1
61
+ 6:
62
+ - 1
63
+ - 1: eval/bleu
64
+ 5: 1
65
+ 6:
66
+ - 1
67
+ - 1: eval/gen_len
68
+ 5: 1
69
+ 6:
70
+ - 1
71
+ - 1: eval/runtime
72
+ 5: 1
73
+ 6:
74
+ - 1
75
+ - 1: eval/samples_per_second
76
+ 5: 1
77
+ 6:
78
+ - 1
79
+ - 1: eval/steps_per_second
80
+ 5: 1
81
+ 6:
82
+ - 1
83
+ - 1: train/epoch
84
+ 5: 1
85
+ 6:
86
+ - 1
87
+ - 1: train/loss
88
+ 5: 1
89
+ 6:
90
+ - 1
91
+ - 1: train/learning_rate
92
+ 5: 1
93
+ 6:
94
+ - 1
95
+ vocab_size:
96
+ desc: null
97
+ value: 32128
98
+ d_model:
99
+ desc: null
100
+ value: 768
101
+ d_kv:
102
+ desc: null
103
+ value: 64
104
+ d_ff:
105
+ desc: null
106
+ value: 2048
107
+ num_layers:
108
+ desc: null
109
+ value: 12
110
+ num_decoder_layers:
111
+ desc: null
112
+ value: 12
113
+ num_heads:
114
+ desc: null
115
+ value: 12
116
+ relative_attention_num_buckets:
117
+ desc: null
118
+ value: 32
119
+ relative_attention_max_distance:
120
+ desc: null
121
+ value: 128
122
+ dropout_rate:
123
+ desc: null
124
+ value: 0.1
125
+ classifier_dropout:
126
+ desc: null
127
+ value: 0.0
128
+ layer_norm_epsilon:
129
+ desc: null
130
+ value: 1.0e-06
131
+ initializer_factor:
132
+ desc: null
133
+ value: 1.0
134
+ feed_forward_proj:
135
+ desc: null
136
+ value: gated-gelu
137
+ use_cache:
138
+ desc: null
139
+ value: true
140
+ dense_act_fn:
141
+ desc: null
142
+ value: gelu_new
143
+ is_gated_act:
144
+ desc: null
145
+ value: true
146
+ return_dict:
147
+ desc: null
148
+ value: true
149
+ output_hidden_states:
150
+ desc: null
151
+ value: false
152
+ output_attentions:
153
+ desc: null
154
+ value: false
155
+ torchscript:
156
+ desc: null
157
+ value: false
158
+ torch_dtype:
159
+ desc: null
160
+ value: float32
161
+ use_bfloat16:
162
+ desc: null
163
+ value: false
164
+ tf_legacy_loss:
165
+ desc: null
166
+ value: false
167
+ pruned_heads:
168
+ desc: null
169
+ value: {}
170
+ tie_word_embeddings:
171
+ desc: null
172
+ value: false
173
+ chunk_size_feed_forward:
174
+ desc: null
175
+ value: 0
176
+ is_encoder_decoder:
177
+ desc: null
178
+ value: true
179
+ is_decoder:
180
+ desc: null
181
+ value: false
182
+ cross_attention_hidden_size:
183
+ desc: null
184
+ value: null
185
+ add_cross_attention:
186
+ desc: null
187
+ value: false
188
+ tie_encoder_decoder:
189
+ desc: null
190
+ value: false
191
+ max_length:
192
+ desc: null
193
+ value: 20
194
+ min_length:
195
+ desc: null
196
+ value: 0
197
+ do_sample:
198
+ desc: null
199
+ value: false
200
+ early_stopping:
201
+ desc: null
202
+ value: false
203
+ num_beams:
204
+ desc: null
205
+ value: 1
206
+ num_beam_groups:
207
+ desc: null
208
+ value: 1
209
+ diversity_penalty:
210
+ desc: null
211
+ value: 0.0
212
+ temperature:
213
+ desc: null
214
+ value: 1.0
215
+ top_k:
216
+ desc: null
217
+ value: 50
218
+ top_p:
219
+ desc: null
220
+ value: 1.0
221
+ typical_p:
222
+ desc: null
223
+ value: 1.0
224
+ repetition_penalty:
225
+ desc: null
226
+ value: 1.0
227
+ length_penalty:
228
+ desc: null
229
+ value: 1.0
230
+ no_repeat_ngram_size:
231
+ desc: null
232
+ value: 0
233
+ encoder_no_repeat_ngram_size:
234
+ desc: null
235
+ value: 0
236
+ bad_words_ids:
237
+ desc: null
238
+ value: null
239
+ num_return_sequences:
240
+ desc: null
241
+ value: 1
242
+ output_scores:
243
+ desc: null
244
+ value: false
245
+ return_dict_in_generate:
246
+ desc: null
247
+ value: false
248
+ forced_bos_token_id:
249
+ desc: null
250
+ value: null
251
+ forced_eos_token_id:
252
+ desc: null
253
+ value: null
254
+ remove_invalid_values:
255
+ desc: null
256
+ value: false
257
+ exponential_decay_length_penalty:
258
+ desc: null
259
+ value: null
260
+ suppress_tokens:
261
+ desc: null
262
+ value: null
263
+ begin_suppress_tokens:
264
+ desc: null
265
+ value: null
266
+ architectures:
267
+ desc: null
268
+ value:
269
+ - T5ForConditionalGeneration
270
+ finetuning_task:
271
+ desc: null
272
+ value: null
273
+ id2label:
274
+ desc: null
275
+ value:
276
+ '0': LABEL_0
277
+ '1': LABEL_1
278
+ label2id:
279
+ desc: null
280
+ value:
281
+ LABEL_0: 0
282
+ LABEL_1: 1
283
+ tokenizer_class:
284
+ desc: null
285
+ value: null
286
+ prefix:
287
+ desc: null
288
+ value: null
289
+ bos_token_id:
290
+ desc: null
291
+ value: null
292
+ pad_token_id:
293
+ desc: null
294
+ value: 0
295
+ eos_token_id:
296
+ desc: null
297
+ value: 1
298
+ sep_token_id:
299
+ desc: null
300
+ value: null
301
+ decoder_start_token_id:
302
+ desc: null
303
+ value: 0
304
+ task_specific_params:
305
+ desc: null
306
+ value: null
307
+ problem_type:
308
+ desc: null
309
+ value: null
310
+ _name_or_path:
311
+ desc: null
312
+ value: csebuetnlp/banglat5
313
+ transformers_version:
314
+ desc: null
315
+ value: 4.37.0
316
+ gradient_checkpointing:
317
+ desc: null
318
+ value: false
319
+ model_type:
320
+ desc: null
321
+ value: t5
322
+ output_past:
323
+ desc: null
324
+ value: true
325
+ output_dir:
326
+ desc: null
327
+ value: /kaggle/working/
328
+ overwrite_output_dir:
329
+ desc: null
330
+ value: false
331
+ do_train:
332
+ desc: null
333
+ value: false
334
+ do_eval:
335
+ desc: null
336
+ value: true
337
+ do_predict:
338
+ desc: null
339
+ value: false
340
+ evaluation_strategy:
341
+ desc: null
342
+ value: epoch
343
+ prediction_loss_only:
344
+ desc: null
345
+ value: false
346
+ per_device_train_batch_size:
347
+ desc: null
348
+ value: 16
349
+ per_device_eval_batch_size:
350
+ desc: null
351
+ value: 16
352
+ per_gpu_train_batch_size:
353
+ desc: null
354
+ value: null
355
+ per_gpu_eval_batch_size:
356
+ desc: null
357
+ value: null
358
+ gradient_accumulation_steps:
359
+ desc: null
360
+ value: 1
361
+ eval_accumulation_steps:
362
+ desc: null
363
+ value: null
364
+ eval_delay:
365
+ desc: null
366
+ value: 0
367
+ learning_rate:
368
+ desc: null
369
+ value: 0.0005
370
+ weight_decay:
371
+ desc: null
372
+ value: 0.1
373
+ adam_beta1:
374
+ desc: null
375
+ value: 0.9
376
+ adam_beta2:
377
+ desc: null
378
+ value: 0.999
379
+ adam_epsilon:
380
+ desc: null
381
+ value: 1.0e-08
382
+ max_grad_norm:
383
+ desc: null
384
+ value: 1.0
385
+ num_train_epochs:
386
+ desc: null
387
+ value: 5
388
+ max_steps:
389
+ desc: null
390
+ value: -1
391
+ lr_scheduler_type:
392
+ desc: null
393
+ value: linear
394
+ lr_scheduler_kwargs:
395
+ desc: null
396
+ value: {}
397
+ warmup_ratio:
398
+ desc: null
399
+ value: 0.0
400
+ warmup_steps:
401
+ desc: null
402
+ value: 0
403
+ log_level:
404
+ desc: null
405
+ value: passive
406
+ log_level_replica:
407
+ desc: null
408
+ value: warning
409
+ log_on_each_node:
410
+ desc: null
411
+ value: true
412
+ logging_dir:
413
+ desc: null
414
+ value: /kaggle/working/runs/Feb26_09-23-32_d79522c38279
415
+ logging_strategy:
416
+ desc: null
417
+ value: steps
418
+ logging_first_step:
419
+ desc: null
420
+ value: false
421
+ logging_steps:
422
+ desc: null
423
+ value: 500
424
+ logging_nan_inf_filter:
425
+ desc: null
426
+ value: true
427
+ save_strategy:
428
+ desc: null
429
+ value: steps
430
+ save_steps:
431
+ desc: null
432
+ value: 500
433
+ save_total_limit:
434
+ desc: null
435
+ value: 3
436
+ save_safetensors:
437
+ desc: null
438
+ value: true
439
+ save_on_each_node:
440
+ desc: null
441
+ value: false
442
+ save_only_model:
443
+ desc: null
444
+ value: false
445
+ no_cuda:
446
+ desc: null
447
+ value: false
448
+ use_cpu:
449
+ desc: null
450
+ value: false
451
+ use_mps_device:
452
+ desc: null
453
+ value: false
454
+ seed:
455
+ desc: null
456
+ value: 42
457
+ data_seed:
458
+ desc: null
459
+ value: null
460
+ jit_mode_eval:
461
+ desc: null
462
+ value: false
463
+ use_ipex:
464
+ desc: null
465
+ value: false
466
+ bf16:
467
+ desc: null
468
+ value: false
469
+ fp16:
470
+ desc: null
471
+ value: false
472
+ fp16_opt_level:
473
+ desc: null
474
+ value: O1
475
+ half_precision_backend:
476
+ desc: null
477
+ value: auto
478
+ bf16_full_eval:
479
+ desc: null
480
+ value: false
481
+ fp16_full_eval:
482
+ desc: null
483
+ value: false
484
+ tf32:
485
+ desc: null
486
+ value: null
487
+ local_rank:
488
+ desc: null
489
+ value: 0
490
+ ddp_backend:
491
+ desc: null
492
+ value: null
493
+ tpu_num_cores:
494
+ desc: null
495
+ value: null
496
+ tpu_metrics_debug:
497
+ desc: null
498
+ value: false
499
+ debug:
500
+ desc: null
501
+ value: []
502
+ dataloader_drop_last:
503
+ desc: null
504
+ value: false
505
+ eval_steps:
506
+ desc: null
507
+ value: null
508
+ dataloader_num_workers:
509
+ desc: null
510
+ value: 0
511
+ past_index:
512
+ desc: null
513
+ value: -1
514
+ run_name:
515
+ desc: null
516
+ value: /kaggle/working/
517
+ disable_tqdm:
518
+ desc: null
519
+ value: false
520
+ remove_unused_columns:
521
+ desc: null
522
+ value: true
523
+ label_names:
524
+ desc: null
525
+ value: null
526
+ load_best_model_at_end:
527
+ desc: null
528
+ value: false
529
+ metric_for_best_model:
530
+ desc: null
531
+ value: null
532
+ greater_is_better:
533
+ desc: null
534
+ value: null
535
+ ignore_data_skip:
536
+ desc: null
537
+ value: false
538
+ fsdp:
539
+ desc: null
540
+ value: []
541
+ fsdp_min_num_params:
542
+ desc: null
543
+ value: 0
544
+ fsdp_config:
545
+ desc: null
546
+ value:
547
+ min_num_params: 0
548
+ xla: false
549
+ xla_fsdp_grad_ckpt: false
550
+ fsdp_transformer_layer_cls_to_wrap:
551
+ desc: null
552
+ value: null
553
+ deepspeed:
554
+ desc: null
555
+ value: null
556
+ label_smoothing_factor:
557
+ desc: null
558
+ value: 0.0
559
+ optim:
560
+ desc: null
561
+ value: adamw_torch
562
+ optim_args:
563
+ desc: null
564
+ value: null
565
+ adafactor:
566
+ desc: null
567
+ value: false
568
+ group_by_length:
569
+ desc: null
570
+ value: false
571
+ length_column_name:
572
+ desc: null
573
+ value: length
574
+ report_to:
575
+ desc: null
576
+ value:
577
+ - tensorboard
578
+ - wandb
579
+ ddp_find_unused_parameters:
580
+ desc: null
581
+ value: null
582
+ ddp_bucket_cap_mb:
583
+ desc: null
584
+ value: null
585
+ ddp_broadcast_buffers:
586
+ desc: null
587
+ value: null
588
+ dataloader_pin_memory:
589
+ desc: null
590
+ value: true
591
+ dataloader_persistent_workers:
592
+ desc: null
593
+ value: false
594
+ skip_memory_metrics:
595
+ desc: null
596
+ value: true
597
+ use_legacy_prediction_loop:
598
+ desc: null
599
+ value: false
600
+ push_to_hub:
601
+ desc: null
602
+ value: true
603
+ resume_from_checkpoint:
604
+ desc: null
605
+ value: null
606
+ hub_model_id:
607
+ desc: null
608
+ value: null
609
+ hub_strategy:
610
+ desc: null
611
+ value: every_save
612
+ hub_token:
613
+ desc: null
614
+ value: <HUB_TOKEN>
615
+ hub_private_repo:
616
+ desc: null
617
+ value: false
618
+ hub_always_push:
619
+ desc: null
620
+ value: false
621
+ gradient_checkpointing_kwargs:
622
+ desc: null
623
+ value: null
624
+ include_inputs_for_metrics:
625
+ desc: null
626
+ value: false
627
+ fp16_backend:
628
+ desc: null
629
+ value: auto
630
+ push_to_hub_model_id:
631
+ desc: null
632
+ value: null
633
+ push_to_hub_organization:
634
+ desc: null
635
+ value: null
636
+ push_to_hub_token:
637
+ desc: null
638
+ value: <PUSH_TO_HUB_TOKEN>
639
+ mp_parameters:
640
+ desc: null
641
+ value: ''
642
+ auto_find_batch_size:
643
+ desc: null
644
+ value: false
645
+ full_determinism:
646
+ desc: null
647
+ value: false
648
+ torchdynamo:
649
+ desc: null
650
+ value: null
651
+ ray_scope:
652
+ desc: null
653
+ value: last
654
+ ddp_timeout:
655
+ desc: null
656
+ value: 1800
657
+ torch_compile:
658
+ desc: null
659
+ value: false
660
+ torch_compile_backend:
661
+ desc: null
662
+ value: null
663
+ torch_compile_mode:
664
+ desc: null
665
+ value: null
666
+ dispatch_batches:
667
+ desc: null
668
+ value: null
669
+ split_batches:
670
+ desc: null
671
+ value: false
672
+ include_tokens_per_second:
673
+ desc: null
674
+ value: false
675
+ include_num_input_tokens_seen:
676
+ desc: null
677
+ value: false
678
+ neftune_noise_alpha:
679
+ desc: null
680
+ value: null
681
+ sortish_sampler:
682
+ desc: null
683
+ value: false
684
+ predict_with_generate:
685
+ desc: null
686
+ value: true
687
+ generation_max_length:
688
+ desc: null
689
+ value: null
690
+ generation_num_beams:
691
+ desc: null
692
+ value: null
693
+ generation_config:
694
+ desc: null
695
+ value: null
wandb/run-20240226_092105-91uhs91a/files/output.log ADDED
@@ -0,0 +1 @@
 
 
1
+ /opt/conda/lib/python3.10/site-packages/transformers/generation/utils.py:1128: UserWarning: Using the model-agnostic default `max_length` (=20) to control the generation length. We recommend setting `max_new_tokens` to control the maximum length of the generation.
wandb/run-20240226_092105-91uhs91a/files/requirements.txt ADDED
@@ -0,0 +1,858 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ absl-py==1.4.0
2
+ accelerate==0.27.2
3
+ access==1.1.9
4
+ affine==2.4.0
5
+ aiobotocore==2.11.0
6
+ aiofiles==22.1.0
7
+ aiohttp-cors==0.7.0
8
+ aiohttp==3.9.1
9
+ aioitertools==0.11.0
10
+ aiorwlock==1.3.0
11
+ aiosignal==1.3.1
12
+ aiosqlite==0.19.0
13
+ albumentations==1.3.1
14
+ alembic==1.13.1
15
+ altair==5.2.0
16
+ annotated-types==0.6.0
17
+ annoy==1.17.3
18
+ anyio==4.2.0
19
+ apache-beam==2.46.0
20
+ aplus==0.11.0
21
+ appdirs==1.4.4
22
+ archspec==0.2.2
23
+ argon2-cffi-bindings==21.2.0
24
+ argon2-cffi==23.1.0
25
+ array-record==0.5.0
26
+ arrow==1.3.0
27
+ arviz==0.17.0
28
+ astroid==3.0.2
29
+ astropy-iers-data==0.2024.1.22.0.30.30
30
+ astropy==6.0.0
31
+ asttokens==2.4.1
32
+ astunparse==1.6.3
33
+ async-lru==2.0.4
34
+ async-timeout==4.0.3
35
+ attrs==23.2.0
36
+ audioread==3.0.1
37
+ autopep8==2.0.4
38
+ babel==2.14.0
39
+ backoff==2.2.1
40
+ bayesian-optimization==1.4.3
41
+ bayespy==0.5.26
42
+ beatrix-jupyterlab==2023.128.151533
43
+ beautifulsoup4==4.12.2
44
+ bidict==0.22.1
45
+ biopython==1.83
46
+ blake3==0.2.1
47
+ bleach==6.1.0
48
+ blessed==1.20.0
49
+ blinker==1.7.0
50
+ blis==0.7.10
51
+ blosc2==2.4.0
52
+ bokeh==3.3.3
53
+ boltons==23.1.1
54
+ boruta==0.3
55
+ boto3==1.26.100
56
+ botocore==1.34.22
57
+ bq-helper==0.4.1
58
+ bqplot==0.12.42
59
+ branca==0.7.0
60
+ brewer2mpl==1.4.1
61
+ brotli==1.0.9
62
+ brotlipy==0.7.0
63
+ cached-property==1.5.2
64
+ cachetools==4.2.4
65
+ cartopy==0.22.0
66
+ catalogue==2.0.10
67
+ catalyst==22.4
68
+ catboost==1.2.2
69
+ category-encoders==2.6.3
70
+ certifi==2023.11.17
71
+ cesium==0.12.1
72
+ cffi==1.16.0
73
+ charset-normalizer==3.3.2
74
+ chex==0.1.85
75
+ cleverhans==4.0.0
76
+ click-plugins==1.1.1
77
+ click==8.1.7
78
+ cligj==0.7.2
79
+ cloud-tpu-client==0.10
80
+ cloud-tpu-profiler==2.4.0
81
+ cloudpathlib==0.16.0
82
+ cloudpickle==2.2.1
83
+ cmdstanpy==1.2.0
84
+ cmudict==1.0.16
85
+ colorama==0.4.6
86
+ colorcet==3.0.1
87
+ colorful==0.5.6
88
+ colorlog==6.8.0
89
+ colorlover==0.3.0
90
+ comm==0.2.1
91
+ conda-libmamba-solver==23.7.0
92
+ conda-package-handling==2.2.0
93
+ conda-package-streaming==0.9.0
94
+ conda==23.7.4
95
+ confection==0.1.4
96
+ contextily==1.5.0
97
+ contourpy==1.2.0
98
+ convertdate==2.4.0
99
+ crcmod==1.7
100
+ cryptography==41.0.7
101
+ cuda-python==12.3.0
102
+ cudf==23.8.0
103
+ cufflinks==0.17.3
104
+ cuml==23.8.0
105
+ cupy==13.0.0
106
+ cvxcanon==0.1.2
107
+ cycler==0.12.1
108
+ cymem==2.0.8
109
+ cysignals==1.11.4
110
+ cython==3.0.8
111
+ cytoolz==0.12.2
112
+ daal4py==2024.1.0
113
+ daal==2024.1.0
114
+ dacite==1.8.1
115
+ dask-cuda==23.8.0
116
+ dask-cudf==23.8.0
117
+ dask==2024.1.0
118
+ dataclasses-json==0.6.3
119
+ dataproc-jupyter-plugin==0.1.66
120
+ datasets==2.1.0
121
+ datashader==0.16.0
122
+ datatile==1.0.3
123
+ db-dtypes==1.2.0
124
+ deap==1.4.1
125
+ debugpy==1.8.0
126
+ decorator==5.1.1
127
+ deepdiff==6.7.1
128
+ defusedxml==0.7.1
129
+ delorean==1.0.0
130
+ deprecated==1.2.14
131
+ deprecation==2.1.0
132
+ descartes==1.1.0
133
+ dill==0.3.7
134
+ dipy==1.8.0
135
+ distlib==0.3.8
136
+ distributed==2024.1.0
137
+ distro==1.9.0
138
+ dm-tree==0.1.8
139
+ docker-pycreds==0.4.0
140
+ docker==7.0.0
141
+ docopt==0.6.2
142
+ docstring-parser==0.15
143
+ docstring-to-markdown==0.13
144
+ docutils==0.20.1
145
+ earthengine-api==0.1.386
146
+ easydict==1.11
147
+ easyocr==1.7.1
148
+ ecos==2.0.12
149
+ eli5==0.13.0
150
+ emoji==1.4.2
151
+ en-core-web-lg==3.7.1
152
+ en-core-web-sm==3.7.1
153
+ entrypoints==0.4
154
+ ephem==4.1.5
155
+ esda==2.5.1
156
+ essentia==2.1b6.dev1110
157
+ et-xmlfile==1.1.0
158
+ etils==1.6.0
159
+ evaluate==0.4.1
160
+ exceptiongroup==1.2.0
161
+ executing==2.0.1
162
+ explainable-ai-sdk==1.3.3
163
+ farama-notifications==0.0.4
164
+ fastai==2.7.13
165
+ fastapi==0.108.0
166
+ fastavro==1.9.3
167
+ fastcore==1.5.29
168
+ fastdownload==0.0.7
169
+ fasteners==0.19
170
+ fastjsonschema==2.19.1
171
+ fastprogress==1.0.3
172
+ fastrlock==0.8.2
173
+ fasttext==0.9.2
174
+ fbpca==1.0
175
+ feather-format==0.4.1
176
+ featuretools==1.28.0
177
+ filelock==3.13.1
178
+ fiona==1.9.5
179
+ fitter==1.7.0
180
+ flake8==7.0.0
181
+ flashtext==2.7
182
+ flask==3.0.1
183
+ flatbuffers==23.5.26
184
+ flax==0.7.5
185
+ folium==0.15.1
186
+ fonttools==4.47.0
187
+ fqdn==1.5.1
188
+ frozendict==2.4.0
189
+ frozenlist==1.4.1
190
+ fsspec==2023.12.2
191
+ ftfy==6.0.3
192
+ funcy==2.0
193
+ fury==0.9.0
194
+ future==0.18.3
195
+ fuzzywuzzy==0.18.0
196
+ gast==0.5.4
197
+ gatspy==0.3
198
+ gcsfs==2023.12.2.post1
199
+ gensim==4.3.2
200
+ geographiclib==2.0
201
+ geohash==1.0
202
+ geojson==3.1.0
203
+ geopandas==0.14.2
204
+ geoplot==0.5.1
205
+ geopy==2.4.1
206
+ geoviews==1.11.0
207
+ ggplot==0.11.5
208
+ giddy==2.3.5
209
+ gitdb==4.0.11
210
+ gitpython==3.1.41
211
+ google-ai-generativelanguage==0.4.0
212
+ google-api-core==2.11.1
213
+ google-api-python-client==2.115.0
214
+ google-apitools==0.5.31
215
+ google-auth-httplib2==0.1.1
216
+ google-auth-oauthlib==1.2.0
217
+ google-auth==2.26.1
218
+ google-cloud-aiplatform==0.6.0a1
219
+ google-cloud-artifact-registry==1.10.0
220
+ google-cloud-automl==1.0.1
221
+ google-cloud-bigquery==2.34.4
222
+ google-cloud-bigtable==1.7.3
223
+ google-cloud-core==2.4.1
224
+ google-cloud-datastore==2.19.0
225
+ google-cloud-dlp==3.14.0
226
+ google-cloud-jupyter-config==0.0.5
227
+ google-cloud-language==2.12.0
228
+ google-cloud-monitoring==2.18.0
229
+ google-cloud-pubsub==2.19.0
230
+ google-cloud-pubsublite==1.9.0
231
+ google-cloud-recommendations-ai==0.7.1
232
+ google-cloud-resource-manager==1.11.0
233
+ google-cloud-spanner==3.40.1
234
+ google-cloud-storage==1.44.0
235
+ google-cloud-translate==3.12.1
236
+ google-cloud-videointelligence==2.12.0
237
+ google-cloud-vision==2.8.0
238
+ google-crc32c==1.5.0
239
+ google-generativeai==0.3.2
240
+ google-pasta==0.2.0
241
+ google-resumable-media==2.7.0
242
+ googleapis-common-protos==1.62.0
243
+ gplearn==0.4.2
244
+ gpustat==1.0.0
245
+ gpxpy==1.6.2
246
+ graphviz==0.20.1
247
+ greenlet==3.0.3
248
+ grpc-google-iam-v1==0.12.7
249
+ grpcio-status==1.48.1
250
+ grpcio==1.51.1
251
+ gviz-api==1.10.0
252
+ gym-notices==0.0.8
253
+ gym==0.26.2
254
+ gymnasium==0.29.0
255
+ h11==0.14.0
256
+ h2o==3.44.0.3
257
+ h5netcdf==1.3.0
258
+ h5py==3.10.0
259
+ haversine==2.8.1
260
+ hdfs==2.7.3
261
+ hep-ml==0.7.2
262
+ hijri-converter==2.3.1
263
+ hmmlearn==0.3.0
264
+ holidays==0.24
265
+ holoviews==1.18.1
266
+ hpsklearn==0.1.0
267
+ html5lib==1.1
268
+ htmlmin==0.1.12
269
+ httplib2==0.21.0
270
+ httptools==0.6.1
271
+ huggingface-hub==0.20.3
272
+ humanize==4.9.0
273
+ hunspell==0.5.5
274
+ husl==4.0.3
275
+ hydra-slayer==0.5.0
276
+ hyperopt==0.2.7
277
+ hypertools==0.8.0
278
+ idna==3.6
279
+ igraph==0.11.3
280
+ imagecodecs==2024.1.1
281
+ imagehash==4.3.1
282
+ imageio==2.33.1
283
+ imbalanced-learn==0.11.0
284
+ imgaug==0.4.0
285
+ importlib-metadata==6.11.0
286
+ importlib-resources==6.1.1
287
+ inequality==1.0.1
288
+ iniconfig==2.0.0
289
+ ipydatawidgets==4.3.5
290
+ ipykernel==6.28.0
291
+ ipyleaflet==0.18.2
292
+ ipympl==0.7.0
293
+ ipython-genutils==0.2.0
294
+ ipython-sql==0.5.0
295
+ ipython==8.20.0
296
+ ipyvolume==0.6.3
297
+ ipyvue==1.10.1
298
+ ipyvuetify==1.8.10
299
+ ipywebrtc==0.6.0
300
+ ipywidgets==7.7.1
301
+ isoduration==20.11.0
302
+ isort==5.13.2
303
+ isoweek==1.3.3
304
+ itsdangerous==2.1.2
305
+ janome==0.5.0
306
+ jaraco.classes==3.3.0
307
+ jax-jumpy==1.0.0
308
+ jax==0.4.23
309
+ jaxlib==0.4.23.dev20240116
310
+ jedi==0.19.1
311
+ jeepney==0.8.0
312
+ jieba==0.42.1
313
+ jinja2==3.1.2
314
+ jmespath==1.0.1
315
+ joblib==1.3.2
316
+ json5==0.9.14
317
+ jsonpatch==1.33
318
+ jsonpointer==2.4
319
+ jsonschema-specifications==2023.12.1
320
+ jsonschema==4.20.0
321
+ jupyter-client==7.4.9
322
+ jupyter-console==6.6.3
323
+ jupyter-core==5.7.1
324
+ jupyter-events==0.9.0
325
+ jupyter-http-over-ws==0.0.8
326
+ jupyter-lsp==1.5.1
327
+ jupyter-server-fileid==0.9.1
328
+ jupyter-server-mathjax==0.2.6
329
+ jupyter-server-proxy==4.1.0
330
+ jupyter-server-terminals==0.5.1
331
+ jupyter-server-ydoc==0.8.0
332
+ jupyter-server==2.12.5
333
+ jupyter-ydoc==0.2.5
334
+ jupyterlab-git==0.44.0
335
+ jupyterlab-lsp==5.0.2
336
+ jupyterlab-pygments==0.3.0
337
+ jupyterlab-server==2.25.2
338
+ jupyterlab-widgets==3.0.9
339
+ jupyterlab==4.0.11
340
+ jupytext==1.16.0
341
+ kaggle-environments==1.14.3
342
+ kaggle==1.6.3
343
+ kagglehub==0.1.6
344
+ keras-core==0.1.7
345
+ keras-cv==0.8.1
346
+ keras-nlp==0.7.0
347
+ keras-tuner==1.4.6
348
+ keras==2.15.0
349
+ kernels-mixer==0.0.7
350
+ keyring==24.3.0
351
+ keyrings.google-artifactregistry-auth==1.1.2
352
+ kfp-pipeline-spec==0.2.2
353
+ kfp-server-api==2.0.5
354
+ kfp==2.5.0
355
+ kiwisolver==1.4.5
356
+ kmapper==2.0.1
357
+ kmodes==0.12.2
358
+ korean-lunar-calendar==0.3.1
359
+ kornia==0.7.1
360
+ kt-legacy==1.0.5
361
+ kubernetes==26.1.0
362
+ langcodes==3.3.0
363
+ langid==1.1.6
364
+ lazy-loader==0.3
365
+ learntools==0.3.4
366
+ leven==1.0.4
367
+ levenshtein==0.23.0
368
+ libclang==16.0.6
369
+ libmambapy==1.5.0
370
+ libpysal==4.9.2
371
+ librosa==0.10.1
372
+ lightgbm==4.2.0
373
+ lightning-utilities==0.10.1
374
+ lime==0.2.0.1
375
+ line-profiler==4.1.2
376
+ linkify-it-py==2.0.2
377
+ llvmlite==0.41.1
378
+ lml==0.1.0
379
+ locket==1.0.0
380
+ loguru==0.7.2
381
+ lunarcalendar==0.0.9
382
+ lxml==5.1.0
383
+ lz4==4.3.3
384
+ mako==1.3.1
385
+ mamba==1.5.0
386
+ mapclassify==2.6.1
387
+ marisa-trie==1.1.0
388
+ markdown-it-py==3.0.0
389
+ markdown==3.5.2
390
+ markovify==0.9.4
391
+ markupsafe==2.1.3
392
+ marshmallow==3.20.2
393
+ matplotlib-inline==0.1.6
394
+ matplotlib-venn==0.11.9
395
+ matplotlib==3.7.4
396
+ mccabe==0.7.0
397
+ mdit-py-plugins==0.4.0
398
+ mdurl==0.1.2
399
+ memory-profiler==0.61.0
400
+ menuinst==2.0.1
401
+ mercantile==1.2.1
402
+ mgwr==2.2.1
403
+ missingno==0.5.2
404
+ mistune==0.8.4
405
+ mizani==0.9.3
406
+ ml-dtypes==0.2.0
407
+ mlcrate==0.2.0
408
+ mlens==0.2.3
409
+ mlxtend==0.23.1
410
+ mmh3==4.1.0
411
+ mne==1.6.1
412
+ mnist==0.2.2
413
+ mock==5.1.0
414
+ momepy==0.7.0
415
+ more-itertools==10.2.0
416
+ mpld3==0.5.10
417
+ mpmath==1.3.0
418
+ msgpack-numpy==0.4.8
419
+ msgpack==1.0.7
420
+ multidict==6.0.4
421
+ multimethod==1.10
422
+ multipledispatch==1.0.0
423
+ multiprocess==0.70.15
424
+ munkres==1.1.4
425
+ murmurhash==1.0.10
426
+ mypy-extensions==1.0.0
427
+ namex==0.0.7
428
+ nb-conda-kernels==2.3.1
429
+ nb-conda==2.2.1
430
+ nbclassic==1.0.0
431
+ nbclient==0.5.13
432
+ nbconvert==6.4.5
433
+ nbdime==3.2.0
434
+ nbformat==5.9.2
435
+ ndindex==1.7
436
+ nest-asyncio==1.5.8
437
+ networkx==3.2.1
438
+ nibabel==5.2.0
439
+ nilearn==0.10.2
440
+ ninja==1.11.1.1
441
+ nltk==3.2.4
442
+ normalizer==0.0.1
443
+ nose==1.3.7
444
+ notebook-executor==0.2
445
+ notebook-shim==0.2.3
446
+ notebook==6.5.4
447
+ numba==0.58.1
448
+ numexpr==2.8.8
449
+ numpy==1.24.4
450
+ nvidia-cublas-cu11==11.10.3.66
451
+ nvidia-cuda-nvrtc-cu11==11.7.99
452
+ nvidia-cuda-runtime-cu11==11.7.99
453
+ nvidia-cudnn-cu11==8.5.0.96
454
+ nvidia-ml-py==11.495.46
455
+ nvtx==0.2.8
456
+ oauth2client==4.1.3
457
+ oauthlib==3.2.2
458
+ objsize==0.6.1
459
+ odfpy==1.4.1
460
+ olefile==0.47
461
+ onnx==1.15.0
462
+ opencensus-context==0.1.3
463
+ opencensus==0.11.4
464
+ opencv-contrib-python==4.9.0.80
465
+ opencv-python-headless==4.9.0.80
466
+ opencv-python==4.9.0.80
467
+ openpyxl==3.1.2
468
+ openslide-python==1.3.1
469
+ opentelemetry-api==1.22.0
470
+ opentelemetry-exporter-otlp-proto-common==1.22.0
471
+ opentelemetry-exporter-otlp-proto-grpc==1.22.0
472
+ opentelemetry-exporter-otlp-proto-http==1.22.0
473
+ opentelemetry-exporter-otlp==1.22.0
474
+ opentelemetry-proto==1.22.0
475
+ opentelemetry-sdk==1.22.0
476
+ opentelemetry-semantic-conventions==0.43b0
477
+ opt-einsum==3.3.0
478
+ optax==0.1.8
479
+ optuna==3.5.0
480
+ orbax-checkpoint==0.4.4
481
+ ordered-set==4.1.0
482
+ orderedmultidict==1.0.1
483
+ orjson==3.9.10
484
+ ortools==9.4.1874
485
+ osmnx==1.8.1
486
+ overrides==7.4.0
487
+ packaging==21.3
488
+ pandas-datareader==0.10.0
489
+ pandas-profiling==3.6.6
490
+ pandas-summary==0.2.0
491
+ pandas==2.1.4
492
+ pandasql==0.7.3
493
+ pandocfilters==1.5.0
494
+ panel==1.3.7
495
+ papermill==2.5.0
496
+ param==2.0.2
497
+ parso==0.8.3
498
+ partd==1.4.1
499
+ path.py==12.5.0
500
+ path==16.9.0
501
+ pathos==0.3.1
502
+ pathy==0.10.3
503
+ patsy==0.5.6
504
+ pdf2image==1.17.0
505
+ pettingzoo==1.24.0
506
+ pexpect==4.8.0
507
+ phik==0.12.4
508
+ pickleshare==0.7.5
509
+ pillow==9.5.0
510
+ pip==23.3.2
511
+ pkgutil-resolve-name==1.3.10
512
+ platformdirs==4.1.0
513
+ plotly-express==0.4.1
514
+ plotly==5.18.0
515
+ plotnine==0.12.4
516
+ pluggy==1.3.0
517
+ pointpats==2.4.0
518
+ polars==0.20.5
519
+ polyglot==16.7.4
520
+ pooch==1.8.0
521
+ portalocker==2.8.2
522
+ pox==0.3.3
523
+ ppca==0.0.4
524
+ ppft==1.7.6.7
525
+ preprocessing==0.1.13
526
+ preshed==3.0.9
527
+ prettytable==3.9.0
528
+ progressbar2==4.3.2
529
+ prometheus-client==0.19.0
530
+ promise==2.3
531
+ prompt-toolkit==3.0.42
532
+ pronouncing==0.2.0
533
+ prophet==1.1.1
534
+ proto-plus==1.23.0
535
+ protobuf==3.20.3
536
+ psutil==5.9.3
537
+ ptyprocess==0.7.0
538
+ pudb==2024.1
539
+ pulp==2.8.0
540
+ pure-eval==0.2.2
541
+ py-cpuinfo==9.0.0
542
+ py-spy==0.3.14
543
+ py4j==0.10.9.7
544
+ pyaml==23.12.0
545
+ pyarabic==0.6.15
546
+ pyarrow==11.0.0
547
+ pyasn1-modules==0.3.0
548
+ pyasn1==0.5.1
549
+ pyastronomy==0.20.0
550
+ pybind11==2.11.1
551
+ pyclipper==1.3.0.post5
552
+ pycodestyle==2.11.1
553
+ pycosat==0.6.6
554
+ pycparser==2.21
555
+ pycryptodome==3.20.0
556
+ pyct==0.5.0
557
+ pycuda==2024.1
558
+ pydantic-core==2.14.6
559
+ pydantic==2.5.3
560
+ pydegensac==0.1.2
561
+ pydicom==2.4.4
562
+ pydocstyle==6.3.0
563
+ pydot==1.4.2
564
+ pydub==0.25.1
565
+ pyemd==1.0.0
566
+ pyerfa==2.0.1.1
567
+ pyexcel-io==0.6.6
568
+ pyexcel-ods==0.6.0
569
+ pyfasttext==0.4.6
570
+ pyflakes==3.2.0
571
+ pygltflib==1.16.1
572
+ pygments==2.17.2
573
+ pyjwt==2.8.0
574
+ pykalman==0.9.5
575
+ pyldavis==3.4.1
576
+ pylibraft==23.8.0
577
+ pylint==3.0.3
578
+ pymc3==3.11.4
579
+ pymeeus==0.5.12
580
+ pymongo==3.13.0
581
+ pympler==1.0.1
582
+ pynndescent==0.5.11
583
+ pynvml==11.4.1
584
+ pynvrtc==9.2
585
+ pyocr==0.8.5
586
+ pyopenssl==23.3.0
587
+ pyparsing==3.1.1
588
+ pypdf==4.0.0
589
+ pyproj==3.6.1
590
+ pysal==23.7
591
+ pyshp==2.3.1
592
+ pysocks==1.7.1
593
+ pytesseract==0.3.10
594
+ pytest==7.4.4
595
+ python-bidi==0.4.2
596
+ python-dateutil==2.8.2
597
+ python-dotenv==1.0.0
598
+ python-json-logger==2.0.7
599
+ python-levenshtein==0.23.0
600
+ python-louvain==0.16
601
+ python-lsp-jsonrpc==1.1.2
602
+ python-lsp-server==1.10.0
603
+ python-slugify==8.0.1
604
+ python-utils==3.8.1
605
+ pythreejs==2.4.2
606
+ pytoolconfig==1.3.1
607
+ pytools==2023.1.1
608
+ pytorch-ignite==0.4.13
609
+ pytorch-lightning==2.1.3
610
+ pytz==2023.3.post1
611
+ pyu2f==0.1.5
612
+ pyupset==0.1.1.post7
613
+ pyviz-comms==3.0.1
614
+ pywavelets==1.5.0
615
+ pyyaml==6.0.1
616
+ pyzmq==24.0.1
617
+ qgrid==1.3.1
618
+ qtconsole==5.5.1
619
+ qtpy==2.4.1
620
+ quantecon==0.7.1
621
+ quantities==0.15.0
622
+ qudida==0.0.4
623
+ raft-dask==23.8.0
624
+ rapidfuzz==3.6.1
625
+ rasterio==1.3.9
626
+ rasterstats==0.19.0
627
+ ray-cpp==2.9.0
628
+ ray==2.9.0
629
+ referencing==0.32.1
630
+ regex==2023.12.25
631
+ requests-oauthlib==1.3.1
632
+ requests-toolbelt==0.10.1
633
+ requests==2.31.0
634
+ responses==0.18.0
635
+ retrying==1.3.3
636
+ rfc3339-validator==0.1.4
637
+ rfc3986-validator==0.1.1
638
+ rgf-python==3.12.0
639
+ rich-click==1.7.3
640
+ rich==13.7.0
641
+ rmm==23.8.0
642
+ rope==1.12.0
643
+ rpds-py==0.16.2
644
+ rsa==4.9
645
+ rtree==1.2.0
646
+ ruamel-yaml-conda==0.15.100
647
+ ruamel.yaml.clib==0.2.7
648
+ ruamel.yaml==0.17.40
649
+ s2sphere==0.2.5
650
+ s3fs==2023.12.2
651
+ s3transfer==0.6.2
652
+ sacrebleu==2.4.0
653
+ safetensors==0.4.2
654
+ scattertext==0.1.19
655
+ scikit-image==0.22.0
656
+ scikit-learn-intelex==2024.1.0
657
+ scikit-learn==1.2.2
658
+ scikit-multilearn==0.2.0
659
+ scikit-optimize==0.9.0
660
+ scikit-plot==0.3.7
661
+ scikit-surprise==1.1.3
662
+ scipy==1.11.4
663
+ seaborn==0.12.2
664
+ secretstorage==3.3.3
665
+ segment-anything==1.0
666
+ segregation==2.5
667
+ semver==3.0.2
668
+ send2trash==1.8.2
669
+ sentencepiece==0.1.99
670
+ sentry-sdk==1.39.2
671
+ setproctitle==1.3.3
672
+ setuptools-git==1.2
673
+ setuptools-scm==8.0.4
674
+ setuptools==69.0.3
675
+ shap==0.44.0
676
+ shapely==1.8.5.post1
677
+ shellingham==1.5.4
678
+ shimmy==1.3.0
679
+ simpervisor==1.0.0
680
+ simpleitk==2.3.1
681
+ simplejson==3.19.2
682
+ six==1.16.0
683
+ sklearn-pandas==2.2.0
684
+ slicer==0.0.7
685
+ smart-open==6.4.0
686
+ smhasher==0.150.1
687
+ smmap==5.0.1
688
+ sniffio==1.3.0
689
+ snowballstemmer==2.2.0
690
+ snuggs==1.4.7
691
+ sortedcontainers==2.4.0
692
+ soundfile==0.12.1
693
+ soupsieve==2.5
694
+ soxr==0.3.7
695
+ spacy-legacy==3.0.12
696
+ spacy-loggers==1.0.5
697
+ spacy==3.7.2
698
+ spaghetti==1.7.4
699
+ spectral==0.23.1
700
+ spglm==1.1.0
701
+ sphinx-rtd-theme==0.2.4
702
+ spint==1.0.7
703
+ splot==1.1.5.post1
704
+ spopt==0.6.0
705
+ spreg==1.4.2
706
+ spvcm==0.3.0
707
+ sqlalchemy==2.0.25
708
+ sqlparse==0.4.4
709
+ squarify==0.4.3
710
+ srsly==2.4.8
711
+ stable-baselines3==2.1.0
712
+ stack-data==0.6.2
713
+ stanio==0.3.0
714
+ starlette==0.32.0.post1
715
+ statsmodels==0.14.1
716
+ stemming==1.0.1
717
+ stop-words==2018.7.23
718
+ stopit==1.1.2
719
+ stumpy==1.12.0
720
+ sympy==1.12
721
+ tables==3.9.2
722
+ tabulate==0.9.0
723
+ tangled-up-in-unicode==0.2.0
724
+ tbb==2021.11.0
725
+ tblib==3.0.0
726
+ tenacity==8.2.3
727
+ tensorboard-data-server==0.7.2
728
+ tensorboard-plugin-profile==2.15.0
729
+ tensorboard==2.15.1
730
+ tensorboardx==2.6.2.2
731
+ tensorflow-cloud==0.1.16
732
+ tensorflow-datasets==4.9.4
733
+ tensorflow-decision-forests==1.8.1
734
+ tensorflow-estimator==2.15.0
735
+ tensorflow-hub==0.15.0
736
+ tensorflow-io-gcs-filesystem==0.35.0
737
+ tensorflow-io==0.35.0
738
+ tensorflow-metadata==0.14.0
739
+ tensorflow-probability==0.23.0
740
+ tensorflow-serving-api==2.14.1
741
+ tensorflow-text==2.15.0
742
+ tensorflow-transform==0.14.0
743
+ tensorflow==2.15.0
744
+ tensorflowjs==4.16.0
745
+ tensorpack==0.11
746
+ tensorstore==0.1.45
747
+ termcolor==2.4.0
748
+ terminado==0.18.0
749
+ testpath==0.6.0
750
+ text-unidecode==1.3
751
+ textblob==0.17.1
752
+ texttable==1.7.0
753
+ theano-pymc==1.1.2
754
+ theano==1.0.5
755
+ thinc==8.2.2
756
+ threadpoolctl==3.2.0
757
+ tifffile==2023.12.9
758
+ timm==0.9.12
759
+ tinycss2==1.2.1
760
+ tobler==0.11.2
761
+ tokenizers==0.15.1
762
+ toml==0.10.2
763
+ tomli==2.0.1
764
+ tomlkit==0.12.3
765
+ toolz==0.12.0
766
+ torch==1.13.1
767
+ torchaudio==2.1.2
768
+ torchdata==0.7.1
769
+ torchinfo==1.8.0
770
+ torchmetrics==1.3.0.post0
771
+ torchtext==0.16.2
772
+ torchvision==0.16.2
773
+ tornado==6.3.3
774
+ tpot==0.12.1
775
+ tqdm==4.66.1
776
+ traceml==1.0.8
777
+ traitlets==5.9.0
778
+ traittypes==0.2.1
779
+ transformers==4.37.0
780
+ treelite-runtime==3.2.0
781
+ treelite==3.2.0
782
+ trueskill==0.4.5
783
+ truststore==0.8.0
784
+ trx-python==0.2.9
785
+ tsfresh==0.20.1
786
+ typeguard==4.1.5
787
+ typer==0.9.0
788
+ types-python-dateutil==2.8.19.20240106
789
+ typing-extensions==4.9.0
790
+ typing-inspect==0.9.0
791
+ typing-utils==0.1.0
792
+ tzdata==2023.4
793
+ tzlocal==5.2
794
+ uc-micro-py==1.0.2
795
+ ucx-py==0.33.0
796
+ ujson==5.9.0
797
+ umap-learn==0.5.5
798
+ unicodedata2==15.1.0
799
+ unidecode==1.3.8
800
+ update-checker==0.18.0
801
+ uri-template==1.3.0
802
+ uritemplate==3.0.1
803
+ urllib3==1.26.18
804
+ urwid-readline==0.13
805
+ urwid==2.4.6
806
+ uvicorn==0.25.0
807
+ uvloop==0.19.0
808
+ vaex-astro==0.9.3
809
+ vaex-core==4.17.1
810
+ vaex-hdf5==0.14.1
811
+ vaex-jupyter==0.8.2
812
+ vaex-ml==0.18.3
813
+ vaex-server==0.9.0
814
+ vaex-viz==0.5.4
815
+ vaex==4.17.0
816
+ vec-noise==1.1.4
817
+ vecstack==0.4.0
818
+ virtualenv==20.21.0
819
+ visions==0.7.5
820
+ vowpalwabbit==9.9.0
821
+ vtk==9.3.0
822
+ wand==0.6.13
823
+ wandb==0.16.2
824
+ wasabi==1.1.2
825
+ watchfiles==0.21.0
826
+ wavio==0.0.8
827
+ wcwidth==0.2.13
828
+ weasel==0.3.4
829
+ webcolors==1.13
830
+ webencodings==0.5.1
831
+ websocket-client==1.7.0
832
+ websockets==12.0
833
+ werkzeug==3.0.1
834
+ wfdb==4.1.2
835
+ whatthepatch==1.0.5
836
+ wheel==0.42.0
837
+ widgetsnbextension==3.6.6
838
+ witwidget==1.8.1
839
+ woodwork==0.27.0
840
+ wordcloud==1.9.3
841
+ wordsegment==1.3.1
842
+ wrapt==1.14.1
843
+ wurlitzer==3.0.3
844
+ xarray-einstats==0.7.0
845
+ xarray==2024.1.0
846
+ xgboost==2.0.3
847
+ xvfbwrapper==0.2.9
848
+ xxhash==3.4.1
849
+ xyzservices==2023.10.1
850
+ y-py==0.6.2
851
+ yapf==0.40.2
852
+ yarl==1.9.3
853
+ ydata-profiling==4.6.4
854
+ yellowbrick==1.5
855
+ ypy-websocket==0.8.4
856
+ zict==3.0.0
857
+ zipp==3.17.0
858
+ zstandard==0.22.0
wandb/run-20240226_092105-91uhs91a/files/wandb-metadata.json ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.133+-x86_64-with-glibc2.31",
3
+ "python": "3.10.13",
4
+ "heartbeatAt": "2024-02-26T09:21:06.484258",
5
+ "startedAt": "2024-02-26T09:21:05.525040",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [],
9
+ "state": "running",
10
+ "program": "kaggle.ipynb",
11
+ "codePathLocal": null,
12
+ "root": "/kaggle/working",
13
+ "host": "d79522c38279",
14
+ "username": "root",
15
+ "executable": "/opt/conda/bin/python3.10",
16
+ "cpu_count": 2,
17
+ "cpu_count_logical": 4,
18
+ "cpu_freq": {
19
+ "current": 2000.186,
20
+ "min": 0.0,
21
+ "max": 0.0
22
+ },
23
+ "cpu_freq_per_core": [
24
+ {
25
+ "current": 2000.186,
26
+ "min": 0.0,
27
+ "max": 0.0
28
+ },
29
+ {
30
+ "current": 2000.186,
31
+ "min": 0.0,
32
+ "max": 0.0
33
+ },
34
+ {
35
+ "current": 2000.186,
36
+ "min": 0.0,
37
+ "max": 0.0
38
+ },
39
+ {
40
+ "current": 2000.186,
41
+ "min": 0.0,
42
+ "max": 0.0
43
+ }
44
+ ],
45
+ "disk": {
46
+ "/": {
47
+ "total": 8062.387607574463,
48
+ "used": 5437.818241119385
49
+ }
50
+ },
51
+ "gpu": "Tesla P100-PCIE-16GB",
52
+ "gpu_count": 1,
53
+ "gpu_devices": [
54
+ {
55
+ "name": "Tesla P100-PCIE-16GB",
56
+ "memory_total": 17179869184
57
+ }
58
+ ],
59
+ "memory": {
60
+ "total": 31.357559204101562
61
+ }
62
+ }
wandb/run-20240226_092105-91uhs91a/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"eval/loss": 1.740322470664978, "eval/bleu": 14.5679, "eval/gen_len": 11.9038, "eval/runtime": 27.3507, "eval/samples_per_second": 29.25, "eval/steps_per_second": 1.828, "train/epoch": 1.11, "train/global_step": 500, "_timestamp": 1708939544.4191306, "_runtime": 278.8868725299835, "_step": 1, "train/loss": 2.1563, "train/learning_rate": 0.0003888888888888889}
wandb/run-20240226_092105-91uhs91a/logs/debug-internal.log ADDED
@@ -0,0 +1,327 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-02-26 09:21:05,531 INFO StreamThr :196 [internal.py:wandb_internal():86] W&B internal server running at pid: 196, started at: 2024-02-26 09:21:05.531199
2
+ 2024-02-26 09:21:05,533 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status
3
+ 2024-02-26 09:21:06,208 INFO WriterThread:196 [datastore.py:open_for_write():87] open: /kaggle/working/wandb/run-20240226_092105-91uhs91a/run-91uhs91a.wandb
4
+ 2024-02-26 09:21:06,208 DEBUG SenderThread:196 [sender.py:send():382] send: header
5
+ 2024-02-26 09:21:06,211 DEBUG SenderThread:196 [sender.py:send():382] send: run
6
+ 2024-02-26 09:21:06,389 INFO SenderThread:196 [dir_watcher.py:__init__():211] watching files in: /kaggle/working/wandb/run-20240226_092105-91uhs91a/files
7
+ 2024-02-26 09:21:06,389 INFO SenderThread:196 [sender.py:_start_run_threads():1136] run started: 91uhs91a with start time 1708939265.532258
8
+ 2024-02-26 09:21:06,397 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: check_version
9
+ 2024-02-26 09:21:06,397 DEBUG SenderThread:196 [sender.py:send_request():409] send_request: check_version
10
+ 2024-02-26 09:21:06,464 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: run_start
11
+ 2024-02-26 09:21:06,475 DEBUG HandlerThread:196 [system_info.py:__init__():27] System info init
12
+ 2024-02-26 09:21:06,475 DEBUG HandlerThread:196 [system_info.py:__init__():42] System info init done
13
+ 2024-02-26 09:21:06,475 INFO HandlerThread:196 [system_monitor.py:start():194] Starting system monitor
14
+ 2024-02-26 09:21:06,475 INFO SystemMonitor:196 [system_monitor.py:_start():158] Starting system asset monitoring threads
15
+ 2024-02-26 09:21:06,475 INFO HandlerThread:196 [system_monitor.py:probe():214] Collecting system info
16
+ 2024-02-26 09:21:06,476 INFO SystemMonitor:196 [interfaces.py:start():190] Started cpu monitoring
17
+ 2024-02-26 09:21:06,476 INFO SystemMonitor:196 [interfaces.py:start():190] Started disk monitoring
18
+ 2024-02-26 09:21:06,477 INFO SystemMonitor:196 [interfaces.py:start():190] Started gpu monitoring
19
+ 2024-02-26 09:21:06,477 INFO SystemMonitor:196 [interfaces.py:start():190] Started memory monitoring
20
+ 2024-02-26 09:21:06,478 INFO SystemMonitor:196 [interfaces.py:start():190] Started network monitoring
21
+ 2024-02-26 09:21:06,484 DEBUG HandlerThread:196 [system_info.py:probe():151] Probing system
22
+ 2024-02-26 09:21:06,486 DEBUG HandlerThread:196 [gitlib.py:_init_repo():56] git repository is invalid
23
+ 2024-02-26 09:21:06,486 DEBUG HandlerThread:196 [system_info.py:probe():199] Probing system done
24
+ 2024-02-26 09:21:06,487 DEBUG HandlerThread:196 [system_monitor.py:probe():223] {'os': 'Linux-5.15.133+-x86_64-with-glibc2.31', 'python': '3.10.13', 'heartbeatAt': '2024-02-26T09:21:06.484258', 'startedAt': '2024-02-26T09:21:05.525040', 'docker': None, 'cuda': None, 'args': (), 'state': 'running', 'program': 'kaggle.ipynb', 'codePathLocal': None, 'root': '/kaggle/working', 'host': 'd79522c38279', 'username': 'root', 'executable': '/opt/conda/bin/python3.10', 'cpu_count': 2, 'cpu_count_logical': 4, 'cpu_freq': {'current': 2000.186, 'min': 0.0, 'max': 0.0}, 'cpu_freq_per_core': [{'current': 2000.186, 'min': 0.0, 'max': 0.0}, {'current': 2000.186, 'min': 0.0, 'max': 0.0}, {'current': 2000.186, 'min': 0.0, 'max': 0.0}, {'current': 2000.186, 'min': 0.0, 'max': 0.0}], 'disk': {'/': {'total': 8062.387607574463, 'used': 5437.818241119385}}, 'gpu': 'Tesla P100-PCIE-16GB', 'gpu_count': 1, 'gpu_devices': [{'name': 'Tesla P100-PCIE-16GB', 'memory_total': 17179869184}], 'memory': {'total': 31.357559204101562}}
25
+ 2024-02-26 09:21:06,487 INFO HandlerThread:196 [system_monitor.py:probe():224] Finished collecting system info
26
+ 2024-02-26 09:21:06,487 INFO HandlerThread:196 [system_monitor.py:probe():227] Publishing system info
27
+ 2024-02-26 09:21:06,487 DEBUG HandlerThread:196 [system_info.py:_save_conda():208] Saving list of conda packages installed into the current environment
28
+ 2024-02-26 09:21:07,391 INFO Thread-12 :196 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240226_092105-91uhs91a/files/conda-environment.yaml
29
+ 2024-02-26 09:21:58,406 INFO Thread-12 :196 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240226_092105-91uhs91a/files/conda-environment.yaml
30
+ 2024-02-26 09:21:58,864 DEBUG HandlerThread:196 [system_info.py:_save_conda():220] Saving conda packages done
31
+ 2024-02-26 09:21:58,865 INFO HandlerThread:196 [system_monitor.py:probe():229] Finished publishing system info
32
+ 2024-02-26 09:21:58,872 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
33
+ 2024-02-26 09:21:58,872 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
34
+ 2024-02-26 09:21:58,873 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
35
+ 2024-02-26 09:21:58,873 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
36
+ 2024-02-26 09:21:58,873 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
37
+ 2024-02-26 09:21:58,873 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
38
+ 2024-02-26 09:21:58,873 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
39
+ 2024-02-26 09:21:58,873 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
40
+ 2024-02-26 09:21:58,873 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
41
+ 2024-02-26 09:21:58,873 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
42
+ 2024-02-26 09:21:58,873 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
43
+ 2024-02-26 09:21:58,873 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: python_packages
44
+ 2024-02-26 09:21:58,873 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: internal_messages
45
+ 2024-02-26 09:21:58,873 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: stop_status
46
+ 2024-02-26 09:21:58,874 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: pause
47
+ 2024-02-26 09:21:58,874 INFO HandlerThread:196 [handler.py:handle_request_pause():708] stopping system metrics thread
48
+ 2024-02-26 09:21:58,874 INFO HandlerThread:196 [system_monitor.py:finish():203] Stopping system monitor
49
+ 2024-02-26 09:21:58,875 DEBUG SystemMonitor:196 [system_monitor.py:_start():172] Starting system metrics aggregation loop
50
+ 2024-02-26 09:21:58,875 DEBUG SystemMonitor:196 [system_monitor.py:_start():179] Finished system metrics aggregation loop
51
+ 2024-02-26 09:21:58,875 DEBUG SystemMonitor:196 [system_monitor.py:_start():183] Publishing last batch of metrics
52
+ 2024-02-26 09:21:58,876 DEBUG SenderThread:196 [sender.py:send_request():409] send_request: python_packages
53
+ 2024-02-26 09:21:58,878 INFO HandlerThread:196 [interfaces.py:finish():202] Joined cpu monitor
54
+ 2024-02-26 09:21:58,879 INFO HandlerThread:196 [interfaces.py:finish():202] Joined disk monitor
55
+ 2024-02-26 09:21:58,879 DEBUG SenderThread:196 [sender.py:send():382] send: telemetry
56
+ 2024-02-26 09:21:58,879 DEBUG SenderThread:196 [sender.py:send():382] send: config
57
+ 2024-02-26 09:21:58,881 DEBUG SenderThread:196 [sender.py:send():382] send: metric
58
+ 2024-02-26 09:21:58,881 DEBUG SenderThread:196 [sender.py:send():382] send: telemetry
59
+ 2024-02-26 09:21:58,881 DEBUG SenderThread:196 [sender.py:send():382] send: metric
60
+ 2024-02-26 09:21:58,881 WARNING SenderThread:196 [sender.py:send_metric():1354] Seen metric with glob (shouldn't happen)
61
+ 2024-02-26 09:21:58,881 DEBUG SenderThread:196 [sender.py:send():382] send: telemetry
62
+ 2024-02-26 09:21:58,882 DEBUG SenderThread:196 [sender.py:send_request():409] send_request: stop_status
63
+ 2024-02-26 09:21:58,891 INFO HandlerThread:196 [interfaces.py:finish():202] Joined gpu monitor
64
+ 2024-02-26 09:21:58,892 INFO HandlerThread:196 [interfaces.py:finish():202] Joined memory monitor
65
+ 2024-02-26 09:21:58,892 INFO HandlerThread:196 [interfaces.py:finish():202] Joined network monitor
66
+ 2024-02-26 09:21:58,892 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
67
+ 2024-02-26 09:21:58,892 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
68
+ 2024-02-26 09:21:58,892 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
69
+ 2024-02-26 09:21:58,892 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
70
+ 2024-02-26 09:21:58,892 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
71
+ 2024-02-26 09:21:58,892 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
72
+ 2024-02-26 09:21:58,893 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
73
+ 2024-02-26 09:21:58,893 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
74
+ 2024-02-26 09:21:58,893 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: resume
75
+ 2024-02-26 09:21:58,893 INFO HandlerThread:196 [handler.py:handle_request_resume():699] starting system metrics thread
76
+ 2024-02-26 09:21:58,893 INFO HandlerThread:196 [system_monitor.py:start():194] Starting system monitor
77
+ 2024-02-26 09:21:58,893 INFO SystemMonitor:196 [system_monitor.py:_start():158] Starting system asset monitoring threads
78
+ 2024-02-26 09:21:58,893 INFO SystemMonitor:196 [interfaces.py:start():190] Started cpu monitoring
79
+ 2024-02-26 09:21:58,894 INFO SystemMonitor:196 [interfaces.py:start():190] Started disk monitoring
80
+ 2024-02-26 09:21:58,895 INFO SystemMonitor:196 [interfaces.py:start():190] Started gpu monitoring
81
+ 2024-02-26 09:21:58,896 INFO SystemMonitor:196 [interfaces.py:start():190] Started memory monitoring
82
+ 2024-02-26 09:21:58,897 INFO SystemMonitor:196 [interfaces.py:start():190] Started network monitoring
83
+ 2024-02-26 09:21:59,035 DEBUG SenderThread:196 [sender.py:send():382] send: config
84
+ 2024-02-26 09:21:59,037 DEBUG SenderThread:196 [sender.py:send():382] send: metric
85
+ 2024-02-26 09:21:59,037 DEBUG SenderThread:196 [sender.py:send():382] send: metric
86
+ 2024-02-26 09:21:59,037 WARNING SenderThread:196 [sender.py:send_metric():1354] Seen metric with glob (shouldn't happen)
87
+ 2024-02-26 09:21:59,037 DEBUG SenderThread:196 [sender.py:send():382] send: files
88
+ 2024-02-26 09:21:59,038 INFO SenderThread:196 [sender.py:_save_file():1403] saving file wandb-metadata.json with policy now
89
+ 2024-02-26 09:21:59,038 DEBUG SenderThread:196 [sender.py:send():382] send: stats
90
+ 2024-02-26 09:21:59,335 INFO wandb-upload_0:196 [upload_job.py:push():131] Uploaded file /tmp/tmp5wprt0czwandb/qmdsgk7x-wandb-metadata.json
91
+ 2024-02-26 09:21:59,406 INFO Thread-12 :196 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240226_092105-91uhs91a/files/wandb-metadata.json
92
+ 2024-02-26 09:21:59,406 INFO Thread-12 :196 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240226_092105-91uhs91a/files/requirements.txt
93
+ 2024-02-26 09:21:59,602 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: pause
94
+ 2024-02-26 09:21:59,602 INFO HandlerThread:196 [handler.py:handle_request_pause():708] stopping system metrics thread
95
+ 2024-02-26 09:21:59,602 INFO HandlerThread:196 [system_monitor.py:finish():203] Stopping system monitor
96
+ 2024-02-26 09:21:59,603 INFO HandlerThread:196 [interfaces.py:finish():202] Joined cpu monitor
97
+ 2024-02-26 09:21:59,603 INFO HandlerThread:196 [interfaces.py:finish():202] Joined disk monitor
98
+ 2024-02-26 09:21:59,603 DEBUG SystemMonitor:196 [system_monitor.py:_start():172] Starting system metrics aggregation loop
99
+ 2024-02-26 09:21:59,604 DEBUG SystemMonitor:196 [system_monitor.py:_start():179] Finished system metrics aggregation loop
100
+ 2024-02-26 09:21:59,604 DEBUG SystemMonitor:196 [system_monitor.py:_start():183] Publishing last batch of metrics
101
+ 2024-02-26 09:21:59,611 INFO HandlerThread:196 [interfaces.py:finish():202] Joined gpu monitor
102
+ 2024-02-26 09:21:59,611 INFO HandlerThread:196 [interfaces.py:finish():202] Joined memory monitor
103
+ 2024-02-26 09:21:59,611 INFO HandlerThread:196 [interfaces.py:finish():202] Joined network monitor
104
+ 2024-02-26 09:21:59,611 DEBUG SenderThread:196 [sender.py:send():382] send: stats
105
+ 2024-02-26 09:22:01,613 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
106
+ 2024-02-26 09:22:01,749 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
107
+ 2024-02-26 09:22:06,619 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
108
+ 2024-02-26 09:22:06,755 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
109
+ 2024-02-26 09:22:07,409 INFO Thread-12 :196 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240226_092105-91uhs91a/files/config.yaml
110
+ 2024-02-26 09:22:08,331 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: resume
111
+ 2024-02-26 09:22:08,331 INFO HandlerThread:196 [handler.py:handle_request_resume():699] starting system metrics thread
112
+ 2024-02-26 09:22:08,331 INFO HandlerThread:196 [system_monitor.py:start():194] Starting system monitor
113
+ 2024-02-26 09:22:08,331 INFO SystemMonitor:196 [system_monitor.py:_start():158] Starting system asset monitoring threads
114
+ 2024-02-26 09:22:08,332 INFO SystemMonitor:196 [interfaces.py:start():190] Started cpu monitoring
115
+ 2024-02-26 09:22:08,332 INFO SystemMonitor:196 [interfaces.py:start():190] Started disk monitoring
116
+ 2024-02-26 09:22:08,335 INFO SystemMonitor:196 [interfaces.py:start():190] Started gpu monitoring
117
+ 2024-02-26 09:22:08,335 INFO SystemMonitor:196 [interfaces.py:start():190] Started memory monitoring
118
+ 2024-02-26 09:22:08,336 INFO SystemMonitor:196 [interfaces.py:start():190] Started network monitoring
119
+ 2024-02-26 09:22:09,404 DEBUG SenderThread:196 [sender.py:send():382] send: config
120
+ 2024-02-26 09:22:09,406 DEBUG SenderThread:196 [sender.py:send():382] send: metric
121
+ 2024-02-26 09:22:09,406 DEBUG SenderThread:196 [sender.py:send():382] send: metric
122
+ 2024-02-26 09:22:09,406 WARNING SenderThread:196 [sender.py:send_metric():1354] Seen metric with glob (shouldn't happen)
123
+ 2024-02-26 09:22:09,927 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: pause
124
+ 2024-02-26 09:22:09,927 INFO HandlerThread:196 [handler.py:handle_request_pause():708] stopping system metrics thread
125
+ 2024-02-26 09:22:09,927 INFO HandlerThread:196 [system_monitor.py:finish():203] Stopping system monitor
126
+ 2024-02-26 09:22:09,927 INFO HandlerThread:196 [interfaces.py:finish():202] Joined cpu monitor
127
+ 2024-02-26 09:22:09,928 INFO HandlerThread:196 [interfaces.py:finish():202] Joined disk monitor
128
+ 2024-02-26 09:22:09,928 DEBUG SystemMonitor:196 [system_monitor.py:_start():172] Starting system metrics aggregation loop
129
+ 2024-02-26 09:22:09,928 DEBUG SystemMonitor:196 [system_monitor.py:_start():179] Finished system metrics aggregation loop
130
+ 2024-02-26 09:22:09,928 DEBUG SystemMonitor:196 [system_monitor.py:_start():183] Publishing last batch of metrics
131
+ 2024-02-26 09:22:09,934 INFO HandlerThread:196 [interfaces.py:finish():202] Joined gpu monitor
132
+ 2024-02-26 09:22:09,934 INFO HandlerThread:196 [interfaces.py:finish():202] Joined memory monitor
133
+ 2024-02-26 09:22:09,934 INFO HandlerThread:196 [interfaces.py:finish():202] Joined network monitor
134
+ 2024-02-26 09:22:09,935 DEBUG SenderThread:196 [sender.py:send():382] send: stats
135
+ 2024-02-26 09:22:11,810 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
136
+ 2024-02-26 09:22:11,936 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
137
+ 2024-02-26 09:22:16,811 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
138
+ 2024-02-26 09:22:16,936 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
139
+ 2024-02-26 09:22:18,738 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: resume
140
+ 2024-02-26 09:22:18,738 INFO HandlerThread:196 [handler.py:handle_request_resume():699] starting system metrics thread
141
+ 2024-02-26 09:22:18,738 INFO HandlerThread:196 [system_monitor.py:start():194] Starting system monitor
142
+ 2024-02-26 09:22:18,738 INFO SystemMonitor:196 [system_monitor.py:_start():158] Starting system asset monitoring threads
143
+ 2024-02-26 09:22:18,739 INFO SystemMonitor:196 [interfaces.py:start():190] Started cpu monitoring
144
+ 2024-02-26 09:22:18,740 INFO SystemMonitor:196 [interfaces.py:start():190] Started disk monitoring
145
+ 2024-02-26 09:22:18,740 INFO SystemMonitor:196 [interfaces.py:start():190] Started gpu monitoring
146
+ 2024-02-26 09:22:18,742 INFO SystemMonitor:196 [interfaces.py:start():190] Started memory monitoring
147
+ 2024-02-26 09:22:18,743 INFO SystemMonitor:196 [interfaces.py:start():190] Started network monitoring
148
+ 2024-02-26 09:22:19,271 DEBUG SenderThread:196 [sender.py:send():382] send: config
149
+ 2024-02-26 09:22:19,272 DEBUG SenderThread:196 [sender.py:send():382] send: metric
150
+ 2024-02-26 09:22:19,273 DEBUG SenderThread:196 [sender.py:send():382] send: metric
151
+ 2024-02-26 09:22:19,273 WARNING SenderThread:196 [sender.py:send_metric():1354] Seen metric with glob (shouldn't happen)
152
+ 2024-02-26 09:22:19,668 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: pause
153
+ 2024-02-26 09:22:19,669 INFO HandlerThread:196 [handler.py:handle_request_pause():708] stopping system metrics thread
154
+ 2024-02-26 09:22:19,669 INFO HandlerThread:196 [system_monitor.py:finish():203] Stopping system monitor
155
+ 2024-02-26 09:22:19,669 DEBUG SystemMonitor:196 [system_monitor.py:_start():172] Starting system metrics aggregation loop
156
+ 2024-02-26 09:22:19,669 DEBUG SystemMonitor:196 [system_monitor.py:_start():179] Finished system metrics aggregation loop
157
+ 2024-02-26 09:22:19,669 DEBUG SystemMonitor:196 [system_monitor.py:_start():183] Publishing last batch of metrics
158
+ 2024-02-26 09:22:19,670 INFO HandlerThread:196 [interfaces.py:finish():202] Joined cpu monitor
159
+ 2024-02-26 09:22:19,670 INFO HandlerThread:196 [interfaces.py:finish():202] Joined disk monitor
160
+ 2024-02-26 09:22:19,675 INFO HandlerThread:196 [interfaces.py:finish():202] Joined gpu monitor
161
+ 2024-02-26 09:22:19,676 INFO HandlerThread:196 [interfaces.py:finish():202] Joined memory monitor
162
+ 2024-02-26 09:22:19,676 INFO HandlerThread:196 [interfaces.py:finish():202] Joined network monitor
163
+ 2024-02-26 09:22:19,676 DEBUG SenderThread:196 [sender.py:send():382] send: stats
164
+ 2024-02-26 09:22:21,813 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
165
+ 2024-02-26 09:22:22,677 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
166
+ 2024-02-26 09:22:26,814 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
167
+ 2024-02-26 09:22:27,678 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
168
+ 2024-02-26 09:22:31,815 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
169
+ 2024-02-26 09:22:32,679 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
170
+ 2024-02-26 09:22:36,816 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
171
+ 2024-02-26 09:22:37,685 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
172
+ 2024-02-26 09:22:37,695 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: resume
173
+ 2024-02-26 09:22:37,696 INFO HandlerThread:196 [handler.py:handle_request_resume():699] starting system metrics thread
174
+ 2024-02-26 09:22:37,696 INFO HandlerThread:196 [system_monitor.py:start():194] Starting system monitor
175
+ 2024-02-26 09:22:37,696 INFO SystemMonitor:196 [system_monitor.py:_start():158] Starting system asset monitoring threads
176
+ 2024-02-26 09:22:37,696 INFO SystemMonitor:196 [interfaces.py:start():190] Started cpu monitoring
177
+ 2024-02-26 09:22:37,697 INFO SystemMonitor:196 [interfaces.py:start():190] Started disk monitoring
178
+ 2024-02-26 09:22:37,699 INFO SystemMonitor:196 [interfaces.py:start():190] Started gpu monitoring
179
+ 2024-02-26 09:22:37,701 INFO SystemMonitor:196 [interfaces.py:start():190] Started memory monitoring
180
+ 2024-02-26 09:22:37,702 INFO SystemMonitor:196 [interfaces.py:start():190] Started network monitoring
181
+ 2024-02-26 09:22:38,419 INFO Thread-12 :196 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240226_092105-91uhs91a/files/config.yaml
182
+ 2024-02-26 09:22:38,766 DEBUG SenderThread:196 [sender.py:send():382] send: config
183
+ 2024-02-26 09:22:38,768 DEBUG SenderThread:196 [sender.py:send():382] send: metric
184
+ 2024-02-26 09:22:38,768 DEBUG SenderThread:196 [sender.py:send():382] send: metric
185
+ 2024-02-26 09:22:38,768 WARNING SenderThread:196 [sender.py:send_metric():1354] Seen metric with glob (shouldn't happen)
186
+ 2024-02-26 09:22:42,174 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
187
+ 2024-02-26 09:22:43,769 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
188
+ 2024-02-26 09:22:47,175 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
189
+ 2024-02-26 09:22:48,770 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
190
+ 2024-02-26 09:22:52,181 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
191
+ 2024-02-26 09:22:53,770 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
192
+ 2024-02-26 09:22:57,184 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
193
+ 2024-02-26 09:22:58,771 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
194
+ 2024-02-26 09:23:02,185 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
195
+ 2024-02-26 09:23:03,772 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
196
+ 2024-02-26 09:23:07,189 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
197
+ 2024-02-26 09:23:08,778 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
198
+ 2024-02-26 09:23:09,430 INFO Thread-12 :196 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240226_092105-91uhs91a/files/config.yaml
199
+ 2024-02-26 09:23:12,191 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
200
+ 2024-02-26 09:23:13,902 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
201
+ 2024-02-26 09:23:17,192 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
202
+ 2024-02-26 09:23:18,902 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
203
+ 2024-02-26 09:23:19,821 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: pause
204
+ 2024-02-26 09:23:19,821 INFO HandlerThread:196 [handler.py:handle_request_pause():708] stopping system metrics thread
205
+ 2024-02-26 09:23:19,821 INFO HandlerThread:196 [system_monitor.py:finish():203] Stopping system monitor
206
+ 2024-02-26 09:23:19,821 DEBUG SystemMonitor:196 [system_monitor.py:_start():172] Starting system metrics aggregation loop
207
+ 2024-02-26 09:23:19,821 DEBUG SystemMonitor:196 [system_monitor.py:_start():179] Finished system metrics aggregation loop
208
+ 2024-02-26 09:23:19,821 DEBUG SystemMonitor:196 [system_monitor.py:_start():183] Publishing last batch of metrics
209
+ 2024-02-26 09:23:19,822 INFO HandlerThread:196 [interfaces.py:finish():202] Joined cpu monitor
210
+ 2024-02-26 09:23:19,822 INFO HandlerThread:196 [interfaces.py:finish():202] Joined disk monitor
211
+ 2024-02-26 09:23:19,830 INFO HandlerThread:196 [interfaces.py:finish():202] Joined gpu monitor
212
+ 2024-02-26 09:23:19,830 INFO HandlerThread:196 [interfaces.py:finish():202] Joined memory monitor
213
+ 2024-02-26 09:23:19,830 INFO HandlerThread:196 [interfaces.py:finish():202] Joined network monitor
214
+ 2024-02-26 09:23:19,830 DEBUG SenderThread:196 [sender.py:send():382] send: stats
215
+ 2024-02-26 09:23:22,193 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
216
+ 2024-02-26 09:23:24,832 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
217
+ 2024-02-26 09:23:27,194 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
218
+ 2024-02-26 09:23:29,832 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
219
+ 2024-02-26 09:23:32,195 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
220
+ 2024-02-26 09:23:32,481 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: resume
221
+ 2024-02-26 09:23:32,481 INFO HandlerThread:196 [handler.py:handle_request_resume():699] starting system metrics thread
222
+ 2024-02-26 09:23:32,481 INFO HandlerThread:196 [system_monitor.py:start():194] Starting system monitor
223
+ 2024-02-26 09:23:32,482 INFO SystemMonitor:196 [system_monitor.py:_start():158] Starting system asset monitoring threads
224
+ 2024-02-26 09:23:32,482 INFO SystemMonitor:196 [interfaces.py:start():190] Started cpu monitoring
225
+ 2024-02-26 09:23:32,483 INFO SystemMonitor:196 [interfaces.py:start():190] Started disk monitoring
226
+ 2024-02-26 09:23:32,484 INFO SystemMonitor:196 [interfaces.py:start():190] Started gpu monitoring
227
+ 2024-02-26 09:23:32,485 INFO SystemMonitor:196 [interfaces.py:start():190] Started memory monitoring
228
+ 2024-02-26 09:23:32,486 INFO SystemMonitor:196 [interfaces.py:start():190] Started network monitoring
229
+ 2024-02-26 09:23:33,088 DEBUG SenderThread:196 [sender.py:send():382] send: config
230
+ 2024-02-26 09:23:33,090 DEBUG SenderThread:196 [sender.py:send():382] send: metric
231
+ 2024-02-26 09:23:33,090 DEBUG SenderThread:196 [sender.py:send():382] send: metric
232
+ 2024-02-26 09:23:33,090 WARNING SenderThread:196 [sender.py:send_metric():1354] Seen metric with glob (shouldn't happen)
233
+ 2024-02-26 09:23:35,091 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
234
+ 2024-02-26 09:23:37,198 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
235
+ 2024-02-26 09:23:40,097 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
236
+ 2024-02-26 09:23:40,440 INFO Thread-12 :196 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240226_092105-91uhs91a/files/config.yaml
237
+ 2024-02-26 09:23:42,199 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
238
+ 2024-02-26 09:23:45,217 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
239
+ 2024-02-26 09:23:47,200 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
240
+ 2024-02-26 09:23:50,218 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
241
+ 2024-02-26 09:23:52,201 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
242
+ 2024-02-26 09:23:55,219 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
243
+ 2024-02-26 09:23:57,202 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
244
+ 2024-02-26 09:24:00,220 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
245
+ 2024-02-26 09:24:02,203 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
246
+ 2024-02-26 09:24:05,221 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
247
+ 2024-02-26 09:24:07,204 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
248
+ 2024-02-26 09:24:10,221 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
249
+ 2024-02-26 09:24:12,209 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
250
+ 2024-02-26 09:24:15,222 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
251
+ 2024-02-26 09:24:17,210 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
252
+ 2024-02-26 09:24:20,223 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
253
+ 2024-02-26 09:24:22,211 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
254
+ 2024-02-26 09:24:25,224 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
255
+ 2024-02-26 09:24:27,212 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
256
+ 2024-02-26 09:24:30,225 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
257
+ 2024-02-26 09:24:32,213 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
258
+ 2024-02-26 09:24:32,486 DEBUG SystemMonitor:196 [system_monitor.py:_start():172] Starting system metrics aggregation loop
259
+ 2024-02-26 09:24:32,488 DEBUG SenderThread:196 [sender.py:send():382] send: stats
260
+ 2024-02-26 09:24:35,489 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
261
+ 2024-02-26 09:24:37,214 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
262
+ 2024-02-26 09:24:40,490 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
263
+ 2024-02-26 09:24:42,215 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
264
+ 2024-02-26 09:24:45,491 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
265
+ 2024-02-26 09:24:47,219 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
266
+ 2024-02-26 09:24:50,491 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
267
+ 2024-02-26 09:24:52,220 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
268
+ 2024-02-26 09:24:55,492 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
269
+ 2024-02-26 09:24:57,221 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
270
+ 2024-02-26 09:25:00,493 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
271
+ 2024-02-26 09:25:02,226 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
272
+ 2024-02-26 09:25:02,488 DEBUG SenderThread:196 [sender.py:send():382] send: stats
273
+ 2024-02-26 09:25:06,490 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
274
+ 2024-02-26 09:25:07,227 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
275
+ 2024-02-26 09:25:07,470 INFO Thread-12 :196 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240226_092105-91uhs91a/files/output.log
276
+ 2024-02-26 09:25:09,470 INFO Thread-12 :196 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240226_092105-91uhs91a/files/output.log
277
+ 2024-02-26 09:25:11,847 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
278
+ 2024-02-26 09:25:12,228 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
279
+ 2024-02-26 09:25:16,848 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
280
+ 2024-02-26 09:25:17,229 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
281
+ 2024-02-26 09:25:21,849 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
282
+ 2024-02-26 09:25:22,230 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
283
+ 2024-02-26 09:25:26,850 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
284
+ 2024-02-26 09:25:27,247 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
285
+ 2024-02-26 09:25:31,851 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
286
+ 2024-02-26 09:25:32,284 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
287
+ 2024-02-26 09:25:32,489 DEBUG SenderThread:196 [sender.py:send():382] send: stats
288
+ 2024-02-26 09:25:34,173 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: partial_history
289
+ 2024-02-26 09:25:34,175 DEBUG SenderThread:196 [sender.py:send():382] send: metric
290
+ 2024-02-26 09:25:34,176 DEBUG SenderThread:196 [sender.py:send():382] send: metric
291
+ 2024-02-26 09:25:34,176 DEBUG SenderThread:196 [sender.py:send():382] send: metric
292
+ 2024-02-26 09:25:34,176 DEBUG SenderThread:196 [sender.py:send():382] send: metric
293
+ 2024-02-26 09:25:34,176 DEBUG SenderThread:196 [sender.py:send():382] send: metric
294
+ 2024-02-26 09:25:34,176 DEBUG SenderThread:196 [sender.py:send():382] send: metric
295
+ 2024-02-26 09:25:34,176 DEBUG SenderThread:196 [sender.py:send():382] send: metric
296
+ 2024-02-26 09:25:34,176 DEBUG SenderThread:196 [sender.py:send():382] send: history
297
+ 2024-02-26 09:25:34,177 DEBUG SenderThread:196 [sender.py:send_request():409] send_request: summary_record
298
+ 2024-02-26 09:25:34,177 INFO SenderThread:196 [sender.py:_save_file():1403] saving file wandb-summary.json with policy end
299
+ 2024-02-26 09:25:34,480 INFO Thread-12 :196 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240226_092105-91uhs91a/files/wandb-summary.json
300
+ 2024-02-26 09:25:37,178 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
301
+ 2024-02-26 09:25:37,295 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
302
+ 2024-02-26 09:25:42,215 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
303
+ 2024-02-26 09:25:42,302 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
304
+ 2024-02-26 09:25:42,484 INFO Thread-12 :196 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240226_092105-91uhs91a/files/config.yaml
305
+ 2024-02-26 09:25:44,419 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: partial_history
306
+ 2024-02-26 09:25:44,420 DEBUG SenderThread:196 [sender.py:send():382] send: metric
307
+ 2024-02-26 09:25:44,421 DEBUG SenderThread:196 [sender.py:send():382] send: metric
308
+ 2024-02-26 09:25:44,421 DEBUG SenderThread:196 [sender.py:send():382] send: history
309
+ 2024-02-26 09:25:44,421 DEBUG SenderThread:196 [sender.py:send_request():409] send_request: summary_record
310
+ 2024-02-26 09:25:44,421 INFO SenderThread:196 [sender.py:_save_file():1403] saving file wandb-summary.json with policy end
311
+ 2024-02-26 09:25:44,485 INFO Thread-12 :196 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240226_092105-91uhs91a/files/wandb-summary.json
312
+ 2024-02-26 09:25:47,401 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
313
+ 2024-02-26 09:25:47,422 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
314
+ 2024-02-26 09:25:52,424 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
315
+ 2024-02-26 09:25:52,494 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
316
+ 2024-02-26 09:25:57,424 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
317
+ 2024-02-26 09:25:57,510 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
318
+ 2024-02-26 09:26:02,425 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
319
+ 2024-02-26 09:26:02,490 DEBUG SenderThread:196 [sender.py:send():382] send: stats
320
+ 2024-02-26 09:26:02,510 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
321
+ 2024-02-26 09:26:07,491 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
322
+ 2024-02-26 09:26:07,512 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
323
+ 2024-02-26 09:26:12,498 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
324
+ 2024-02-26 09:26:12,518 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
325
+ 2024-02-26 09:26:13,499 INFO Thread-12 :196 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240226_092105-91uhs91a/files/config.yaml
326
+ 2024-02-26 09:26:17,520 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: keepalive
327
+ 2024-02-26 09:26:17,670 DEBUG HandlerThread:196 [handler.py:handle_request():146] handle_request: status_report
wandb/run-20240226_092105-91uhs91a/logs/debug.log ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-02-26 09:21:05,526 INFO MainThread:34 [wandb_setup.py:_flush():76] Current SDK version is 0.16.2
2
+ 2024-02-26 09:21:05,526 INFO MainThread:34 [wandb_setup.py:_flush():76] Configure stats pid to 34
3
+ 2024-02-26 09:21:05,526 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-02-26 09:21:05,526 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-02-26 09:21:05,526 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-02-26 09:21:05,527 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-02-26 09:21:05,527 INFO MainThread:34 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-02-26 09:21:05,527 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
9
+ 2024-02-26 09:21:05,527 INFO MainThread:34 [wandb_init.py:_log_setup():526] Logging user logs to /kaggle/working/wandb/run-20240226_092105-91uhs91a/logs/debug.log
10
+ 2024-02-26 09:21:05,527 INFO MainThread:34 [wandb_init.py:_log_setup():527] Logging internal logs to /kaggle/working/wandb/run-20240226_092105-91uhs91a/logs/debug-internal.log
11
+ 2024-02-26 09:21:05,527 INFO MainThread:34 [wandb_init.py:_jupyter_setup():472] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x77fcc2873700>
12
+ 2024-02-26 09:21:05,527 INFO MainThread:34 [wandb_init.py:init():566] calling init triggers
13
+ 2024-02-26 09:21:05,527 INFO MainThread:34 [wandb_init.py:init():573] wandb.init called with sweep_config: {}
14
+ config: {}
15
+ 2024-02-26 09:21:05,527 INFO MainThread:34 [wandb_init.py:init():616] starting backend
16
+ 2024-02-26 09:21:05,528 INFO MainThread:34 [wandb_init.py:init():620] setting up manager
17
+ 2024-02-26 09:21:05,529 INFO MainThread:34 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
+ 2024-02-26 09:21:05,532 INFO MainThread:34 [wandb_init.py:init():628] backend started and connected
19
+ 2024-02-26 09:21:05,543 INFO MainThread:34 [wandb_run.py:_label_probe_notebook():1294] probe notebook
20
+ 2024-02-26 09:21:06,207 INFO MainThread:34 [wandb_init.py:init():720] updated telemetry
21
+ 2024-02-26 09:21:06,210 INFO MainThread:34 [wandb_init.py:init():753] communicating run to backend with 90.0 second timeout
22
+ 2024-02-26 09:21:06,396 INFO MainThread:34 [wandb_run.py:_on_init():2254] communicating current version
23
+ 2024-02-26 09:21:06,458 INFO MainThread:34 [wandb_run.py:_on_init():2263] got version response upgrade_message: "wandb version 0.16.3 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
24
+
25
+ 2024-02-26 09:21:06,458 INFO MainThread:34 [wandb_init.py:init():804] starting run threads in backend
26
+ 2024-02-26 09:21:36,490 INFO MainThread:34 [wandb_run.py:_console_start():2233] atexit reg
27
+ 2024-02-26 09:21:36,490 INFO MainThread:34 [wandb_run.py:_redirect():2088] redirect: wrap_raw
28
+ 2024-02-26 09:21:36,491 INFO MainThread:34 [wandb_run.py:_redirect():2153] Wrapping output streams.
29
+ 2024-02-26 09:21:36,491 INFO MainThread:34 [wandb_run.py:_redirect():2178] Redirects installed.
30
+ 2024-02-26 09:21:36,492 INFO MainThread:34 [wandb_init.py:init():847] run started, returning control to user process
31
+ 2024-02-26 09:21:36,500 INFO MainThread:34 [wandb_run.py:_config_callback():1342] config_cb None None {'vocab_size': 32128, 'd_model': 768, 'd_kv': 64, 'd_ff': 2048, 'num_layers': 12, 'num_decoder_layers': 12, 'num_heads': 12, 'relative_attention_num_buckets': 32, 'relative_attention_max_distance': 128, 'dropout_rate': 0.1, 'classifier_dropout': 0.0, 'layer_norm_epsilon': 1e-06, 'initializer_factor': 1.0, 'feed_forward_proj': 'gated-gelu', 'use_cache': True, 'dense_act_fn': 'gelu_new', 'is_gated_act': True, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['T5ForConditionalGeneration'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': 1, 'sep_token_id': None, 'decoder_start_token_id': 0, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'csebuetnlp/banglat5', 'transformers_version': '4.37.0', 'gradient_checkpointing': False, 'model_type': 't5', 'output_past': True, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 300, 'per_device_eval_batch_size': 32, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0005, 'weight_decay': 0.1, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 5, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Feb26_09-20-49_d79522c38279', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 3, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': False, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': None, 'generation_num_beams': None, 'generation_config': None}
32
+ 2024-02-26 09:21:37,880 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
33
+ 2024-02-26 09:21:37,880 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
34
+ 2024-02-26 09:21:58,243 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
35
+ 2024-02-26 09:21:58,790 INFO MainThread:34 [wandb_run.py:_config_callback():1342] config_cb None None {'vocab_size': 32128, 'd_model': 768, 'd_kv': 64, 'd_ff': 2048, 'num_layers': 12, 'num_decoder_layers': 12, 'num_heads': 12, 'relative_attention_num_buckets': 32, 'relative_attention_max_distance': 128, 'dropout_rate': 0.1, 'classifier_dropout': 0.0, 'layer_norm_epsilon': 1e-06, 'initializer_factor': 1.0, 'feed_forward_proj': 'gated-gelu', 'use_cache': True, 'dense_act_fn': 'gelu_new', 'is_gated_act': True, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['T5ForConditionalGeneration'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': 1, 'sep_token_id': None, 'decoder_start_token_id': 0, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'csebuetnlp/banglat5', 'transformers_version': '4.37.0', 'gradient_checkpointing': False, 'model_type': 't5', 'output_past': True, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 300, 'per_device_eval_batch_size': 16, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0005, 'weight_decay': 0.1, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 5, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Feb26_09-21-58_d79522c38279', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 3, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': False, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': None, 'generation_num_beams': None, 'generation_config': None}
36
+ 2024-02-26 09:21:59,601 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
37
+ 2024-02-26 09:21:59,601 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
38
+ 2024-02-26 09:22:08,330 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
39
+ 2024-02-26 09:22:09,400 INFO MainThread:34 [wandb_run.py:_config_callback():1342] config_cb None None {'vocab_size': 32128, 'd_model': 768, 'd_kv': 64, 'd_ff': 2048, 'num_layers': 12, 'num_decoder_layers': 12, 'num_heads': 12, 'relative_attention_num_buckets': 32, 'relative_attention_max_distance': 128, 'dropout_rate': 0.1, 'classifier_dropout': 0.0, 'layer_norm_epsilon': 1e-06, 'initializer_factor': 1.0, 'feed_forward_proj': 'gated-gelu', 'use_cache': True, 'dense_act_fn': 'gelu_new', 'is_gated_act': True, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['T5ForConditionalGeneration'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': 1, 'sep_token_id': None, 'decoder_start_token_id': 0, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'csebuetnlp/banglat5', 'transformers_version': '4.37.0', 'gradient_checkpointing': False, 'model_type': 't5', 'output_past': True, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 300, 'per_device_eval_batch_size': 16, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0005, 'weight_decay': 0.1, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 5, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Feb26_09-22-08_d79522c38279', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 3, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': False, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': None, 'generation_num_beams': None, 'generation_config': None}
40
+ 2024-02-26 09:22:09,926 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
41
+ 2024-02-26 09:22:09,926 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
42
+ 2024-02-26 09:22:18,737 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
43
+ 2024-02-26 09:22:19,266 INFO MainThread:34 [wandb_run.py:_config_callback():1342] config_cb None None {'vocab_size': 32128, 'd_model': 768, 'd_kv': 64, 'd_ff': 2048, 'num_layers': 12, 'num_decoder_layers': 12, 'num_heads': 12, 'relative_attention_num_buckets': 32, 'relative_attention_max_distance': 128, 'dropout_rate': 0.1, 'classifier_dropout': 0.0, 'layer_norm_epsilon': 1e-06, 'initializer_factor': 1.0, 'feed_forward_proj': 'gated-gelu', 'use_cache': True, 'dense_act_fn': 'gelu_new', 'is_gated_act': True, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['T5ForConditionalGeneration'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': 1, 'sep_token_id': None, 'decoder_start_token_id': 0, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'csebuetnlp/banglat5', 'transformers_version': '4.37.0', 'gradient_checkpointing': False, 'model_type': 't5', 'output_past': True, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 300, 'per_device_eval_batch_size': 10, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0005, 'weight_decay': 0.1, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 5, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Feb26_09-22-18_d79522c38279', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 3, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': False, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': None, 'generation_num_beams': None, 'generation_config': None}
44
+ 2024-02-26 09:22:19,668 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
45
+ 2024-02-26 09:22:19,668 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
46
+ 2024-02-26 09:22:37,695 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
47
+ 2024-02-26 09:22:38,762 INFO MainThread:34 [wandb_run.py:_config_callback():1342] config_cb None None {'vocab_size': 32128, 'd_model': 768, 'd_kv': 64, 'd_ff': 2048, 'num_layers': 12, 'num_decoder_layers': 12, 'num_heads': 12, 'relative_attention_num_buckets': 32, 'relative_attention_max_distance': 128, 'dropout_rate': 0.1, 'classifier_dropout': 0.0, 'layer_norm_epsilon': 1e-06, 'initializer_factor': 1.0, 'feed_forward_proj': 'gated-gelu', 'use_cache': True, 'dense_act_fn': 'gelu_new', 'is_gated_act': True, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['T5ForConditionalGeneration'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': 1, 'sep_token_id': None, 'decoder_start_token_id': 0, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'csebuetnlp/banglat5', 'transformers_version': '4.37.0', 'gradient_checkpointing': False, 'model_type': 't5', 'output_past': True, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0005, 'weight_decay': 0.1, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 5, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Feb26_09-22-37_d79522c38279', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 3, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': False, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': None, 'generation_num_beams': None, 'generation_config': None}
48
+ 2024-02-26 09:23:19,820 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
49
+ 2024-02-26 09:23:19,820 INFO MainThread:34 [wandb_init.py:_pause_backend():437] pausing backend
50
+ 2024-02-26 09:23:32,480 INFO MainThread:34 [wandb_init.py:_resume_backend():442] resuming backend
51
+ 2024-02-26 09:23:33,079 INFO MainThread:34 [wandb_run.py:_config_callback():1342] config_cb None None {'vocab_size': 32128, 'd_model': 768, 'd_kv': 64, 'd_ff': 2048, 'num_layers': 12, 'num_decoder_layers': 12, 'num_heads': 12, 'relative_attention_num_buckets': 32, 'relative_attention_max_distance': 128, 'dropout_rate': 0.1, 'classifier_dropout': 0.0, 'layer_norm_epsilon': 1e-06, 'initializer_factor': 1.0, 'feed_forward_proj': 'gated-gelu', 'use_cache': True, 'dense_act_fn': 'gelu_new', 'is_gated_act': True, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': False, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['T5ForConditionalGeneration'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': 1, 'sep_token_id': None, 'decoder_start_token_id': 0, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'csebuetnlp/banglat5', 'transformers_version': '4.37.0', 'gradient_checkpointing': False, 'model_type': 't5', 'output_past': True, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 16, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0005, 'weight_decay': 0.1, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 5, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Feb26_09-23-32_d79522c38279', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 3, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': False, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': None, 'generation_num_beams': None, 'generation_config': None}
wandb/run-20240226_092105-91uhs91a/run-91uhs91a.wandb ADDED
Binary file (37.8 kB). View file