sagarpatelcompstak commited on
Commit
1ec280f
1 Parent(s): c77cdaf

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+ tags:
4
+ - autotrain
5
+ - image-classification
6
+ base_model: google/vit-base-patch16-224
7
+ widget:
8
+ - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/tiger.jpg
9
+ example_title: Tiger
10
+ - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/teapot.jpg
11
+ example_title: Teapot
12
+ - src: https://huggingface.co/datasets/mishig/sample_images/resolve/main/palace.jpg
13
+ example_title: Palace
14
+ ---
15
+
16
+ # Model Trained Using AutoTrain
17
+
18
+ - Problem type: Image Classification
19
+
20
+ ## Validation Metrics
21
+ loss: 0.326568067073822
22
+
23
+ f1_macro: 0.574992771283407
24
+
25
+ f1_micro: 0.8994565217391305
26
+
27
+ f1_weighted: 0.8962362631935303
28
+
29
+ precision_macro: 0.5852941598199051
30
+
31
+ precision_micro: 0.8994565217391305
32
+
33
+ precision_weighted: 0.8948775064010129
34
+
35
+ recall_macro: 0.566717099325795
36
+
37
+ recall_micro: 0.8994565217391305
38
+
39
+ recall_weighted: 0.8994565217391305
40
+
41
+ accuracy: 0.8994565217391305
checkpoint-184/config.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google/vit-base-patch16-224",
3
+ "_num_labels": 3,
4
+ "architectures": [
5
+ "ViTForImageClassification"
6
+ ],
7
+ "attention_probs_dropout_prob": 0.0,
8
+ "encoder_stride": 16,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.0,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "lease",
14
+ "1": "rentRoll",
15
+ "2": "sales"
16
+ },
17
+ "image_size": 224,
18
+ "initializer_range": 0.02,
19
+ "intermediate_size": 3072,
20
+ "label2id": {
21
+ "lease": 0,
22
+ "rentRoll": 1,
23
+ "sales": 2
24
+ },
25
+ "layer_norm_eps": 1e-12,
26
+ "model_type": "vit",
27
+ "num_attention_heads": 12,
28
+ "num_channels": 3,
29
+ "num_hidden_layers": 12,
30
+ "patch_size": 16,
31
+ "problem_type": "single_label_classification",
32
+ "qkv_bias": true,
33
+ "torch_dtype": "float32",
34
+ "transformers_version": "4.44.2"
35
+ }
checkpoint-184/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:321aec42847fa5df155c6b2a74779d170ea2a904c4cddb94fb32fba21372cb09
3
+ size 343227052
checkpoint-184/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b0086819a23da6a96416002441ceb1038f1bbc702c187ae2edc4ac48438a2e9
3
+ size 686569402
checkpoint-184/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eaaac4e2bcc9096629c5b9cd790ebb3b0f642e25b055f079a52061bb44c4f5ce
3
+ size 13990
checkpoint-184/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e7d646de7dbb44840773538562e8f963df7089795a67618e3ba1dc9107749d8f
3
+ size 1064
checkpoint-184/trainer_state.json ADDED
@@ -0,0 +1,200 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.326568067073822,
3
+ "best_model_checkpoint": "sep-20/checkpoint-184",
4
+ "epoch": 1.0,
5
+ "eval_steps": 500,
6
+ "global_step": 184,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.04891304347826087,
13
+ "grad_norm": 6.508527755737305,
14
+ "learning_rate": 2.368421052631579e-05,
15
+ "loss": 0.9057,
16
+ "step": 9
17
+ },
18
+ {
19
+ "epoch": 0.09782608695652174,
20
+ "grad_norm": 6.560257434844971,
21
+ "learning_rate": 4.736842105263158e-05,
22
+ "loss": 0.4782,
23
+ "step": 18
24
+ },
25
+ {
26
+ "epoch": 0.14673913043478262,
27
+ "grad_norm": 5.121021270751953,
28
+ "learning_rate": 4.7575757575757576e-05,
29
+ "loss": 0.3964,
30
+ "step": 27
31
+ },
32
+ {
33
+ "epoch": 0.1956521739130435,
34
+ "grad_norm": 5.416026592254639,
35
+ "learning_rate": 4.484848484848485e-05,
36
+ "loss": 0.461,
37
+ "step": 36
38
+ },
39
+ {
40
+ "epoch": 0.24456521739130435,
41
+ "grad_norm": 3.697016716003418,
42
+ "learning_rate": 4.212121212121212e-05,
43
+ "loss": 0.6009,
44
+ "step": 45
45
+ },
46
+ {
47
+ "epoch": 0.29347826086956524,
48
+ "grad_norm": 3.856910467147827,
49
+ "learning_rate": 3.939393939393939e-05,
50
+ "loss": 0.2506,
51
+ "step": 54
52
+ },
53
+ {
54
+ "epoch": 0.3423913043478261,
55
+ "grad_norm": 14.140103340148926,
56
+ "learning_rate": 3.6666666666666666e-05,
57
+ "loss": 0.396,
58
+ "step": 63
59
+ },
60
+ {
61
+ "epoch": 0.391304347826087,
62
+ "grad_norm": 4.570571422576904,
63
+ "learning_rate": 3.3939393939393945e-05,
64
+ "loss": 0.3124,
65
+ "step": 72
66
+ },
67
+ {
68
+ "epoch": 0.44021739130434784,
69
+ "grad_norm": 11.288423538208008,
70
+ "learning_rate": 3.121212121212122e-05,
71
+ "loss": 0.5017,
72
+ "step": 81
73
+ },
74
+ {
75
+ "epoch": 0.4891304347826087,
76
+ "grad_norm": 10.446467399597168,
77
+ "learning_rate": 2.8484848484848486e-05,
78
+ "loss": 0.2368,
79
+ "step": 90
80
+ },
81
+ {
82
+ "epoch": 0.5380434782608695,
83
+ "grad_norm": 1.5973994731903076,
84
+ "learning_rate": 2.575757575757576e-05,
85
+ "loss": 0.2517,
86
+ "step": 99
87
+ },
88
+ {
89
+ "epoch": 0.5869565217391305,
90
+ "grad_norm": 0.7517912983894348,
91
+ "learning_rate": 2.3030303030303034e-05,
92
+ "loss": 0.571,
93
+ "step": 108
94
+ },
95
+ {
96
+ "epoch": 0.6358695652173914,
97
+ "grad_norm": 7.41497278213501,
98
+ "learning_rate": 2.0303030303030303e-05,
99
+ "loss": 0.2846,
100
+ "step": 117
101
+ },
102
+ {
103
+ "epoch": 0.6847826086956522,
104
+ "grad_norm": 4.097425937652588,
105
+ "learning_rate": 1.7575757575757576e-05,
106
+ "loss": 0.4801,
107
+ "step": 126
108
+ },
109
+ {
110
+ "epoch": 0.7336956521739131,
111
+ "grad_norm": 4.8678669929504395,
112
+ "learning_rate": 1.484848484848485e-05,
113
+ "loss": 0.3254,
114
+ "step": 135
115
+ },
116
+ {
117
+ "epoch": 0.782608695652174,
118
+ "grad_norm": 0.7251734733581543,
119
+ "learning_rate": 1.2121212121212122e-05,
120
+ "loss": 0.1278,
121
+ "step": 144
122
+ },
123
+ {
124
+ "epoch": 0.8315217391304348,
125
+ "grad_norm": 10.660299301147461,
126
+ "learning_rate": 9.393939393939394e-06,
127
+ "loss": 0.5605,
128
+ "step": 153
129
+ },
130
+ {
131
+ "epoch": 0.8804347826086957,
132
+ "grad_norm": 0.7391616702079773,
133
+ "learning_rate": 6.666666666666667e-06,
134
+ "loss": 0.3178,
135
+ "step": 162
136
+ },
137
+ {
138
+ "epoch": 0.9293478260869565,
139
+ "grad_norm": 0.3952378034591675,
140
+ "learning_rate": 3.939393939393939e-06,
141
+ "loss": 0.3842,
142
+ "step": 171
143
+ },
144
+ {
145
+ "epoch": 0.9782608695652174,
146
+ "grad_norm": 5.111515522003174,
147
+ "learning_rate": 1.2121212121212122e-06,
148
+ "loss": 0.2226,
149
+ "step": 180
150
+ },
151
+ {
152
+ "epoch": 1.0,
153
+ "eval_accuracy": 0.8994565217391305,
154
+ "eval_f1_macro": 0.574992771283407,
155
+ "eval_f1_micro": 0.8994565217391305,
156
+ "eval_f1_weighted": 0.8962362631935303,
157
+ "eval_loss": 0.326568067073822,
158
+ "eval_precision_macro": 0.5852941598199051,
159
+ "eval_precision_micro": 0.8994565217391305,
160
+ "eval_precision_weighted": 0.8948775064010129,
161
+ "eval_recall_macro": 0.566717099325795,
162
+ "eval_recall_micro": 0.8994565217391305,
163
+ "eval_recall_weighted": 0.8994565217391305,
164
+ "eval_runtime": 108.1675,
165
+ "eval_samples_per_second": 3.402,
166
+ "eval_steps_per_second": 0.213,
167
+ "step": 184
168
+ }
169
+ ],
170
+ "logging_steps": 9,
171
+ "max_steps": 184,
172
+ "num_input_tokens_seen": 0,
173
+ "num_train_epochs": 1,
174
+ "save_steps": 500,
175
+ "stateful_callbacks": {
176
+ "EarlyStoppingCallback": {
177
+ "args": {
178
+ "early_stopping_patience": 5,
179
+ "early_stopping_threshold": 0.01
180
+ },
181
+ "attributes": {
182
+ "early_stopping_patience_counter": 0
183
+ }
184
+ },
185
+ "TrainerControl": {
186
+ "args": {
187
+ "should_epoch_stop": false,
188
+ "should_evaluate": false,
189
+ "should_log": false,
190
+ "should_save": true,
191
+ "should_training_stop": true
192
+ },
193
+ "attributes": {}
194
+ }
195
+ },
196
+ "total_flos": 1.138367530175017e+17,
197
+ "train_batch_size": 8,
198
+ "trial_name": null,
199
+ "trial_params": null
200
+ }
checkpoint-184/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:559a65ed63eece12124be46dd04ca3a3e8f546fdff3a88642240a16b7465f684
3
+ size 5240
config.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google/vit-base-patch16-224",
3
+ "_num_labels": 3,
4
+ "architectures": [
5
+ "ViTForImageClassification"
6
+ ],
7
+ "attention_probs_dropout_prob": 0.0,
8
+ "encoder_stride": 16,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.0,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "lease",
14
+ "1": "rentRoll",
15
+ "2": "sales"
16
+ },
17
+ "image_size": 224,
18
+ "initializer_range": 0.02,
19
+ "intermediate_size": 3072,
20
+ "label2id": {
21
+ "lease": 0,
22
+ "rentRoll": 1,
23
+ "sales": 2
24
+ },
25
+ "layer_norm_eps": 1e-12,
26
+ "model_type": "vit",
27
+ "num_attention_heads": 12,
28
+ "num_channels": 3,
29
+ "num_hidden_layers": 12,
30
+ "patch_size": 16,
31
+ "problem_type": "single_label_classification",
32
+ "qkv_bias": true,
33
+ "torch_dtype": "float32",
34
+ "transformers_version": "4.44.2"
35
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:321aec42847fa5df155c6b2a74779d170ea2a904c4cddb94fb32fba21372cb09
3
+ size 343227052
preprocessor_config.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_normalize": true,
3
+ "do_rescale": true,
4
+ "do_resize": true,
5
+ "image_mean": [
6
+ 0.5,
7
+ 0.5,
8
+ 0.5
9
+ ],
10
+ "image_processor_type": "ViTImageProcessor",
11
+ "image_std": [
12
+ 0.5,
13
+ 0.5,
14
+ 0.5
15
+ ],
16
+ "resample": 2,
17
+ "rescale_factor": 0.00392156862745098,
18
+ "size": {
19
+ "height": 224,
20
+ "width": 224
21
+ }
22
+ }
runs/Sep20_19-10-03_r-sagarpatelcompstak-submission-classificaiton-lekwi2-57c52-aqq/events.out.tfevents.1726859405.r-sagarpatelcompstak-submission-classificaiton-lekwi2-57c52-aqq.178.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f9ce3c5af6234ac027a62cfa2d8a72e60cf214a44ae676ffdbb7c59b88dea1ed
3
- size 7045
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:56e1f1d1fe3a5ea2292128d6d0ac6498ab7a30e2d104a5372c5d3fbe1f962f36
3
+ size 10326
runs/Sep20_19-10-03_r-sagarpatelcompstak-submission-classificaiton-lekwi2-57c52-aqq/events.out.tfevents.1726860900.r-sagarpatelcompstak-submission-classificaiton-lekwi2-57c52-aqq.178.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ffceb48fbeba8b546959c6536125905ff057efb63cc350ab80211363d017f5d
3
+ size 921
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:559a65ed63eece12124be46dd04ca3a3e8f546fdff3a88642240a16b7465f684
3
+ size 5240
training_params.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "data_path": "sep-20/autotrain-data",
3
+ "model": "google/vit-base-patch16-224",
4
+ "username": "sagarpatelcompstak",
5
+ "lr": 5e-05,
6
+ "epochs": 1,
7
+ "batch_size": 8,
8
+ "warmup_ratio": 0.1,
9
+ "gradient_accumulation": 1,
10
+ "optimizer": "adamw_torch",
11
+ "scheduler": "linear",
12
+ "weight_decay": 0.0,
13
+ "max_grad_norm": 1.0,
14
+ "seed": 42,
15
+ "train_split": "train",
16
+ "valid_split": "validation",
17
+ "logging_steps": -1,
18
+ "project_name": "sep-20",
19
+ "auto_find_batch_size": false,
20
+ "mixed_precision": "fp16",
21
+ "save_total_limit": 1,
22
+ "push_to_hub": true,
23
+ "eval_strategy": "epoch",
24
+ "image_column": "autotrain_image",
25
+ "target_column": "autotrain_label",
26
+ "log": "tensorboard",
27
+ "early_stopping_patience": 5,
28
+ "early_stopping_threshold": 0.01
29
+ }