Shresthadev403 commited on
Commit
cae3d5c
1 Parent(s): ec78dbb

End of training

Browse files
README.md CHANGED
@@ -17,13 +17,13 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on the food101 dataset.
19
  It achieves the following results on the evaluation set:
20
- - eval_loss: 1.9754
21
- - eval_accuracy: 0.7336
22
- - eval_runtime: 156.8793
23
- - eval_samples_per_second: 96.571
24
- - eval_steps_per_second: 6.036
25
- - epoch: 8.45
26
- - step: 8000
27
 
28
  ## Model description
29
 
 
17
 
18
  This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on the food101 dataset.
19
  It achieves the following results on the evaluation set:
20
+ - eval_loss: 1.6087
21
+ - eval_accuracy: 0.7577
22
+ - eval_runtime: 154.5661
23
+ - eval_samples_per_second: 98.016
24
+ - eval_steps_per_second: 6.127
25
+ - epoch: 9.5
26
+ - step: 9000
27
 
28
  ## Model description
29
 
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b20edeb0fe7d3b161c0a46a58cae188fa221d032d8c7766255b9a02137a73c09
3
  size 343528508
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7003b62c05cc4a3a4b6bbe5ba0e3ccfad6a308b200d2885c2120cebe195b2270
3
  size 343528508
runs/Feb05_03-08-58_983d148b451b/events.out.tfevents.1707102539.983d148b451b.26.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a033a1a586ec82c34e1ea9c30d4c0e829e843a9fd4da33e9e4e02374982ce418
3
- size 13141
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64e1b4c293cf33cdc2de20d5b790df43e32e72d52f1f69cd50b68e6278944da9
3
+ size 13621
trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "best_metric": 0.7070627062706271,
3
- "best_model_checkpoint": "food-image-classification/checkpoint-7000",
4
- "epoch": 8.447729672650475,
5
  "eval_steps": 1000,
6
- "global_step": 8000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -127,6 +127,21 @@
127
  "eval_samples_per_second": 96.571,
128
  "eval_steps_per_second": 6.036,
129
  "step": 8000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
130
  }
131
  ],
132
  "logging_steps": 1000,
@@ -134,7 +149,7 @@
134
  "num_input_tokens_seen": 0,
135
  "num_train_epochs": 500,
136
  "save_steps": 1000,
137
- "total_flos": 3.970613944525455e+19,
138
  "train_batch_size": 16,
139
  "trial_name": null,
140
  "trial_params": null
 
1
  {
2
+ "best_metric": 0.7335973597359736,
3
+ "best_model_checkpoint": "food-image-classification/checkpoint-8000",
4
+ "epoch": 9.503695881731785,
5
  "eval_steps": 1000,
6
+ "global_step": 9000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
127
  "eval_samples_per_second": 96.571,
128
  "eval_steps_per_second": 6.036,
129
  "step": 8000
130
+ },
131
+ {
132
+ "epoch": 9.5,
133
+ "learning_rate": 9.503695881731786e-06,
134
+ "loss": 1.7383,
135
+ "step": 9000
136
+ },
137
+ {
138
+ "epoch": 9.5,
139
+ "eval_accuracy": 0.7576897689768977,
140
+ "eval_loss": 1.608676552772522,
141
+ "eval_runtime": 154.5661,
142
+ "eval_samples_per_second": 98.016,
143
+ "eval_steps_per_second": 6.127,
144
+ "step": 9000
145
  }
146
  ],
147
  "logging_steps": 1000,
 
149
  "num_input_tokens_seen": 0,
150
  "num_train_epochs": 500,
151
  "save_steps": 1000,
152
+ "total_flos": 4.466940687591137e+19,
153
  "train_batch_size": 16,
154
  "trial_name": null,
155
  "trial_params": null