khalilUoM commited on
Commit
c133389
1 Parent(s): 011ea93

End of training

Browse files
README.md ADDED
@@ -0,0 +1,229 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: google/vit-base-patch16-224
4
+ tags:
5
+ - generated_from_trainer
6
+ datasets:
7
+ - imagefolder
8
+ metrics:
9
+ - accuracy
10
+ - f1
11
+ - precision
12
+ - recall
13
+ model-index:
14
+ - name: physiotheraphy-E2
15
+ results:
16
+ - task:
17
+ name: Image Classification
18
+ type: image-classification
19
+ dataset:
20
+ name: imagefolder
21
+ type: imagefolder
22
+ config: default
23
+ split: train
24
+ args: default
25
+ metrics:
26
+ - name: Accuracy
27
+ type: accuracy
28
+ value: 0.9673024523160763
29
+ - name: F1
30
+ type: f1
31
+ value: 0.9684234987255815
32
+ - name: Precision
33
+ type: precision
34
+ value: 0.9707593418301198
35
+ - name: Recall
36
+ type: recall
37
+ value: 0.9667053446477023
38
+ ---
39
+
40
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
41
+ should probably proofread and complete it, then remove this comment. -->
42
+
43
+ # physiotheraphy-E2
44
+
45
+ This model is a fine-tuned version of [google/vit-base-patch16-224](https://huggingface.co/google/vit-base-patch16-224) on the imagefolder dataset.
46
+ It achieves the following results on the evaluation set:
47
+ - Accuracy: 0.9673
48
+ - F1: 0.9684
49
+ - Precision: 0.9708
50
+ - Recall: 0.9667
51
+ - Loss: 0.1718
52
+ - Classification Report: precision recall f1-score support
53
+
54
+ 0 0.92 0.95 0.93 57
55
+ 1 0.97 0.99 0.98 70
56
+ 2 0.97 1.00 0.99 33
57
+ 3 1.00 0.95 0.98 43
58
+ 4 0.97 1.00 0.99 34
59
+ 5 1.00 0.97 0.98 32
60
+ 6 0.97 0.97 0.97 65
61
+ 7 0.97 0.91 0.94 33
62
+
63
+ accuracy 0.97 367
64
+ macro avg 0.97 0.97 0.97 367
65
+ weighted avg 0.97 0.97 0.97 367
66
+
67
+ - Confusion Matrix: [[0.9473684210526315, 0.0, 0.0, 0.0, 0.017543859649122806, 0.0, 0.017543859649122806, 0.017543859649122806], [0.014285714285714285, 0.9857142857142858, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0], [0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0], [0.0, 0.046511627906976744, 0.0, 0.9534883720930233, 0.0, 0.0, 0.0, 0.0], [0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0], [0.03125, 0.0, 0.0, 0.0, 0.0, 0.96875, 0.0, 0.0], [0.03076923076923077, 0.0, 0.0, 0.0, 0.0, 0.0, 0.9692307692307692, 0.0], [0.030303030303030304, 0.0, 0.030303030303030304, 0.0, 0.0, 0.0, 0.030303030303030304, 0.9090909090909091]]
68
+
69
+ ## Model description
70
+
71
+ More information needed
72
+
73
+ ## Intended uses & limitations
74
+
75
+ More information needed
76
+
77
+ ## Training and evaluation data
78
+
79
+ More information needed
80
+
81
+ ## Training procedure
82
+
83
+ ### Training hyperparameters
84
+
85
+ The following hyperparameters were used during training:
86
+ - learning_rate: 0.0005
87
+ - train_batch_size: 4
88
+ - eval_batch_size: 4
89
+ - seed: 42
90
+ - gradient_accumulation_steps: 2
91
+ - total_train_batch_size: 8
92
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
93
+ - lr_scheduler_type: linear
94
+ - lr_scheduler_warmup_ratio: 0.1
95
+ - num_epochs: 8
96
+ - mixed_precision_training: Native AMP
97
+
98
+ ### Training results
99
+
100
+ | Training Loss | Epoch | Step | Accuracy | F1 | Precision | Recall | Validation Loss | Classification Report | Confusion Matrix |
101
+ |:-------------:|:------:|:----:|:--------:|:------:|:---------:|:------:|:---------------:|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------:|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------:|
102
+ | 0.9941 | 0.9973 | 182 | 0.6975 | 0.6724 | 0.7754 | 0.6769 | 0.9489 | precision recall f1-score support
103
+
104
+ 0 0.93 0.46 0.61 57
105
+ 1 0.92 0.79 0.85 70
106
+ 2 0.80 0.48 0.60 33
107
+ 3 0.86 0.70 0.77 43
108
+ 4 0.39 1.00 0.56 34
109
+ 5 0.74 0.72 0.73 32
110
+ 6 0.66 0.94 0.77 65
111
+ 7 0.92 0.33 0.49 33
112
+
113
+ accuracy 0.70 367
114
+ macro avg 0.78 0.68 0.67 367
115
+ weighted avg 0.79 0.70 0.70 367
116
+ | [[0.45614035087719296, 0.0, 0.05263157894736842, 0.03508771929824561, 0.24561403508771928, 0.05263157894736842, 0.15789473684210525, 0.0], [0.0, 0.7857142857142857, 0.0, 0.04285714285714286, 0.08571428571428572, 0.0, 0.07142857142857142, 0.014285714285714285], [0.0, 0.0, 0.48484848484848486, 0.0, 0.48484848484848486, 0.030303030303030304, 0.0, 0.0], [0.0, 0.023255813953488372, 0.0, 0.6976744186046512, 0.09302325581395349, 0.023255813953488372, 0.16279069767441862, 0.0], [0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0], [0.0, 0.0625, 0.0, 0.0, 0.21875, 0.71875, 0.0, 0.0], [0.015384615384615385, 0.0, 0.0, 0.0, 0.046153846153846156, 0.0, 0.9384615384615385, 0.0], [0.030303030303030304, 0.06060606060606061, 0.030303030303030304, 0.0, 0.12121212121212122, 0.09090909090909091, 0.3333333333333333, 0.3333333333333333]] |
117
+ | 0.6919 | 2.0 | 365 | 0.8665 | 0.8633 | 0.8600 | 0.8742 | 0.4393 | precision recall f1-score support
118
+
119
+ 0 0.84 0.63 0.72 57
120
+ 1 0.86 0.93 0.89 70
121
+ 2 0.84 0.97 0.90 33
122
+ 3 1.00 0.95 0.98 43
123
+ 4 0.89 1.00 0.94 34
124
+ 5 0.85 0.91 0.88 32
125
+ 6 0.97 0.88 0.92 65
126
+ 7 0.63 0.73 0.68 33
127
+
128
+ accuracy 0.87 367
129
+ macro avg 0.86 0.87 0.86 367
130
+ weighted avg 0.87 0.87 0.86 367
131
+ | [[0.631578947368421, 0.08771929824561403, 0.08771929824561403, 0.0, 0.03508771929824561, 0.07017543859649122, 0.0, 0.08771929824561403], [0.02857142857142857, 0.9285714285714286, 0.0, 0.0, 0.014285714285714285, 0.014285714285714285, 0.0, 0.014285714285714285], [0.0, 0.0, 0.9696969696969697, 0.0, 0.030303030303030304, 0.0, 0.0, 0.0], [0.023255813953488372, 0.0, 0.0, 0.9534883720930233, 0.0, 0.0, 0.0, 0.023255813953488372], [0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0], [0.0, 0.0, 0.0, 0.0, 0.0, 0.90625, 0.0, 0.09375], [0.03076923076923077, 0.03076923076923077, 0.0, 0.0, 0.0, 0.0, 0.8769230769230769, 0.06153846153846154], [0.06060606060606061, 0.12121212121212122, 0.030303030303030304, 0.0, 0.0, 0.0, 0.06060606060606061, 0.7272727272727273]] |
132
+ | 0.4322 | 2.9973 | 547 | 0.8501 | 0.8412 | 0.8687 | 0.8387 | 0.6005 | precision recall f1-score support
133
+
134
+ 0 0.78 0.95 0.86 57
135
+ 1 1.00 0.76 0.86 70
136
+ 2 0.96 0.76 0.85 33
137
+ 3 0.83 0.91 0.87 43
138
+ 4 0.72 1.00 0.84 34
139
+ 5 0.92 0.75 0.83 32
140
+ 6 0.82 0.95 0.88 65
141
+ 7 0.91 0.64 0.75 33
142
+
143
+ accuracy 0.85 367
144
+ macro avg 0.87 0.84 0.84 367
145
+ weighted avg 0.87 0.85 0.85 367
146
+ | [[0.9473684210526315, 0.0, 0.0, 0.03508771929824561, 0.0, 0.0, 0.017543859649122806, 0.0], [0.05714285714285714, 0.7571428571428571, 0.0, 0.04285714285714286, 0.05714285714285714, 0.014285714285714285, 0.04285714285714286, 0.02857142857142857], [0.030303030303030304, 0.0, 0.7575757575757576, 0.0, 0.12121212121212122, 0.030303030303030304, 0.06060606060606061, 0.0], [0.046511627906976744, 0.0, 0.0, 0.9069767441860465, 0.0, 0.0, 0.046511627906976744, 0.0], [0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0], [0.09375, 0.0, 0.0, 0.0, 0.15625, 0.75, 0.0, 0.0], [0.015384615384615385, 0.0, 0.0, 0.03076923076923077, 0.0, 0.0, 0.9538461538461539, 0.0], [0.12121212121212122, 0.0, 0.030303030303030304, 0.030303030303030304, 0.0, 0.0, 0.18181818181818182, 0.6363636363636364]] |
147
+ | 0.2358 | 4.0 | 730 | 0.9401 | 0.9392 | 0.9461 | 0.9370 | 0.2496 | precision recall f1-score support
148
+
149
+ 0 0.82 0.96 0.89 57
150
+ 1 1.00 0.91 0.96 70
151
+ 2 1.00 0.94 0.97 33
152
+ 3 0.95 0.98 0.97 43
153
+ 4 1.00 0.85 0.92 34
154
+ 5 0.89 1.00 0.94 32
155
+ 6 0.97 0.97 0.97 65
156
+ 7 0.94 0.88 0.91 33
157
+
158
+ accuracy 0.94 367
159
+ macro avg 0.95 0.94 0.94 367
160
+ weighted avg 0.95 0.94 0.94 367
161
+ | [[0.9649122807017544, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.03508771929824561], [0.02857142857142857, 0.9142857142857143, 0.0, 0.014285714285714285, 0.0, 0.04285714285714286, 0.0, 0.0], [0.06060606060606061, 0.0, 0.9393939393939394, 0.0, 0.0, 0.0, 0.0, 0.0], [0.023255813953488372, 0.0, 0.0, 0.9767441860465116, 0.0, 0.0, 0.0, 0.0], [0.11764705882352941, 0.0, 0.0, 0.0, 0.8529411764705882, 0.029411764705882353, 0.0, 0.0], [0.0, 0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0], [0.03076923076923077, 0.0, 0.0, 0.0, 0.0, 0.0, 0.9692307692307692, 0.0], [0.030303030303030304, 0.0, 0.0, 0.030303030303030304, 0.0, 0.0, 0.06060606060606061, 0.8787878787878788]] |
162
+ | 0.0904 | 4.9973 | 912 | 0.9401 | 0.9448 | 0.9506 | 0.9429 | 0.2831 | precision recall f1-score support
163
+
164
+ 0 0.79 0.98 0.88 57
165
+ 1 0.98 0.93 0.96 70
166
+ 2 1.00 0.94 0.97 33
167
+ 3 1.00 0.95 0.98 43
168
+ 4 0.97 1.00 0.99 34
169
+ 5 0.97 0.94 0.95 32
170
+ 6 0.98 0.89 0.94 65
171
+ 7 0.91 0.91 0.91 33
172
+
173
+ accuracy 0.94 367
174
+ macro avg 0.95 0.94 0.94 367
175
+ weighted avg 0.95 0.94 0.94 367
176
+ | [[0.9824561403508771, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.017543859649122806], [0.04285714285714286, 0.9285714285714286, 0.0, 0.0, 0.0, 0.014285714285714285, 0.0, 0.014285714285714285], [0.030303030303030304, 0.0, 0.9393939393939394, 0.0, 0.030303030303030304, 0.0, 0.0, 0.0], [0.046511627906976744, 0.0, 0.0, 0.9534883720930233, 0.0, 0.0, 0.0, 0.0], [0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0], [0.03125, 0.03125, 0.0, 0.0, 0.0, 0.9375, 0.0, 0.0], [0.09230769230769231, 0.0, 0.0, 0.0, 0.0, 0.0, 0.8923076923076924, 0.015384615384615385], [0.06060606060606061, 0.0, 0.0, 0.0, 0.0, 0.0, 0.030303030303030304, 0.9090909090909091]] |
177
+ | 0.0313 | 6.0 | 1095 | 0.9673 | 0.9684 | 0.9708 | 0.9667 | 0.1718 | precision recall f1-score support
178
+
179
+ 0 0.92 0.95 0.93 57
180
+ 1 0.97 0.99 0.98 70
181
+ 2 0.97 1.00 0.99 33
182
+ 3 1.00 0.95 0.98 43
183
+ 4 0.97 1.00 0.99 34
184
+ 5 1.00 0.97 0.98 32
185
+ 6 0.97 0.97 0.97 65
186
+ 7 0.97 0.91 0.94 33
187
+
188
+ accuracy 0.97 367
189
+ macro avg 0.97 0.97 0.97 367
190
+ weighted avg 0.97 0.97 0.97 367
191
+ | [[0.9473684210526315, 0.0, 0.0, 0.0, 0.017543859649122806, 0.0, 0.017543859649122806, 0.017543859649122806], [0.014285714285714285, 0.9857142857142858, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0], [0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0], [0.0, 0.046511627906976744, 0.0, 0.9534883720930233, 0.0, 0.0, 0.0, 0.0], [0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0], [0.03125, 0.0, 0.0, 0.0, 0.0, 0.96875, 0.0, 0.0], [0.03076923076923077, 0.0, 0.0, 0.0, 0.0, 0.0, 0.9692307692307692, 0.0], [0.030303030303030304, 0.0, 0.030303030303030304, 0.0, 0.0, 0.0, 0.030303030303030304, 0.9090909090909091]] |
192
+ | 0.0047 | 6.9973 | 1277 | 0.9646 | 0.9630 | 0.9675 | 0.9604 | 0.1481 | precision recall f1-score support
193
+
194
+ 0 0.92 0.96 0.94 57
195
+ 1 1.00 0.99 0.99 70
196
+ 2 0.97 1.00 0.99 33
197
+ 3 0.98 0.98 0.98 43
198
+ 4 0.97 1.00 0.99 34
199
+ 5 1.00 0.97 0.98 32
200
+ 6 0.94 0.97 0.95 65
201
+ 7 0.96 0.82 0.89 33
202
+
203
+ accuracy 0.96 367
204
+ macro avg 0.97 0.96 0.96 367
205
+ weighted avg 0.97 0.96 0.96 367
206
+ | [[0.9649122807017544, 0.0, 0.0, 0.0, 0.017543859649122806, 0.0, 0.0, 0.017543859649122806], [0.0, 0.9857142857142858, 0.0, 0.014285714285714285, 0.0, 0.0, 0.0, 0.0], [0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0], [0.023255813953488372, 0.0, 0.0, 0.9767441860465116, 0.0, 0.0, 0.0, 0.0], [0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0], [0.03125, 0.0, 0.0, 0.0, 0.0, 0.96875, 0.0, 0.0], [0.03076923076923077, 0.0, 0.0, 0.0, 0.0, 0.0, 0.9692307692307692, 0.0], [0.030303030303030304, 0.0, 0.030303030303030304, 0.0, 0.0, 0.0, 0.12121212121212122, 0.8181818181818182]] |
207
+ | 0.0019 | 7.9781 | 1456 | 0.9646 | 0.9630 | 0.9675 | 0.9604 | 0.1477 | precision recall f1-score support
208
+
209
+ 0 0.92 0.96 0.94 57
210
+ 1 1.00 0.99 0.99 70
211
+ 2 0.97 1.00 0.99 33
212
+ 3 0.98 0.98 0.98 43
213
+ 4 0.97 1.00 0.99 34
214
+ 5 1.00 0.97 0.98 32
215
+ 6 0.94 0.97 0.95 65
216
+ 7 0.96 0.82 0.89 33
217
+
218
+ accuracy 0.96 367
219
+ macro avg 0.97 0.96 0.96 367
220
+ weighted avg 0.97 0.96 0.96 367
221
+ | [[0.9649122807017544, 0.0, 0.0, 0.0, 0.017543859649122806, 0.0, 0.0, 0.017543859649122806], [0.0, 0.9857142857142858, 0.0, 0.014285714285714285, 0.0, 0.0, 0.0, 0.0], [0.0, 0.0, 1.0, 0.0, 0.0, 0.0, 0.0, 0.0], [0.023255813953488372, 0.0, 0.0, 0.9767441860465116, 0.0, 0.0, 0.0, 0.0], [0.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0], [0.03125, 0.0, 0.0, 0.0, 0.0, 0.96875, 0.0, 0.0], [0.03076923076923077, 0.0, 0.0, 0.0, 0.0, 0.0, 0.9692307692307692, 0.0], [0.030303030303030304, 0.0, 0.030303030303030304, 0.0, 0.0, 0.0, 0.12121212121212122, 0.8181818181818182]] |
222
+
223
+
224
+ ### Framework versions
225
+
226
+ - Transformers 4.43.3
227
+ - Pytorch 2.3.1+cu121
228
+ - Datasets 2.20.0
229
+ - Tokenizers 0.19.1
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9395d2365280ddf6fa37711bc7d46cec9b90b77a588e2c199255312d2dc4839a
3
  size 343242432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e95e52bb0e50a7540b3960a6516e389621ac4e674f392f3c1715f194adfbc955
3
  size 343242432
runs/Jul28_08-18-42_792806acba88/events.out.tfevents.1722155627.792806acba88.672.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43b962b06be2900e3ee4ce6b1e3050604f87c118f6a4c6395ca3727b535aa50b
3
+ size 1032