IlyasMoutawwakil HF staff commited on
Commit
1c5c3b1
·
verified ·
1 Parent(s): c3e8f3d

Upload cuda_inference_transformers_image-classification_google/vit-base-patch16-224/benchmark_report.json with huggingface_hub

Browse files
cuda_inference_transformers_image-classification_google/vit-base-patch16-224/benchmark_report.json ADDED
@@ -0,0 +1,174 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "forward": {
3
+ "memory": {
4
+ "unit": "MB",
5
+ "max_ram": 1062.670336,
6
+ "max_global_vram": 778.469376,
7
+ "max_process_vram": 209335.001088,
8
+ "max_reserved": 406.847488,
9
+ "max_allocated": 355.912704
10
+ },
11
+ "latency": {
12
+ "unit": "s",
13
+ "count": 143,
14
+ "total": 0.9990407528877254,
15
+ "mean": 0.006986298971242837,
16
+ "stdev": 0.0001443188814090477,
17
+ "p50": 0.006934846878051758,
18
+ "p90": 0.007149181747436523,
19
+ "p95": 0.007313628101348877,
20
+ "p99": 0.007505675249099732,
21
+ "values": [
22
+ 0.006942845821380615,
23
+ 0.007180604934692383,
24
+ 0.007111485004425049,
25
+ 0.0073497238159179685,
26
+ 0.007481083869934082,
27
+ 0.00752348279953003,
28
+ 0.007455644130706787,
29
+ 0.007475004196166992,
30
+ 0.007381723880767822,
31
+ 0.007321084022521972,
32
+ 0.007548762798309326,
33
+ 0.007246524810791016,
34
+ 0.0071033258438110355,
35
+ 0.007045566082000732,
36
+ 0.007053246021270752,
37
+ 0.006943486213684082,
38
+ 0.006884127140045166,
39
+ 0.006886847019195556,
40
+ 0.006859167098999024,
41
+ 0.006889246940612793,
42
+ 0.006915326118469238,
43
+ 0.006969245910644531,
44
+ 0.006926365852355957,
45
+ 0.006936446189880371,
46
+ 0.006874046802520752,
47
+ 0.006913725852966309,
48
+ 0.006844606876373291,
49
+ 0.0069386868476867675,
50
+ 0.006914525985717773,
51
+ 0.006893246173858642,
52
+ 0.006950205802917481,
53
+ 0.00695948600769043,
54
+ 0.006927486896514893,
55
+ 0.006968286991119385,
56
+ 0.006970046997070312,
57
+ 0.00697516679763794,
58
+ 0.006980125904083252,
59
+ 0.0069804458618164065,
60
+ 0.006936286926269531,
61
+ 0.006927486896514893,
62
+ 0.006871325969696045,
63
+ 0.006905726909637451,
64
+ 0.006955326080322266,
65
+ 0.006953566074371338,
66
+ 0.006881566047668457,
67
+ 0.0069706859588623046,
68
+ 0.006882846832275391,
69
+ 0.006891646862030029,
70
+ 0.006857087135314941,
71
+ 0.006958207130432129,
72
+ 0.006931807041168213,
73
+ 0.0069130859375,
74
+ 0.0069130859375,
75
+ 0.006920445919036866,
76
+ 0.006903807163238526,
77
+ 0.006994845867156982,
78
+ 0.00696908712387085,
79
+ 0.006927166938781738,
80
+ 0.006939645767211914,
81
+ 0.006991806030273437,
82
+ 0.00690540599822998,
83
+ 0.006910366058349609,
84
+ 0.006935006141662598,
85
+ 0.006928606986999512,
86
+ 0.006905887126922607,
87
+ 0.006901725769042969,
88
+ 0.006863966941833496,
89
+ 0.006896286964416504,
90
+ 0.00695996618270874,
91
+ 0.007136925220489502,
92
+ 0.007140604972839355,
93
+ 0.00717148494720459,
94
+ 0.007081884860992432,
95
+ 0.007000606060028076,
96
+ 0.007147005081176758,
97
+ 0.007090365886688233,
98
+ 0.007124125957489014,
99
+ 0.00711084508895874,
100
+ 0.0070540461540222164,
101
+ 0.007067485809326172,
102
+ 0.007103324890136719,
103
+ 0.007208765029907227,
104
+ 0.007125085830688477,
105
+ 0.007149725914001465,
106
+ 0.007080605983734131,
107
+ 0.007035165786743164,
108
+ 0.007005886077880859,
109
+ 0.006896286964416504,
110
+ 0.006956285953521728,
111
+ 0.006862686157226562,
112
+ 0.006885087013244629,
113
+ 0.006934045791625976,
114
+ 0.006912605762481689,
115
+ 0.006877087116241455,
116
+ 0.006843007087707519,
117
+ 0.006908445835113525,
118
+ 0.006909566879272461,
119
+ 0.006879967212677002,
120
+ 0.006884925842285157,
121
+ 0.007120766162872314,
122
+ 0.0071666851043701175,
123
+ 0.00717420482635498,
124
+ 0.007039004802703857,
125
+ 0.006972445964813233,
126
+ 0.006934846878051758,
127
+ 0.006939486980438232,
128
+ 0.006928926944732666,
129
+ 0.006963645935058594,
130
+ 0.0069663658142089845,
131
+ 0.006972286224365234,
132
+ 0.00694364595413208,
133
+ 0.0069138860702514645,
134
+ 0.006944126129150391,
135
+ 0.006881086826324463,
136
+ 0.006894365787506103,
137
+ 0.006926846027374267,
138
+ 0.006875967025756836,
139
+ 0.0068996472358703615,
140
+ 0.0068814058303833005,
141
+ 0.006894207000732422,
142
+ 0.006886846065521241,
143
+ 0.00689212703704834,
144
+ 0.006879167079925537,
145
+ 0.007129724979400635,
146
+ 0.007118205070495606,
147
+ 0.0071254057884216305,
148
+ 0.006903326988220214,
149
+ 0.006890845775604248,
150
+ 0.006922206878662109,
151
+ 0.006946207046508789,
152
+ 0.006916767120361328,
153
+ 0.006866846084594726,
154
+ 0.0068623671531677245,
155
+ 0.0068497262001037595,
156
+ 0.006874046802520752,
157
+ 0.0068759660720825195,
158
+ 0.006901247024536133,
159
+ 0.006893086910247803,
160
+ 0.006881246089935303,
161
+ 0.006918207168579101,
162
+ 0.006835165977478027,
163
+ 0.006860126972198486,
164
+ 0.006869886875152588
165
+ ]
166
+ },
167
+ "throughput": {
168
+ "unit": "samples/s",
169
+ "value": 143.13730404556443
170
+ },
171
+ "energy": null,
172
+ "efficiency": null
173
+ }
174
+ }