Upload cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json with huggingface_hub
Browse files
cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json
CHANGED
@@ -84,7 +84,7 @@
|
|
84 |
],
|
85 |
"gpu_count": 1,
|
86 |
"gpu_vram_mb": 24146608128,
|
87 |
-
"optimum_benchmark_version": "0.
|
88 |
"optimum_benchmark_commit": null,
|
89 |
"transformers_version": "4.42.3",
|
90 |
"transformers_commit": null,
|
@@ -104,7 +104,7 @@
|
|
104 |
"forward": {
|
105 |
"memory": {
|
106 |
"unit": "MB",
|
107 |
-
"max_ram": 976.
|
108 |
"max_global_vram": 1434.976256,
|
109 |
"max_process_vram": 0.0,
|
110 |
"max_reserved": 794.820608,
|
@@ -112,101 +112,104 @@
|
|
112 |
},
|
113 |
"latency": {
|
114 |
"unit": "s",
|
115 |
-
"count":
|
116 |
-
"total": 1.
|
117 |
-
"mean": 0.
|
118 |
-
"stdev": 0.
|
119 |
-
"p50": 0.
|
120 |
-
"p90": 0.
|
121 |
-
"p95": 0.
|
122 |
-
"p99": 0.
|
123 |
"values": [
|
124 |
-
0.
|
125 |
-
0.
|
126 |
-
0.
|
127 |
-
0.
|
128 |
-
0.
|
129 |
-
0.
|
130 |
-
0.
|
131 |
-
0.
|
132 |
-
0.
|
133 |
-
0.
|
134 |
-
0.
|
135 |
-
0.
|
136 |
-
0.
|
137 |
-
0.
|
138 |
-
0.
|
139 |
-
0.
|
140 |
-
0.
|
141 |
-
0.
|
142 |
-
0.014375935554504395,
|
143 |
-
0.014260224342346191,
|
144 |
-
0.014311424255371094,
|
145 |
-
0.014268416404724121,
|
146 |
-
0.014050304412841797,
|
147 |
-
0.014303232192993164,
|
148 |
-
0.014225407600402832,
|
149 |
-
0.01407487964630127,
|
150 |
-
0.014136320114135742,
|
151 |
-
0.014467071533203125,
|
152 |
-
0.014335040092468262,
|
153 |
-
0.014097439765930175,
|
154 |
-
0.01406668758392334,
|
155 |
-
0.01491763210296631,
|
156 |
-
0.01459712028503418,
|
157 |
-
0.014526464462280274,
|
158 |
-
0.014592000007629394,
|
159 |
-
0.015723520278930665,
|
160 |
-
0.01721343994140625,
|
161 |
-
0.01546444797515869,
|
162 |
-
0.014766079902648926,
|
163 |
-
0.014264320373535156,
|
164 |
-
0.014220288276672363,
|
165 |
-
0.014145536422729492,
|
166 |
-
0.014108672142028808,
|
167 |
-
0.014125056266784668,
|
168 |
-
0.014125056266784668,
|
169 |
-
0.01426534366607666,
|
170 |
-
0.014252032279968262,
|
171 |
-
0.014897151947021485,
|
172 |
-
0.01417728042602539,
|
173 |
-
0.014160896301269531,
|
174 |
-
0.014171135902404786,
|
175 |
-
0.014114815711975098,
|
176 |
-
0.014152704238891601,
|
177 |
-
0.01451417636871338,
|
178 |
-
0.014017536163330077,
|
179 |
-
0.013896703720092773,
|
180 |
-
0.013942784309387207,
|
181 |
-
0.014303168296813966,
|
182 |
-
0.014135295867919923,
|
183 |
0.014082048416137695,
|
184 |
-
0.
|
185 |
-
0.
|
186 |
-
0.
|
187 |
-
0.
|
188 |
-
0.
|
189 |
-
0.
|
190 |
-
0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
191 |
0.014101504325866699,
|
192 |
-
0.
|
193 |
-
0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
194 |
]
|
195 |
},
|
196 |
"throughput": {
|
197 |
"unit": "samples/s",
|
198 |
-
"value":
|
199 |
},
|
200 |
"energy": {
|
201 |
"unit": "kWh",
|
202 |
-
"cpu": 1.
|
203 |
-
"ram":
|
204 |
-
"gpu": 3.
|
205 |
-
"total": 6.
|
206 |
},
|
207 |
"efficiency": {
|
208 |
"unit": "samples/kWh",
|
209 |
-
"value":
|
210 |
}
|
211 |
}
|
212 |
}
|
|
|
84 |
],
|
85 |
"gpu_count": 1,
|
86 |
"gpu_vram_mb": 24146608128,
|
87 |
+
"optimum_benchmark_version": "0.3.0",
|
88 |
"optimum_benchmark_commit": null,
|
89 |
"transformers_version": "4.42.3",
|
90 |
"transformers_commit": null,
|
|
|
104 |
"forward": {
|
105 |
"memory": {
|
106 |
"unit": "MB",
|
107 |
+
"max_ram": 976.187392,
|
108 |
"max_global_vram": 1434.976256,
|
109 |
"max_process_vram": 0.0,
|
110 |
"max_reserved": 794.820608,
|
|
|
112 |
},
|
113 |
"latency": {
|
114 |
"unit": "s",
|
115 |
+
"count": 73,
|
116 |
+
"total": 1.0094481916427611,
|
117 |
+
"mean": 0.013828057419763852,
|
118 |
+
"stdev": 0.000524276007705457,
|
119 |
+
"p50": 0.013813759803771973,
|
120 |
+
"p90": 0.014313472175598144,
|
121 |
+
"p95": 0.01499320316314697,
|
122 |
+
"p99": 0.01558552604675293,
|
123 |
"values": [
|
124 |
+
0.015321087837219239,
|
125 |
+
0.015568896293640137,
|
126 |
+
0.01562828826904297,
|
127 |
+
0.015215616226196289,
|
128 |
+
0.014110783576965331,
|
129 |
+
0.013884415626525879,
|
130 |
+
0.01420083236694336,
|
131 |
+
0.014844927787780762,
|
132 |
+
0.014370911598205567,
|
133 |
+
0.014386176109313965,
|
134 |
+
0.014109696388244629,
|
135 |
+
0.013880255699157715,
|
136 |
+
0.013993984222412109,
|
137 |
+
0.014191616058349609,
|
138 |
+
0.014065664291381836,
|
139 |
+
0.013813759803771973,
|
140 |
+
0.013923328399658203,
|
141 |
+
0.014142463684082032,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
142 |
0.014082048416137695,
|
143 |
+
0.013999103546142578,
|
144 |
+
0.014004223823547364,
|
145 |
+
0.01387724781036377,
|
146 |
+
0.013910016059875489,
|
147 |
+
0.01399500846862793,
|
148 |
+
0.013827072143554688,
|
149 |
+
0.013780991554260253,
|
150 |
+
0.014105600357055664,
|
151 |
+
0.014211071968078613,
|
152 |
+
0.014019583702087402,
|
153 |
+
0.013859840393066406,
|
154 |
+
0.013947903633117676,
|
155 |
+
0.013727744102478028,
|
156 |
+
0.013845503807067871,
|
157 |
+
0.013601792335510255,
|
158 |
+
0.013552639961242676,
|
159 |
+
0.01394380760192871,
|
160 |
+
0.014152704238891601,
|
161 |
+
0.014339072227478027,
|
162 |
0.014101504325866699,
|
163 |
+
0.013966336250305175,
|
164 |
+
0.014003199577331543,
|
165 |
+
0.013811712265014648,
|
166 |
+
0.013805567741394043,
|
167 |
+
0.013789183616638183,
|
168 |
+
0.01339907169342041,
|
169 |
+
0.01335807991027832,
|
170 |
+
0.013360128402709961,
|
171 |
+
0.013364224433898926,
|
172 |
+
0.013415424346923829,
|
173 |
+
0.013340671539306641,
|
174 |
+
0.01337548828125,
|
175 |
+
0.013418496131896973,
|
176 |
+
0.013376511573791505,
|
177 |
+
0.013401087760925292,
|
178 |
+
0.013342783927917481,
|
179 |
+
0.013384703636169434,
|
180 |
+
0.013378560066223144,
|
181 |
+
0.013367360115051269,
|
182 |
+
0.013371456146240234,
|
183 |
+
0.013373439788818359,
|
184 |
+
0.013353919982910157,
|
185 |
+
0.013350879669189453,
|
186 |
+
0.01338368034362793,
|
187 |
+
0.013339648246765137,
|
188 |
+
0.013351936340332032,
|
189 |
+
0.013336607933044434,
|
190 |
+
0.013355008125305176,
|
191 |
+
0.013327360153198242,
|
192 |
+
0.013293567657470704,
|
193 |
+
0.013340671539306641,
|
194 |
+
0.013365247726440429,
|
195 |
+
0.013389823913574218,
|
196 |
+
0.013319168090820312
|
197 |
]
|
198 |
},
|
199 |
"throughput": {
|
200 |
"unit": "samples/s",
|
201 |
+
"value": 72.31673760413685
|
202 |
},
|
203 |
"energy": {
|
204 |
"unit": "kWh",
|
205 |
+
"cpu": 1.600326630804274e-07,
|
206 |
+
"ram": 8.738030096339772e-08,
|
207 |
+
"gpu": 3.615151040266736e-07,
|
208 |
+
"total": 6.089280680704987e-07
|
209 |
},
|
210 |
"efficiency": {
|
211 |
"unit": "samples/kWh",
|
212 |
+
"value": 1642230.0965181077
|
213 |
}
|
214 |
}
|
215 |
}
|