IlyasMoutawwakil HF staff commited on
Commit
ad829dc
·
verified ·
1 Parent(s): dd6e14a

Upload cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.3.1+rocm5.7",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
9
  "library": "transformers",
@@ -11,7 +11,7 @@
11
  "model": "google-bert/bert-base-uncased",
12
  "processor": "google-bert/bert-base-uncased",
13
  "device": "cuda",
14
- "device_ids": "5",
15
  "seed": 42,
16
  "inter_op_num_threads": null,
17
  "intra_op_num_threads": null,
@@ -111,24 +111,24 @@
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
- "max_ram": 1034.399744,
115
  "max_global_vram": 68702.69952,
116
- "max_process_vram": 44947.423232,
117
  "max_reserved": 589.299712,
118
  "max_allocated": 533.571072
119
  },
120
  "latency": {
121
  "unit": "s",
122
  "count": 1,
123
- "total": 8.1447783203125,
124
- "mean": 8.1447783203125,
125
  "stdev": 0.0,
126
- "p50": 8.1447783203125,
127
- "p90": 8.1447783203125,
128
- "p95": 8.1447783203125,
129
- "p99": 8.1447783203125,
130
  "values": [
131
- 8.1447783203125
132
  ]
133
  },
134
  "throughput": null,
@@ -138,161 +138,168 @@
138
  "forward": {
139
  "memory": {
140
  "unit": "MB",
141
- "max_ram": 1164.197888,
142
  "max_global_vram": 68702.69952,
143
- "max_process_vram": 225674.031104,
144
  "max_reserved": 589.299712,
145
  "max_allocated": 439.456768
146
  },
147
  "latency": {
148
  "unit": "s",
149
- "count": 133,
150
- "total": 1.0039466438293458,
151
- "mean": 0.007548471006235681,
152
- "stdev": 0.0015299985621818849,
153
- "p50": 0.0075959739685058595,
154
- "p90": 0.007701221942901611,
155
- "p95": 0.007799845504760742,
156
- "p99": 0.014211061820983901,
157
  "values": [
158
- 0.007720452785491943,
159
- 0.007671974182128906,
160
- 0.007859333038330078,
161
- 0.007679014205932617,
162
- 0.007774373054504394,
163
- 0.007655014038085937,
164
- 0.007681253910064697,
165
- 0.007644773960113526,
166
- 0.007890052795410156,
167
- 0.007679654121398926,
168
- 0.007702534198760986,
169
- 0.007703972816467285,
170
- 0.007672294139862061,
171
- 0.0076724538803100584,
172
- 0.007660614013671875,
173
- 0.007695972919464111,
174
- 0.007727014064788818,
175
- 0.007329734802246093,
176
- 0.007442214012145996,
177
- 0.007626853942871094,
178
- 0.007566693782806396,
179
- 0.00751005506515503,
180
- 0.007329574108123779,
181
- 0.007838054180145264,
182
- 0.007270215034484863,
183
- 0.007506534099578858,
184
- 0.0076724538803100584,
185
- 0.007668774127960205,
186
- 0.007631653785705566,
187
- 0.0076154937744140625,
188
- 0.00762973403930664,
189
- 0.0076154937744140625,
190
- 0.007652453899383545,
191
- 0.007616773128509522,
192
- 0.0076572542190551755,
193
- 0.0076460537910461425,
194
- 0.007645733833312988,
195
- 0.014887309074401856,
196
- 0.007743173122406006,
197
- 0.007675494194030762,
198
- 0.007645413875579834,
199
- 0.0076455740928649905,
200
- 0.007639974117279053,
201
- 0.0075959739685058595,
202
- 0.007656453132629394,
203
- 0.007647493839263916,
204
- 0.007611653804779053,
205
- 0.007625574111938476,
206
- 0.00762317419052124,
207
- 0.007669573783874512,
208
- 0.007641893863677978,
209
- 0.0076433329582214354,
210
- 0.007661573886871338,
211
- 0.007643814086914062,
212
- 0.0076081337928771975,
213
- 0.007668454170227051,
214
- 0.007643972873687744,
215
- 0.007654533863067627,
216
- 0.007647654056549072,
217
- 0.0076284542083740236,
218
- 0.007663814067840577,
219
- 0.007670053005218506,
220
- 0.007670534133911133,
221
- 0.00767997407913208,
222
- 0.012774036407470704,
223
- 0.0076154937744140625,
224
- 0.0076388540267944335,
225
- 0.007662213802337646,
226
- 0.007636134147644043,
227
- 0.007420293807983399,
228
- 0.007280294895172119,
229
- 0.007115015983581543,
230
- 0.007062215805053711,
231
- 0.006970536231994629,
232
- 0.0070658960342407226,
233
- 0.007170215129852295,
234
- 0.007293894767761231,
235
- 0.007220294952392578,
236
- 0.007059336185455322,
237
- 0.007167816162109375,
238
- 0.007138694763183594,
239
- 0.007119815826416015,
240
- 0.006989416122436524,
241
- 0.007128614902496338,
242
- 0.007118216037750244,
243
- 0.007131976127624512,
244
- 0.007019336223602295,
245
- 0.007042215824127197,
246
- 0.007117414951324463,
247
- 0.00675741720199585,
248
- 0.0067727770805358885,
249
- 0.006748136997222901,
250
- 0.006700296878814697,
251
- 0.011966038703918457,
252
- 0.0066943769454956056,
253
- 0.0067110171318054196,
254
- 0.006693737030029297,
255
- 0.006752457141876221,
256
- 0.00672957706451416,
257
- 0.006779496192932129,
258
- 0.006751817226409912,
259
- 0.006746216773986817,
260
- 0.006955656051635742,
261
- 0.006740616798400879,
262
- 0.006810216903686523,
263
- 0.006742856979370117,
264
- 0.00679069709777832,
265
- 0.006767816066741943,
266
- 0.006730377197265625,
267
- 0.006733256816864014,
268
- 0.006739017009735108,
269
- 0.006819177150726319,
270
- 0.0069974160194396975,
271
- 0.00680877685546875,
272
- 0.0068414158821105955,
273
- 0.006712296962738037,
274
- 0.006774217128753662,
275
- 0.006783017158508301,
276
- 0.00676013708114624,
277
- 0.0067679758071899415,
278
- 0.006782697200775146,
279
- 0.00673181676864624,
280
- 0.006772936820983887,
281
- 0.006762217044830323,
282
- 0.006781257152557373,
283
- 0.00677069616317749,
284
- 0.007255015850067138,
285
- 0.0077034921646118165,
286
- 0.007631814002990723,
287
- 0.0076810941696166994,
288
- 0.007327014923095703,
289
- 0.00725949478149414,
290
- 0.021094968795776367
 
 
 
 
 
 
 
291
  ]
292
  },
293
  "throughput": {
294
  "unit": "samples/s",
295
- "value": 132.4771598346095
296
  },
297
  "energy": null,
298
  "efficiency": null
 
3
  "name": "cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.2.0.dev20231010+rocm5.7",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
9
  "library": "transformers",
 
11
  "model": "google-bert/bert-base-uncased",
12
  "processor": "google-bert/bert-base-uncased",
13
  "device": "cuda",
14
+ "device_ids": "6",
15
  "seed": 42,
16
  "inter_op_num_threads": null,
17
  "intra_op_num_threads": null,
 
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
+ "max_ram": 946.982912,
115
  "max_global_vram": 68702.69952,
116
+ "max_process_vram": 47096.79104,
117
  "max_reserved": 589.299712,
118
  "max_allocated": 533.571072
119
  },
120
  "latency": {
121
  "unit": "s",
122
  "count": 1,
123
+ "total": 8.0222158203125,
124
+ "mean": 8.0222158203125,
125
  "stdev": 0.0,
126
+ "p50": 8.0222158203125,
127
+ "p90": 8.0222158203125,
128
+ "p95": 8.0222158203125,
129
+ "p99": 8.0222158203125,
130
  "values": [
131
+ 8.0222158203125
132
  ]
133
  },
134
  "throughput": null,
 
138
  "forward": {
139
  "memory": {
140
  "unit": "MB",
141
+ "max_ram": 1084.862464,
142
  "max_global_vram": 68702.69952,
143
+ "max_process_vram": 237212.884992,
144
  "max_reserved": 589.299712,
145
  "max_allocated": 439.456768
146
  },
147
  "latency": {
148
  "unit": "s",
149
+ "count": 140,
150
+ "total": 1.0030574531555176,
151
+ "mean": 0.0071646960939679835,
152
+ "stdev": 0.00037430606419756976,
153
+ "p50": 0.007102169990539551,
154
+ "p90": 0.007583194828033448,
155
+ "p95": 0.007632730889320373,
156
+ "p99": 0.008038914513587948,
157
  "values": [
158
+ 0.007846251964569092,
159
+ 0.007688490867614746,
160
+ 0.007654891014099121,
161
+ 0.007640330791473388,
162
+ 0.007684330940246582,
163
+ 0.007304969787597656,
164
+ 0.007356811046600342,
165
+ 0.007289610862731934,
166
+ 0.007249929904937744,
167
+ 0.007470890998840332,
168
+ 0.008162092208862305,
169
+ 0.006934410095214844,
170
+ 0.006910409927368164,
171
+ 0.006923369884490967,
172
+ 0.006878729820251465,
173
+ 0.006957930088043213,
174
+ 0.006864009857177734,
175
+ 0.006900650024414062,
176
+ 0.007057291030883789,
177
+ 0.007127049922943115,
178
+ 0.007180970191955566,
179
+ 0.006889450073242188,
180
+ 0.006898890018463135,
181
+ 0.0068900899887084964,
182
+ 0.0068545699119567875,
183
+ 0.006871530055999756,
184
+ 0.006911690235137939,
185
+ 0.006983689785003662,
186
+ 0.007002729892730713,
187
+ 0.006895209789276123,
188
+ 0.006999050140380859,
189
+ 0.006968810081481934,
190
+ 0.008366251945495605,
191
+ 0.007415209770202637,
192
+ 0.007457291126251221,
193
+ 0.007448970794677734,
194
+ 0.007292651176452637,
195
+ 0.007418409824371338,
196
+ 0.007416650772094726,
197
+ 0.007372011184692382,
198
+ 0.007317131042480469,
199
+ 0.007518089771270752,
200
+ 0.007526731014251709,
201
+ 0.00756273078918457,
202
+ 0.0075163311958312985,
203
+ 0.007611371040344238,
204
+ 0.007517611026763916,
205
+ 0.0076011309623718265,
206
+ 0.007590250968933105,
207
+ 0.007579211235046387,
208
+ 0.007569611072540283,
209
+ 0.007548810958862305,
210
+ 0.007562891006469726,
211
+ 0.007561611175537109,
212
+ 0.007538889884948731,
213
+ 0.0075764918327331544,
214
+ 0.007570250988006592,
215
+ 0.007591210842132569,
216
+ 0.00756865119934082,
217
+ 0.0075588908195495606,
218
+ 0.0075779309272766115,
219
+ 0.007391049861907959,
220
+ 0.007177771091461182,
221
+ 0.007008490085601807,
222
+ 0.006701930046081543,
223
+ 0.006764490127563477,
224
+ 0.006726569175720215,
225
+ 0.006678090095520019,
226
+ 0.006694570064544678,
227
+ 0.0067024102210998535,
228
+ 0.006730888843536377,
229
+ 0.006672490119934082,
230
+ 0.006723209857940674,
231
+ 0.006744329929351807,
232
+ 0.006723689079284668,
233
+ 0.006959691047668457,
234
+ 0.007156970024108887,
235
+ 0.006916170120239258,
236
+ 0.006933290004730225,
237
+ 0.007370730876922607,
238
+ 0.007219850063323974,
239
+ 0.006695209980010986,
240
+ 0.0069420900344848635,
241
+ 0.00688001012802124,
242
+ 0.0069089698791503906,
243
+ 0.006768330097198487,
244
+ 0.006665929794311524,
245
+ 0.0069635300636291505,
246
+ 0.0067622499465942384,
247
+ 0.006724009990692139,
248
+ 0.006731529235839844,
249
+ 0.006728970050811768,
250
+ 0.006694570064544678,
251
+ 0.006697450160980225,
252
+ 0.006749448776245118,
253
+ 0.0067428889274597165,
254
+ 0.006732009887695313,
255
+ 0.00672992992401123,
256
+ 0.006764490127563477,
257
+ 0.0067267298698425295,
258
+ 0.00670288896560669,
259
+ 0.00701521110534668,
260
+ 0.006711369037628174,
261
+ 0.0069139299392700195,
262
+ 0.0067144098281860356,
263
+ 0.006903210163116455,
264
+ 0.006924489974975586,
265
+ 0.006775050163269043,
266
+ 0.006709449768066407,
267
+ 0.006687530040740967,
268
+ 0.006681768894195556,
269
+ 0.006758729934692383,
270
+ 0.006681290149688721,
271
+ 0.006770890235900879,
272
+ 0.0075540909767150876,
273
+ 0.007632330894470215,
274
+ 0.007596170902252198,
275
+ 0.007408650875091552,
276
+ 0.0070913701057434084,
277
+ 0.0071011300086975095,
278
+ 0.007200331211090088,
279
+ 0.0070889701843261715,
280
+ 0.0075294508934021,
281
+ 0.007523050785064697,
282
+ 0.007602571010589599,
283
+ 0.00758241081237793,
284
+ 0.007561611175537109,
285
+ 0.00756705093383789,
286
+ 0.007548810958862305,
287
+ 0.0075712108612060545,
288
+ 0.00755233097076416,
289
+ 0.0075609698295593265,
290
+ 0.007525771141052246,
291
+ 0.007570411205291748,
292
+ 0.007545451164245606,
293
+ 0.00754033088684082,
294
+ 0.007550411224365235,
295
+ 0.0072220897674560545,
296
+ 0.007169610977172851,
297
+ 0.007103209972381592
298
  ]
299
  },
300
  "throughput": {
301
  "unit": "samples/s",
302
+ "value": 139.57326129183738
303
  },
304
  "energy": null,
305
  "efficiency": null