IlyasMoutawwakil HF staff commited on
Commit
af7fdea
·
verified ·
1 Parent(s): ce04625

Upload cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.2.2+rocm5.7",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
9
  "library": "transformers",
@@ -11,7 +11,7 @@
11
  "model": "google-bert/bert-base-uncased",
12
  "processor": "google-bert/bert-base-uncased",
13
  "device": "cuda",
14
- "device_ids": "6",
15
  "seed": 42,
16
  "inter_op_num_threads": null,
17
  "intra_op_num_threads": null,
@@ -97,7 +97,7 @@
97
  "transformers_commit": null,
98
  "accelerate_version": "0.33.0",
99
  "accelerate_commit": null,
100
- "diffusers_version": "0.30.1",
101
  "diffusers_commit": null,
102
  "optimum_version": null,
103
  "optimum_commit": null,
@@ -111,24 +111,24 @@
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
- "max_ram": 904.454144,
115
- "max_global_vram": 875.425792,
116
- "max_process_vram": 48006.356992,
117
  "max_reserved": 589.299712,
118
  "max_allocated": 533.571072
119
  },
120
  "latency": {
121
  "unit": "s",
122
  "count": 1,
123
- "total": 8.6468671875,
124
- "mean": 8.6468671875,
125
  "stdev": 0.0,
126
- "p50": 8.6468671875,
127
- "p90": 8.6468671875,
128
- "p95": 8.6468671875,
129
- "p99": 8.6468671875,
130
  "values": [
131
- 8.6468671875
132
  ]
133
  },
134
  "throughput": null,
@@ -138,153 +138,177 @@
138
  "forward": {
139
  "memory": {
140
  "unit": "MB",
141
- "max_ram": 1023.791104,
142
- "max_global_vram": 939.872256,
143
- "max_process_vram": 240932.306944,
144
  "max_reserved": 589.299712,
145
  "max_allocated": 439.456768
146
  },
147
  "latency": {
148
  "unit": "s",
149
- "count": 125,
150
- "total": 1.0017708358764645,
151
- "mean": 0.008014166687011718,
152
- "stdev": 0.0007465907271951295,
153
- "p50": 0.007964495182037354,
154
- "p90": 0.008243439483642577,
155
- "p95": 0.008332015800476074,
156
- "p99": 0.00884146957397461,
157
  "values": [
158
- 0.007716013908386231,
159
- 0.007700973987579345,
160
- 0.007814895153045655,
161
- 0.00794785499572754,
162
- 0.00785521411895752,
163
- 0.007820495128631592,
164
- 0.007783054828643799,
165
- 0.0077804942131042484,
166
- 0.007849454879760742,
167
- 0.007880495071411133,
168
- 0.00891569709777832,
169
- 0.008011215209960938,
170
- 0.00803569507598877,
171
- 0.00799649477005005,
172
- 0.007924174785614014,
173
- 0.00786705493927002,
174
- 0.0078100948333740235,
175
- 0.007739374160766602,
176
- 0.0077337751388549805,
177
- 0.007784173965454101,
178
- 0.007783695220947266,
179
- 0.007960494995117187,
180
- 0.007923695087432861,
181
- 0.007801134109497071,
182
- 0.007723054885864258,
183
- 0.0076532940864562984,
184
- 0.007686735153198242,
185
- 0.0076459341049194334,
186
- 0.007636013984680176,
187
- 0.007864494800567627,
188
- 0.007649134159088135,
189
- 0.00760353422164917,
190
- 0.007625773906707763,
191
- 0.007637615203857422,
192
- 0.007648173809051514,
193
- 0.0076409740447998045,
194
- 0.007554574012756348,
195
- 0.007689455032348633,
196
- 0.007588334083557129,
197
- 0.0075956940650939945,
198
- 0.007597455024719238,
199
- 0.007678733825683594,
200
- 0.007626093864440918,
201
- 0.007636175155639648,
202
- 0.007637613773345947,
203
- 0.007556494235992432,
204
- 0.0076553750038146975,
205
- 0.0076030540466308595,
206
- 0.00758609390258789,
207
- 0.015914429664611818,
208
- 0.008352175712585449,
209
- 0.008254255294799805,
210
- 0.00803665542602539,
211
- 0.008123375892639161,
212
- 0.008118254661560059,
213
- 0.0080820951461792,
214
- 0.008066254615783692,
215
- 0.007988494873046875,
216
- 0.008142575263977051,
217
- 0.008245615005493164,
218
- 0.008339056015014648,
219
- 0.008295695304870605,
220
- 0.008203535079956055,
221
- 0.00827361488342285,
222
- 0.008212815284729004,
223
- 0.008240176200866698,
224
- 0.008163535118103027,
225
- 0.008411855697631835,
226
- 0.008543215751647949,
227
- 0.008606415748596191,
228
- 0.00826801586151123,
229
- 0.00818033504486084,
230
- 0.008144015312194825,
231
- 0.008093134880065918,
232
- 0.008043054580688477,
233
- 0.0080820951461792,
234
- 0.007995534896850585,
235
- 0.007928494930267333,
236
- 0.007967053890228271,
237
- 0.007926255226135255,
238
- 0.008009775161743164,
239
- 0.008067376136779785,
240
- 0.007951854228973388,
241
- 0.008009296417236328,
242
- 0.00803073501586914,
243
- 0.008039534568786621,
244
- 0.00805809497833252,
245
- 0.008303854942321777,
246
- 0.008195535659790039,
247
- 0.008096175193786621,
248
- 0.008050894737243653,
249
- 0.008007214546203614,
250
- 0.008077614784240723,
251
- 0.008115376472473144,
252
- 0.00813553524017334,
253
- 0.008003535270690918,
254
- 0.008082096099853515,
255
- 0.008153935432434083,
256
- 0.00814321517944336,
257
- 0.008051694869995117,
258
- 0.008064974784851074,
259
- 0.00799969482421875,
260
- 0.008000494956970215,
261
- 0.007990254878997802,
262
- 0.007962894916534425,
263
- 0.008016494750976563,
264
- 0.00799041509628296,
265
- 0.008016494750976563,
266
- 0.007925294876098632,
267
- 0.007879855155944825,
268
- 0.007880815029144287,
269
- 0.007964495182037354,
270
- 0.007877614974975586,
271
- 0.007858573913574219,
272
- 0.007896494865417481,
273
- 0.007846415042877197,
274
- 0.007865135192871093,
275
- 0.007880333900451661,
276
- 0.007892815113067627,
277
- 0.007893294811248779,
278
- 0.007920014858245849,
279
- 0.00799041509628296,
280
- 0.007978254795074463,
281
- 0.007911214828491211,
282
- 0.007987053871154785
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
283
  ]
284
  },
285
  "throughput": {
286
  "unit": "samples/s",
287
- "value": 124.779036804995
288
  },
289
  "energy": null,
290
  "efficiency": null
 
3
  "name": "cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.4.0+rocm6.1",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
9
  "library": "transformers",
 
11
  "model": "google-bert/bert-base-uncased",
12
  "processor": "google-bert/bert-base-uncased",
13
  "device": "cuda",
14
+ "device_ids": "0",
15
  "seed": 42,
16
  "inter_op_num_threads": null,
17
  "intra_op_num_threads": null,
 
97
  "transformers_commit": null,
98
  "accelerate_version": "0.33.0",
99
  "accelerate_commit": null,
100
+ "diffusers_version": "0.30.2",
101
  "diffusers_commit": null,
102
  "optimum_version": null,
103
  "optimum_commit": null,
 
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
+ "max_ram": 1386.295296,
115
+ "max_global_vram": 11.075584,
116
+ "max_process_vram": 0.0,
117
  "max_reserved": 589.299712,
118
  "max_allocated": 533.571072
119
  },
120
  "latency": {
121
  "unit": "s",
122
  "count": 1,
123
+ "total": 7.72250390625,
124
+ "mean": 7.72250390625,
125
  "stdev": 0.0,
126
+ "p50": 7.72250390625,
127
+ "p90": 7.72250390625,
128
+ "p95": 7.72250390625,
129
+ "p99": 7.72250390625,
130
  "values": [
131
+ 7.72250390625
132
  ]
133
  },
134
  "throughput": null,
 
138
  "forward": {
139
  "memory": {
140
  "unit": "MB",
141
+ "max_ram": 1485.74208,
142
+ "max_global_vram": 11.14112,
143
+ "max_process_vram": 0.0,
144
  "max_reserved": 589.299712,
145
  "max_allocated": 439.456768
146
  },
147
  "latency": {
148
  "unit": "s",
149
+ "count": 149,
150
+ "total": 1.0048536100387568,
151
+ "mean": 0.006743983959991659,
152
+ "stdev": 0.00046732393598980873,
153
+ "p50": 0.006559018135070801,
154
+ "p90": 0.007237319850921631,
155
+ "p95": 0.007262568092346191,
156
+ "p99": 0.00734000078201294,
157
  "values": [
158
+ 0.006815018177032471,
159
+ 0.006684617042541504,
160
+ 0.006481898784637451,
161
+ 0.006476297855377198,
162
+ 0.006494218826293945,
163
+ 0.006513738155364991,
164
+ 0.006502217769622802,
165
+ 0.006486538887023926,
166
+ 0.006455177783966064,
167
+ 0.006764936923980713,
168
+ 0.006500938892364502,
169
+ 0.006439017772674561,
170
+ 0.006785098075866699,
171
+ 0.006964776039123535,
172
+ 0.006782698154449463,
173
+ 0.0066874980926513675,
174
+ 0.006369099140167236,
175
+ 0.006727017879486084,
176
+ 0.007182536125183106,
177
+ 0.007213575839996338,
178
+ 0.007224616050720215,
179
+ 0.007224616050720215,
180
+ 0.007223176002502441,
181
+ 0.007225574970245362,
182
+ 0.007190375804901123,
183
+ 0.007231336116790771,
184
+ 0.007218056201934815,
185
+ 0.007196616172790527,
186
+ 0.0071759757995605465,
187
+ 0.007201255798339844,
188
+ 0.007254216194152832,
189
+ 0.00724781608581543,
190
+ 0.007245096206665039,
191
+ 0.00728349494934082,
192
+ 0.007304615974426269,
193
+ 0.00725213623046875,
194
+ 0.007277575016021729,
195
+ 0.007268136024475098,
196
+ 0.0072420558929443355,
197
+ 0.007286375999450684,
198
+ 0.007242856025695801,
199
+ 0.007219655990600586,
200
+ 0.007212775230407715,
201
+ 0.00719325590133667,
202
+ 0.007175016880035401,
203
+ 0.00717901611328125,
204
+ 0.007186056137084961,
205
+ 0.007171815872192383,
206
+ 0.007171815872192383,
207
+ 0.007130856037139892,
208
+ 0.007136295795440674,
209
+ 0.010255166053771973,
210
+ 0.007326855182647705,
211
+ 0.007132297039031983,
212
+ 0.007190375804901123,
213
+ 0.007184135913848877,
214
+ 0.007171976089477539,
215
+ 0.007209735870361328,
216
+ 0.007218376159667968,
217
+ 0.007237256050109863,
218
+ 0.0072058959007263185,
219
+ 0.007213575839996338,
220
+ 0.0067975778579711914,
221
+ 0.006574217796325684,
222
+ 0.006611978054046631,
223
+ 0.006545097827911377,
224
+ 0.00678205680847168,
225
+ 0.006546538829803467,
226
+ 0.006555498123168945,
227
+ 0.006612937927246093,
228
+ 0.006559018135070801,
229
+ 0.006589417934417725,
230
+ 0.006545417785644532,
231
+ 0.006769736766815185,
232
+ 0.006558857917785645,
233
+ 0.006496778964996338,
234
+ 0.006350698947906494,
235
+ 0.006378858089447021,
236
+ 0.006356459140777588,
237
+ 0.006362858772277832,
238
+ 0.006408298969268799,
239
+ 0.006370058059692383,
240
+ 0.006371979236602783,
241
+ 0.006386538982391357,
242
+ 0.006367178916931152,
243
+ 0.0063726181983947755,
244
+ 0.0063545389175415036,
245
+ 0.006347818851470947,
246
+ 0.0063430190086364744,
247
+ 0.006319979190826416,
248
+ 0.0065198178291320805,
249
+ 0.0064250979423522946,
250
+ 0.006331018924713135,
251
+ 0.00634653902053833,
252
+ 0.0063716588020324705,
253
+ 0.006351978778839112,
254
+ 0.006345738887786865,
255
+ 0.0063511781692504885,
256
+ 0.006352458953857422,
257
+ 0.006482698917388916,
258
+ 0.0067703771591186524,
259
+ 0.006564457893371582,
260
+ 0.006645738124847412,
261
+ 0.006682218074798584,
262
+ 0.0065151782035827634,
263
+ 0.006714858055114746,
264
+ 0.006686057090759277,
265
+ 0.006536777973175049,
266
+ 0.006493739128112793,
267
+ 0.006346219062805176,
268
+ 0.006347817897796631,
269
+ 0.006329739093780517,
270
+ 0.0063466992378234865,
271
+ 0.006368779182434082,
272
+ 0.006377738952636719,
273
+ 0.0063521389961242675,
274
+ 0.006351978778839112,
275
+ 0.0063614192008972165,
276
+ 0.0063594989776611325,
277
+ 0.0063783788681030275,
278
+ 0.006355177879333496,
279
+ 0.006360619068145752,
280
+ 0.006335178852081299,
281
+ 0.006329898834228516,
282
+ 0.006363338947296142,
283
+ 0.006362538814544677,
284
+ 0.006342058181762695,
285
+ 0.006369739055633545,
286
+ 0.0063447790145874025,
287
+ 0.006381578922271728,
288
+ 0.006357577800750733,
289
+ 0.006341578960418701,
290
+ 0.006348299026489258,
291
+ 0.006372138977050781,
292
+ 0.006352778911590576,
293
+ 0.006365738868713379,
294
+ 0.006667977809906006,
295
+ 0.0066521382331848145,
296
+ 0.006762857913970947,
297
+ 0.00667229700088501,
298
+ 0.006439339160919189,
299
+ 0.006351179122924805,
300
+ 0.006547018051147461,
301
+ 0.006341578960418701,
302
+ 0.007352135181427002,
303
+ 0.007237575054168701,
304
+ 0.007236135959625244,
305
+ 0.0071850957870483395,
306
+ 0.007216135978698731
307
  ]
308
  },
309
  "throughput": {
310
  "unit": "samples/s",
311
+ "value": 148.2803052220244
312
  },
313
  "energy": null,
314
  "efficiency": null