IlyasMoutawwakil HF staff commited on
Commit
7c1fbcb
1 Parent(s): b878e1c

Upload cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_token-classification_microsoft/deberta-v3-base",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.4.0+rocm6.1",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "token-classification",
9
  "library": "transformers",
@@ -104,24 +104,24 @@
104
  "load": {
105
  "memory": {
106
  "unit": "MB",
107
- "max_ram": 1374.171136,
108
- "max_global_vram": 0.0,
109
- "max_process_vram": 0.0,
110
  "max_reserved": 773.849088,
111
  "max_allocated": 736.603648
112
  },
113
  "latency": {
114
  "unit": "s",
115
  "count": 1,
116
- "total": 7.8211630859375,
117
- "mean": 7.8211630859375,
118
  "stdev": 0.0,
119
- "p50": 7.8211630859375,
120
- "p90": 7.8211630859375,
121
- "p95": 7.8211630859375,
122
- "p99": 7.8211630859375,
123
  "values": [
124
- 7.8211630859375
125
  ]
126
  },
127
  "throughput": null,
@@ -131,98 +131,104 @@
131
  "forward": {
132
  "memory": {
133
  "unit": "MB",
134
- "max_ram": 1537.859584,
135
- "max_global_vram": 0.0,
136
- "max_process_vram": 0.0,
137
  "max_reserved": 773.849088,
138
  "max_allocated": 745.086976
139
  },
140
  "latency": {
141
  "unit": "s",
142
- "count": 70,
143
- "total": 1.0048157081604003,
144
- "mean": 0.014354510116577148,
145
- "stdev": 0.0006998602608446172,
146
- "p50": 0.014274752616882325,
147
- "p90": 0.015382380676269531,
148
- "p95": 0.015417909240722656,
149
- "p99": 0.016172993774414068,
150
  "values": [
151
- 0.015660429000854494,
152
- 0.014923311233520508,
153
- 0.014743631362915038,
154
- 0.014742991447448731,
155
- 0.01459083080291748,
156
- 0.014586671829223633,
157
- 0.014473072052001953,
158
- 0.014281392097473145,
159
- 0.014183793067932129,
160
- 0.014553711891174317,
161
- 0.014383312225341796,
162
- 0.014314191818237305,
163
- 0.014232592582702637,
164
- 0.01731386375427246,
165
- 0.014493391990661622,
166
- 0.013963633537292481,
167
- 0.013950353622436523,
168
- 0.014131954193115234,
169
- 0.014268113136291503,
170
- 0.014454031944274902,
171
- 0.013852913856506348,
172
- 0.014144753456115723,
173
- 0.013671955108642579,
174
- 0.013778034210205079,
175
- 0.013447154998779296,
176
- 0.013763154983520508,
177
- 0.014186033248901367,
178
- 0.014009714126586914,
179
- 0.014020913124084473,
180
- 0.01437707233428955,
181
- 0.013947473526000977,
182
- 0.013909073829650878,
183
- 0.01348731517791748,
184
- 0.013493715286254883,
185
- 0.013534355163574218,
186
- 0.014307473182678223,
187
- 0.013509394645690918,
188
- 0.013873554229736329,
189
- 0.01377147388458252,
190
- 0.013480114936828613,
191
- 0.013523795127868653,
192
- 0.013473715782165527,
193
- 0.013797233581542969,
194
- 0.013511795043945313,
195
- 0.013507155418395997,
196
- 0.013532915115356446,
197
- 0.013525554656982422,
198
- 0.014065552711486816,
199
- 0.014231152534484863,
200
- 0.014463791847229004,
201
- 0.014546192169189454,
202
- 0.014261391639709473,
203
- 0.014397872924804687,
204
- 0.013843793869018555,
205
- 0.01484986972808838,
206
- 0.014529071807861327,
207
- 0.014553071975708008,
208
- 0.014667470932006836,
209
- 0.014703150749206542,
210
- 0.01469227123260498,
211
- 0.014854670524597168,
212
- 0.015328108787536621,
213
- 0.015322189331054687,
214
- 0.015443469047546387,
215
- 0.015391308784484864,
216
- 0.015381388664245606,
217
- 0.015379469871520996,
218
- 0.015392108917236328,
219
- 0.015420429229736328,
220
- 0.015414829254150391
 
 
 
 
 
 
221
  ]
222
  },
223
  "throughput": {
224
  "unit": "samples/s",
225
- "value": 69.66451602170393
226
  },
227
  "energy": null,
228
  "efficiency": null
 
3
  "name": "cuda_inference_transformers_token-classification_microsoft/deberta-v3-base",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.2.2+rocm5.7",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "token-classification",
9
  "library": "transformers",
 
104
  "load": {
105
  "memory": {
106
  "unit": "MB",
107
+ "max_ram": 893.31712,
108
+ "max_global_vram": 1059.885056,
109
+ "max_process_vram": 45893.763072,
110
  "max_reserved": 773.849088,
111
  "max_allocated": 736.603648
112
  },
113
  "latency": {
114
  "unit": "s",
115
  "count": 1,
116
+ "total": 7.09972509765625,
117
+ "mean": 7.09972509765625,
118
  "stdev": 0.0,
119
+ "p50": 7.09972509765625,
120
+ "p90": 7.09972509765625,
121
+ "p95": 7.09972509765625,
122
+ "p99": 7.09972509765625,
123
  "values": [
124
+ 7.09972509765625
125
  ]
126
  },
127
  "throughput": null,
 
131
  "forward": {
132
  "memory": {
133
  "unit": "MB",
134
+ "max_ram": 1038.180352,
135
+ "max_global_vram": 1122.811904,
136
+ "max_process_vram": 292149.198848,
137
  "max_reserved": 773.849088,
138
  "max_allocated": 745.086976
139
  },
140
  "latency": {
141
  "unit": "s",
142
+ "count": 76,
143
+ "total": 1.0010555467605589,
144
+ "mean": 0.013171783510007357,
145
+ "stdev": 0.0005228528429792025,
146
+ "p50": 0.012988920688629151,
147
+ "p90": 0.013722038269042967,
148
+ "p95": 0.013880318880081177,
149
+ "p99": 0.01504787540435791,
150
  "values": [
151
+ 0.012939001083374023,
152
+ 0.013142040252685547,
153
+ 0.012951001167297364,
154
+ 0.013017400741577149,
155
+ 0.012907960891723632,
156
+ 0.012662841796875,
157
+ 0.012517722129821777,
158
+ 0.012533562660217285,
159
+ 0.012498682022094726,
160
+ 0.012547162055969239,
161
+ 0.012569561958312988,
162
+ 0.0125994815826416,
163
+ 0.0125567626953125,
164
+ 0.015943312644958497,
165
+ 0.013363479614257812,
166
+ 0.012921401023864746,
167
+ 0.01278828239440918,
168
+ 0.013097399711608887,
169
+ 0.012783162117004395,
170
+ 0.013227160453796388,
171
+ 0.013100920677185058,
172
+ 0.013227319717407227,
173
+ 0.013104280471801759,
174
+ 0.013266519546508789,
175
+ 0.013140280723571778,
176
+ 0.01324507999420166,
177
+ 0.013232760429382324,
178
+ 0.013538839340209961,
179
+ 0.013443639755249023,
180
+ 0.013765077590942382,
181
+ 0.013094361305236817,
182
+ 0.012924120903015137,
183
+ 0.013038360595703125,
184
+ 0.012930200576782227,
185
+ 0.012898361206054688,
186
+ 0.013159481048583985,
187
+ 0.01294636058807373,
188
+ 0.012942360877990724,
189
+ 0.012850521087646484,
190
+ 0.013222200393676758,
191
+ 0.012908921241760254,
192
+ 0.012913081169128418,
193
+ 0.01291020107269287,
194
+ 0.013130041122436523,
195
+ 0.01291420078277588,
196
+ 0.012888920783996583,
197
+ 0.012922361373901367,
198
+ 0.013024121284484864,
199
+ 0.012953241348266602,
200
+ 0.013331480026245117,
201
+ 0.012916601181030274,
202
+ 0.01292988109588623,
203
+ 0.012914840698242187,
204
+ 0.012917560577392578,
205
+ 0.012870201110839844,
206
+ 0.014749396324157714,
207
+ 0.014464435577392578,
208
+ 0.013452919960021972,
209
+ 0.013532439231872558,
210
+ 0.01391739845275879,
211
+ 0.0138546781539917,
212
+ 0.013867959022521973,
213
+ 0.013477239608764648,
214
+ 0.013828437805175781,
215
+ 0.013610038757324219,
216
+ 0.013660438537597656,
217
+ 0.012942041397094727,
218
+ 0.013398839950561523,
219
+ 0.013678998947143554,
220
+ 0.013543319702148438,
221
+ 0.012952120780944824,
222
+ 0.012960440635681152,
223
+ 0.012936281204223634,
224
+ 0.01295820140838623,
225
+ 0.013284760475158692,
226
+ 0.012901080131530763
227
  ]
228
  },
229
  "throughput": {
230
  "unit": "samples/s",
231
+ "value": 75.91986303451183
232
  },
233
  "energy": null,
234
  "efficiency": null