IlyasMoutawwakil HF staff commited on
Commit
ea17b28
·
verified ·
1 Parent(s): afdd2ab

Upload cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json CHANGED
@@ -85,7 +85,7 @@
85
  "gpu_count": 1,
86
  "gpu_vram_mb": 68702699520,
87
  "optimum_benchmark_version": "0.2.1",
88
- "optimum_benchmark_commit": "43734676b09236e2ae4d7c19f90e8e23d1f2f201",
89
  "transformers_version": "4.42.3",
90
  "transformers_commit": null,
91
  "accelerate_version": "0.31.0",
@@ -104,91 +104,95 @@
104
  "forward": {
105
  "memory": {
106
  "unit": "MB",
107
- "max_ram": 1029.459968,
108
- "max_global_vram": 1122.885632,
109
- "max_process_vram": 251069.059072,
110
  "max_reserved": 773.849088,
111
  "max_allocated": 745.087488
112
  },
113
  "latency": {
114
  "unit": "s",
115
- "count": 63,
116
- "total": 1.0025160417556764,
117
- "mean": 0.015912953043740897,
118
- "stdev": 0.0010932306697772273,
119
- "p50": 0.015534128189086914,
120
- "p90": 0.017585762786865234,
121
- "p95": 0.017694481468200685,
122
- "p99": 0.0180983508682251,
123
  "values": [
124
- 0.014831733703613282,
125
- 0.015563729286193848,
126
- 0.01578708839416504,
127
- 0.015681167602539063,
128
- 0.015457009315490723,
129
- 0.015333970069885254,
130
- 0.015452849388122558,
131
- 0.015350609779357911,
132
- 0.016276363372802735,
133
- 0.016875240325927736,
134
- 0.017390275955200197,
135
- 0.01722947692871094,
136
- 0.017317796707153322,
137
- 0.01686164093017578,
138
- 0.017059078216552736,
139
- 0.017020519256591798,
140
- 0.017574756622314454,
141
- 0.017700674057006836,
142
- 0.01758851432800293,
143
- 0.01743107604980469,
144
- 0.017507875442504883,
145
- 0.01759155464172363,
146
- 0.01765603446960449,
147
- 0.017429155349731447,
148
- 0.01783843231201172,
149
- 0.017698753356933595,
150
- 0.016020206451416014,
151
- 0.01852242851257324,
152
- 0.015258770942687988,
153
- 0.015051093101501465,
154
- 0.01456341552734375,
155
- 0.014558135986328125,
156
- 0.014932852745056152,
157
- 0.014809814453125,
158
- 0.014706134796142578,
159
- 0.015048373222351074,
160
- 0.014690935134887695,
161
- 0.014668535232543946,
162
- 0.014894933700561523,
163
- 0.014695574760437011,
164
- 0.01465637493133545,
165
- 0.01497557258605957,
166
- 0.014666134834289551,
167
- 0.014717655181884766,
168
- 0.015068853378295898,
169
- 0.014717974662780763,
170
- 0.016144363403320313,
171
- 0.01673748016357422,
172
- 0.016272363662719728,
173
- 0.01557972812652588,
174
- 0.015890767097473144,
175
- 0.015024212837219239,
176
- 0.01536596965789795,
177
- 0.015742127418518066,
178
- 0.015599568367004395,
179
- 0.015377809524536133,
180
- 0.015506929397583007,
181
- 0.015337809562683106,
182
- 0.015443569183349609,
183
- 0.015534128189086914,
184
- 0.015490288734436036,
185
- 0.015886606216430663,
186
- 0.014853174209594727
 
 
 
 
187
  ]
188
  },
189
  "throughput": {
190
  "unit": "samples/s",
191
- "value": 62.841887187829904
192
  },
193
  "energy": null,
194
  "efficiency": null
 
85
  "gpu_count": 1,
86
  "gpu_vram_mb": 68702699520,
87
  "optimum_benchmark_version": "0.2.1",
88
+ "optimum_benchmark_commit": "074fc31a7e13e70facbebbf8b363a0c312f75f69",
89
  "transformers_version": "4.42.3",
90
  "transformers_commit": null,
91
  "accelerate_version": "0.31.0",
 
104
  "forward": {
105
  "memory": {
106
  "unit": "MB",
107
+ "max_ram": 1029.001216,
108
+ "max_global_vram": 1122.865152,
109
+ "max_process_vram": 228421.677056,
110
  "max_reserved": 773.849088,
111
  "max_allocated": 745.087488
112
  },
113
  "latency": {
114
  "unit": "s",
115
+ "count": 67,
116
+ "total": 1.006473002433777,
117
+ "mean": 0.015021985110951892,
118
+ "stdev": 0.000630600390596163,
119
+ "p50": 0.014817052841186524,
120
+ "p90": 0.015568344879150391,
121
+ "p95": 0.01573995990753174,
122
+ "p99": 0.016974796028137215,
123
  "values": [
124
+ 0.014683452606201172,
125
+ 0.015511928558349609,
126
+ 0.01576136779785156,
127
+ 0.01569000816345215,
128
+ 0.015431289672851562,
129
+ 0.015311610221862794,
130
+ 0.015291930198669434,
131
+ 0.015457049369812011,
132
+ 0.01540312957763672,
133
+ 0.015316729545593261,
134
+ 0.015537368774414062,
135
+ 0.01548616886138916,
136
+ 0.015488248825073242,
137
+ 0.015533848762512207,
138
+ 0.015161211013793946,
139
+ 0.015273050308227539,
140
+ 0.014954011917114257,
141
+ 0.015674009323120117,
142
+ 0.015243450164794921,
143
+ 0.01851591682434082,
144
+ 0.014888092041015625,
145
+ 0.014817052841186524,
146
+ 0.01448409366607666,
147
+ 0.014719613075256347,
148
+ 0.014489534378051758,
149
+ 0.014750811576843261,
150
+ 0.014644252777099609,
151
+ 0.014729052543640138,
152
+ 0.014446173667907715,
153
+ 0.014843932151794434,
154
+ 0.014374493598937988,
155
+ 0.014473053932189941,
156
+ 0.014821533203125,
157
+ 0.014422333717346192,
158
+ 0.014786492347717286,
159
+ 0.014548733711242675,
160
+ 0.014808571815490723,
161
+ 0.014443934440612793,
162
+ 0.014772412300109863,
163
+ 0.014438014030456543,
164
+ 0.014798172950744629,
165
+ 0.014488574028015136,
166
+ 0.014764573097229004,
167
+ 0.014508733749389649,
168
+ 0.014788891792297363,
169
+ 0.015580728530883789,
170
+ 0.01616312599182129,
171
+ 0.016180885314941405,
172
+ 0.015428089141845703,
173
+ 0.015242330551147461,
174
+ 0.015538969039916993,
175
+ 0.015315290451049804,
176
+ 0.015560089111328125,
177
+ 0.01475001335144043,
178
+ 0.014794492721557617,
179
+ 0.014467293739318848,
180
+ 0.014886172294616699,
181
+ 0.014485533714294434,
182
+ 0.014864252090454102,
183
+ 0.014458654403686523,
184
+ 0.014810332298278809,
185
+ 0.014430334091186523,
186
+ 0.014818653106689453,
187
+ 0.014563613891601563,
188
+ 0.014378334045410156,
189
+ 0.014545052528381348,
190
+ 0.01443385410308838
191
  ]
192
  },
193
  "throughput": {
194
  "unit": "samples/s",
195
+ "value": 66.56909806620314
196
  },
197
  "energy": null,
198
  "efficiency": null