IlyasMoutawwakil HF staff commited on
Commit
610961d
·
verified ·
1 Parent(s): 3353870

Upload cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark_report.json with huggingface_hub

Browse files
cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark_report.json CHANGED
@@ -2,7 +2,7 @@
2
  "forward": {
3
  "memory": {
4
  "unit": "MB",
5
- "max_ram": 907.698176,
6
  "max_global_vram": 1195.900928,
7
  "max_process_vram": 0.0,
8
  "max_reserved": 555.74528,
@@ -10,167 +10,163 @@
10
  },
11
  "latency": {
12
  "unit": "s",
13
- "count": 136,
14
- "total": 0.9994755530357362,
15
- "mean": 0.007349084948792177,
16
- "stdev": 0.0003792756581833165,
17
- "p50": 0.007301120042800903,
18
- "p90": 0.007598079919815063,
19
- "p95": 0.008035840272903442,
20
- "p99": 0.008403537750244142,
21
  "values": [
22
- 0.010314751625061035,
23
- 0.008490976333618163,
24
- 0.008166399955749512,
25
- 0.008033280372619628,
26
- 0.008032192230224609,
27
- 0.008043519973754883,
28
- 0.0082227201461792,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
29
  0.007953407764434815,
30
- 0.008241151809692383,
31
- 0.007333888053894043,
32
- 0.007159808158874512,
33
- 0.007098368167877197,
34
- 0.007106560230255127,
35
- 0.0071198720932006835,
36
- 0.007112671852111816,
37
- 0.007093247890472412,
38
- 0.00769536018371582,
39
- 0.007529471874237061,
40
- 0.007667712211608887,
41
- 0.008084447860717773,
42
- 0.007585792064666748,
43
- 0.007507967948913574,
44
- 0.0074997758865356446,
45
- 0.007477248191833496,
46
- 0.007534527778625488,
47
- 0.007610367774963379,
48
- 0.0074967041015625,
49
- 0.007423999786376953,
50
- 0.0074700798988342285,
51
- 0.007478271961212158,
52
- 0.007290880203247071,
53
- 0.007166975975036621,
54
- 0.007173120021820068,
55
- 0.007111680030822754,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
56
  0.007165952205657959,
57
- 0.007112703800201416,
58
- 0.007122943878173828,
59
- 0.007132127761840821,
60
- 0.00709119987487793,
 
 
 
 
 
 
61
  0.007120895862579346,
62
- 0.007101439952850342,
63
- 0.00714137601852417,
64
- 0.007098368167877197,
65
- 0.007101439952850342,
66
- 0.007134208202362061,
67
- 0.007095295906066895,
68
- 0.007123968124389648,
69
- 0.00710041618347168,
70
- 0.007105535984039306,
71
- 0.0071198720932006835,
72
- 0.007127039909362793,
73
- 0.007135231971740722,
74
- 0.007097343921661377,
75
- 0.0071157760620117185,
76
- 0.007090176105499267,
77
- 0.007107583999633789,
78
- 0.007072768211364746,
79
- 0.007087103843688965,
80
- 0.00710041618347168,
81
- 0.007102464199066162,
82
- 0.007109632015228271,
83
- 0.007048096179962158,
84
- 0.007110655784606934,
85
- 0.007104512214660645,
86
- 0.007130112171173096,
87
- 0.007090176105499267,
88
- 0.007677951812744141,
89
- 0.007126016139984131,
90
- 0.007117824077606201,
91
- 0.00709119987487793,
92
- 0.007107583999633789,
93
- 0.007158783912658692,
94
- 0.007016448020935059,
95
- 0.007031807899475098,
96
- 0.007076863765716553,
97
- 0.0070594558715820314,
98
- 0.007203839778900147,
99
- 0.0071198720932006835,
100
- 0.007114783763885498,
101
- 0.007094272136688232,
102
- 0.0070696320533752445,
103
- 0.007101439952850342,
104
- 0.007097248077392578,
105
- 0.00713318395614624,
106
- 0.007074816226959229,
107
- 0.007047167778015137,
108
- 0.007093247890472412,
109
- 0.007098368167877197,
110
- 0.00708403205871582,
111
- 0.007057407855987549,
112
- 0.007116799831390381,
113
- 0.007096320152282715,
114
- 0.0074741759300231934,
115
- 0.00748748779296875,
116
- 0.007504896163940429,
117
- 0.007418879985809326,
118
- 0.0074332160949707035,
119
- 0.007423999786376953,
120
- 0.007457791805267334,
121
- 0.007437312126159668,
122
- 0.007344128131866455,
123
- 0.007351295948028564,
124
- 0.00729804801940918,
125
- 0.007367680072784424,
126
- 0.007304192066192627,
127
- 0.007365568161010742,
128
- 0.007395328044891358,
129
- 0.0074414081573486324,
130
- 0.007409632205963135,
131
- 0.007386112213134765,
132
- 0.007419904232025146,
133
- 0.007468031883239746,
134
- 0.007437312126159668,
135
- 0.0074414081573486324,
136
- 0.007425024032592774,
137
- 0.007431168079376221,
138
- 0.007395328044891358,
139
- 0.007450623989105225,
140
- 0.007457791805267334,
141
- 0.00746284818649292,
142
- 0.00745472002029419,
143
- 0.007435264110565186,
144
- 0.007419904232025146,
145
- 0.007398399829864502,
146
- 0.007386112213134765,
147
- 0.007457791805267334,
148
- 0.007423999786376953,
149
- 0.007426047801971435,
150
- 0.007421919822692871,
151
- 0.007434239864349365,
152
- 0.007459839820861816,
153
- 0.007425024032592774,
154
- 0.007428095817565918,
155
- 0.007431168079376221,
156
- 0.007445504188537597,
157
- 0.00744652795791626
158
  ]
159
  },
160
  "throughput": {
161
  "unit": "samples/s",
162
- "value": 136.07136221283577
163
  },
164
  "energy": {
165
  "unit": "kWh",
166
- "cpu": 8.59822675993669e-08,
167
- "ram": 4.7003675358016906e-08,
168
- "gpu": 1.4988135984057874e-07,
169
- "total": 2.8286730279796253e-07
170
  },
171
  "efficiency": {
172
  "unit": "samples/kWh",
173
- "value": 3535226.55361213
174
  }
175
  }
176
  }
 
2
  "forward": {
3
  "memory": {
4
  "unit": "MB",
5
+ "max_ram": 907.358208,
6
  "max_global_vram": 1195.900928,
7
  "max_process_vram": 0.0,
8
  "max_reserved": 555.74528,
 
10
  },
11
  "latency": {
12
  "unit": "s",
13
+ "count": 132,
14
+ "total": 1.0019654092788697,
15
+ "mean": 0.007590647039991437,
16
+ "stdev": 0.0005809214402145674,
17
+ "p50": 0.007497215986251832,
18
+ "p90": 0.007849881744384765,
19
+ "p95": 0.008180326747894286,
20
+ "p99": 0.010368777799606323,
21
  "values": [
22
+ 0.010809344291687012,
23
+ 0.010365951538085937,
24
+ 0.010370047569274902,
25
+ 0.010334207534790038,
26
+ 0.008263680458068847,
27
+ 0.008077247619628906,
28
+ 0.007772160053253174,
29
+ 0.007849984169006348,
30
+ 0.007860223770141601,
31
+ 0.00775980806350708,
32
+ 0.007813119888305664,
33
+ 0.007738368034362793,
34
+ 0.00760422420501709,
35
+ 0.007623680114746094,
36
+ 0.007647232055664062,
37
+ 0.007670783996582031,
38
+ 0.0077281279563903805,
39
+ 0.007759871959686279,
40
+ 0.007775231838226319,
41
+ 0.007679999828338623,
42
+ 0.007666687965393066,
43
+ 0.00744755220413208,
44
+ 0.007442431926727295,
45
+ 0.007530432224273681,
46
+ 0.007450623989105225,
47
+ 0.00744755220413208,
48
+ 0.0073697280883789065,
49
+ 0.007373824119567871,
50
+ 0.00738099193572998,
51
+ 0.007883776187896728,
52
+ 0.007751679897308349,
53
+ 0.007715839862823487,
54
+ 0.007575551986694336,
55
+ 0.007740416049957275,
56
+ 0.007640063762664795,
57
+ 0.007618559837341309,
58
+ 0.007633920192718506,
59
+ 0.007709695816040039,
60
+ 0.007714848041534424,
61
+ 0.007730175971984863,
62
+ 0.007783423900604248,
63
+ 0.007670783996582031,
64
+ 0.007659520149230957,
65
+ 0.007577600002288819,
66
+ 0.007701504230499268,
67
+ 0.007727136135101318,
68
+ 0.007686143875122071,
69
+ 0.007473152160644531,
70
+ 0.007404543876647949,
71
+ 0.007310336112976074,
72
+ 0.007361536026000977,
73
+ 0.007288832187652588,
74
+ 0.007318528175354004,
75
+ 0.007401408195495606,
76
+ 0.007353343963623047,
77
+ 0.007592959880828858,
78
+ 0.007602176189422607,
79
+ 0.007684095859527588,
80
+ 0.007584767818450928,
81
+ 0.007624703884124756,
82
+ 0.007709695816040039,
83
+ 0.007574528217315674,
84
+ 0.007775231838226319,
85
+ 0.007469056129455566,
86
+ 0.007442431926727295,
87
+ 0.007435264110565186,
88
+ 0.00808140754699707,
89
+ 0.008516608238220215,
90
  0.007953407764434815,
91
+ 0.007694272041320801,
92
+ 0.00774348783493042,
93
+ 0.007689216136932373,
94
+ 0.007505919933319092,
95
+ 0.007451648235321045,
96
+ 0.007628896236419678,
97
+ 0.007448575973510742,
98
+ 0.0072499198913574215,
99
+ 0.007223296165466309,
100
+ 0.0074403839111328125,
101
+ 0.007488512039184571,
102
+ 0.007554048061370849,
103
+ 0.007553023815155029,
104
+ 0.007576576232910156,
105
+ 0.007797760009765625,
106
+ 0.007848959922790527,
107
+ 0.007590911865234375,
108
+ 0.007327744007110596,
109
+ 0.008112128257751466,
110
+ 0.007512063980102539,
111
+ 0.0072837119102478025,
112
+ 0.007245823860168457,
113
+ 0.00724889612197876,
114
+ 0.0071495680809020995,
115
+ 0.007219200134277344,
116
+ 0.007226367950439453,
117
+ 0.007227424144744873,
118
+ 0.007808000087738037,
119
+ 0.007456768035888672,
120
+ 0.007237631797790528,
121
+ 0.0071823358535766605,
122
+ 0.007185408115386963,
123
+ 0.00719974422454834,
124
+ 0.007180287837982178,
125
+ 0.007193600177764893,
126
+ 0.007158783912658692,
127
+ 0.007160799980163574,
128
+ 0.00838144016265869,
129
+ 0.007326720237731933,
130
+ 0.007231488227844239,
131
+ 0.0071987838745117185,
132
  0.007165952205657959,
133
+ 0.007201791763305664,
134
+ 0.007181312084197998,
135
+ 0.007197696208953858,
136
+ 0.007192575931549072,
137
+ 0.0071792640686035155,
138
+ 0.007153664112091064,
139
+ 0.007153632164001465,
140
+ 0.007222271919250488,
141
+ 0.0072120318412780765,
142
+ 0.007187456130981445,
143
  0.007120895862579346,
144
+ 0.007156735897064209,
145
+ 0.007221216201782226,
146
+ 0.007202816009521484,
147
+ 0.007206912040710449,
148
+ 0.007145472049713135,
149
+ 0.007170048236846924,
150
+ 0.007160831928253173,
151
+ 0.007176191806793213,
152
+ 0.007171072006225586,
153
+ 0.007153664112091064
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
154
  ]
155
  },
156
  "throughput": {
157
  "unit": "samples/s",
158
+ "value": 131.74107486904413
159
  },
160
  "energy": {
161
  "unit": "kWh",
162
+ "cpu": 8.530816633543142e-08,
163
+ "ram": 4.662791786273805e-08,
164
+ "gpu": 1.6067358976977895e-07,
165
+ "total": 2.926096739679484e-07
166
  },
167
  "efficiency": {
168
  "unit": "samples/kWh",
169
+ "value": 3417522.0061573805
170
  }
171
  }
172
  }