IlyasMoutawwakil HF staff commited on
Commit
eeef4a3
·
verified ·
1 Parent(s): 545521b

Upload cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark_report.json with huggingface_hub

Browse files
cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark_report.json CHANGED
@@ -2,161 +2,151 @@
2
  "forward": {
3
  "memory": {
4
  "unit": "MB",
5
- "max_ram": 1004.060672,
6
- "max_global_vram": 1878.831104,
7
- "max_process_vram": 202205.937664,
8
  "max_reserved": 555.74528,
9
  "max_allocated": 499.507712
10
  },
11
  "latency": {
12
  "unit": "s",
13
- "count": 133,
14
- "total": 0.999595571517944,
15
- "mean": 0.007515756176826649,
16
- "stdev": 0.00037991401469624283,
17
- "p50": 0.007288486957550049,
18
- "p90": 0.008080583381652831,
19
- "p95": 0.008092551231384278,
20
- "p99": 0.008191104660034181,
21
  "values": [
22
- 0.008225767135620118,
23
- 0.007568325996398926,
24
- 0.00800352668762207,
25
- 0.007811526775360108,
26
- 0.00797616720199585,
27
- 0.008106085777282715,
28
- 0.007792645931243896,
29
- 0.007912967205047607,
30
- 0.007996807098388672,
31
- 0.008070087432861328,
32
- 0.00806896686553955,
33
- 0.00767456579208374,
34
- 0.007627206802368164,
35
- 0.007576645851135254,
36
- 0.007590726852416992,
37
- 0.007584805965423584,
38
- 0.0074118471145629885,
39
- 0.007368966102600098,
40
- 0.007229606151580811,
41
- 0.008690407752990723,
42
- 0.008053927421569825,
43
- 0.008087845802307128,
44
- 0.007979366779327392,
45
- 0.00809360694885254,
46
- 0.008077767372131348,
47
- 0.00806736660003662,
48
- 0.00808432674407959,
49
- 0.008091847419738769,
50
- 0.008076327323913574,
51
- 0.008084807395935058,
52
- 0.008085447311401367,
53
- 0.008057126998901367,
54
- 0.008077607154846192,
55
- 0.008054885864257812,
56
- 0.008065446853637696,
57
- 0.008063687324523925,
58
- 0.008056487083435059,
59
- 0.008086887359619141,
60
- 0.008102087020874024,
61
- 0.008071526527404786,
62
- 0.008071367263793944,
63
- 0.00805424690246582,
64
- 0.008117446899414063,
65
- 0.008081287384033202,
66
- 0.007522885799407959,
67
- 0.007453446865081787,
68
- 0.007149126052856445,
69
- 0.007150246143341065,
70
- 0.007166726112365723,
71
- 0.007152006149291992,
72
- 0.007177765846252441,
73
- 0.007143846035003662,
74
- 0.0071752071380615234,
75
- 0.007156325817108154,
76
- 0.007157286167144775,
77
- 0.00738304615020752,
78
- 0.00713904619216919,
79
- 0.007140006065368652,
80
- 0.0071592068672180174,
81
- 0.007123526096343994,
82
- 0.007137765884399414,
83
- 0.007131525993347168,
84
- 0.007157125949859619,
85
- 0.007187205791473389,
86
- 0.007147046089172363,
87
- 0.007154726028442383,
88
- 0.007170086860656739,
89
- 0.00721872615814209,
90
- 0.007186245918273926,
91
- 0.007188965797424316,
92
- 0.0071680059432983395,
93
- 0.007166245937347412,
94
- 0.007188646793365478,
95
- 0.0071624059677124026,
96
- 0.0071716861724853516,
97
- 0.007185445785522461,
98
- 0.007166245937347412,
99
- 0.007152805805206299,
100
- 0.0072016057968139644,
101
- 0.007190086841583252,
102
- 0.007169285774230957,
103
- 0.007172646045684815,
104
- 0.0071680059432983395,
105
- 0.007216485977172851,
106
- 0.0071747260093688965,
107
- 0.007216967105865478,
108
- 0.0071985659599304195,
109
- 0.007160165786743164,
110
- 0.0072016057968139644,
111
- 0.007183526039123535,
112
- 0.0071766462326049805,
113
- 0.007157286167144775,
114
- 0.007152966976165772,
115
- 0.007152326107025146,
116
- 0.007469605922698974,
117
- 0.007752007007598877,
118
- 0.00743440580368042,
119
- 0.007499046802520752,
120
- 0.00749936580657959,
121
- 0.007599845886230468,
122
- 0.00764128589630127,
123
- 0.00765360689163208,
124
- 0.00753536605834961,
125
- 0.0075347270965576175,
126
- 0.007536166191101075,
127
- 0.007217926025390625,
128
- 0.007260806083679199,
129
- 0.007604805946350097,
130
- 0.008114727020263671,
131
- 0.007839527130126953,
132
- 0.007663045883178711,
133
- 0.0076612868309020995,
134
- 0.00767456579208374,
135
- 0.007666886806488037,
136
- 0.007692166805267334,
137
- 0.007256966114044189,
138
- 0.007283205986022949,
139
- 0.007240325927734375,
140
- 0.00726352596282959,
141
- 0.007240807056427002,
142
- 0.007284485816955567,
143
- 0.007263366222381592,
144
- 0.007277445793151855,
145
- 0.007256007194519043,
146
- 0.007244326114654541,
147
- 0.00723952579498291,
148
- 0.0072670459747314455,
149
- 0.0072480058670043946,
150
- 0.007293286800384521,
151
- 0.007263366222381592,
152
- 0.007238565921783447,
153
- 0.00728176498413086,
154
- 0.007288486957550049
155
  ]
156
  },
157
  "throughput": {
158
  "unit": "samples/s",
159
- "value": 133.05381075071364
160
  },
161
  "energy": null,
162
  "efficiency": null
 
2
  "forward": {
3
  "memory": {
4
  "unit": "MB",
5
+ "max_ram": 1004.847104,
6
+ "max_global_vram": 3097.423872,
7
+ "max_process_vram": 224252.178432,
8
  "max_reserved": 555.74528,
9
  "max_allocated": 499.507712
10
  },
11
  "latency": {
12
  "unit": "s",
13
+ "count": 123,
14
+ "total": 0.9959080824851987,
15
+ "mean": 0.008096813678741455,
16
+ "stdev": 0.00035558714977267883,
17
+ "p50": 0.008307499885559082,
18
+ "p90": 0.008389163398742676,
19
+ "p95": 0.008405723190307617,
20
+ "p99": 0.008563761672973633,
21
  "values": [
22
+ 0.008784618377685547,
23
+ 0.007700459957122803,
24
+ 0.007709421157836914,
25
+ 0.007648621082305908,
26
+ 0.007698861122131348,
27
+ 0.0075878210067749025,
28
+ 0.007505260944366455,
29
+ 0.007494861125946045,
30
+ 0.007476781845092774,
31
+ 0.007473421096801758,
32
+ 0.007453580856323242,
33
+ 0.007416141986846924,
34
+ 0.00749678087234497,
35
+ 0.007409582138061524,
36
+ 0.007482380867004395,
37
+ 0.007416141033172607,
38
+ 0.007426541805267334,
39
+ 0.007459500789642334,
40
+ 0.007426382064819336,
41
+ 0.007426701068878174,
42
+ 0.007417901992797852,
43
+ 0.007495660781860352,
44
+ 0.007872620105743407,
45
+ 0.007932139873504638,
46
+ 0.00794878101348877,
47
+ 0.007902700901031495,
48
+ 0.007895180225372314,
49
+ 0.007774701118469238,
50
+ 0.007754539966583252,
51
+ 0.007497422218322754,
52
+ 0.00745838212966919,
53
+ 0.007452301025390625,
54
+ 0.008156299591064453,
55
+ 0.007790380001068116,
56
+ 0.008057740211486817,
57
+ 0.0082006196975708,
58
+ 0.008325899124145508,
59
+ 0.008333418846130371,
60
+ 0.008337418556213378,
61
+ 0.008352779388427734,
62
+ 0.008363018989562989,
63
+ 0.008345418930053711,
64
+ 0.008342219352722167,
65
+ 0.00835901927947998,
66
+ 0.00832446002960205,
67
+ 0.008363658905029298,
68
+ 0.008385579109191894,
69
+ 0.008368939399719237,
70
+ 0.008378539085388183,
71
+ 0.008345099449157714,
72
+ 0.008390059471130371,
73
+ 0.008313419342041016,
74
+ 0.008358058929443359,
75
+ 0.008361578941345215,
76
+ 0.00839021873474121,
77
+ 0.008300780296325683,
78
+ 0.008415658950805665,
79
+ 0.008336459159851074,
80
+ 0.008266058921813965,
81
+ 0.008293739318847656,
82
+ 0.008330060005187987,
83
+ 0.008393259048461913,
84
+ 0.007939499855041504,
85
+ 0.007575981140136719,
86
+ 0.007746860027313233,
87
+ 0.007942541122436524,
88
+ 0.007696619987487793,
89
+ 0.007824141025543213,
90
+ 0.00787805986404419,
91
+ 0.007769101142883301,
92
+ 0.007687820911407471,
93
+ 0.008096940040588379,
94
+ 0.007825261116027831,
95
+ 0.008015179634094238,
96
+ 0.008361259460449218,
97
+ 0.008350058555603028,
98
+ 0.008290059089660644,
99
+ 0.008375338554382324,
100
+ 0.008452939033508302,
101
+ 0.008396939277648926,
102
+ 0.008319178581237792,
103
+ 0.008332460403442384,
104
+ 0.008392780303955079,
105
+ 0.008296939849853515,
106
+ 0.00834269905090332,
107
+ 0.008325419425964355,
108
+ 0.008333418846130371,
109
+ 0.008595019340515138,
110
+ 0.00831533908843994,
111
+ 0.008262378692626953,
112
+ 0.008346059799194335,
113
+ 0.008302539825439452,
114
+ 0.008293099403381347,
115
+ 0.00827453899383545,
116
+ 0.008373899459838867,
117
+ 0.008350700378417968,
118
+ 0.00842333984375,
119
+ 0.00831037998199463,
120
+ 0.008361259460449218,
121
+ 0.008324778556823731,
122
+ 0.008312458992004395,
123
+ 0.00831037998199463,
124
+ 0.008322380065917968,
125
+ 0.008304299354553223,
126
+ 0.008285738945007325,
127
+ 0.00834269905090332,
128
+ 0.008324459075927734,
129
+ 0.00834973907470703,
130
+ 0.00826509952545166,
131
+ 0.008318058967590332,
132
+ 0.008269739151000977,
133
+ 0.008307499885559082,
134
+ 0.008340459823608398,
135
+ 0.00833661937713623,
136
+ 0.008356939315795899,
137
+ 0.008351498603820801,
138
+ 0.008368297576904297,
139
+ 0.008411178588867188,
140
+ 0.008395018577575684,
141
+ 0.008363658905029298,
142
+ 0.008406699180603027,
143
+ 0.008255660057067871,
144
+ 0.008332619667053223
 
 
 
 
 
 
 
 
 
 
145
  ]
146
  },
147
  "throughput": {
148
  "unit": "samples/s",
149
+ "value": 123.50537380223341
150
  },
151
  "energy": null,
152
  "efficiency": null