Aaron Mueller commited on
Commit
43332b1
·
1 Parent(s): 59a9012

next iteration

Browse files
Files changed (3) hide show
  1. app.py +0 -8
  2. src/leaderboard/read_evals.py +0 -5
  3. src/populate.py +2 -2
app.py CHANGED
@@ -73,14 +73,6 @@ def init_leaderboard(dataframe, track):
73
  hide_columns=[c.name for c in fields(AutoEvalColumn) if c.hidden],
74
  filter_columns=[
75
  ColumnFilter(AutoEvalColumn.model_type.name, type="checkboxgroup", label="Model types"),
76
- ColumnFilter(AutoEvalColumn.precision.name, type="checkboxgroup", label="Precision"),
77
- ColumnFilter(
78
- AutoEvalColumn.params.name,
79
- type="slider",
80
- min=0.01,
81
- max=150,
82
- label="Select the number of parameters (B)",
83
- ),
84
  ColumnFilter(
85
  AutoEvalColumn.still_on_hub.name, type="boolean", label="Deleted/incomplete", default=True
86
  ),
 
73
  hide_columns=[c.name for c in fields(AutoEvalColumn) if c.hidden],
74
  filter_columns=[
75
  ColumnFilter(AutoEvalColumn.model_type.name, type="checkboxgroup", label="Model types"),
 
 
 
 
 
 
 
 
76
  ColumnFilter(
77
  AutoEvalColumn.still_on_hub.name, type="boolean", label="Deleted/incomplete", default=True
78
  ),
src/leaderboard/read_evals.py CHANGED
@@ -154,11 +154,6 @@ def get_raw_eval_results(results_path: str, requests_path: str) -> list[EvalResu
154
 
155
  eval_results = {}
156
  for model_result_filepath in model_result_filepaths:
157
- with open(model_result_filepath, 'r') as f:
158
- this_track = f["track"]
159
- if this_track != track:
160
- continue
161
-
162
  # Creation of result
163
  eval_result = EvalResult.init_from_json_file(model_result_filepath)
164
  eval_result.update_with_request_file(requests_path)
 
154
 
155
  eval_results = {}
156
  for model_result_filepath in model_result_filepaths:
 
 
 
 
 
157
  # Creation of result
158
  eval_result = EvalResult.init_from_json_file(model_result_filepath)
159
  eval_result.update_with_request_file(requests_path)
src/populate.py CHANGED
@@ -14,8 +14,8 @@ def get_leaderboard_df(results_path: str, requests_path: str, cols: list, benchm
14
  all_data_json = [v.to_dict() for v in raw_data]
15
 
16
  df = pd.DataFrame.from_records(all_data_json)
17
- df = df.sort_values(by=[AutoEvalColumn.average.name], ascending=False)
18
- df = df[cols].round(decimals=2)
19
 
20
  # filter out if any of the benchmarks have not been produced
21
  df = df[has_no_nan_values(df, benchmark_cols)]
 
14
  all_data_json = [v.to_dict() for v in raw_data]
15
 
16
  df = pd.DataFrame.from_records(all_data_json)
17
+ # df = df.sort_values(by=[AutoEvalColumn.average.name], ascending=False)
18
+ df = df[cols].round(decimals=1)
19
 
20
  # filter out if any of the benchmarks have not been produced
21
  df = df[has_no_nan_values(df, benchmark_cols)]