{ "results": { "exams_dialy_vi": { "acc": 0.46869565217391307, "acc_stderr": 0.01040752976753813, "acc_norm": 0.52, "acc_norm_stderr": 0.010419642173256296 }, "exams_hoahoc_vi": { "acc": 0.27578947368421053, "acc_stderr": 0.014507352152603738, "acc_norm": 0.2926315789473684, "acc_norm_stderr": 0.014768973899438454 }, "exams_lichsu_vi": { "acc": 0.45532710280373834, "acc_stderr": 0.006809156793452153, "acc_norm": 0.5196261682242991, "acc_norm_stderr": 0.0068312295416031125 }, "exams_sinhhoc_vi": { "acc": 0.3506451612903226, "acc_stderr": 0.008571645026979423, "acc_norm": 0.3996774193548387, "acc_norm_stderr": 0.008799061593983897 }, "exams_toan_vi": { "acc": 0.22842105263157894, "acc_stderr": 0.007865240816946034, "acc_norm": 0.2536842105263158, "acc_norm_stderr": 0.008151955663264111 }, "exams_vatly_vi": { "acc": 0.3211764705882353, "acc_stderr": 0.016024941739325747, "acc_norm": 0.3729411764705882, "acc_norm_stderr": 0.01659665112186545 }, "exams_van_vi": { "acc": 0.38704225352112676, "acc_stderr": 0.008176010850435296, "acc_norm": 0.4445070422535211, "acc_norm_stderr": 0.008341143676748432 }, "exams_vi": { "acc": 0.376822866565892, "acc_norm": 0.426607118317988 } }, "versions": { "exams_dialy_vi": 0, "exams_hoahoc_vi": 0, "exams_lichsu_vi": 0, "exams_sinhhoc_vi": 0, "exams_toan_vi": 0, "exams_vatly_vi": 0, "exams_van_vi": 0 }, "config": { "model": "hf-causal", "model_args": "pretrained=vietgpt/dama-2-7b", "num_fewshot": 5, "batch_size": null, "batch_sizes": [], "device": "cuda:0", "no_cache": true, "limit": null, "bootstrap_iters": 100000, "description_dict": {} } }