AI Model Name: Llama 3 8B "Built with Meta Llama 3" https://llama.meta.com/llama3/license/ Full walkthrough to reproduce these results here: https://github.com/catid/AQLM/blob/main/catid_readme.md Baseline evaluation results: ``` hf (pretrained=meta-llama/Meta-Llama-3-8B-Instruct), gen_kwargs: (None), limit: None, num_fewshot: None, batch_size: 16 | Tasks |Version|Filter|n-shot| Metric |Value | |Stderr| |-------------|------:|------|-----:|--------|-----:|---|-----:| |winogrande | 1|none | 0|acc |0.7198|± |0.0126| |piqa | 1|none | 0|acc |0.7873|± |0.0095| | | |none | 0|acc_norm|0.7867|± |0.0096| |hellaswag | 1|none | 0|acc |0.5767|± |0.0049| | | |none | 0|acc_norm|0.7585|± |0.0043| |arc_easy | 1|none | 0|acc |0.8140|± |0.0080| | | |none | 0|acc_norm|0.7971|± |0.0083| |arc_challenge| 1|none | 0|acc |0.5290|± |0.0146| | | |none | 0|acc_norm|0.5674|± |0.0145| ``` This repo evaluation results (AQLM with global fine-tuning): ``` hf (pretrained=catid/cat-llama-3-8b-instruct-aqlm), gen_kwargs: (None), limit: None, num_fewshot: None, batch_size: 16 | Tasks |Version|Filter|n-shot| Metric |Value | |Stderr| |-------------|------:|------|-----:|--------|-----:|---|-----:| |winogrande | 1|none | 0|acc |0.7119|± |0.0127| |piqa | 1|none | 0|acc |0.7807|± |0.0097| | | |none | 0|acc_norm|0.7824|± |0.0096| |hellaswag | 1|none | 0|acc |0.5716|± |0.0049| | | |none | 0|acc_norm|0.7539|± |0.0043| |arc_easy | 1|none | 0|acc |0.8152|± |0.0080| | | |none | 0|acc_norm|0.7866|± |0.0084| |arc_challenge| 1|none | 0|acc |0.5043|± |0.0146| | | |none | 0|acc_norm|0.5555|± |0.0145| ``` To reproduce evaluation results: ```bash git clone https://github.com/EleutherAI/lm-evaluation-harness cd lm-evaluation-harness conda create -n lmeval python=3.10 -y && conda activate lmeval pip install -e . pip install accelerate aqlm"[gpu,cpu]" accelerate launch lm_eval --model hf \ --model_args pretrained=catid/cat-llama-3-8b-instruct-aqlm \ --tasks winogrande,piqa,hellaswag,arc_easy,arc_challenge \ --batch_size 16 ``` You can run this model as a `transformers` model using https://github.com/oobabooga/text-generation-webui