--- language: - en license: other tags: - chat license_name: tongyi-qianwen license_link: https://huggingface.co/Qwen/Qwen2-72B-Instruct/blob/main/LICENSE pipeline_tag: text-generation --- # Dracarys-72B-Instruct # Introduction We introduce the latest in the Smaug series, the Dracarys family of finetunes targeting coding performance improvements across a variety of base models. This variant is a finetune of [Qwen2-72B-Instruct](https://huggingface.co/Qwen/Qwen2-72B-Instruct) Compared to Qwen2-72B-Instruct, Dracarys has better LiveCodeBench scores (see evaluation results below). ### Model Description - **Developed by:** [Abacus.AI](https://abacus.ai) - **License:** https://huggingface.co/Qwen/Qwen2-72B-Instruct/blob/main/LICENSE - **Finetuned from model:** [Qwen2-72B-Instruct](https://huggingface.co/Qwen/Qwen2-72B-Instruct). ## How to use The prompt format is unchanged from Qwen2-72B-Instruct (see evaluations for prompt details for LCB) ### Use with transformers See the snippet below for usage with Transformers: ```python import transformers import torch model_id = "abacusai/Dracarys-72B-Instruct" pipeline = transformers.pipeline( "text-generation", model=model_id, model_kwargs={"torch_dtype": torch.bfloat16}, device_map="auto", ) messages = [ {"role": "system", "content": "You are data science coding assistant that generates Python code using Pandas and Numpy."}, {"role": "user", "content": "Write code to select rows from the dataframe `df` having the maximum `temp` for each `city`"}, ] prompt = pipeline.tokenizer.apply_chat_template( messages, tokenize=False, add_generation_prompt=True ) terminators = [ pipeline.tokenizer.eos_token_id, pipeline.tokenizer.convert_tokens_to_ids("<|eot_id|>") ] outputs = pipeline( prompt, max_new_tokens=256, eos_token_id=terminators, do_sample=True, temperature=0.6, top_p=0.9, ) print(outputs[0]["generated_text"][len(prompt):]) ``` # Evaluation Results ## LiveCodeBench | Model | Code Generation | Code Execution |Test Output Prediction | |---------------------------|-----------------|----------------|-----------------------| | **Dracarys-72B-Instruct** | **33.57** | **62.96** | **58.93** | | Qwen2-72B-Instruct | 32.92 | 58.95 | 55.88 | ## Breakdown of LiveCodeBench CodeGeneration | Model | Easy | Medium | Hard | |---------------------------|-----------------|----------------|-----------------------| | **Dracarys-72B-Instruct** | 64.16 | **25.06** | **3.64** | | Qwen2-72B-Instruct | 65.83 | 22.28 | 3.11 | ## Breakdown of LiveCodeBench TestOutputPrediction | Model | Easy | Medium | Hard | |---------------------------|-----------------|----------------|-----------------------| | **Dracarys-72B-Instruct** | **65.37** | **58.74** | **46.38** | | Qwen2-72B-Instruct | 63.19 | 54.08 | 46.52 | ## LiveBench (July update) | Model | Global Average | Coding Average | Language Average | Mathematics Average | Data Analysis Average | Reasoning Average | IF Average | |---------------------------|----------------|----------------|------------------|---------------------|-----------------------|------------------|-------------| | **Dracarys-72B-Instruct** | **41.20** | **38.95** | **31.17** | 42.77 | 26.24 | 40 | 68.08 | | Qwen2-72B-Instruct | 40.15 | 32.38 | 29.21 | 43.44 | 26.24 | 41.33 | 68.27 |