rwmasood commited on
Commit
da88633
·
verified ·
1 Parent(s): 7e326b5

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +37 -0
README.md ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: llama3.3
3
+ datasets:
4
+ - allenai/c4
5
+ language:
6
+ - en
7
+ base_model:
8
+ - meta-llama/Llama-3.3-70B-Instruct
9
+ ---
10
+
11
+
12
+ # Overview
13
+ This document presents the evaluation results of `DeepSeek-R1-Distill-Llama-70B`, a **4-bit quantized model using GPTQ**, evaluated with the **Language Model Evaluation Harness** on the **ARC-Challenge** benchmark.
14
+
15
+
16
+ ## ⚙️ Model Configuration
17
+
18
+ - **Model:** `Llama-3.3-70B-Instruct`
19
+ - **Parameters:** `70 billion`
20
+ - **Quantization:** `4-bit GPTQ`
21
+ - **Source:** Hugging Face (`hf`)
22
+ - **Precision:** `torch.float16`
23
+ - **Hardware:** `NVIDIA A100 80GB PCIe`
24
+ - **CUDA Version:** `12.4`
25
+ - **PyTorch Version:** `2.6.0+cu124`
26
+ - **Batch Size:** `1`
27
+
28
+
29
+ 📌 **Interpretation:**
30
+ - The evaluation was performed on a **high-performance GPU (A100 80GB)**.
31
+ - The model is significantly larger than the previous 8B version, with **GPTQ 4-bit quantization reducing memory footprint**.
32
+ - A **single-sample batch size** was used, which might slow evaluation speed.
33
+
34
+ ---
35
+
36
+
37
+ 📌 Let us know if you need further analysis or model tuning! 🚀