jinjieyuan
commited on
Commit
•
0e4a2e4
1
Parent(s):
3b8e6a9
Update README.md
Browse filesSigned-off-by: jinjieyuan <[email protected]>
README.md
CHANGED
@@ -3,26 +3,27 @@ language: en
|
|
3 |
license: apache-2.0
|
4 |
---
|
5 |
|
6 |
-
# Shears Model Card:
|
7 |
|
8 |
-
|
9 |
|
10 |
## Model Details
|
11 |
|
12 |
### Information
|
13 |
|
14 |
-
- **Model name:**
|
15 |
- **Base model:** [LLaMA-13b](https://huggingface.co/yahma/llama-13b-hf)
|
16 |
- **Sparsity:** 50%
|
17 |
- **Domain:** Math
|
18 |
- **Subnetwork version:** Heuristic
|
|
|
19 |
|
20 |
### Adapter Configuration
|
21 |
|
22 |
- **LoRA rank:** 32 (24 in the heuristic subnetwork)
|
23 |
- **LoRA alpha:** 64
|
24 |
- **LoRA target modules:** q_proj, k_proj, v_proj, up_proj, down_proj
|
25 |
-
- **LoRA rank search space:** [32, 24, 16]
|
26 |
|
27 |
### Training Hyperparameters
|
28 |
|
@@ -40,6 +41,12 @@ Unified math reasoning dataset: [math_10k.json](https://github.com/AGI-Edgerunne
|
|
40 |
|
41 |
## How to use
|
42 |
|
|
|
|
|
|
|
|
|
|
|
|
|
43 |
```python
|
44 |
import torch
|
45 |
from peft import PeftModel
|
|
|
3 |
license: apache-2.0
|
4 |
---
|
5 |
|
6 |
+
# Shears Model Card: shears-llama-13b-50-math-heuristic
|
7 |
|
8 |
+
The heuristic subnetwork discovered from the [super-network](https://huggingface.co/IntelLabs/shears-llama-13b-50-math-super) fine-tuned on LLaMA-13B with some math reasoning datasets using Shears.
|
9 |
|
10 |
## Model Details
|
11 |
|
12 |
### Information
|
13 |
|
14 |
+
- **Model name:** shears-llama-13b-50-math-heuristic
|
15 |
- **Base model:** [LLaMA-13b](https://huggingface.co/yahma/llama-13b-hf)
|
16 |
- **Sparsity:** 50%
|
17 |
- **Domain:** Math
|
18 |
- **Subnetwork version:** Heuristic
|
19 |
+
- **NNCF Configuration:** [nncf_shears_llama_13b_sparsity50.json](https://github.com/IntelLabs/Hardware-Aware-Automated-Machine-Learning/tree/main/Shears/nncf_config/unified_math/nncf_shears_llama_13b_sparsity50.json)
|
20 |
|
21 |
### Adapter Configuration
|
22 |
|
23 |
- **LoRA rank:** 32 (24 in the heuristic subnetwork)
|
24 |
- **LoRA alpha:** 64
|
25 |
- **LoRA target modules:** q_proj, k_proj, v_proj, up_proj, down_proj
|
26 |
+
- **LoRA rank search space:** [32, 24, 16] (for each LoRA module)
|
27 |
|
28 |
### Training Hyperparameters
|
29 |
|
|
|
41 |
|
42 |
## How to use
|
43 |
|
44 |
+
Use our modified PEFT library (apply [patch](https://github.com/IntelLabs/Hardware-Aware-Automated-Machine-Learning/tree/main/Shears/patches/peft-modifications-for-shears-inference-usage.patch)):
|
45 |
+
```bash
|
46 |
+
git clone https://github.com/huggingface/peft.git
|
47 |
+
pushd peft && git checkout v0.5.0 && git apply --ignore-space-change --ignore-whitespace peft-modifications-for-shears-inference-usage.patch && pip install -e . && popd
|
48 |
+
```
|
49 |
+
|
50 |
```python
|
51 |
import torch
|
52 |
from peft import PeftModel
|