PEFT
Safetensors
English
jinjieyuan commited on
Commit
0e4a2e4
1 Parent(s): 3b8e6a9

Update README.md

Browse files

Signed-off-by: jinjieyuan <[email protected]>

Files changed (1) hide show
  1. README.md +11 -4
README.md CHANGED
@@ -3,26 +3,27 @@ language: en
3
  license: apache-2.0
4
  ---
5
 
6
- # Shears Model Card: Shears-llama-13b-50-math-heuristic
7
 
8
- Fine tuned model on LLaMA-13B with some math reasoning datasets using Shears.
9
 
10
  ## Model Details
11
 
12
  ### Information
13
 
14
- - **Model name:** Shears-llama-13b-50-math-heuristic
15
  - **Base model:** [LLaMA-13b](https://huggingface.co/yahma/llama-13b-hf)
16
  - **Sparsity:** 50%
17
  - **Domain:** Math
18
  - **Subnetwork version:** Heuristic
 
19
 
20
  ### Adapter Configuration
21
 
22
  - **LoRA rank:** 32 (24 in the heuristic subnetwork)
23
  - **LoRA alpha:** 64
24
  - **LoRA target modules:** q_proj, k_proj, v_proj, up_proj, down_proj
25
- - **LoRA rank search space:** [32, 24, 16]
26
 
27
  ### Training Hyperparameters
28
 
@@ -40,6 +41,12 @@ Unified math reasoning dataset: [math_10k.json](https://github.com/AGI-Edgerunne
40
 
41
  ## How to use
42
 
 
 
 
 
 
 
43
  ```python
44
  import torch
45
  from peft import PeftModel
 
3
  license: apache-2.0
4
  ---
5
 
6
+ # Shears Model Card: shears-llama-13b-50-math-heuristic
7
 
8
+ The heuristic subnetwork discovered from the [super-network](https://huggingface.co/IntelLabs/shears-llama-13b-50-math-super) fine-tuned on LLaMA-13B with some math reasoning datasets using Shears.
9
 
10
  ## Model Details
11
 
12
  ### Information
13
 
14
+ - **Model name:** shears-llama-13b-50-math-heuristic
15
  - **Base model:** [LLaMA-13b](https://huggingface.co/yahma/llama-13b-hf)
16
  - **Sparsity:** 50%
17
  - **Domain:** Math
18
  - **Subnetwork version:** Heuristic
19
+ - **NNCF Configuration:** [nncf_shears_llama_13b_sparsity50.json](https://github.com/IntelLabs/Hardware-Aware-Automated-Machine-Learning/tree/main/Shears/nncf_config/unified_math/nncf_shears_llama_13b_sparsity50.json)
20
 
21
  ### Adapter Configuration
22
 
23
  - **LoRA rank:** 32 (24 in the heuristic subnetwork)
24
  - **LoRA alpha:** 64
25
  - **LoRA target modules:** q_proj, k_proj, v_proj, up_proj, down_proj
26
+ - **LoRA rank search space:** [32, 24, 16] (for each LoRA module)
27
 
28
  ### Training Hyperparameters
29
 
 
41
 
42
  ## How to use
43
 
44
+ Use our modified PEFT library (apply [patch](https://github.com/IntelLabs/Hardware-Aware-Automated-Machine-Learning/tree/main/Shears/patches/peft-modifications-for-shears-inference-usage.patch)):
45
+ ```bash
46
+ git clone https://github.com/huggingface/peft.git
47
+ pushd peft && git checkout v0.5.0 && git apply --ignore-space-change --ignore-whitespace peft-modifications-for-shears-inference-usage.patch && pip install -e . && popd
48
+ ```
49
+
50
  ```python
51
  import torch
52
  from peft import PeftModel