File size: 695 Bytes
2d76661
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
- Training Parameters: 
  - base_model: TinyLlama/TinyLlama_v1.1
  - save_ckpt_log_name: tinyllama_prune_log
  - pruning_ratio: 0.39
  - pruner_type: taylor
  - temperature: 1.0
  - top_p: 0.95
  - max_seq_len: 2048
  - channel_wise: False
  - block_wise: True
  - layer_wise: False
  - layer: 12
  - block_attention_layer_start: 3
  - block_attention_layer_end: 19
  - block_mlp_layer_start: 3
  - block_mlp_layer_end: 19
  - iterative_steps: 1
  - grouping_strategy: sum
  - global_pruning: False
  - taylor: param_first
  - num_examples: 10
  - device: cuda
  - test_before_train: True
  - eval_device: cuda
  - test_after_train: True
  - seed: 42
  - save_model: True
  - torch_version: 2.4