andrewhinh commited on
Commit
5b363a4
·
verified ·
1 Parent(s): cd4b00f

Upload Qwen2VLForConditionalGeneration

Browse files
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "andrewhinh/qwen2-vl-7b-instruct-full-sft",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
@@ -19,7 +19,7 @@
19
  "r": 8,
20
  "rank_pattern": {},
21
  "revision": null,
22
- "target_modules": "^(?!.*visual).*(?:up_proj|o_proj|v_proj|q_proj|k_proj|down_proj|gate_proj).*",
23
  "task_type": "CAUSAL_LM",
24
  "use_dora": false,
25
  "use_rslora": false
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "andrewhinh/qwen2-vl-7b-instruct-lora-sft-merged",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
19
  "r": 8,
20
  "rank_pattern": {},
21
  "revision": null,
22
+ "target_modules": "^(?!.*visual.patch_embed|visual.blocks).*(?:q_proj|k_proj|down_proj|gate_proj|v_proj|o_proj|up_proj).*",
23
  "task_type": "CAUSAL_LM",
24
  "use_dora": false,
25
  "use_rslora": false
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ea9065b7ba5b5a916e8395b0ef6197f45665e4a04220317801246ba17abdcb70
3
- size 80792096
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ce7c858abfdc39ed6d2ee9b516725a5cb008d869661661321673bff7f51d634
3
+ size 40422208
generation_config.json CHANGED
@@ -1,4 +1,5 @@
1
  {
 
2
  "bos_token_id": 151643,
3
  "do_sample": true,
4
  "eos_token_id": [
@@ -9,5 +10,5 @@
9
  "temperature": 0.01,
10
  "top_k": 1,
11
  "top_p": 0.001,
12
- "transformers_version": "4.46.1"
13
  }
 
1
  {
2
+ "attn_implementation": "flash_attention_2",
3
  "bos_token_id": 151643,
4
  "do_sample": true,
5
  "eos_token_id": [
 
10
  "temperature": 0.01,
11
  "top_k": 1,
12
  "top_p": 0.001,
13
+ "transformers_version": "4.45.2"
14
  }