MingComplex commited on
Commit
06127fa
·
1 Parent(s): 37e5c41

update config

Browse files
Files changed (1) hide show
  1. config.json +24 -18
config.json CHANGED
@@ -5,9 +5,13 @@
5
  "attention_dropout": 0.0,
6
  "bos_token_id": 151643,
7
  "eos_token_id": 151645,
 
 
 
 
 
8
  "hidden_act": "silu",
9
  "hidden_size": 3584,
10
- "image_token_id": 151655,
11
  "initializer_range": 0.02,
12
  "intermediate_size": 18944,
13
  "max_position_embeddings": 32768,
@@ -17,30 +21,32 @@
17
  "num_hidden_layers": 28,
18
  "num_key_value_heads": 4,
19
  "rms_norm_eps": 1e-06,
20
- "rope_scaling": {
21
- "mrope_section": [
22
- 16,
23
- 24,
24
- 24
25
- ],
26
- "rope_type": "default",
27
- "type": "default"
28
- },
29
  "rope_theta": 1000000.0,
30
  "sliding_window": 32768,
31
  "tie_word_embeddings": false,
32
  "torch_dtype": "bfloat16",
33
- "transformers_version": "4.47.1",
34
  "use_cache": true,
35
  "use_sliding_window": false,
36
- "video_token_id": 151656,
37
  "vision_config": {
 
 
 
 
38
  "in_chans": 3,
39
- "model_type": "qwen2_vl",
40
- "spatial_patch_size": 14
 
 
 
 
 
 
 
 
 
 
 
41
  },
42
- "vision_end_token_id": 151653,
43
- "vision_start_token_id": 151652,
44
- "vision_token_id": 151654,
45
  "vocab_size": 152064
46
- }
 
5
  "attention_dropout": 0.0,
6
  "bos_token_id": 151643,
7
  "eos_token_id": 151645,
8
+ "vision_start_token_id": 151652,
9
+ "vision_end_token_id": 151653,
10
+ "vision_token_id": 151654,
11
+ "image_token_id": 151655,
12
+ "video_token_id": 151656,
13
  "hidden_act": "silu",
14
  "hidden_size": 3584,
 
15
  "initializer_range": 0.02,
16
  "intermediate_size": 18944,
17
  "max_position_embeddings": 32768,
 
21
  "num_hidden_layers": 28,
22
  "num_key_value_heads": 4,
23
  "rms_norm_eps": 1e-06,
 
 
 
 
 
 
 
 
 
24
  "rope_theta": 1000000.0,
25
  "sliding_window": 32768,
26
  "tie_word_embeddings": false,
27
  "torch_dtype": "bfloat16",
28
+ "transformers_version": "4.41.2",
29
  "use_cache": true,
30
  "use_sliding_window": false,
 
31
  "vision_config": {
32
+ "depth": 32,
33
+ "embed_dim": 1280,
34
+ "mlp_ratio": 4,
35
+ "num_heads": 16,
36
  "in_chans": 3,
37
+ "hidden_size": 3584,
38
+ "patch_size": 14,
39
+ "spatial_merge_size": 2,
40
+ "spatial_patch_size": 14,
41
+ "temporal_patch_size": 2
42
+ },
43
+ "rope_scaling": {
44
+ "type": "mrope",
45
+ "mrope_section": [
46
+ 16,
47
+ 24,
48
+ 24
49
+ ]
50
  },
 
 
 
51
  "vocab_size": 152064
52
+ }