echo840 commited on
Commit
157a078
1 Parent(s): a85de91

change config.json

Browse files
Files changed (4) hide show
  1. config.json +4 -9
  2. modeling_monkey.py +1 -1
  3. trainer_state.json +0 -0
  4. training_args.bin +0 -3
config.json CHANGED
@@ -1,12 +1,11 @@
1
  {
2
- "_name_or_path": "/home/kas/data/llm/qwen/",
3
  "architectures": [
4
- "QWenLMHeadModel"
5
  ],
6
  "attn_dropout_prob": 0.0,
7
  "auto_map": {
8
  "AutoConfig": "configuration_qwen.QWenConfig",
9
- "AutoModelForCausalLM": "modeling_qwen.QWenLMHeadModel"
10
  },
11
  "bf16": true,
12
  "emb_dropout_prob": 0.0,
@@ -18,7 +17,7 @@
18
  "kv_channels": 128,
19
  "layer_norm_epsilon": 1e-06,
20
  "max_position_embeddings": 8192,
21
- "model_type": "qwen",
22
  "no_bias": true,
23
  "num_attention_heads": 32,
24
  "num_hidden_layers": 32,
@@ -36,19 +35,15 @@
36
  "use_flash_attn": false,
37
  "use_logn_attn": true,
38
  "visual": {
39
- "add_window": false,
40
  "heads": 16,
41
  "image_size": 896,
42
  "image_start_id": 151857,
43
  "layers": 48,
44
- "lora_repeat_num": 4,
45
  "mlp_ratio": 4.9231,
46
- "n_queries": 256,
47
  "output_dim": 4096,
48
  "patch_size": 14,
49
- "use_global": true,
50
  "width": 1664,
51
- "window_all": false
52
  },
53
  "vocab_size": 151936
54
  }
 
1
  {
 
2
  "architectures": [
3
+ "MonkeyLMHeadModel"
4
  ],
5
  "attn_dropout_prob": 0.0,
6
  "auto_map": {
7
  "AutoConfig": "configuration_qwen.QWenConfig",
8
+ "AutoModelForCausalLM": "modeling_monkey.MonkeyLMHeadModel"
9
  },
10
  "bf16": true,
11
  "emb_dropout_prob": 0.0,
 
17
  "kv_channels": 128,
18
  "layer_norm_epsilon": 1e-06,
19
  "max_position_embeddings": 8192,
20
+ "model_type": "monkey",
21
  "no_bias": true,
22
  "num_attention_heads": 32,
23
  "num_hidden_layers": 32,
 
35
  "use_flash_attn": false,
36
  "use_logn_attn": true,
37
  "visual": {
 
38
  "heads": 16,
39
  "image_size": 896,
40
  "image_start_id": 151857,
41
  "layers": 48,
 
42
  "mlp_ratio": 4.9231,
 
43
  "output_dim": 4096,
44
  "patch_size": 14,
 
45
  "width": 1664,
46
+ "lora_repeat_num":4
47
  },
48
  "vocab_size": 151936
49
  }
modeling_monkey.py CHANGED
@@ -26,7 +26,7 @@ try:
26
  except ImportError:
27
  rearrange = None
28
  from torch import nn
29
- from monkey_model.modeling_qwen import QWenModel,QWenPreTrainedModel,QWenLMHeadModel
30
  SUPPORT_CUDA = torch.cuda.is_available()
31
  SUPPORT_BF16 = SUPPORT_CUDA and torch.cuda.is_bf16_supported()
32
  SUPPORT_FP16 = SUPPORT_CUDA and torch.cuda.get_device_capability(0)[0] >= 7
 
26
  except ImportError:
27
  rearrange = None
28
  from torch import nn
29
+ from .modeling_qwen import QWenModel,QWenPreTrainedModel,QWenLMHeadModel
30
  SUPPORT_CUDA = torch.cuda.is_available()
31
  SUPPORT_BF16 = SUPPORT_CUDA and torch.cuda.is_bf16_supported()
32
  SUPPORT_FP16 = SUPPORT_CUDA and torch.cuda.get_device_capability(0)[0] >= 7
trainer_state.json DELETED
The diff for this file is too large to render. See raw diff
 
training_args.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:60d8ad6ecdf0d6dc385ea49113e3c6c6ace92ef3a6f3c9707e5b528a592734d2
3
- size 5563