anahita-b commited on
Commit
aa27c77
1 Parent(s): 8884f5d

Add vision_config, cleanup

Browse files
Files changed (1) hide show
  1. config.json +14 -9
config.json CHANGED
@@ -14,17 +14,8 @@
14
  "mlp_ratio":4,
15
  "num_attention_heads":12,
16
  "num_hidden_layers":6,
17
- "resolution_before":224,
18
  "stop_gradient":false,
19
  "tie_word_embeddings":false,
20
- "vit_embed_dim":512,
21
- "vit_num_hidden_layers":12,
22
- "vit_layernorm_init_from_vit":false,
23
- "vit_layernorm_shared":true,
24
- "vit_patch_size":16,
25
- "vit_remove_last":false,
26
- "vit_intermediate_size":512,
27
- "vit_hidden_size":768,
28
  "vocab_size":50265,
29
  "text_config_dict": null,
30
  "text_config":{
@@ -47,5 +38,19 @@
47
  "eos_token_id": 2,
48
  "position_embedding_type": "absolute",
49
  "use_cache": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
50
  }
51
  }
 
14
  "mlp_ratio":4,
15
  "num_attention_heads":12,
16
  "num_hidden_layers":6,
 
17
  "stop_gradient":false,
18
  "tie_word_embeddings":false,
 
 
 
 
 
 
 
 
19
  "vocab_size":50265,
20
  "text_config_dict": null,
21
  "text_config":{
 
38
  "eos_token_id": 2,
39
  "position_embedding_type": "absolute",
40
  "use_cache": true
41
+ },
42
+ "vision_config_dict": null,
43
+ "vision_config":{
44
+ "architectures": ["BridgeTowerVisionModel"],
45
+ "embed_dim": 512,
46
+ "input_resolution": 224,
47
+ "width": 768,
48
+ "layers": 12,
49
+ "patch_size": 16,
50
+ "transformer_width": 512,
51
+ "output_resolution": 288,
52
+ "stop_gradient": false,
53
+ "vit_layernorm_shared": true,
54
+ "vit_remove_last": false,
55
  }
56
  }