{ "attn_l2_norm": true, "depth": 30, "rope": true, "rope_size": 128, "rope_theta": 10000, "shared_aln": false, "text_encoder_2_path": null, "text_encoder_path": null, "use_swiglu_ffn": true, "vae_kwargs": { "ch": 160, "share_quant_resi": 4, "test_mode": true, "v_patch_nums": [ 1, 2, 3, 4, 6, 9, 13, 18, 24, 32 ], "vocab_size": 4096, "z_channels": 32 } }