{ "_attn_implementation_autoset": true, "ae_mode": "token", "attn_implementation": null, "auto_map": { "AutoConfig": "modelling_pisco.COCOMConfig", "AutoModel": "modelling_pisco.COCOM" }, "compr_base_model_name": "mistralai/Mistral-7B-Instruct-v0.2", "compr_every_n_layer": null, "compr_linear_type": "concat", "compr_mlp_hidden_dim": 8096, "compr_model_name": null, "compr_n_layers": null, "compr_rate": 16, "compr_rms_norm": false, "compr_use_mlp": true, "decoder_model_name": "mistralai/Mistral-7B-Instruct-v0.2", "device_map": null, "different_mem_tokens": true, "doc_max_length": 128, "generation_top_k": 1, "kbtc_training": false, "load_adapters": true, "lora": true, "lora_compressor": false, "lora_r": 16, "lora_r_compressor": 16, "max_new_tokens": 128, "model_type": "COCOM", "optimize_mem_tokens": true, "quantization": "no", "sep": true, "training_form": "both_separately", "transformers_version": "4.48.0" }