{ "auto_mapping":null, "base_model_name_or_path":"lmsys/vicuna-7b-v1.3", "bias":"none", "fan_in_fan_out":false, "inference_mode":true, "init_lora_weights":true, "layers_pattern":null, "layers_to_transform":null, "lora_alpha":32, "lora_dropout":0.05, "modules_to_save":null, "peft_type":"LORA", "r":8, "revision":null, "target_modules":[ "q_proj", "v_proj" ], "model_type": "llama", "task_type":"CAUSAL_LM" }