{ "_attn_implementation": "flash_attention_2", "attn_implementation": "flash_attention_2", "bos_token_id": 151643, "do_sample": true, "eos_token_id": [ 151645, 151643 ], "pad_token_id": 151643, "repetition_penalty": 1.05, "rope_theta": 1000000000.0, "temperature": 0.7, "top_k": 20, "top_p": 0.8, "transformers_version": "4.44.0" }