--- license: llama2 base_model: Photolens/llama-2-7b-langchain-chat tags: - generated_from_trainer model-index: - name: assistant-llama2-7b-chat results: [] datasets: - wasertech/OneOS --- # assistant-llama2-7b-chat This model is a fine-tuned version of [Photolens/llama-2-7b-langchain-chat](https://huggingface.co/Photolens/llama-2-7b-langchain-chat) on the OneOS dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1.41e-05 - train_batch_size: 2 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 4 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 ### Training results Model is able to mostly answer to complexe queries given its output is parsed propely (accounting for hallucinations). For more information look at the [Prompt Example Section](#prompt-example) ### Framework versions - Transformers 4.33.2 - Pytorch 2.0.1+cu117 - Datasets 2.14.5 - Tokenizers 0.13.3 # Training Report https://wandb.ai/wasertech/huggingface/reports/Llama2-langchain-chat-fine-tune--Vmlldzo1NTQ2OTAy # Prompt Example The above link leads to a conversation log between [Assistant using this model](https://gitlab.com/waser-technologies/technologies/assistant/-/merge_requests/3/diffs?commit_id=e17a80de631686f7fe94c5ca136214be73bafd8a) and me. https://gist.github.com/wasertech/342cd167ba78060336b3328e9eea0eca