--- license: other language: - en library_name: transformers pipeline_tag: conversational --- ## Model Card for Model ID Finetuned depacoda-research/llamma-13b-hf on conversations ## Model Details ### Model Description The depacoda-research/llamma-13b-hf model was finetuned on conversations and question answering prompts **Developed by:** [More Information Needed] **Shared by:** [More Information Needed] **Model type:** Causal LM **Language(s) (NLP):** English, multilingual **License:** Research **Finetuned from model:** depacoda-research/llamma-13b-hf ## Model Sources [optional] **Repository:** [More Information Needed] **Paper:** [More Information Needed] **Demo:** [More Information Needed] ## Uses The model can be used for prompt answering ### Direct Use The model can be used for prompt answering ### Downstream Use Generating text and prompt answering ## Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ``` from transformers import LlamaTokenizer, LlamaForCausalLM from peft import PeftModel MODEL_NAME = "decapoda-research/llama-13b-hf" tokenizer = LlamaTokenizer.from_pretrained(MODEL_NAME, add_eos_token=True) tokenizer.pad_token_id = 0 model = LlamaForCausalLM.from_pretrained(MODEL_NAME, load_in_8bit=True, device_map="auto") model = PeftModel.from_pretrained(model, "Sandiago21/public-ai-model") ``` ## Training Details ### Training Data The decapoda-research/llama-13b-hf was finetuned on conversations and question answering data ### Training Procedure The decapoda-research/llama-13b-hf model was further trained and finetuned on question answering and prompts data ## Model Architecture and Objective The model is based on decapoda-research/llama-13b-hf model and finetuned adapters on top of the main model on conversations and question answering data.