Alpaca-LoRA is an open-source project that reproduces results from Stanford Alpaca using Low-Rank Adaptation (LoRA) techniques. It provides an Instruct model of similar quality to text-davinci-003. Alpaca-LoRA uses the resource-efficient low-rank adaptation (LoRA) method, also widely used in Stable Diffusion, with Meta’s LLaMA to achieve results comparable to Alpaca Alpaca formula is open source, but may not be used commercially. However, the LLaMA model used for Alpaca is not released for commercial use, and the OpenAI GPT-3.5 terms of use prohibit using the model to develop AI models that compete with OpenAI. Stanford has therefore not yet released the model, only the training data and the code to generate the data and fine-tune the model.
Github link:
Inference: click red bar Space
The labeled dataset I used to fine-tune and train the Alpaca model can be found at: https://www.kaggle.com/datasets/praveengovi/emotions-dataset-for-nlp?select=train.txt
To create Space in HuggingFace: https://huggingface.co/new-space ( Select for CPU Upgrade or above)
https://huggingface.co/spaces/RinInori/alpaca_finetune_6_sentiments
Upload app.py and requirements.txt to https://huggingface.co/spaces/RinInori/alpaca_finetune_6_sentiments/tree/main
Alpaca ref: https://github.com/tloen/alpaca-lora
- Downloads last month
- 6