gajesh's picture
Re-upload model with updated configuration
25833d8 verified
---
library_name: transformers
tags:
- causal-lm
- llama
- fine-tuned
- text-generation
---
# Fine-Tuned LLaMA 3.2 1B Model
This model is a fine-tuned version of [meta-llama/Llama-3.2-1B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct) on custom data. It has been trained to generate coherent and contextually relevant responses based on the input prompt.
## Model Description
- **Model Type**: LLaMA (Large Language Model for AI Assistants)
- **Architecture**: Causal Language Model (LlamaForCausalLM)
- **Base Model**: `meta-llama/Llama-3.2-1B-Instruct`
- **Fine-Tuning**: Fine-tuned on domain-specific data to enhance performance on targeted tasks.
- **Intended Use**: Suitable for various NLP tasks such as text generation, question answering, and code analysis.
## Training Data
The model was fine-tuned on a dataset containing domain-specific examples designed to improve its understanding and generation capabilities within specific contexts. The training data included:
- **Code Samples**: Various programming languages for code analysis and explanation.
- **Technical Documentation**: To improve technical writing and explanation capabilities.
## Training Details
- **Fine-Tuning Epochs**: 5
- **Batch Size**: 1 (with gradient accumulation)
- **Learning Rate**: 1e-5
- **Hardware**: Fine-tuned using an NVIDIA A10G on an `g5.16xlarge` instance.
- **Optimizer**: AdamW with weight decay
### Model Configuration
- **Hidden Size**: 2048
- **Number of Layers**: 16
- **Number of Attention Heads**: 32
- **Intermediate Size**: 8192
## Usage
To use this model, you can either download it and run locally using the `transformers` library or use the Hugging Face Inference API.
### Using with `transformers`
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
# Load the fine-tuned model and tokenizer
tokenizer = AutoTokenizer.from_pretrained("username/your-fine-tuned-llama")
model = AutoModelForCausalLM.from_pretrained("username/your-fine-tuned-llama")
# Generate text
prompt = "What does EigenLayer do exactly?"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_length=150, num_beams=4, temperature=0.5, do_sample=True)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
```
### Using with the Hugging Face Inference API
You can also use the model via the Hugging Face API endpoint:
```python
import requests
API_URL = "https://api-inference.huggingface.co/models/username/your-fine-tuned-llama"
headers = {"Authorization": "Bearer YOUR_HUGGING_FACE_API_TOKEN"}
def query(prompt):
response = requests.post(API_URL, headers=headers, json={"inputs": prompt})
return response.json()
print(query("Explain how EigenLayer functions."))
```
## Limitations
- The model may generate incorrect or biased information. Users should verify the outputs for critical applications.
- Due to fine-tuning, there might be domain-specific biases in the generation.
## Ethical Considerations
Please ensure that the outputs of this model are used responsibly. The model may generate unintended or harmful content, so it should be used with caution in sensitive applications.
## Acknowledgements
This model was fine-tuned based on [meta-llama/Llama-3.2-1B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct). Special thanks to the open-source community and contributors to the `transformers` library.