File size: 2,297 Bytes
672bf87 aa7c09c 672bf87 aa7c09c 672bf87 aa7c09c 672bf87 aa7c09c 672bf87 aa7c09c 672bf87 aa7c09c 672bf87 aa7c09c 672bf87 aa7c09c 672bf87 aa7c09c 672bf87 aa7c09c 672bf87 ecf8113 aa7c09c |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 |
---
library_name: transformers
pipeline_tag: text-generation
tags:
- BGPT
- meta
- pytorch
- llama
- llama-3
---
### Model Description
This model is a finetuned version of Llama3.2-3B-Instruct specifically designed for generating multilingual outputs across multiple Indic languages. The model has been trained on a diverse and curated dataset comprising Hindi, Punjabi, Marathi, Malayalam, Oriya, Kannada, Gujarati, Bengali, Urdu, Tamil, and Telugu. It is optimized to handle natural language tasks such as translation, summarization, and conversational generation across these languages effectively.
- **Developed by:** [More Information Needed]
- **Model type:** Finetuned LLaMA (Language Model for Multilingual Text Generation)
- **Language(s) (NLP):** Hindi, Punjabi, Marathi, Malayalam, Oriya, Kannada, Gujarati, Bengali, Urdu, Tamil, Telugu
- **Finetuned from model:** Llama3.2-3B-Instruct
## How to Get Started with the Model
Make sure to update your transformers installation via `pip install --upgrade transformers`.
Use the code below to get started with the model.
```python
import torch
from transformers import pipeline
model_id = "Onkarn/ML-Test-v01"
pipe = pipeline(
"text-generation",
model=model_id,
torch_dtype=torch.bfloat16,
device_map="auto",
)
messages = [
{"role": "system", "content": "You are a helpful assistant who responds in hindi"},
{"role": "user", "content": "कर्नाटक की राजधानी क्या है?"},
]
outputs = pipe(
messages,
max_new_tokens=256,
)
print(outputs[0]["generated_text"][-1])
```
## Training Details
### Training Data
The training dataset included a diverse collection of text sources in:
- Hindi, Punjabi, Marathi, Malayalam, Oriya, Kannada, Gujarati, Bengali, Urdu, Tamil, and Telugu.
### Training Parameters
- **Optimization Technique**: LoRA (Low-Rank Adaptation)
- **Epochs**: 3.0
- **Batch Size**: 2.0 (per device train batch size)
- **Learning Rate**: 5e-05
## Environmental Impact
- **Hardware Type:** T4
- **Hours used:** 29 hours
- **Cloud Provider:** Google Cloud Platform
- **Compute Region:** asia-southeast1
- **Carbon Emitted:** Total emissions are estimated to be 0.85 kgCO$_2$eq of which 100 percents were directly offset by the cloud provider.
|