You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

qwen_1.5_odia_7b (Pre-trained)

Qwen_1.5_Odia_7B is a pre-trained Odia large language model with 7 billion parameters, and it is based on Qwen 1.5-7B. The model is pre-trained on the Culturex-Odia dataset, a filtered version of the original CulturaX dataset for Odia text. The training dataset contains 49 million tokens. The CulturaX-Odia dataset is sourced from mc4 and four distinct OSCAR corpora.

For more details about the model, data, training procedure, and evaluations, go through the blog post.

Model Description

  • Model type: A 7B pre-trained decoder-only model
  • Primary Language(s): Odia and English
  • License: Apache-2.0 (Commercial)

NOTE

This is not an instruction-tuned model, so it may not be able to follow human instructions without using one/few-shot learning or instruction fine-tuning. The model has no moderation mechanisms and may generate harmful or inappropriate responses. It is recommended to first fine-tune it on the task(s) you are interested in.

Citation Information

If you find this model useful, please consider giving 👏 and citing:

@misc{Qwen1.5_odia_7b,
  author = {Sambit Sekhar and Shantipriya Parida and Debasish Dhal},
  title = {Introducing OdiaGenAI's Qwen-Based Pre-trained LLM for Odia Language},
  year = {2023},
  publisher = {Hugging Face},
  journal = {Hugging Face repository},
  howpublished = {\url{https://huggingface.co/OdiaGenAI}},
}

Contributions

  • Sambit Sekhar
  • Shantipriya Parida
  • Debasish Dhal
Downloads last month
0
Inference API
Unable to determine this model’s pipeline type. Check the docs .

Model tree for OdiaGenAI-LLM/qwen_1.5_odia_7b

Base model

Qwen/Qwen1.5-7B
Adapter
(6181)
this model
Adapters
1 model