VishnuPJ commited on
Commit
8b52af7
1 Parent(s): 241c33f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -7
README.md CHANGED
@@ -5,22 +5,22 @@ language:
5
  - ml
6
  pipeline_tag: text-generation
7
  ---
8
- # **MalayaLLM [മലയാളം/Malayalam]**
9
 
10
  <img src="https://cdn-uploads.huggingface.co/production/uploads/64e65800e44b2668a56f9731/bipVMulaNJ9um46ecYpR4.png" alt="Baby MalayaLLM" width="300" height="200">
11
 
12
- # **MalayaLLM_7B_Base**
13
 
14
  This is an attempt to construct a Language Model (LLM) focused on **generative AI for Malayalam language**. While several LLMs are proficient in supporting multiple languages, including Malayalam, enhancing their performance for specific tasks such as content generation and question answering specifically in Malayalam can be achieved through dedicated training on a Malayalam dataset. In pursuit of this, I've undertaken the **continuous pre-training of the LLAMA2 model using a comprehensive Malayalam dataset**.
15
 
16
  The model is currently in its early stages, and ongoing training and fine-tuning with a more comprehensive dataset are necessary to enhance its performance. I will consistently provide updated revisions to the model.
17
- # **Github Repo**:
18
  For comprehensive insights into model training, fine-tuning, and other advanced techniques, refer to the MalayaLLM GitHub repository at the following link:
19
  https://github.com/VishnuPJ/MalayaLLM
20
- # **Introducing the Developer**:
21
  Discover the mind behind this model and stay updated on their contributions to the field
22
  https://www.linkedin.com/in/vishnu-prasad-j/
23
- # **Model description**
24
  The MalayaLLM models have been improved and customized to incorporate a comprehensive Malayalam vocabulary comprising approximately 18,000 tokens, expanding upon the groundwork laid by the original LLaMA-2.
25
 
26
  - **Model type:** A 7B LLaMA2 pretrained model on Malayalam .
@@ -31,14 +31,14 @@ The MalayaLLM models have been improved and customized to incorporate a comprehe
31
  - **Training Precision:** `float16`
32
  - **Code:** [GitHub](https://github.com/VishnuPJ/MalayaLLM)
33
 
34
- ## **Available Models**
35
  | Model | Type | Data | Base Model | # Params | Download Links |
36
  |--------------------------|-----------------------------|-------------------|----------------------|------|------------------------------------------------------------------------|
37
  | Malayalam LLaMA 7B Base #v0.1 | Base model | 12GB | LLaMA 7B | 7B | [HF Hub](https://huggingface.co/VishnuPJ/MalayaLLM_7B_Base) |
38
  | Malayalam LLaMA 7B Instruct #v0.1| Instruction following model | 52k instructions | Malayalam LLaMA 7B Base | 7B | [HF Hub](https://huggingface.co/VishnuPJ/MalayaLLM_7B_Instruct_v0.1) |
39
  | Malayalam LLaMA 7B Instruct #v0.2| Instruction following model | 52k instructions | Malayalam LLaMA 7B Base | 7B | [HF Hub](https://huggingface.co/VishnuPJ/MalayaLLM_7B_Instruct_v0.2) |
40
 
41
- ### **Quantized Version of Available Models**
42
  | Model | Format | Bits | Download Links |
43
  |--------------------------|--------|----------------------|------------------------------------------------------------------------------|
44
  | Malayalam LLaMA 7B Instruct #v0.1 | GGUF | Q8_0 | [HF Hub](https://huggingface.co/VishnuPJ/MalayaLLM_7B_Instruct_v0.1_GGUF) |
 
5
  - ml
6
  pipeline_tag: text-generation
7
  ---
8
+ # MalayaLLM [മലയാളം/Malayalam]
9
 
10
  <img src="https://cdn-uploads.huggingface.co/production/uploads/64e65800e44b2668a56f9731/bipVMulaNJ9um46ecYpR4.png" alt="Baby MalayaLLM" width="300" height="200">
11
 
12
+ # MalayaLLM_7B_Base
13
 
14
  This is an attempt to construct a Language Model (LLM) focused on **generative AI for Malayalam language**. While several LLMs are proficient in supporting multiple languages, including Malayalam, enhancing their performance for specific tasks such as content generation and question answering specifically in Malayalam can be achieved through dedicated training on a Malayalam dataset. In pursuit of this, I've undertaken the **continuous pre-training of the LLAMA2 model using a comprehensive Malayalam dataset**.
15
 
16
  The model is currently in its early stages, and ongoing training and fine-tuning with a more comprehensive dataset are necessary to enhance its performance. I will consistently provide updated revisions to the model.
17
+ # Github Repo:
18
  For comprehensive insights into model training, fine-tuning, and other advanced techniques, refer to the MalayaLLM GitHub repository at the following link:
19
  https://github.com/VishnuPJ/MalayaLLM
20
+ # Introducing the Developer:
21
  Discover the mind behind this model and stay updated on their contributions to the field
22
  https://www.linkedin.com/in/vishnu-prasad-j/
23
+ # Model description
24
  The MalayaLLM models have been improved and customized to incorporate a comprehensive Malayalam vocabulary comprising approximately 18,000 tokens, expanding upon the groundwork laid by the original LLaMA-2.
25
 
26
  - **Model type:** A 7B LLaMA2 pretrained model on Malayalam .
 
31
  - **Training Precision:** `float16`
32
  - **Code:** [GitHub](https://github.com/VishnuPJ/MalayaLLM)
33
 
34
+ ## Available Models
35
  | Model | Type | Data | Base Model | # Params | Download Links |
36
  |--------------------------|-----------------------------|-------------------|----------------------|------|------------------------------------------------------------------------|
37
  | Malayalam LLaMA 7B Base #v0.1 | Base model | 12GB | LLaMA 7B | 7B | [HF Hub](https://huggingface.co/VishnuPJ/MalayaLLM_7B_Base) |
38
  | Malayalam LLaMA 7B Instruct #v0.1| Instruction following model | 52k instructions | Malayalam LLaMA 7B Base | 7B | [HF Hub](https://huggingface.co/VishnuPJ/MalayaLLM_7B_Instruct_v0.1) |
39
  | Malayalam LLaMA 7B Instruct #v0.2| Instruction following model | 52k instructions | Malayalam LLaMA 7B Base | 7B | [HF Hub](https://huggingface.co/VishnuPJ/MalayaLLM_7B_Instruct_v0.2) |
40
 
41
+ ### Quantized Version of Available Models
42
  | Model | Format | Bits | Download Links |
43
  |--------------------------|--------|----------------------|------------------------------------------------------------------------------|
44
  | Malayalam LLaMA 7B Instruct #v0.1 | GGUF | Q8_0 | [HF Hub](https://huggingface.co/VishnuPJ/MalayaLLM_7B_Instruct_v0.1_GGUF) |