ajibawa-2023
commited on
Commit
•
e2595df
1
Parent(s):
5911e77
Update README.md
Browse files
README.md
CHANGED
@@ -13,7 +13,7 @@ tags:
|
|
13 |
|
14 |
Large Language Models (LLMs) are good with code generations. Sometimes they do make mistakes in code generation. How about if they can give detailed explanation along with the code.
|
15 |
This is what I have tried over here. The base Llama-2 model was used for training purpose. It is trained on around **290000** set of codes. Each set having 2 conversations.
|
16 |
-
Along with Python, Java, JavaScript, GO, C++, Rust, Ruby, Sql, MySql, R, Julia, Haskell, etc. code with detailed explanation is used for training purpose. It is built upon using my existing
|
17 |
This conversation is in Vicuna/ShareGPT format. Each set, along with code, has detailed explanation.
|
18 |
|
19 |
I have released the new data [Code-290k-ShareGPT](https://huggingface.co/datasets/ajibawa-2023/Code-290k-ShareGPT) on which this Model is trained.
|
|
|
13 |
|
14 |
Large Language Models (LLMs) are good with code generations. Sometimes they do make mistakes in code generation. How about if they can give detailed explanation along with the code.
|
15 |
This is what I have tried over here. The base Llama-2 model was used for training purpose. It is trained on around **290000** set of codes. Each set having 2 conversations.
|
16 |
+
Along with Python, Java, JavaScript, GO, C++, Rust, Ruby, Sql, MySql, R, Julia, Haskell, etc. code with detailed explanation is used for training purpose. It is built upon using my existing Datasets [Python-Code-23k-ShareGPT](https://huggingface.co/datasets/ajibawa-2023/Python-Code-23k-ShareGPT) and [Code-74k-ShareGPT](https://huggingface.co/datasets/ajibawa-2023/Code-74k-ShareGPT) .
|
17 |
This conversation is in Vicuna/ShareGPT format. Each set, along with code, has detailed explanation.
|
18 |
|
19 |
I have released the new data [Code-290k-ShareGPT](https://huggingface.co/datasets/ajibawa-2023/Code-290k-ShareGPT) on which this Model is trained.
|