ArmelR commited on
Commit
c16bd6d
1 Parent(s): 462ebd9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +24 -1
README.md CHANGED
@@ -1,4 +1,27 @@
1
- - Pretraining
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  Gradio - ready 50 steps
3
  - Fine-tuning
4
  Oasst Guanaco 100 steps
 
1
+ ---
2
+ task_categories:
3
+ - text-generation
4
+ ---
5
+ # Description
6
+ This language model is the version 0.0 of a Gradio Coding Assistant. It is an instruction fine-tuned version of [StarCoder](https://huggingface.co/bigcode/starcoder) that is
7
+ design to provide assistance to developers who are using [gradio](https://www.gradio.app).
8
+
9
+ # Dataset
10
+ The dataset is multi-source. Its content comes from the following sources
11
+ - The stack
12
+ More precisely, we looked into [the-stack-dedup](https://huggingface.co/datasets/bigcode/the-stack-dedup) which contain codes permissive licenses. We shortlisted the files whose
13
+ content incorporated the keyword `gradio`.
14
+ - GitHub Issues
15
+ We scrapped all the issues of the official repository [the-gradio-app/gradio](https://github.com/gradio-app/gradio) and added them to our training dataset.
16
+ - Spaces on Hugging Face Hub
17
+ We used the [huggingface_hub api](https://huggingface.co/docs/huggingface_hub/package_reference/hf_api) to scrape the data from the spaces which are designed with gradio. We kept track of those
18
+ with permissive licenses, namely MIT and Apache 2.0. This set of code was further deduplicated.
19
+
20
+ # Training setting and hyperparameters
21
+ For our fine-tuning, we decided to follow a 2-step strategy.
22
+ - Pretraining (Fine-tuning) with next token prediction on the previously built gradio dataset (this step should familiarize the model with the gradio syntax.)
23
+ - Instruction fine-tuning on an instruction fine-tuning (this step should make the model conversational)
24
+ ## Pretraining
25
  Gradio - ready 50 steps
26
  - Fine-tuning
27
  Oasst Guanaco 100 steps