Prasanna Dhungana commited on
Commit
fbd2075
1 Parent(s): 01bc9e4

update readme.md

Browse files
Files changed (1) hide show
  1. README.md +6 -8
README.md CHANGED
@@ -16,21 +16,22 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  # finetune_starcoder2_with_R_data
18
 
19
- This model is a fine-tuned version of [bigcode/starcoder2-3b](https://huggingface.co/bigcode/starcoder2-3b) on an unknown dataset.
20
 
21
  ## Model description
22
 
23
- More information needed
24
 
25
  ## Intended uses & limitations
26
 
27
- More information needed
28
 
29
  ## Training and evaluation data
30
 
31
- More information needed
32
 
33
  ## Training procedure
 
34
 
35
  ### Training hyperparameters
36
 
@@ -46,10 +47,7 @@ The following hyperparameters were used during training:
46
  - lr_scheduler_warmup_steps: 100
47
  - training_steps: 1000
48
  - mixed_precision_training: Native AMP
49
-
50
- ### Training results
51
-
52
-
53
 
54
  ### Framework versions
55
 
 
16
 
17
  # finetune_starcoder2_with_R_data
18
 
19
+ This model is a variant of the bigcode/starcoder2-3b architecture, adapted and fine-tuned specifically for generating R programming code.
20
 
21
  ## Model description
22
 
23
+ Model is
24
 
25
  ## Intended uses & limitations
26
 
27
+ Tailored for R programming tasks, this model is optimized for generating code snippets, functions, or scripts in the R language. Its limitations may include its applicability solely within the domain of R programming and potential constraints related to the size and diversity of the training data.
28
 
29
  ## Training and evaluation data
30
 
31
+ The model was trained and evaluated on a subset of the bigcode/Stack dataset containing R programming language data.
32
 
33
  ## Training procedure
34
+ Fine-tuning was performed using the PEFT (Parameter Efficient Fine Tuning) method over 1000 epochs on the R dataset. Additionally, the model was loaded with 4-bit quantization using the LoRA library to optimize memory usage and inference speed, enhancing its efficiency for generating R code.
35
 
36
  ### Training hyperparameters
37
 
 
47
  - lr_scheduler_warmup_steps: 100
48
  - training_steps: 1000
49
  - mixed_precision_training: Native AMP
50
+ =
 
 
 
51
 
52
  ### Framework versions
53