Ar4l commited on
Commit
0baf521
1 Parent(s): 98b97ad

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -13,7 +13,7 @@ Model for the paper [**"A Transformer-Based Approach for Smart Invocation of Aut
13
  This model is fine-tuned on a code-completion dataset collected from the open-source [Code4Me](https://github.com/code4me-me/code4me) plugin. The training objective is to have a small, lightweight transformer model to filter out unnecessary and unhelpful code completions. To this end, we leverage the in-IDE telemetry data, and integrate it with the textual code data in the transformer's attention module.
14
 
15
  - **Developed by:** [AISE Lab](https://www.linkedin.com/company/aise-tudelft/) @ [SERG](https://se.ewi.tudelft.nl/), Delft University of Technology
16
- - **Model type:** [JonBERTa](https://github.com/Ar4l/curating-code-completions/blob/main/modeling_jonberta.py)
17
  - **Language:** Code
18
  - **Finetuned from model:** [`CodeBERTa-small-v1`](https://huggingface.co/huggingface/CodeBERTa-small-v1).
19
 
@@ -51,8 +51,8 @@ To cite, please use
51
  This model was trained with the following hyperparameters, everything else being `TrainingArguments`' default. The dataset was prepared identically across all models as detailed in the paper.
52
 
53
  ```python
54
- num_train_epochs : int = 3
55
- learning_rate : float = 2e-5
56
- batch_size : int = 16
57
  ```
58
 
 
13
  This model is fine-tuned on a code-completion dataset collected from the open-source [Code4Me](https://github.com/code4me-me/code4me) plugin. The training objective is to have a small, lightweight transformer model to filter out unnecessary and unhelpful code completions. To this end, we leverage the in-IDE telemetry data, and integrate it with the textual code data in the transformer's attention module.
14
 
15
  - **Developed by:** [AISE Lab](https://www.linkedin.com/company/aise-tudelft/) @ [SERG](https://se.ewi.tudelft.nl/), Delft University of Technology
16
+ - **Model type:** [RoBERTa](https://huggingface.co/FacebookAI/roberta-base)
17
  - **Language:** Code
18
  - **Finetuned from model:** [`CodeBERTa-small-v1`](https://huggingface.co/huggingface/CodeBERTa-small-v1).
19
 
 
51
  This model was trained with the following hyperparameters, everything else being `TrainingArguments`' default. The dataset was prepared identically across all models as detailed in the paper.
52
 
53
  ```python
54
+ num_train_epochs : int = 6
55
+ learning_rate : float = search([2e-5, 1e-5, 5e-5])
56
+ batch_size : int = 16
57
  ```
58