Upload README.md with huggingface_hub
Browse files
README.md
CHANGED
@@ -13,7 +13,7 @@ Model for the paper [**"A Transformer-Based Approach for Smart Invocation of Aut
|
|
13 |
This model is fine-tuned on a code-completion dataset collected from the open-source [Code4Me](https://github.com/code4me-me/code4me) plugin. The training objective is to have a small, lightweight transformer model to filter out unnecessary and unhelpful code completions. To this end, we leverage the in-IDE telemetry data, and integrate it with the textual code data in the transformer's attention module.
|
14 |
|
15 |
- **Developed by:** [AISE Lab](https://www.linkedin.com/company/aise-tudelft/) @ [SERG](https://se.ewi.tudelft.nl/), Delft University of Technology
|
16 |
-
- **Model type:** [
|
17 |
- **Language:** Code
|
18 |
- **Finetuned from model:** [`CodeBERTa-small-v1`](https://huggingface.co/huggingface/CodeBERTa-small-v1).
|
19 |
|
@@ -51,8 +51,8 @@ To cite, please use
|
|
51 |
This model was trained with the following hyperparameters, everything else being `TrainingArguments`' default. The dataset was prepared identically across all models as detailed in the paper.
|
52 |
|
53 |
```python
|
54 |
-
num_train_epochs : int
|
55 |
-
learning_rate : float
|
56 |
-
batch_size : int
|
57 |
```
|
58 |
|
|
|
13 |
This model is fine-tuned on a code-completion dataset collected from the open-source [Code4Me](https://github.com/code4me-me/code4me) plugin. The training objective is to have a small, lightweight transformer model to filter out unnecessary and unhelpful code completions. To this end, we leverage the in-IDE telemetry data, and integrate it with the textual code data in the transformer's attention module.
|
14 |
|
15 |
- **Developed by:** [AISE Lab](https://www.linkedin.com/company/aise-tudelft/) @ [SERG](https://se.ewi.tudelft.nl/), Delft University of Technology
|
16 |
+
- **Model type:** [RoBERTa](https://huggingface.co/FacebookAI/roberta-base)
|
17 |
- **Language:** Code
|
18 |
- **Finetuned from model:** [`CodeBERTa-small-v1`](https://huggingface.co/huggingface/CodeBERTa-small-v1).
|
19 |
|
|
|
51 |
This model was trained with the following hyperparameters, everything else being `TrainingArguments`' default. The dataset was prepared identically across all models as detailed in the paper.
|
52 |
|
53 |
```python
|
54 |
+
num_train_epochs : int = 6
|
55 |
+
learning_rate : float = search([2e-5, 1e-5, 5e-5])
|
56 |
+
batch_size : int = 16
|
57 |
```
|
58 |
|