chathuru commited on
Commit
e07014b
1 Parent(s): d1936d0

CuATR-distilbert-LoRA

Browse files
README.md CHANGED
@@ -18,9 +18,9 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 0.7010
22
- - Accuracy: 0.4348
23
- - F1: 0.0
24
 
25
  ## Model description
26
 
@@ -52,11 +52,11 @@ The following hyperparameters were used during training:
52
 
53
  ### Training results
54
 
55
- | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
56
- |:-------------:|:-----:|:----:|:---------------:|:--------:|:---:|
57
- | 0.6952 | 0.67 | 1 | 0.7016 | 0.4348 | 0.0 |
58
- | 0.6907 | 2.0 | 3 | 0.7012 | 0.4348 | 0.0 |
59
- | 0.719 | 2.67 | 4 | 0.7010 | 0.4348 | 0.0 |
60
 
61
 
62
  ### Framework versions
 
18
 
19
  This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.6890
22
+ - Accuracy: 0.5652
23
+ - F1: 0.7222
24
 
25
  ## Model description
26
 
 
52
 
53
  ### Training results
54
 
55
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
56
+ |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
57
+ | 0.7122 | 0.67 | 1 | 0.6891 | 0.5652 | 0.7222 |
58
+ | 0.7112 | 2.0 | 3 | 0.6890 | 0.5652 | 0.7222 |
59
+ | 0.6955 | 2.67 | 4 | 0.6890 | 0.5652 | 0.7222 |
60
 
61
 
62
  ### Framework versions
adapter_config.json CHANGED
@@ -2,22 +2,22 @@
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
  "base_model_name_or_path": "distilbert-base-uncased",
5
- "bias": "all",
6
  "fan_in_fan_out": false,
7
  "inference_mode": false,
8
  "init_lora_weights": true,
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
11
- "lora_alpha": 32,
12
- "lora_dropout": 0.05,
13
  "modules_to_save": null,
14
  "peft_type": "LORA",
15
  "r": 16,
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
- "k_lin",
20
  "v_lin",
 
21
  "q_lin"
22
  ],
23
  "task_type": "TOKEN_CLS"
 
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
  "base_model_name_or_path": "distilbert-base-uncased",
5
+ "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": false,
8
  "init_lora_weights": true,
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
11
+ "lora_alpha": 16,
12
+ "lora_dropout": 0.1,
13
  "modules_to_save": null,
14
  "peft_type": "LORA",
15
  "r": 16,
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
 
19
  "v_lin",
20
+ "k_lin",
21
  "q_lin"
22
  ],
23
  "task_type": "TOKEN_CLS"
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2bcd6bb8a5da04a0abf779344bb4cddd008046bb80a11ee65f83cb468e12a606
3
- size 1990240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b42e653456701e10c61c62e1d9ee2a365ae65a2669fde89dfc4d8dd35ae27cc
3
+ size 1774720
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b5c0d322e926af0aab09913b96d93544750b6d0852ba0b4f735925101ec3d6d2
3
  size 4600
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c85194ab166096f43ad5356cbba09fbe00e9f9707ef5784691ae1932410d2900
3
  size 4600