alenatz commited on
Commit
9c44356
1 Parent(s): 7ca55ab

alenatz/relation-extraction-biobert-biocause

Browse files
README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: dmis-lab/biobert-v1.1
3
  tags:
4
  - generated_from_trainer
5
  metrics:
@@ -17,15 +17,16 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  # relation-biobert-biocause
19
 
20
- This model is a fine-tuned version of [dmis-lab/biobert-v1.1](https://huggingface.co/dmis-lab/biobert-v1.1) on the None dataset.
21
  It achieves the following results on the evaluation set:
22
- - Loss: 0.0398
23
- - Precision: 0.3308
24
- - Recall: 0.3333
25
- - F1: 0.3321
26
- - Accuracy: 0.9925
27
- - Balanced Accuracy: 0.3333
28
- - Kappa: 0.0
 
29
 
30
  ## Model description
31
 
@@ -44,26 +45,30 @@ More information needed
44
  ### Training hyperparameters
45
 
46
  The following hyperparameters were used during training:
47
- - learning_rate: 2e-05
48
- - train_batch_size: 8
49
- - eval_batch_size: 8
50
  - seed: 42
51
- - gradient_accumulation_steps: 2
52
- - total_train_batch_size: 16
53
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
54
  - lr_scheduler_type: linear
55
  - num_epochs: 1
56
 
57
  ### Training results
58
 
59
- | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | Balanced Accuracy | Kappa |
60
- |:-------------:|:------:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|:-----------------:|:-----:|
61
- | 0.0983 | 0.5764 | 100 | 0.0462 | 0.3308 | 0.3333 | 0.3321 | 0.9925 | 0.3333 | 0.0 |
 
 
 
 
 
 
62
 
63
 
64
  ### Framework versions
65
 
66
- - Transformers 4.41.2
67
  - Pytorch 2.3.0+cu121
68
  - Datasets 2.20.0
69
  - Tokenizers 0.19.1
 
1
  ---
2
+ base_model: dmis-lab/biobert-base-cased-v1.2
3
  tags:
4
  - generated_from_trainer
5
  metrics:
 
17
 
18
  # relation-biobert-biocause
19
 
20
+ This model is a fine-tuned version of [dmis-lab/biobert-base-cased-v1.2](https://huggingface.co/dmis-lab/biobert-base-cased-v1.2) on an unknown dataset.
21
  It achieves the following results on the evaluation set:
22
+ - Loss: 0.2103
23
+ - Precision: 0.1164
24
+ - Recall: 0.625
25
+ - F1: 0.1963
26
+ - Accuracy: 0.9448
27
+ - Relation P: 0.1164
28
+ - Relation R: 0.625
29
+ - Relation F1: 0.1963
30
 
31
  ## Model description
32
 
 
45
  ### Training hyperparameters
46
 
47
  The following hyperparameters were used during training:
48
+ - learning_rate: 4e-05
49
+ - train_batch_size: 16
50
+ - eval_batch_size: 16
51
  - seed: 42
 
 
52
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
53
  - lr_scheduler_type: linear
54
  - num_epochs: 1
55
 
56
  ### Training results
57
 
58
+ | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | Relation P | Relation R | Relation F1 |
59
+ |:-------------:|:------:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|:----------:|:----------:|:-----------:|
60
+ | 0.6563 | 0.1282 | 20 | 0.2984 | 0.0211 | 0.2105 | 0.0384 | 0.8265 | 0.0211 | 0.2105 | 0.0384 |
61
+ | 0.6563 | 0.2564 | 40 | 0.2302 | 0.0763 | 0.4605 | 0.1308 | 0.9266 | 0.0763 | 0.4605 | 0.1308 |
62
+ | 0.6563 | 0.3846 | 60 | 0.4003 | 0.1406 | 0.5921 | 0.2273 | 0.9617 | 0.1406 | 0.5921 | 0.2273 |
63
+ | 0.6563 | 0.5128 | 80 | 0.2185 | 0.0554 | 0.4671 | 0.0990 | 0.8933 | 0.0554 | 0.4671 | 0.0990 |
64
+ | 0.6563 | 0.6410 | 100 | 0.2261 | 0.1345 | 0.7105 | 0.2262 | 0.9510 | 0.1345 | 0.7105 | 0.2262 |
65
+ | 0.6563 | 0.7692 | 120 | 0.2315 | 0.1259 | 0.6579 | 0.2114 | 0.9502 | 0.1259 | 0.6579 | 0.2114 |
66
+ | 0.6563 | 0.8974 | 140 | 0.2324 | 0.1417 | 0.6711 | 0.2339 | 0.9547 | 0.1417 | 0.6711 | 0.2339 |
67
 
68
 
69
  ### Framework versions
70
 
71
+ - Transformers 4.42.4
72
  - Pytorch 2.3.0+cu121
73
  - Datasets 2.20.0
74
  - Tokenizers 0.19.1
config.json CHANGED
@@ -1,24 +1,21 @@
1
  {
2
- "_name_or_path": "dmis-lab/biobert-v1.1",
3
  "architectures": [
4
- "BertForTokenClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
  "classifier_dropout": null,
8
- "gradient_checkpointing": false,
9
  "hidden_act": "gelu",
10
  "hidden_dropout_prob": 0.1,
11
  "hidden_size": 768,
12
  "id2label": {
13
  "0": "O",
14
- "1": "B-RELATION",
15
- "2": "I-RELATION"
16
  },
17
  "initializer_range": 0.02,
18
  "intermediate_size": 3072,
19
  "label2id": {
20
- "B-RELATION": 1,
21
- "I-RELATION": 2,
22
  "O": 0
23
  },
24
  "layer_norm_eps": 1e-12,
@@ -29,7 +26,7 @@
29
  "pad_token_id": 0,
30
  "position_embedding_type": "absolute",
31
  "torch_dtype": "float32",
32
- "transformers_version": "4.41.2",
33
  "type_vocab_size": 2,
34
  "use_cache": true,
35
  "vocab_size": 28996
 
1
  {
2
+ "_name_or_path": "dmis-lab/biobert-base-cased-v1.2",
3
  "architectures": [
4
+ "BertForUnbalancedTokenClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
  "classifier_dropout": null,
 
8
  "hidden_act": "gelu",
9
  "hidden_dropout_prob": 0.1,
10
  "hidden_size": 768,
11
  "id2label": {
12
  "0": "O",
13
+ "1": "I-REL"
 
14
  },
15
  "initializer_range": 0.02,
16
  "intermediate_size": 3072,
17
  "label2id": {
18
+ "I-REL": 1,
 
19
  "O": 0
20
  },
21
  "layer_norm_eps": 1e-12,
 
26
  "pad_token_id": 0,
27
  "position_embedding_type": "absolute",
28
  "torch_dtype": "float32",
29
+ "transformers_version": "4.42.4",
30
  "type_vocab_size": 2,
31
  "use_cache": true,
32
  "vocab_size": 28996
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7915356e386023376eb43a853a5fa069dbcc93a74b99a9536ab30a42379e0f78
3
- size 430911284
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:234c6a38c4fd4ed65f28c898d49ddf69a5a77cbb2459f7807e060ee20ae9ea8b
3
+ size 433270744
runs/Jul13_20-11-13_a518198a24e3/events.out.tfevents.1720901476.a518198a24e3.2562.14 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a63b8cc941fa6f43d26b1fac05a8b0a21ca140bd2b2da92dc7a0394f46a138b9
3
+ size 9867
runs/Jul13_20-11-13_a518198a24e3/events.out.tfevents.1720901584.a518198a24e3.2562.15 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4c08c1b10d53a64130a354a73f610568c43fe07947721eb85e212797e9b84d2c
3
+ size 723
tokenizer.json CHANGED
@@ -1,11 +1,6 @@
1
  {
2
  "version": "1.0",
3
- "truncation": {
4
- "direction": "Right",
5
- "max_length": 512,
6
- "strategy": "LongestFirst",
7
- "stride": 0
8
- },
9
  "padding": null,
10
  "added_tokens": [
11
  {
@@ -59,7 +54,7 @@
59
  "clean_text": true,
60
  "handle_chinese_chars": true,
61
  "strip_accents": null,
62
- "lowercase": false
63
  },
64
  "pre_tokenizer": {
65
  "type": "BertPreTokenizer"
 
1
  {
2
  "version": "1.0",
3
+ "truncation": null,
 
 
 
 
 
4
  "padding": null,
5
  "added_tokens": [
6
  {
 
54
  "clean_text": true,
55
  "handle_chinese_chars": true,
56
  "strip_accents": null,
57
+ "lowercase": true
58
  },
59
  "pre_tokenizer": {
60
  "type": "BertPreTokenizer"
tokenizer_config.json CHANGED
@@ -44,9 +44,9 @@
44
  "clean_up_tokenization_spaces": true,
45
  "cls_token": "[CLS]",
46
  "do_basic_tokenize": true,
47
- "do_lower_case": false,
48
  "mask_token": "[MASK]",
49
- "model_max_length": 512,
50
  "never_split": null,
51
  "pad_token": "[PAD]",
52
  "sep_token": "[SEP]",
 
44
  "clean_up_tokenization_spaces": true,
45
  "cls_token": "[CLS]",
46
  "do_basic_tokenize": true,
47
+ "do_lower_case": true,
48
  "mask_token": "[MASK]",
49
+ "model_max_length": 1000000000000000019884624838656,
50
  "never_split": null,
51
  "pad_token": "[PAD]",
52
  "sep_token": "[SEP]",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:91044e2900a7f2a0b7e49f54311f5995df446ecc659b5083369e8efb6d06d627
3
- size 5048
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44604e4feb03bd233f157d2b3797dea1025e6fc6ce9089c21c0bd797adb0f5ca
3
+ size 5112