Zero-Shot Classification
Transformers
PyTorch
Safetensors
bert
text-classification
Inference Endpoints
saattrupdan commited on
Commit
7bd18ab
1 Parent(s): 7ba7453

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +13 -5
README.md CHANGED
@@ -32,9 +32,13 @@ inference:
32
 
33
  This model is a fine-tuned version of [NbAiLab/nb-bert-base](https://huggingface.co/NbAiLab/nb-bert-base) for Natural Language Inference in Danish, Norwegian Bokmål and Swedish.
34
 
35
- It has been fine-tuned on a dataset composed of [DanFEVER](https://aclanthology.org/2021.nodalida-main.pdf#page=439) as well as machine translated versions of [MultiNLI](https://cims.nyu.edu/~sbowman/multinli/) and [CommitmentBank](https://doi.org/10.18148/sub/2019.v23i2.601) into all three languages, and machine translated versions of [FEVER](https://aclanthology.org/N18-1074/) and [Adversarial NLI](https://aclanthology.org/2020.acl-main.441/) into Swedish.
36
 
37
- The three languages are sampled equally during training, and they're validated on validation splits of [DanFEVER](https://aclanthology.org/2021.nodalida-main.pdf#page=439) and machine translated versions of [MultiNLI](https://cims.nyu.edu/~sbowman/multinli/) for Swedish and Norwegian Bokmål, sampled equally.
 
 
 
 
38
 
39
 
40
  ## Quick start
@@ -69,13 +73,17 @@ We report Matthew's Correlation Coefficient (MCC), macro-average F1-score as wel
69
 
70
  | **Model** | **MCC** | **Macro-F1** | **Accuracy** | **Number of Parameters** |
71
  | :-------- | :------------ | :--------- | :----------- | :----------- |
72
- | [`alexandrainst/nb-bert-large-nli-scandi`](https://huggingface.co/alexandrainst/nb-bert-large-nli-scandi) | **73.80%** | **58.41%** | **86.98%** | 354M |
73
- | `alexandrainst/nb-bert-base-nli-scandi` (this) | 62.44% | 55.00% | 80.42% | 178M |
74
- | [`alexandrainst/electra-small-nordic-nli-scandi`](https://huggingface.co/alexandrainst/electra-small-nordic-nli-scandi) | 47.28% | 48.88% | 73.46% | **22M** |
75
 
76
 
77
  ## Training procedure
78
 
 
 
 
 
79
  ### Training hyperparameters
80
 
81
  The following hyperparameters were used during training:
 
32
 
33
  This model is a fine-tuned version of [NbAiLab/nb-bert-base](https://huggingface.co/NbAiLab/nb-bert-base) for Natural Language Inference in Danish, Norwegian Bokmål and Swedish.
34
 
35
+ We have released three models for Scandinavian NLI, of different sizes:
36
 
37
+ - [alexandrainst/scandi-nli-large](https://huggingface.co/alexandrainst/scandi-nli-large)
38
+ - [alexandrainst/scandi-nli-base](https://huggingface.co/alexandrainst/scandi-nli-base)
39
+ - [alexandrainst/scandi-nli-small](https://huggingface.co/alexandrainst/scandi-nli-small)
40
+
41
+ The performance and model size of each of them can be found in the Performance section below.
42
 
43
 
44
  ## Quick start
 
73
 
74
  | **Model** | **MCC** | **Macro-F1** | **Accuracy** | **Number of Parameters** |
75
  | :-------- | :------------ | :--------- | :----------- | :----------- |
76
+ | [`alexandrainst/scandi-nli-large`](https://huggingface.co/alexandrainst/scandi-nli-large) | **73.80%** | **58.41%** | **86.98%** | 354M |
77
+ | `alexandrainst/scandi-nli-base` (this) | 62.44% | 55.00% | 80.42% | 178M |
78
+ | [`alexandrainst/scandi-nli-small`](https://huggingface.co/alexandrainst/scandi-nli-small) | 47.28% | 48.88% | 73.46% | **22M** |
79
 
80
 
81
  ## Training procedure
82
 
83
+ It has been fine-tuned on a dataset composed of [DanFEVER](https://aclanthology.org/2021.nodalida-main.pdf#page=439) as well as machine translated versions of [MultiNLI](https://cims.nyu.edu/~sbowman/multinli/) and [CommitmentBank](https://doi.org/10.18148/sub/2019.v23i2.601) into all three languages, and machine translated versions of [FEVER](https://aclanthology.org/N18-1074/) and [Adversarial NLI](https://aclanthology.org/2020.acl-main.441/) into Swedish.
84
+
85
+ The three languages are sampled equally during training, and they're validated on validation splits of [DanFEVER](https://aclanthology.org/2021.nodalida-main.pdf#page=439) and machine translated versions of [MultiNLI](https://cims.nyu.edu/~sbowman/multinli/) for Swedish and Norwegian Bokmål, sampled equally.
86
+
87
  ### Training hyperparameters
88
 
89
  The following hyperparameters were used during training: