Update README.md
Browse files
README.md
CHANGED
@@ -318,9 +318,9 @@ pipeline_tag: zero-shot-classification
|
|
318 |
|
319 |
This is [DeBERTa-v3-base](https://hf.co/microsoft/deberta-v3-base) fine-tuned with multi-task learning on 600 tasks of the [tasksource collection](https://github.com/sileod/tasksource/).
|
320 |
This checkpoint has strong zero-shot validation performance on many tasks (e.g. 70% on WNLI), and can be used for:
|
321 |
-
- Zero-shot entailment-based classification pipeline (similar to bart-mnli), see [ZS].
|
322 |
-
- Natural language inference, and many other tasks with tasksource-adapters, see [TA]
|
323 |
-
- Further fine-tuning with a new task (classification, token classification or multiple-choice).
|
324 |
|
325 |
# [ZS] Zero-shot classification pipeline
|
326 |
```python
|
@@ -336,7 +336,7 @@ NLI training data of this model includes [label-nli](https://huggingface.co/data
|
|
336 |
# [TA] Tasksource-adapters: 1 line access to hundreds of tasks
|
337 |
|
338 |
```python
|
339 |
-
!pip install tasknet
|
340 |
import tasknet as tn
|
341 |
pipe = tn.load_pipeline('sileod/deberta-v3-base-tasksource-nli','glue/sst2') # works for 500+ tasksource tasks
|
342 |
pipe(['That movie was great !', 'Awful movie.'])
|
@@ -346,6 +346,16 @@ The list of tasks is available in model config.json.
|
|
346 |
This is more efficient than ZS since it requires only one forward pass per example, but it is less flexible.
|
347 |
|
348 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
349 |
## Evaluation
|
350 |
This model ranked 1st among all models with the microsoft/deberta-v3-base architecture according to the IBM model recycling evaluation.
|
351 |
https://ibm.github.io/model-recycling/
|
|
|
318 |
|
319 |
This is [DeBERTa-v3-base](https://hf.co/microsoft/deberta-v3-base) fine-tuned with multi-task learning on 600 tasks of the [tasksource collection](https://github.com/sileod/tasksource/).
|
320 |
This checkpoint has strong zero-shot validation performance on many tasks (e.g. 70% on WNLI), and can be used for:
|
321 |
+
- Zero-shot entailment-based classification pipeline (similar to bart-mnli), see [ZS]([ZS] Zero-shot classification pipeline).
|
322 |
+
- Natural language inference, and many other tasks with tasksource-adapters, see [TA]([TA] Tasksource-adapters: 1 line access to hundreds of tasks).
|
323 |
+
- Further fine-tuning with a new task (classification, token classification or multiple-choice) [FT]([FT] Tasknet: 3 lines fine-tuning).
|
324 |
|
325 |
# [ZS] Zero-shot classification pipeline
|
326 |
```python
|
|
|
336 |
# [TA] Tasksource-adapters: 1 line access to hundreds of tasks
|
337 |
|
338 |
```python
|
339 |
+
# !pip install tasknet
|
340 |
import tasknet as tn
|
341 |
pipe = tn.load_pipeline('sileod/deberta-v3-base-tasksource-nli','glue/sst2') # works for 500+ tasksource tasks
|
342 |
pipe(['That movie was great !', 'Awful movie.'])
|
|
|
346 |
This is more efficient than ZS since it requires only one forward pass per example, but it is less flexible.
|
347 |
|
348 |
|
349 |
+
# [FT] Tasknet: 3 lines fine-tuning
|
350 |
+
|
351 |
+
```python
|
352 |
+
# !pip install tasknet
|
353 |
+
import tasknet as tn
|
354 |
+
hparams=dict(model_name='sileod/deberta-v3-base-tasksource-nli', learning_rate=2e-5)
|
355 |
+
model, trainer = tn.Model_Trainer([tn.AutoTask("glue/rte")], hparams)
|
356 |
+
trainer.train()
|
357 |
+
```
|
358 |
+
|
359 |
## Evaluation
|
360 |
This model ranked 1st among all models with the microsoft/deberta-v3-base architecture according to the IBM model recycling evaluation.
|
361 |
https://ibm.github.io/model-recycling/
|