sileod commited on
Commit
67bb00c
1 Parent(s): d3640a5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +14 -4
README.md CHANGED
@@ -318,9 +318,9 @@ pipeline_tag: zero-shot-classification
318
 
319
  This is [DeBERTa-v3-base](https://hf.co/microsoft/deberta-v3-base) fine-tuned with multi-task learning on 600 tasks of the [tasksource collection](https://github.com/sileod/tasksource/).
320
  This checkpoint has strong zero-shot validation performance on many tasks (e.g. 70% on WNLI), and can be used for:
321
- - Zero-shot entailment-based classification pipeline (similar to bart-mnli), see [ZS].
322
- - Natural language inference, and many other tasks with tasksource-adapters, see [TA]
323
- - Further fine-tuning with a new task (classification, token classification or multiple-choice).
324
 
325
  # [ZS] Zero-shot classification pipeline
326
  ```python
@@ -336,7 +336,7 @@ NLI training data of this model includes [label-nli](https://huggingface.co/data
336
  # [TA] Tasksource-adapters: 1 line access to hundreds of tasks
337
 
338
  ```python
339
- !pip install tasknet tasksource
340
  import tasknet as tn
341
  pipe = tn.load_pipeline('sileod/deberta-v3-base-tasksource-nli','glue/sst2') # works for 500+ tasksource tasks
342
  pipe(['That movie was great !', 'Awful movie.'])
@@ -346,6 +346,16 @@ The list of tasks is available in model config.json.
346
  This is more efficient than ZS since it requires only one forward pass per example, but it is less flexible.
347
 
348
 
 
 
 
 
 
 
 
 
 
 
349
  ## Evaluation
350
  This model ranked 1st among all models with the microsoft/deberta-v3-base architecture according to the IBM model recycling evaluation.
351
  https://ibm.github.io/model-recycling/
 
318
 
319
  This is [DeBERTa-v3-base](https://hf.co/microsoft/deberta-v3-base) fine-tuned with multi-task learning on 600 tasks of the [tasksource collection](https://github.com/sileod/tasksource/).
320
  This checkpoint has strong zero-shot validation performance on many tasks (e.g. 70% on WNLI), and can be used for:
321
+ - Zero-shot entailment-based classification pipeline (similar to bart-mnli), see [ZS]([ZS] Zero-shot classification pipeline).
322
+ - Natural language inference, and many other tasks with tasksource-adapters, see [TA]([TA] Tasksource-adapters: 1 line access to hundreds of tasks).
323
+ - Further fine-tuning with a new task (classification, token classification or multiple-choice) [FT]([FT] Tasknet: 3 lines fine-tuning).
324
 
325
  # [ZS] Zero-shot classification pipeline
326
  ```python
 
336
  # [TA] Tasksource-adapters: 1 line access to hundreds of tasks
337
 
338
  ```python
339
+ # !pip install tasknet
340
  import tasknet as tn
341
  pipe = tn.load_pipeline('sileod/deberta-v3-base-tasksource-nli','glue/sst2') # works for 500+ tasksource tasks
342
  pipe(['That movie was great !', 'Awful movie.'])
 
346
  This is more efficient than ZS since it requires only one forward pass per example, but it is less flexible.
347
 
348
 
349
+ # [FT] Tasknet: 3 lines fine-tuning
350
+
351
+ ```python
352
+ # !pip install tasknet
353
+ import tasknet as tn
354
+ hparams=dict(model_name='sileod/deberta-v3-base-tasksource-nli', learning_rate=2e-5)
355
+ model, trainer = tn.Model_Trainer([tn.AutoTask("glue/rte")], hparams)
356
+ trainer.train()
357
+ ```
358
+
359
  ## Evaluation
360
  This model ranked 1st among all models with the microsoft/deberta-v3-base architecture according to the IBM model recycling evaluation.
361
  https://ibm.github.io/model-recycling/