Narsil HF staff commited on
Commit
5a3287e
1 Parent(s): dfa909d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -3
README.md CHANGED
@@ -8,6 +8,7 @@ thumbnail: https://huggingface.co/front/thumbnails/microsoft.png
8
  license: mit
9
  widget:
10
  - text: "[CLS] I love you. [SEP] I like you. [SEP]"
 
11
  ---
12
 
13
  ## DeBERTa: Decoding-enhanced BERT with Disentangled Attention
@@ -40,9 +41,7 @@ We present the dev results on SQuAD 1.1/2.0 and several GLUE benchmark tasks.
40
  ```bash
41
  cd transformers/examples/text-classification/
42
  export TASK_NAME=mrpc
43
- python -m torch.distributed.launch --nproc_per_node=8 run_glue.py --model_name_or_path microsoft/deberta-v2-xxlarge \\
44
- --task_name $TASK_NAME --do_train --do_eval --max_seq_length 128 --per_device_train_batch_size 4 \\
45
- --learning_rate 3e-6 --num_train_epochs 3 --output_dir /tmp/$TASK_NAME/ --overwrite_output_dir --sharded_ddp --fp16
46
  ```
47
 
48
  ### Citation
 
8
  license: mit
9
  widget:
10
  - text: "[CLS] I love you. [SEP] I like you. [SEP]"
11
+ pipeline_tag: zero-shot-classification
12
  ---
13
 
14
  ## DeBERTa: Decoding-enhanced BERT with Disentangled Attention
 
41
  ```bash
42
  cd transformers/examples/text-classification/
43
  export TASK_NAME=mrpc
44
+ python -m torch.distributed.launch --nproc_per_node=8 run_glue.py --model_name_or_path microsoft/deberta-v2-xxlarge \\\n--task_name $TASK_NAME --do_train --do_eval --max_seq_length 128 --per_device_train_batch_size 4 \\\n--learning_rate 3e-6 --num_train_epochs 3 --output_dir /tmp/$TASK_NAME/ --overwrite_output_dir --sharded_ddp --fp16
 
 
45
  ```
46
 
47
  ### Citation