Update README.md
Browse files
README.md
CHANGED
|
@@ -8,6 +8,7 @@ thumbnail: https://huggingface.co/front/thumbnails/microsoft.png
|
|
| 8 |
license: mit
|
| 9 |
widget:
|
| 10 |
- text: "[CLS] I love you. [SEP] I like you. [SEP]"
|
|
|
|
| 11 |
---
|
| 12 |
|
| 13 |
## DeBERTa: Decoding-enhanced BERT with Disentangled Attention
|
|
@@ -40,9 +41,7 @@ We present the dev results on SQuAD 1.1/2.0 and several GLUE benchmark tasks.
|
|
| 40 |
```bash
|
| 41 |
cd transformers/examples/text-classification/
|
| 42 |
export TASK_NAME=mrpc
|
| 43 |
-
python -m torch.distributed.launch --nproc_per_node=8 run_glue.py --model_name_or_path microsoft/deberta-v2-xxlarge
|
| 44 |
-
--task_name $TASK_NAME --do_train --do_eval --max_seq_length 128 --per_device_train_batch_size 4 \\
|
| 45 |
-
--learning_rate 3e-6 --num_train_epochs 3 --output_dir /tmp/$TASK_NAME/ --overwrite_output_dir --sharded_ddp --fp16
|
| 46 |
```
|
| 47 |
|
| 48 |
### Citation
|
|
|
|
| 8 |
license: mit
|
| 9 |
widget:
|
| 10 |
- text: "[CLS] I love you. [SEP] I like you. [SEP]"
|
| 11 |
+
pipeline_tag: zero-shot-classification
|
| 12 |
---
|
| 13 |
|
| 14 |
## DeBERTa: Decoding-enhanced BERT with Disentangled Attention
|
|
|
|
| 41 |
```bash
|
| 42 |
cd transformers/examples/text-classification/
|
| 43 |
export TASK_NAME=mrpc
|
| 44 |
+
python -m torch.distributed.launch --nproc_per_node=8 run_glue.py --model_name_or_path microsoft/deberta-v2-xxlarge \\\n--task_name $TASK_NAME --do_train --do_eval --max_seq_length 128 --per_device_train_batch_size 4 \\\n--learning_rate 3e-6 --num_train_epochs 3 --output_dir /tmp/$TASK_NAME/ --overwrite_output_dir --sharded_ddp --fp16
|
|
|
|
|
|
|
| 45 |
```
|
| 46 |
|
| 47 |
### Citation
|