Update README.md
Browse filesUpdate link to distillation process
README.md
CHANGED
|
@@ -11,7 +11,7 @@ datasets:
|
|
| 11 |
This model is a distilled version of the [BERT base model](https://huggingface.co/bert-base-cased).
|
| 12 |
It was introduced in [this paper](https://arxiv.org/abs/1910.01108).
|
| 13 |
The code for the distillation process can be found
|
| 14 |
-
[here](https://github.com/huggingface/transformers/tree/
|
| 15 |
This model is cased: it does make a difference between english and English.
|
| 16 |
|
| 17 |
All the training details on the pre-training, the uses, limitations and potential biases (included below) are the same as for [DistilBERT-base-uncased](https://huggingface.co/distilbert-base-uncased).
|
|
|
|
| 11 |
This model is a distilled version of the [BERT base model](https://huggingface.co/bert-base-cased).
|
| 12 |
It was introduced in [this paper](https://arxiv.org/abs/1910.01108).
|
| 13 |
The code for the distillation process can be found
|
| 14 |
+
[here](https://github.com/huggingface/transformers/tree/main/examples/research_projects/distillation).
|
| 15 |
This model is cased: it does make a difference between english and English.
|
| 16 |
|
| 17 |
All the training details on the pre-training, the uses, limitations and potential biases (included below) are the same as for [DistilBERT-base-uncased](https://huggingface.co/distilbert-base-uncased).
|