mobilebert-uncased-squadv1-14blocks-structured39.8-int8

This model is a fine-tuned version of google/mobilebert-uncased on the squad dataset.

Notice that this model only has the first 14 transformer blocks. It is quantized and structually pruned by NNCF. The sparsity in remaining linear layers is 39.8%.

  • Torch f1: 90.15
  • IR f1: 89.8414

Framework versions

  • Transformers 4.25.1
  • Pytorch 1.13.1+cu116
  • Datasets 2.8.0
  • Tokenizers 0.13.2
Downloads last month
2
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Dataset used to train yujiepan/mobilebert-uncased-squadv1-14blocks-structured39.8-int8