ONNX
DaniAffCH commited on
Commit
31b7bd1
·
1 Parent(s): 31f3c1a

[GSoC] Add block quantized models (#270)

Browse files

* Gemm and MatMul block quantization support

* refactoring

* fix indentation

* node name independent

* Block quantization tool:
- constant weight category supported
- add data type saturation
- handled the case in which all the elements within a block are the same

benchmark script modified to support block quantized models

block quantized some models

* add missing block quantized models

* formatting

* add blocked models to eval script. Evaluation yunet

* Add sface and pphumanseg evaluation, block quantization tool fix, handpose blocked model fix, removed blocked CRNN EN,

* changed evaluation metric in block_quantize script and add verbose mode

* Add evaluation for PP-ResNet and Mobilenet

* changed file suffix and update readmes

* renamed int8bq

Files changed (1) hide show
  1. README.md +3 -1
README.md CHANGED
@@ -8,8 +8,10 @@ Key features of the YOLOX object detector
8
  - **SimOTA advanced label assignment strategy** reduces training time and avoids additional solver hyperparameters
9
  - **Strong data augmentations like MixUp and Mosiac** to boost YOLOX performance
10
 
11
- Note:
12
  - This version of YoloX: YoloX_s
 
 
13
 
14
  ## Demo
15
 
 
8
  - **SimOTA advanced label assignment strategy** reduces training time and avoids additional solver hyperparameters
9
  - **Strong data augmentations like MixUp and Mosiac** to boost YOLOX performance
10
 
11
+ **Note**:
12
  - This version of YoloX: YoloX_s
13
+ - `object_detection_yolox_2022nov_int8bq.onnx` represents the block-quantized version in int8 precision and is generated using [block_quantize.py](../../tools/quantize/block_quantize.py) with `block_size=64`.
14
+
15
 
16
  ## Demo
17