Update README.md
Browse files
README.md
CHANGED
@@ -179,7 +179,7 @@ if __name__ == '__main__':
|
|
179 |
## Training and Optimization Details
|
180 |
|
181 |
**Quantization Process:**
|
182 |
-
This model employs SmoothQuant quantization implemented through LLM Compressor. SmoothQuant redistributes quantization difficulty from activations to weights by applying mathematically equivalent transformations, enabling effective FP8 quantization. The quantization calibration was performed using the WikiText
|
183 |
|
184 |
**Calibration Dataset:**
|
185 |
- **WikiText**: Used for SmoothQuant calibration to optimize quantization parameters
|
|
|
179 |
## Training and Optimization Details
|
180 |
|
181 |
**Quantization Process:**
|
182 |
+
This model employs SmoothQuant quantization implemented through LLM Compressor. SmoothQuant redistributes quantization difficulty from activations to weights by applying mathematically equivalent transformations, enabling effective FP8 quantization. The quantization calibration was performed using the WikiText dataset. The quantization process specifically targets the weights and activations of linear operators within transformer blocks, preserving model accuracy while significantly reducing computational requirements.
|
183 |
|
184 |
**Calibration Dataset:**
|
185 |
- **WikiText**: Used for SmoothQuant calibration to optimize quantization parameters
|