Update README.md
Browse files
README.md
CHANGED
@@ -54,7 +54,7 @@ The model is quantized with nvidia-modelopt **v0.15.1** <br>
|
|
54 |
**Test Hardware:** H100 <br>
|
55 |
|
56 |
## Post Training Quantization
|
57 |
-
This model was obtained by quantizing the weights and activations of Meta-Llama-3.1-8B to FP8 data type, ready for inference with TensorRT-LLM and vLLM. Only the weights and activations of the linear operators within transformers blocks are quantized. This optimization reduces the number of bits per parameter from 16 to 8, reducing the disk size and GPU memory requirements by approximately 50%. On H100, we achieved speedup
|
58 |
|
59 |
## Usage
|
60 |
|
|
|
54 |
**Test Hardware:** H100 <br>
|
55 |
|
56 |
## Post Training Quantization
|
57 |
+
This model was obtained by quantizing the weights and activations of Meta-Llama-3.1-8B to FP8 data type, ready for inference with TensorRT-LLM and vLLM. Only the weights and activations of the linear operators within transformers blocks are quantized. This optimization reduces the number of bits per parameter from 16 to 8, reducing the disk size and GPU memory requirements by approximately 50%. On H100, we achieved 1.3x speedup.
|
58 |
|
59 |
## Usage
|
60 |
|