zhiyucheng commited on
Commit
adb8efe
·
verified ·
1 Parent(s): b64a154

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -54,7 +54,7 @@ The model is quantized with nvidia-modelopt **v0.15.1** <br>
54
  **Test Hardware:** H100 <br>
55
 
56
  ## Post Training Quantization
57
- This model was obtained by quantizing the weights and activations of Meta-Llama-3.1-8B to FP8 data type, ready for inference with TensorRT-LLM and vLLM. Only the weights and activations of the linear operators within transformers blocks are quantized. This optimization reduces the number of bits per parameter from 16 to 8, reducing the disk size and GPU memory requirements by approximately 50%. On H100, we achieved speedup
58
 
59
  ## Usage
60
 
 
54
  **Test Hardware:** H100 <br>
55
 
56
  ## Post Training Quantization
57
+ This model was obtained by quantizing the weights and activations of Meta-Llama-3.1-8B to FP8 data type, ready for inference with TensorRT-LLM and vLLM. Only the weights and activations of the linear operators within transformers blocks are quantized. This optimization reduces the number of bits per parameter from 16 to 8, reducing the disk size and GPU memory requirements by approximately 50%. On H100, we achieved 1.3x speedup.
58
 
59
  ## Usage
60