YAML Metadata
Warning:
empty or missing yaml metadata in repo card
(https://huggingface.co/docs/hub/model-cards#model-card-metadata)
olmOCR-7B 8-bit Quantized Model
This is the olmOCR-7B model quantized to 8-bit precision.
Usage with SGLang
To use this 8-bit quantized model with SGLang:
python -m sglang.launch_server \
--model-path /path/to/./olmOCR-7B-8bit \
--trust-remote-code \
--port 30000 --host 0.0.0.0
Note: The --trust-remote-code
flag is necessary because the model uses a custom tokenizer.
Original Model Information
The original model is a qwen2_vl
architecture variant, which is a vision-language model
based on Qwen2. The model has been modified for OCR capabilities.
Quantization Information
This model has been quantized to 8-bit precision to reduce memory usage.
- Downloads last month
- 2
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support