Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Posts
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up

cognitivecomputations
/
DeepSeek-R1-AWQ

Text Generation
Transformers
Safetensors
English
Chinese
deepseek_v3
conversational
custom_code
text-generation-inference
4-bit precision
awq
Model card Files Files and versions
xet
Community
34
New discussion
Resources
  • PR & discussions documentation
  • Code of Conduct
  • Hub documentation

update vllm to 0.8.x and meet some trouble

3
#34 opened 8 days ago by
HuggingLianWang

AMD Instinct MI210 + vllm fail to run this model, any solutions please? Is there any other deepseek-r1-671b models that can run succesfully on AMD Instinct MI210 + vllm? Thanks!

5
#33 opened about 1 month ago by
luciagan

More stable startup command, not easy oom.

1
#31 opened about 2 months ago by
Piekey

The awq quantization model may encounter garbled characters when performing inference on long texts.

9
#24 opened 2 months ago by
wx111

Add instructions to run R1-AWQ on SGLang

2
#22 opened 3 months ago by
ganler

requests get stuck when sending long prompts (already solved, but still don't know why?)

1
1
#18 opened 3 months ago by
uv0xab

Is there any accuracy results comparing to original DeepSeek-R1?

2
#15 opened 3 months ago by
traphix

Any one can run this model with SGlang framework?

5
#13 opened 3 months ago by
muziyongshixin

Regarding the issue of inconsistent calculation of tokens

#12 opened 3 months ago by
liguoyu3564

Max-Batch-Size, max-num-sequence, and fp_cache fp8_e4m3

#11 opened 3 months ago by
BenFogerty

The inference performance of the DeepSeek-R1-AWQ model is weak compared to the DeepSeek-R1 model

3
8
#3 opened 3 months ago by
qingqingz916
Company
TOS Privacy About Jobs
Website
Models Datasets Spaces Pricing Docs