Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up

unsloth
/
DeepSeek-V3-0324-GGUF

Text Generation
Transformers
GGUF
English
deepseek_v3
deepseek
unsloth
custom_code
fp8
conversational
Model card Files Files and versions
xet
Community
17
New discussion
Resources
  • PR & discussions documentation
  • Code of Conduct
  • Hub documentation

IQ2_XXS optimal for me

#17 opened about 1 month ago by
jweb

World's Largest Dataset

#16 opened about 2 months ago by
UJJAWAL-TYAGI

Re-converting the GGUF for MLA?

👍 6
2
#15 opened about 2 months ago by
Silver267

What tool/framework to test gguf models?

1
#14 opened about 2 months ago by
bobchenyx

Request: DOI

#13 opened 2 months ago by
jeffhoule01

How to run ollama using these new quantized weights?

👀 1
2
#12 opened 2 months ago by
vadimkantorov

Running Model "unsloth/DeepSeek-V3-0324-GGUF" with vLLM does not working

2
#11 opened 2 months ago by
puppadas

The UD-IQ2_XXS is surprisingly good, but it's good to know that it degrades gradually but significantly after about 1000 tokens.

1
#9 opened 2 months ago by
mmbela

671B params or 685B params?

6
#8 opened 2 months ago by
createthis

how to run tools use correctly

#7 opened 2 months ago by
rockcat-miao

How many bits of Quantization is enough for Code Generation Tasks?

1
#5 opened 2 months ago by
luweigen

Added IQ1_S version to Ollama

3
#4 opened 2 months ago by
Muhammadreza

Is the 2.51bit model using imatrix?

7
#3 opened 3 months ago by
daweiba12

Will you release the imatrix.dat used for the quants?

2
#2 opened 3 months ago by
tdh111

Would There be Dynamic Qunatized Versions like 2.51bit

8
#1 opened 3 months ago by
MotorBottle
Company
TOS Privacy About Jobs
Website
Models Datasets Spaces Pricing Docs