Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up

TheBloke
/
Mixtral-8x7B-Instruct-v0.1-AWQ

Text Generation
Transformers
Safetensors
mixtral
conversational
text-generation-inference
4-bit precision
awq
Model card Files Files and versions Community
10
New discussion
Resources
  • PR & discussions documentation
  • Code of Conduct
  • Hub documentation

Use ybelkada/Mixtral-8x7B-Instruct-v0.1-AWQ with VLLM instead

πŸ‘ 3
1
#10 opened about 1 year ago by
blobpenguin

Inference taking too much time

3
#9 opened over 1 year ago by
tariksetia

Update README.md

#8 opened over 1 year ago by
skoita

RuntimeError: probability tensor contains either `inf`, `nan` or element < 0

2
#7 opened over 1 year ago by
aaganaie

TGI - response is an empty string

2
#6 opened over 1 year ago by
p-christ

OC is not a multiple of cta_N = 64

2
#5 opened over 1 year ago by
lazyDataScientist

Not supporting with TGI

1
#4 opened over 1 year ago by
abhishek3jangid

always getting 0 in output

πŸ‘ 4
15
#3 opened over 1 year ago by
xubuild

OOM under vLLM even with 80GB GPU

5
#2 opened over 1 year ago by
mike-ravkine

Not supported for TGI > 1.3 ?

πŸ‘ 1
20
#1 opened over 1 year ago by
paulcx
Company
TOS Privacy About Jobs
Website
Models Datasets Spaces Pricing Docs