Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up

TheBloke
/
Mixtral-8x7B-v0.1-GPTQ

Text Generation
Transformers
Safetensors
mixtral
text-generation-inference
4-bit precision
gptq
Model card Files Files and versions Community
11
New discussion
Resources
  • PR & discussions documentation
  • Code of Conduct
  • Hub documentation

The generative output is strange

#11 opened 11 months ago by
tangpeng

Speeds compared to llama_cpp_python?

2
#10 opened over 1 year ago by
SpaceCowboy850

Unable to start TGI service for TheBloke/Mixtral-8x7B-v0.1-GPTQ with num_shard as 4

#9 opened over 1 year ago by
swapnil3597

What would be the minimal Sagemaker instance to deploy this model ?

2
#7 opened over 1 year ago by
CarlosAndrea

ValueError: Unsupported model type mixtral

1
#6 opened over 1 year ago by
seabasshn

RuntimeError: shape '[32, 8]' is invalid for input of size 0

7
#5 opened over 1 year ago by
woldeM

Are you going to release mixtral-8x7B-v0.1-awq

👍 4
#4 opened over 1 year ago by
HelloJiang

Running the model using "pip install auto-gptq" still results in "CUDA extension not installed"

#3 opened over 1 year ago by
mvetter

TypeError: mixtral isn't supported yet.

2
#2 opened over 1 year ago by
luv2261

Build AutoGPTQ from source

3
#1 opened over 1 year ago by
PeePants
Company
TOS Privacy About Jobs
Website
Models Datasets Spaces Pricing Docs