Nice performance
#1
by
Evi1ran
- opened
Good job, nice performance! π
Could you please release the full weights? (Not quantized with AWQ int4)
Good job, nice performance! π
Could you please release the full weights? (Not quantized with AWQ int4)
it's the quantization of DavidAU/Qwen2.5-QwQ-35B-Eureka-Cubed-abliterated-uncensored.
i just want to test lmdeploy's performance compared to vllm so i create this.
sadly my gpu with 24GB VRAM still can not run this quantization on lmdeploy.
Good job, nice performance! π
Could you please release the full weights? (Not quantized with AWQ int4)
by the way, if you have enough RAM and VRAM for full weights, could you make gptq int4 and awq int4 quantization of hf format?
i encouter some problems hard to solve with AutoGPTQ and AutoAWQ