DeepQ-1.5B GGUF

Quick Start

We advise you to clone llama.cpp and install it following the official guide. We follow the latest version of llama.cpp. In the following demonstration, we assume that you are running commands under the repository llama.cpp.

Since cloning the entire repo may be inefficient, you can manually download the GGUF file that you need or use huggingface-cli:

Install

pip install -U huggingface_hub

Download:

huggingface-cli download  Linzes/DeepQ-1.5B-gguf DeepQ-1.5B_Q5_K_M.gguf --local-dir . --local-dir-use-symlinks False

For users, to achieve chatbot-like experience, it is recommended to commence in the conversation mode:

./llama-cli -m <gguf-file-path> \
    -co -cnv -p "You are a helpful assistant." \
    -fa -ngl 80 -n 512
Downloads last month
2
GGUF
Model size
1.54B params
Architecture
qwen2
Hardware compatibility
Log In to view the estimation

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for Linzes/DeepQ-1.5B-gguf

Base model

Qwen/Qwen2.5-1.5B
Quantized
(56)
this model

Collection including Linzes/DeepQ-1.5B-gguf