bnjmnmarie commited on
Commit
f0d4540
·
verified ·
1 Parent(s): 9cb2c9a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +24 -3
README.md CHANGED
@@ -1,3 +1,24 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model:
4
+ - Qwen/Qwen3-8B
5
+ tags:
6
+ - autoround
7
+ ---
8
+ This is [Qwen/Qwen3-8B](https://huggingface.co/Qwen/Qwen3-8B) quantized with [AutoRound](https://github.com/intel/auto-round/tree/main/auto_round) in 2-bit (symmetric + gptq format). The model has been created, tested, and evaluated by The Kaitchup.
9
+ The model is compatible with vLLM and Transformers.
10
+
11
+ More details in this article:
12
+ [How Well Does Qwen3 Handle 4-bit and 2-bit Quantization?](https://kaitchup.substack.com/p/how-well-does-qwen3-handle-4-bit)
13
+
14
+
15
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/64b93e6bd6c468ac7536607e/3J5BLZXRl6eT8g11r1JDQ.png)
16
+
17
+
18
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/64b93e6bd6c468ac7536607e/0wvK6MwnngzKA8m2qs7qS.png)
19
+
20
+ - **Developed by:** [The Kaitchup](https://kaitchup.substack.com/)
21
+ - **License:** Apache 2.0 license
22
+
23
+ ## How to Support My Work
24
+ Subscribe to [The Kaitchup](https://kaitchup.substack.com/subscribe). This helps me a lot to continue quantizing and evaluating models for free.