bnjmnmarie commited on
Commit
95420e8
·
verified ·
1 Parent(s): 1700227

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +23 -3
README.md CHANGED
@@ -1,3 +1,23 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model:
4
+ - Qwen/Qwen2.5-72B-Instruct
5
+ tags:
6
+ - autoround
7
+ ---
8
+ This is [Qwen/Qwen2.5-72B-Instruct](https://huggingface.co/Qwen/Qwen2.5-72B-Instruct) quantized with [AutoRound](https://github.com/intel/auto-round/tree/main/auto_round) in 2-bit (symmetric + gptq format) with a group size of 64 and calibration samples of 2048 tokens. The model has been created, tested, and evaluated by The Kaitchup.
9
+ The model is compatible with vLLM and Transformers.
10
+
11
+ More details in this article:
12
+ [Accurate 2-bit Quantization: Run Massive LLMs on a Single Consumer GPU](https://kaitchup.substack.com/p/accurate-2-bit-quantization-run-massive)
13
+
14
+
15
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/64b93e6bd6c468ac7536607e/hOlFr-7E3oIZvNHHOuy-K.png)
16
+
17
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/64b93e6bd6c468ac7536607e/MfZPTnkvXp9elT_UE5DnE.png)
18
+
19
+ - **Developed by:** [The Kaitchup](https://kaitchup.substack.com/)
20
+ - **License:** Apache 2.0 license
21
+
22
+ ## How to Support My Work
23
+ Subscribe to [The Kaitchup](https://kaitchup.substack.com/subscribe). This helps me a lot to continue quantizing and evaluating models for free.