-
The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits
Paper • 2402.17764 • Published • 618 -
Beyond Language Models: Byte Models are Digital World Simulators
Paper • 2402.19155 • Published • 54 -
BitNet v2: Native 4-bit Activations with Hadamard Transformation for 1-bit LLMs
Paper • 2504.18415 • Published • 43 -
Kijai/PrecompiledWheels
Updated • 15
kas
shing3232
AI & ML interests
None yet
Recent Activity
updated
a collection
22 days ago
sakura
new activity
22 days ago
Qwen/Qwen1.5-MoE-A2.7B-Chat-GPTQ-Int4:Int4为什么比没量化的float32和float16还慢
upvoted
a
paper
about 1 month ago
TransMLA: Multi-head Latent Attention Is All You Need
Organizations
None yet
Collections
1
spaces
1
models
9
shing3232/Sakura-1.5B-Qwen2.5-v1.0-GGUF-IMX
Updated
•
50
•
1
shing3232/sakura-14b-qwen2beta-v0.9.2-IMX
Updated
•
7
•
3
shing3232/Sakura13B-LNovel-v0.9-qwen1.5-GGUF-IMX
Updated
•
67
•
7
shing3232/Sakura1.8B-LNovel-v0.9pre2-qwen1_GGUF-IMX
Updated
•
43
shing3232/Sakura13B-LNovel-v0.9b-GGUF-IMX-2.33_re
Updated
•
5
shing3232/Sakura1.8B-LNovel-v0.9-qwen1.5_GGUF-IMX_re
Updated
•
4
•
1
shing3232/Sakura13B-LNovel-v0.9b-GGUF-IMX-2.33
Updated
•
47
•
3
shing3232/Sakura-LNovel-v0.9b-GGUF-IMX-JPZH
Updated
•
241
shing3232/Sakura-13B-LNovel-v0.9b-GGUF-IMX-wikitest
Updated
•
6