metadata
license: llama3.1
language:
- en
- ja
base_model:
- tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.5
base_model_relation: quantized
Swallow-8B-it-v05-gguf-q6_k-mixed-v1
- Quantization Type: Mixed Precision (
q5_K
,q6_K
,q8_0
) - Bits Per Weight (BPW):
7.13
Swallow-8B-it-v05-gguf-q6_k-mixed-v2
- Quantization Type: Mixed Precision (
q6_K
,q8_0
) - Bits Per Weight (BPW):
7.50
Swallow-8B-it-v05-gguf-q8_0-mixed-v1
- Quantization Type: Mixed Precision (
bf16
,q4_K
,q5_K
,q6_K
,q8_0
) - Bits Per Weight (BPW):
8.01
Swallow-8B-it-v05-gguf-q8_0-mixed-v2
- Quantization Type: Mixed Precision (
bf16
,q5_K
,q6_K
,q8_0
) - Bits Per Weight (BPW):
9.31
Swallow-8B-it-v05-gguf-q8_0-mixed-v3
- Quantization Type: Mixed Precision (
bf16
,q6_K
,q8_0
) - Bits Per Weight (BPW):
11.44
Swallow-8B-it-v05-gguf-q8_0-mixed-v4
- Quantization Type: Mixed Precision (
bf16
,q8_0
) - Bits Per Weight (BPW):
13.38