Add files using upload-large-folder tool
Browse files- Llama-3.1-Nemotron-Nano-8B-v1-Q4_K_M.gguf +2 -2
- Llama-3.1-Nemotron-Nano-8B-v1-UD-IQ1_S.gguf +2 -2
- Llama-3.1-Nemotron-Nano-8B-v1-UD-Q2_K_XL.gguf +2 -2
- Llama-3.1-Nemotron-Nano-8B-v1-UD-Q3_K_XL.gguf +2 -2
- Llama-3.1-Nemotron-Nano-8B-v1-UD-Q4_K_XL.gguf +2 -2
- Llama-3.1-Nemotron-Nano-8B-v1-UD-Q5_K_XL.gguf +2 -2
- README.md +17 -0
- config.json +3 -2
Llama-3.1-Nemotron-Nano-8B-v1-Q4_K_M.gguf
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3fc180a9b0310146a86d1a0ac79e0fa7b932fbe5ad697b0aaf73f931f38eac3c
|
3 |
+
size 4920737344
|
Llama-3.1-Nemotron-Nano-8B-v1-UD-IQ1_S.gguf
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d00bbcaae67af32a221a2a173402e9411f13dc098e11fbeca6ba554af984c58e
|
3 |
+
size 2164670016
|
Llama-3.1-Nemotron-Nano-8B-v1-UD-Q2_K_XL.gguf
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:366adf3f9dddeb82256b730cc1fd8ec7558990d876b9e182296b822bf9cabcb6
|
3 |
+
size 3388767808
|
Llama-3.1-Nemotron-Nano-8B-v1-UD-Q3_K_XL.gguf
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:10ee118ebe4293541b546a93cb772d482be7c6cac94e27ec17aa295bc06d0b67
|
3 |
+
size 4199611968
|
Llama-3.1-Nemotron-Nano-8B-v1-UD-Q4_K_XL.gguf
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c62e39d9170993d5eab0260b0935429f11af85100689f6dd40a4743bd5ac1476
|
3 |
+
size 4994203200
|
Llama-3.1-Nemotron-Nano-8B-v1-UD-Q5_K_XL.gguf
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bfcb0f81d1fa19274914c027ed9007881f81a5f1ec81217ab4f0c89bad3f709d
|
3 |
+
size 5738135104
|
README.md
CHANGED
@@ -16,6 +16,23 @@ tags:
|
|
16 |
- llama-3
|
17 |
- pytorch
|
18 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
19 |
|
20 |
# Llama-3.1-Nemotron-Nano-8B-v1
|
21 |
|
|
|
16 |
- llama-3
|
17 |
- pytorch
|
18 |
---
|
19 |
+
<div>
|
20 |
+
<p style="margin-top: 0;margin-bottom: 0;">
|
21 |
+
<em><a href="https://docs.unsloth.ai/basics/unsloth-dynamic-v2.0-gguf">Unsloth Dynamic 2.0</a> achieves superior accuracy & outperforms other leading quants.</em>
|
22 |
+
</p>
|
23 |
+
<div style="display: flex; gap: 5px; align-items: center; ">
|
24 |
+
<a href="https://github.com/unslothai/unsloth/">
|
25 |
+
<img src="https://github.com/unslothai/unsloth/raw/main/images/unsloth%20new%20logo.png" width="133">
|
26 |
+
</a>
|
27 |
+
<a href="https://discord.gg/unsloth">
|
28 |
+
<img src="https://github.com/unslothai/unsloth/raw/main/images/Discord%20button.png" width="173">
|
29 |
+
</a>
|
30 |
+
<a href="https://docs.unsloth.ai/basics/qwen3-how-to-run-and-fine-tune">
|
31 |
+
<img src="https://raw.githubusercontent.com/unslothai/unsloth/refs/heads/main/images/documentation%20green%20button.png" width="143">
|
32 |
+
</a>
|
33 |
+
</div>
|
34 |
+
</div>
|
35 |
+
|
36 |
|
37 |
# Llama-3.1-Nemotron-Nano-8B-v1
|
38 |
|
config.json
CHANGED
@@ -1,5 +1,4 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "nvidia/Llama-3.1-Nemotron-Nano-8B-v1",
|
3 |
"architectures": [
|
4 |
"LlamaForCausalLM"
|
5 |
],
|
@@ -22,6 +21,7 @@
|
|
22 |
"num_attention_heads": 32,
|
23 |
"num_hidden_layers": 32,
|
24 |
"num_key_value_heads": 8,
|
|
|
25 |
"pretraining_tp": 1,
|
26 |
"rms_norm_eps": 1e-05,
|
27 |
"rope_scaling": {
|
@@ -34,7 +34,8 @@
|
|
34 |
"rope_theta": 500000.0,
|
35 |
"tie_word_embeddings": false,
|
36 |
"torch_dtype": "bfloat16",
|
37 |
-
"transformers_version": "4.
|
|
|
38 |
"use_cache": true,
|
39 |
"vocab_size": 128256
|
40 |
}
|
|
|
1 |
{
|
|
|
2 |
"architectures": [
|
3 |
"LlamaForCausalLM"
|
4 |
],
|
|
|
21 |
"num_attention_heads": 32,
|
22 |
"num_hidden_layers": 32,
|
23 |
"num_key_value_heads": 8,
|
24 |
+
"pad_token_id": 128004,
|
25 |
"pretraining_tp": 1,
|
26 |
"rms_norm_eps": 1e-05,
|
27 |
"rope_scaling": {
|
|
|
34 |
"rope_theta": 500000.0,
|
35 |
"tie_word_embeddings": false,
|
36 |
"torch_dtype": "bfloat16",
|
37 |
+
"transformers_version": "4.52.2",
|
38 |
+
"unsloth_fixed": true,
|
39 |
"use_cache": true,
|
40 |
"vocab_size": 128256
|
41 |
}
|