Sweaterdog commited on
Commit
346ad16
·
verified ·
0 Parent(s):

Duplicate from Sweaterdog/MindCraft-LLM-tuning

Browse files
.gitattributes ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
+ unsloth.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
38
+ unsloth.F16.gguf filter=lfs diff=lfs merge=lfs -text
39
+ Hermes-v1(qwen).Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
40
+ Hermes-v1(qwen).F16.gguf filter=lfs diff=lfs merge=lfs -text
41
+ unsloth.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
42
+ unsloth.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
43
+ unsloth.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
44
+ Andy-v2-llama.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
45
+ Andy-v2-llama.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
46
+ Andy-v2-llama.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
47
+ Andy-v2-llama.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
48
+ Andy-v2-qwen.F16.gguf filter=lfs diff=lfs merge=lfs -text
49
+ Andy-v2-qwen.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
50
+ Andy-v2-qwen.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
51
+ Andy-v2-qwen.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
52
+ Hermes-v1(qwen)-(deprecated).F16.gguf filter=lfs diff=lfs merge=lfs -text
53
+ Hermes-v1(qwen)-(deprecated).Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
54
+ Andy-v2-qwen.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
55
+ Andy-v3-llama.F16.gguf filter=lfs diff=lfs merge=lfs -text
56
+ Andy-v3-llama.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
57
+ Andy-v3-llama.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
58
+ Andy-v3-llama.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
59
+ Andy-v3-llama.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
60
+ deprecated-models/Andy-v2-llama.F16.gguf filter=lfs diff=lfs merge=lfs -text
61
+ Andy-v3-qwen.F16.gguf filter=lfs diff=lfs merge=lfs -text
62
+ Andy-v3-qwen.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
63
+ Andy-v3-qwen.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
64
+ Andy-v3-qwen.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
65
+ Andy-v3-qwen.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
66
+ Andy-3.5-beta.F16.gguf filter=lfs diff=lfs merge=lfs -text
Andy-3.5-beta.F16.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4fe9f303717f606050093411c103c7bc9336ba63879efc622350b0e301b94713
3
+ size 15237851168
Andy-v2-qwen.F16.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:29db043f30a0650f1976bcaf60ec559a90dae594b747556c7c073e74f55551eb
3
+ size 15237851168
Andy-v2-qwen.Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1268d05080abdd55e3cc307c9a467520e5a03dbed81d8fd3ea4e50d73cdd1dcf
3
+ size 3015938080
Andy-v2-qwen.Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8283e32ca75cf803f79bc777cdb4b6750460a98cd95b3423a425d03dadf4669b
3
+ size 4683071520
Andy-v2-qwen.Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c626afc0731be9dc3a4583b80634e78d77275b68ead2c9001cc83e71693b058
3
+ size 5444829216
Andy-v2-qwen.Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db2f55958e065de12f7a73dab246ecbaa5b0dcd6ed3fc20f758be39220dd50ab
3
+ size 8098523168
Andy-v3-llama.F16.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:560fa7270b61306b4091c201ec74bc633678b90254b5f68c628e2577246aa9f3
3
+ size 6433683744
Andy-v3-llama.Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02032272eb3e1421ab4ac73ba79e3660f4d735c3f62dc0c3ef69f48105f875cf
3
+ size 1363931424
Andy-v3-llama.Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b66a8dd197d0ec4ba4179651fce80854700613e449d3bfddf8caf871672200ae
3
+ size 2019373344
Andy-v3-llama.Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f568c16fdd3377a7ad7ed68a722ab0c8f81710fc9fa5103df1c5f604d487fc74
3
+ size 2322149664
Andy-v3-llama.Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df74d06b3ed4855178d8830080349413748d1eba09e6d01c643221311ad65995
3
+ size 3421894944
README.md ADDED
@@ -0,0 +1,116 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - unsloth/Qwen2.5-7B-bnb-4bit
4
+ - unsloth/Llama-3.2-3B-Instruct
5
+ tags:
6
+ - text-generation-inference
7
+ - transformers
8
+ - unsloth
9
+ - qwen2
10
+ - llama3
11
+ - trl
12
+ license: apache-2.0
13
+ language:
14
+ - en
15
+ datasets:
16
+ - Sweaterdog/MindCraft-LLM-tuning
17
+ ---
18
+
19
+ # Uploaded models
20
+
21
+ - **Developed by:** Sweaterdog
22
+ - **License:** apache-2.0
23
+ - **Finetuned from model :** unsloth/Qwen2.5-7B-bnb-4bit and unsloth/Llama-3.2-3B-Instruct
24
+
25
+ The MindCraft LLM tuning CSV file can be found here, this can be tweaked as needed. [MindCraft-LLM](https://huggingface.co/datasets/Sweaterdog/MindCraft-LLM-tuning)
26
+
27
+ # What is the Purpose?
28
+
29
+ This model is built and designed to play Minecraft via the extension named "[MindCraft](https://github.com/kolbytn/mindcraft)" Which allows language models, like the ones provided in the files section, to play Minecraft.
30
+ - Why a new model?
31
+ #
32
+ While, yes, models that aren't fine tuned to play Minecraft *Can* play Minecraft, most are slow, innaccurate, and not as smart, in the fine tuning, it expands reasoning, conversation examples, and command (tool) usage.
33
+ - What kind of Dataset was used?
34
+ #
35
+ I'm deeming the first generation of this model, Hermesv1, for future generations, they will be named ***"Andy"*** based from the actual MindCraft plugin's default character. it was trained for reasoning by using examples of in-game "Vision" as well as examples of spatial reasoning, for expanding thinking, I also added puzzle examples where the model broke down the process step by step to reach the goal.
36
+ - Why choose Qwen2.5 for the base model?
37
+ #
38
+ During testing, to find the best local LLM for playing Minecraft, I came across two, Gemma 2, and Qwen2.5, these two were by far the best at playing Minecraft before fine-tuning, and I knew, once tuned, it would become better.
39
+ - If Gemma 2 and Qwen 2.5 are the best before fine tuning, why include Llama 3.2, especially the lower intelligence, 3B parameter version?
40
+ #
41
+ That is a great question, I know since Llama 3.2 3b has low amounts of parameters, it is dumb, and doesn't play minecraft well without fine tuning, but, it is a lot smaller than other models which are for people with less powerful computers, and the hope is, once the model is tuned, it will become much better at minecraft.
42
+
43
+ - Why is it taking so long to release more tuned models?
44
+ #
45
+ Well, you see, I do not have the most powerful computer, and Unsloth, the thing I'm using for fine tuning, has a google colab set up, so I am waiting for GPU time to tune the models, but they will be released ASAP, I promise.
46
+
47
+ - Will there ever be vision fine tuning?
48
+ #
49
+ Yes! In MindCraft there will be vision support for VLM's *(vision language models)*, Most likely, the model will be Qwen2-VL-7b, or LLaMa3.2-11b-vision since they are relatively new, yes, I am still holding out hope for llama3.2
50
+
51
+ # How to Use
52
+ In order to use this model, A, download the GGUF file of the version you want, either a Qwen, or Llama model, and then the Modelfile, after you download both, in the Modelfile, change the directory of the model, to your model. Here is a simple guide if needed for the rest:
53
+ #
54
+ 1. Download the .gguf Model u want. For this example it is in the standard Windows "Download" Folder
55
+
56
+ 2. Download the Modelfile
57
+
58
+ 3. Open the Modelfile with / in notepad, or you can rename it to Modelfile.txt, and change the GGUF path, for example, this is my PATH "C:\Users\SweaterDog\OneDrive\Documents\Raw GGUF Files\Hermes-1.0\Hermes-1.Q8_0.gguf"
59
+
60
+ 4. Safe + Close Modelfile
61
+
62
+ 5. Rename "Modelfile.txt" into "Modelfile" if you changed it before-hand
63
+
64
+ 6. Open CMD and type in "ollama create Hermes1 -f Modelfile" (You can change the name to anything you'd like, for this example, I am just using the same name as the GGUF)
65
+
66
+ 7. Wait until finished
67
+
68
+ 8. In the CMD window, type "ollama run Hermes1" (replace the name with whatever you called it)
69
+
70
+ 9. (Optional, needed for versions after the 11/15/24 update) If you downloaded a model that was tuned from Qwen, and in the model name you kept Qwen, you need to go into the file "prompter.js" and remove the qwen section, if you named it something that doesn't include qwen in the name, you can skip this step.
71
+
72
+ # How to fine tune a Gemini Model
73
+ 1. Download the CSV for [MindCraft-LLM-tuning](https://huggingface.co/datasets/Sweaterdog/MindCraft-LLM-tuning)
74
+ 2. Open sheet.google.com, and upload the CSV file
75
+ 3. Go to [API keys and Services](https://aistudio.google.com/app/apikey), then click on "New Tuned Model" on the left popup bar
76
+ 4. Press "Import" and then select the CSV file you uploaded to google sheets
77
+ 5. Rename the model to whatever you want, set the training settings, epochs, learning rate, and batch size
78
+ 6. Change the model to either Gemini-1.0-pro or Gemini-1.5-flash **NOTE** Gemini 1.0 pro will be deprecated on February 15, 2025, meaning the model WILL BE deleted!
79
+ 7. Hit tune and wait.
80
+ 8. After the model is finished training, hit "Add API access" and select the google project you'd like to connect it to
81
+ 9. Copy the model ID, and paste it into the Gemini.json file in MindCraft, then name the model to whatever you want.
82
+ 10. (Optional) Test the model by pressing "Use in chat" and ask it basic actions, such as "Grapevine_eater: Come here!" and see the output, if it is not to your liking, train the model again with different settings,
83
+ 11. (Optional) Since the rates for Gemini models are limited (If you do not have billing enabled) I recommend making a launch.bat file in the MindCraft folder, instead of crashing and having you need to manually start the program every time the rate limit is reached. Here is the code I use in launch.bat
84
+ ```
85
+ @echo off
86
+ setlocal enabledelayedexpansion
87
+
88
+ :loop
89
+ node main.js
90
+ timeout /t 10 /nobreak
91
+
92
+ echo Restarting...
93
+ goto loop
94
+ ```
95
+ 12. Enjoy having a model play Minecraft with you, hopefully it is smarter than regular Gemini models!
96
+ #
97
+
98
+ **WARNING** The new v3 generation of models suck! That is because they were also trained for building *(coding)* and often do not use commands! I recommend using the v2 generation still, the LLaMa version is in the [deprecated models folder](https://huggingface.co/Sweaterdog/MindCraft-LLM-tuning/tree/main/deprecated-models).
99
+
100
+ #
101
+
102
+ For Anybody who is wondering what the context length is, the Qwen version, has a length of 64000 tokens, for the Llama version, it has a 128000 token context window. *(***NOTE*** Any model can support a longer context, but these are the supported values in training)*
103
+
104
+ #
105
+
106
+ I wanted to include the google colab link, in case you wanted to know how to train models via CSV, or use my dataset to train your own model, on your own settings, on a different model. [Google Colab](https://colab.research.google.com/drive/1VYkncZMfGFkeCEgN2IzbZIKEDkyQuJAS#scrollTo=2eSvM9zX_2d3)
107
+
108
+ #
109
+
110
+ **UPDATE** The Qwen and Llama models are out, with the expanded dataset! I have found the llama models are incredibly dumb, but changing the Modelfile may provide better results, With the Qwen version of Andy, the Q4_K_M, it took 2 minutes to craft a wooden pickaxe, collected stone after that, took 5 minutes.
111
+
112
+ #
113
+
114
+ This qwen2 and llama3.2 model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
115
+
116
+ [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "model_type": "qwen2"
3
+ }
deprecated-models/Andy-v2-llama.F16.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70d7076b5e3841d4e0bf378e23ddfe2d9219a075e1043234be8c1040cf035d52
3
+ size 6433687648
deprecated-models/Andy-v2-llama.Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0726b826ca3413968bcf7dd978b2b383dbd6c083ee50cb90cfb804808c3aa713
3
+ size 1363935328
deprecated-models/Andy-v2-llama.Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a81b6d19a9897e9fef012438045f22da0c301e1444f824d64949a6eacc807ca3
3
+ size 2019377248
deprecated-models/Andy-v2-llama.Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d82b1bda997782f6dd78096987f2f637b1740866adcb6ddaf32f690da5d751b2
3
+ size 2322153568
deprecated-models/Andy-v2-llama.Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae67678cf945763fee2b4d46123b5b8d84e0612e1c7cfc438d461b7ec61ae297
3
+ size 3421898848
deprecated-models/Andy-v3-qwen.F16.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a570efbab566dd5f768d2dc0d68c5940111811b558607b04cfdd79375be3a477
3
+ size 15237850592
deprecated-models/Andy-v3-qwen.Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9185c8d4bc4376a53c652b458187cabc7658463c318bdad870cf1996995ee3bd
3
+ size 3015937504
deprecated-models/Andy-v3-qwen.Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0edec17f27d36baeda8965ab006bbf9d211eb04504cab3e99620ee374a41890a
3
+ size 4683070944
deprecated-models/Andy-v3-qwen.Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a48bf7b7069075908766f84a84ae1347d082af18325c844c3a296cda67013eb
3
+ size 5444828640
deprecated-models/Andy-v3-qwen.Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:300dc49708437a1cefdf1697f22f534761c740b9fb0d9d004cd98bc0f345f994
3
+ size 8098522592
deprecated-models/Hermes-v1(qwen)-(deprecated).F16.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3a6035859851cc54bfcb25bd8295fe7785370cb596385cb83a14ce798e2ba8c
3
+ size 7816423456
deprecated-models/Hermes-v1(qwen)-(deprecated).Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d3f2c5af4c021cadb5259c1dc56102def07b5062b2e85d68d14d64e730e48fc
3
+ size 8098523168
extras/Modelfile ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ FROM /content/Sweaterdog/MindCraft-LLM-tuning/unsloth.F16.gguf
2
+
3
+ TEMPLATE """Below is an input from the system or yourself, no matter what, DO NOT use the same command with the same syntax 3 messages in a row.
4
+ Prompt or Next action:{{ if .Prompt }}
5
+ {{ .Prompt }}{{ end }}
6
+ What do you do:
7
+ {{ .Response }}<|endoftext|>"""
8
+
9
+ PARAMETER stop "<|object_ref_end|>"
10
+ PARAMETER stop "<|video_pad|>"
11
+ PARAMETER stop "<|im_end|>"
12
+ PARAMETER stop "<|file_sep|>"
13
+ PARAMETER stop "<|quad_end|>"
14
+ PARAMETER stop "<|repo_name|>"
15
+ PARAMETER stop "<|object_ref_start|>"
16
+ PARAMETER stop "<|box_start|>"
17
+ PARAMETER stop "</tool_call>"
18
+ PARAMETER stop "<|box_end|>"
19
+ PARAMETER stop "<|im_start|>"
20
+ PARAMETER stop "<tool_call>"
21
+ PARAMETER stop "<|PAD_TOKEN|>"
22
+ PARAMETER stop "<|image_pad|>"
23
+ PARAMETER stop "<|quad_start|>"
24
+ PARAMETER stop "<|endoftext|>"
25
+ PARAMETER stop "<|vision_"
26
+ PARAMETER stop "<|fim_"
27
+ PARAMETER temperature 0.9
28
+ PARAMETER min_p 0.8
extras/Modelfile [EXPERIMENTAL] ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ FROM "YOUR/PATH/HERE"
2
+
3
+ TEMPLATE """{{- if .Messages }}
4
+ {{- if or .System .Tools }}<|im_start|>system
5
+ {{- if .System }}
6
+ {{ .System }}
7
+ {{- end }}
8
+ {{- if .Tools }}
9
+
10
+ For each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:
11
+ <tool_call>
12
+ {"name": <function-name>, "arguments": <args-json-object>}
13
+ </tool_call>
14
+ {{- end }}<|im_end|>
15
+ {{ end }}
16
+ {{- range $i, $_ := .Messages }}
17
+ {{- $last := eq (len (slice $.Messages $i)) 1 -}}
18
+ {{- if eq .Role "user" }}<|im_start|>user
19
+ {{ .Content }}<|im_end|>
20
+ {{ else if eq .Role "assistant" }}<|im_start|>assistant
21
+ {{ if .Content }}{{ .Content }}
22
+ {{- else if .ToolCalls }}<tool_call>
23
+ {{ range .ToolCalls }}{"name": "{{ .Function.Name }}", "arguments": {{ .Function.Arguments }}}
24
+ {{ end }}</tool_call>
25
+ {{- end }}{{ if not $last }}<|im_end|>
26
+ {{ end }}
27
+ {{- else if eq .Role "tool" }}<|im_start|>user
28
+ <tool_response>
29
+ {{ .Content }}
30
+ </tool_response><|im_end|>
31
+ {{ end }}
32
+ {{- if and (ne .Role "assistant") $last }}<|im_start|>assistant
33
+ {{ end }}
34
+ {{- end }}
35
+ {{- else }}
36
+ {{- if .System }}<|im_start|>system
37
+ {{ .System }}<|im_end|>
38
+ {{ end }}{{ if .Prompt }}<|im_start|>user
39
+ {{ .Prompt }}<|im_end|>
40
+ {{ end }}<|im_start|>assistant
41
+ {{ end }}{{ .Response }}{{ if .Response }}<|im_end|>{{ end }}"""
42
+ SYSTEM You are Andy, a fine tuned model for playing Minecraft, made by SweaterDog
extras/adapter_config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "unsloth/Qwen2.5-7B-Instruct-bnb-4bit",
5
+ "bias": "none",
6
+ "fan_in_fan_out": false,
7
+ "inference_mode": true,
8
+ "init_lora_weights": true,
9
+ "layer_replication": null,
10
+ "layers_pattern": null,
11
+ "layers_to_transform": null,
12
+ "loftq_config": {},
13
+ "lora_alpha": 16,
14
+ "lora_dropout": 0,
15
+ "megatron_config": null,
16
+ "megatron_core": "megatron.core",
17
+ "modules_to_save": null,
18
+ "peft_type": "LORA",
19
+ "r": 16,
20
+ "rank_pattern": {},
21
+ "revision": null,
22
+ "target_modules": [
23
+ "o_proj",
24
+ "down_proj",
25
+ "gate_proj",
26
+ "q_proj",
27
+ "k_proj",
28
+ "up_proj",
29
+ "v_proj"
30
+ ],
31
+ "task_type": "CAUSAL_LM",
32
+ "use_dora": false,
33
+ "use_rslora": false
34
+ }
extras/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0546a5fe725500c4e87dbca874c5a0212435693a779186ed7bc61d264e4de148
3
+ size 161533192
extras/added_tokens.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
+ "<|PAD_TOKEN|>": 151665,
5
+ "<|box_end|>": 151649,
6
+ "<|box_start|>": 151648,
7
+ "<|endoftext|>": 151643,
8
+ "<|file_sep|>": 151664,
9
+ "<|fim_middle|>": 151660,
10
+ "<|fim_pad|>": 151662,
11
+ "<|fim_prefix|>": 151659,
12
+ "<|fim_suffix|>": 151661,
13
+ "<|im_end|>": 151645,
14
+ "<|im_start|>": 151644,
15
+ "<|image_pad|>": 151655,
16
+ "<|object_ref_end|>": 151647,
17
+ "<|object_ref_start|>": 151646,
18
+ "<|quad_end|>": 151651,
19
+ "<|quad_start|>": 151650,
20
+ "<|repo_name|>": 151663,
21
+ "<|video_pad|>": 151656,
22
+ "<|vision_end|>": 151653,
23
+ "<|vision_pad|>": 151654,
24
+ "<|vision_start|>": 151652
25
+ }
extras/config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "model_type": "llama"
3
+ }
extras/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
extras/special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|im_end|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|PAD_TOKEN|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
extras/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fab42efe8d17406525a9154b728cf9e957629a8ed7ce997770efdd71128c6a1a
3
+ size 11422086
extras/tokenizer_config.json ADDED
@@ -0,0 +1,216 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ },
181
+ "151665": {
182
+ "content": "<|PAD_TOKEN|>",
183
+ "lstrip": false,
184
+ "normalized": false,
185
+ "rstrip": false,
186
+ "single_word": false,
187
+ "special": true
188
+ }
189
+ },
190
+ "additional_special_tokens": [
191
+ "<|im_start|>",
192
+ "<|im_end|>",
193
+ "<|object_ref_start|>",
194
+ "<|object_ref_end|>",
195
+ "<|box_start|>",
196
+ "<|box_end|>",
197
+ "<|quad_start|>",
198
+ "<|quad_end|>",
199
+ "<|vision_start|>",
200
+ "<|vision_end|>",
201
+ "<|vision_pad|>",
202
+ "<|image_pad|>",
203
+ "<|video_pad|>"
204
+ ],
205
+ "bos_token": null,
206
+ "chat_template": "{{ 'Below is an input from the system or yourself, no matter what, will play minecraft, and reward yourself every time you successfully finish an action, DO NOT use the same command with the same syntax 3 messages in a row. You are an Ai named Hermes, built for playing Minecraft.\nPrompt or Next action:' }}{% for message in messages %}{% if message['role'] == 'user' %}{{ '\n' + message['content'] }}{% elif message['role'] == 'assistant' %}{{ '\nWhat do you do:\n' + message['content'] + '<|im_end|>' }}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '\nWhat do you do:\n' }}{% endif %}",
207
+ "clean_up_tokenization_spaces": false,
208
+ "eos_token": "<|im_end|>",
209
+ "errors": "replace",
210
+ "model_max_length": 128000,
211
+ "pad_token": "<|PAD_TOKEN|>",
212
+ "padding_side": "left",
213
+ "split_special_tokens": false,
214
+ "tokenizer_class": "Qwen2Tokenizer",
215
+ "unk_token": null
216
+ }
extras/vocab.json ADDED
The diff for this file is too large to render. See raw diff