File size: 3,939 Bytes
dc5e166
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
e095e9d
dc5e166
e095e9d
dc5e166
e095e9d
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
dc5e166
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
---
tags:
- merge
- mergekit
- lazymergekit
- KoboldAI/LLaMA2-13B-Tiefighter
- KoboldAI/LLaMA2-13B-Holomax
base_model:
- KoboldAI/LLaMA2-13B-Tiefighter
- KoboldAI/LLaMA2-13B-Holomax
- KoboldAI/LLaMA2-13B-Tiefighter
- KoboldAI/LLaMA2-13B-Holomax
- KoboldAI/LLaMA2-13B-Tiefighter
- KoboldAI/LLaMA2-13B-Holomax
- KoboldAI/LLaMA2-13B-Tiefighter
- KoboldAI/LLaMA2-13B-Holomax
- KoboldAI/LLaMA2-13B-Tiefighter
---
# Unhinged and Happy:

Second experiment in building an unpredicatable model based on TWO creative based LLMs.

After blowing up LazyMergekit in Colab 3 times... here we are.

This merge takes "Layer Theory", "Merge Rules" and throws some of it out the window.

Bottom line if you want an unhinged creative journey... start here.

This merged model is uncensored, freak of nature creative and slightly crazy.

It will break from time to time, and it will not likely win any LLM "testing" awards.

But when it works... oh boy.

It will follow your instructions (mostly) but I suggest lowering temp a bit if it starts going off the rails.

If you are looking for a model for Roleplay, Creative writing, or Entertainment value this is it.

Keep in mind the prose can be slightly "off" at times, which seems to be a by-product of this type of merge.

Tuned (a bit de-hinged?) to be built and tested later.

Context Length: 4096

Please see orginal model cards (Tiefighter and Holomax) for more details below.

# D_AU-Tiefighter-Holomax-15B-UNHINGED-V1

D_AU-Tiefighter-Holomax-15B-UNHINGED-V1 is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
* [KoboldAI/LLaMA2-13B-Tiefighter](https://huggingface.co/KoboldAI/LLaMA2-13B-Tiefighter)
* [KoboldAI/LLaMA2-13B-Holomax](https://huggingface.co/KoboldAI/LLaMA2-13B-Holomax)
* [KoboldAI/LLaMA2-13B-Tiefighter](https://huggingface.co/KoboldAI/LLaMA2-13B-Tiefighter)
* [KoboldAI/LLaMA2-13B-Holomax](https://huggingface.co/KoboldAI/LLaMA2-13B-Holomax)
* [KoboldAI/LLaMA2-13B-Tiefighter](https://huggingface.co/KoboldAI/LLaMA2-13B-Tiefighter)
* [KoboldAI/LLaMA2-13B-Holomax](https://huggingface.co/KoboldAI/LLaMA2-13B-Holomax)
* [KoboldAI/LLaMA2-13B-Tiefighter](https://huggingface.co/KoboldAI/LLaMA2-13B-Tiefighter)
* [KoboldAI/LLaMA2-13B-Holomax](https://huggingface.co/KoboldAI/LLaMA2-13B-Holomax)
* [KoboldAI/LLaMA2-13B-Tiefighter](https://huggingface.co/KoboldAI/LLaMA2-13B-Tiefighter)

## 🧩 Configuration

```yaml
slices:
  - sources:
    - model: KoboldAI/LLaMA2-13B-Tiefighter
      layer_range: [0, 10]
  - sources:
    - model: KoboldAI/LLaMA2-13B-Holomax
      layer_range: [11,15]
  - sources:
    - model: KoboldAI/LLaMA2-13B-Tiefighter
      layer_range: [16,20]
  - sources:
    - model: KoboldAI/LLaMA2-13B-Holomax
      layer_range: [16,22]
  - sources:
    - model: KoboldAI/LLaMA2-13B-Tiefighter
      layer_range: [21, 30]
  - sources:
    - model: KoboldAI/LLaMA2-13B-Holomax
      layer_range: [31,33]
  - sources:
    - model: KoboldAI/LLaMA2-13B-Tiefighter
      layer_range: [31,35]
  - sources:
    - model: KoboldAI/LLaMA2-13B-Holomax
      layer_range: [36,40]
  - sources:
    - model: KoboldAI/LLaMA2-13B-Tiefighter
      layer_range: [36,40]
merge_method: passthrough
dtype: bfloat16
```

## 💻 Usage

```python
!pip install -qU transformers accelerate

from transformers import AutoTokenizer
import transformers
import torch

model = "DavidAU/D_AU-Tiefighter-Holomax-20B-V1"
messages = [{"role": "user", "content": "What is a large language model?"}]

tokenizer = AutoTokenizer.from_pretrained(model)
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
pipeline = transformers.pipeline(
    "text-generation",
    model=model,
    torch_dtype=torch.float16,
    device_map="auto",
)

outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
print(outputs[0]["generated_text"])
```