File size: 3,226 Bytes
51d9458
 
 
6459a5c
51d9458
6459a5c
 
 
51d9458
 
6459a5c
 
 
51d9458
 
 
 
04691e0
51d9458
6459a5c
51d9458
 
 
 
 
 
 
 
 
 
 
6459a5c
51d9458
6459a5c
 
 
51d9458
6459a5c
 
51d9458
 
 
 
 
 
6459a5c
51d9458
 
 
 
 
 
 
 
 
 
6459a5c
 
 
51d9458
 
6459a5c
 
 
04691e0
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
---
base_model:
- sthenno/tempesthenno-ppo-ckpt40
- sometimesanotion/LoRA-64-Chocolatine-2-14B-Instruct-v2.0b3
- jpacifico/Chocolatine-2-14B-Instruct-v2.0b3
- huihui-ai/DeepSeek-R1-Distill-Qwen-14B-abliterated-v2
- EVA-UNIT-01/EVA-Qwen2.5-14B-v0.2
- underwoods/medius-erebus-magnum-14b
- sometimesanotion/Qwenvergence-14B-v3-Prose
- sometimesanotion/LoRA-64-Chocolatine-2-14B-Instruct-v2.0b3
- oxyapi/oxy-1-small
- allura-org/TQ2.5-14B-Sugarquill-v1
- sometimesanotion/Base-Chocolatine-2-14B-Instruct-v2.0b3
library_name: transformers
tags:
- mergekit
- merge
new_version: sometimesanotion/Qwenvergence-14B-v13-Prose-DS
---
# merge

This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).

## Merge Details
### Merge Method

This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [sometimesanotion/Base-Chocolatine-2-14B-Instruct-v2.0b3](https://huggingface.co/sometimesanotion/Base-Chocolatine-2-14B-Instruct-v2.0b3) as a base.

### Models Merged

The following models were included in the merge:
* [sthenno/tempesthenno-ppo-ckpt40](https://huggingface.co/sthenno/tempesthenno-ppo-ckpt40) + [sometimesanotion/LoRA-64-Chocolatine-2-14B-Instruct-v2.0b3](https://huggingface.co/sometimesanotion/LoRA-64-Chocolatine-2-14B-Instruct-v2.0b3)
* [jpacifico/Chocolatine-2-14B-Instruct-v2.0b3](https://huggingface.co/jpacifico/Chocolatine-2-14B-Instruct-v2.0b3)
* [huihui-ai/DeepSeek-R1-Distill-Qwen-14B-abliterated-v2](https://huggingface.co/huihui-ai/DeepSeek-R1-Distill-Qwen-14B-abliterated-v2)
* [EVA-UNIT-01/EVA-Qwen2.5-14B-v0.2](https://huggingface.co/EVA-UNIT-01/EVA-Qwen2.5-14B-v0.2)
* [underwoods/medius-erebus-magnum-14b](https://huggingface.co/underwoods/medius-erebus-magnum-14b)
* [sometimesanotion/Qwenvergence-14B-v3-Prose](https://huggingface.co/sometimesanotion/Qwenvergence-14B-v3-Prose) + [sometimesanotion/LoRA-64-Chocolatine-2-14B-Instruct-v2.0b3](https://huggingface.co/sometimesanotion/LoRA-64-Chocolatine-2-14B-Instruct-v2.0b3)
* [oxyapi/oxy-1-small](https://huggingface.co/oxyapi/oxy-1-small)
* [allura-org/TQ2.5-14B-Sugarquill-v1](https://huggingface.co/allura-org/TQ2.5-14B-Sugarquill-v1)

### Configuration

The following YAML configuration was used to produce this model:

```yaml
name:                Qwenvergence-14B-v12-Prose-DS
merge_method:        model_stock
base_model:          sometimesanotion/Base-Chocolatine-2-14B-Instruct-v2.0b3
tokenizer_source:    base
dtype:               float32
out_dtype:           bfloat16
parameters:
  int8_mask:         true
  normalize:         true
  rescale:           false
models:
  - model:           EVA-UNIT-01/EVA-Qwen2.5-14B-v0.2
  - model:           oxyapi/oxy-1-small
  - model:           allura-org/TQ2.5-14B-Sugarquill-v1
  - model:           jpacifico/Chocolatine-2-14B-Instruct-v2.0b3
  - model:           sometimesanotion/Qwenvergence-14B-v3-Prose+sometimesanotion/LoRA-64-Chocolatine-2-14B-Instruct-v2.0b3
  - model:           underwoods/medius-erebus-magnum-14b
  - model:           sthenno/tempesthenno-ppo-ckpt40+sometimesanotion/LoRA-64-Chocolatine-2-14B-Instruct-v2.0b3
  - model:           huihui-ai/DeepSeek-R1-Distill-Qwen-14B-abliterated-v2
```