|
--- |
|
license: cc-by-nc-4.0 |
|
tags: |
|
- not-for-all-audiences |
|
- roleplay |
|
- merge |
|
- nsfw |
|
--- |
|
# Arconte-13B |
|
|
|
Arconte is Llama-2 merge. Arconte has many iterations, trying different recipes/models/merge-methods, in particular, iteration I and iteration Z are both models which showed promise. This version of Arconte is variation I redone with a more experimental approach to the merge recipe, and it shows great results. |
|
|
|
Originally, the idea was to do one of those fancy Dare Ties model A, Dare Ties model B, Slerp model A + model B. I already did the Slerp model C, but it's flawed due to the flawed iterations I and Z. I still plan to do model C, so now I am remaking iteration Z. |
|
|
|
Models used: |
|
|
|
[NeverSleep/X-NoroChronos-13B](https://huggingface.co/NeverSleep/X-NoroChronos-13B) |
|
|
|
[Undi95/Emerhyst-13B](https://huggingface.co/Undi95/Emerhyst-13B) |
|
|
|
[Henk717/echidna-tiefigther-25](https://huggingface.co/Henk717/echidna-tiefigther-25) |
|
|
|
After completing model C, current roadmap is to either go into mistral merges, or trying my hand at making loras/qloras. No mixtral, nor anything above 13B parameters in the future due to hardware limitations. |
|
|
|
All testing was done with Q5_K_M GUFF. I'll upload the full GUFF range along with an Imatrix version soon. |
|
|
|
# Update 3/30/24 |
|
|
|
I have tested this model further and I concluded that I find it boring. I remember I greenlighted this model because it was coherent (as much as a Q5_K_M can be), but now I think it's just not that good. But perhaps it is just my taste in models? or maybe my sampling settings are bad? I would like some feedback to know how good or bad this model is. I still plan to cook that C model, but I don't know if I will use this one to do it. |
|
|
|
I will be releasing another model soon, an older model that I think is better than this one. |
|
|