datasets:
- Delta-Vector/Hydrus-Instruct-SmolTalk-V2
- Delta-Vector/Hydrus-SonnetOrca-V2
- Delta-Vector/Hydrus-FeedSum-ShareGPT
- Delta-Vector/Hydrus-Tulu-Personas-Filtered-Sharegpt
- Delta-Vector/Hydrus-No_Robots-R1-Filtered
- Delta-Vector/Hydrus-Chat_error-Pure-Dove-sharegpt
- Delta-Vector/Hydrus-HelpSteer2
- Delta-Vector/Hydrus-R1-Thinking-Sharegpt
- Delta-Vector/Hydrus-Science-QA-sharegpt
- Delta-Vector/Hydrus-Claude-Instruct-2.7K
- Delta-Vector/Hydrus-Claude-Instruct-5K
- PocketDoc/Dans-Assistantmaxx-UnnaturalInstructions-GPT4
- PocketDoc/Dans-Toolmaxx-ShellCommands
- PocketDoc/Dans-MemoryCore-CoreCurriculum-Small
- PocketDoc/Dans-Logicmaxx-SAT-AP
- PocketDoc/Dans-Benchmaxx
- Nitral-AI/ARES-ShareGPT
- PocketDoc/Dans-Taskmaxx-TableGPT
- Delta-Vector/Ursa-Erebus-16K
- Delta-Vector/Ursa-Books-Light-Novels-V1
- NewEden/Orion-LIT
- Delta-Vector/Ursa-Asstr-V2-18k
- Delta-Vector/Ursa-Books-V2
- Delta-Vector/Ursa-Scribblehub-7k
- Delta-Vector/Ursa-Orion-EA-Comp-Filtered
- Delta-Vector/Ursa-HoneyFeed
- Delta-Vector/Ursa-Falling-through-the-world
base_model:
- Delta-Vector/Sol-Reaver-15B-Instruct
base_model_relation: quantized
quantized_by: ArtusDev
tags:
- roleplay
- instruct
- creative_writing
- story-writing
- mistral
- exl3
Sol Reaver 15B

Model Information
Sol-Reaver-15B-Instruct
The first in the line of a New series of Roleplay / Adventure / Co-writer Models - Finetuned ontop of Sol-Reaver-15B-Pretrain
This model has been trained on 200M tokens of high quality Instruct data, It's focus is to provide a base for further finetuning|Merging
It's goal is to have refreshing Prose, Creativity, Good Instruct following and the *Brains*.
Support me on Ko-Fi: https://ko-fi.com/deltavector
Quantized Versions
Available Downloads
- GGUF FormatFor use with LLama.cpp & Forks(Coming Soon!)
- EXL2 FormatFor use with TabbyAPI (Coming Soon!)
- EXL3 FormatFor use with TabbyAPI (Slower on Ampere))
Prompting
Model has been tuned with the ChatML formatting. A typical input would look like this:
<|im_start|>user
Hi there!<|im_end|>
<|im_start|>assistant
Nice to meet you!<|im_end|>
<|im_start|>user
Can I ask a question?<|im_end|>
<|im_start|>assistant
Samplers
For testing of this model, I used Temp=1, 0.1 Min-P.
See Axolotl Config
https://files.catbox.moe/u9dakg.yml
Training
The training was done for 2 epoch using 8 x H200s GPUs graciously provided by Kalomaze for the fine-tuning of the model.
Credits
Thank you to Lucy Knada, Ateron, Alicat, Intervitens, Cgato, Kubernetes Bad and the rest of Anthracite.