File size: 2,576 Bytes
0999c3b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
---
base_model:
- nvidia/Llama-3_3-Nemotron-Super-49B-v1
---
# Join our Discord! https://discord.gg/Nbv9pQ88Xb
## More than 5500 members of helpful, LLM enthusiasts! A hub for players and makers alike!
### We need testers!
---

Drummer proudly presents...

#  Valkyrie 49B v1

![image/png](https://cdn-uploads.huggingface.co/production/uploads/65f2fd1c25b848bd061b5c2e/8I-AvB0bFSoEcxlLU7dtY.png)

## Description

> it swears unprompted 10/10 model

> ... characters work well, groups work well, scenarios also work really well so great model overall

> This is pretty exciting though. GLM-4 already had me on the verge of deleting all of my other 32b and lower models. I got to test this more but I think this model at Q3m is the death blow lol

> Smart Nemotron 49b learned how to roleplay

> Even without thinking it rock solid at 4qm.

> Without thinking is like 40-70b level. With thinking is 100+b level

> This model would have been AGI if it were named properly with a name like "Bob". Alas, it was not.

> I think this model is nice. It follows prompts very well. I didn't really note any major issues or repetition

> Yeah this is good. I think its clearly smart enough, close to the other L3.3 70b models. It follows directions and formatting very well. I asked it to create the intro message, my first response was formatted differently, and it immediately followed my format on the second message. I also have max tokens at 2k cause I like the model to finish it's thought. But I started trimming the models responses when I felt the last bit was unnecessary and it started replying closer to that length. It's pretty much uncensored.

> [Nemotron is my favorite model, and I think you fixed it!!](https://huggingface.co/TheDrummer/Valkyrie-49B-v1-GGUF/discussions/1)

## Usage
- Llama 3 Chat Template
- \<think\> capable upon prefill or `detailed thinking on` on top of the system prompt

## Links
- Original: https://huggingface.co/TheDrummer/Valkyrie-49B-v1
- GGUF: https://huggingface.co/TheDrummer/Valkyrie-49B-v1-GGUF
- iMatrix (recommended): https://huggingface.co/bartowski/TheDrummer_Valkyrie-49B-v1-GGUF
- EXL3: https://huggingface.co/collections/ReadyArt/thedrummer-valkyrie-49b-v1-exl3-682b634353aba9f40d906185

## Special Thanks
- Thank you to the testers at BeaverAI! You da MVP!
- Thank you to each and everyone who donated and subscribed in [Patreon](https://www.patreon.com/TheDrummer) and [Ko-Fi](https://ko-fi.com/thedrummer) to make our venture a little bit easier.
- [Subscribe to my Patreon!](https://www.patreon.com/TheDrummer)

`config-v1a`