John Leimgruber III
ubergarm
AI & ML interests
Open LLMs and Astrophotography image processing.
Recent Activity
new activity
5 days ago
unsloth/DeepSeek-R1-GGUF:No think tokens visible
upvoted
an
article
6 days ago
Open-R1: a fully open reproduction of DeepSeek-R1
Organizations
None yet
ubergarm's activity
No think tokens visible
4
#15 opened 6 days ago
by
sudkamath
Over 2 tok/sec agg backed by NVMe SSD on 96GB RAM + 24GB VRAM AM5 rig with llama.cpp
9
#13 opened 7 days ago
by
ubergarm
Got it running after downloading some RAM!
4
#7 opened 8 days ago
by
ubergarm
Over 128k context on 1x 3090 TI FE 24GB VRAM!
#1 opened 7 days ago
by
ubergarm
Inference speed
2
#9 opened 8 days ago
by
Iker
Control over output
1
#12 opened 10 days ago
by
TeachableMachine
Emotions
2
#3 opened 11 days ago
by
jujutechnology
What advantage does this have over normal algorithmic ways of turning HTML to Markdown ?
5
#5 opened 20 days ago
by
MohamedRashad
FuseO1-DeekSeekR1-QwQ-SkyT1-32B-Preview
2
#1 opened 16 days ago
by
AaronFeng753
System Prompt
17
#2 opened 17 days ago
by
Wanfq
FIXED: Error with llama-server `unknown pre-tokenizer type: 'deepseek-r1-qwen'`
4
#1 opened 16 days ago
by
ubergarm
The `tokenizer_config.json` is missing the `chat_template` jinja?
1
#1 opened 16 days ago
by
ubergarm
Great RP model in only 12B! A few notes and sampler settings for llama.cpp server inside.
2
#2 opened 20 days ago
by
ubergarm
Nice ~90x real-time generation on 3090TI. Quickstart provided.
5
#20 opened about 1 month ago
by
ubergarm
Observation: 4-bit quantization can't answer the Strawberry prompt
12
#2 opened 4 months ago
by
ThePabli
63.17 MMLU-Pro Computer Science with `Q8_0`
#2 opened 4 months ago
by
ubergarm
Benchmarks worse than Qwen2.5-7B-Instruct on MMLU-Pro Computer Science in limited testing.
#1 opened 4 months ago
by
ubergarm
Promising looking results on 24GB VRAM folks!
9
#3 opened 5 months ago
by
ubergarm
Awesome model
6
#5 opened 5 months ago
by
dillfrescott
vram usage of each?
3
#1 opened 5 months ago
by
jasonden