John Leimgruber III
ubergarm
AI & ML interests
Open LLMs and Astrophotography image processing.
Recent Activity
new activity
about 3 hours ago
unsloth/DeepSeek-R1-GGUF:Over 2 tok/sec agg backed by NVMe SSD on 96GB RAM + 24GB VRAM AM5 rig with llama.cpp
new activity
about 21 hours ago
unsloth/DeepSeek-R1-GGUF:Got it running after downloading some RAM!
new activity
1 day ago
mradermacher/Qwen2.5-14B-Instruct-1M-i1-GGUF:Over 128k context on 1x 3090 TI FE 24GB VRAM!
Organizations
None yet
ubergarm's activity
Over 2 tok/sec agg backed by NVMe SSD on 96GB RAM + 24GB VRAM AM5 rig with llama.cpp
3
#13 opened about 21 hours ago
by
ubergarm
Got it running after downloading some RAM!
4
#7 opened 2 days ago
by
ubergarm
Over 128k context on 1x 3090 TI FE 24GB VRAM!
#1 opened 1 day ago
by
ubergarm
Inference speed
2
#9 opened 2 days ago
by
Iker
Control over output
1
#12 opened 3 days ago
by
TeachableMachine
Emotions
2
#3 opened 5 days ago
by
jujutechnology
What advantage does this have over normal algorithmic ways of turning HTML to Markdown ?
5
#5 opened 14 days ago
by
MohamedRashad
FuseO1-DeekSeekR1-QwQ-SkyT1-32B-Preview
2
#1 opened 10 days ago
by
AaronFeng753
System Prompt
16
#2 opened 10 days ago
by
Wanfq
FIXED: Error with llama-server `unknown pre-tokenizer type: 'deepseek-r1-qwen'`
4
#1 opened 10 days ago
by
ubergarm
The `tokenizer_config.json` is missing the `chat_template` jinja?
1
#1 opened 10 days ago
by
ubergarm
Great RP model in only 12B! A few notes and sampler settings for llama.cpp server inside.
2
#2 opened 13 days ago
by
ubergarm
Nice ~90x real-time generation on 3090TI. Quickstart provided.
5
#20 opened 25 days ago
by
ubergarm
Observation: 4-bit quantization can't answer the Strawberry prompt
12
#2 opened 4 months ago
by
ThePabli
63.17 MMLU-Pro Computer Science with `Q8_0`
#2 opened 4 months ago
by
ubergarm
Benchmarks worse than Qwen2.5-7B-Instruct on MMLU-Pro Computer Science in limited testing.
#1 opened 4 months ago
by
ubergarm
Promising looking results on 24GB VRAM folks!
9
#3 opened 4 months ago
by
ubergarm
Awesome model
6
#5 opened 5 months ago
by
dillfrescott
vram usage of each?
3
#1 opened 5 months ago
by
jasonden
Works good generating python on my 64GB RAM w/ 3090TI 24GB VRAM dev box
3
#2 opened 6 months ago
by
ubergarm