Jared Sulzdorf PRO
jsulz
AI & ML interests
Infrastructure, law, policy
Recent Activity
reacted
to
danielhanchen's
post
with 🚀
1 day ago
🦥 Introducing Unsloth Dynamic v2.0 GGUFs!
Our v2.0 quants set new benchmarks on 5-shot MMLU and KL Divergence, meaning you can now run & fine-tune quantized LLMs while preserving as much accuracy as possible.
Llama 4: https://huggingface.co/unsloth/Llama-4-Scout-17B-16E-Instruct-GGUF
DeepSeek-R1: https://huggingface.co/unsloth/DeepSeek-R1-GGUF-UD
Gemma 3: https://huggingface.co/unsloth/gemma-3-27b-it-GGUF
We made selective layer quantization much smarter. Instead of modifying only a subset of layers, we now dynamically quantize all layers so every layer has a different bit. Now, our dynamic method can be applied to all LLM architectures, not just MoE's.
Blog with Details: https://docs.unsloth.ai/basics/dynamic-v2.0
All our future GGUF uploads will leverage Dynamic 2.0 and our hand curated 300K–1.5M token calibration dataset to improve conversational chat performance.
For accurate benchmarking, we built an evaluation framework to match the reported 5-shot MMLU scores of Llama 4 and Gemma 3. This allowed apples-to-apples comparisons between full-precision vs. Dynamic v2.0, QAT and standard iMatrix quants.
Dynamic v2.0 aims to minimize the performance gap between full-precision models and their quantized counterparts.
reacted
to
danielhanchen's
post
with ❤️
1 day ago
🦥 Introducing Unsloth Dynamic v2.0 GGUFs!
Our v2.0 quants set new benchmarks on 5-shot MMLU and KL Divergence, meaning you can now run & fine-tune quantized LLMs while preserving as much accuracy as possible.
Llama 4: https://huggingface.co/unsloth/Llama-4-Scout-17B-16E-Instruct-GGUF
DeepSeek-R1: https://huggingface.co/unsloth/DeepSeek-R1-GGUF-UD
Gemma 3: https://huggingface.co/unsloth/gemma-3-27b-it-GGUF
We made selective layer quantization much smarter. Instead of modifying only a subset of layers, we now dynamically quantize all layers so every layer has a different bit. Now, our dynamic method can be applied to all LLM architectures, not just MoE's.
Blog with Details: https://docs.unsloth.ai/basics/dynamic-v2.0
All our future GGUF uploads will leverage Dynamic 2.0 and our hand curated 300K–1.5M token calibration dataset to improve conversational chat performance.
For accurate benchmarking, we built an evaluation framework to match the reported 5-shot MMLU scores of Llama 4 and Gemma 3. This allowed apples-to-apples comparisons between full-precision vs. Dynamic v2.0, QAT and standard iMatrix quants.
Dynamic v2.0 aims to minimize the performance gap between full-precision models and their quantized counterparts.
reacted
to
danielhanchen's
post
with 🤗
1 day ago
🦥 Introducing Unsloth Dynamic v2.0 GGUFs!
Our v2.0 quants set new benchmarks on 5-shot MMLU and KL Divergence, meaning you can now run & fine-tune quantized LLMs while preserving as much accuracy as possible.
Llama 4: https://huggingface.co/unsloth/Llama-4-Scout-17B-16E-Instruct-GGUF
DeepSeek-R1: https://huggingface.co/unsloth/DeepSeek-R1-GGUF-UD
Gemma 3: https://huggingface.co/unsloth/gemma-3-27b-it-GGUF
We made selective layer quantization much smarter. Instead of modifying only a subset of layers, we now dynamically quantize all layers so every layer has a different bit. Now, our dynamic method can be applied to all LLM architectures, not just MoE's.
Blog with Details: https://docs.unsloth.ai/basics/dynamic-v2.0
All our future GGUF uploads will leverage Dynamic 2.0 and our hand curated 300K–1.5M token calibration dataset to improve conversational chat performance.
For accurate benchmarking, we built an evaluation framework to match the reported 5-shot MMLU scores of Llama 4 and Gemma 3. This allowed apples-to-apples comparisons between full-precision vs. Dynamic v2.0, QAT and standard iMatrix quants.
Dynamic v2.0 aims to minimize the performance gap between full-precision models and their quantized counterparts.
Organizations
jsulz's activity
xet repositories
3
3
#1 opened 2 months ago
by
Tonic

[bot] Conversion to Parquet
#1 opened 5 months ago
by
parquet-converter

The monthly addition figure has empty extension?
1
#5 opened 7 months ago
by
1a1a11a
Compressed -> Deduped column header
2
#4 opened 7 months ago
by
erinys

LFS Analysis Roadmap
3
#3 opened 7 months ago
by
jsulz

Suggested text changes
1
#1 opened 7 months ago
by
erinys

Filter by file extension chart potential bug
1
#2 opened 7 months ago
by
Awhildy

[bot] Conversion to Parquet
#1 opened 8 months ago
by
parquet-converter

[bot] Conversion to Parquet
#1 opened 8 months ago
by
parquet-converter
