The aim of this model is to retain the reasoning capabilities of DeepSeek-R1-Distill-Llama-70B, while aligning more with the original Llama 3.3 model on which it is based.

As this model derives from Llama 3.3, the Llama 3.3 Community License Agreement applies.

Use the DeepSeek Chat Prompt Template when working with this model.

70B Safetensors BF16 format:

Use with transformers as you would Llama 3.3, but use the DeepSeek Chat Prompt Template as you would with the original DeepSeek-R1-Distill-Llama models.

Or download files from here

70B GGUF Quantised versions:

Use these with Llama.cpp, LM Studio or Kobold.cpp. Thanks to mradermacher for converting these from the safetensors format.

Filename Type Size Quality
LlamaAligned-DeepSeekR1-Distill-70b-Q4_K_M.gguf Q4_K_M 42.5GB OK quality, default.
LlamaAligned-DeepSeekR1-Distill-70b-Q8_0.gguf part1 part2 Q8_0 75.0GB Best quality quantised version.
LlamaAligned-DeepSeekR1-Distill-70b-Q3_K_S.gguf Q3_K_S 30.9GB Lower quality.

8B Safetensors BF16 format:

Use with transformers as you would Llama 3.1, but use the DeepSeek Chat Prompt Template as you would with the original DeepSeek-R1-Distill-Llama models.

Use model id BlueBeck/LlamaAligned-DeepSeekR1-Distill-8b

Or download files from here

8B GGUF Quantised versions:

Use these with Llama.cpp, LM Studio or Kobold.cpp. Thanks to mradermacher for converting these from the safetensors format.

Filename Type Size Quality
LlamaAligned-DeepSeekR1-Distill-8b-Q4_K_M.gguf Q4_K_M 4.92GB OK quality, default.
LlamaAligned-DeepSeekR1-Distill-8b-Q8_0.gguf Q8_0 8.54GB Best quality quantised version.
LlamaAligned-DeepSeekR1-Distill-8b-Q6_K.gguf Q6_K 6.6GB High quality.
LlamaAligned-DeepSeekR1-Distill-8b-Q5_K_M>.gguf Q5_K_M> 5.73GB Good quality.
LlamaAligned-DeepSeekR1-Distill-8b-Q3_K_S.gguf Q3_K_S 3.66GB Lower quality.
Downloads last month
23
Safetensors
Model size
70.6B params
Tensor type
BF16
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Model tree for BlueBeck/LlamaAligned-DeepSeekR1-Distill-70b

Finetuned
(11)
this model
Quantizations
1 model

Space using BlueBeck/LlamaAligned-DeepSeekR1-Distill-70b 1