TurnPhi Project
- Developed by: Quazim0t0
- Finetuned from model : unsloth/phi-4-unsloth-bnb-4bit
- GGUF
- Trained for 8 Hours on A800 with the Bespoke Stratos 17k Dataset.
- Trained for 6 Hours on A800 with the Bespoke Stratos 35k Dataset.
- Trained for 2 Hours on A800 with the Benford's Law Reasoning Small 430 Row Dataset, ensuring no overfitting.
- Trained for 4 Hours on A800 with the Sky-T1_data_17k Dataset
- Trained for 6 Hours on A800 with the Openthoughts 114k Dataset.
- 18$ Training...I'm actually amazed by the results.
OpenWeb UI Function
If using this model for Open WebUI here is a simple function to organize the models responses: https://openwebui.com/f/quaz93/phi4_turn_r1_distill_thought_function_v1
Phi4 Turn R1Distill LoRA Adapters
Overview
These LoRA adapters were trained using diverse reasoning datasets that incorporate structured Thought and Solution responses to enhance logical inference. This project was designed to test the R1 dataset on Phi-4, aiming to create a lightweight, fast, and efficient reasoning model.
All adapters were fine-tuned using an NVIDIA A800 GPU, ensuring high performance and compatibility for continued training, merging, or direct deployment.
As part of an open-source initiative, all resources are made publicly available for unrestricted research and development.
LoRA Adapters
Below are the currently available LoRA fine-tuned adapters (as of January 30, 2025):
- Phi4.Turn.R1Distill-Lora1
- Phi4.Turn.R1Distill-Lora2
- Phi4.Turn.R1Distill-Lora3
- Phi4.Turn.R1Distill-Lora4
- Phi4.Turn.R1Distill-Lora5
- Phi4.Turn.R1Distill-Lora6
- Phi4.Turn.R1Distill-Lora7
- Phi4.Turn.R1Distill-Lora8
GGUF Full & Quantized Models
To facilitate broader testing and real-world inference, GGUF Full and Quantized versions have been provided for evaluation on Open WebUI and other LLM interfaces.
Version 1
Version 1.1
Version 1.2
Version 1.3
Version 1.4
Version 1.5
OpenLLM Leadboards
Score:
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/Quazim0t0__Phi4.Turn.R1Distill_v1.5.1-Tensors-details)
| Metric |Value|
|-------------------|----:|
|Avg. |22.67|
|IFEval (0-Shot) |29.95|
|BBH (3-Shot) |49.22|
|MATH Lvl 5 (4-Shot)| 1.59|
|GPQA (0-shot) | 2.46|
|MuSR (0-shot) | 7.04|
|MMLU-PRO (5-shot) |45.75|
- Downloads last month
- 409
Model tree for Quazim0t0/Phi4.Turn.R1Distill_v1.5.1_Q4_k-GGUF
Datasets used to train Quazim0t0/Phi4.Turn.R1Distill_v1.5.1_Q4_k-GGUF
Collection including Quazim0t0/Phi4.Turn.R1Distill_v1.5.1_Q4_k-GGUF
Evaluation results
- strict accuracy on IFEval (0-Shot)Open LLM Leaderboard29.950
- normalized accuracy on BBH (3-Shot)Open LLM Leaderboard49.220
- exact match on MATH Lvl 5 (4-Shot)Open LLM Leaderboard1.590
- acc_norm on GPQA (0-shot)Open LLM Leaderboard2.460
- acc_norm on MuSR (0-shot)Open LLM Leaderboard7.040
- accuracy on MMLU-PRO (5-shot)test set Open LLM Leaderboard45.750