--- license: mit language: - en base_model: - microsoft/phi-4 pipeline_tag: question-answering tags: - unsloth datasets: - mlabonne/FineTome-100k metrics: - accuracy new_version: microsoft/phi-4-gguf library_name: diffusers --- My First Huggingface Model - Default UnSloth phi4 template with LoRA fine tuner Locally trained for around 2 hours, utilized around 16 GB RAM to store the data. I also used 8 GB RAM to train the model with my GPU