--- base_model: unsloth/Qwen3-4B library_name: peft license: mit datasets: - DeepMount00/o1-ITA-REASONING language: - it pipeline_tag: question-answering --- # Model Card for Model ID ### Model Description - **Training objective**: Fine-tuned on Italian instruction-style reasoning dataset for better performance in logical, educational, and chain-of-thought tasks. - **Language(s) (NLP):** Italian - **License:** MIT - **Finetuned from model:** unsloth/Qwen3-4B ## Uses ### Direct Use This model is intended for reasoning-intensive tasks in Italian ## Bias, Risks, and Limitations - May hallucinate or make factual errors in complex logic chains. - Not safe for unsupervised use in high-stakes domains like medical/legal reasoning. - Output quality depends on instruction clarity. # Training Data The DeepMount00/o1-ITA-REASONING dataset is crafted to train language models in providing structured, methodical responses to questions in Italian. Each entry follows a four-step reasoning approach: - Reasoning: Initial thought process - Verification: Self-review of the reasoning - Correction: Amendments if needed - Final Answer: Conclusive response The dataset is formatted using XML-like tags to delineate each component, promoting transparency and structured thinking. It is particularly beneficial for educational purposes, encouraging systematic problem-solving and critical thinking in the Italian language. ## How to Get Started with the Model Use the code below to get started with the model. ```python from huggingface_hub import login from transformers import AutoTokenizer, AutoModelForCausalLM from peft import PeftModel login(token="") tokenizer = AutoTokenizer.from_pretrained("unsloth/Qwen3-4B",) base_model = AutoModelForCausalLM.from_pretrained( "unsloth/Qwen3-4B", device_map={"": 0}, token="" ) model = PeftModel.from_pretrained(base_model,"Rustamshry/ITA-Reasoning-o1") question = "Quali sono i costi e i benefici ambientali, sociali ed economici dell'energia solare?" messages = [ {"role" : "user", "content" : question} ] text = tokenizer.apply_chat_template( messages, tokenize = False, add_generation_prompt = True, # Must add for generation enable_thinking = True, # Disable thinking ) from transformers import TextStreamer _ = model.generate( **tokenizer(text, return_tensors = "pt").to("cuda"), max_new_tokens = 2048, temperature = 0.6, top_p = 0.95, top_k = 20, streamer = TextStreamer(tokenizer, skip_prompt = True), ) ``` ### Framework versions - PEFT 0.14.0