Model Card for Model ID

Model Description

  • Training objective: Fine-tuned on Italian instruction-style reasoning dataset for better performance in logical, educational, and chain-of-thought tasks.
  • Language(s) (NLP): Italian
  • License: MIT
  • Finetuned from model: unsloth/Qwen3-4B

Uses

Direct Use

This model is intended for reasoning-intensive tasks in Italian

Bias, Risks, and Limitations

  • May hallucinate or make factual errors in complex logic chains.
  • Not safe for unsupervised use in high-stakes domains like medical/legal reasoning.
  • Output quality depends on instruction clarity.

Training Data

The DeepMount00/o1-ITA-REASONING dataset is crafted to train language models in providing structured, methodical responses to questions in Italian. Each entry follows a four-step reasoning approach:

  • Reasoning: Initial thought process
  • Verification: Self-review of the reasoning
  • Correction: Amendments if needed
  • Final Answer: Conclusive response

The dataset is formatted using XML-like tags to delineate each component, promoting transparency and structured thinking. It is particularly beneficial for educational purposes, encouraging systematic problem-solving and critical thinking in the Italian language.

How to Get Started with the Model

Use the code below to get started with the model.

from huggingface_hub import login
from transformers import AutoTokenizer, AutoModelForCausalLM
from peft import PeftModel


login(token="")

tokenizer = AutoTokenizer.from_pretrained("unsloth/Qwen3-4B",)
base_model = AutoModelForCausalLM.from_pretrained(
    "unsloth/Qwen3-4B",
    device_map={"": 0}, token=""
)

model = PeftModel.from_pretrained(base_model,"Rustamshry/ITA-Reasoning-o1")


question = "Quali sono i costi e i benefici ambientali, sociali ed economici dell'energia solare?"

messages = [
    {"role" : "user", "content" : question}
]
text = tokenizer.apply_chat_template(
    messages,
    tokenize = False,
    add_generation_prompt = True, # Must add for generation
    enable_thinking = True, # Disable thinking
)

from transformers import TextStreamer
_ = model.generate(
    **tokenizer(text, return_tensors = "pt").to("cuda"),
    max_new_tokens = 2048,
    temperature = 0.6, 
    top_p = 0.95, 
    top_k = 20,
    streamer = TextStreamer(tokenizer, skip_prompt = True),
)

Framework versions

  • PEFT 0.14.0
Downloads last month
8
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Rustamshry/ITA-Reasoning-o1

Base model

Qwen/Qwen3-4B-Base
Finetuned
Qwen/Qwen3-4B
Finetuned
unsloth/Qwen3-4B
Adapter
(1)
this model

Dataset used to train Rustamshry/ITA-Reasoning-o1