Model Card for Model ID
Model Description
- Training objective: Fine-tuned on Italian instruction-style reasoning dataset for better performance in logical, educational, and chain-of-thought tasks.
- Language(s) (NLP): Italian
- License: MIT
- Finetuned from model: unsloth/Qwen3-4B
Uses
Direct Use
This model is intended for reasoning-intensive tasks in Italian
Bias, Risks, and Limitations
- May hallucinate or make factual errors in complex logic chains.
- Not safe for unsupervised use in high-stakes domains like medical/legal reasoning.
- Output quality depends on instruction clarity.
Training Data
The DeepMount00/o1-ITA-REASONING dataset is crafted to train language models in providing structured, methodical responses to questions in Italian. Each entry follows a four-step reasoning approach:
- Reasoning: Initial thought process
- Verification: Self-review of the reasoning
- Correction: Amendments if needed
- Final Answer: Conclusive response
The dataset is formatted using XML-like tags to delineate each component, promoting transparency and structured thinking. It is particularly beneficial for educational purposes, encouraging systematic problem-solving and critical thinking in the Italian language.
How to Get Started with the Model
Use the code below to get started with the model.
from huggingface_hub import login
from transformers import AutoTokenizer, AutoModelForCausalLM
from peft import PeftModel
login(token="")
tokenizer = AutoTokenizer.from_pretrained("unsloth/Qwen3-4B",)
base_model = AutoModelForCausalLM.from_pretrained(
"unsloth/Qwen3-4B",
device_map={"": 0}, token=""
)
model = PeftModel.from_pretrained(base_model,"Rustamshry/ITA-Reasoning-o1")
question = "Quali sono i costi e i benefici ambientali, sociali ed economici dell'energia solare?"
messages = [
{"role" : "user", "content" : question}
]
text = tokenizer.apply_chat_template(
messages,
tokenize = False,
add_generation_prompt = True, # Must add for generation
enable_thinking = True, # Disable thinking
)
from transformers import TextStreamer
_ = model.generate(
**tokenizer(text, return_tensors = "pt").to("cuda"),
max_new_tokens = 2048,
temperature = 0.6,
top_p = 0.95,
top_k = 20,
streamer = TextStreamer(tokenizer, skip_prompt = True),
)
Framework versions
- PEFT 0.14.0
- Downloads last month
- 8
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support