File size: 4,125 Bytes
9e83164 051aeec 259fd8e 051aeec 000f765 051aeec a3db867 051aeec a3db867 051aeec a3db867 051aeec a3db867 051aeec a3db867 051aeec a3db867 051aeec a3db867 051aeec a3db867 051aeec a3db867 051aeec a3db867 051aeec a3db867 2a8917f |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 |
---
license: apache-2.0
datasets:
- nvidia/OpenScienceReasoning-2
language:
- en
base_model:
- Qwen/Qwen3-1.7B
pipeline_tag: text-generation
library_name: transformers
tags:
- trl
- text-generation-inference
- medical
- science
---

# **OpenScienceReasoning-Qwen-e10**
> OpenScienceReasoning-Qwen-e10 is a high-efficiency, science-focused reasoning model fine-tuned on **Qwen3-1.7B** using the [**nvidia/OpenScienceReasoning-2**](https://huggingface.co/datasets/nvidia/OpenScienceReasoning-2) dataset. It incorporates **10,000 distinct entries** for scientific reasoning, chain-of-thought exploration, and analytical problem solving.
> The model blends symbolic precision, scientific logic, and structured output fluency—making it an ideal tool for researchers, educators, and developers seeking advanced reasoning under constrained compute.
> \[!note]
> GGUF: [https://huggingface.co/prithivMLmods/OpenScienceReasoning-Qwen-e10-GGUF](https://huggingface.co/prithivMLmods/OpenScienceReasoning-Qwen-e10-GGUF)
---
## **Key Features**
1. **Scientific Reasoning & Chain-of-Thought**
Fine-tuned on **10,000 curated entries** from the **OpenScienceReasoning-2** dataset, designed to enhance step-by-step analytical reasoning in science and mathematics.
2. **Advanced Code Reasoning & Generation**
Supports multi-language coding with explanations, optimization hints, and error detection—ideal for algorithm synthesis, debugging, and prototyping.
3. **Mathematical & Scientific Problem Solving**
Performs analytical reasoning in physics, biology, chemistry, and mathematics—explaining concepts, solving equations, and handling symbolic derivations.
4. **Hybrid Symbolic-AI Thinking**
Combines structured logic, chain-of-thought reasoning, and open-ended inference, delivering robust performance on STEM-related tasks.
5. **Structured Output Mastery**
Seamlessly generates output in **LaTeX**, **Markdown**, **JSON**, **CSV**, and **YAML**, suited for technical documentation, research papers, and structured data.
6. **Optimized Lightweight Footprint for Versatile Deployment**
Balances performance and efficiency, making it deployable on **mid-range GPUs**, **offline clusters**, and **edge AI systems**.
---
## **Quickstart with Transformers**
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model_name = "prithivMLmods/OpenScienceReasoning-Qwen-e10"
model = AutoModelForCausalLM.from_pretrained(
model_name,
torch_dtype="auto",
device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained(model_name)
prompt = "Explain the difference between Newtonian mechanics and quantum mechanics with examples."
messages = [
{"role": "system", "content": "You are a scientific tutor skilled in reasoning, math, and coding."},
{"role": "user", "content": prompt}
]
text = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
generated_ids = model.generate(
**model_inputs,
max_new_tokens=512
)
generated_ids = [
output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
]
response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
print(response)
```
---
## **Intended Use**
* Scientific tutoring, computational reasoning, and mathematical education
* Research assistant for physics, chemistry, biology, and interdisciplinary domains
* Structured technical data generation in multiple formats
* STEM-focused chatbot or API for research and education tools
* Deployment in mid-resource environments requiring high reasoning fidelity
## **Limitations**
* Not tuned for general-purpose or long-form creative writing
* Context limitations may hinder multi-document or full codebase analysis
* Specialized for scientific and technical reasoning—general chat may underperform
* Prioritizes structured logic over casual or emotional tone generation |