prithivMLmods's picture
Update README.md
0f2906f verified
---
license: apache-2.0
base_model:
- Qwen/Qwen3-4B-Instruct-2507
language:
- en
pipeline_tag: text-generation
library_name: transformers
tags:
- text-generation-inference
- trl
- event-driven
- abliterated
- smoothing
---
![5](https://cdn-uploads.huggingface.co/production/uploads/65bb837dbfb878f46c77de4c/aMY6z1xw0HkVK-DRbRrUW.png)
# **Kepler-186f-Qwen3-Instruct-4B**
> **Kepler-186f-Qwen3-Instruct-4B** is a reasoning-focused model fine-tuned on **Qwen** for **Abliterated Reasoning** and **polished token probabilities**, enhancing balanced **multilingual generation** across mathematics and general-purpose reasoning.
> It specializes in **event-driven logic**, **structured analysis**, and precise probabilistic modeling—making it an ideal tool for researchers, educators, and developers working with uncertainty and structured reasoning.
> [!note]
> GGUF: [https://huggingface.co/prithivMLmods/Kepler-186f-Qwen3-Instruct-4B-GGUF](https://huggingface.co/prithivMLmods/Kepler-186f-Qwen3-Instruct-4B-GGUF)
---
## **Key Features**
1. **Abliterated Reasoning**
Enhanced reasoning precision through polished token probability distributions in Qwen and similar models, ensuring balanced and context-aware outputs.
2. **Event Simulation & Logical Analysis**
Models random events, probability-driven reasoning, and logical decision-making with strong consistency.
3. **Multilingual Mathematical & General-Purpose Problem Solving**
Delivers robust performance in **math**, **probability**, and **structured multilingual tasks**, enabling wide applicability in global research and education.
4. **Hybrid Symbolic-Probabilistic Thinking**
Combines structured logic, probabilistic inference, and reasoning fluency, providing accuracy across uncertainty-driven tasks.
5. **Structured Output Mastery**
Generates well-structured outputs in **LaTeX**, **Markdown**, **JSON**, **CSV**, and **YAML**, supporting technical workflows and data-driven research.
6. **Optimized Lightweight Footprint**
Large **4B parameter size**, deployable on **mid-range GPUs**, **offline clusters**, and **edge devices**, while maintaining reasoning quality.
---
## **Quickstart with Transformers**
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model_name = "prithivMLmods/Kepler-186f-Qwen3-Instruct-4B"
model = AutoModelForCausalLM.from_pretrained(
model_name,
torch_dtype="auto",
device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained(model_name)
prompt = "Simulate the probability of rolling two dice and getting a sum greater than 9. Show the reasoning."
messages = [
{"role": "system", "content": "You are a reasoning tutor skilled in probability, logic, and multilingual problem-solving."},
{"role": "user", "content": prompt}
]
text = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
generated_ids = model.generate(
**model_inputs,
max_new_tokens=512
)
generated_ids = [
output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
]
response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
print(response)
```
---
## **Intended Use**
* Balanced multilingual reasoning and probability modeling
* Event simulation, uncertainty analysis, and structured problem solving
* Educational and research-focused reasoning tasks
* Deployment on mid-resource environments with efficient reasoning
* Technical content and structured data generation
---
## **Limitations**
* Focused on reasoning and mathematics—less suited for creative writing
* Despite 4B size, very complex multi-hop tasks may still challenge the model
* Prioritizes structured reasoning and probabilistic accuracy over conversational or emotional tone
* May produce inconsistent outputs when handling **very long contexts** or cross-domain multi-document inputs