Update README.md
Browse files
README.md
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
---
|
2 |
base_model:
|
3 |
-
- saishshinde15/
|
4 |
tags:
|
5 |
- text-generation-inference
|
6 |
- transformers
|
@@ -19,12 +19,12 @@ pipeline_tag: text-generation
|
|
19 |
|
20 |
- **Developed by:** clyrai
|
21 |
- **License:** apache-2.0
|
22 |
-
- **Fine-tuned from:** [saishshinde15/
|
23 |
- **Category:** Experimental, Research
|
24 |
|
25 |
## **Introduction**
|
26 |
|
27 |
-
TethysAI Vortex Reasoning is an **experimental model** that advances the structured reasoning capabilities pioneered by [
|
28 |
|
29 |
The core objective was to investigate whether **deep reasoning and self-questioning behavior could emerge purely through SFT on high-quality datasets**. The results were highly promising: the model successfully **questions itself internally**, improves reasoning depth, and consistently generates structured, logical responses.
|
30 |
|
|
|
1 |
---
|
2 |
base_model:
|
3 |
+
- saishshinde15/Clyrai_Base_Reasoning
|
4 |
tags:
|
5 |
- text-generation-inference
|
6 |
- transformers
|
|
|
19 |
|
20 |
- **Developed by:** clyrai
|
21 |
- **License:** apache-2.0
|
22 |
+
- **Fine-tuned from:** [saishshinde15/Clyrai_Base_Reasoning](https://huggingface.co/saishshinde15/TethysAI_Base_Reasoning)
|
23 |
- **Category:** Experimental, Research
|
24 |
|
25 |
## **Introduction**
|
26 |
|
27 |
+
TethysAI Vortex Reasoning is an **experimental model** that advances the structured reasoning capabilities pioneered by [Clyrai_Base Reasoning](https://huggingface.co/saishshinde15/TethysAI_Base_Reasoning). While the Base Reasoning model utilized **Generalized Reinforced Policy Optimization (GRPO)** to enhance step-by-step logical thought processes similar to **DeepSeek-R1**, this model takes a different approach—**eliminating GRPO and instead relying on high-end Supervised Fine-Tuning (SFT) techniques**.
|
28 |
|
29 |
The core objective was to investigate whether **deep reasoning and self-questioning behavior could emerge purely through SFT on high-quality datasets**. The results were highly promising: the model successfully **questions itself internally**, improves reasoning depth, and consistently generates structured, logical responses.
|
30 |
|