saishshinde15 commited on
Commit
9492a8c
·
verified ·
1 Parent(s): 0e85df1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  base_model:
3
- - saishshinde15/TethysAI_Base_Reasoning
4
  tags:
5
  - text-generation-inference
6
  - transformers
@@ -19,12 +19,12 @@ pipeline_tag: text-generation
19
 
20
  - **Developed by:** clyrai
21
  - **License:** apache-2.0
22
- - **Fine-tuned from:** [saishshinde15/TBH.AI_Base_Reasoning](https://huggingface.co/saishshinde15/TethysAI_Base_Reasoning)
23
  - **Category:** Experimental, Research
24
 
25
  ## **Introduction**
26
 
27
- TethysAI Vortex Reasoning is an **experimental model** that advances the structured reasoning capabilities pioneered by [Clyrai Base Reasoning](https://huggingface.co/saishshinde15/TethysAI_Base_Reasoning). While the Base Reasoning model utilized **Generalized Reinforced Policy Optimization (GRPO)** to enhance step-by-step logical thought processes similar to **DeepSeek-R1**, this model takes a different approach—**eliminating GRPO and instead relying on high-end Supervised Fine-Tuning (SFT) techniques**.
28
 
29
  The core objective was to investigate whether **deep reasoning and self-questioning behavior could emerge purely through SFT on high-quality datasets**. The results were highly promising: the model successfully **questions itself internally**, improves reasoning depth, and consistently generates structured, logical responses.
30
 
 
1
  ---
2
  base_model:
3
+ - saishshinde15/Clyrai_Base_Reasoning
4
  tags:
5
  - text-generation-inference
6
  - transformers
 
19
 
20
  - **Developed by:** clyrai
21
  - **License:** apache-2.0
22
+ - **Fine-tuned from:** [saishshinde15/Clyrai_Base_Reasoning](https://huggingface.co/saishshinde15/TethysAI_Base_Reasoning)
23
  - **Category:** Experimental, Research
24
 
25
  ## **Introduction**
26
 
27
+ TethysAI Vortex Reasoning is an **experimental model** that advances the structured reasoning capabilities pioneered by [Clyrai_Base Reasoning](https://huggingface.co/saishshinde15/TethysAI_Base_Reasoning). While the Base Reasoning model utilized **Generalized Reinforced Policy Optimization (GRPO)** to enhance step-by-step logical thought processes similar to **DeepSeek-R1**, this model takes a different approach—**eliminating GRPO and instead relying on high-end Supervised Fine-Tuning (SFT) techniques**.
28
 
29
  The core objective was to investigate whether **deep reasoning and self-questioning behavior could emerge purely through SFT on high-quality datasets**. The results were highly promising: the model successfully **questions itself internally**, improves reasoning depth, and consistently generates structured, logical responses.
30