eagle0504 commited on
Commit
6affacd
·
verified ·
1 Parent(s): 1881b48

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +104 -0
README.md ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ license: mit
4
+ datasets:
5
+ - eagle0504/augmented_codealpaca-20k-using-together-ai-deepseek-v1
6
+ language:
7
+ - en
8
+ new_version: eagle0504/finetuned-deepseek-r1-distill-qwen-1.5b-by-openai-gsm8k-enhanced-v2
9
+ pipeline_tag: question-answering
10
+ ---
11
+
12
+ # Model Card for CodeAlpaca-20k Dataset Enhanced with Reasoning
13
+
14
+ This model is fine-tuned to answer questions based on the CodeAlpaca-20k dataset enhanced with reasoning provided from Deepseek R1.
15
+
16
+ Invoke notebook shared [here](https://colab.research.google.com/drive/1B_Fbz0w76QxHbo9zAOf_pyZKKNI0EJJ9?usp=sharing), a publicly available Colab notebook for tests.
17
+
18
+ ---
19
+
20
+ ## Model Details
21
+
22
+ ### Model Description
23
+
24
+ This is a transformer-based question-answering model fine-tuned from `eagle0504/finetuned-deepseek-r1-distill-qwen-1.5b-by-openai-gsm8k-enhanced-v2`. It was trained on a dataset derived from the OpenAI GSM8K benchmark, enhanced with chain-of-thought reasoning to encourage intermediate logical steps. The dataset pairs math word problems with structured answers, using `<think>...</think>` and `<answer>...</answer>` tags.
25
+
26
+ - **Developed by:** Yiqiao Yin
27
+ - **Model type:** Causal Language Model (fine-tuned for Q&A with reasoning)
28
+ - **Language(s):** English
29
+ - **License:** MIT
30
+ - **Finetuned from model:** eagle0504/finetuned-deepseek-r1-distill-qwen-1.5b-by-openai-gsm8k-enhanced-v2
31
+
32
+ ---
33
+
34
+ ## Training Configuration
35
+
36
+ - 🖥️ **Hardware:** Trained on a RunPod instance with:
37
+ - 🔥 4 x A100 SXM
38
+ - 🧠 146 vCPU
39
+ - 🧮 1144 GB RAM
40
+ - 💽 20 GB disk per GPU
41
+ - 🐳 **Container Image:** `runpod/pytorch:2.1.0-py3.10-cuda11.8.0-devel-ubuntu22.04`
42
+ - ⏱️ **Total Training Time:** 2 hours
43
+ - 💸 **Cost:** ~$7.56/hour × 2 hours = **$14+ USD**
44
+ - ⚙️ **Zero Redundancy Optimization:** DeepSpeed Stage 1
45
+ - 🎯 **Precision:** FP16 mixed-precision training
46
+
47
+ ---
48
+
49
+ ## Performance
50
+
51
+ - **Mean token-level accuracy:** **98%**
52
+ - Evaluation based on in-training token match accuracy over the formatted `<think>...</think><answer>...</answer>` structure.
53
+ - Model demonstrates strong reasoning capability in multi-step arithmetic and logic problems.
54
+
55
+ ---
56
+
57
+ ## Inference Format
58
+
59
+ To generate accurate completions, prompt the model in the following structure:
60
+
61
+ ```
62
+ <question>Question: If Sally has 3 apples and buys 2 more, how many does she have in total?</question>
63
+ ```
64
+
65
+ Be aware that this token `</question>` will prompt the answer to start with `<think>` which is trained into the model based on training data.
66
+
67
+ The model will continue reasoning within `<think>...</think>` and provide a final answer inside `<answer>...</answer>`.
68
+
69
+ ---
70
+
71
+ ## Intended Use
72
+
73
+ This model is intended for educational and research purposes in:
74
+ - Chain-of-thought prompting
75
+ - Math reasoning and logical inference
76
+ - Question-answering with intermediate steps
77
+
78
+ ---
79
+
80
+ ## Limitations
81
+
82
+ - Trained on structured synthetic data — real-world generalization may vary
83
+ - Best performance achieved when following the exact inference format
84
+ - Does not support multilingual inputs
85
+
86
+ ---
87
+
88
+ ## Citation
89
+
90
+ If you use this model, please cite:
91
+
92
+ ```
93
+ @misc{yin2024gsm8k,
94
+ author = {Yiqiao Yin},
95
+ title = {TBD},
96
+ year = 2025,
97
+ note = {TBD}
98
+ }
99
+ ```
100
+
101
+ ## Model Card Contact
102
+
103
+ Author: Yiqiao Yin
104
+ Connect with me on [LinkedIn](https://www.linkedin.com/in/yiqiaoyin/)