rawsun00001 commited on
Commit
882a53f
·
verified ·
1 Parent(s): 01be675

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. README.md +207 -0
  2. adapter_config.json +37 -0
  3. adapter_model.safetensors +3 -0
  4. checkpoint-1000/README.md +207 -0
  5. checkpoint-1000/adapter_config.json +37 -0
  6. checkpoint-1000/adapter_model.safetensors +3 -0
  7. checkpoint-1000/merges.txt +0 -0
  8. checkpoint-1000/optimizer.pt +3 -0
  9. checkpoint-1000/rng_state.pth +3 -0
  10. checkpoint-1000/scaler.pt +3 -0
  11. checkpoint-1000/scheduler.pt +3 -0
  12. checkpoint-1000/special_tokens_map.json +30 -0
  13. checkpoint-1000/tokenizer.json +0 -0
  14. checkpoint-1000/tokenizer_config.json +28 -0
  15. checkpoint-1000/trainer_state.json +85 -0
  16. checkpoint-1000/training_args.bin +3 -0
  17. checkpoint-1000/vocab.json +0 -0
  18. checkpoint-10000/README.md +207 -0
  19. checkpoint-10000/adapter_config.json +37 -0
  20. checkpoint-10000/adapter_model.safetensors +3 -0
  21. checkpoint-10000/merges.txt +0 -0
  22. checkpoint-10000/optimizer.pt +3 -0
  23. checkpoint-10000/rng_state.pth +3 -0
  24. checkpoint-10000/scaler.pt +3 -0
  25. checkpoint-10000/scheduler.pt +3 -0
  26. checkpoint-10000/special_tokens_map.json +30 -0
  27. checkpoint-10000/tokenizer.json +0 -0
  28. checkpoint-10000/tokenizer_config.json +28 -0
  29. checkpoint-10000/trainer_state.json +544 -0
  30. checkpoint-10000/training_args.bin +3 -0
  31. checkpoint-10000/vocab.json +0 -0
  32. checkpoint-11000/README.md +207 -0
  33. checkpoint-11000/adapter_config.json +37 -0
  34. checkpoint-11000/adapter_model.safetensors +3 -0
  35. checkpoint-11000/merges.txt +0 -0
  36. checkpoint-11000/optimizer.pt +3 -0
  37. checkpoint-11000/rng_state.pth +3 -0
  38. checkpoint-11000/scaler.pt +3 -0
  39. checkpoint-11000/scheduler.pt +3 -0
  40. checkpoint-11000/special_tokens_map.json +30 -0
  41. checkpoint-11000/tokenizer.json +0 -0
  42. checkpoint-11000/tokenizer_config.json +28 -0
  43. checkpoint-11000/trainer_state.json +595 -0
  44. checkpoint-11000/training_args.bin +3 -0
  45. checkpoint-11000/vocab.json +0 -0
  46. checkpoint-12000/README.md +207 -0
  47. checkpoint-12000/adapter_config.json +37 -0
  48. checkpoint-12000/adapter_model.safetensors +3 -0
  49. checkpoint-12000/merges.txt +0 -0
  50. checkpoint-12000/optimizer.pt +3 -0
README.md ADDED
@@ -0,0 +1,207 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: distilgpt2
3
+ library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - base_model:adapter:distilgpt2
7
+ - lora
8
+ - transformers
9
+ ---
10
+
11
+ # Model Card for Model ID
12
+
13
+ <!-- Provide a quick summary of what the model is/does. -->
14
+
15
+
16
+
17
+ ## Model Details
18
+
19
+ ### Model Description
20
+
21
+ <!-- Provide a longer summary of what this model is. -->
22
+
23
+
24
+
25
+ - **Developed by:** [More Information Needed]
26
+ - **Funded by [optional]:** [More Information Needed]
27
+ - **Shared by [optional]:** [More Information Needed]
28
+ - **Model type:** [More Information Needed]
29
+ - **Language(s) (NLP):** [More Information Needed]
30
+ - **License:** [More Information Needed]
31
+ - **Finetuned from model [optional]:** [More Information Needed]
32
+
33
+ ### Model Sources [optional]
34
+
35
+ <!-- Provide the basic links for the model. -->
36
+
37
+ - **Repository:** [More Information Needed]
38
+ - **Paper [optional]:** [More Information Needed]
39
+ - **Demo [optional]:** [More Information Needed]
40
+
41
+ ## Uses
42
+
43
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
44
+
45
+ ### Direct Use
46
+
47
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
48
+
49
+ [More Information Needed]
50
+
51
+ ### Downstream Use [optional]
52
+
53
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
54
+
55
+ [More Information Needed]
56
+
57
+ ### Out-of-Scope Use
58
+
59
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
60
+
61
+ [More Information Needed]
62
+
63
+ ## Bias, Risks, and Limitations
64
+
65
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
66
+
67
+ [More Information Needed]
68
+
69
+ ### Recommendations
70
+
71
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
72
+
73
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
74
+
75
+ ## How to Get Started with the Model
76
+
77
+ Use the code below to get started with the model.
78
+
79
+ [More Information Needed]
80
+
81
+ ## Training Details
82
+
83
+ ### Training Data
84
+
85
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
86
+
87
+ [More Information Needed]
88
+
89
+ ### Training Procedure
90
+
91
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
92
+
93
+ #### Preprocessing [optional]
94
+
95
+ [More Information Needed]
96
+
97
+
98
+ #### Training Hyperparameters
99
+
100
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
101
+
102
+ #### Speeds, Sizes, Times [optional]
103
+
104
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
105
+
106
+ [More Information Needed]
107
+
108
+ ## Evaluation
109
+
110
+ <!-- This section describes the evaluation protocols and provides the results. -->
111
+
112
+ ### Testing Data, Factors & Metrics
113
+
114
+ #### Testing Data
115
+
116
+ <!-- This should link to a Dataset Card if possible. -->
117
+
118
+ [More Information Needed]
119
+
120
+ #### Factors
121
+
122
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
123
+
124
+ [More Information Needed]
125
+
126
+ #### Metrics
127
+
128
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
129
+
130
+ [More Information Needed]
131
+
132
+ ### Results
133
+
134
+ [More Information Needed]
135
+
136
+ #### Summary
137
+
138
+
139
+
140
+ ## Model Examination [optional]
141
+
142
+ <!-- Relevant interpretability work for the model goes here -->
143
+
144
+ [More Information Needed]
145
+
146
+ ## Environmental Impact
147
+
148
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
149
+
150
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
151
+
152
+ - **Hardware Type:** [More Information Needed]
153
+ - **Hours used:** [More Information Needed]
154
+ - **Cloud Provider:** [More Information Needed]
155
+ - **Compute Region:** [More Information Needed]
156
+ - **Carbon Emitted:** [More Information Needed]
157
+
158
+ ## Technical Specifications [optional]
159
+
160
+ ### Model Architecture and Objective
161
+
162
+ [More Information Needed]
163
+
164
+ ### Compute Infrastructure
165
+
166
+ [More Information Needed]
167
+
168
+ #### Hardware
169
+
170
+ [More Information Needed]
171
+
172
+ #### Software
173
+
174
+ [More Information Needed]
175
+
176
+ ## Citation [optional]
177
+
178
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
179
+
180
+ **BibTeX:**
181
+
182
+ [More Information Needed]
183
+
184
+ **APA:**
185
+
186
+ [More Information Needed]
187
+
188
+ ## Glossary [optional]
189
+
190
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
191
+
192
+ [More Information Needed]
193
+
194
+ ## More Information [optional]
195
+
196
+ [More Information Needed]
197
+
198
+ ## Model Card Authors [optional]
199
+
200
+ [More Information Needed]
201
+
202
+ ## Model Card Contact
203
+
204
+ [More Information Needed]
205
+ ### Framework versions
206
+
207
+ - PEFT 0.17.0
adapter_config.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "distilgpt2",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 32,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.05,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "qalora_group_size": 16,
24
+ "r": 16,
25
+ "rank_pattern": {},
26
+ "revision": null,
27
+ "target_modules": [
28
+ "c_proj",
29
+ "c_attn"
30
+ ],
31
+ "target_parameters": null,
32
+ "task_type": "CAUSAL_LM",
33
+ "trainable_token_indices": null,
34
+ "use_dora": false,
35
+ "use_qalora": false,
36
+ "use_rslora": false
37
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3efcec786a7e5e5420f5989a58a6947f312a4c8a0a88d0c1f2a9044c040bb261
3
+ size 3248616
checkpoint-1000/README.md ADDED
@@ -0,0 +1,207 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: distilgpt2
3
+ library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - base_model:adapter:distilgpt2
7
+ - lora
8
+ - transformers
9
+ ---
10
+
11
+ # Model Card for Model ID
12
+
13
+ <!-- Provide a quick summary of what the model is/does. -->
14
+
15
+
16
+
17
+ ## Model Details
18
+
19
+ ### Model Description
20
+
21
+ <!-- Provide a longer summary of what this model is. -->
22
+
23
+
24
+
25
+ - **Developed by:** [More Information Needed]
26
+ - **Funded by [optional]:** [More Information Needed]
27
+ - **Shared by [optional]:** [More Information Needed]
28
+ - **Model type:** [More Information Needed]
29
+ - **Language(s) (NLP):** [More Information Needed]
30
+ - **License:** [More Information Needed]
31
+ - **Finetuned from model [optional]:** [More Information Needed]
32
+
33
+ ### Model Sources [optional]
34
+
35
+ <!-- Provide the basic links for the model. -->
36
+
37
+ - **Repository:** [More Information Needed]
38
+ - **Paper [optional]:** [More Information Needed]
39
+ - **Demo [optional]:** [More Information Needed]
40
+
41
+ ## Uses
42
+
43
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
44
+
45
+ ### Direct Use
46
+
47
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
48
+
49
+ [More Information Needed]
50
+
51
+ ### Downstream Use [optional]
52
+
53
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
54
+
55
+ [More Information Needed]
56
+
57
+ ### Out-of-Scope Use
58
+
59
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
60
+
61
+ [More Information Needed]
62
+
63
+ ## Bias, Risks, and Limitations
64
+
65
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
66
+
67
+ [More Information Needed]
68
+
69
+ ### Recommendations
70
+
71
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
72
+
73
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
74
+
75
+ ## How to Get Started with the Model
76
+
77
+ Use the code below to get started with the model.
78
+
79
+ [More Information Needed]
80
+
81
+ ## Training Details
82
+
83
+ ### Training Data
84
+
85
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
86
+
87
+ [More Information Needed]
88
+
89
+ ### Training Procedure
90
+
91
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
92
+
93
+ #### Preprocessing [optional]
94
+
95
+ [More Information Needed]
96
+
97
+
98
+ #### Training Hyperparameters
99
+
100
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
101
+
102
+ #### Speeds, Sizes, Times [optional]
103
+
104
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
105
+
106
+ [More Information Needed]
107
+
108
+ ## Evaluation
109
+
110
+ <!-- This section describes the evaluation protocols and provides the results. -->
111
+
112
+ ### Testing Data, Factors & Metrics
113
+
114
+ #### Testing Data
115
+
116
+ <!-- This should link to a Dataset Card if possible. -->
117
+
118
+ [More Information Needed]
119
+
120
+ #### Factors
121
+
122
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
123
+
124
+ [More Information Needed]
125
+
126
+ #### Metrics
127
+
128
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
129
+
130
+ [More Information Needed]
131
+
132
+ ### Results
133
+
134
+ [More Information Needed]
135
+
136
+ #### Summary
137
+
138
+
139
+
140
+ ## Model Examination [optional]
141
+
142
+ <!-- Relevant interpretability work for the model goes here -->
143
+
144
+ [More Information Needed]
145
+
146
+ ## Environmental Impact
147
+
148
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
149
+
150
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
151
+
152
+ - **Hardware Type:** [More Information Needed]
153
+ - **Hours used:** [More Information Needed]
154
+ - **Cloud Provider:** [More Information Needed]
155
+ - **Compute Region:** [More Information Needed]
156
+ - **Carbon Emitted:** [More Information Needed]
157
+
158
+ ## Technical Specifications [optional]
159
+
160
+ ### Model Architecture and Objective
161
+
162
+ [More Information Needed]
163
+
164
+ ### Compute Infrastructure
165
+
166
+ [More Information Needed]
167
+
168
+ #### Hardware
169
+
170
+ [More Information Needed]
171
+
172
+ #### Software
173
+
174
+ [More Information Needed]
175
+
176
+ ## Citation [optional]
177
+
178
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
179
+
180
+ **BibTeX:**
181
+
182
+ [More Information Needed]
183
+
184
+ **APA:**
185
+
186
+ [More Information Needed]
187
+
188
+ ## Glossary [optional]
189
+
190
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
191
+
192
+ [More Information Needed]
193
+
194
+ ## More Information [optional]
195
+
196
+ [More Information Needed]
197
+
198
+ ## Model Card Authors [optional]
199
+
200
+ [More Information Needed]
201
+
202
+ ## Model Card Contact
203
+
204
+ [More Information Needed]
205
+ ### Framework versions
206
+
207
+ - PEFT 0.17.0
checkpoint-1000/adapter_config.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "distilgpt2",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 32,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.05,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "qalora_group_size": 16,
24
+ "r": 16,
25
+ "rank_pattern": {},
26
+ "revision": null,
27
+ "target_modules": [
28
+ "c_proj",
29
+ "c_attn"
30
+ ],
31
+ "target_parameters": null,
32
+ "task_type": "CAUSAL_LM",
33
+ "trainable_token_indices": null,
34
+ "use_dora": false,
35
+ "use_qalora": false,
36
+ "use_rslora": false
37
+ }
checkpoint-1000/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e8ee6d4fa553a2205fc2ed4b4c0689a9f4326d5bb00e3141f133960c2b8e19a
3
+ size 3248616
checkpoint-1000/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-1000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37207bf789a542293a103be9ff481f7bb2af2224225965f624ac804e8b2e1003
3
+ size 6518714
checkpoint-1000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ef8e0419b49b0ef5350645fe4a8fecb5ac089d0189be793fc01e31fb99505eb
3
+ size 14244
checkpoint-1000/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d8fdcd0311eba9854fff738038ed4c1a269832665b4d88ba4e4e3d02a1a7e0e
3
+ size 988
checkpoint-1000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:72668d829dbdb7f9fcf02b26ca0fd56bd5ec9500d2675771791441d2766c2e2c
3
+ size 1064
checkpoint-1000/special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|endoftext|>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<|endoftext|>",
18
+ "lstrip": false,
19
+ "normalized": true,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<|endoftext|>",
25
+ "lstrip": false,
26
+ "normalized": true,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
checkpoint-1000/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-1000/tokenizer_config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "50256": {
5
+ "content": "<|endoftext|>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ }
12
+ },
13
+ "bos_token": "<|endoftext|>",
14
+ "clean_up_tokenization_spaces": false,
15
+ "eos_token": "<|endoftext|>",
16
+ "extra_special_tokens": {},
17
+ "max_length": 180,
18
+ "model_max_length": 1024,
19
+ "pad_to_multiple_of": null,
20
+ "pad_token": "<|endoftext|>",
21
+ "pad_token_type_id": 0,
22
+ "padding_side": "right",
23
+ "stride": 0,
24
+ "tokenizer_class": "GPT2Tokenizer",
25
+ "truncation_side": "right",
26
+ "truncation_strategy": "longest_first",
27
+ "unk_token": "<|endoftext|>"
28
+ }
checkpoint-1000/trainer_state.json ADDED
@@ -0,0 +1,85 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 1000,
3
+ "best_metric": 0.9456237554550171,
4
+ "best_model_checkpoint": "./banking-sms-parser-v10/checkpoint-1000",
5
+ "epoch": 0.2962962962962963,
6
+ "eval_steps": 500,
7
+ "global_step": 1000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.05925925925925926,
14
+ "grad_norm": 0.47084560990333557,
15
+ "learning_rate": 0.0002955777777777778,
16
+ "loss": 2.0436,
17
+ "step": 200
18
+ },
19
+ {
20
+ "epoch": 0.11851851851851852,
21
+ "grad_norm": 0.40365636348724365,
22
+ "learning_rate": 0.0002911333333333333,
23
+ "loss": 1.3594,
24
+ "step": 400
25
+ },
26
+ {
27
+ "epoch": 0.14814814814814814,
28
+ "eval_loss": 1.1683876514434814,
29
+ "eval_runtime": 29.5095,
30
+ "eval_samples_per_second": 101.662,
31
+ "eval_steps_per_second": 12.708,
32
+ "step": 500
33
+ },
34
+ {
35
+ "epoch": 0.17777777777777778,
36
+ "grad_norm": 0.5327613353729248,
37
+ "learning_rate": 0.0002866888888888889,
38
+ "loss": 1.2262,
39
+ "step": 600
40
+ },
41
+ {
42
+ "epoch": 0.23703703703703705,
43
+ "grad_norm": 0.5959504246711731,
44
+ "learning_rate": 0.0002822444444444444,
45
+ "loss": 1.1381,
46
+ "step": 800
47
+ },
48
+ {
49
+ "epoch": 0.2962962962962963,
50
+ "grad_norm": 0.4996171295642853,
51
+ "learning_rate": 0.0002778,
52
+ "loss": 1.066,
53
+ "step": 1000
54
+ },
55
+ {
56
+ "epoch": 0.2962962962962963,
57
+ "eval_loss": 0.9456237554550171,
58
+ "eval_runtime": 29.4299,
59
+ "eval_samples_per_second": 101.937,
60
+ "eval_steps_per_second": 12.742,
61
+ "step": 1000
62
+ }
63
+ ],
64
+ "logging_steps": 200,
65
+ "max_steps": 13500,
66
+ "num_input_tokens_seen": 0,
67
+ "num_train_epochs": 4,
68
+ "save_steps": 1000,
69
+ "stateful_callbacks": {
70
+ "TrainerControl": {
71
+ "args": {
72
+ "should_epoch_stop": false,
73
+ "should_evaluate": false,
74
+ "should_log": false,
75
+ "should_save": true,
76
+ "should_training_stop": false
77
+ },
78
+ "attributes": {}
79
+ }
80
+ },
81
+ "total_flos": 665698959360000.0,
82
+ "train_batch_size": 8,
83
+ "trial_name": null,
84
+ "trial_params": null
85
+ }
checkpoint-1000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5b2e2a527ba622332dbb2398486338b6fc135483c9c1c937ef5cefadf332325
3
+ size 5304
checkpoint-1000/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-10000/README.md ADDED
@@ -0,0 +1,207 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: distilgpt2
3
+ library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - base_model:adapter:distilgpt2
7
+ - lora
8
+ - transformers
9
+ ---
10
+
11
+ # Model Card for Model ID
12
+
13
+ <!-- Provide a quick summary of what the model is/does. -->
14
+
15
+
16
+
17
+ ## Model Details
18
+
19
+ ### Model Description
20
+
21
+ <!-- Provide a longer summary of what this model is. -->
22
+
23
+
24
+
25
+ - **Developed by:** [More Information Needed]
26
+ - **Funded by [optional]:** [More Information Needed]
27
+ - **Shared by [optional]:** [More Information Needed]
28
+ - **Model type:** [More Information Needed]
29
+ - **Language(s) (NLP):** [More Information Needed]
30
+ - **License:** [More Information Needed]
31
+ - **Finetuned from model [optional]:** [More Information Needed]
32
+
33
+ ### Model Sources [optional]
34
+
35
+ <!-- Provide the basic links for the model. -->
36
+
37
+ - **Repository:** [More Information Needed]
38
+ - **Paper [optional]:** [More Information Needed]
39
+ - **Demo [optional]:** [More Information Needed]
40
+
41
+ ## Uses
42
+
43
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
44
+
45
+ ### Direct Use
46
+
47
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
48
+
49
+ [More Information Needed]
50
+
51
+ ### Downstream Use [optional]
52
+
53
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
54
+
55
+ [More Information Needed]
56
+
57
+ ### Out-of-Scope Use
58
+
59
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
60
+
61
+ [More Information Needed]
62
+
63
+ ## Bias, Risks, and Limitations
64
+
65
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
66
+
67
+ [More Information Needed]
68
+
69
+ ### Recommendations
70
+
71
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
72
+
73
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
74
+
75
+ ## How to Get Started with the Model
76
+
77
+ Use the code below to get started with the model.
78
+
79
+ [More Information Needed]
80
+
81
+ ## Training Details
82
+
83
+ ### Training Data
84
+
85
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
86
+
87
+ [More Information Needed]
88
+
89
+ ### Training Procedure
90
+
91
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
92
+
93
+ #### Preprocessing [optional]
94
+
95
+ [More Information Needed]
96
+
97
+
98
+ #### Training Hyperparameters
99
+
100
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
101
+
102
+ #### Speeds, Sizes, Times [optional]
103
+
104
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
105
+
106
+ [More Information Needed]
107
+
108
+ ## Evaluation
109
+
110
+ <!-- This section describes the evaluation protocols and provides the results. -->
111
+
112
+ ### Testing Data, Factors & Metrics
113
+
114
+ #### Testing Data
115
+
116
+ <!-- This should link to a Dataset Card if possible. -->
117
+
118
+ [More Information Needed]
119
+
120
+ #### Factors
121
+
122
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
123
+
124
+ [More Information Needed]
125
+
126
+ #### Metrics
127
+
128
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
129
+
130
+ [More Information Needed]
131
+
132
+ ### Results
133
+
134
+ [More Information Needed]
135
+
136
+ #### Summary
137
+
138
+
139
+
140
+ ## Model Examination [optional]
141
+
142
+ <!-- Relevant interpretability work for the model goes here -->
143
+
144
+ [More Information Needed]
145
+
146
+ ## Environmental Impact
147
+
148
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
149
+
150
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
151
+
152
+ - **Hardware Type:** [More Information Needed]
153
+ - **Hours used:** [More Information Needed]
154
+ - **Cloud Provider:** [More Information Needed]
155
+ - **Compute Region:** [More Information Needed]
156
+ - **Carbon Emitted:** [More Information Needed]
157
+
158
+ ## Technical Specifications [optional]
159
+
160
+ ### Model Architecture and Objective
161
+
162
+ [More Information Needed]
163
+
164
+ ### Compute Infrastructure
165
+
166
+ [More Information Needed]
167
+
168
+ #### Hardware
169
+
170
+ [More Information Needed]
171
+
172
+ #### Software
173
+
174
+ [More Information Needed]
175
+
176
+ ## Citation [optional]
177
+
178
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
179
+
180
+ **BibTeX:**
181
+
182
+ [More Information Needed]
183
+
184
+ **APA:**
185
+
186
+ [More Information Needed]
187
+
188
+ ## Glossary [optional]
189
+
190
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
191
+
192
+ [More Information Needed]
193
+
194
+ ## More Information [optional]
195
+
196
+ [More Information Needed]
197
+
198
+ ## Model Card Authors [optional]
199
+
200
+ [More Information Needed]
201
+
202
+ ## Model Card Contact
203
+
204
+ [More Information Needed]
205
+ ### Framework versions
206
+
207
+ - PEFT 0.17.0
checkpoint-10000/adapter_config.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "distilgpt2",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 32,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.05,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "qalora_group_size": 16,
24
+ "r": 16,
25
+ "rank_pattern": {},
26
+ "revision": null,
27
+ "target_modules": [
28
+ "c_proj",
29
+ "c_attn"
30
+ ],
31
+ "target_parameters": null,
32
+ "task_type": "CAUSAL_LM",
33
+ "trainable_token_indices": null,
34
+ "use_dora": false,
35
+ "use_qalora": false,
36
+ "use_rslora": false
37
+ }
checkpoint-10000/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e469b5b67b9cfb1e04ccb7dfc3e2a21c24e5d78f4240d8b68a775f4b8603fe01
3
+ size 3248616
checkpoint-10000/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-10000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b446e97e2a7c87bc3aad2815aa8460f724773ccd21cdd0d5511c7d66aaa56ef4
3
+ size 6518714
checkpoint-10000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:281ee8a77a8e6d9db96e3a879e0288c8233833642fc0cf3abcf042aa7b35e32b
3
+ size 14244
checkpoint-10000/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c072a285b6f4670d281023d2fdc827b249e4a784e3ebc49cfbe50a54914c99b
3
+ size 988
checkpoint-10000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7173c7464031a760641151afa8096faf47fe8bf592067a6c5fec7ca4e6d8ec9a
3
+ size 1064
checkpoint-10000/special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|endoftext|>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<|endoftext|>",
18
+ "lstrip": false,
19
+ "normalized": true,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<|endoftext|>",
25
+ "lstrip": false,
26
+ "normalized": true,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
checkpoint-10000/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-10000/tokenizer_config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "50256": {
5
+ "content": "<|endoftext|>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ }
12
+ },
13
+ "bos_token": "<|endoftext|>",
14
+ "clean_up_tokenization_spaces": false,
15
+ "eos_token": "<|endoftext|>",
16
+ "extra_special_tokens": {},
17
+ "max_length": 180,
18
+ "model_max_length": 1024,
19
+ "pad_to_multiple_of": null,
20
+ "pad_token": "<|endoftext|>",
21
+ "pad_token_type_id": 0,
22
+ "padding_side": "right",
23
+ "stride": 0,
24
+ "tokenizer_class": "GPT2Tokenizer",
25
+ "truncation_side": "right",
26
+ "truncation_strategy": "longest_first",
27
+ "unk_token": "<|endoftext|>"
28
+ }
checkpoint-10000/trainer_state.json ADDED
@@ -0,0 +1,544 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 10000,
3
+ "best_metric": 0.46695125102996826,
4
+ "best_model_checkpoint": "./banking-sms-parser-v10/checkpoint-10000",
5
+ "epoch": 2.962962962962963,
6
+ "eval_steps": 500,
7
+ "global_step": 10000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.05925925925925926,
14
+ "grad_norm": 0.47084560990333557,
15
+ "learning_rate": 0.0002955777777777778,
16
+ "loss": 2.0436,
17
+ "step": 200
18
+ },
19
+ {
20
+ "epoch": 0.11851851851851852,
21
+ "grad_norm": 0.40365636348724365,
22
+ "learning_rate": 0.0002911333333333333,
23
+ "loss": 1.3594,
24
+ "step": 400
25
+ },
26
+ {
27
+ "epoch": 0.14814814814814814,
28
+ "eval_loss": 1.1683876514434814,
29
+ "eval_runtime": 29.5095,
30
+ "eval_samples_per_second": 101.662,
31
+ "eval_steps_per_second": 12.708,
32
+ "step": 500
33
+ },
34
+ {
35
+ "epoch": 0.17777777777777778,
36
+ "grad_norm": 0.5327613353729248,
37
+ "learning_rate": 0.0002866888888888889,
38
+ "loss": 1.2262,
39
+ "step": 600
40
+ },
41
+ {
42
+ "epoch": 0.23703703703703705,
43
+ "grad_norm": 0.5959504246711731,
44
+ "learning_rate": 0.0002822444444444444,
45
+ "loss": 1.1381,
46
+ "step": 800
47
+ },
48
+ {
49
+ "epoch": 0.2962962962962963,
50
+ "grad_norm": 0.4996171295642853,
51
+ "learning_rate": 0.0002778,
52
+ "loss": 1.066,
53
+ "step": 1000
54
+ },
55
+ {
56
+ "epoch": 0.2962962962962963,
57
+ "eval_loss": 0.9456237554550171,
58
+ "eval_runtime": 29.4299,
59
+ "eval_samples_per_second": 101.937,
60
+ "eval_steps_per_second": 12.742,
61
+ "step": 1000
62
+ },
63
+ {
64
+ "epoch": 0.35555555555555557,
65
+ "grad_norm": 0.47130197286605835,
66
+ "learning_rate": 0.0002733555555555555,
67
+ "loss": 1.0067,
68
+ "step": 1200
69
+ },
70
+ {
71
+ "epoch": 0.4148148148148148,
72
+ "grad_norm": 0.693696916103363,
73
+ "learning_rate": 0.0002689111111111111,
74
+ "loss": 0.969,
75
+ "step": 1400
76
+ },
77
+ {
78
+ "epoch": 0.4444444444444444,
79
+ "eval_loss": 0.8003639578819275,
80
+ "eval_runtime": 29.3553,
81
+ "eval_samples_per_second": 102.196,
82
+ "eval_steps_per_second": 12.775,
83
+ "step": 1500
84
+ },
85
+ {
86
+ "epoch": 0.4740740740740741,
87
+ "grad_norm": 0.5786986351013184,
88
+ "learning_rate": 0.0002644666666666666,
89
+ "loss": 0.9081,
90
+ "step": 1600
91
+ },
92
+ {
93
+ "epoch": 0.5333333333333333,
94
+ "grad_norm": 0.5209688544273376,
95
+ "learning_rate": 0.0002600222222222222,
96
+ "loss": 0.8801,
97
+ "step": 1800
98
+ },
99
+ {
100
+ "epoch": 0.5925925925925926,
101
+ "grad_norm": 0.7497162818908691,
102
+ "learning_rate": 0.0002555777777777778,
103
+ "loss": 0.8183,
104
+ "step": 2000
105
+ },
106
+ {
107
+ "epoch": 0.5925925925925926,
108
+ "eval_loss": 0.7105592489242554,
109
+ "eval_runtime": 29.7273,
110
+ "eval_samples_per_second": 100.917,
111
+ "eval_steps_per_second": 12.615,
112
+ "step": 2000
113
+ },
114
+ {
115
+ "epoch": 0.6518518518518519,
116
+ "grad_norm": 0.7358675599098206,
117
+ "learning_rate": 0.00025113333333333333,
118
+ "loss": 0.809,
119
+ "step": 2200
120
+ },
121
+ {
122
+ "epoch": 0.7111111111111111,
123
+ "grad_norm": 0.5220760107040405,
124
+ "learning_rate": 0.0002466888888888889,
125
+ "loss": 0.7799,
126
+ "step": 2400
127
+ },
128
+ {
129
+ "epoch": 0.7407407407407407,
130
+ "eval_loss": 0.6572133898735046,
131
+ "eval_runtime": 29.3661,
132
+ "eval_samples_per_second": 102.158,
133
+ "eval_steps_per_second": 12.77,
134
+ "step": 2500
135
+ },
136
+ {
137
+ "epoch": 0.7703703703703704,
138
+ "grad_norm": 0.4607163965702057,
139
+ "learning_rate": 0.00024224444444444443,
140
+ "loss": 0.7557,
141
+ "step": 2600
142
+ },
143
+ {
144
+ "epoch": 0.8296296296296296,
145
+ "grad_norm": 0.5450932383537292,
146
+ "learning_rate": 0.00023779999999999998,
147
+ "loss": 0.7264,
148
+ "step": 2800
149
+ },
150
+ {
151
+ "epoch": 0.8888888888888888,
152
+ "grad_norm": 0.5149521231651306,
153
+ "learning_rate": 0.00023335555555555556,
154
+ "loss": 0.7054,
155
+ "step": 3000
156
+ },
157
+ {
158
+ "epoch": 0.8888888888888888,
159
+ "eval_loss": 0.6189373731613159,
160
+ "eval_runtime": 29.6913,
161
+ "eval_samples_per_second": 101.04,
162
+ "eval_steps_per_second": 12.63,
163
+ "step": 3000
164
+ },
165
+ {
166
+ "epoch": 0.9481481481481482,
167
+ "grad_norm": 0.5345056056976318,
168
+ "learning_rate": 0.0002289111111111111,
169
+ "loss": 0.7026,
170
+ "step": 3200
171
+ },
172
+ {
173
+ "epoch": 1.0074074074074073,
174
+ "grad_norm": 0.5385560989379883,
175
+ "learning_rate": 0.00022446666666666666,
176
+ "loss": 0.6888,
177
+ "step": 3400
178
+ },
179
+ {
180
+ "epoch": 1.037037037037037,
181
+ "eval_loss": 0.5899662971496582,
182
+ "eval_runtime": 29.4218,
183
+ "eval_samples_per_second": 101.965,
184
+ "eval_steps_per_second": 12.746,
185
+ "step": 3500
186
+ },
187
+ {
188
+ "epoch": 1.0666666666666667,
189
+ "grad_norm": 0.5609071850776672,
190
+ "learning_rate": 0.0002200222222222222,
191
+ "loss": 0.6671,
192
+ "step": 3600
193
+ },
194
+ {
195
+ "epoch": 1.125925925925926,
196
+ "grad_norm": 0.5616484880447388,
197
+ "learning_rate": 0.00021557777777777776,
198
+ "loss": 0.6636,
199
+ "step": 3800
200
+ },
201
+ {
202
+ "epoch": 1.1851851851851851,
203
+ "grad_norm": 0.5596534609794617,
204
+ "learning_rate": 0.0002111333333333333,
205
+ "loss": 0.6515,
206
+ "step": 4000
207
+ },
208
+ {
209
+ "epoch": 1.1851851851851851,
210
+ "eval_loss": 0.5668333172798157,
211
+ "eval_runtime": 29.3385,
212
+ "eval_samples_per_second": 102.255,
213
+ "eval_steps_per_second": 12.782,
214
+ "step": 4000
215
+ },
216
+ {
217
+ "epoch": 1.2444444444444445,
218
+ "grad_norm": 0.5900132656097412,
219
+ "learning_rate": 0.00020668888888888888,
220
+ "loss": 0.6409,
221
+ "step": 4200
222
+ },
223
+ {
224
+ "epoch": 1.3037037037037038,
225
+ "grad_norm": 0.6175922751426697,
226
+ "learning_rate": 0.00020224444444444443,
227
+ "loss": 0.641,
228
+ "step": 4400
229
+ },
230
+ {
231
+ "epoch": 1.3333333333333333,
232
+ "eval_loss": 0.5480825304985046,
233
+ "eval_runtime": 29.457,
234
+ "eval_samples_per_second": 101.843,
235
+ "eval_steps_per_second": 12.73,
236
+ "step": 4500
237
+ },
238
+ {
239
+ "epoch": 1.362962962962963,
240
+ "grad_norm": 0.42820265889167786,
241
+ "learning_rate": 0.00019779999999999998,
242
+ "loss": 0.6212,
243
+ "step": 4600
244
+ },
245
+ {
246
+ "epoch": 1.4222222222222223,
247
+ "grad_norm": 0.7179502248764038,
248
+ "learning_rate": 0.00019335555555555553,
249
+ "loss": 0.6268,
250
+ "step": 4800
251
+ },
252
+ {
253
+ "epoch": 1.4814814814814814,
254
+ "grad_norm": 0.4201723635196686,
255
+ "learning_rate": 0.00018891111111111108,
256
+ "loss": 0.6083,
257
+ "step": 5000
258
+ },
259
+ {
260
+ "epoch": 1.4814814814814814,
261
+ "eval_loss": 0.538067638874054,
262
+ "eval_runtime": 29.3858,
263
+ "eval_samples_per_second": 102.09,
264
+ "eval_steps_per_second": 12.761,
265
+ "step": 5000
266
+ },
267
+ {
268
+ "epoch": 1.5407407407407407,
269
+ "grad_norm": 0.6730241775512695,
270
+ "learning_rate": 0.00018446666666666666,
271
+ "loss": 0.6148,
272
+ "step": 5200
273
+ },
274
+ {
275
+ "epoch": 1.6,
276
+ "grad_norm": 0.5300161242485046,
277
+ "learning_rate": 0.0001800222222222222,
278
+ "loss": 0.6001,
279
+ "step": 5400
280
+ },
281
+ {
282
+ "epoch": 1.6296296296296298,
283
+ "eval_loss": 0.5220749378204346,
284
+ "eval_runtime": 29.4236,
285
+ "eval_samples_per_second": 101.959,
286
+ "eval_steps_per_second": 12.745,
287
+ "step": 5500
288
+ },
289
+ {
290
+ "epoch": 1.6592592592592592,
291
+ "grad_norm": 0.45725536346435547,
292
+ "learning_rate": 0.00017557777777777776,
293
+ "loss": 0.605,
294
+ "step": 5600
295
+ },
296
+ {
297
+ "epoch": 1.7185185185185186,
298
+ "grad_norm": 0.49325183033943176,
299
+ "learning_rate": 0.00017115555555555554,
300
+ "loss": 0.6024,
301
+ "step": 5800
302
+ },
303
+ {
304
+ "epoch": 1.7777777777777777,
305
+ "grad_norm": 0.5796788930892944,
306
+ "learning_rate": 0.00016671111111111109,
307
+ "loss": 0.5941,
308
+ "step": 6000
309
+ },
310
+ {
311
+ "epoch": 1.7777777777777777,
312
+ "eval_loss": 0.5122686624526978,
313
+ "eval_runtime": 29.5133,
314
+ "eval_samples_per_second": 101.649,
315
+ "eval_steps_per_second": 12.706,
316
+ "step": 6000
317
+ },
318
+ {
319
+ "epoch": 1.837037037037037,
320
+ "grad_norm": 0.5793591737747192,
321
+ "learning_rate": 0.00016226666666666664,
322
+ "loss": 0.5809,
323
+ "step": 6200
324
+ },
325
+ {
326
+ "epoch": 1.8962962962962964,
327
+ "grad_norm": 0.4882641136646271,
328
+ "learning_rate": 0.00015782222222222219,
329
+ "loss": 0.5869,
330
+ "step": 6400
331
+ },
332
+ {
333
+ "epoch": 1.925925925925926,
334
+ "eval_loss": 0.5018052458763123,
335
+ "eval_runtime": 29.2669,
336
+ "eval_samples_per_second": 102.505,
337
+ "eval_steps_per_second": 12.813,
338
+ "step": 6500
339
+ },
340
+ {
341
+ "epoch": 1.9555555555555557,
342
+ "grad_norm": 0.5069060325622559,
343
+ "learning_rate": 0.0001533777777777778,
344
+ "loss": 0.5812,
345
+ "step": 6600
346
+ },
347
+ {
348
+ "epoch": 2.0148148148148146,
349
+ "grad_norm": 0.44775187969207764,
350
+ "learning_rate": 0.00014893333333333334,
351
+ "loss": 0.5397,
352
+ "step": 6800
353
+ },
354
+ {
355
+ "epoch": 2.074074074074074,
356
+ "grad_norm": 0.5677600502967834,
357
+ "learning_rate": 0.0001444888888888889,
358
+ "loss": 0.5567,
359
+ "step": 7000
360
+ },
361
+ {
362
+ "epoch": 2.074074074074074,
363
+ "eval_loss": 0.49609723687171936,
364
+ "eval_runtime": 29.4825,
365
+ "eval_samples_per_second": 101.755,
366
+ "eval_steps_per_second": 12.719,
367
+ "step": 7000
368
+ },
369
+ {
370
+ "epoch": 2.1333333333333333,
371
+ "grad_norm": 0.5369960069656372,
372
+ "learning_rate": 0.00014004444444444444,
373
+ "loss": 0.5591,
374
+ "step": 7200
375
+ },
376
+ {
377
+ "epoch": 2.1925925925925926,
378
+ "grad_norm": 0.353333055973053,
379
+ "learning_rate": 0.0001356,
380
+ "loss": 0.5717,
381
+ "step": 7400
382
+ },
383
+ {
384
+ "epoch": 2.2222222222222223,
385
+ "eval_loss": 0.49044421315193176,
386
+ "eval_runtime": 29.573,
387
+ "eval_samples_per_second": 101.444,
388
+ "eval_steps_per_second": 12.68,
389
+ "step": 7500
390
+ },
391
+ {
392
+ "epoch": 2.251851851851852,
393
+ "grad_norm": 0.45505258440971375,
394
+ "learning_rate": 0.00013115555555555554,
395
+ "loss": 0.5576,
396
+ "step": 7600
397
+ },
398
+ {
399
+ "epoch": 2.311111111111111,
400
+ "grad_norm": 0.4602493643760681,
401
+ "learning_rate": 0.00012671111111111112,
402
+ "loss": 0.5407,
403
+ "step": 7800
404
+ },
405
+ {
406
+ "epoch": 2.3703703703703702,
407
+ "grad_norm": 0.4852631688117981,
408
+ "learning_rate": 0.00012226666666666667,
409
+ "loss": 0.5523,
410
+ "step": 8000
411
+ },
412
+ {
413
+ "epoch": 2.3703703703703702,
414
+ "eval_loss": 0.48393580317497253,
415
+ "eval_runtime": 29.3448,
416
+ "eval_samples_per_second": 102.233,
417
+ "eval_steps_per_second": 12.779,
418
+ "step": 8000
419
+ },
420
+ {
421
+ "epoch": 2.4296296296296296,
422
+ "grad_norm": 0.4038563668727875,
423
+ "learning_rate": 0.0001178222222222222,
424
+ "loss": 0.5313,
425
+ "step": 8200
426
+ },
427
+ {
428
+ "epoch": 2.488888888888889,
429
+ "grad_norm": 0.5276423692703247,
430
+ "learning_rate": 0.00011339999999999999,
431
+ "loss": 0.5489,
432
+ "step": 8400
433
+ },
434
+ {
435
+ "epoch": 2.5185185185185186,
436
+ "eval_loss": 0.47875121235847473,
437
+ "eval_runtime": 29.4918,
438
+ "eval_samples_per_second": 101.723,
439
+ "eval_steps_per_second": 12.715,
440
+ "step": 8500
441
+ },
442
+ {
443
+ "epoch": 2.5481481481481483,
444
+ "grad_norm": 0.539804995059967,
445
+ "learning_rate": 0.00010895555555555554,
446
+ "loss": 0.5456,
447
+ "step": 8600
448
+ },
449
+ {
450
+ "epoch": 2.6074074074074076,
451
+ "grad_norm": 0.629550039768219,
452
+ "learning_rate": 0.0001045111111111111,
453
+ "loss": 0.5523,
454
+ "step": 8800
455
+ },
456
+ {
457
+ "epoch": 2.6666666666666665,
458
+ "grad_norm": 0.6091557741165161,
459
+ "learning_rate": 0.00010006666666666666,
460
+ "loss": 0.5358,
461
+ "step": 9000
462
+ },
463
+ {
464
+ "epoch": 2.6666666666666665,
465
+ "eval_loss": 0.4742780327796936,
466
+ "eval_runtime": 29.4044,
467
+ "eval_samples_per_second": 102.026,
468
+ "eval_steps_per_second": 12.753,
469
+ "step": 9000
470
+ },
471
+ {
472
+ "epoch": 2.725925925925926,
473
+ "grad_norm": 0.5220003724098206,
474
+ "learning_rate": 9.56222222222222e-05,
475
+ "loss": 0.5398,
476
+ "step": 9200
477
+ },
478
+ {
479
+ "epoch": 2.785185185185185,
480
+ "grad_norm": 0.4543071389198303,
481
+ "learning_rate": 9.117777777777777e-05,
482
+ "loss": 0.53,
483
+ "step": 9400
484
+ },
485
+ {
486
+ "epoch": 2.814814814814815,
487
+ "eval_loss": 0.4698003828525543,
488
+ "eval_runtime": 29.5022,
489
+ "eval_samples_per_second": 101.687,
490
+ "eval_steps_per_second": 12.711,
491
+ "step": 9500
492
+ },
493
+ {
494
+ "epoch": 2.8444444444444446,
495
+ "grad_norm": 0.4849318265914917,
496
+ "learning_rate": 8.673333333333332e-05,
497
+ "loss": 0.5278,
498
+ "step": 9600
499
+ },
500
+ {
501
+ "epoch": 2.9037037037037035,
502
+ "grad_norm": 0.40453964471817017,
503
+ "learning_rate": 8.228888888888888e-05,
504
+ "loss": 0.5448,
505
+ "step": 9800
506
+ },
507
+ {
508
+ "epoch": 2.962962962962963,
509
+ "grad_norm": 0.5250543355941772,
510
+ "learning_rate": 7.784444444444443e-05,
511
+ "loss": 0.538,
512
+ "step": 10000
513
+ },
514
+ {
515
+ "epoch": 2.962962962962963,
516
+ "eval_loss": 0.46695125102996826,
517
+ "eval_runtime": 29.5114,
518
+ "eval_samples_per_second": 101.656,
519
+ "eval_steps_per_second": 12.707,
520
+ "step": 10000
521
+ }
522
+ ],
523
+ "logging_steps": 200,
524
+ "max_steps": 13500,
525
+ "num_input_tokens_seen": 0,
526
+ "num_train_epochs": 4,
527
+ "save_steps": 1000,
528
+ "stateful_callbacks": {
529
+ "TrainerControl": {
530
+ "args": {
531
+ "should_epoch_stop": false,
532
+ "should_evaluate": false,
533
+ "should_log": false,
534
+ "should_save": true,
535
+ "should_training_stop": false
536
+ },
537
+ "attributes": {}
538
+ }
539
+ },
540
+ "total_flos": 6656989593600000.0,
541
+ "train_batch_size": 8,
542
+ "trial_name": null,
543
+ "trial_params": null
544
+ }
checkpoint-10000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5b2e2a527ba622332dbb2398486338b6fc135483c9c1c937ef5cefadf332325
3
+ size 5304
checkpoint-10000/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-11000/README.md ADDED
@@ -0,0 +1,207 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: distilgpt2
3
+ library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - base_model:adapter:distilgpt2
7
+ - lora
8
+ - transformers
9
+ ---
10
+
11
+ # Model Card for Model ID
12
+
13
+ <!-- Provide a quick summary of what the model is/does. -->
14
+
15
+
16
+
17
+ ## Model Details
18
+
19
+ ### Model Description
20
+
21
+ <!-- Provide a longer summary of what this model is. -->
22
+
23
+
24
+
25
+ - **Developed by:** [More Information Needed]
26
+ - **Funded by [optional]:** [More Information Needed]
27
+ - **Shared by [optional]:** [More Information Needed]
28
+ - **Model type:** [More Information Needed]
29
+ - **Language(s) (NLP):** [More Information Needed]
30
+ - **License:** [More Information Needed]
31
+ - **Finetuned from model [optional]:** [More Information Needed]
32
+
33
+ ### Model Sources [optional]
34
+
35
+ <!-- Provide the basic links for the model. -->
36
+
37
+ - **Repository:** [More Information Needed]
38
+ - **Paper [optional]:** [More Information Needed]
39
+ - **Demo [optional]:** [More Information Needed]
40
+
41
+ ## Uses
42
+
43
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
44
+
45
+ ### Direct Use
46
+
47
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
48
+
49
+ [More Information Needed]
50
+
51
+ ### Downstream Use [optional]
52
+
53
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
54
+
55
+ [More Information Needed]
56
+
57
+ ### Out-of-Scope Use
58
+
59
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
60
+
61
+ [More Information Needed]
62
+
63
+ ## Bias, Risks, and Limitations
64
+
65
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
66
+
67
+ [More Information Needed]
68
+
69
+ ### Recommendations
70
+
71
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
72
+
73
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
74
+
75
+ ## How to Get Started with the Model
76
+
77
+ Use the code below to get started with the model.
78
+
79
+ [More Information Needed]
80
+
81
+ ## Training Details
82
+
83
+ ### Training Data
84
+
85
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
86
+
87
+ [More Information Needed]
88
+
89
+ ### Training Procedure
90
+
91
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
92
+
93
+ #### Preprocessing [optional]
94
+
95
+ [More Information Needed]
96
+
97
+
98
+ #### Training Hyperparameters
99
+
100
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
101
+
102
+ #### Speeds, Sizes, Times [optional]
103
+
104
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
105
+
106
+ [More Information Needed]
107
+
108
+ ## Evaluation
109
+
110
+ <!-- This section describes the evaluation protocols and provides the results. -->
111
+
112
+ ### Testing Data, Factors & Metrics
113
+
114
+ #### Testing Data
115
+
116
+ <!-- This should link to a Dataset Card if possible. -->
117
+
118
+ [More Information Needed]
119
+
120
+ #### Factors
121
+
122
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
123
+
124
+ [More Information Needed]
125
+
126
+ #### Metrics
127
+
128
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
129
+
130
+ [More Information Needed]
131
+
132
+ ### Results
133
+
134
+ [More Information Needed]
135
+
136
+ #### Summary
137
+
138
+
139
+
140
+ ## Model Examination [optional]
141
+
142
+ <!-- Relevant interpretability work for the model goes here -->
143
+
144
+ [More Information Needed]
145
+
146
+ ## Environmental Impact
147
+
148
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
149
+
150
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
151
+
152
+ - **Hardware Type:** [More Information Needed]
153
+ - **Hours used:** [More Information Needed]
154
+ - **Cloud Provider:** [More Information Needed]
155
+ - **Compute Region:** [More Information Needed]
156
+ - **Carbon Emitted:** [More Information Needed]
157
+
158
+ ## Technical Specifications [optional]
159
+
160
+ ### Model Architecture and Objective
161
+
162
+ [More Information Needed]
163
+
164
+ ### Compute Infrastructure
165
+
166
+ [More Information Needed]
167
+
168
+ #### Hardware
169
+
170
+ [More Information Needed]
171
+
172
+ #### Software
173
+
174
+ [More Information Needed]
175
+
176
+ ## Citation [optional]
177
+
178
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
179
+
180
+ **BibTeX:**
181
+
182
+ [More Information Needed]
183
+
184
+ **APA:**
185
+
186
+ [More Information Needed]
187
+
188
+ ## Glossary [optional]
189
+
190
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
191
+
192
+ [More Information Needed]
193
+
194
+ ## More Information [optional]
195
+
196
+ [More Information Needed]
197
+
198
+ ## Model Card Authors [optional]
199
+
200
+ [More Information Needed]
201
+
202
+ ## Model Card Contact
203
+
204
+ [More Information Needed]
205
+ ### Framework versions
206
+
207
+ - PEFT 0.17.0
checkpoint-11000/adapter_config.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "distilgpt2",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 32,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.05,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "qalora_group_size": 16,
24
+ "r": 16,
25
+ "rank_pattern": {},
26
+ "revision": null,
27
+ "target_modules": [
28
+ "c_proj",
29
+ "c_attn"
30
+ ],
31
+ "target_parameters": null,
32
+ "task_type": "CAUSAL_LM",
33
+ "trainable_token_indices": null,
34
+ "use_dora": false,
35
+ "use_qalora": false,
36
+ "use_rslora": false
37
+ }
checkpoint-11000/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:584793f3f1527007903559128e2fe5b814b9b5f4891a3af426913102ce2eb4af
3
+ size 3248616
checkpoint-11000/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-11000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9751901d1352fac6f2e365d2202a3dc2d03e11f9562f8ab51b7008943368a95
3
+ size 6518714
checkpoint-11000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:963ca21a536f8365ddeacfa23786fde574fe723ceac54374eed1c1ee982bb3ee
3
+ size 14244
checkpoint-11000/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93eff9e1f9576ce4414c6b4d8adfd7479e9fd6585ac64d4041c7aeec38be28c7
3
+ size 988
checkpoint-11000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7aa624420f96a2e9654a0864e1ec4d155282f0b0eacfc3ce38b8b5bbbf1b8d57
3
+ size 1064
checkpoint-11000/special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|endoftext|>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<|endoftext|>",
18
+ "lstrip": false,
19
+ "normalized": true,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<|endoftext|>",
25
+ "lstrip": false,
26
+ "normalized": true,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
checkpoint-11000/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-11000/tokenizer_config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "50256": {
5
+ "content": "<|endoftext|>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ }
12
+ },
13
+ "bos_token": "<|endoftext|>",
14
+ "clean_up_tokenization_spaces": false,
15
+ "eos_token": "<|endoftext|>",
16
+ "extra_special_tokens": {},
17
+ "max_length": 180,
18
+ "model_max_length": 1024,
19
+ "pad_to_multiple_of": null,
20
+ "pad_token": "<|endoftext|>",
21
+ "pad_token_type_id": 0,
22
+ "padding_side": "right",
23
+ "stride": 0,
24
+ "tokenizer_class": "GPT2Tokenizer",
25
+ "truncation_side": "right",
26
+ "truncation_strategy": "longest_first",
27
+ "unk_token": "<|endoftext|>"
28
+ }
checkpoint-11000/trainer_state.json ADDED
@@ -0,0 +1,595 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 11000,
3
+ "best_metric": 0.4608609676361084,
4
+ "best_model_checkpoint": "./banking-sms-parser-v10/checkpoint-11000",
5
+ "epoch": 3.259259259259259,
6
+ "eval_steps": 500,
7
+ "global_step": 11000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.05925925925925926,
14
+ "grad_norm": 0.47084560990333557,
15
+ "learning_rate": 0.0002955777777777778,
16
+ "loss": 2.0436,
17
+ "step": 200
18
+ },
19
+ {
20
+ "epoch": 0.11851851851851852,
21
+ "grad_norm": 0.40365636348724365,
22
+ "learning_rate": 0.0002911333333333333,
23
+ "loss": 1.3594,
24
+ "step": 400
25
+ },
26
+ {
27
+ "epoch": 0.14814814814814814,
28
+ "eval_loss": 1.1683876514434814,
29
+ "eval_runtime": 29.5095,
30
+ "eval_samples_per_second": 101.662,
31
+ "eval_steps_per_second": 12.708,
32
+ "step": 500
33
+ },
34
+ {
35
+ "epoch": 0.17777777777777778,
36
+ "grad_norm": 0.5327613353729248,
37
+ "learning_rate": 0.0002866888888888889,
38
+ "loss": 1.2262,
39
+ "step": 600
40
+ },
41
+ {
42
+ "epoch": 0.23703703703703705,
43
+ "grad_norm": 0.5959504246711731,
44
+ "learning_rate": 0.0002822444444444444,
45
+ "loss": 1.1381,
46
+ "step": 800
47
+ },
48
+ {
49
+ "epoch": 0.2962962962962963,
50
+ "grad_norm": 0.4996171295642853,
51
+ "learning_rate": 0.0002778,
52
+ "loss": 1.066,
53
+ "step": 1000
54
+ },
55
+ {
56
+ "epoch": 0.2962962962962963,
57
+ "eval_loss": 0.9456237554550171,
58
+ "eval_runtime": 29.4299,
59
+ "eval_samples_per_second": 101.937,
60
+ "eval_steps_per_second": 12.742,
61
+ "step": 1000
62
+ },
63
+ {
64
+ "epoch": 0.35555555555555557,
65
+ "grad_norm": 0.47130197286605835,
66
+ "learning_rate": 0.0002733555555555555,
67
+ "loss": 1.0067,
68
+ "step": 1200
69
+ },
70
+ {
71
+ "epoch": 0.4148148148148148,
72
+ "grad_norm": 0.693696916103363,
73
+ "learning_rate": 0.0002689111111111111,
74
+ "loss": 0.969,
75
+ "step": 1400
76
+ },
77
+ {
78
+ "epoch": 0.4444444444444444,
79
+ "eval_loss": 0.8003639578819275,
80
+ "eval_runtime": 29.3553,
81
+ "eval_samples_per_second": 102.196,
82
+ "eval_steps_per_second": 12.775,
83
+ "step": 1500
84
+ },
85
+ {
86
+ "epoch": 0.4740740740740741,
87
+ "grad_norm": 0.5786986351013184,
88
+ "learning_rate": 0.0002644666666666666,
89
+ "loss": 0.9081,
90
+ "step": 1600
91
+ },
92
+ {
93
+ "epoch": 0.5333333333333333,
94
+ "grad_norm": 0.5209688544273376,
95
+ "learning_rate": 0.0002600222222222222,
96
+ "loss": 0.8801,
97
+ "step": 1800
98
+ },
99
+ {
100
+ "epoch": 0.5925925925925926,
101
+ "grad_norm": 0.7497162818908691,
102
+ "learning_rate": 0.0002555777777777778,
103
+ "loss": 0.8183,
104
+ "step": 2000
105
+ },
106
+ {
107
+ "epoch": 0.5925925925925926,
108
+ "eval_loss": 0.7105592489242554,
109
+ "eval_runtime": 29.7273,
110
+ "eval_samples_per_second": 100.917,
111
+ "eval_steps_per_second": 12.615,
112
+ "step": 2000
113
+ },
114
+ {
115
+ "epoch": 0.6518518518518519,
116
+ "grad_norm": 0.7358675599098206,
117
+ "learning_rate": 0.00025113333333333333,
118
+ "loss": 0.809,
119
+ "step": 2200
120
+ },
121
+ {
122
+ "epoch": 0.7111111111111111,
123
+ "grad_norm": 0.5220760107040405,
124
+ "learning_rate": 0.0002466888888888889,
125
+ "loss": 0.7799,
126
+ "step": 2400
127
+ },
128
+ {
129
+ "epoch": 0.7407407407407407,
130
+ "eval_loss": 0.6572133898735046,
131
+ "eval_runtime": 29.3661,
132
+ "eval_samples_per_second": 102.158,
133
+ "eval_steps_per_second": 12.77,
134
+ "step": 2500
135
+ },
136
+ {
137
+ "epoch": 0.7703703703703704,
138
+ "grad_norm": 0.4607163965702057,
139
+ "learning_rate": 0.00024224444444444443,
140
+ "loss": 0.7557,
141
+ "step": 2600
142
+ },
143
+ {
144
+ "epoch": 0.8296296296296296,
145
+ "grad_norm": 0.5450932383537292,
146
+ "learning_rate": 0.00023779999999999998,
147
+ "loss": 0.7264,
148
+ "step": 2800
149
+ },
150
+ {
151
+ "epoch": 0.8888888888888888,
152
+ "grad_norm": 0.5149521231651306,
153
+ "learning_rate": 0.00023335555555555556,
154
+ "loss": 0.7054,
155
+ "step": 3000
156
+ },
157
+ {
158
+ "epoch": 0.8888888888888888,
159
+ "eval_loss": 0.6189373731613159,
160
+ "eval_runtime": 29.6913,
161
+ "eval_samples_per_second": 101.04,
162
+ "eval_steps_per_second": 12.63,
163
+ "step": 3000
164
+ },
165
+ {
166
+ "epoch": 0.9481481481481482,
167
+ "grad_norm": 0.5345056056976318,
168
+ "learning_rate": 0.0002289111111111111,
169
+ "loss": 0.7026,
170
+ "step": 3200
171
+ },
172
+ {
173
+ "epoch": 1.0074074074074073,
174
+ "grad_norm": 0.5385560989379883,
175
+ "learning_rate": 0.00022446666666666666,
176
+ "loss": 0.6888,
177
+ "step": 3400
178
+ },
179
+ {
180
+ "epoch": 1.037037037037037,
181
+ "eval_loss": 0.5899662971496582,
182
+ "eval_runtime": 29.4218,
183
+ "eval_samples_per_second": 101.965,
184
+ "eval_steps_per_second": 12.746,
185
+ "step": 3500
186
+ },
187
+ {
188
+ "epoch": 1.0666666666666667,
189
+ "grad_norm": 0.5609071850776672,
190
+ "learning_rate": 0.0002200222222222222,
191
+ "loss": 0.6671,
192
+ "step": 3600
193
+ },
194
+ {
195
+ "epoch": 1.125925925925926,
196
+ "grad_norm": 0.5616484880447388,
197
+ "learning_rate": 0.00021557777777777776,
198
+ "loss": 0.6636,
199
+ "step": 3800
200
+ },
201
+ {
202
+ "epoch": 1.1851851851851851,
203
+ "grad_norm": 0.5596534609794617,
204
+ "learning_rate": 0.0002111333333333333,
205
+ "loss": 0.6515,
206
+ "step": 4000
207
+ },
208
+ {
209
+ "epoch": 1.1851851851851851,
210
+ "eval_loss": 0.5668333172798157,
211
+ "eval_runtime": 29.3385,
212
+ "eval_samples_per_second": 102.255,
213
+ "eval_steps_per_second": 12.782,
214
+ "step": 4000
215
+ },
216
+ {
217
+ "epoch": 1.2444444444444445,
218
+ "grad_norm": 0.5900132656097412,
219
+ "learning_rate": 0.00020668888888888888,
220
+ "loss": 0.6409,
221
+ "step": 4200
222
+ },
223
+ {
224
+ "epoch": 1.3037037037037038,
225
+ "grad_norm": 0.6175922751426697,
226
+ "learning_rate": 0.00020224444444444443,
227
+ "loss": 0.641,
228
+ "step": 4400
229
+ },
230
+ {
231
+ "epoch": 1.3333333333333333,
232
+ "eval_loss": 0.5480825304985046,
233
+ "eval_runtime": 29.457,
234
+ "eval_samples_per_second": 101.843,
235
+ "eval_steps_per_second": 12.73,
236
+ "step": 4500
237
+ },
238
+ {
239
+ "epoch": 1.362962962962963,
240
+ "grad_norm": 0.42820265889167786,
241
+ "learning_rate": 0.00019779999999999998,
242
+ "loss": 0.6212,
243
+ "step": 4600
244
+ },
245
+ {
246
+ "epoch": 1.4222222222222223,
247
+ "grad_norm": 0.7179502248764038,
248
+ "learning_rate": 0.00019335555555555553,
249
+ "loss": 0.6268,
250
+ "step": 4800
251
+ },
252
+ {
253
+ "epoch": 1.4814814814814814,
254
+ "grad_norm": 0.4201723635196686,
255
+ "learning_rate": 0.00018891111111111108,
256
+ "loss": 0.6083,
257
+ "step": 5000
258
+ },
259
+ {
260
+ "epoch": 1.4814814814814814,
261
+ "eval_loss": 0.538067638874054,
262
+ "eval_runtime": 29.3858,
263
+ "eval_samples_per_second": 102.09,
264
+ "eval_steps_per_second": 12.761,
265
+ "step": 5000
266
+ },
267
+ {
268
+ "epoch": 1.5407407407407407,
269
+ "grad_norm": 0.6730241775512695,
270
+ "learning_rate": 0.00018446666666666666,
271
+ "loss": 0.6148,
272
+ "step": 5200
273
+ },
274
+ {
275
+ "epoch": 1.6,
276
+ "grad_norm": 0.5300161242485046,
277
+ "learning_rate": 0.0001800222222222222,
278
+ "loss": 0.6001,
279
+ "step": 5400
280
+ },
281
+ {
282
+ "epoch": 1.6296296296296298,
283
+ "eval_loss": 0.5220749378204346,
284
+ "eval_runtime": 29.4236,
285
+ "eval_samples_per_second": 101.959,
286
+ "eval_steps_per_second": 12.745,
287
+ "step": 5500
288
+ },
289
+ {
290
+ "epoch": 1.6592592592592592,
291
+ "grad_norm": 0.45725536346435547,
292
+ "learning_rate": 0.00017557777777777776,
293
+ "loss": 0.605,
294
+ "step": 5600
295
+ },
296
+ {
297
+ "epoch": 1.7185185185185186,
298
+ "grad_norm": 0.49325183033943176,
299
+ "learning_rate": 0.00017115555555555554,
300
+ "loss": 0.6024,
301
+ "step": 5800
302
+ },
303
+ {
304
+ "epoch": 1.7777777777777777,
305
+ "grad_norm": 0.5796788930892944,
306
+ "learning_rate": 0.00016671111111111109,
307
+ "loss": 0.5941,
308
+ "step": 6000
309
+ },
310
+ {
311
+ "epoch": 1.7777777777777777,
312
+ "eval_loss": 0.5122686624526978,
313
+ "eval_runtime": 29.5133,
314
+ "eval_samples_per_second": 101.649,
315
+ "eval_steps_per_second": 12.706,
316
+ "step": 6000
317
+ },
318
+ {
319
+ "epoch": 1.837037037037037,
320
+ "grad_norm": 0.5793591737747192,
321
+ "learning_rate": 0.00016226666666666664,
322
+ "loss": 0.5809,
323
+ "step": 6200
324
+ },
325
+ {
326
+ "epoch": 1.8962962962962964,
327
+ "grad_norm": 0.4882641136646271,
328
+ "learning_rate": 0.00015782222222222219,
329
+ "loss": 0.5869,
330
+ "step": 6400
331
+ },
332
+ {
333
+ "epoch": 1.925925925925926,
334
+ "eval_loss": 0.5018052458763123,
335
+ "eval_runtime": 29.2669,
336
+ "eval_samples_per_second": 102.505,
337
+ "eval_steps_per_second": 12.813,
338
+ "step": 6500
339
+ },
340
+ {
341
+ "epoch": 1.9555555555555557,
342
+ "grad_norm": 0.5069060325622559,
343
+ "learning_rate": 0.0001533777777777778,
344
+ "loss": 0.5812,
345
+ "step": 6600
346
+ },
347
+ {
348
+ "epoch": 2.0148148148148146,
349
+ "grad_norm": 0.44775187969207764,
350
+ "learning_rate": 0.00014893333333333334,
351
+ "loss": 0.5397,
352
+ "step": 6800
353
+ },
354
+ {
355
+ "epoch": 2.074074074074074,
356
+ "grad_norm": 0.5677600502967834,
357
+ "learning_rate": 0.0001444888888888889,
358
+ "loss": 0.5567,
359
+ "step": 7000
360
+ },
361
+ {
362
+ "epoch": 2.074074074074074,
363
+ "eval_loss": 0.49609723687171936,
364
+ "eval_runtime": 29.4825,
365
+ "eval_samples_per_second": 101.755,
366
+ "eval_steps_per_second": 12.719,
367
+ "step": 7000
368
+ },
369
+ {
370
+ "epoch": 2.1333333333333333,
371
+ "grad_norm": 0.5369960069656372,
372
+ "learning_rate": 0.00014004444444444444,
373
+ "loss": 0.5591,
374
+ "step": 7200
375
+ },
376
+ {
377
+ "epoch": 2.1925925925925926,
378
+ "grad_norm": 0.353333055973053,
379
+ "learning_rate": 0.0001356,
380
+ "loss": 0.5717,
381
+ "step": 7400
382
+ },
383
+ {
384
+ "epoch": 2.2222222222222223,
385
+ "eval_loss": 0.49044421315193176,
386
+ "eval_runtime": 29.573,
387
+ "eval_samples_per_second": 101.444,
388
+ "eval_steps_per_second": 12.68,
389
+ "step": 7500
390
+ },
391
+ {
392
+ "epoch": 2.251851851851852,
393
+ "grad_norm": 0.45505258440971375,
394
+ "learning_rate": 0.00013115555555555554,
395
+ "loss": 0.5576,
396
+ "step": 7600
397
+ },
398
+ {
399
+ "epoch": 2.311111111111111,
400
+ "grad_norm": 0.4602493643760681,
401
+ "learning_rate": 0.00012671111111111112,
402
+ "loss": 0.5407,
403
+ "step": 7800
404
+ },
405
+ {
406
+ "epoch": 2.3703703703703702,
407
+ "grad_norm": 0.4852631688117981,
408
+ "learning_rate": 0.00012226666666666667,
409
+ "loss": 0.5523,
410
+ "step": 8000
411
+ },
412
+ {
413
+ "epoch": 2.3703703703703702,
414
+ "eval_loss": 0.48393580317497253,
415
+ "eval_runtime": 29.3448,
416
+ "eval_samples_per_second": 102.233,
417
+ "eval_steps_per_second": 12.779,
418
+ "step": 8000
419
+ },
420
+ {
421
+ "epoch": 2.4296296296296296,
422
+ "grad_norm": 0.4038563668727875,
423
+ "learning_rate": 0.0001178222222222222,
424
+ "loss": 0.5313,
425
+ "step": 8200
426
+ },
427
+ {
428
+ "epoch": 2.488888888888889,
429
+ "grad_norm": 0.5276423692703247,
430
+ "learning_rate": 0.00011339999999999999,
431
+ "loss": 0.5489,
432
+ "step": 8400
433
+ },
434
+ {
435
+ "epoch": 2.5185185185185186,
436
+ "eval_loss": 0.47875121235847473,
437
+ "eval_runtime": 29.4918,
438
+ "eval_samples_per_second": 101.723,
439
+ "eval_steps_per_second": 12.715,
440
+ "step": 8500
441
+ },
442
+ {
443
+ "epoch": 2.5481481481481483,
444
+ "grad_norm": 0.539804995059967,
445
+ "learning_rate": 0.00010895555555555554,
446
+ "loss": 0.5456,
447
+ "step": 8600
448
+ },
449
+ {
450
+ "epoch": 2.6074074074074076,
451
+ "grad_norm": 0.629550039768219,
452
+ "learning_rate": 0.0001045111111111111,
453
+ "loss": 0.5523,
454
+ "step": 8800
455
+ },
456
+ {
457
+ "epoch": 2.6666666666666665,
458
+ "grad_norm": 0.6091557741165161,
459
+ "learning_rate": 0.00010006666666666666,
460
+ "loss": 0.5358,
461
+ "step": 9000
462
+ },
463
+ {
464
+ "epoch": 2.6666666666666665,
465
+ "eval_loss": 0.4742780327796936,
466
+ "eval_runtime": 29.4044,
467
+ "eval_samples_per_second": 102.026,
468
+ "eval_steps_per_second": 12.753,
469
+ "step": 9000
470
+ },
471
+ {
472
+ "epoch": 2.725925925925926,
473
+ "grad_norm": 0.5220003724098206,
474
+ "learning_rate": 9.56222222222222e-05,
475
+ "loss": 0.5398,
476
+ "step": 9200
477
+ },
478
+ {
479
+ "epoch": 2.785185185185185,
480
+ "grad_norm": 0.4543071389198303,
481
+ "learning_rate": 9.117777777777777e-05,
482
+ "loss": 0.53,
483
+ "step": 9400
484
+ },
485
+ {
486
+ "epoch": 2.814814814814815,
487
+ "eval_loss": 0.4698003828525543,
488
+ "eval_runtime": 29.5022,
489
+ "eval_samples_per_second": 101.687,
490
+ "eval_steps_per_second": 12.711,
491
+ "step": 9500
492
+ },
493
+ {
494
+ "epoch": 2.8444444444444446,
495
+ "grad_norm": 0.4849318265914917,
496
+ "learning_rate": 8.673333333333332e-05,
497
+ "loss": 0.5278,
498
+ "step": 9600
499
+ },
500
+ {
501
+ "epoch": 2.9037037037037035,
502
+ "grad_norm": 0.40453964471817017,
503
+ "learning_rate": 8.228888888888888e-05,
504
+ "loss": 0.5448,
505
+ "step": 9800
506
+ },
507
+ {
508
+ "epoch": 2.962962962962963,
509
+ "grad_norm": 0.5250543355941772,
510
+ "learning_rate": 7.784444444444443e-05,
511
+ "loss": 0.538,
512
+ "step": 10000
513
+ },
514
+ {
515
+ "epoch": 2.962962962962963,
516
+ "eval_loss": 0.46695125102996826,
517
+ "eval_runtime": 29.5114,
518
+ "eval_samples_per_second": 101.656,
519
+ "eval_steps_per_second": 12.707,
520
+ "step": 10000
521
+ },
522
+ {
523
+ "epoch": 3.022222222222222,
524
+ "grad_norm": 0.46158769726753235,
525
+ "learning_rate": 7.34e-05,
526
+ "loss": 0.5379,
527
+ "step": 10200
528
+ },
529
+ {
530
+ "epoch": 3.0814814814814815,
531
+ "grad_norm": 0.5191985368728638,
532
+ "learning_rate": 6.895555555555556e-05,
533
+ "loss": 0.522,
534
+ "step": 10400
535
+ },
536
+ {
537
+ "epoch": 3.111111111111111,
538
+ "eval_loss": 0.46375858783721924,
539
+ "eval_runtime": 29.4058,
540
+ "eval_samples_per_second": 102.021,
541
+ "eval_steps_per_second": 12.753,
542
+ "step": 10500
543
+ },
544
+ {
545
+ "epoch": 3.140740740740741,
546
+ "grad_norm": 0.5155138969421387,
547
+ "learning_rate": 6.451111111111111e-05,
548
+ "loss": 0.5137,
549
+ "step": 10600
550
+ },
551
+ {
552
+ "epoch": 3.2,
553
+ "grad_norm": 0.6184846758842468,
554
+ "learning_rate": 6.006666666666666e-05,
555
+ "loss": 0.5398,
556
+ "step": 10800
557
+ },
558
+ {
559
+ "epoch": 3.259259259259259,
560
+ "grad_norm": 0.48932573199272156,
561
+ "learning_rate": 5.5622222222222215e-05,
562
+ "loss": 0.5293,
563
+ "step": 11000
564
+ },
565
+ {
566
+ "epoch": 3.259259259259259,
567
+ "eval_loss": 0.4608609676361084,
568
+ "eval_runtime": 29.8123,
569
+ "eval_samples_per_second": 100.63,
570
+ "eval_steps_per_second": 12.579,
571
+ "step": 11000
572
+ }
573
+ ],
574
+ "logging_steps": 200,
575
+ "max_steps": 13500,
576
+ "num_input_tokens_seen": 0,
577
+ "num_train_epochs": 4,
578
+ "save_steps": 1000,
579
+ "stateful_callbacks": {
580
+ "TrainerControl": {
581
+ "args": {
582
+ "should_epoch_stop": false,
583
+ "should_evaluate": false,
584
+ "should_log": false,
585
+ "should_save": true,
586
+ "should_training_stop": false
587
+ },
588
+ "attributes": {}
589
+ }
590
+ },
591
+ "total_flos": 7322688552960000.0,
592
+ "train_batch_size": 8,
593
+ "trial_name": null,
594
+ "trial_params": null
595
+ }
checkpoint-11000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5b2e2a527ba622332dbb2398486338b6fc135483c9c1c937ef5cefadf332325
3
+ size 5304
checkpoint-11000/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-12000/README.md ADDED
@@ -0,0 +1,207 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: distilgpt2
3
+ library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - base_model:adapter:distilgpt2
7
+ - lora
8
+ - transformers
9
+ ---
10
+
11
+ # Model Card for Model ID
12
+
13
+ <!-- Provide a quick summary of what the model is/does. -->
14
+
15
+
16
+
17
+ ## Model Details
18
+
19
+ ### Model Description
20
+
21
+ <!-- Provide a longer summary of what this model is. -->
22
+
23
+
24
+
25
+ - **Developed by:** [More Information Needed]
26
+ - **Funded by [optional]:** [More Information Needed]
27
+ - **Shared by [optional]:** [More Information Needed]
28
+ - **Model type:** [More Information Needed]
29
+ - **Language(s) (NLP):** [More Information Needed]
30
+ - **License:** [More Information Needed]
31
+ - **Finetuned from model [optional]:** [More Information Needed]
32
+
33
+ ### Model Sources [optional]
34
+
35
+ <!-- Provide the basic links for the model. -->
36
+
37
+ - **Repository:** [More Information Needed]
38
+ - **Paper [optional]:** [More Information Needed]
39
+ - **Demo [optional]:** [More Information Needed]
40
+
41
+ ## Uses
42
+
43
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
44
+
45
+ ### Direct Use
46
+
47
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
48
+
49
+ [More Information Needed]
50
+
51
+ ### Downstream Use [optional]
52
+
53
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
54
+
55
+ [More Information Needed]
56
+
57
+ ### Out-of-Scope Use
58
+
59
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
60
+
61
+ [More Information Needed]
62
+
63
+ ## Bias, Risks, and Limitations
64
+
65
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
66
+
67
+ [More Information Needed]
68
+
69
+ ### Recommendations
70
+
71
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
72
+
73
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
74
+
75
+ ## How to Get Started with the Model
76
+
77
+ Use the code below to get started with the model.
78
+
79
+ [More Information Needed]
80
+
81
+ ## Training Details
82
+
83
+ ### Training Data
84
+
85
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
86
+
87
+ [More Information Needed]
88
+
89
+ ### Training Procedure
90
+
91
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
92
+
93
+ #### Preprocessing [optional]
94
+
95
+ [More Information Needed]
96
+
97
+
98
+ #### Training Hyperparameters
99
+
100
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
101
+
102
+ #### Speeds, Sizes, Times [optional]
103
+
104
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
105
+
106
+ [More Information Needed]
107
+
108
+ ## Evaluation
109
+
110
+ <!-- This section describes the evaluation protocols and provides the results. -->
111
+
112
+ ### Testing Data, Factors & Metrics
113
+
114
+ #### Testing Data
115
+
116
+ <!-- This should link to a Dataset Card if possible. -->
117
+
118
+ [More Information Needed]
119
+
120
+ #### Factors
121
+
122
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
123
+
124
+ [More Information Needed]
125
+
126
+ #### Metrics
127
+
128
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
129
+
130
+ [More Information Needed]
131
+
132
+ ### Results
133
+
134
+ [More Information Needed]
135
+
136
+ #### Summary
137
+
138
+
139
+
140
+ ## Model Examination [optional]
141
+
142
+ <!-- Relevant interpretability work for the model goes here -->
143
+
144
+ [More Information Needed]
145
+
146
+ ## Environmental Impact
147
+
148
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
149
+
150
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
151
+
152
+ - **Hardware Type:** [More Information Needed]
153
+ - **Hours used:** [More Information Needed]
154
+ - **Cloud Provider:** [More Information Needed]
155
+ - **Compute Region:** [More Information Needed]
156
+ - **Carbon Emitted:** [More Information Needed]
157
+
158
+ ## Technical Specifications [optional]
159
+
160
+ ### Model Architecture and Objective
161
+
162
+ [More Information Needed]
163
+
164
+ ### Compute Infrastructure
165
+
166
+ [More Information Needed]
167
+
168
+ #### Hardware
169
+
170
+ [More Information Needed]
171
+
172
+ #### Software
173
+
174
+ [More Information Needed]
175
+
176
+ ## Citation [optional]
177
+
178
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
179
+
180
+ **BibTeX:**
181
+
182
+ [More Information Needed]
183
+
184
+ **APA:**
185
+
186
+ [More Information Needed]
187
+
188
+ ## Glossary [optional]
189
+
190
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
191
+
192
+ [More Information Needed]
193
+
194
+ ## More Information [optional]
195
+
196
+ [More Information Needed]
197
+
198
+ ## Model Card Authors [optional]
199
+
200
+ [More Information Needed]
201
+
202
+ ## Model Card Contact
203
+
204
+ [More Information Needed]
205
+ ### Framework versions
206
+
207
+ - PEFT 0.17.0
checkpoint-12000/adapter_config.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "distilgpt2",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 32,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.05,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "qalora_group_size": 16,
24
+ "r": 16,
25
+ "rank_pattern": {},
26
+ "revision": null,
27
+ "target_modules": [
28
+ "c_proj",
29
+ "c_attn"
30
+ ],
31
+ "target_parameters": null,
32
+ "task_type": "CAUSAL_LM",
33
+ "trainable_token_indices": null,
34
+ "use_dora": false,
35
+ "use_qalora": false,
36
+ "use_rslora": false
37
+ }
checkpoint-12000/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52b377b2b2769700bb41ee946eb12f0becc601b221b3a839cccaa392495830f6
3
+ size 3248616
checkpoint-12000/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-12000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf85d0480c0f2c80ca06340328baf17ed7f743a49a085abb27f750563eb4c1d0
3
+ size 6518714