mhenrichsen commited on
Commit
a3f20a8
·
verified ·
1 Parent(s): 60d6f7c

End of training

Browse files
Files changed (2) hide show
  1. README.md +146 -0
  2. generation_config.json +9 -0
README.md ADDED
@@ -0,0 +1,146 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ base_model: syvai/tts-v1-pretrained
4
+ tags:
5
+ - axolotl
6
+ - generated_from_trainer
7
+ datasets:
8
+ - syvai/zac-coral-tts
9
+ - syvai/zac-dk-voice-pro
10
+ - syvai/zac-dk-voice-single-speaker
11
+ model-index:
12
+ - name: tts-v0.3-finetuned
13
+ results: []
14
+ ---
15
+
16
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
+ should probably proofread and complete it, then remove this comment. -->
18
+
19
+ [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
20
+ <details><summary>See axolotl config</summary>
21
+
22
+ axolotl version: `0.10.0`
23
+ ```yaml
24
+ base_model: syvai/tts-v1-pretrained
25
+ # Automatically upload checkpoint and final model to HF
26
+ hub_model_id: syvai/tts-v0.3-finetuned
27
+
28
+ plugins:
29
+ - axolotl.integrations.liger.LigerPlugin
30
+ liger_rope: true
31
+ liger_rms_norm: true
32
+ liger_glu_activation: true
33
+ liger_fused_linear_cross_entropy: true
34
+
35
+ datasets:
36
+ - path: syvai/zac-coral-tts
37
+ type:
38
+ - path: syvai/zac-dk-voice-pro
39
+ type:
40
+ - path: syvai/zac-dk-voice-single-speaker
41
+ type:
42
+ dataset_prepared_path: last_run_prepared
43
+ val_set_size: 0.01
44
+ eval_sample_packing: False
45
+ output_dir: ./outputs/finetuned
46
+
47
+ sequence_len: 8196
48
+ sample_packing: true
49
+ pad_to_sequence_len: true
50
+
51
+ wandb_project: orph
52
+ wandb_entity:
53
+ wandb_watch:
54
+ wandb_name:
55
+ wandb_log_model:
56
+
57
+ gradient_accumulation_steps: 16
58
+ micro_batch_size: 2
59
+ num_epochs: 3
60
+ optimizer: adamw_torch_fused
61
+ lr_scheduler: cosine
62
+ learning_rate: 2e-5
63
+
64
+ bf16: auto
65
+ tf32: false
66
+
67
+ gradient_checkpointing: true
68
+ gradient_checkpointing_kwargs:
69
+ use_reentrant: false
70
+ resume_from_checkpoint:
71
+ logging_steps: 1
72
+ flash_attention: true
73
+
74
+ warmup_steps: 3
75
+ evals_per_epoch: 5
76
+ saves_per_epoch: 5
77
+ weight_decay: 0.05
78
+
79
+ special_tokens:
80
+ pad_token: <custom_token_7>
81
+
82
+ ```
83
+
84
+ </details><br>
85
+
86
+ # tts-v0.3-finetuned
87
+
88
+ This model is a fine-tuned version of [syvai/tts-v1-pretrained](https://huggingface.co/syvai/tts-v1-pretrained) on the syvai/zac-coral-tts, the syvai/zac-dk-voice-pro and the syvai/zac-dk-voice-single-speaker datasets.
89
+ It achieves the following results on the evaluation set:
90
+ - Loss: 4.3861
91
+
92
+ ## Model description
93
+
94
+ More information needed
95
+
96
+ ## Intended uses & limitations
97
+
98
+ More information needed
99
+
100
+ ## Training and evaluation data
101
+
102
+ More information needed
103
+
104
+ ## Training procedure
105
+
106
+ ### Training hyperparameters
107
+
108
+ The following hyperparameters were used during training:
109
+ - learning_rate: 2e-05
110
+ - train_batch_size: 2
111
+ - eval_batch_size: 2
112
+ - seed: 42
113
+ - gradient_accumulation_steps: 16
114
+ - total_train_batch_size: 32
115
+ - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
116
+ - lr_scheduler_type: cosine
117
+ - lr_scheduler_warmup_steps: 3
118
+ - training_steps: 159
119
+
120
+ ### Training results
121
+
122
+ | Training Loss | Epoch | Step | Validation Loss |
123
+ |:-------------:|:------:|:----:|:---------------:|
124
+ | No log | 0 | 0 | 4.9462 |
125
+ | 4.6427 | 0.2042 | 11 | 4.6525 |
126
+ | 4.5404 | 0.4084 | 22 | 4.5458 |
127
+ | 4.4701 | 0.6125 | 33 | 4.4809 |
128
+ | 4.4139 | 0.8167 | 44 | 4.4405 |
129
+ | 4.4146 | 1.0186 | 55 | 4.4182 |
130
+ | 4.4021 | 1.2227 | 66 | 4.4045 |
131
+ | 4.3951 | 1.4269 | 77 | 4.3957 |
132
+ | 4.3845 | 1.6311 | 88 | 4.3906 |
133
+ | 4.3651 | 1.8353 | 99 | 4.3880 |
134
+ | 4.3924 | 2.0371 | 110 | 4.3867 |
135
+ | 4.3551 | 2.2413 | 121 | 4.3865 |
136
+ | 4.3478 | 2.4455 | 132 | 4.3861 |
137
+ | 4.3894 | 2.6497 | 143 | 4.3859 |
138
+ | 4.3801 | 2.8538 | 154 | 4.3861 |
139
+
140
+
141
+ ### Framework versions
142
+
143
+ - Transformers 4.52.3
144
+ - Pytorch 2.6.0+cu124
145
+ - Datasets 3.6.0
146
+ - Tokenizers 0.21.1
generation_config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 128000,
4
+ "do_sample": true,
5
+ "eos_token_id": 128001,
6
+ "temperature": 0.6,
7
+ "top_p": 0.9,
8
+ "transformers_version": "4.52.3"
9
+ }