Update README.md
Browse files
README.md
CHANGED
|
@@ -8,11 +8,17 @@ tags:
|
|
| 8 |
|
| 9 |
This is an on-going project. it is a modified version of Higgs-Boson audio tokenizer, you can fully train it. all scripts have been tested.
|
| 10 |
a Few notes however:
|
| 11 |
-
|
| 12 |
-
|
| 13 |
-
|
| 14 |
-
3- for the semantic teacher, I am using ```utter-project/mHuBERT-147``` which has a good multilingual support. if you want the original setup you can change it in the config.
|
| 15 |
|
| 16 |
I will train a checkpoint on a larger enough dataset one of these days after figuring out a few things first. but the setup is solid.
|
| 17 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 18 |
Happy training (~~inshallah~~).
|
|
|
|
| 8 |
|
| 9 |
This is an on-going project. it is a modified version of Higgs-Boson audio tokenizer, you can fully train it. all scripts have been tested.
|
| 10 |
a Few notes however:
|
| 11 |
+
- this is not backward compatible with the original checkpoint (I think you can tweak it to be, but you have to adhere to Boson community license if you do.)
|
| 12 |
+
- I highly recommend you to pretrain the model without the mel and adversarial setup first. it saves you a significant amount of compute, time and speed-up your convergence. raise the batch size as much as you can before the adversarial phase.
|
| 13 |
+
- for the semantic teacher, I am using ```utter-project/mHuBERT-147``` which has a good multilingual support. if you want the original setup you can change it in the config.
|
|
|
|
| 14 |
|
| 15 |
I will train a checkpoint on a larger enough dataset one of these days after figuring out a few things first. but the setup is solid.
|
| 16 |
|
| 17 |
+
```bash
|
| 18 |
+
python train_boson_mixed_precision.py --data_csv "yourdata.csv" \
|
| 19 |
+
--config config.json --batch_size 42 \
|
| 20 |
+
--use_mixed_precision \
|
| 21 |
+
--use_discriminator
|
| 22 |
+
```
|
| 23 |
+
|
| 24 |
Happy training (~~inshallah~~).
|