| import os | |
| from run_model_downloader import download_models | |
| if not os.path.exists("Models/ToucanTTS_Meta/best.pt"): | |
| download_models() | |
| import gradio as gr | |
| from Preprocessing.multilinguality.SimilaritySolver import load_json_from_path | |
| from Utility.utils import float2pcm | |
| import os | |
| import torch | |
| from Architectures.ControllabilityGAN.GAN import GanWrapper | |
| from InferenceInterfaces.ToucanTTSInterface import ToucanTTSInterface | |
| from Utility.storage_config import MODELS_DIR | |
| class ControllableInterface(torch.nn.Module): | |
| def __init__(self, available_artificial_voices=1000): | |
| super().__init__() | |
| self.model = ToucanTTSInterface(device="cuda", tts_model_path="Meta", language="eng") | |
| self.wgan = GanWrapper(os.path.join(MODELS_DIR, "Embedding", "embedding_gan.pt"), device="cuda") | |
| self.generated_speaker_embeds = list() | |
| self.available_artificial_voices = available_artificial_voices | |
| self.current_language = "" | |
| self.current_accent = "" | |
| def read(self, | |
| prompt, | |
| language, | |
| accent, | |
| voice_seed, | |
| duration_scaling_factor, | |
| pause_duration_scaling_factor, | |
| pitch_variance_scale, | |
| energy_variance_scale, | |
| emb_slider_1, | |
| emb_slider_2, | |
| emb_slider_3, | |
| emb_slider_4, | |
| emb_slider_5, | |
| emb_slider_6, | |
| loudness_in_db | |
| ): | |
| if self.current_language != language: | |
| self.model = ToucanTTSInterface(device="cpu", tts_model_path="Meta", language=language) | |
| self.current_language = language | |
| self.wgan.set_latent(voice_seed) | |
| controllability_vector = torch.tensor([emb_slider_1, | |
| emb_slider_2, | |
| emb_slider_3, | |
| emb_slider_4, | |
| emb_slider_5, | |
| emb_slider_6], dtype=torch.float32) | |
| embedding = self.wgan.modify_embed(controllability_vector) | |
| self.model.set_utterance_embedding(embedding=embedding) | |
| phones = self.model.text2phone.get_phone_string(prompt) | |
| if len(phones) > 1800: | |
| return | |
| print(prompt) | |
| wav, sr, fig = self.model(prompt, | |
| input_is_phones=False, | |
| duration_scaling_factor=duration_scaling_factor, | |
| pitch_variance_scale=pitch_variance_scale, | |
| energy_variance_scale=energy_variance_scale, | |
| pause_duration_scaling_factor=pause_duration_scaling_factor, | |
| return_plot_as_filepath=True, | |
| loudness_in_db=loudness_in_db) | |
| return sr, wav, fig | |
| title = "Controllable Text-to-Speech for over 7000 Languages" | |
| article = "Check out the IMS Toucan TTS Toolkit at https://github.com/DigitalPhonetics/IMS-Toucan" | |
| available_artificial_voices = 1000 | |
| path_to_iso_list = "Preprocessing/multilinguality/iso_to_fullname.json" | |
| iso_to_name = load_json_from_path(path_to_iso_list) | |
| text_selection = [f"{iso_to_name[iso_code]} Text ({iso_code})" for iso_code in iso_to_name] | |
| controllable_ui = ControllableInterface(available_artificial_voices=available_artificial_voices) | |
| def read(prompt, | |
| language, | |
| voice_seed, | |
| duration_scaling_factor, | |
| pitch_variance_scale, | |
| energy_variance_scale, | |
| emb1, | |
| emb2 | |
| ): | |
| with torch.no_grad(): | |
| sr, wav, fig = controllable_ui.read(prompt, | |
| language.split(" ")[-1].split("(")[1].split(")")[0], | |
| language.split(" ")[-1].split("(")[1].split(")")[0], | |
| voice_seed, | |
| duration_scaling_factor, | |
| 1., | |
| pitch_variance_scale, | |
| energy_variance_scale, | |
| emb1, | |
| emb2, | |
| 0., | |
| 0., | |
| 0., | |
| 0., | |
| -24.) | |
| return (sr, float2pcm(wav)), fig | |
| iface = gr.Interface(fn=read, | |
| inputs=[gr.Textbox(lines=2, | |
| placeholder="write what you want the synthesis to read here...", | |
| value="The woods are lovely, dark and deep, but I have promises to keep, and miles to go, before I sleep.", | |
| label="Text input"), | |
| gr.Dropdown(text_selection, | |
| type="value", | |
| value='English Text (eng)', | |
| label="Select the Language of the Text (type on your keyboard to find it quickly)"), | |
| gr.Slider(minimum=0, maximum=available_artificial_voices, step=1, | |
| value=279, | |
| label="Random Seed for the artificial Voice"), | |
| gr.Slider(minimum=0.7, maximum=1.3, step=0.1, value=1.0, label="Duration Scale"), | |
| gr.Slider(minimum=0.5, maximum=1.5, step=0.1, value=1.0, label="Pitch Variance Scale"), | |
| gr.Slider(minimum=0.5, maximum=1.5, step=0.1, value=1.0, label="Energy Variance Scale"), | |
| gr.Slider(minimum=-10.0, maximum=10.0, step=0.1, value=0.0, label="Femininity / Masculinity"), | |
| gr.Slider(minimum=-10.0, maximum=10.0, step=0.1, value=0.0, label="Voice Depth") | |
| ], | |
| outputs=[gr.Audio(type="numpy", label="Speech"), | |
| gr.Image(label="Visualization")], | |
| title=title, | |
| theme="default", | |
| allow_flagging="never", | |
| article=article) | |
| iface.launch() | |