fffiloni commited on
Commit
20c6475
·
verified ·
1 Parent(s): 01ce88b

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +7 -5
app.py CHANGED
@@ -60,8 +60,9 @@ def infer_img2img(prompt, audio_path, desired_strength, progress=gr.Progress(tra
60
  norm_spec = pad_spec(norm_spec, 1024)
61
  norm_spec = normalize(norm_spec) # normalize to [-1, 1], because pipeline do not normalize for torch.Tensor input
62
 
63
- raw_image = image_add_color(torch_to_pil(norm_spec[:,:,:width]))
64
-
 
65
  # Generation for different strength
66
  image_list = []
67
  audio_list = []
@@ -179,9 +180,10 @@ with gr.Blocks(css=css) as demo:
179
  submit_btn_img2img = gr.Button("Submit")
180
  audio_out_img2img = gr.Audio(label="Audio Ressult")
181
 
182
- with gr.Row():
183
- input_spectrogram = gr.Image(label="Input Spectrogram")
184
- output_spectrogram = gr.Image(label="Output Spectrogram")
 
185
 
186
  gr.Examples(
187
  examples = [
 
60
  norm_spec = pad_spec(norm_spec, 1024)
61
  norm_spec = normalize(norm_spec) # normalize to [-1, 1], because pipeline do not normalize for torch.Tensor input
62
 
63
+ # raw_image = image_add_color(torch_to_pil(norm_spec[:,:,:width]))
64
+ raw_image = image_add_color(torch_to_pil(norm_spec))
65
+
66
  # Generation for different strength
67
  image_list = []
68
  audio_list = []
 
180
  submit_btn_img2img = gr.Button("Submit")
181
  audio_out_img2img = gr.Audio(label="Audio Ressult")
182
 
183
+ with gr.Accordion("Compare Spectrograms", open=False):
184
+ with gr.Column():
185
+ input_spectrogram = gr.Image(label="Input Spectrogram")
186
+ output_spectrogram = gr.Image(label="Output Spectrogram")
187
 
188
  gr.Examples(
189
  examples = [