Spaces:
Build error
Build error
yama
commited on
Commit
·
7943b58
1
Parent(s):
6006f4b
Update app.py
Browse files
app.py
CHANGED
@@ -26,6 +26,7 @@ import wave
|
|
26 |
import contextlib
|
27 |
from transformers import pipeline
|
28 |
import psutil
|
|
|
29 |
|
30 |
whisper_models = ["tiny", "base", "small", "medium", "large-v1", "large-v2"]
|
31 |
source_languages = {
|
@@ -351,6 +352,22 @@ def speech_to_text(video_file_path, selected_source_lang, whisper_model, num_spe
|
|
351 |
raise RuntimeError("Error Running inference with local model", e)
|
352 |
|
353 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
354 |
# ---- Gradio Layout -----
|
355 |
# Inspiration from https://huggingface.co/spaces/RASMUS/Whisper-youtube-crosslingual-subtitles
|
356 |
video_in = gr.Video(label="Video file", mirror_webcam=False)
|
@@ -375,7 +392,7 @@ openai_prompt_in = gr.TextArea(label="openai_prompt", value="""会議の文字
|
|
375 |
- 会議の目的
|
376 |
- 会議の内容
|
377 |
- 会議の結果""")
|
378 |
-
|
379 |
title = "Whisper speaker diarization"
|
380 |
demo = gr.Blocks(title=title)
|
381 |
demo.encrypt = False
|
@@ -445,7 +462,8 @@ with demo:
|
|
445 |
transcription_df.render()
|
446 |
openai_key_in.render()
|
447 |
openai_prompt_in.render()
|
448 |
-
|
|
|
449 |
system_info.render()
|
450 |
gr.Markdown(
|
451 |
'''<center><img src='https://visitor-badge.glitch.me/badge?page_id=WhisperDiarizationSpeakers' alt='visitor badge'><a href="https://opensource.org/licenses/Apache-2.0"><img src='https://img.shields.io/badge/License-Apache_2.0-blue.svg' alt='License: Apache 2.0'></center>''')
|
|
|
26 |
import contextlib
|
27 |
from transformers import pipeline
|
28 |
import psutil
|
29 |
+
import openai
|
30 |
|
31 |
whisper_models = ["tiny", "base", "small", "medium", "large-v1", "large-v2"]
|
32 |
source_languages = {
|
|
|
352 |
raise RuntimeError("Error Running inference with local model", e)
|
353 |
|
354 |
|
355 |
+
def create_transcription_summary(openai_key, prompt):
|
356 |
+
openai.api_key = openai_key
|
357 |
+
system_template = prompt
|
358 |
+
|
359 |
+
transcript_text = ""
|
360 |
+
completion = openai.ChatCompletion.create(
|
361 |
+
model="gpt-3.5-turbo",
|
362 |
+
messages=[
|
363 |
+
{"role": "system", "content": system_template},
|
364 |
+
{"role": "user", "content": transcript_text}
|
365 |
+
]
|
366 |
+
)
|
367 |
+
transcript_summary = completion.choices[0].message.content
|
368 |
+
return transcript_summary
|
369 |
+
|
370 |
+
|
371 |
# ---- Gradio Layout -----
|
372 |
# Inspiration from https://huggingface.co/spaces/RASMUS/Whisper-youtube-crosslingual-subtitles
|
373 |
video_in = gr.Video(label="Video file", mirror_webcam=False)
|
|
|
392 |
- 会議の目的
|
393 |
- 会議の内容
|
394 |
- 会議の結果""")
|
395 |
+
transcription_summary_out = gr.Textbox(label="transcription_summary")
|
396 |
title = "Whisper speaker diarization"
|
397 |
demo = gr.Blocks(title=title)
|
398 |
demo.encrypt = False
|
|
|
462 |
transcription_df.render()
|
463 |
openai_key_in.render()
|
464 |
openai_prompt_in.render()
|
465 |
+
transcription_summary_btn = gr.Button("Evaluate and analyze transcription content")
|
466 |
+
transcription_summary_out.render()
|
467 |
system_info.render()
|
468 |
gr.Markdown(
|
469 |
'''<center><img src='https://visitor-badge.glitch.me/badge?page_id=WhisperDiarizationSpeakers' alt='visitor badge'><a href="https://opensource.org/licenses/Apache-2.0"><img src='https://img.shields.io/badge/License-Apache_2.0-blue.svg' alt='License: Apache 2.0'></center>''')
|