import whisper | |
import gradio as gr | |
model = whisper.load_hf_model(repo_id="jerpint/whisper", filename="base.pt") | |
def transcribe(audio, translate): | |
task = "translate" if translate else None | |
result = model.transcribe(audio, task=task) | |
return result["text"] | |
title = "BabelFish" | |
description = "Record your voice in any language, babelfish will output a transcript of what was said. Check 'Translate to english' to get an english transcription. Based on the OpenAI Whisper model" | |
gr.Interface( | |
fn=transcribe, | |
inputs=[ | |
gr.Audio(source="microphone", type="filepath"), | |
gr.Checkbox(label="Translate to english"), | |
], | |
title=title, | |
description=description, | |
examples=[ | |
["samples/french_hello.wav", True], | |
["samples/english_hello.wav", True], | |
["samples/hebrew_hello.wav", True], | |
["samples/spanish_hello.wav", True], | |
], | |
outputs="text", | |
).launch() | |