Skip to content

A nearly-live implementation of OpenAI's Whisper.

License

Notifications You must be signed in to change notification settings

positivewon/WhisperLive

Repository files navigation

whisper-live

A nearly-live implementation of OpenAI's Whisper.

This project is a real-time transcription application that uses the OpenAI Whisper model to convert speech input into text output. It can be used to transcribe both live audio input from microphone and pre-recorded audio files.

Unlike traditional speech recognition systems that rely on continuous audio streaming, we use voice activity detection (VAD) to detect the presence of speech and only send the audio data to whisper when speech is detected. This helps to reduce the amount of data sent to the whisper model and improves the accuracy of the transcription output.

Installation

  • Install PyAudio and ffmpeg
 bash setup.sh
  • To install client requirements
 pip install -r requirements/client.txt
  • To install server requirements
 pip install -r requirements/server.txt

Getting Started

  • Run the server
 python server.py
  • On the client side

    • To transcribe an audio file:
      python client.py --audio "audio.wav" --host "localhost" --port "9090" --multilingual --language "hi" --task "transcribe"
                                                                                                                  "translate"

    This command transcribes the specified audio file (audio.wav) using the Whisper model. It connects to the server running on localhost at port 9090. It also enables the multilingual feature, allowing transcription in multiple languages. The --language flag specifies the target language for transcription, in this case, Hindi ("hi"). The --task flag is set to "transcribe" to indicate that transcription is the desired task. Also, --task can be set to "translate" to translate source language to English.

    • To transcribe from microphone:
     python client.py --host "localhost" --port "9090" --multilingual --language "en" --task "transcribe"

    This command captures audio from the microphone and sends it to the server for transcription. It uses the same options as the previous command, enabling the multilingual feature and specifying the target language and task.

Transcribe audio from browser

  • Run the server
 python server.py

This would start the websocket server on port 9090.

Chrome Extension

Firefox Extension

Whisper Live Server in Docker

  • Build docker container
 docker build . -t whisper-live
  • Run docker container
 docker run -it --gpus all -p 9090:9090 whisper-live:latest

Future Work

  • Update Documentation.
  • Keep only a single server implementation i.e. websockets and get rid of the socket implementation in server.py. Also, update client.py to websockets-client implemenation.
  • Add translation to other languages on top of transcription.

Citations

@article{Whisper
  title = {Robust Speech Recognition via Large-Scale Weak Supervision},
  url = {https://arxiv.org/abs/2212.04356},
  author = {Radford, Alec and Kim, Jong Wook and Xu, Tao and Brockman, Greg and McLeavey, Christine and Sutskever, Ilya},
  publisher = {arXiv},
  year = {2022},
}
@misc{Silero VAD,
  author = {Silero Team},
  title = {Silero VAD: pre-trained enterprise-grade Voice Activity Detector (VAD), Number Detector and Language Classifier},
  year = {2021},
  publisher = {GitHub},
  journal = {GitHub repository},
  howpublished = {\url{https://github.com/snakers4/silero-vad}},
  commit = {insert_some_commit_here},
  email = {hello@silero.ai}
}

About

A nearly-live implementation of OpenAI's Whisper.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 67.0%
  • JavaScript 23.4%
  • HTML 6.5%
  • Shell 1.8%
  • CSS 1.3%