Skip to content

Commit

Permalink
Merge remote-tracking branch 'upstream/main' into save_transcript
Browse files Browse the repository at this point in the history
  • Loading branch information
makaveli10 committed Feb 1, 2024
2 parents f4027de + e64bc9f commit f590446
Show file tree
Hide file tree
Showing 6 changed files with 15 additions and 10 deletions.
2 changes: 1 addition & 1 deletion Audio-Transcription-Chrome/options.js
Original file line number Diff line number Diff line change
Expand Up @@ -103,7 +103,7 @@ async function startRecord(option) {
multilingual: option.multilingual,
language: option.language,
task: option.task,
model_size: option.modelSize
model: option.modelSize
})
);
};
Expand Down
2 changes: 1 addition & 1 deletion Audio-Transcription-Firefox/content.js
Original file line number Diff line number Diff line change
Expand Up @@ -78,7 +78,7 @@ function startRecording(data) {
multilingual: data.useMultilingual,
language: data.language,
task: data.task,
model_size: data.modelSize
model: data.modelSize
})
);
};
Expand Down
5 changes: 3 additions & 2 deletions TensorRT_whisper.md
Original file line number Diff line number Diff line change
Expand Up @@ -37,17 +37,18 @@ python -c "import torch; import tensorrt; import tensorrt_llm"
- We build `small.en` and `small` multilingual TensorRT engine. The script logs the path of the directory with Whisper TensorRT engine. We need the model_path to run the server.
```bash
# convert small.en
bash build_whisper_tensorrt /root/TensorRT-LLM-examples small.en
bash scripts/build_whisper_tensorrt.sh /root/TensorRT-LLM-examples small.en

# convert small multilingual model
bash build_whisper_tensorrt /root/TensorRT-LLM-examples small
bash scripts/build_whisper_tensorrt.sh /root/TensorRT-LLM-examples small
```

## Run WhisperLive Server with TensorRT Backend
```bash
cd /home/WhisperLive

# Install requirements
bash scripts/setup.sh
pip install -r requirements/server.txt

# Required to create mel spectogram
Expand Down
6 changes: 5 additions & 1 deletion requirements/server.txt
Original file line number Diff line number Diff line change
Expand Up @@ -2,4 +2,8 @@ faster-whisper==0.10.0
torch
websockets
onnxruntime==1.16.0
numba
numba
openai-whisper
kaldialign
soundfile
ffmpeg-python
6 changes: 3 additions & 3 deletions whisper_live/server.py
Original file line number Diff line number Diff line change
Expand Up @@ -414,10 +414,10 @@ def __init__(

def warmup(self, warmup_steps=10):
logging.info("[INFO:] Warming up TensorRT engine..")
mel, duration = self.transcriber.log_mel_spectrogram("tests/jfk.flac")
mel, _ = self.transcriber.log_mel_spectrogram("tests/jfk.flac")
for i in range(warmup_steps):
last_segment = self.transcriber.transcribe(mel)

self.transcriber.transcribe(mel)
def set_eos(self, eos):
self.lock.acquire()
self.eos = eos
Expand Down
4 changes: 2 additions & 2 deletions whisper_live/transcriber_tensorrt.py
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@
from whisper.tokenizer import get_tokenizer
from whisper_live.tensorrt_utils import (mel_filters, store_transcripts,
write_error_stats, load_audio_wav_format,
pad_or_trim)
pad_or_trim, load_audio)

import tensorrt_llm
import tensorrt_llm.logger as logger
Expand Down Expand Up @@ -337,4 +337,4 @@ def decode_wav_file(
if normalizer:
prediction = normalizer(prediction)

return prediction.strip()
return prediction.strip()

0 comments on commit f590446

Please sign in to comment.