From c1197c490e884adf1fa3fafcf56327ed7b6cf849 Mon Sep 17 00:00:00 2001 From: Paul F <48354391+dotgrid@users.noreply.github.com> Date: Sat, 19 Aug 2023 08:19:49 +0100 Subject: [PATCH] Document --compute_type command line option --- README.md | 4 ++++ 1 file changed, 4 insertions(+) diff --git a/README.md b/README.md index 3be1a3c..66193a1 100644 --- a/README.md +++ b/README.md @@ -127,6 +127,10 @@ To label the transcript with speaker ID's (set number of speakers if known e.g. whisperx examples/sample01.wav --model large-v2 --diarize --highlight_words True +To run on CPU instead of GPU (and for running on Mac OS X): + + whisperx examples/sample01.wav --compute_type int8 + ### Other languages The phoneme ASR alignment model is *language-specific*, for tested languages these models are [automatically picked from torchaudio pipelines or huggingface](https://github.com/m-bain/whisperX/blob/e909f2f766b23b2000f2d95df41f9b844ac53e49/whisperx/transcribe.py#L22).