From 48e7caad77a731c9de84beb4346357466e62bc7e Mon Sep 17 00:00:00 2001 From: Mark Berger Date: Mon, 24 Jul 2023 11:45:38 +0200 Subject: [PATCH] Update transcribe.py -> small change in `batch_size` description Changed the description of the `batch_size` parameter. --- whisperx/transcribe.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/whisperx/transcribe.py b/whisperx/transcribe.py index 1855178..1cc144e 100644 --- a/whisperx/transcribe.py +++ b/whisperx/transcribe.py @@ -22,7 +22,7 @@ def cli(): parser.add_argument("--model_dir", type=str, default=None, help="the path to save model files; uses ~/.cache/whisper by default") parser.add_argument("--device", default="cuda" if torch.cuda.is_available() else "cpu", help="device to use for PyTorch inference") parser.add_argument("--device_index", default=0, type=int, help="device index to use for FasterWhisper inference") - parser.add_argument("--batch_size", default=8, type=int, help="device to use for PyTorch inference") + parser.add_argument("--batch_size", default=8, type=int, help="the preferred batch size for inference") parser.add_argument("--compute_type", default="float16", type=str, choices=["float16", "float32", "int8"], help="compute type for computation") parser.add_argument("--output_dir", "-o", type=str, default=".", help="directory to save the outputs")