From 535b75f487e42f5077c9453f0d589525408d4a63 Mon Sep 17 00:00:00 2001 From: McCloudS <64094529+McCloudS@users.noreply.github.com> Date: Thu, 26 Oct 2023 09:36:00 -0600 Subject: [PATCH] Update README.md --- README.md | 3 +-- 1 file changed, 1 insertion(+), 2 deletions(-) diff --git a/README.md b/README.md index daafe48..384535c 100644 --- a/README.md +++ b/README.md @@ -96,8 +96,7 @@ The following environment variables are available in Docker. They will default | Variable | Default Value | Description | |---------------------------|------------------------|-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------| | TRANSCRIBE_DEVICE | 'cpu' | Can transcribe via gpu (Cuda only) or cpu. Takes option of "cpu", "gpu", "cuda". You must be running a cuda dockerfile to use the cuda/gpu options without failing. | -| WHISPER_MODEL | 'medium' | Can be:'tiny', 'tiny.en', 'base', 'base.en', 'small', 'small.en', 'medium', 'medium.en', 'large-v1', - 'large-v2', or 'large' | +| WHISPER_MODEL | 'medium' | Can be:'tiny', 'tiny.en', 'base', 'base.en', 'small', 'small.en', 'medium', 'medium.en', 'large-v1','large-v2', or 'large' | | CONCURRENT_TRANSCRIPTIONS | 2 | Number of files it will transcribe in parallel | | WHISPER_THREADS | 4 | number of threads to use during computation | | MODEL_PATH | '.' | This is where the WHISPER_MODEL will be stored. This defaults to placing it where you execute the script |