Update README with info on Faster Whisper
Browse files
README.md
CHANGED
@@ -46,6 +46,35 @@ python cli.py --model large --vad silero-vad --language Japanese "https://www.yo
|
|
46 |
Rather than supplying arguments to `app.py` or `cli.py`, you can also use the configuration file [config.json5](config.json5). See that file for more information.
|
47 |
If you want to use a different configuration file, you can use the `WHISPER_WEBUI_CONFIG` environment variable to specify the path to another file.
|
48 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
49 |
## Google Colab
|
50 |
|
51 |
You can also run this Web UI directly on [Google Colab](https://colab.research.google.com/drive/1qeTSvi7Bt_5RMm88ipW4fkcsMOKlDDss?usp=sharing), if you haven't got a GPU powerful enough to run the larger models.
|
@@ -87,12 +116,6 @@ cores (up to 8):
|
|
87 |
python app.py --input_audio_max_duration -1 --auto_parallel True
|
88 |
```
|
89 |
|
90 |
-
### Multiple Files
|
91 |
-
|
92 |
-
You can upload multiple files either through the "Upload files" option, or as a playlist on YouTube.
|
93 |
-
Each audio file will then be processed in turn, and the resulting SRT/VTT/Transcript will be made available in the "Download" section.
|
94 |
-
When more than one file is processed, the UI will also generate a "All_Output" zip file containing all the text output files.
|
95 |
-
|
96 |
# Docker
|
97 |
|
98 |
To run it in Docker, first install Docker and optionally the NVIDIA Container Toolkit in order to use the GPU.
|
|
|
46 |
Rather than supplying arguments to `app.py` or `cli.py`, you can also use the configuration file [config.json5](config.json5). See that file for more information.
|
47 |
If you want to use a different configuration file, you can use the `WHISPER_WEBUI_CONFIG` environment variable to specify the path to another file.
|
48 |
|
49 |
+
### Multiple Files
|
50 |
+
|
51 |
+
You can upload multiple files either through the "Upload files" option, or as a playlist on YouTube.
|
52 |
+
Each audio file will then be processed in turn, and the resulting SRT/VTT/Transcript will be made available in the "Download" section.
|
53 |
+
When more than one file is processed, the UI will also generate a "All_Output" zip file containing all the text output files.
|
54 |
+
|
55 |
+
## Faster Whisper
|
56 |
+
|
57 |
+
You can also use [Faster Whisper](https://github.com/guillaumekln/faster-whisper) as a drop-in replacement for the default Whisper which achieves up to a 4x speedup
|
58 |
+
and 2x reduction in memory usage.
|
59 |
+
|
60 |
+
To use Faster Whisper, install the requirements in `requirements-fastWhisper.txt`:
|
61 |
+
```
|
62 |
+
pip install -r requirements-fastWhisper.txt
|
63 |
+
```
|
64 |
+
And then run the App or the CLI with the `--whisper_implementation fast-whisper` flag:
|
65 |
+
```
|
66 |
+
python app.py --whisper_implementation fast-whisper --input_audio_max_duration -1 --server_name 127.0.0.1 --auto_parallel True
|
67 |
+
```
|
68 |
+
You can also select the whisper implementation in `config.json5`:
|
69 |
+
```json5
|
70 |
+
{
|
71 |
+
"whisper_implementation": "fast-whisper"
|
72 |
+
}
|
73 |
+
```
|
74 |
+
### GPU Acceleration
|
75 |
+
|
76 |
+
In order to use GPU acceleration with Faster Whisper, both CUDA 11.2 and cuDNN 8 must be installed. You may want to install it in a virtual environment like Anaconda.
|
77 |
+
|
78 |
## Google Colab
|
79 |
|
80 |
You can also run this Web UI directly on [Google Colab](https://colab.research.google.com/drive/1qeTSvi7Bt_5RMm88ipW4fkcsMOKlDDss?usp=sharing), if you haven't got a GPU powerful enough to run the larger models.
|
|
|
116 |
python app.py --input_audio_max_duration -1 --auto_parallel True
|
117 |
```
|
118 |
|
|
|
|
|
|
|
|
|
|
|
|
|
119 |
# Docker
|
120 |
|
121 |
To run it in Docker, first install Docker and optionally the NVIDIA Container Toolkit in order to use the GPU.
|