Sync from GitHub repo
Browse filesThis Space is synced from the GitHub repo: https://github.com/SWivid/F5-TTS. Please submit contributions to the Space there
- README_REPO.md +5 -3
- pyproject.toml +1 -1
- src/f5_tts/runtime/triton_trtllm/README.md +5 -5
README_REPO.md
CHANGED
|
@@ -114,9 +114,11 @@ Deployment solution with Triton and TensorRT-LLM.
|
|
| 114 |
#### Benchmark Results
|
| 115 |
Decoding on a single L20 GPU, using 26 different prompt_audio & target_text pairs.
|
| 116 |
|
| 117 |
-
| Model
|
| 118 |
-
|
| 119 |
-
| F5-TTS Base (Vocos) |
|
|
|
|
|
|
|
| 120 |
|
| 121 |
See [detailed instructions](src/f5_tts/runtime/triton_trtllm/README.md) for more information.
|
| 122 |
|
|
|
|
| 114 |
#### Benchmark Results
|
| 115 |
Decoding on a single L20 GPU, using 26 different prompt_audio & target_text pairs.
|
| 116 |
|
| 117 |
+
| Model | Concurrency | Avg Latency | RTF | Mode |
|
| 118 |
+
|---------------------|----------------|-------------|--------|-----------------|
|
| 119 |
+
| F5-TTS Base (Vocos) | 2 | 253 ms | 0.0394 | Client-Server |
|
| 120 |
+
| F5-TTS Base (Vocos) | 1 (Batch_size) | - | 0.0402 | Offline TRT-LLM |
|
| 121 |
+
| F5-TTS Base (Vocos) | 1 (Batch_size) | - | 0.1467 | Offline Pytorch |
|
| 122 |
|
| 123 |
See [detailed instructions](src/f5_tts/runtime/triton_trtllm/README.md) for more information.
|
| 124 |
|
pyproject.toml
CHANGED
|
@@ -4,7 +4,7 @@ build-backend = "setuptools.build_meta"
|
|
| 4 |
|
| 5 |
[project]
|
| 6 |
name = "f5-tts"
|
| 7 |
-
version = "1.1.
|
| 8 |
description = "F5-TTS: A Fairytaler that Fakes Fluent and Faithful Speech with Flow Matching"
|
| 9 |
readme = "README.md"
|
| 10 |
license = {text = "MIT License"}
|
|
|
|
| 4 |
|
| 5 |
[project]
|
| 6 |
name = "f5-tts"
|
| 7 |
+
version = "1.1.1"
|
| 8 |
description = "F5-TTS: A Fairytaler that Fakes Fluent and Faithful Speech with Flow Matching"
|
| 9 |
readme = "README.md"
|
| 10 |
license = {text = "MIT License"}
|
src/f5_tts/runtime/triton_trtllm/README.md
CHANGED
|
@@ -59,11 +59,11 @@ benchmark.py --output-dir $log_dir \
|
|
| 59 |
### Benchmark Results
|
| 60 |
Decoding on a single L20 GPU, using 26 different prompt_audio/target_text pairs.
|
| 61 |
|
| 62 |
-
| Model
|
| 63 |
-
|
| 64 |
-
| F5-TTS Base (Vocos) | 2
|
| 65 |
-
| F5-TTS Base (Vocos) | 1 (Batch_size)
|
| 66 |
-
| F5-TTS Base (Vocos) | 1 (Batch_size)
|
| 67 |
|
| 68 |
### Credits
|
| 69 |
1. [F5-TTS-TRTLLM](https://github.com/Bigfishering/f5-tts-trtllm)
|
|
|
|
| 59 |
### Benchmark Results
|
| 60 |
Decoding on a single L20 GPU, using 26 different prompt_audio/target_text pairs.
|
| 61 |
|
| 62 |
+
| Model | Concurrency | Avg Latency | RTF | Mode |
|
| 63 |
+
|---------------------|----------------|-------------|--------|-----------------|
|
| 64 |
+
| F5-TTS Base (Vocos) | 2 | 253 ms | 0.0394 | Client-Server |
|
| 65 |
+
| F5-TTS Base (Vocos) | 1 (Batch_size) | - | 0.0402 | Offline TRT-LLM |
|
| 66 |
+
| F5-TTS Base (Vocos) | 1 (Batch_size) | - | 0.1467 | Offline Pytorch |
|
| 67 |
|
| 68 |
### Credits
|
| 69 |
1. [F5-TTS-TRTLLM](https://github.com/Bigfishering/f5-tts-trtllm)
|