Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -8,6 +8,7 @@ import gradio as gr
|
|
8 |
import numpy as np
|
9 |
import soundfile as sf
|
10 |
import torch
|
|
|
11 |
|
12 |
from dia.model import Dia
|
13 |
|
@@ -21,21 +22,6 @@ parser.add_argument("--share", action="store_true", help="Enable Gradio sharing"
|
|
21 |
|
22 |
args = parser.parse_args()
|
23 |
|
24 |
-
|
25 |
-
# Determine device
|
26 |
-
if args.device:
|
27 |
-
device = torch.device(args.device)
|
28 |
-
elif torch.cuda.is_available():
|
29 |
-
device = torch.device("cuda")
|
30 |
-
# Simplified MPS check for broader compatibility
|
31 |
-
elif hasattr(torch.backends, "mps") and torch.backends.mps.is_available():
|
32 |
-
# Basic check is usually sufficient, detailed check can be problematic
|
33 |
-
device = torch.device("mps")
|
34 |
-
else:
|
35 |
-
device = torch.device("cpu")
|
36 |
-
|
37 |
-
print(f"Using device: {device}")
|
38 |
-
|
39 |
# Load Nari model and config
|
40 |
print("Loading Nari model...")
|
41 |
try:
|
@@ -46,6 +32,7 @@ except Exception as e:
|
|
46 |
raise
|
47 |
|
48 |
|
|
|
49 |
def run_inference(
|
50 |
text_input: str,
|
51 |
audio_prompt_input: Optional[Tuple[int, np.ndarray]],
|
|
|
8 |
import numpy as np
|
9 |
import soundfile as sf
|
10 |
import torch
|
11 |
+
import spaces
|
12 |
|
13 |
from dia.model import Dia
|
14 |
|
|
|
22 |
|
23 |
args = parser.parse_args()
|
24 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
25 |
# Load Nari model and config
|
26 |
print("Loading Nari model...")
|
27 |
try:
|
|
|
32 |
raise
|
33 |
|
34 |
|
35 |
+
@spaces.GPU
|
36 |
def run_inference(
|
37 |
text_input: str,
|
38 |
audio_prompt_input: Optional[Tuple[int, np.ndarray]],
|