Spaces:
Runtime error
Runtime error
Boning c
commited on
Update app.py
Browse files
app.py
CHANGED
|
@@ -8,7 +8,7 @@ from html import escape
|
|
| 8 |
# Model config
|
| 9 |
PRIMARY_MODEL = "Smilyai-labs/Sam-reason-A1"
|
| 10 |
FALLBACK_MODEL = "Smilyai-labs/Sam-reason-S2.1"
|
| 11 |
-
USAGE_LIMIT =
|
| 12 |
RESET_AFTER_SECONDS = 20 * 60 # 20 minutes
|
| 13 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
| 14 |
|
|
@@ -27,7 +27,7 @@ def load_models():
|
|
| 27 |
return f"✅ Loaded: {PRIMARY_MODEL} with fallback {FALLBACK_MODEL}"
|
| 28 |
|
| 29 |
# Token streaming
|
| 30 |
-
def generate_stream(prompt, use_fallback=False, max_length=
|
| 31 |
model = fallback_model if use_fallback else primary_model
|
| 32 |
tokenizer = fallback_tokenizer if use_fallback else primary_tokenizer
|
| 33 |
input_ids = tokenizer(prompt, return_tensors="pt").input_ids.to(device)
|
|
|
|
| 8 |
# Model config
|
| 9 |
PRIMARY_MODEL = "Smilyai-labs/Sam-reason-A1"
|
| 10 |
FALLBACK_MODEL = "Smilyai-labs/Sam-reason-S2.1"
|
| 11 |
+
USAGE_LIMIT = 5
|
| 12 |
RESET_AFTER_SECONDS = 20 * 60 # 20 minutes
|
| 13 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
| 14 |
|
|
|
|
| 27 |
return f"✅ Loaded: {PRIMARY_MODEL} with fallback {FALLBACK_MODEL}"
|
| 28 |
|
| 29 |
# Token streaming
|
| 30 |
+
def generate_stream(prompt, use_fallback=False, max_length=100, temperature=0.2, top_p=0.9):
|
| 31 |
model = fallback_model if use_fallback else primary_model
|
| 32 |
tokenizer = fallback_tokenizer if use_fallback else primary_tokenizer
|
| 33 |
input_ids = tokenizer(prompt, return_tensors="pt").input_ids.to(device)
|