Boning c commited on
Commit
acff8f4
·
verified ·
1 Parent(s): 7bb17e1

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -2
app.py CHANGED
@@ -8,7 +8,7 @@ from html import escape
8
  # Model config
9
  PRIMARY_MODEL = "Smilyai-labs/Sam-reason-A1"
10
  FALLBACK_MODEL = "Smilyai-labs/Sam-reason-S2.1"
11
- USAGE_LIMIT = 2
12
  RESET_AFTER_SECONDS = 20 * 60 # 20 minutes
13
  device = "cuda" if torch.cuda.is_available() else "cpu"
14
 
@@ -27,7 +27,7 @@ def load_models():
27
  return f"✅ Loaded: {PRIMARY_MODEL} with fallback {FALLBACK_MODEL}"
28
 
29
  # Token streaming
30
- def generate_stream(prompt, use_fallback=False, max_length=10, temperature=0.6, top_p=0.9):
31
  model = fallback_model if use_fallback else primary_model
32
  tokenizer = fallback_tokenizer if use_fallback else primary_tokenizer
33
  input_ids = tokenizer(prompt, return_tensors="pt").input_ids.to(device)
 
8
  # Model config
9
  PRIMARY_MODEL = "Smilyai-labs/Sam-reason-A1"
10
  FALLBACK_MODEL = "Smilyai-labs/Sam-reason-S2.1"
11
+ USAGE_LIMIT = 5
12
  RESET_AFTER_SECONDS = 20 * 60 # 20 minutes
13
  device = "cuda" if torch.cuda.is_available() else "cpu"
14
 
 
27
  return f"✅ Loaded: {PRIMARY_MODEL} with fallback {FALLBACK_MODEL}"
28
 
29
  # Token streaming
30
+ def generate_stream(prompt, use_fallback=False, max_length=100, temperature=0.2, top_p=0.9):
31
  model = fallback_model if use_fallback else primary_model
32
  tokenizer = fallback_tokenizer if use_fallback else primary_tokenizer
33
  input_ids = tokenizer(prompt, return_tensors="pt").input_ids.to(device)