Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -20,7 +20,7 @@ def generate_text(prompt, temperature, top_p):
|
|
20 |
input_tokens = input_tokens.to('cpu')
|
21 |
|
22 |
generated_text = prompt_with_eos # Start with the initial prompt plus "EOS"
|
23 |
-
prompt_length = len(generated_text)
|
24 |
|
25 |
for _ in range(80): # Adjust the range to control the number of tokens generated
|
26 |
with torch.no_grad():
|
@@ -48,7 +48,7 @@ interface = gr.Interface(
|
|
48 |
fn=generate_text,
|
49 |
inputs=[
|
50 |
gr.Textbox(lines=2, placeholder="Enter your prompt here..."),
|
51 |
-
gr.Slider(minimum=0.
|
52 |
gr.Slider(minimum=0.05, maximum=1.0, value=0.1, label="Top-p")
|
53 |
],
|
54 |
outputs=gr.Textbox(),
|
|
|
20 |
input_tokens = input_tokens.to('cpu')
|
21 |
|
22 |
generated_text = prompt_with_eos # Start with the initial prompt plus "EOS"
|
23 |
+
prompt_length = len(generated_text)
|
24 |
|
25 |
for _ in range(80): # Adjust the range to control the number of tokens generated
|
26 |
with torch.no_grad():
|
|
|
48 |
fn=generate_text,
|
49 |
inputs=[
|
50 |
gr.Textbox(lines=2, placeholder="Enter your prompt here..."),
|
51 |
+
gr.Slider(minimum=0.1, maximum=1.0, value=0.1, label="Temperature"),
|
52 |
gr.Slider(minimum=0.05, maximum=1.0, value=0.1, label="Top-p")
|
53 |
],
|
54 |
outputs=gr.Textbox(),
|