Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -19,8 +19,8 @@ hf_hub_download(
|
|
19 |
local_dir="./models"
|
20 |
)
|
21 |
hf_hub_download(
|
22 |
-
repo_id="
|
23 |
-
filename="
|
24 |
local_dir="./models"
|
25 |
)
|
26 |
|
@@ -126,7 +126,7 @@ def respond(
|
|
126 |
|
127 |
llm = Llama(
|
128 |
model_path=f"models/{model}",
|
129 |
-
flash_attn=
|
130 |
n_threads=40,
|
131 |
n_gpu_layers=81,
|
132 |
n_batch=1024,
|
@@ -213,7 +213,7 @@ demo = gr.ChatInterface(
|
|
213 |
),
|
214 |
gr.Dropdown([
|
215 |
'Mistral-7B-Instruct-v0.3-f32.gguf',
|
216 |
-
'
|
217 |
],
|
218 |
value="Mistral-7B-Instruct-v0.3-f32.gguf",
|
219 |
label="Model"
|
|
|
19 |
local_dir="./models"
|
20 |
)
|
21 |
hf_hub_download(
|
22 |
+
repo_id="bartowski/Einstein-v6-7B-GGUF",
|
23 |
+
filename="Einstein-v6-7B-Q6_K.gguf",
|
24 |
local_dir="./models"
|
25 |
)
|
26 |
|
|
|
126 |
|
127 |
llm = Llama(
|
128 |
model_path=f"models/{model}",
|
129 |
+
flash_attn=True,
|
130 |
n_threads=40,
|
131 |
n_gpu_layers=81,
|
132 |
n_batch=1024,
|
|
|
213 |
),
|
214 |
gr.Dropdown([
|
215 |
'Mistral-7B-Instruct-v0.3-f32.gguf',
|
216 |
+
'Einstein-v6-7B-Q6_K.gguf'
|
217 |
],
|
218 |
value="Mistral-7B-Instruct-v0.3-f32.gguf",
|
219 |
label="Model"
|