Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -19,8 +19,8 @@ hf_hub_download(
|
|
| 19 |
local_dir="./models"
|
| 20 |
)
|
| 21 |
hf_hub_download(
|
| 22 |
-
repo_id="
|
| 23 |
-
filename="
|
| 24 |
local_dir="./models"
|
| 25 |
)
|
| 26 |
|
|
@@ -126,7 +126,7 @@ def respond(
|
|
| 126 |
|
| 127 |
llm = Llama(
|
| 128 |
model_path=f"models/{model}",
|
| 129 |
-
flash_attn=
|
| 130 |
n_threads=40,
|
| 131 |
n_gpu_layers=81,
|
| 132 |
n_batch=1024,
|
|
@@ -213,7 +213,7 @@ demo = gr.ChatInterface(
|
|
| 213 |
),
|
| 214 |
gr.Dropdown([
|
| 215 |
'Mistral-7B-Instruct-v0.3-f32.gguf',
|
| 216 |
-
'
|
| 217 |
],
|
| 218 |
value="Mistral-7B-Instruct-v0.3-f32.gguf",
|
| 219 |
label="Model"
|
|
|
|
| 19 |
local_dir="./models"
|
| 20 |
)
|
| 21 |
hf_hub_download(
|
| 22 |
+
repo_id="bartowski/Einstein-v6-7B-GGUF",
|
| 23 |
+
filename="Einstein-v6-7B-Q6_K.gguf",
|
| 24 |
local_dir="./models"
|
| 25 |
)
|
| 26 |
|
|
|
|
| 126 |
|
| 127 |
llm = Llama(
|
| 128 |
model_path=f"models/{model}",
|
| 129 |
+
flash_attn=True,
|
| 130 |
n_threads=40,
|
| 131 |
n_gpu_layers=81,
|
| 132 |
n_batch=1024,
|
|
|
|
| 213 |
),
|
| 214 |
gr.Dropdown([
|
| 215 |
'Mistral-7B-Instruct-v0.3-f32.gguf',
|
| 216 |
+
'Einstein-v6-7B-Q6_K.gguf'
|
| 217 |
],
|
| 218 |
value="Mistral-7B-Instruct-v0.3-f32.gguf",
|
| 219 |
label="Model"
|