Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -23,11 +23,6 @@ hf_hub_download(
|
|
23 |
filename="Meta-Llama-3-8B-Instruct-Q6_K.gguf",
|
24 |
local_dir="./models"
|
25 |
)
|
26 |
-
hf_hub_download(
|
27 |
-
repo_id="bartowski/aya-23-8B-GGUF",
|
28 |
-
filename="aya-23-8B-Q6_K.gguf",
|
29 |
-
local_dir="./models"
|
30 |
-
)
|
31 |
|
32 |
css = """
|
33 |
.message-row {
|
@@ -76,8 +71,7 @@ PLACEHOLDER = """
|
|
76 |
def get_context_by_model(model_name):
|
77 |
model_context_limits = {
|
78 |
"Mistral-7B-Instruct-v0.3-Q6_K.gguf": 32768,
|
79 |
-
"Meta-Llama-3-8B-Instruct-Q6_K.gguf": 8192
|
80 |
-
"aya-23-8B-Q6_K.gguf": 8192
|
81 |
}
|
82 |
return model_context_limits.get(model_name, None)
|
83 |
|
@@ -254,8 +248,7 @@ demo = gr.ChatInterface(
|
|
254 |
),
|
255 |
gr.Dropdown([
|
256 |
'Mistral-7B-Instruct-v0.3-Q6_K.gguf',
|
257 |
-
'Meta-Llama-3-8B-Instruct-Q6_K.gguf'
|
258 |
-
'aya-23-8B-Q6_K.gguf'
|
259 |
],
|
260 |
value="Mistral-7B-Instruct-v0.3-Q6_K.gguf",
|
261 |
label="Model"
|
|
|
23 |
filename="Meta-Llama-3-8B-Instruct-Q6_K.gguf",
|
24 |
local_dir="./models"
|
25 |
)
|
|
|
|
|
|
|
|
|
|
|
26 |
|
27 |
css = """
|
28 |
.message-row {
|
|
|
71 |
def get_context_by_model(model_name):
|
72 |
model_context_limits = {
|
73 |
"Mistral-7B-Instruct-v0.3-Q6_K.gguf": 32768,
|
74 |
+
"Meta-Llama-3-8B-Instruct-Q6_K.gguf": 8192
|
|
|
75 |
}
|
76 |
return model_context_limits.get(model_name, None)
|
77 |
|
|
|
248 |
),
|
249 |
gr.Dropdown([
|
250 |
'Mistral-7B-Instruct-v0.3-Q6_K.gguf',
|
251 |
+
'Meta-Llama-3-8B-Instruct-Q6_K.gguf'
|
|
|
252 |
],
|
253 |
value="Mistral-7B-Instruct-v0.3-Q6_K.gguf",
|
254 |
label="Model"
|