Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -13,14 +13,14 @@ import ollama
|
|
| 13 |
|
| 14 |
# Model from run.sh
|
| 15 |
MODEL_ID_MAP = {
|
|
|
|
|
|
|
| 16 |
"(ιΏιεε)Qwen3-4B-Instruct-2507": 'hf.co/bartowski/Qwen_Qwen3-4B-Instruct-2507-GGUF:Q4_K_M',
|
| 17 |
#"(ιΏιεε)Qwen3-4B-Thinking-2507": 'hf.co/bartowski/Qwen_Qwen3-4B-Thinking-2507-GGUF:Q4_K_M',
|
| 18 |
-
"(HuggingFace)SmolLM2-360M": 'smollm2:360m-instruct-q5_K_M',
|
| 19 |
-
"(Meta)Llama3.2-3B-Instruct": 'hf.co/bartowski/Llama-3.2-3B-Instruct-GGUF:Q4_K_M', # OK speed with CPU
|
| 20 |
#"(Google)Gemma3n-e2b-it": 'gemma3n:e2b-it-q4_K_M',
|
| 21 |
-
"(IBM)Granite3.3-2B": 'granite3.3:2b',
|
| 22 |
#"(Tencent)ζ··ε
-1.8B-Instruct":'hf.co/bartowski/tencent_Hunyuan-1.8B-Instruct-GGUF:Q4_K_M',
|
| 23 |
-
#"(Tencent)ζ··ε
-4B-Instruct": 'hf.co/bartowski/tencent_Hunyuan-4B-Instruct-GGUF:Q4_K_M'
|
|
|
|
| 24 |
}
|
| 25 |
|
| 26 |
|
|
@@ -32,7 +32,7 @@ DEFAULT_SYSTEM_PROMPT = """Answer everything in simple, smart, relevant and accu
|
|
| 32 |
|
| 33 |
# --- Gradio Interface ---
|
| 34 |
with gr.Blocks(theme=gr.themes.Default(primary_hue="blue", secondary_hue="neutral")) as demo:
|
| 35 |
-
gr.Markdown(f"## LLM/SLM running with CPU") # Changed title to be more generic
|
| 36 |
gr.Markdown(f"(Run-Location-As: `{check_ipinfo}`)")
|
| 37 |
gr.Markdown("Chat with the model, customize its behavior with a system prompt, and toggle streaming output.")
|
| 38 |
|
|
|
|
| 13 |
|
| 14 |
# Model from run.sh
|
| 15 |
MODEL_ID_MAP = {
|
| 16 |
+
"(IBM)Granite3.3-2B": 'granite3.3:2b',
|
| 17 |
+
"(Meta)Llama3.2-3B-Instruct": 'hf.co/bartowski/Llama-3.2-3B-Instruct-GGUF:Q4_K_M', # OK speed with CPU
|
| 18 |
"(ιΏιεε)Qwen3-4B-Instruct-2507": 'hf.co/bartowski/Qwen_Qwen3-4B-Instruct-2507-GGUF:Q4_K_M',
|
| 19 |
#"(ιΏιεε)Qwen3-4B-Thinking-2507": 'hf.co/bartowski/Qwen_Qwen3-4B-Thinking-2507-GGUF:Q4_K_M',
|
|
|
|
|
|
|
| 20 |
#"(Google)Gemma3n-e2b-it": 'gemma3n:e2b-it-q4_K_M',
|
|
|
|
| 21 |
#"(Tencent)ζ··ε
-1.8B-Instruct":'hf.co/bartowski/tencent_Hunyuan-1.8B-Instruct-GGUF:Q4_K_M',
|
| 22 |
+
#"(Tencent)ζ··ε
-4B-Instruct": 'hf.co/bartowski/tencent_Hunyuan-4B-Instruct-GGUF:Q4_K_M',
|
| 23 |
+
"(HuggingFace)SmolLM2-360M": 'smollm2:360m-instruct-q5_K_M'
|
| 24 |
}
|
| 25 |
|
| 26 |
|
|
|
|
| 32 |
|
| 33 |
# --- Gradio Interface ---
|
| 34 |
with gr.Blocks(theme=gr.themes.Default(primary_hue="blue", secondary_hue="neutral")) as demo:
|
| 35 |
+
gr.Markdown(f"## HH demo: LLM/SLM chatbot running with CPU only.") # Changed title to be more generic
|
| 36 |
gr.Markdown(f"(Run-Location-As: `{check_ipinfo}`)")
|
| 37 |
gr.Markdown("Chat with the model, customize its behavior with a system prompt, and toggle streaming output.")
|
| 38 |
|