Spaces:
No application file
No application file
Update app.py
Browse files
app.py
CHANGED
@@ -91,8 +91,8 @@ primary_client = InferenceClient("meta-llama/Llama-3.3-70B-Instruct", token=HF_T
|
|
91 |
print("Primary model (LLaMA-3.3-70B-Instruct) client initialized.")
|
92 |
|
93 |
# Initialize InferenceClient for fallback model (Gemma-2-9b-it)
|
94 |
-
fallback_client = InferenceClient("
|
95 |
-
print("Fallback model (
|
96 |
|
97 |
|
98 |
# Load spacy model for sentence splitting
|
@@ -1184,7 +1184,7 @@ if __name__ == "__main__":
|
|
1184 |
|
1185 |
DESCRIPTION = """
|
1186 |
# LLM with Tools (DuckDuckGo Search, Date Calculation, Business Info RAG, Hugging Face Dataset Logging) and Two-Tier Model System
|
1187 |
-
Ask me anything! I can perform web searches, calculate dates, retrieve business information using RAG, and conversation data will be logged to a Hugging Face Dataset. I use a primary LLaMA-70B model for complex queries and a fallback
|
1188 |
"""
|
1189 |
|
1190 |
demo = gr.ChatInterface(
|
|
|
91 |
print("Primary model (LLaMA-3.3-70B-Instruct) client initialized.")
|
92 |
|
93 |
# Initialize InferenceClient for fallback model (Gemma-2-9b-it)
|
94 |
+
fallback_client = InferenceClient("meta-llama/Llama-4-Scout-17B-16E-Instruct", token=HF_TOKEN)
|
95 |
+
print("Fallback model (Llama-4-Scout-17B-16E-Instruct) client initialized.")
|
96 |
|
97 |
|
98 |
# Load spacy model for sentence splitting
|
|
|
1184 |
|
1185 |
DESCRIPTION = """
|
1186 |
# LLM with Tools (DuckDuckGo Search, Date Calculation, Business Info RAG, Hugging Face Dataset Logging) and Two-Tier Model System
|
1187 |
+
Ask me anything! I can perform web searches, calculate dates, retrieve business information using RAG, and conversation data will be logged to a Hugging Face Dataset. I use a primary LLaMA-70B model for complex queries and a fallback Llama-4-Scout-17B-16E-Instruct model for simpler ones and RAG synthesis.
|
1188 |
"""
|
1189 |
|
1190 |
demo = gr.ChatInterface(
|