Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
|
@@ -7,7 +7,7 @@ from llama_index.indices.managed.llama_cloud import (
|
|
| 7 |
LlamaCloudCompositeRetriever,
|
| 8 |
)
|
| 9 |
from llama_index.core import Settings
|
| 10 |
-
from llama_index.llms.
|
| 11 |
from llama_cloud.types import CompositeRetrievalMode
|
| 12 |
from llama_index.core.memory import ChatMemoryBuffer
|
| 13 |
from llama_index.core.chat_engine import CondensePlusContextChatEngine
|
|
@@ -26,9 +26,12 @@ from openinference.instrumentation.llama_index import LlamaIndexInstrumentor
|
|
| 26 |
# Replace with your actual LlamaCloud Project Name
|
| 27 |
LLAMA_CLOUD_PROJECT_NAME = "CustomerSupportProject"
|
| 28 |
|
| 29 |
-
# Configure
|
| 30 |
-
# Ensure
|
| 31 |
-
Settings.llm =
|
|
|
|
|
|
|
|
|
|
| 32 |
print(f"[INFO] Configured LLM: {Settings.llm.model}")
|
| 33 |
|
| 34 |
# Configure LlamaTrace (Arize Phoenix)
|
|
|
|
| 7 |
LlamaCloudCompositeRetriever,
|
| 8 |
)
|
| 9 |
from llama_index.core import Settings
|
| 10 |
+
from llama_index.llms.nebius import NebiusLLM
|
| 11 |
from llama_cloud.types import CompositeRetrievalMode
|
| 12 |
from llama_index.core.memory import ChatMemoryBuffer
|
| 13 |
from llama_index.core.chat_engine import CondensePlusContextChatEngine
|
|
|
|
| 26 |
# Replace with your actual LlamaCloud Project Name
|
| 27 |
LLAMA_CLOUD_PROJECT_NAME = "CustomerSupportProject"
|
| 28 |
|
| 29 |
+
# Configure NebiusLLM
|
| 30 |
+
# Ensure NEBIUS_API_KEY is set in your environment variables
|
| 31 |
+
Settings.llm = NebiusLLM(
|
| 32 |
+
model="meta-llama/Meta-Llama-3.1-405B-Instruct",
|
| 33 |
+
temperature=0
|
| 34 |
+
)
|
| 35 |
print(f"[INFO] Configured LLM: {Settings.llm.model}")
|
| 36 |
|
| 37 |
# Configure LlamaTrace (Arize Phoenix)
|