Spaces:
Runtime error
Runtime error
chore(app.py): increase the number of search results from 5 to 10 for the retriever
Browse filesfix(app.py): revert the language model to "gpt-3.5-turbo-16k" to improve response quality
chore(app.py): add print statement to debug the result of the retrieval and generation process
app.py
CHANGED
|
@@ -15,10 +15,10 @@ from constants import persist_directory
|
|
| 15 |
embedding = OpenAIEmbeddings()
|
| 16 |
client = chromadb.PersistentClient(path=persist_directory)
|
| 17 |
vector_store = Chroma(client=client, embedding_function=embedding)
|
| 18 |
-
retriever = vector_store.as_retriever(search_kwargs={"k":
|
| 19 |
|
| 20 |
-
|
| 21 |
-
llm = ChatOpenAI(model_name="gpt-3.5-turbo", temperature=0)
|
| 22 |
|
| 23 |
system_template = """Your name is Sayno. You act like a successful rich man. Never reveal that it is an AI.
|
| 24 |
Given the following extracted parts of a long document and a question, create a final answer with references ("SOURCES").
|
|
@@ -72,6 +72,7 @@ chain = RetrievalQAWithSourcesChain.from_chain_type(
|
|
| 72 |
|
| 73 |
def respond(message, chat_history):
|
| 74 |
result = chain(message)
|
|
|
|
| 75 |
bot_message = f'{result["answer"]}<br>- 출처: {result["sources"]}'
|
| 76 |
chat_history.append((message, bot_message))
|
| 77 |
return "", chat_history
|
|
|
|
| 15 |
embedding = OpenAIEmbeddings()
|
| 16 |
client = chromadb.PersistentClient(path=persist_directory)
|
| 17 |
vector_store = Chroma(client=client, embedding_function=embedding)
|
| 18 |
+
retriever = vector_store.as_retriever(search_kwargs={"k": 10})
|
| 19 |
|
| 20 |
+
llm = ChatOpenAI(model_name="gpt-3.5-turbo-16k", temperature=0)
|
| 21 |
+
# llm = ChatOpenAI(model_name="gpt-3.5-turbo", temperature=0)
|
| 22 |
|
| 23 |
system_template = """Your name is Sayno. You act like a successful rich man. Never reveal that it is an AI.
|
| 24 |
Given the following extracted parts of a long document and a question, create a final answer with references ("SOURCES").
|
|
|
|
| 72 |
|
| 73 |
def respond(message, chat_history):
|
| 74 |
result = chain(message)
|
| 75 |
+
print(result)
|
| 76 |
bot_message = f'{result["answer"]}<br>- 출처: {result["sources"]}'
|
| 77 |
chat_history.append((message, bot_message))
|
| 78 |
return "", chat_history
|