Commit
·
4a6114a
1
Parent(s):
c1f91db
fix latex
Browse files- app.py +2 -5
- utils/openai_utils.py +2 -2
app.py
CHANGED
|
@@ -6,7 +6,6 @@ import streamlit as st
|
|
| 6 |
|
| 7 |
# from openai import OpenAI
|
| 8 |
import random
|
| 9 |
-
# import prompts
|
| 10 |
|
| 11 |
from utils.help import get_disclaimer
|
| 12 |
from utils.format import sec_to_time, fix_latex, get_youtube_embed
|
|
@@ -171,9 +170,6 @@ text_data_YT, context_embeddings_YT = load_youtube_data(base_path, model_name, y
|
|
| 171 |
text_data_Latex, context_embeddings_Latex = load_book_data(base_path, model_name, latex_chunk_tokens, latex_overlap_tokens)
|
| 172 |
summary = load_summary('data/KG_FEM_summary.json')
|
| 173 |
|
| 174 |
-
|
| 175 |
-
|
| 176 |
-
|
| 177 |
if 'question_answered' not in st.session_state:
|
| 178 |
st.session_state.question_answered = False
|
| 179 |
if 'context_by_video' not in st.session_state:
|
|
@@ -268,6 +264,7 @@ if submit_button_placeholder.button("AI Answer", type="primary"):
|
|
| 268 |
temperature=integration_temperature,
|
| 269 |
top_p=integration_top_p
|
| 270 |
)
|
|
|
|
| 271 |
|
| 272 |
if answer.split()[0] == "NOT_ENOUGH_INFO":
|
| 273 |
st.markdown("")
|
|
@@ -297,7 +294,7 @@ if submit_button_placeholder.button("AI Answer", type="primary"):
|
|
| 297 |
if st.session_state.question_answered:
|
| 298 |
st.markdown("")
|
| 299 |
st.markdown("#### Query:")
|
| 300 |
-
st.markdown(
|
| 301 |
if show_expert_responce:
|
| 302 |
st.markdown("#### Initial Expert Answer:")
|
| 303 |
st.markdown(st.session_state.expert_answer)
|
|
|
|
| 6 |
|
| 7 |
# from openai import OpenAI
|
| 8 |
import random
|
|
|
|
| 9 |
|
| 10 |
from utils.help import get_disclaimer
|
| 11 |
from utils.format import sec_to_time, fix_latex, get_youtube_embed
|
|
|
|
| 170 |
text_data_Latex, context_embeddings_Latex = load_book_data(base_path, model_name, latex_chunk_tokens, latex_overlap_tokens)
|
| 171 |
summary = load_summary('data/KG_FEM_summary.json')
|
| 172 |
|
|
|
|
|
|
|
|
|
|
| 173 |
if 'question_answered' not in st.session_state:
|
| 174 |
st.session_state.question_answered = False
|
| 175 |
if 'context_by_video' not in st.session_state:
|
|
|
|
| 264 |
temperature=integration_temperature,
|
| 265 |
top_p=integration_top_p
|
| 266 |
)
|
| 267 |
+
answer = fix_latex(answer)
|
| 268 |
|
| 269 |
if answer.split()[0] == "NOT_ENOUGH_INFO":
|
| 270 |
st.markdown("")
|
|
|
|
| 294 |
if st.session_state.question_answered:
|
| 295 |
st.markdown("")
|
| 296 |
st.markdown("#### Query:")
|
| 297 |
+
st.markdown(fix_latex(st.session_state.question))
|
| 298 |
if show_expert_responce:
|
| 299 |
st.markdown("#### Initial Expert Answer:")
|
| 300 |
st.markdown(st.session_state.expert_answer)
|
utils/openai_utils.py
CHANGED
|
@@ -41,7 +41,7 @@ def openai_domain_specific_answer_generation(system_prompt, question, model="gpt
|
|
| 41 |
presence_penalty=0.0 # Prevent introduction of unrelated ideas.
|
| 42 |
)
|
| 43 |
|
| 44 |
-
return
|
| 45 |
|
| 46 |
def openai_context_integration(system_prompt, query, expert_answer, retrieved_context, model="gpt4o-mini", temperature=0.3, top_p=0.3):
|
| 47 |
|
|
@@ -76,4 +76,4 @@ def openai_context_integration(system_prompt, query, expert_answer, retrieved_co
|
|
| 76 |
presence_penalty=0.0 # Neutral to avoid introducing unrelated ideas.
|
| 77 |
)
|
| 78 |
|
| 79 |
-
return
|
|
|
|
| 41 |
presence_penalty=0.0 # Prevent introduction of unrelated ideas.
|
| 42 |
)
|
| 43 |
|
| 44 |
+
return response.choices[0].message.content
|
| 45 |
|
| 46 |
def openai_context_integration(system_prompt, query, expert_answer, retrieved_context, model="gpt4o-mini", temperature=0.3, top_p=0.3):
|
| 47 |
|
|
|
|
| 76 |
presence_penalty=0.0 # Neutral to avoid introducing unrelated ideas.
|
| 77 |
)
|
| 78 |
|
| 79 |
+
return response.choices[0].message.content
|