from utils.prompts import GENERATOR_PROMPT from utils.config import GENERATOR_MODEL from langchain_openai import ChatOpenAI from langchain_core.prompts import ChatPromptTemplate # Prompt prompt = ChatPromptTemplate.from_messages([("human", GENERATOR_PROMPT)]) prompt # Generator LLM llm = ChatOpenAI(model_name=GENERATOR_MODEL, temperature=0.7, streaming=True) # Post-processing def format_docs(docs): return "\n\n".join(doc.page_content for doc in docs) # Chain rag_chain = prompt | llm # generation = rag_chain.invoke({"context": docs, "question": question}) # print(generation)