Spaces:
Running
Running
Disable max tokens for Gemini & Azure OpenAI
Browse files- helpers/llm_helper.py +2 -2
helpers/llm_helper.py
CHANGED
@@ -157,7 +157,7 @@ def get_langchain_llm(
|
|
157 |
return GoogleGenerativeAI(
|
158 |
model=model,
|
159 |
temperature=GlobalConfig.LLM_MODEL_TEMPERATURE,
|
160 |
-
max_tokens=max_new_tokens,
|
161 |
timeout=None,
|
162 |
max_retries=2,
|
163 |
google_api_key=api_key,
|
@@ -182,7 +182,7 @@ def get_langchain_llm(
|
|
182 |
api_version=azure_api_version,
|
183 |
azure_endpoint=azure_endpoint_url,
|
184 |
temperature=GlobalConfig.LLM_MODEL_TEMPERATURE,
|
185 |
-
max_tokens=max_new_tokens,
|
186 |
timeout=None,
|
187 |
max_retries=1,
|
188 |
api_key=api_key,
|
|
|
157 |
return GoogleGenerativeAI(
|
158 |
model=model,
|
159 |
temperature=GlobalConfig.LLM_MODEL_TEMPERATURE,
|
160 |
+
# max_tokens=max_new_tokens,
|
161 |
timeout=None,
|
162 |
max_retries=2,
|
163 |
google_api_key=api_key,
|
|
|
182 |
api_version=azure_api_version,
|
183 |
azure_endpoint=azure_endpoint_url,
|
184 |
temperature=GlobalConfig.LLM_MODEL_TEMPERATURE,
|
185 |
+
# max_tokens=max_new_tokens,
|
186 |
timeout=None,
|
187 |
max_retries=1,
|
188 |
api_key=api_key,
|