Unverified Commit 607f9eda authored by nan jiang's avatar nan jiang Committed by GitHub

Fix/app runner typo (#2661)

parent f25cec26
...@@ -84,7 +84,7 @@ class AppRunner: ...@@ -84,7 +84,7 @@ class AppRunner:
return rest_tokens return rest_tokens
def recale_llm_max_tokens(self, model_config: ModelConfigEntity, def recalc_llm_max_tokens(self, model_config: ModelConfigEntity,
prompt_messages: list[PromptMessage]): prompt_messages: list[PromptMessage]):
# recalc max_tokens if sum(prompt_token + max_tokens) over model token limit # recalc max_tokens if sum(prompt_token + max_tokens) over model token limit
model_type_instance = model_config.provider_model_bundle.model_type_instance model_type_instance = model_config.provider_model_bundle.model_type_instance
......
...@@ -181,7 +181,7 @@ class BasicApplicationRunner(AppRunner): ...@@ -181,7 +181,7 @@ class BasicApplicationRunner(AppRunner):
return return
# Re-calculate the max tokens if sum(prompt_token + max_tokens) over model token limit # Re-calculate the max tokens if sum(prompt_token + max_tokens) over model token limit
self.recale_llm_max_tokens( self.recalc_llm_max_tokens(
model_config=app_orchestration_config.model_config, model_config=app_orchestration_config.model_config,
prompt_messages=prompt_messages prompt_messages=prompt_messages
) )
......
...@@ -131,7 +131,7 @@ class AssistantCotApplicationRunner(BaseAssistantApplicationRunner): ...@@ -131,7 +131,7 @@ class AssistantCotApplicationRunner(BaseAssistantApplicationRunner):
) )
# recale llm max tokens # recale llm max tokens
self.recale_llm_max_tokens(self.model_config, prompt_messages) self.recalc_llm_max_tokens(self.model_config, prompt_messages)
# invoke model # invoke model
chunks: Generator[LLMResultChunk, None, None] = model_instance.invoke_llm( chunks: Generator[LLMResultChunk, None, None] = model_instance.invoke_llm(
prompt_messages=prompt_messages, prompt_messages=prompt_messages,
......
...@@ -106,7 +106,7 @@ class AssistantFunctionCallApplicationRunner(BaseAssistantApplicationRunner): ...@@ -106,7 +106,7 @@ class AssistantFunctionCallApplicationRunner(BaseAssistantApplicationRunner):
) )
# recale llm max tokens # recale llm max tokens
self.recale_llm_max_tokens(self.model_config, prompt_messages) self.recalc_llm_max_tokens(self.model_config, prompt_messages)
# invoke model # invoke model
chunks: Union[Generator[LLMResultChunk, None, None], LLMResult] = model_instance.invoke_llm( chunks: Union[Generator[LLMResultChunk, None, None], LLMResult] = model_instance.invoke_llm(
prompt_messages=prompt_messages, prompt_messages=prompt_messages,
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment