Commit 7b738e04 authored by takatost's avatar takatost

fix typo

parent 3f6c1724
...@@ -134,7 +134,7 @@ class CotAgentRunner(BaseAgentRunner): ...@@ -134,7 +134,7 @@ class CotAgentRunner(BaseAgentRunner):
input=query input=query
) )
# recalc llm max tokens # recale llm max tokens
self.recalc_llm_max_tokens(self.model_config, prompt_messages) self.recalc_llm_max_tokens(self.model_config, prompt_messages)
# invoke model # invoke model
chunks: Generator[LLMResultChunk, None, None] = model_instance.invoke_llm( chunks: Generator[LLMResultChunk, None, None] = model_instance.invoke_llm(
......
...@@ -107,7 +107,7 @@ class FunctionCallAgentRunner(BaseAgentRunner): ...@@ -107,7 +107,7 @@ class FunctionCallAgentRunner(BaseAgentRunner):
messages_ids=message_file_ids messages_ids=message_file_ids
) )
# recalc llm max tokens # recale llm max tokens
self.recalc_llm_max_tokens(self.model_config, prompt_messages) self.recalc_llm_max_tokens(self.model_config, prompt_messages)
# invoke model # invoke model
chunks: Union[Generator[LLMResultChunk, None, None], LLMResult] = model_instance.invoke_llm( chunks: Union[Generator[LLMResultChunk, None, None], LLMResult] = model_instance.invoke_llm(
......
...@@ -84,7 +84,7 @@ class AppRunner: ...@@ -84,7 +84,7 @@ class AppRunner:
return rest_tokens return rest_tokens
def recale_llm_max_tokens(self, model_config: ModelConfigWithCredentialsEntity, def recalc_llm_max_tokens(self, model_config: ModelConfigWithCredentialsEntity,
prompt_messages: list[PromptMessage]): prompt_messages: list[PromptMessage]):
# recalc max_tokens if sum(prompt_token + max_tokens) over model token limit # recalc max_tokens if sum(prompt_token + max_tokens) over model token limit
model_type_instance = model_config.provider_model_bundle.model_type_instance model_type_instance = model_config.provider_model_bundle.model_type_instance
......
...@@ -189,7 +189,7 @@ class ChatAppRunner(AppRunner): ...@@ -189,7 +189,7 @@ class ChatAppRunner(AppRunner):
return return
# Re-calculate the max tokens if sum(prompt_token + max_tokens) over model token limit # Re-calculate the max tokens if sum(prompt_token + max_tokens) over model token limit
self.recale_llm_max_tokens( self.recalc_llm_max_tokens(
model_config=application_generate_entity.model_config, model_config=application_generate_entity.model_config,
prompt_messages=prompt_messages prompt_messages=prompt_messages
) )
......
...@@ -149,7 +149,7 @@ class CompletionAppRunner(AppRunner): ...@@ -149,7 +149,7 @@ class CompletionAppRunner(AppRunner):
return return
# Re-calculate the max tokens if sum(prompt_token + max_tokens) over model token limit # Re-calculate the max tokens if sum(prompt_token + max_tokens) over model token limit
self.recale_llm_max_tokens( self.recalc_llm_max_tokens(
model_config=application_generate_entity.model_config, model_config=application_generate_entity.model_config,
prompt_messages=prompt_messages prompt_messages=prompt_messages
) )
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment