fix: llm invoke condition

This commit is contained in:
Novice
2026-01-27 10:12:51 +08:00
parent 54fce5e903
commit 585e11a1fc

View File

@ -298,17 +298,17 @@ class LLMNode(Node[LLMNodeData]):
structured_output: LLMStructuredOutput | None = None
sandbox = self.graph_runtime_state.sandbox
if sandbox:
tool_dependencies = self._extract_tool_dependencies()
generator = self._invoke_llm_with_sandbox(
sandbox=sandbox,
model_instance=model_instance,
prompt_messages=prompt_messages,
stop=stop,
variable_pool=variable_pool,
tool_dependencies=tool_dependencies,
)
elif self.tool_call_enabled:
if self.tool_call_enabled:
if sandbox:
tool_dependencies = self._extract_tool_dependencies()
generator = self._invoke_llm_with_sandbox(
sandbox=sandbox,
model_instance=model_instance,
prompt_messages=prompt_messages,
stop=stop,
variable_pool=variable_pool,
tool_dependencies=tool_dependencies,
)
generator = self._invoke_llm_with_tools(
model_instance=model_instance,
prompt_messages=prompt_messages,