From a3e1a15ef6745f3a4ff66bdd3c59deccb65af24a Mon Sep 17 00:00:00 2001 From: stream Date: Tue, 8 Jul 2025 15:00:43 +0800 Subject: [PATCH] chore: fix ruff lint check errors --- api/controllers/console/app/workflow.py | 2 +- api/core/llm_generator/llm_generator.py | 4 ++-- api/core/llm_generator/prompts.py | 2 +- 3 files changed, 4 insertions(+), 4 deletions(-) diff --git a/api/controllers/console/app/workflow.py b/api/controllers/console/app/workflow.py index 22d2940e3f..fc91462e2b 100644 --- a/api/controllers/console/app/workflow.py +++ b/api/controllers/console/app/workflow.py @@ -27,7 +27,7 @@ from core.llm_generator.llm_generator import LLMGenerator from extensions.ext_database import db from factories import file_factory, variable_factory from fields.workflow_fields import workflow_fields, workflow_pagination_fields -from fields.workflow_run_fields import workflow_run_node_execution_fields, workflow_node_ai_modify_fields +from fields.workflow_run_fields import workflow_node_ai_modify_fields, workflow_run_node_execution_fields from libs import helper from libs.helper import TimestampField, uuid_value from libs.login import current_user, login_required diff --git a/api/core/llm_generator/llm_generator.py b/api/core/llm_generator/llm_generator.py index c39c3a6741..1e73500cef 100644 --- a/api/core/llm_generator/llm_generator.py +++ b/api/core/llm_generator/llm_generator.py @@ -4,7 +4,6 @@ import re from typing import Optional, cast import json_repair -from pydantic import BaseModel from core.llm_generator.output_parser.rule_config_generator import RuleConfigGeneratorOutputParser from core.llm_generator.output_parser.suggested_questions_after_answer import SuggestedQuestionsAfterAnswerOutputParser @@ -12,9 +11,10 @@ from core.llm_generator.prompts import ( CONVERSATION_TITLE_PROMPT, GENERATOR_QA_PROMPT, JAVASCRIPT_CODE_GENERATOR_PROMPT_TEMPLATE, + PROMPT_OPTIMIZATION_METAPROMPT_SYSTEM, PYTHON_CODE_GENERATOR_PROMPT_TEMPLATE, SYSTEM_STRUCTURED_OUTPUT_GENERATE, - WORKFLOW_RULE_CONFIG_PROMPT_GENERATE_TEMPLATE, PROMPT_OPTIMIZATION_METAPROMPT_SYSTEM, + WORKFLOW_RULE_CONFIG_PROMPT_GENERATE_TEMPLATE, ) from core.model_manager import ModelManager from core.model_runtime.entities.llm_entities import LLMResult diff --git a/api/core/llm_generator/prompts.py b/api/core/llm_generator/prompts.py index 595877e533..eb59ccabcb 100644 --- a/api/core/llm_generator/prompts.py +++ b/api/core/llm_generator/prompts.py @@ -361,4 +361,4 @@ Both your input and output should be in JSON format. ! Above is the schema for output content ! Your output must strictly follow the schema format, do not output any content outside of the JSON body. -""" +""" # noqa: E501