From ce2281d31b87e59ba71cf49657dde616c5c1dd39 Mon Sep 17 00:00:00 2001 From: Ding <44717411+ding113@users.noreply.github.com> Date: Mon, 8 Sep 2025 10:29:12 +0800 Subject: [PATCH] Fix: Parameter Extractor Uses Correct Prompt for Prompt Mode in Chat Models (#24636) Co-authored-by: -LAN- --- .../nodes/parameter_extractor/parameter_extractor_node.py | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/api/core/workflow/nodes/parameter_extractor/parameter_extractor_node.py b/api/core/workflow/nodes/parameter_extractor/parameter_extractor_node.py index a854c7e87..1e1c10a11 100644 --- a/api/core/workflow/nodes/parameter_extractor/parameter_extractor_node.py +++ b/api/core/workflow/nodes/parameter_extractor/parameter_extractor_node.py @@ -52,6 +52,7 @@ from .exc import ( ) from .prompts import ( CHAT_EXAMPLE, + CHAT_GENERATE_JSON_PROMPT, CHAT_GENERATE_JSON_USER_MESSAGE_TEMPLATE, COMPLETION_GENERATE_JSON_PROMPT, FUNCTION_CALLING_EXTRACTOR_EXAMPLE, @@ -752,7 +753,7 @@ class ParameterExtractorNode(BaseNode): if model_mode == ModelMode.CHAT: system_prompt_messages = ChatModelMessage( role=PromptMessageRole.SYSTEM, - text=FUNCTION_CALLING_EXTRACTOR_SYSTEM_PROMPT.format(histories=memory_str, instruction=instruction), + text=CHAT_GENERATE_JSON_PROMPT.format(histories=memory_str).replace("{{instructions}}", instruction), ) user_prompt_message = ChatModelMessage(role=PromptMessageRole.USER, text=input_text) return [system_prompt_messages, user_prompt_message]