根据开发组建议重命名,移除多余的Prompt后缀
This commit is contained in:
@@ -68,7 +68,7 @@ class ExpressionLearner:
|
||||
# 学习用(开启行编号,便于溯源)
|
||||
random_msg_str: str = await build_anonymous_messages(random_msg, show_ids=True)
|
||||
|
||||
prompt_template = prompt_manager.get_prompt("learn_style_prompt")
|
||||
prompt_template = prompt_manager.get_prompt("learn_style")
|
||||
prompt_template.add_context("bot_name", global_config.bot.nickname)
|
||||
prompt_template.add_context("chat_str", random_msg_str)
|
||||
|
||||
|
||||
@@ -378,7 +378,7 @@ class ExpressionSelector:
|
||||
reply_reason_block = ""
|
||||
|
||||
# 3. 构建prompt(只包含情境,不包含完整的表达方式)
|
||||
prompt_template = prompt_manager.get_prompt("expression_evaluation_prompt")
|
||||
prompt_template = prompt_manager.get_prompt("expression_evaluation")
|
||||
prompt_template.add_context("bot_name", global_config.bot.nickname)
|
||||
prompt_template.add_context("chat_observe_info", chat_context)
|
||||
prompt_template.add_context("all_situations", all_situations_str)
|
||||
|
||||
@@ -200,7 +200,7 @@ class JargonExplainer:
|
||||
explanations_text = "\n".join(jargon_explanations)
|
||||
|
||||
# 使用LLM概括黑话解释
|
||||
prompt_of_summarize = prompt_manager.get_prompt("jargon_explainer_summarize_prompt")
|
||||
prompt_of_summarize = prompt_manager.get_prompt("jargon_explainer_summarize")
|
||||
prompt_of_summarize.add_context("chat_context", lambda _: chat_context)
|
||||
prompt_of_summarize.add_context("jargon_explanations", lambda _: explanations_text)
|
||||
summarize_prompt = await prompt_manager.render_prompt(prompt_of_summarize)
|
||||
|
||||
@@ -193,7 +193,7 @@ class JargonMiner:
|
||||
"- 请参考上一次推断的含义,结合新的上下文信息,给出更准确或更新的推断结果"
|
||||
)
|
||||
|
||||
prompt1_template = prompt_manager.get_prompt("jargon_inference_with_context_prompt")
|
||||
prompt1_template = prompt_manager.get_prompt("jargon_inference_with_context")
|
||||
prompt1_template.add_context("bot_name", global_config.bot.nickname)
|
||||
prompt1_template.add_context("content", str(content))
|
||||
prompt1_template.add_context("raw_content_list", raw_content_text)
|
||||
@@ -233,7 +233,7 @@ class JargonMiner:
|
||||
return
|
||||
|
||||
# 步骤2: 仅基于content推断
|
||||
prompt2_template = prompt_manager.get_prompt("jargon_inference_content_only_prompt")
|
||||
prompt2_template = prompt_manager.get_prompt("jargon_inference_content_only")
|
||||
prompt2_template.add_context("content", str(content))
|
||||
prompt2 = await prompt_manager.render_prompt(prompt2_template)
|
||||
|
||||
@@ -275,7 +275,7 @@ class JargonMiner:
|
||||
logger.debug(f"jargon {content} 推断1结果: {response1}")
|
||||
|
||||
# 步骤3: 比较两个推断结果
|
||||
prompt3_template = prompt_manager.get_prompt("jargon_compare_inference_prompt")
|
||||
prompt3_template = prompt_manager.get_prompt("jargon_compare_inference")
|
||||
prompt3_template.add_context("inference1", json.dumps(inference1, ensure_ascii=False))
|
||||
prompt3_template.add_context("inference2", json.dumps(inference2, ensure_ascii=False))
|
||||
prompt3 = await prompt_manager.render_prompt(prompt3_template)
|
||||
|
||||
@@ -72,7 +72,7 @@ class ReflectTracker:
|
||||
|
||||
# LLM Judge
|
||||
try:
|
||||
prompt_template = prompt_manager.get_prompt("reflect_judge_prompt")
|
||||
prompt_template = prompt_manager.get_prompt("reflect_judge")
|
||||
prompt_template.add_context("situation", str(self.expression.situation))
|
||||
prompt_template.add_context("style", str(self.expression.style))
|
||||
prompt_template.add_context("context_block", context_block)
|
||||
|
||||
Reference in New Issue
Block a user