917 lines
38 KiB
Python
917 lines
38 KiB
Python
"""Maisaka 推理引擎。"""
|
||
|
||
from datetime import datetime
|
||
from typing import TYPE_CHECKING, Any, Optional, cast
|
||
|
||
import asyncio
|
||
import difflib
|
||
import json
|
||
import time
|
||
import traceback
|
||
|
||
from sqlmodel import col, select
|
||
|
||
from src.chat.heart_flow.heartFC_utils import CycleDetail
|
||
from src.chat.message_receive.message import SessionMessage
|
||
from src.chat.replyer.replyer_manager import replyer_manager
|
||
from src.chat.utils.utils import process_llm_response
|
||
from src.common.data_models.message_component_data_model import MessageSequence, TextComponent
|
||
from src.common.database.database import get_db_session
|
||
from src.common.database.database_model import PersonInfo
|
||
from src.common.logger import get_logger
|
||
from src.config.config import global_config
|
||
from src.know_u.knowledge_store import get_knowledge_store
|
||
from src.learners.jargon_explainer import search_jargon
|
||
from src.llm_models.exceptions import ReqAbortException
|
||
from src.llm_models.payload_content.tool_option import ToolCall
|
||
from src.services import database_service as database_api, send_service
|
||
|
||
from .context_messages import (
|
||
AssistantMessage,
|
||
LLMContextMessage,
|
||
SessionBackedMessage,
|
||
ToolResultMessage,
|
||
)
|
||
from .message_adapter import (
|
||
build_visible_text_from_sequence,
|
||
clone_message_sequence,
|
||
format_speaker_content,
|
||
)
|
||
from .tool_handlers import (
|
||
handle_mcp_tool,
|
||
handle_unknown_tool,
|
||
)
|
||
|
||
if TYPE_CHECKING:
|
||
from .runtime import MaisakaHeartFlowChatting
|
||
|
||
logger = get_logger("maisaka_reasoning_engine")
|
||
|
||
|
||
class MaisakaReasoningEngine:
|
||
"""负责内部思考、推理与工具执行。"""
|
||
|
||
def __init__(self, runtime: "MaisakaHeartFlowChatting") -> None:
|
||
self._runtime = runtime
|
||
self._last_reasoning_content: str = ""
|
||
|
||
async def run_loop(self) -> None:
|
||
"""独立消费消息批次,并执行对应的内部思考轮次。"""
|
||
try:
|
||
while self._runtime._running:
|
||
cached_messages = await self._runtime._internal_turn_queue.get()
|
||
timeout_triggered = cached_messages is None
|
||
if not timeout_triggered and not cached_messages:
|
||
self._runtime._internal_turn_queue.task_done()
|
||
continue
|
||
|
||
self._runtime._agent_state = self._runtime._STATE_RUNNING
|
||
if cached_messages:
|
||
self._append_wait_interrupted_message_if_needed()
|
||
await self._ingest_messages(cached_messages)
|
||
anchor_message = cached_messages[-1]
|
||
else:
|
||
anchor_message = self._get_timeout_anchor_message()
|
||
if anchor_message is None:
|
||
logger.warning(
|
||
f"{self._runtime.log_prefix} 等待超时后缺少可复用的锚点消息,跳过本轮继续思考"
|
||
)
|
||
self._runtime._internal_turn_queue.task_done()
|
||
continue
|
||
logger.info(f"{self._runtime.log_prefix} 等待超时后开始新一轮思考")
|
||
self._runtime._chat_history.append(self._build_wait_timeout_message())
|
||
self._trim_chat_history()
|
||
try:
|
||
for round_index in range(self._runtime._max_internal_rounds):
|
||
cycle_detail = self._start_cycle()
|
||
self._runtime._log_cycle_started(cycle_detail, round_index)
|
||
planner_started_at = time.time()
|
||
try:
|
||
logger.info(
|
||
f"{self._runtime.log_prefix} 规划器开始执行: "
|
||
f"回合={round_index + 1} "
|
||
f"历史消息数={len(self._runtime._chat_history)} "
|
||
f"开始时间={planner_started_at:.3f}"
|
||
)
|
||
interrupt_flag = asyncio.Event()
|
||
self._runtime._planner_interrupt_flag = interrupt_flag
|
||
self._runtime._chat_loop_service.set_interrupt_flag(interrupt_flag)
|
||
try:
|
||
response = await self._runtime._chat_loop_service.chat_loop_step(self._runtime._chat_history)
|
||
finally:
|
||
if self._runtime._planner_interrupt_flag is interrupt_flag:
|
||
self._runtime._planner_interrupt_flag = None
|
||
self._runtime._chat_loop_service.set_interrupt_flag(None)
|
||
cycle_detail.time_records["planner"] = time.time() - planner_started_at
|
||
logger.info(
|
||
f"{self._runtime.log_prefix} 规划器执行完成: "
|
||
f"回合={round_index + 1} "
|
||
f"耗时={cycle_detail.time_records['planner']:.3f} 秒"
|
||
)
|
||
|
||
reasoning_content = response.content or ""
|
||
if self._should_replace_reasoning(reasoning_content):
|
||
response.content = "让我根据新情况重新思考:"
|
||
response.raw_message.content = "让我根据新情况重新思考:"
|
||
logger.info(f"{self._runtime.log_prefix} 当前思考与上一轮过于相似,已替换为重新思考提示")
|
||
|
||
self._last_reasoning_content = reasoning_content
|
||
self._runtime._chat_history.append(response.raw_message)
|
||
|
||
if response.tool_calls:
|
||
tool_started_at = time.time()
|
||
should_pause = await self._handle_tool_calls(
|
||
response.tool_calls,
|
||
response.content or "",
|
||
anchor_message,
|
||
)
|
||
cycle_detail.time_records["tool_calls"] = time.time() - tool_started_at
|
||
if should_pause:
|
||
break
|
||
continue
|
||
|
||
if response.content:
|
||
continue
|
||
|
||
break
|
||
except ReqAbortException:
|
||
interrupted_at = time.time()
|
||
logger.info(
|
||
f"{self._runtime.log_prefix} 规划器打断成功: "
|
||
f"回合={round_index + 1} "
|
||
f"开始时间={planner_started_at:.3f} "
|
||
f"打断时间={interrupted_at:.3f} "
|
||
f"耗时={interrupted_at - planner_started_at:.3f} 秒"
|
||
)
|
||
break
|
||
finally:
|
||
self._end_cycle(cycle_detail)
|
||
finally:
|
||
if self._runtime._agent_state == self._runtime._STATE_RUNNING:
|
||
self._runtime._agent_state = self._runtime._STATE_STOP
|
||
self._runtime._internal_turn_queue.task_done()
|
||
except asyncio.CancelledError:
|
||
self._runtime._log_internal_loop_cancelled()
|
||
raise
|
||
except Exception:
|
||
logger.exception(f"{self._runtime.log_prefix} Maisaka 内部循环发生异常")
|
||
logger.error(traceback.format_exc())
|
||
raise
|
||
|
||
def _get_timeout_anchor_message(self) -> Optional[SessionMessage]:
|
||
"""在 wait 超时后复用最近一条真实用户消息作为锚点。"""
|
||
if self._runtime.message_cache:
|
||
return self._runtime.message_cache[-1]
|
||
return None
|
||
|
||
def _build_wait_timeout_message(self) -> ToolResultMessage:
|
||
"""构造 wait 超时后的工具结果消息。"""
|
||
tool_call_id = self._runtime._pending_wait_tool_call_id or "wait_timeout"
|
||
self._runtime._pending_wait_tool_call_id = None
|
||
return ToolResultMessage(
|
||
content="等待已超时,期间没有收到新的用户输入。请基于现有上下文继续下一轮思考。",
|
||
timestamp=datetime.now(),
|
||
tool_call_id=tool_call_id,
|
||
tool_name="wait",
|
||
)
|
||
|
||
def _append_wait_interrupted_message_if_needed(self) -> None:
|
||
"""如果 wait 被新消息打断,则补一条对应的工具结果消息。"""
|
||
tool_call_id = self._runtime._pending_wait_tool_call_id
|
||
if not tool_call_id:
|
||
return
|
||
|
||
self._runtime._pending_wait_tool_call_id = None
|
||
self._runtime._chat_history.append(
|
||
ToolResultMessage(
|
||
content="等待过程被新的用户输入打断,已继续处理最新消息。",
|
||
timestamp=datetime.now(),
|
||
tool_call_id=tool_call_id,
|
||
tool_name="wait",
|
||
)
|
||
)
|
||
|
||
async def _ingest_messages(self, messages: list[SessionMessage]) -> None:
|
||
"""处理传入消息列表,将其转换为历史消息并加入聊天历史缓存。"""
|
||
for message in messages:
|
||
# 构建用户消息序列
|
||
user_sequence, visible_text = await self._build_message_sequence(message)
|
||
if not user_sequence.components:
|
||
continue
|
||
|
||
history_message = SessionBackedMessage.from_session_message(
|
||
message,
|
||
raw_message=user_sequence,
|
||
visible_text=visible_text,
|
||
source_kind="user",
|
||
)
|
||
self._insert_chat_history_message(history_message)
|
||
self._trim_chat_history()
|
||
|
||
async def _build_message_sequence(self, message: SessionMessage) -> tuple[MessageSequence, str]:
|
||
message_sequence = MessageSequence([])
|
||
planner_prefix = self._build_planner_user_prefix(message)
|
||
|
||
appended_component = False
|
||
if global_config.maisaka.direct_image_input:
|
||
source_sequence = getattr(message, "maisaka_original_raw_message", message.raw_message)
|
||
else:
|
||
source_sequence = message.raw_message
|
||
|
||
planner_components = clone_message_sequence(source_sequence).components
|
||
if planner_components and isinstance(planner_components[0], TextComponent):
|
||
planner_components[0].text = planner_prefix + planner_components[0].text
|
||
else:
|
||
planner_components.insert(0, TextComponent(planner_prefix))
|
||
|
||
for component in planner_components:
|
||
message_sequence.components.append(component)
|
||
appended_component = True
|
||
|
||
legacy_visible_text = self._build_legacy_visible_text(message, source_sequence)
|
||
if not appended_component:
|
||
if not message.processed_plain_text:
|
||
await message.process()
|
||
content = (message.processed_plain_text or "").strip()
|
||
if content:
|
||
message_sequence.text(planner_prefix + content)
|
||
legacy_visible_text = self._build_legacy_visible_text_from_text(message, content)
|
||
|
||
return message_sequence, legacy_visible_text
|
||
|
||
@staticmethod
|
||
def _build_planner_user_prefix(message: SessionMessage) -> str:
|
||
user_info = message.message_info.user_info
|
||
timestamp_text = message.timestamp.strftime("%H:%M:%S")
|
||
user_name = user_info.user_nickname or user_info.user_id
|
||
group_card = user_info.user_cardname or ""
|
||
message_id = message.message_id or ""
|
||
return (
|
||
f"[时间]{timestamp_text}\n"
|
||
f"[用户]{user_name}\n"
|
||
f"[用户群昵称]{group_card}\n"
|
||
f"[msg_id]{message_id}\n"
|
||
"[发言内容]"
|
||
)
|
||
|
||
def _build_legacy_visible_text(self, message: SessionMessage, source_sequence: MessageSequence) -> str:
|
||
user_info = message.message_info.user_info
|
||
speaker_name = user_info.user_cardname or user_info.user_nickname or user_info.user_id
|
||
legacy_sequence = MessageSequence([])
|
||
legacy_sequence.text(format_speaker_content(speaker_name, "", message.timestamp, message.message_id))
|
||
for component in clone_message_sequence(source_sequence).components:
|
||
legacy_sequence.components.append(component)
|
||
return build_visible_text_from_sequence(legacy_sequence).strip()
|
||
|
||
def _build_legacy_visible_text_from_text(self, message: SessionMessage, content: str) -> str:
|
||
user_info = message.message_info.user_info
|
||
speaker_name = user_info.user_cardname or user_info.user_nickname or user_info.user_id
|
||
return format_speaker_content(speaker_name, content, message.timestamp, message.message_id).strip()
|
||
|
||
def _insert_chat_history_message(self, message: LLMContextMessage) -> int:
|
||
"""将消息按处理顺序追加到聊天历史末尾。"""
|
||
self._runtime._chat_history.append(message)
|
||
return len(self._runtime._chat_history) - 1
|
||
|
||
def _start_cycle(self) -> CycleDetail:
|
||
"""开始一轮 Maisaka 思考循环。"""
|
||
self._runtime._cycle_counter += 1
|
||
self._runtime._current_cycle_detail = CycleDetail(cycle_id=self._runtime._cycle_counter)
|
||
self._runtime._current_cycle_detail.thinking_id = f"maisaka_tid{round(time.time(), 2)}"
|
||
return self._runtime._current_cycle_detail
|
||
|
||
def _end_cycle(self, cycle_detail: CycleDetail, only_long_execution: bool = True) -> CycleDetail:
|
||
"""结束并记录一轮 Maisaka 思考循环。"""
|
||
cycle_detail.end_time = time.time()
|
||
self._runtime.history_loop.append(cycle_detail)
|
||
|
||
timer_strings = [
|
||
f"{name}: {duration:.2f}s"
|
||
for name, duration in cycle_detail.time_records.items()
|
||
if not only_long_execution or duration >= 0.1
|
||
]
|
||
self._runtime._log_cycle_completed(cycle_detail, timer_strings)
|
||
return cycle_detail
|
||
|
||
def _trim_chat_history(self) -> None:
|
||
"""裁剪聊天历史,保证用户消息数量不超过配置限制。"""
|
||
conversation_message_count = sum(1 for message in self._runtime._chat_history if message.count_in_context)
|
||
if conversation_message_count <= self._runtime._max_context_size:
|
||
return
|
||
|
||
trimmed_history = list(self._runtime._chat_history)
|
||
removed_count = 0
|
||
|
||
while conversation_message_count >= self._runtime._max_context_size and trimmed_history:
|
||
removed_message = trimmed_history.pop(0)
|
||
removed_count += 1
|
||
if removed_message.count_in_context:
|
||
conversation_message_count -= 1
|
||
|
||
self._runtime._chat_history = trimmed_history
|
||
self._runtime._log_history_trimmed(removed_count, conversation_message_count)
|
||
|
||
@staticmethod
|
||
def _calculate_similarity(text1: str, text2: str) -> float:
|
||
"""计算两个文本之间的相似度。
|
||
|
||
Args:
|
||
text1: 第一个文本
|
||
text2: 第二个文本
|
||
|
||
Returns:
|
||
float: 相似度值,范围 0-1,1 表示完全相同
|
||
"""
|
||
return difflib.SequenceMatcher(None, text1, text2).ratio()
|
||
|
||
def _should_replace_reasoning(self, current_content: str) -> bool:
|
||
"""判断是否需要替换推理内容。
|
||
|
||
当当前推理内容与上一次相似度大于90%时,返回True。
|
||
|
||
Args:
|
||
current_content: 当前的推理内容
|
||
|
||
Returns:
|
||
bool: 是否需要替换
|
||
"""
|
||
if not self._last_reasoning_content or not current_content:
|
||
logger.info(
|
||
f"{self._runtime.log_prefix} 跳过思考相似度判定: "
|
||
f"上一轮为空={not bool(self._last_reasoning_content)} "
|
||
f"当前为空={not bool(current_content)} 相似度=0.00"
|
||
)
|
||
return False
|
||
|
||
similarity = self._calculate_similarity(current_content, self._last_reasoning_content)
|
||
logger.info(f"{self._runtime.log_prefix} 思考内容相似度: {similarity:.2f}")
|
||
return similarity > 0.9
|
||
|
||
@staticmethod
|
||
def _post_process_reply_text(reply_text: str) -> list[str]:
|
||
"""沿用旧回复链的文本后处理,执行分段与错别字注入。"""
|
||
processed_segments: list[str] = []
|
||
for segment in process_llm_response(reply_text):
|
||
normalized_segment = segment.strip()
|
||
if normalized_segment:
|
||
processed_segments.append(normalized_segment)
|
||
|
||
if processed_segments:
|
||
return processed_segments
|
||
return [reply_text.strip()]
|
||
|
||
async def _handle_tool_calls(
|
||
self,
|
||
tool_calls: list[ToolCall],
|
||
latest_thought: str,
|
||
anchor_message: SessionMessage,
|
||
) -> bool:
|
||
for tool_call in tool_calls:
|
||
if tool_call.func_name == "reply":
|
||
reply_sent = await self._handle_reply(tool_call, latest_thought, anchor_message)
|
||
if not reply_sent:
|
||
logger.warning(
|
||
f"{self._runtime.log_prefix} 回复工具未生成可见消息,将继续下一轮循环"
|
||
)
|
||
continue
|
||
|
||
if tool_call.func_name == "no_reply":
|
||
self._runtime._chat_history.append(
|
||
self._build_tool_message(
|
||
tool_call,
|
||
"本轮未发送可见回复。",
|
||
)
|
||
)
|
||
continue
|
||
|
||
if tool_call.func_name == "query_jargon":
|
||
await self._handle_query_jargon(tool_call)
|
||
continue
|
||
|
||
if tool_call.func_name == "query_person_info":
|
||
await self._handle_query_person_info(tool_call)
|
||
continue
|
||
|
||
if tool_call.func_name == "wait":
|
||
seconds = (tool_call.args or {}).get("seconds", 30)
|
||
try:
|
||
wait_seconds = int(seconds)
|
||
except (TypeError, ValueError):
|
||
wait_seconds = 30
|
||
wait_seconds = max(0, wait_seconds)
|
||
self._runtime._enter_wait_state(seconds=wait_seconds, tool_call_id=tool_call.call_id)
|
||
return True
|
||
|
||
if tool_call.func_name == "stop":
|
||
self._runtime._chat_history.append(
|
||
self._build_tool_message(
|
||
tool_call,
|
||
"当前对话循环已暂停,等待新消息到来。",
|
||
)
|
||
)
|
||
self._runtime._enter_stop_state()
|
||
return True
|
||
|
||
if tool_call.func_name == "send_emoji":
|
||
await self._handle_send_emoji(tool_call, anchor_message)
|
||
continue
|
||
|
||
if self._runtime._mcp_manager and self._runtime._mcp_manager.is_mcp_tool(tool_call.func_name):
|
||
await handle_mcp_tool(tool_call, self._runtime._chat_history, self._runtime._mcp_manager)
|
||
continue
|
||
|
||
await handle_unknown_tool(tool_call, self._runtime._chat_history)
|
||
|
||
return False
|
||
|
||
async def _handle_query_jargon(self, tool_call: ToolCall) -> None:
|
||
tool_args = tool_call.args or {}
|
||
raw_words = tool_args.get("words")
|
||
|
||
if not isinstance(raw_words, list):
|
||
self._runtime._chat_history.append(
|
||
self._build_tool_message(tool_call, "查询黑话工具需要提供 `words` 数组参数。")
|
||
)
|
||
return
|
||
|
||
words: list[str] = []
|
||
seen_words: set[str] = set()
|
||
for item in raw_words:
|
||
if not isinstance(item, str):
|
||
continue
|
||
word = item.strip()
|
||
if not word or word in seen_words:
|
||
continue
|
||
seen_words.add(word)
|
||
words.append(word)
|
||
|
||
if not words:
|
||
self._runtime._chat_history.append(
|
||
self._build_tool_message(tool_call, "查询黑话工具至少需要一个非空词条。")
|
||
)
|
||
return
|
||
|
||
logger.info(f"{self._runtime.log_prefix} 已触发黑话查询: 词条={words!r}")
|
||
|
||
results: list[dict[str, object]] = []
|
||
for word in words:
|
||
exact_matches = search_jargon(
|
||
keyword=word,
|
||
chat_id=self._runtime.session_id,
|
||
limit=5,
|
||
case_sensitive=False,
|
||
fuzzy=False,
|
||
)
|
||
matched_entries = exact_matches or search_jargon(
|
||
keyword=word,
|
||
chat_id=self._runtime.session_id,
|
||
limit=5,
|
||
case_sensitive=False,
|
||
fuzzy=True,
|
||
)
|
||
|
||
results.append(
|
||
{
|
||
"word": word,
|
||
"found": bool(matched_entries),
|
||
"matches": matched_entries,
|
||
}
|
||
)
|
||
|
||
logger.info(f"{self._runtime.log_prefix} 黑话查询完成: 结果={results!r}")
|
||
self._runtime._chat_history.append(
|
||
self._build_tool_message(
|
||
tool_call,
|
||
json.dumps({"results": results}, ensure_ascii=False),
|
||
)
|
||
)
|
||
|
||
async def _handle_query_person_info(self, tool_call: ToolCall) -> None:
|
||
"""查询指定人物的档案和相关知识。"""
|
||
tool_args = tool_call.args or {}
|
||
raw_person_name = tool_args.get("person_name")
|
||
raw_limit = tool_args.get("limit", 3)
|
||
|
||
if not isinstance(raw_person_name, str):
|
||
self._runtime._chat_history.append(
|
||
self._build_tool_message(tool_call, "查询人物信息工具需要提供字符串类型的 `person_name` 参数。")
|
||
)
|
||
return
|
||
|
||
person_name = raw_person_name.strip()
|
||
if not person_name:
|
||
self._runtime._chat_history.append(
|
||
self._build_tool_message(tool_call, "查询人物信息工具需要提供非空的 `person_name` 参数。")
|
||
)
|
||
return
|
||
|
||
try:
|
||
limit = max(1, min(int(raw_limit), 10))
|
||
except (TypeError, ValueError):
|
||
limit = 3
|
||
|
||
logger.info(
|
||
f"{self._runtime.log_prefix} 已触发人物信息查询: "
|
||
f"人物名={person_name!r} 限制条数={limit}"
|
||
)
|
||
|
||
persons = self._query_person_records(person_name, limit)
|
||
result = {
|
||
"query": person_name,
|
||
"persons": persons,
|
||
"related_knowledge": self._query_related_knowledge(person_name, persons, limit),
|
||
}
|
||
|
||
logger.info(
|
||
f"{self._runtime.log_prefix} 人物信息查询完成: "
|
||
f"人物记录数={len(result['persons'])} 相关知识数={len(result['related_knowledge'])}"
|
||
)
|
||
self._runtime._chat_history.append(
|
||
self._build_tool_message(
|
||
tool_call,
|
||
json.dumps(result, ensure_ascii=False),
|
||
)
|
||
)
|
||
|
||
def _query_person_records(self, person_name: str, limit: int) -> list[dict[str, Any]]:
|
||
"""按名称、昵称或用户 ID 查询人物档案。"""
|
||
with get_db_session() as session:
|
||
records = session.exec(
|
||
select(PersonInfo)
|
||
.where(
|
||
col(PersonInfo.person_name).contains(person_name)
|
||
| col(PersonInfo.user_nickname).contains(person_name)
|
||
| col(PersonInfo.user_id).contains(person_name)
|
||
)
|
||
.order_by(col(PersonInfo.last_known_time).desc(), col(PersonInfo.id).desc())
|
||
.limit(limit)
|
||
).all()
|
||
|
||
persons: list[dict[str, Any]] = []
|
||
for record in records:
|
||
memory_points: list[str] = []
|
||
if record.memory_points:
|
||
try:
|
||
parsed_points = json.loads(record.memory_points)
|
||
if isinstance(parsed_points, list):
|
||
memory_points = [str(point).strip() for point in parsed_points if str(point).strip()]
|
||
except (json.JSONDecodeError, TypeError, ValueError):
|
||
memory_points = []
|
||
|
||
persons.append(
|
||
{
|
||
"person_id": record.person_id,
|
||
"person_name": record.person_name or "",
|
||
"user_nickname": record.user_nickname,
|
||
"user_id": record.user_id,
|
||
"platform": record.platform,
|
||
"name_reason": record.name_reason or "",
|
||
"is_known": record.is_known,
|
||
"know_counts": record.know_counts,
|
||
"memory_points": memory_points[:20],
|
||
"last_known_time": (
|
||
record.last_known_time.isoformat() if record.last_known_time is not None else None
|
||
),
|
||
}
|
||
)
|
||
|
||
return persons
|
||
|
||
def _query_related_knowledge(
|
||
self,
|
||
person_name: str,
|
||
persons: list[dict[str, Any]],
|
||
limit: int,
|
||
) -> list[dict[str, Any]]:
|
||
"""从 Maisaka knowledge 中补充检索与该人物相关的条目。"""
|
||
store = get_knowledge_store()
|
||
knowledge_items: list[dict[str, Any]] = []
|
||
seen_ids: set[str] = set()
|
||
|
||
for person in persons:
|
||
matched_items = store.get_knowledge_by_user(
|
||
platform=str(person.get("platform", "")).strip(),
|
||
user_id=str(person.get("user_id", "")).strip(),
|
||
user_nickname=str(person.get("user_nickname", "")).strip(),
|
||
person_name=str(person.get("person_name", "")).strip(),
|
||
limit=max(limit, 5),
|
||
)
|
||
for item in matched_items:
|
||
item_id = str(item.get("id", "")).strip()
|
||
if item_id and item_id in seen_ids:
|
||
continue
|
||
if item_id:
|
||
seen_ids.add(item_id)
|
||
knowledge_items.append(item)
|
||
|
||
if not knowledge_items:
|
||
fallback_items = store.search_knowledge(person_name, limit=max(limit, 5))
|
||
for item in fallback_items:
|
||
item_id = str(item.get("id", "")).strip()
|
||
if item_id and item_id in seen_ids:
|
||
continue
|
||
if item_id:
|
||
seen_ids.add(item_id)
|
||
knowledge_items.append(item)
|
||
|
||
results: list[dict[str, Any]] = []
|
||
for item in knowledge_items:
|
||
results.append(
|
||
{
|
||
"id": str(item.get("id", "")).strip(),
|
||
"category_id": str(item.get("category_id", "")).strip(),
|
||
"category_name": str(item.get("category_name", "")).strip(),
|
||
"content": str(item.get("content", "")).strip(),
|
||
"metadata": item.get("metadata", {}),
|
||
"created_at": item.get("created_at"),
|
||
}
|
||
)
|
||
return results
|
||
|
||
async def _handle_reply(
|
||
self,
|
||
tool_call: ToolCall,
|
||
latest_thought: str,
|
||
anchor_message: SessionMessage,
|
||
) -> bool:
|
||
tool_args = tool_call.args or {}
|
||
target_message_id = str(tool_args.get("msg_id") or "").strip()
|
||
quote_reply = bool(tool_args.get("quote", True))
|
||
raw_unknown_words = tool_args.get("unknown_words")
|
||
unknown_words = raw_unknown_words if isinstance(raw_unknown_words, list) else None
|
||
if not target_message_id:
|
||
self._runtime._chat_history.append(
|
||
self._build_tool_message(tool_call, "回复工具需要提供有效的 `msg_id` 参数。")
|
||
)
|
||
return False
|
||
|
||
target_message = self._runtime._source_messages_by_id.get(target_message_id)
|
||
if target_message is None:
|
||
self._runtime._chat_history.append(
|
||
self._build_tool_message(tool_call, f"未找到要回复的目标消息,msg_id={target_message_id}")
|
||
)
|
||
return False
|
||
|
||
logger.info(
|
||
f"{self._runtime.log_prefix} 已触发回复工具: "
|
||
f"目标消息编号={target_message_id} 引用回复={quote_reply} 最新思考={latest_thought!r}"
|
||
)
|
||
logger.info(f"{self._runtime.log_prefix} 正在获取 Maisaka 回复生成器")
|
||
try:
|
||
replyer = replyer_manager.get_replyer(
|
||
chat_stream=self._runtime.chat_stream,
|
||
request_type="maisaka_replyer",
|
||
replyer_type="maisaka",
|
||
)
|
||
except Exception:
|
||
logger.exception(
|
||
f"{self._runtime.log_prefix} 获取回复生成器时发生异常: "
|
||
f"目标消息编号={target_message_id}"
|
||
)
|
||
self._runtime._chat_history.append(
|
||
self._build_tool_message(tool_call, "获取 Maisaka 回复生成器时发生异常。")
|
||
)
|
||
return False
|
||
|
||
if replyer is None:
|
||
logger.error(f"{self._runtime.log_prefix} 获取 Maisaka 回复生成器失败")
|
||
self._runtime._chat_history.append(
|
||
self._build_tool_message(tool_call, "Maisaka 回复生成器当前不可用。")
|
||
)
|
||
return False
|
||
|
||
from src.chat.replyer.maisaka_generator import MaisakaReplyGenerator
|
||
|
||
replyer = cast(MaisakaReplyGenerator, replyer)
|
||
logger.info(f"{self._runtime.log_prefix} 已成功获取 Maisaka 回复生成器")
|
||
|
||
logger.info(f"{self._runtime.log_prefix} 正在调用回复生成接口: 目标消息编号={target_message_id}")
|
||
try:
|
||
success, reply_result = await replyer.generate_reply_with_context(
|
||
reply_reason=latest_thought,
|
||
stream_id=self._runtime.session_id,
|
||
reply_message=target_message,
|
||
chat_history=self._runtime._chat_history,
|
||
unknown_words=unknown_words,
|
||
log_reply=False,
|
||
)
|
||
except Exception as exc:
|
||
import traceback
|
||
logger.error(
|
||
f"{self._runtime.log_prefix} 回复生成器执行异常: 目标消息编号={target_message_id} "
|
||
f"异常类型={type(exc).__name__} 异常信息={str(exc)}\n{traceback.format_exc()}"
|
||
)
|
||
self._runtime._chat_history.append(
|
||
self._build_tool_message(tool_call, "生成可见回复时发生异常。")
|
||
)
|
||
return False
|
||
|
||
logger.info(
|
||
f"{self._runtime.log_prefix} 回复生成完成: "
|
||
f"成功={success} 回复文本={reply_result.completion.response_text!r} "
|
||
f"错误信息={reply_result.error_message!r}"
|
||
)
|
||
reply_text = reply_result.completion.response_text.strip() if success else ""
|
||
if not reply_text:
|
||
logger.warning(
|
||
f"{self._runtime.log_prefix} 回复生成器返回空文本: "
|
||
f"目标消息编号={target_message_id} 错误信息={reply_result.error_message!r}"
|
||
)
|
||
self._runtime._chat_history.append(
|
||
self._build_tool_message(tool_call, "生成可见回复失败。")
|
||
)
|
||
return False
|
||
|
||
reply_segments = self._post_process_reply_text(reply_text)
|
||
combined_reply_text = "".join(reply_segments)
|
||
logger.info(
|
||
f"{self._runtime.log_prefix} 回复后处理完成: "
|
||
f"目标消息编号={target_message_id} 分段数={len(reply_segments)} "
|
||
f"分段内容={reply_segments!r}"
|
||
)
|
||
|
||
logger.info(
|
||
f"{self._runtime.log_prefix} 正在发送引导回复: "
|
||
f"目标消息编号={target_message_id} 引用回复={quote_reply} 回复分段={reply_segments!r}"
|
||
)
|
||
try:
|
||
sent = False
|
||
for index, segment in enumerate(reply_segments):
|
||
sent = await send_service.text_to_stream(
|
||
text=segment,
|
||
stream_id=self._runtime.session_id,
|
||
set_reply=quote_reply if index == 0 else False,
|
||
reply_message=target_message if quote_reply and index == 0 else None,
|
||
selected_expressions=reply_result.selected_expression_ids or None,
|
||
typing=index > 0,
|
||
)
|
||
if not sent:
|
||
break
|
||
except Exception:
|
||
logger.exception(
|
||
f"{self._runtime.log_prefix} 发送文字消息时发生异常,目标消息编号={target_message_id}"
|
||
)
|
||
self._runtime._chat_history.append(
|
||
self._build_tool_message(tool_call, "发送可见回复时发生异常。")
|
||
)
|
||
return False
|
||
|
||
logger.info(
|
||
f"{self._runtime.log_prefix} 引导回复发送结果: "
|
||
f"目标消息编号={target_message_id} 发送成功={sent}"
|
||
)
|
||
tool_result = "可见回复已生成并发送。" if sent else "可见回复生成成功,但发送失败。"
|
||
self._runtime._chat_history.append(self._build_tool_message(tool_call, tool_result))
|
||
if not sent:
|
||
return False
|
||
|
||
target_user_info = target_message.message_info.user_info
|
||
target_user_name = (
|
||
target_user_info.user_cardname
|
||
or target_user_info.user_nickname
|
||
or target_user_info.user_id
|
||
)
|
||
if self._runtime.chat_stream is not None:
|
||
await database_api.store_tool_info(
|
||
chat_stream=self._runtime.chat_stream,
|
||
display_prompt=f"你对{target_user_name}进行了回复:{combined_reply_text}",
|
||
tool_data={
|
||
"msg_id": target_message_id,
|
||
"quote": quote_reply,
|
||
"reply_text": combined_reply_text,
|
||
"reply_segments": reply_segments,
|
||
},
|
||
tool_name="reply",
|
||
tool_reasoning=latest_thought,
|
||
)
|
||
|
||
bot_name = global_config.bot.nickname.strip() or "MaiSaka"
|
||
reply_timestamp = datetime.now()
|
||
planner_prefix = (
|
||
f"[时间]{reply_timestamp.strftime('%H:%M:%S')}\n"
|
||
f"[用户]{bot_name}\n"
|
||
"[用户群昵称]\n"
|
||
"[msg_id]\n"
|
||
"[发言内容]"
|
||
)
|
||
history_message = SessionBackedMessage(
|
||
raw_message=MessageSequence([TextComponent(f"{planner_prefix}{combined_reply_text}")]),
|
||
visible_text="",
|
||
timestamp=reply_timestamp,
|
||
source_kind="guided_reply",
|
||
)
|
||
visible_reply_text = format_speaker_content(
|
||
bot_name,
|
||
combined_reply_text,
|
||
reply_timestamp,
|
||
)
|
||
history_message.visible_text = visible_reply_text
|
||
self._runtime._chat_history.append(history_message)
|
||
return True
|
||
|
||
async def _handle_send_emoji(self, tool_call: ToolCall, anchor_message: SessionMessage) -> None:
|
||
"""处理发送表情包的工具调用。
|
||
|
||
Args:
|
||
tool_call: 工具调用对象
|
||
anchor_message: 锚点消息
|
||
"""
|
||
from src.chat.emoji_system.emoji_manager import emoji_manager
|
||
from src.common.utils.utils_image import ImageUtils
|
||
import random
|
||
|
||
tool_args = tool_call.args or {}
|
||
emotion = str(tool_args.get("emotion") or "").strip()
|
||
|
||
logger.info(f"{self._runtime.log_prefix} 已触发表情包发送工具: 情绪={emotion!r}")
|
||
|
||
# 获取表情包列表
|
||
if not emoji_manager.emojis:
|
||
self._runtime._chat_history.append(
|
||
self._build_tool_message(tool_call, "当前表情包库中没有可用表情。")
|
||
)
|
||
return
|
||
|
||
# 根据情感选择表情包
|
||
selected_emoji = None
|
||
if emotion:
|
||
# 尝试找到匹配情感的表情包
|
||
matching_emojis = [
|
||
emoji for emoji in emoji_manager.emojis
|
||
if emotion.lower() in (e.lower() for e in emoji.emotion)
|
||
]
|
||
if matching_emojis:
|
||
selected_emoji = random.choice(matching_emojis)
|
||
logger.info(
|
||
f"{self._runtime.log_prefix} 找到 {len(matching_emojis)} 个匹配情绪 {emotion!r} 的表情包,"
|
||
f"已选择:{selected_emoji.description}"
|
||
)
|
||
|
||
# 如果没有找到匹配的情感表情包,随机选择一个
|
||
if selected_emoji is None:
|
||
selected_emoji = random.choice(emoji_manager.emojis)
|
||
logger.info(
|
||
f"{self._runtime.log_prefix} 没有表情包匹配情绪 {emotion!r},"
|
||
f"已随机选择:{selected_emoji.description}"
|
||
)
|
||
|
||
# 更新表情包使用次数
|
||
emoji_manager.update_emoji_usage(selected_emoji)
|
||
|
||
# 获取表情包的 base64 数据
|
||
try:
|
||
emoji_base64 = ImageUtils.image_path_to_base64(str(selected_emoji.full_path))
|
||
if not emoji_base64:
|
||
raise ValueError("表情图片转换为 base64 失败")
|
||
except Exception as exc:
|
||
logger.error(
|
||
f"{self._runtime.log_prefix} 表情图片转换为 base64 失败: {exc}"
|
||
)
|
||
self._runtime._chat_history.append(
|
||
self._build_tool_message(tool_call, f"发送表情包失败:{exc}")
|
||
)
|
||
return
|
||
|
||
# 发送表情包
|
||
try:
|
||
sent = await send_service.emoji_to_stream(
|
||
emoji_base64=emoji_base64,
|
||
stream_id=self._runtime.session_id,
|
||
storage_message=True,
|
||
set_reply=False,
|
||
reply_message=None,
|
||
)
|
||
except Exception as exc:
|
||
logger.exception(
|
||
f"{self._runtime.log_prefix} 发送表情包时发生异常: {exc}"
|
||
)
|
||
self._runtime._chat_history.append(
|
||
self._build_tool_message(tool_call, f"发送表情包时发生异常:{exc}")
|
||
)
|
||
return
|
||
|
||
if sent:
|
||
logger.info(
|
||
f"{self._runtime.log_prefix} 表情包发送成功: "
|
||
f"描述={selected_emoji.description!r} 情绪标签={selected_emoji.emotion}"
|
||
)
|
||
self._runtime._chat_history.append(
|
||
self._build_tool_message(
|
||
tool_call,
|
||
f"已发送表情包:{selected_emoji.description}(情绪:{', '.join(selected_emoji.emotion)})"
|
||
)
|
||
)
|
||
else:
|
||
logger.warning(f"{self._runtime.log_prefix} 表情包发送失败")
|
||
self._runtime._chat_history.append(
|
||
self._build_tool_message(tool_call, "发送表情包失败。")
|
||
)
|
||
|
||
def _build_tool_message(self, tool_call: ToolCall, content: str) -> ToolResultMessage:
|
||
return ToolResultMessage(
|
||
content=content,
|
||
timestamp=datetime.now(),
|
||
tool_call_id=tool_call.call_id,
|
||
tool_name=tool_call.func_name,
|
||
)
|