feat:缓存调试信息开关
This commit is contained in:
@@ -56,7 +56,7 @@ BOT_CONFIG_PATH: Path = (CONFIG_DIR / "bot_config.toml").resolve().absolute()
|
||||
MODEL_CONFIG_PATH: Path = (CONFIG_DIR / "model_config.toml").resolve().absolute()
|
||||
LEGACY_ENV_PATH: Path = (PROJECT_ROOT / ".env").resolve().absolute()
|
||||
MMC_VERSION: str = "1.0.0"
|
||||
CONFIG_VERSION: str = "8.9.19"
|
||||
CONFIG_VERSION: str = "8.9.20"
|
||||
MODEL_CONFIG_VERSION: str = "1.14.3"
|
||||
|
||||
logger = get_logger("config")
|
||||
|
||||
@@ -1324,6 +1324,15 @@ class DebugConfig(ConfigBase):
|
||||
)
|
||||
"""是否记录 Replyer 请求体,默认关闭"""
|
||||
|
||||
enable_llm_cache_stats: bool = Field(
|
||||
default=False,
|
||||
json_schema_extra={
|
||||
"x-widget": "switch",
|
||||
"x-icon": "chart-no-axes-column",
|
||||
},
|
||||
)
|
||||
"""是否记录 LLM prompt cache 调试统计,默认关闭"""
|
||||
|
||||
|
||||
class ExtraPromptItem(ConfigBase):
|
||||
platform: str = Field(
|
||||
|
||||
@@ -182,6 +182,16 @@ class _LLMCacheStatsStore:
|
||||
_store = _LLMCacheStatsStore()
|
||||
|
||||
|
||||
def _is_llm_cache_stats_enabled() -> bool:
|
||||
"""读取调试配置,默认关闭 LLM prompt cache 统计。"""
|
||||
|
||||
try:
|
||||
from src.config.config import global_config
|
||||
return bool(global_config.debug.enable_llm_cache_stats)
|
||||
except Exception:
|
||||
return False
|
||||
|
||||
|
||||
def _normalize_request_type(request_type: str) -> str:
|
||||
normalized = str(request_type or "").strip()
|
||||
return normalized or "unknown"
|
||||
@@ -1313,6 +1323,9 @@ def record_llm_cache_usage(
|
||||
) -> None:
|
||||
"""Record one LLM prompt cache usage event."""
|
||||
|
||||
if not _is_llm_cache_stats_enabled():
|
||||
return
|
||||
|
||||
normalized_task_name = str(task_name or "").strip()
|
||||
if normalized_task_name not in FOCUSED_TASK_NAMES:
|
||||
return
|
||||
|
||||
Reference in New Issue
Block a user