后端: 1. AIHub 模型分级从 Worker/Strategist 两级重构为 Lite/Pro/Max 三级 - AIHub 结构体从 Worker + Strategist 改为 Lite + Pro + Max,分别对应轻量(标题生成)、标准(Chat 路由/闲聊/交付总结)、高能力(Plan 规划/Execute ReAct)三个能力层级 - config.example.yaml 新增 liteModel / proModel / maxModel 三个模型配置项,替代原 workerModel / strategistModel - 启动层 InitEino 改为创建三个独立模型实例,抽取公共 baseURL 和 apiKey 减少重复 - pickChatModel 统一返回 Pro 模型,旧 strategist 参数不再生效;pickTitleModel 从 Worker 切到 Lite - runNewAgentGraph 按 Plan/Execute→Max、Chat/Deliver→Pro 分级注入;Graph 出错回退也切到 Pro - Memory 模块初始化从 Worker 改为 Pro 2. Plan 节点从"两阶段评估"简化为"单轮深度规划",thinking 开关改为全配置化 - 移除 Phase 1(快速评估 1600 token)+ Phase 2(深度规划 3200 token)的两轮调用逻辑,改为单轮不限 token 深度规划 - PlanDecision 移除 need_thinking 字段,prompt 规则和 JSON contract 同步删除该字段 - 各节点(Plan / Execute / Deliver)thinking 开关从硬编码改为从 AgentGraphDeps 读取,由 config.yaml 的 agent.thinking 段按节点注入 - 新增 agent.thinking 配置段(plan / execute / deliver / memory 四个独立布尔开关),config.example.yaml 补齐默认值 - 新增 resolveThinkingMode 公共函数,plan / execute / deliver 和 memory 决策/抽取链路统一使用 3. Memory 模块 LLM 调用支持 thinking 开关 - Config 新增 LLMThinking 字段,config_loader 从 agent.thinking.memory 读取 - LLMDecisionOrchestrator.Compare 和 LLMWriteOrchestrator.ExtractFacts 的 thinking 模式从硬编码 Disabled 改为读取配置 前端: 1. 移除助手输入区模型选择器及全部偏好持久化逻辑 - 删除 ModelType 类型、selectedModel ref、MODEL_PREFERENCE_STORAGE_KEY 常量 - 删除 isModelType / loadModelPreferenceMap / persistModelPreferenceMap / savePreferredModel / resolvePreferredModel / applyPreferredModelForConversation 六个函数及 modelPreferenceMap ref - 删除 selectedModel watch 监听、发送消息时的 savePreferredModel 调用、切会话时的 applyPreferredModelForConversation 调用、会话迁移时的模型偏好迁移 - fetchChatStream 的 model 参数硬编码为 'worker' - 删除模板中"模型"下拉选择器(标准/策略)及对应的全局样式 .assistant-model-select-panel 2. 上下文窗口指示器简化为仅显示总占用 - ContextWindowMeter 移除 msg0~msg3 四段彩色分段逻辑(ContextSegment 接口、segments computed、v-for 渲染) - 进度条改为单一蓝色条,按 total/budget 比例填充;超预算时变红 - Tooltip 简化为仅显示"总计 X / 预算 Y(Z%)" 仓库:无
94 lines
3.6 KiB
Go
94 lines
3.6 KiB
Go
package service
|
|
|
|
import (
|
|
"time"
|
|
|
|
memorymodel "github.com/LoveLosita/smartflow/backend/memory/model"
|
|
"github.com/spf13/viper"
|
|
)
|
|
|
|
// LoadConfigFromViper 读取记忆模块配置并做默认值兜底。
|
|
//
|
|
// 默认策略:
|
|
// 1. temperature/top_p 使用低随机参数,提升可复现性;
|
|
// 2. Day1 先提供参数位,不强制所有参数立即生效;
|
|
// 3. 轮询与重试参数给出保守默认值,避免对主链路造成压力。
|
|
func LoadConfigFromViper() memorymodel.Config {
|
|
cfg := memorymodel.Config{
|
|
Enabled: viper.GetBool("memory.enabled"),
|
|
RAGEnabled: viper.GetBool("memory.rag.enabled"),
|
|
ReadMode: memorymodel.NormalizeReadMode(viper.GetString("memory.read.mode")),
|
|
InjectRenderMode: memorymodel.NormalizeInjectRenderMode(viper.GetString("memory.inject.renderMode")),
|
|
ExtractPrompt: viper.GetString("memory.prompt.extract"),
|
|
DecisionPrompt: viper.GetString("memory.prompt.decision"),
|
|
Threshold: viper.GetFloat64("memory.threshold"),
|
|
EnableReranker: viper.GetBool("memory.enableReranker"),
|
|
LLMTemperature: viper.GetFloat64("memory.llm.temperature"),
|
|
LLMTopP: viper.GetFloat64("memory.llm.topP"),
|
|
JobMaxRetry: viper.GetInt("memory.job.maxRetry"),
|
|
WorkerPollEvery: viper.GetDuration("memory.worker.pollEvery"),
|
|
WorkerClaimBatch: viper.GetInt("memory.worker.claimBatch"),
|
|
ReadConstraintLimit: viper.GetInt("memory.read.constraintLimit"),
|
|
ReadPreferenceLimit: viper.GetInt("memory.read.preferenceLimit"),
|
|
ReadFactLimit: viper.GetInt("memory.read.factLimit"),
|
|
ReadTodoHintLimit: viper.GetInt("memory.read.todoHintLimit"),
|
|
|
|
// 决策层配置:默认关闭,灰度开启后才会生效。
|
|
DecisionEnabled: viper.GetBool("memory.decision.enabled"),
|
|
DecisionCandidateTopK: viper.GetInt("memory.decision.candidateTopK"),
|
|
DecisionCandidateMinScore: viper.GetFloat64("memory.decision.candidateMinScore"),
|
|
DecisionFallbackMode: viper.GetString("memory.decision.fallbackMode"),
|
|
WriteMode: viper.GetString("memory.write.mode"),
|
|
WriteMinConfidence: viper.GetFloat64("memory.write.minConfidence"),
|
|
LLMThinking: viper.GetBool("agent.thinking.memory"),
|
|
}
|
|
|
|
if cfg.Threshold <= 0 {
|
|
cfg.Threshold = 0.55
|
|
}
|
|
if cfg.LLMTemperature <= 0 {
|
|
cfg.LLMTemperature = 0.1
|
|
}
|
|
if cfg.LLMTopP <= 0 {
|
|
cfg.LLMTopP = 0.2
|
|
}
|
|
if cfg.JobMaxRetry <= 0 {
|
|
cfg.JobMaxRetry = 6
|
|
}
|
|
if cfg.WorkerPollEvery <= 0 {
|
|
cfg.WorkerPollEvery = 2 * time.Second
|
|
}
|
|
if cfg.WorkerClaimBatch <= 0 {
|
|
cfg.WorkerClaimBatch = 1
|
|
}
|
|
cfg.ReadConstraintLimit = cfg.EffectiveReadConstraintLimit()
|
|
cfg.ReadPreferenceLimit = cfg.EffectiveReadPreferenceLimit()
|
|
cfg.ReadFactLimit = cfg.EffectiveReadFactLimit()
|
|
cfg.ReadTodoHintLimit = cfg.EffectiveReadTodoHintLimit()
|
|
cfg.ReadMode = cfg.EffectiveReadMode()
|
|
cfg.InjectRenderMode = cfg.EffectiveInjectRenderMode()
|
|
|
|
// 决策层配置默认值兜底。
|
|
// 说明:
|
|
// 1. TopK 和 MinScore 是 Milvus 召回参数,需要保守默认值避免召回过多噪声候选;
|
|
// 2. FallbackMode 默认退回旧路径新增,保证决策流程异常时不丢数据;
|
|
// 3. WriteMode 由 DecisionEnabled 隐式决定,这里不做强制联动。
|
|
if cfg.DecisionCandidateTopK <= 0 {
|
|
cfg.DecisionCandidateTopK = 5
|
|
}
|
|
if cfg.DecisionCandidateMinScore <= 0 {
|
|
cfg.DecisionCandidateMinScore = 0.6
|
|
}
|
|
if cfg.DecisionFallbackMode == "" {
|
|
cfg.DecisionFallbackMode = "legacy_add"
|
|
}
|
|
if cfg.WriteMode == "" {
|
|
cfg.WriteMode = "legacy"
|
|
}
|
|
if cfg.WriteMinConfidence <= 0 {
|
|
cfg.WriteMinConfidence = 0.5
|
|
}
|
|
|
|
return cfg
|
|
}
|