后端:
1.阶段 6 CP4/CP5 目录收口与共享边界纯化
- 将 backend 根目录收口为 services、client、gateway、cmd、shared 五个一级目录
- 收拢 bootstrap、inits、infra/kafka、infra/outbox、conv、respond、pkg、middleware,移除根目录旧实现与空目录
- 将 utils 下沉到 services/userauth/internal/auth,将 logic 下沉到 services/schedule/core/planning
- 将迁移期 runtime 桥接实现统一收拢到 services/runtime/{conv,dao,eventsvc,model},删除 shared/legacy 与未再被 import 的旧 service 实现
- 将 gateway/shared/respond 收口为 HTTP/Gin 错误写回适配,shared/respond 仅保留共享错误语义与状态映射
- 将 HTTP IdempotencyMiddleware 与 RateLimitMiddleware 收口到 gateway/middleware
- 将 GormCachePlugin 下沉到 shared/infra/gormcache,将共享 RateLimiter 下沉到 shared/infra/ratelimit,将 agent token budget 下沉到 services/agent/shared
- 删除 InitEino 兼容壳,收缩 cmd/internal/coreinit 仅保留旧组合壳残留域初始化语义
- 更新微服务迁移计划与桌面 checklist,补齐 CP4/CP5 当前切流点、目录终态与验证结果
- 完成 go test ./...、git diff --check 与最终真实 smoke;health、register/login、task/create+get、schedule/today、task-class/list、memory/items、agent chat/meta/timeline/context-stats 全部 200,SSE 合并结果为 CP5_OK 且 [DONE] 只有 1 个
84 lines
2.3 KiB
Go
84 lines
2.3 KiB
Go
package kafka
|
||
|
||
import (
|
||
"context"
|
||
"errors"
|
||
"fmt"
|
||
"time"
|
||
|
||
segmentkafka "github.com/segmentio/kafka-go"
|
||
)
|
||
|
||
// WaitTopicReady 在指定超时时间内等待 Kafka topic 可用。
|
||
// 背景:初次部署时 broker 可能已启动,但 topic/partition 还没就绪。
|
||
// 这里启动前先探测,可减少“应用已启动但实际无法消费”的静默窗口。
|
||
func WaitTopicReady(parent context.Context, brokers []string, topic string, timeout time.Duration) error {
|
||
if len(brokers) == 0 {
|
||
return errors.New("kafka brokers is empty")
|
||
}
|
||
if topic == "" {
|
||
return errors.New("kafka topic is empty")
|
||
}
|
||
if timeout <= 0 {
|
||
timeout = 30 * time.Second
|
||
}
|
||
|
||
ctx, cancel := context.WithTimeout(parent, timeout)
|
||
defer cancel()
|
||
|
||
ticker := time.NewTicker(1 * time.Second)
|
||
defer ticker.Stop()
|
||
|
||
var lastErr error
|
||
for {
|
||
if err := probeTopic(ctx, brokers, topic); err == nil {
|
||
return nil
|
||
} else {
|
||
lastErr = err
|
||
}
|
||
|
||
select {
|
||
case <-ctx.Done():
|
||
if lastErr != nil {
|
||
return fmt.Errorf("wait topic ready timeout, topic=%s: %w", topic, lastErr)
|
||
}
|
||
return fmt.Errorf("wait topic ready timeout, topic=%s", topic)
|
||
case <-ticker.C:
|
||
}
|
||
}
|
||
}
|
||
|
||
// probeTopic 轮询所有 broker,只要任一 broker 能读到 topic 分区信息即视为就绪。
|
||
func probeTopic(ctx context.Context, brokers []string, topic string) error {
|
||
var lastErr error
|
||
for _, broker := range brokers {
|
||
conn, err := segmentkafka.DialContext(ctx, "tcp", broker)
|
||
if err != nil {
|
||
lastErr = err
|
||
continue
|
||
}
|
||
|
||
// 1. segmentio/kafka-go 的 ReadPartitions 不直接接收 context。
|
||
// 2. 这里必须给底层连接设置 I/O deadline,避免 broker 已接受连接但 metadata 响应卡住时,
|
||
// 上层 WaitTopicReady 永远阻塞,导致 outbox dispatch / consume 循环无法启动。
|
||
// 3. deadline 命中后本轮探测失败,外层 ticker 会继续重试直到总 timeout 到期。
|
||
_ = conn.SetDeadline(time.Now().Add(2 * time.Second))
|
||
partitions, readErr := conn.ReadPartitions(topic)
|
||
_ = conn.Close()
|
||
if readErr != nil {
|
||
lastErr = readErr
|
||
continue
|
||
}
|
||
if len(partitions) == 0 {
|
||
lastErr = fmt.Errorf("topic %s has no partitions yet", topic)
|
||
continue
|
||
}
|
||
return nil
|
||
}
|
||
|
||
if lastErr != nil {
|
||
return lastErr
|
||
}
|
||
return errors.New("unable to probe topic readiness")
|
||
}
|