feat(sync): implement Outbox Pattern for reward-service to wallet-service sync

Add event synchronization infrastructure between reward-service and wallet-service:

reward-service changes:
- Add OutboxEvent model to prisma schema for reliable event publishing
- Add outbox.repository.ts for outbox table CRUD operations
- Add outbox-publisher.service.ts for polling and publishing events to Kafka
- Add event-ack.controller.ts to receive consumer confirmations

wallet-service changes:
- Add ProcessedEvent model to prisma schema for idempotency checking
- Add reward-event-consumer.controller.ts to consume reward.summary.updated events
- Add event-ack.publisher.ts to send ACK to reward-service
- Update kafka.module.ts with Kafka client configuration
- Update main.ts to connect Kafka microservice on startup

Event flow: reward-service -> Kafka (reward.summary.updated) -> wallet-service -> Kafka (reward.events.ack) -> reward-service

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com>
This commit is contained in:
hailin 2025-12-10 23:26:49 -08:00
parent e2b3f25dbc
commit 538aae4ef0
13 changed files with 1084 additions and 8 deletions

View File

@ -0,0 +1,31 @@
-- CreateTable
CREATE TABLE "outbox_events" (
"outbox_id" BIGSERIAL NOT NULL,
"event_type" VARCHAR(100) NOT NULL,
"topic" VARCHAR(100) NOT NULL,
"key" VARCHAR(200) NOT NULL,
"payload" JSONB NOT NULL,
"aggregate_id" VARCHAR(100) NOT NULL,
"aggregate_type" VARCHAR(50) NOT NULL,
"status" VARCHAR(20) NOT NULL DEFAULT 'PENDING',
"retry_count" INTEGER NOT NULL DEFAULT 0,
"max_retries" INTEGER NOT NULL DEFAULT 5,
"last_error" TEXT,
"created_at" TIMESTAMP(3) NOT NULL DEFAULT CURRENT_TIMESTAMP,
"published_at" TIMESTAMP(3),
"next_retry_at" TIMESTAMP(3),
CONSTRAINT "outbox_events_pkey" PRIMARY KEY ("outbox_id")
);
-- CreateIndex
CREATE INDEX "idx_outbox_status_created" ON "outbox_events"("status", "created_at");
-- CreateIndex
CREATE INDEX "idx_outbox_status_retry" ON "outbox_events"("status", "next_retry_at");
-- CreateIndex
CREATE INDEX "idx_outbox_aggregate" ON "outbox_events"("aggregate_type", "aggregate_id");
-- CreateIndex
CREATE INDEX "idx_outbox_topic" ON "outbox_events"("topic");

View File

@ -181,3 +181,41 @@ model RewardEvent {
@@index([userId], name: "idx_reward_event_user")
@@index([occurredAt], name: "idx_reward_event_occurred")
}
// ============================================
// Outbox 事件发件箱表 (Outbox Pattern)
// 保证事件发布的可靠性:
// 1. 业务数据和 Outbox 记录在同一个事务中写入
// 2. 后台任务轮询 Outbox 表并发布到 Kafka
// 3. 消费方确认后标记为 CONFIRMED
// ============================================
model OutboxEvent {
id BigInt @id @default(autoincrement()) @map("outbox_id")
// 事件信息
eventType String @map("event_type") @db.VarChar(100)
topic String @map("topic") @db.VarChar(100)
key String @map("key") @db.VarChar(200)
payload Json @map("payload")
// 聚合根信息 (用于幂等性检查)
aggregateId String @map("aggregate_id") @db.VarChar(100)
aggregateType String @map("aggregate_type") @db.VarChar(50)
// 发布状态: PENDING -> SENT -> CONFIRMED 或 FAILED
status String @default("PENDING") @map("status") @db.VarChar(20)
retryCount Int @default(0) @map("retry_count")
maxRetries Int @default(5) @map("max_retries")
lastError String? @map("last_error") @db.Text
// 时间戳
createdAt DateTime @default(now()) @map("created_at")
publishedAt DateTime? @map("published_at")
nextRetryAt DateTime? @map("next_retry_at")
@@map("outbox_events")
@@index([status, createdAt], name: "idx_outbox_status_created")
@@index([status, nextRetryAt], name: "idx_outbox_status_retry")
@@index([aggregateType, aggregateId], name: "idx_outbox_aggregate")
@@index([topic], name: "idx_outbox_topic")
}

View File

@ -3,6 +3,7 @@ import { ConfigModule } from '@nestjs/config';
import { PrismaService } from './persistence/prisma/prisma.service';
import { RewardLedgerEntryRepositoryImpl } from './persistence/repositories/reward-ledger-entry.repository.impl';
import { RewardSummaryRepositoryImpl } from './persistence/repositories/reward-summary.repository.impl';
import { OutboxRepository } from './persistence/repositories/outbox.repository';
import { ReferralServiceClient } from './external/referral-service/referral-service.client';
import { AuthorizationServiceClient } from './external/authorization-service/authorization-service.client';
import { WalletServiceClient } from './external/wallet-service/wallet-service.client';
@ -16,6 +17,7 @@ import { REFERRAL_SERVICE_CLIENT, AUTHORIZATION_SERVICE_CLIENT } from '../domain
imports: [ConfigModule, KafkaModule, RedisModule],
providers: [
PrismaService,
OutboxRepository,
{
provide: REWARD_LEDGER_ENTRY_REPOSITORY,
useClass: RewardLedgerEntryRepositoryImpl,
@ -40,6 +42,7 @@ import { REFERRAL_SERVICE_CLIENT, AUTHORIZATION_SERVICE_CLIENT } from '../domain
],
exports: [
PrismaService,
OutboxRepository,
REWARD_LEDGER_ENTRY_REPOSITORY,
REWARD_SUMMARY_REPOSITORY,
REFERRAL_SERVICE_CLIENT,

View File

@ -0,0 +1,84 @@
import { Controller, Logger } from '@nestjs/common';
import { MessagePattern, Payload, Ctx, KafkaContext } from '@nestjs/microservices';
import { OutboxRepository } from '../persistence/repositories/outbox.repository';
/**
*
*/
interface EventAckMessage {
/** 原始事件的 aggregateId */
eventId: string;
/** 原始事件类型 */
eventType: string;
/** 消费服务名称 */
consumerService: string;
/** 处理结果 */
success: boolean;
/** 错误信息(如果失败) */
error?: string;
/** 确认时间 */
confirmedAt: string;
}
/**
* Kafka
*
* B方案核心组件(wallet-service)
* 使 @MessagePattern Kafka
*/
@Controller()
export class EventAckController {
private readonly logger = new Logger(EventAckController.name);
constructor(private readonly outboxRepository: OutboxRepository) {}
/**
*
*
* (wallet-service) topic
*/
@MessagePattern('reward.events.ack')
async handleEventAck(
@Payload() message: EventAckMessage,
@Ctx() context: KafkaContext,
): Promise<void> {
const partition = context.getPartition();
const offset = context.getMessage().offset;
this.logger.debug(
`[ACK] Received ack from ${message.consumerService} for event ${message.eventId} ` +
`[partition=${partition}, offset=${offset}]`,
);
try {
if (message.success) {
// 标记事件为已确认(使用 eventId + eventType 精确匹配)
const confirmed = await this.outboxRepository.markAsConfirmed(
message.eventId,
message.eventType,
);
if (confirmed) {
this.logger.log(
`[ACK] ✓ Event ${message.eventId} (${message.eventType}) confirmed by ${message.consumerService}`,
);
} else {
this.logger.warn(
`[ACK] Event ${message.eventId} (${message.eventType}) not found or already confirmed`,
);
}
} else {
// 消费方处理失败
this.logger.warn(
`[ACK] ✗ Event ${message.eventId} failed in ${message.consumerService}: ${message.error}`,
);
// 不改变状态,等待超时重发
}
} catch (error) {
this.logger.error(
`[ACK] Error processing ack for event ${message.eventId}:`,
error,
);
}
}
}

View File

@ -4,6 +4,10 @@ import { ClientsModule, Transport } from '@nestjs/microservices';
import { EventPublisherService } from './event-publisher.service';
import { EventConsumerController } from './event-consumer.controller';
import { EventAckPublisher } from './event-ack.publisher';
import { EventAckController } from './event-ack.controller';
import { OutboxPublisherService } from './outbox-publisher.service';
import { OutboxRepository } from '../persistence/repositories/outbox.repository';
import { PrismaService } from '../persistence/prisma/prisma.service';
import { ApplicationModule } from '../../application/application.module';
@Module({
@ -29,8 +33,8 @@ import { ApplicationModule } from '../../application/application.module';
},
]),
],
controllers: [EventConsumerController],
providers: [EventPublisherService, EventAckPublisher],
exports: [EventPublisherService, EventAckPublisher, ClientsModule],
controllers: [EventConsumerController, EventAckController],
providers: [PrismaService, OutboxRepository, EventPublisherService, EventAckPublisher, OutboxPublisherService],
exports: [EventPublisherService, EventAckPublisher, OutboxPublisherService, ClientsModule],
})
export class KafkaModule {}

View File

@ -0,0 +1,280 @@
import { Injectable, Logger, OnModuleInit, OnModuleDestroy, Inject } from '@nestjs/common';
import { ConfigService } from '@nestjs/config';
import { ClientKafka } from '@nestjs/microservices';
import { OutboxRepository, OutboxEvent, OutboxStatus } from '../persistence/repositories/outbox.repository';
/**
* Outbox Publisher Service (B方案 - )
*
* Outbox Kafka
* 使100%
*
* :
* 1. PENDING
* 2. Kafka SENT
* 3. (wallet-service) reward.events.ack
* 4. CONFIRMED
* 5. PENDING
*/
@Injectable()
export class OutboxPublisherService implements OnModuleInit, OnModuleDestroy {
private readonly logger = new Logger(OutboxPublisherService.name);
private isRunning = false;
private pollInterval: NodeJS.Timeout | null = null;
private timeoutCheckInterval: NodeJS.Timeout | null = null;
private cleanupInterval: NodeJS.Timeout | null = null;
private isConnected = false;
// 配置
private readonly pollIntervalMs: number;
private readonly batchSize: number;
private readonly cleanupIntervalMs: number;
private readonly confirmationTimeoutMinutes: number;
private readonly timeoutCheckIntervalMs: number;
constructor(
@Inject('KAFKA_SERVICE')
private readonly kafkaClient: ClientKafka,
private readonly outboxRepository: OutboxRepository,
private readonly configService: ConfigService,
) {
this.pollIntervalMs = this.configService.get<number>('OUTBOX_POLL_INTERVAL_MS', 1000);
this.batchSize = this.configService.get<number>('OUTBOX_BATCH_SIZE', 100);
this.cleanupIntervalMs = this.configService.get<number>('OUTBOX_CLEANUP_INTERVAL_MS', 3600000); // 1小时
this.confirmationTimeoutMinutes = this.configService.get<number>('OUTBOX_CONFIRMATION_TIMEOUT_MINUTES', 5);
this.timeoutCheckIntervalMs = this.configService.get<number>('OUTBOX_TIMEOUT_CHECK_INTERVAL_MS', 60000); // 1分钟
this.logger.log(
`[OUTBOX] OutboxPublisher (B方案) configured: ` +
`pollInterval=${this.pollIntervalMs}ms, batchSize=${this.batchSize}, ` +
`confirmationTimeout=${this.confirmationTimeoutMinutes}min`,
);
}
async onModuleInit() {
this.logger.log('[OUTBOX] Connecting to Kafka...');
try {
await this.kafkaClient.connect();
this.isConnected = true;
this.logger.log('[OUTBOX] Connected to Kafka');
this.start();
} catch (error) {
this.logger.error('[OUTBOX] Failed to connect to Kafka:', error);
this.logger.warn('[OUTBOX] OutboxPublisher will not start - events will accumulate in outbox table');
}
}
async onModuleDestroy() {
this.stop();
if (this.isConnected) {
await this.kafkaClient.close();
}
}
/**
*
*/
start(): void {
if (this.isRunning) {
this.logger.warn('[OUTBOX] Publisher already running');
return;
}
this.isRunning = true;
this.logger.log('[OUTBOX] Starting outbox publisher (B方案)...');
// 启动发布轮询
this.pollInterval = setInterval(() => {
this.processOutbox().catch((err) => {
this.logger.error('[OUTBOX] Error processing outbox:', err);
});
}, this.pollIntervalMs);
// 启动超时检查任务B方案核心
this.timeoutCheckInterval = setInterval(() => {
this.checkConfirmationTimeouts().catch((err) => {
this.logger.error('[OUTBOX] Error checking confirmation timeouts:', err);
});
}, this.timeoutCheckIntervalMs);
// 启动清理任务
this.cleanupInterval = setInterval(() => {
this.cleanup().catch((err) => {
this.logger.error('[OUTBOX] Error cleaning up outbox:', err);
});
}, this.cleanupIntervalMs);
this.logger.log('[OUTBOX] Outbox publisher started (B方案 - 消费方确认模式)');
}
/**
*
*/
stop(): void {
if (!this.isRunning) return;
this.isRunning = false;
if (this.pollInterval) {
clearInterval(this.pollInterval);
this.pollInterval = null;
}
if (this.timeoutCheckInterval) {
clearInterval(this.timeoutCheckInterval);
this.timeoutCheckInterval = null;
}
if (this.cleanupInterval) {
clearInterval(this.cleanupInterval);
this.cleanupInterval = null;
}
this.logger.log('[OUTBOX] Outbox publisher stopped');
}
/**
* Outbox
*/
async processOutbox(): Promise<void> {
if (!this.isConnected) {
return;
}
try {
// 1. 获取待发布事件
const pendingEvents = await this.outboxRepository.findPendingEvents(this.batchSize);
// 2. 获取需要重试的事件
const retryEvents = await this.outboxRepository.findEventsForRetry(Math.floor(this.batchSize / 2));
const allEvents = [...pendingEvents, ...retryEvents];
if (allEvents.length === 0) {
return;
}
this.logger.debug(`[OUTBOX] Processing ${allEvents.length} events (${pendingEvents.length} pending, ${retryEvents.length} retry)`);
// 3. 逐个发布
for (const event of allEvents) {
await this.publishEvent(event);
}
} catch (error) {
this.logger.error('[OUTBOX] Error in processOutbox:', error);
}
}
/**
* (B方案)
*
* 使 emit() Kafka SENT
* CONFIRMED
*/
private async publishEvent(event: OutboxEvent): Promise<void> {
try {
this.logger.debug(`[OUTBOX] Publishing event ${event.id} to topic ${event.topic}`);
// 构造 Kafka 消息,包含 outboxId 用于确认
const payload = {
...(event.payload as Record<string, unknown>),
_outbox: {
id: event.id.toString(),
aggregateId: event.aggregateId,
eventType: event.eventType,
},
};
const message = {
key: event.key,
value: JSON.stringify(payload),
};
// 发布到 Kafka
this.kafkaClient.emit(event.topic, message);
// B方案标记为 SENT等待消费方确认
await this.outboxRepository.markAsSent(event.id);
this.logger.log(
`[OUTBOX] → Event ${event.id} sent to ${event.topic} (awaiting consumer confirmation)`,
);
} catch (error) {
const errorMessage = error instanceof Error ? error.message : String(error);
this.logger.error(`[OUTBOX] ✗ Failed to publish event ${event.id}: ${errorMessage}`);
// 标记为失败并安排重试
await this.outboxRepository.markAsFailed(event.id, errorMessage);
}
}
/**
* (B方案核心)
*
* SENT PENDING 便
*/
private async checkConfirmationTimeouts(): Promise<void> {
if (!this.isConnected) {
return;
}
try {
const timedOutEvents = await this.outboxRepository.findSentEventsTimedOut(
this.confirmationTimeoutMinutes,
this.batchSize,
);
if (timedOutEvents.length === 0) {
return;
}
this.logger.warn(
`[OUTBOX] Found ${timedOutEvents.length} events without confirmation after ${this.confirmationTimeoutMinutes} minutes`,
);
for (const event of timedOutEvents) {
await this.outboxRepository.resetSentToPending(event.id);
this.logger.warn(
`[OUTBOX] Event ${event.id} reset to PENDING for retry (retry #${event.retryCount + 1})`,
);
}
} catch (error) {
this.logger.error('[OUTBOX] Error checking confirmation timeouts:', error);
}
}
/**
*
*/
private async cleanup(): Promise<void> {
const retentionDays = this.configService.get<number>('OUTBOX_RETENTION_DAYS', 7);
await this.outboxRepository.cleanupOldEvents(retentionDays);
}
/**
*
*/
async triggerProcessing(): Promise<void> {
this.logger.log('[OUTBOX] Manual processing triggered');
await this.processOutbox();
}
/**
*
*/
async getStats(): Promise<{
isRunning: boolean;
isConnected: boolean;
pending: number;
sent: number;
confirmed: number;
failed: number;
}> {
const stats = await this.outboxRepository.getStats();
return {
isRunning: this.isRunning,
isConnected: this.isConnected,
...stats,
};
}
}

View File

@ -0,0 +1,316 @@
import { Injectable, Logger } from '@nestjs/common';
import { PrismaService } from '../prisma/prisma.service';
import { Prisma } from '@prisma/client';
export enum OutboxStatus {
PENDING = 'PENDING', // 待发送
SENT = 'SENT', // 已发送到 Kafka等待消费方确认
CONFIRMED = 'CONFIRMED', // 消费方已确认处理成功
FAILED = 'FAILED', // 发送失败,等待重试
}
export interface OutboxEventData {
eventType: string;
topic: string;
key: string;
payload: Record<string, unknown>;
aggregateId: string;
aggregateType: string;
}
export interface OutboxEvent extends OutboxEventData {
id: bigint;
status: OutboxStatus;
retryCount: number;
maxRetries: number;
lastError: string | null;
createdAt: Date;
publishedAt: Date | null;
nextRetryAt: Date | null;
}
@Injectable()
export class OutboxRepository {
private readonly logger = new Logger(OutboxRepository.name);
constructor(private readonly prisma: PrismaService) {}
/**
* Outbox
*/
async saveInTransaction(
tx: Prisma.TransactionClient,
events: OutboxEventData[],
): Promise<void> {
if (events.length === 0) return;
this.logger.debug(`[OUTBOX] Saving ${events.length} events to outbox (in transaction)`);
await tx.outboxEvent.createMany({
data: events.map((event) => ({
eventType: event.eventType,
topic: event.topic,
key: event.key,
payload: event.payload as Prisma.JsonObject,
aggregateId: event.aggregateId,
aggregateType: event.aggregateType,
status: OutboxStatus.PENDING,
})),
});
this.logger.debug(`[OUTBOX] Saved ${events.length} events to outbox`);
}
/**
* Outbox
*/
async saveEvents(events: OutboxEventData[]): Promise<void> {
if (events.length === 0) return;
this.logger.debug(`[OUTBOX] Saving ${events.length} events to outbox`);
await this.prisma.outboxEvent.createMany({
data: events.map((event) => ({
eventType: event.eventType,
topic: event.topic,
key: event.key,
payload: event.payload as Prisma.JsonObject,
aggregateId: event.aggregateId,
aggregateType: event.aggregateType,
status: OutboxStatus.PENDING,
})),
});
this.logger.log(`[OUTBOX] ✓ Saved ${events.length} events to outbox`);
}
/**
*
*/
async findPendingEvents(limit: number = 100): Promise<OutboxEvent[]> {
const events = await this.prisma.outboxEvent.findMany({
where: {
status: OutboxStatus.PENDING,
},
orderBy: {
createdAt: 'asc',
},
take: limit,
});
return events.map((e) => this.mapToOutboxEvent(e));
}
/**
*
*/
async findEventsForRetry(limit: number = 50): Promise<OutboxEvent[]> {
const now = new Date();
const events = await this.prisma.outboxEvent.findMany({
where: {
status: OutboxStatus.FAILED,
retryCount: {
lt: 5,
},
nextRetryAt: {
lte: now,
},
},
orderBy: {
nextRetryAt: 'asc',
},
take: limit,
});
return events.map((e) => this.mapToOutboxEvent(e));
}
/**
*
*/
async markAsSent(id: bigint): Promise<void> {
await this.prisma.outboxEvent.update({
where: { id },
data: {
status: OutboxStatus.SENT,
publishedAt: new Date(),
},
});
this.logger.debug(`[OUTBOX] Marked event ${id} as SENT (awaiting consumer confirmation)`);
}
/**
*
*/
async markAsConfirmed(eventId: string, eventType?: string): Promise<boolean> {
const whereClause: Prisma.OutboxEventWhereInput = {
aggregateId: eventId,
status: OutboxStatus.SENT,
};
if (eventType) {
whereClause.eventType = eventType;
}
const result = await this.prisma.outboxEvent.updateMany({
where: whereClause,
data: {
status: OutboxStatus.CONFIRMED,
},
});
if (result.count > 0) {
this.logger.log(`[OUTBOX] ✓ Event ${eventId} (${eventType || 'all types'}) confirmed by consumer`);
return true;
}
this.logger.warn(`[OUTBOX] Event ${eventId} (${eventType || 'any'}) not found or not in SENT status`);
return false;
}
/**
*
*/
async findSentEventsTimedOut(timeoutMinutes: number = 5, limit: number = 50): Promise<OutboxEvent[]> {
const cutoffTime = new Date();
cutoffTime.setMinutes(cutoffTime.getMinutes() - timeoutMinutes);
const events = await this.prisma.outboxEvent.findMany({
where: {
status: OutboxStatus.SENT,
publishedAt: {
lt: cutoffTime,
},
retryCount: {
lt: 5,
},
},
orderBy: {
publishedAt: 'asc',
},
take: limit,
});
return events.map((e) => this.mapToOutboxEvent(e));
}
/**
* SENT PENDING 便
*/
async resetSentToPending(id: bigint): Promise<void> {
const event = await this.prisma.outboxEvent.findUnique({
where: { id },
});
if (!event) return;
await this.prisma.outboxEvent.update({
where: { id },
data: {
status: OutboxStatus.PENDING,
retryCount: event.retryCount + 1,
lastError: 'Consumer confirmation timeout',
},
});
this.logger.warn(`[OUTBOX] Event ${id} reset to PENDING for retry (confirmation timeout)`);
}
/**
*
*/
async markAsFailed(id: bigint, error: string): Promise<void> {
const event = await this.prisma.outboxEvent.findUnique({
where: { id },
});
if (!event) return;
const newRetryCount = event.retryCount + 1;
const isFinalFailure = newRetryCount >= event.maxRetries;
// 指数退避1min, 2min, 4min, 8min, 16min
const delayMinutes = Math.pow(2, newRetryCount - 1);
const nextRetryAt = new Date();
nextRetryAt.setMinutes(nextRetryAt.getMinutes() + delayMinutes);
await this.prisma.outboxEvent.update({
where: { id },
data: {
status: OutboxStatus.FAILED,
retryCount: newRetryCount,
lastError: error,
nextRetryAt: isFinalFailure ? null : nextRetryAt,
},
});
if (isFinalFailure) {
this.logger.error(`[OUTBOX] Event ${id} permanently failed after ${newRetryCount} retries: ${error}`);
} else {
this.logger.warn(`[OUTBOX] Event ${id} failed, retry ${newRetryCount}/${event.maxRetries}, next retry at ${nextRetryAt}`);
}
}
/**
* 7
*/
async cleanupOldEvents(retentionDays: number = 7): Promise<number> {
const cutoffDate = new Date();
cutoffDate.setDate(cutoffDate.getDate() - retentionDays);
const result = await this.prisma.outboxEvent.deleteMany({
where: {
status: OutboxStatus.CONFIRMED,
publishedAt: {
lt: cutoffDate,
},
},
});
if (result.count > 0) {
this.logger.log(`[OUTBOX] Cleaned up ${result.count} old events`);
}
return result.count;
}
/**
* Outbox
*/
async getStats(): Promise<{
pending: number;
sent: number;
confirmed: number;
failed: number;
}> {
const [pending, sent, confirmed, failed] = await Promise.all([
this.prisma.outboxEvent.count({ where: { status: OutboxStatus.PENDING } }),
this.prisma.outboxEvent.count({ where: { status: OutboxStatus.SENT } }),
this.prisma.outboxEvent.count({ where: { status: OutboxStatus.CONFIRMED } }),
this.prisma.outboxEvent.count({ where: { status: OutboxStatus.FAILED } }),
]);
return { pending, sent, confirmed, failed };
}
private mapToOutboxEvent(record: any): OutboxEvent {
return {
id: record.id,
eventType: record.eventType,
topic: record.topic,
key: record.key,
payload: record.payload as Record<string, unknown>,
aggregateId: record.aggregateId,
aggregateType: record.aggregateType,
status: record.status as OutboxStatus,
retryCount: record.retryCount,
maxRetries: record.maxRetries,
lastError: record.lastError,
createdAt: record.createdAt,
publishedAt: record.publishedAt,
nextRetryAt: record.nextRetryAt,
};
}
}

View File

@ -0,0 +1,22 @@
-- CreateTable
CREATE TABLE "processed_events" (
"processed_id" BIGSERIAL NOT NULL,
"event_id" VARCHAR(200) NOT NULL,
"event_type" VARCHAR(100) NOT NULL,
"source_service" VARCHAR(50) NOT NULL,
"processed_at" TIMESTAMP(3) NOT NULL DEFAULT CURRENT_TIMESTAMP,
CONSTRAINT "processed_events_pkey" PRIMARY KEY ("processed_id")
);
-- CreateIndex
CREATE INDEX "processed_events_event_id_idx" ON "processed_events"("event_id");
-- CreateIndex
CREATE INDEX "processed_events_event_type_idx" ON "processed_events"("event_type");
-- CreateIndex
CREATE INDEX "processed_events_processed_at_idx" ON "processed_events"("processed_at");
-- CreateIndex
CREATE UNIQUE INDEX "processed_events_event_id_event_type_key" ON "processed_events"("event_id", "event_type");

View File

@ -199,3 +199,27 @@ model WithdrawalOrder {
@@index([txHash])
@@index([createdAt])
}
// ============================================
// 已处理事件表 (幂等性检查)
// 用于确保 Kafka 事件不会被重复处理
// ============================================
model ProcessedEvent {
id BigInt @id @default(autoincrement()) @map("processed_id")
// 事件标识 (聚合根ID + 事件类型)
eventId String @map("event_id") @db.VarChar(200)
eventType String @map("event_type") @db.VarChar(100)
// 来源服务
sourceService String @map("source_service") @db.VarChar(50)
// 处理时间
processedAt DateTime @default(now()) @map("processed_at")
@@unique([eventId, eventType])
@@map("processed_events")
@@index([eventId])
@@index([eventType])
@@index([processedAt])
}

View File

@ -0,0 +1,86 @@
import { Injectable, Logger, Inject } from '@nestjs/common';
import { ClientKafka } from '@nestjs/microservices';
/**
*
*/
interface EventAckMessage {
/** 原始事件的 aggregateId */
eventId: string;
/** 原始事件类型 */
eventType: string;
/** 消费服务名称 */
consumerService: string;
/** 处理结果 */
success: boolean;
/** 错误信息(如果失败) */
error?: string;
/** 确认时间 */
confirmedAt: string;
}
/**
*
*
* B方案核心组件
* reward.events.ack topic
*/
@Injectable()
export class EventAckPublisher {
private readonly logger = new Logger(EventAckPublisher.name);
private readonly serviceName = 'wallet-service';
constructor(
@Inject('KAFKA_SERVICE')
private readonly kafkaClient: ClientKafka,
) {}
/**
*
*/
async sendSuccess(eventId: string, eventType: string): Promise<void> {
const ackMessage: EventAckMessage = {
eventId,
eventType,
consumerService: this.serviceName,
success: true,
confirmedAt: new Date().toISOString(),
};
try {
this.kafkaClient.emit('reward.events.ack', {
key: eventId,
value: JSON.stringify(ackMessage),
});
this.logger.log(`[ACK] ✓ Sent success confirmation for event ${eventId} (${eventType})`);
} catch (error) {
this.logger.error(`[ACK] Failed to send confirmation for event ${eventId}:`, error);
}
}
/**
*
*/
async sendFailure(eventId: string, eventType: string, errorMessage: string): Promise<void> {
const ackMessage: EventAckMessage = {
eventId,
eventType,
consumerService: this.serviceName,
success: false,
error: errorMessage,
confirmedAt: new Date().toISOString(),
};
try {
this.kafkaClient.emit('reward.events.ack', {
key: eventId,
value: JSON.stringify(ackMessage),
});
this.logger.warn(`[ACK] ✗ Sent failure confirmation for event ${eventId}: ${errorMessage}`);
} catch (error) {
this.logger.error(`[ACK] Failed to send failure confirmation for event ${eventId}:`, error);
}
}
}

View File

@ -1,10 +1,37 @@
import { Module, Global } from '@nestjs/common';
import { ConfigModule, ConfigService } from '@nestjs/config';
import { ClientsModule, Transport } from '@nestjs/microservices';
import { EventPublisherService } from './event-publisher.service';
import { DepositEventConsumerService } from './deposit-event-consumer.service';
import { RewardEventConsumerController } from './reward-event-consumer.controller';
import { EventAckPublisher } from './event-ack.publisher';
import { PrismaService } from '../persistence/prisma/prisma.service';
@Global()
@Module({
providers: [EventPublisherService, DepositEventConsumerService],
exports: [EventPublisherService, DepositEventConsumerService],
imports: [
ClientsModule.registerAsync([
{
name: 'KAFKA_SERVICE',
imports: [ConfigModule],
useFactory: (configService: ConfigService) => ({
transport: Transport.KAFKA,
options: {
client: {
clientId: configService.get<string>('KAFKA_CLIENT_ID', 'wallet-service'),
brokers: configService.get<string>('KAFKA_BROKERS', 'localhost:9092').split(','),
},
consumer: {
groupId: configService.get<string>('KAFKA_GROUP_ID', 'wallet-service-group'),
},
},
}),
inject: [ConfigService],
},
]),
],
controllers: [RewardEventConsumerController],
providers: [PrismaService, EventPublisherService, DepositEventConsumerService, EventAckPublisher],
exports: [EventPublisherService, DepositEventConsumerService, EventAckPublisher, ClientsModule],
})
export class KafkaModule {}

View File

@ -0,0 +1,139 @@
import { Controller, Logger } from '@nestjs/common';
import { MessagePattern, Payload, Ctx, KafkaContext } from '@nestjs/microservices';
import { PrismaService } from '../persistence/prisma/prisma.service';
import { EventAckPublisher } from './event-ack.publisher';
import Decimal from 'decimal.js';
/**
*
*/
interface RewardSummaryUpdatedEvent {
eventType: string;
aggregateId: string;
payload: {
accountSequence: string;
userId: string;
pendingUsdt: number;
pendingHashpower: number;
pendingExpireAt: string | null;
settleableUsdt: number;
settleableHashpower: number;
settledTotalUsdt: number;
settledTotalHashpower: number;
expiredTotalUsdt: number;
expiredTotalHashpower: number;
};
_outbox?: {
id: string;
aggregateId: string;
eventType: string;
};
}
/**
* Kafka
*
* reward-service
* 使
*/
@Controller()
export class RewardEventConsumerController {
private readonly logger = new Logger(RewardEventConsumerController.name);
constructor(
private readonly prisma: PrismaService,
private readonly eventAckPublisher: EventAckPublisher,
) {}
/**
*
*/
@MessagePattern('reward.summary.updated')
async handleRewardSummaryUpdated(
@Payload() message: RewardSummaryUpdatedEvent,
@Ctx() context: KafkaContext,
): Promise<void> {
const partition = context.getPartition();
const offset = context.getMessage().offset;
const outboxInfo = message._outbox;
const eventId = outboxInfo?.aggregateId || message.aggregateId;
const eventType = outboxInfo?.eventType || message.eventType;
this.logger.log(
`[REWARD-EVENT] Received reward.summary.updated for ${eventId} ` +
`[partition=${partition}, offset=${offset}]`,
);
try {
// 1. 幂等性检查
const alreadyProcessed = await this.prisma.processedEvent.findUnique({
where: {
eventId_eventType: {
eventId,
eventType,
},
},
});
if (alreadyProcessed) {
this.logger.warn(`[REWARD-EVENT] Event ${eventId} (${eventType}) already processed, skipping`);
// 仍然发送确认,避免重复发送
if (outboxInfo) {
await this.eventAckPublisher.sendSuccess(eventId, eventType);
}
return;
}
// 2. 更新 wallet_accounts 表的 rewards 数据
const payload = message.payload;
const accountSequence = BigInt(payload.accountSequence);
await this.prisma.$transaction(async (tx) => {
// 更新钱包账户的奖励数据
await tx.walletAccount.update({
where: { accountSequence },
data: {
pendingUsdt: new Decimal(payload.pendingUsdt),
pendingHashpower: new Decimal(payload.pendingHashpower),
pendingExpireAt: payload.pendingExpireAt ? new Date(payload.pendingExpireAt) : null,
settleableUsdt: new Decimal(payload.settleableUsdt),
settleableHashpower: new Decimal(payload.settleableHashpower),
settledTotalUsdt: new Decimal(payload.settledTotalUsdt),
settledTotalHashpower: new Decimal(payload.settledTotalHashpower),
expiredTotalUsdt: new Decimal(payload.expiredTotalUsdt),
expiredTotalHashpower: new Decimal(payload.expiredTotalHashpower),
},
});
// 记录已处理事件(幂等性)
await tx.processedEvent.create({
data: {
eventId,
eventType,
sourceService: 'reward-service',
},
});
});
this.logger.log(
`[REWARD-EVENT] ✓ Updated rewards for accountSequence ${accountSequence}: ` +
`settleable=${payload.settleableUsdt}, pending=${payload.pendingUsdt}`,
);
// 3. 发送确认
if (outboxInfo) {
await this.eventAckPublisher.sendSuccess(eventId, eventType);
}
} catch (error) {
const errorMessage = error instanceof Error ? error.message : String(error);
this.logger.error(`[REWARD-EVENT] Error processing event ${eventId}: ${errorMessage}`);
// 发送失败确认
if (outboxInfo) {
await this.eventAckPublisher.sendFailure(eventId, eventType, errorMessage);
}
throw error;
}
}
}

View File

@ -1,9 +1,11 @@
import { NestFactory } from '@nestjs/core';
import { ValidationPipe } from '@nestjs/common';
import { ValidationPipe, Logger } from '@nestjs/common';
import { SwaggerModule, DocumentBuilder } from '@nestjs/swagger';
import { MicroserviceOptions, Transport } from '@nestjs/microservices';
import { AppModule } from './app.module';
async function bootstrap() {
const logger = new Logger('Bootstrap');
const app = await NestFactory.create(AppModule);
// Global prefix
@ -36,10 +38,30 @@ async function bootstrap() {
const document = SwaggerModule.createDocument(app, config);
SwaggerModule.setup('api/docs', app, document);
// Kafka 微服务 - 用于 @MessagePattern 消费消息
const kafkaBrokers = process.env.KAFKA_BROKERS?.split(',') || ['localhost:9092'];
const kafkaGroupId = process.env.KAFKA_GROUP_ID || 'wallet-service-group';
app.connectMicroservice<MicroserviceOptions>({
transport: Transport.KAFKA,
options: {
client: {
clientId: 'wallet-service',
brokers: kafkaBrokers,
},
consumer: {
groupId: kafkaGroupId,
},
},
});
await app.startAllMicroservices();
logger.log('Kafka microservice started');
const port = process.env.APP_PORT || 3002;
await app.listen(port);
console.log(`Wallet Service is running on port ${port}`);
console.log(`Swagger docs: http://localhost:${port}/api/docs`);
logger.log(`Wallet Service is running on port ${port}`);
logger.log(`Swagger docs: http://localhost:${port}/api/docs`);
}
bootstrap();