fix(ops): /health DB+Redis checks, validated env.REDIS_URL across workers, error_events 90d retention
Three audit-pass-#3 findings, all in the "wakes you at 3am" category.
- /api/public/health now runs DB SELECT 1 + Redis PING in parallel and
returns 503 + a degraded payload when either fails. Anonymous probes
(no X-Intake-Secret) still get a flat {status:'ok'} so generic uptime
monitors keep working; authenticated probes see the dep results.
- All worker entrypoints (ai, bulk, documents, email, export, import,
maintenance, notifications, reports, webhooks) and src/lib/redis.ts
now use env.REDIS_URL (Zod-validated at boot) instead of
process.env.REDIS_URL!. Previously a missing env let the app start
silently and fail at first job pickup.
- maintenance worker gains an `error-events-retention` case that
delete()s rows older than 90 days from error_events. scheduler.ts
registers it at 06:00 daily. Closes the contract from migration
0040 which declared the table "pruned at 90 days" but had no
implementation.
Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
This commit is contained in:
@@ -1,28 +1,56 @@
|
|||||||
import { NextRequest, NextResponse } from 'next/server';
|
import { NextRequest, NextResponse } from 'next/server';
|
||||||
import { timingSafeEqual } from 'node:crypto';
|
import { timingSafeEqual } from 'node:crypto';
|
||||||
|
import { sql } from 'drizzle-orm';
|
||||||
|
|
||||||
|
import { db } from '@/lib/db';
|
||||||
|
import { redis } from '@/lib/redis';
|
||||||
import { env } from '@/lib/env';
|
import { env } from '@/lib/env';
|
||||||
|
import { logger } from '@/lib/logger';
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* GET /api/public/health
|
* GET /api/public/health
|
||||||
*
|
*
|
||||||
* Health probe used by the marketing-website server on startup to verify
|
* Liveness + readiness probe. Two response shapes:
|
||||||
* it's pointed at a CRM matching its own deployment env (plan §14.8
|
|
||||||
* critical: prevent staging-website-talking-to-prod-CRM).
|
|
||||||
*
|
*
|
||||||
* Auditor-K §41 flagged that the previous response disclosed `NODE_ENV`
|
* 1. **Anonymous (no header):** minimal `{status:'ok'}` so uptime
|
||||||
* and `APP_URL` to anonymous internet — mirrors the website's own intake
|
* monitors can poll without leaking deployment fingerprints.
|
||||||
* secret gate so we don't leak deployment fingerprints. When
|
*
|
||||||
* `WEBSITE_INTAKE_SECRET` is set and the caller presents the matching
|
* 2. **Authenticated (`X-Intake-Secret`, timing-safe compared):** full
|
||||||
* `X-Intake-Secret` header we return the full payload; otherwise return
|
* payload including env + dependency check results. This is what
|
||||||
* a minimal `{status:'ok'}` so generic uptime monitors still get a 200.
|
* the marketing site uses on startup AND what k8s readiness
|
||||||
|
* probes should hit, because it returns 503 on hard dep failures.
|
||||||
|
*
|
||||||
|
* The dep checks (DB SELECT 1, Redis PING) run on every request — they
|
||||||
|
* are <5ms each. If either fails, the response is 503 so a load balancer
|
||||||
|
* stops routing to this instance.
|
||||||
*/
|
*/
|
||||||
export function GET(req: NextRequest): Response {
|
|
||||||
|
type HealthCheck = { ok: true; latencyMs: number } | { ok: false; error: string };
|
||||||
|
|
||||||
|
async function checkDb(): Promise<HealthCheck> {
|
||||||
|
const start = Date.now();
|
||||||
|
try {
|
||||||
|
await db.execute(sql`SELECT 1`);
|
||||||
|
return { ok: true, latencyMs: Date.now() - start };
|
||||||
|
} catch (err) {
|
||||||
|
return { ok: false, error: err instanceof Error ? err.message : 'unknown' };
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
async function checkRedis(): Promise<HealthCheck> {
|
||||||
|
const start = Date.now();
|
||||||
|
try {
|
||||||
|
const pong = await redis.ping();
|
||||||
|
if (pong !== 'PONG') return { ok: false, error: `unexpected: ${pong}` };
|
||||||
|
return { ok: true, latencyMs: Date.now() - start };
|
||||||
|
} catch (err) {
|
||||||
|
return { ok: false, error: err instanceof Error ? err.message : 'unknown' };
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
export async function GET(req: NextRequest): Promise<Response> {
|
||||||
const expected = env.WEBSITE_INTAKE_SECRET;
|
const expected = env.WEBSITE_INTAKE_SECRET;
|
||||||
const provided = req.headers.get('x-intake-secret');
|
const provided = req.headers.get('x-intake-secret');
|
||||||
// Use timingSafeEqual rather than a `===` comparison — string equality
|
|
||||||
// is not constant-time and lets a remote attacker enumerate the secret
|
|
||||||
// byte-by-byte via response-time differences.
|
|
||||||
const matched =
|
const matched =
|
||||||
!!expected &&
|
!!expected &&
|
||||||
!!provided &&
|
!!provided &&
|
||||||
@@ -35,6 +63,8 @@ export function GET(req: NextRequest): Response {
|
|||||||
}
|
}
|
||||||
})();
|
})();
|
||||||
|
|
||||||
|
// Anonymous probe: no dep checks, never 503. Uptime monitors that
|
||||||
|
// can't carry the secret keep working.
|
||||||
if (!matched) {
|
if (!matched) {
|
||||||
return NextResponse.json(
|
return NextResponse.json(
|
||||||
{ status: 'ok', timestamp: new Date().toISOString() },
|
{ status: 'ok', timestamp: new Date().toISOString() },
|
||||||
@@ -42,13 +72,23 @@ export function GET(req: NextRequest): Response {
|
|||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Authenticated probe: run dep checks in parallel and surface a 503
|
||||||
|
// when anything required is down.
|
||||||
|
const [dbCheck, redisCheck] = await Promise.all([checkDb(), checkRedis()]);
|
||||||
|
const allOk = dbCheck.ok && redisCheck.ok;
|
||||||
|
|
||||||
|
if (!allOk) {
|
||||||
|
logger.warn({ db: dbCheck, redis: redisCheck }, 'Health probe found an unhealthy dependency');
|
||||||
|
}
|
||||||
|
|
||||||
return NextResponse.json(
|
return NextResponse.json(
|
||||||
{
|
{
|
||||||
status: 'ok',
|
status: allOk ? 'ok' : 'degraded',
|
||||||
env: env.NODE_ENV,
|
env: env.NODE_ENV,
|
||||||
appUrl: env.APP_URL,
|
appUrl: env.APP_URL,
|
||||||
timestamp: new Date().toISOString(),
|
timestamp: new Date().toISOString(),
|
||||||
|
checks: { db: dbCheck, redis: redisCheck },
|
||||||
},
|
},
|
||||||
{ headers: { 'cache-control': 'no-store' } },
|
{ status: allOk ? 200 : 503, headers: { 'cache-control': 'no-store' } },
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,19 +1,20 @@
|
|||||||
import { Queue, type ConnectionOptions } from 'bullmq';
|
import { Queue, type ConnectionOptions } from 'bullmq';
|
||||||
|
import { env } from '@/lib/env';
|
||||||
|
|
||||||
const redisUrl = process.env.REDIS_URL!;
|
const redisUrl = env.REDIS_URL;
|
||||||
|
|
||||||
// 10 queues matching 11-REALTIME-AND-BACKGROUND-JOBS.md Section 3.1
|
// 10 queues matching 11-REALTIME-AND-BACKGROUND-JOBS.md Section 3.1
|
||||||
const QUEUE_CONFIGS = {
|
const QUEUE_CONFIGS = {
|
||||||
email: { concurrency: 5, maxAttempts: 5 },
|
email: { concurrency: 5, maxAttempts: 5 },
|
||||||
documents: { concurrency: 3, maxAttempts: 5 },
|
documents: { concurrency: 3, maxAttempts: 5 },
|
||||||
notifications: { concurrency: 10, maxAttempts: 3 },
|
notifications: { concurrency: 10, maxAttempts: 3 },
|
||||||
import: { concurrency: 1, maxAttempts: 1 },
|
import: { concurrency: 1, maxAttempts: 1 },
|
||||||
export: { concurrency: 2, maxAttempts: 3 },
|
export: { concurrency: 2, maxAttempts: 3 },
|
||||||
reports: { concurrency: 1, maxAttempts: 3 },
|
reports: { concurrency: 1, maxAttempts: 3 },
|
||||||
webhooks: { concurrency: 5, maxAttempts: 3 },
|
webhooks: { concurrency: 5, maxAttempts: 3 },
|
||||||
maintenance: { concurrency: 1, maxAttempts: 3 },
|
maintenance: { concurrency: 1, maxAttempts: 3 },
|
||||||
ai: { concurrency: 2, maxAttempts: 3 },
|
ai: { concurrency: 2, maxAttempts: 3 },
|
||||||
bulk: { concurrency: 2, maxAttempts: 3 },
|
bulk: { concurrency: 2, maxAttempts: 3 },
|
||||||
} as const;
|
} as const;
|
||||||
|
|
||||||
export type QueueName = keyof typeof QUEUE_CONFIGS;
|
export type QueueName = keyof typeof QUEUE_CONFIGS;
|
||||||
@@ -28,8 +29,8 @@ export function getQueue(name: QueueName): Queue {
|
|||||||
defaultJobOptions: {
|
defaultJobOptions: {
|
||||||
attempts: QUEUE_CONFIGS[name].maxAttempts,
|
attempts: QUEUE_CONFIGS[name].maxAttempts,
|
||||||
backoff: { type: 'exponential', delay: 1000 },
|
backoff: { type: 'exponential', delay: 1000 },
|
||||||
removeOnComplete: { age: 24 * 3600 }, // keep completed jobs 24 hours
|
removeOnComplete: { age: 24 * 3600 }, // keep completed jobs 24 hours
|
||||||
removeOnFail: { age: 7 * 24 * 3600 }, // keep failed jobs 7 days
|
removeOnFail: { age: 7 * 24 * 3600 }, // keep failed jobs 7 days
|
||||||
},
|
},
|
||||||
});
|
});
|
||||||
queues.set(name, queue);
|
queues.set(name, queue);
|
||||||
|
|||||||
@@ -57,6 +57,8 @@ export async function registerRecurringJobs(): Promise<void> {
|
|||||||
{ queue: 'maintenance', name: 'gdpr-export-cleanup', pattern: '0 4 * * *' },
|
{ queue: 'maintenance', name: 'gdpr-export-cleanup', pattern: '0 4 * * *' },
|
||||||
// Phase 3b: AI usage ledger retention (90-day rolling window)
|
// Phase 3b: AI usage ledger retention (90-day rolling window)
|
||||||
{ queue: 'maintenance', name: 'ai-usage-retention', pattern: '0 5 * * *' },
|
{ queue: 'maintenance', name: 'ai-usage-retention', pattern: '0 5 * * *' },
|
||||||
|
// Migration 0040 contract: error_events older than 90 days get pruned.
|
||||||
|
{ queue: 'maintenance', name: 'error-events-retention', pattern: '0 6 * * *' },
|
||||||
];
|
];
|
||||||
|
|
||||||
for (const job of recurring) {
|
for (const job of recurring) {
|
||||||
|
|||||||
@@ -1,4 +1,5 @@
|
|||||||
import { Worker, type Job } from 'bullmq';
|
import { Worker, type Job } from 'bullmq';
|
||||||
|
import { env } from '@/lib/env';
|
||||||
|
|
||||||
import type { ConnectionOptions } from 'bullmq';
|
import type { ConnectionOptions } from 'bullmq';
|
||||||
import { logger } from '@/lib/logger';
|
import { logger } from '@/lib/logger';
|
||||||
@@ -310,7 +311,7 @@ export const aiWorker = new Worker(
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
connection: { url: process.env.REDIS_URL! } as ConnectionOptions,
|
connection: { url: env.REDIS_URL } as ConnectionOptions,
|
||||||
concurrency: QUEUE_CONFIGS.ai.concurrency,
|
concurrency: QUEUE_CONFIGS.ai.concurrency,
|
||||||
},
|
},
|
||||||
);
|
);
|
||||||
|
|||||||
@@ -1,4 +1,5 @@
|
|||||||
import { Worker, type Job } from 'bullmq';
|
import { Worker, type Job } from 'bullmq';
|
||||||
|
import { env } from '@/lib/env';
|
||||||
|
|
||||||
import type { ConnectionOptions } from 'bullmq';
|
import type { ConnectionOptions } from 'bullmq';
|
||||||
import { logger } from '@/lib/logger';
|
import { logger } from '@/lib/logger';
|
||||||
@@ -19,7 +20,7 @@ export const documentsWorker = new Worker(
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
connection: { url: process.env.REDIS_URL! } as ConnectionOptions,
|
connection: { url: env.REDIS_URL } as ConnectionOptions,
|
||||||
concurrency: QUEUE_CONFIGS.documents.concurrency,
|
concurrency: QUEUE_CONFIGS.documents.concurrency,
|
||||||
},
|
},
|
||||||
);
|
);
|
||||||
|
|||||||
@@ -1,4 +1,5 @@
|
|||||||
import { Worker, type Job } from 'bullmq';
|
import { Worker, type Job } from 'bullmq';
|
||||||
|
import { env } from '@/lib/env';
|
||||||
|
|
||||||
import type { ConnectionOptions } from 'bullmq';
|
import type { ConnectionOptions } from 'bullmq';
|
||||||
import { logger } from '@/lib/logger';
|
import { logger } from '@/lib/logger';
|
||||||
@@ -56,7 +57,7 @@ export const emailWorker = new Worker(
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
connection: { url: process.env.REDIS_URL! } as ConnectionOptions,
|
connection: { url: env.REDIS_URL } as ConnectionOptions,
|
||||||
concurrency: QUEUE_CONFIGS.email.concurrency,
|
concurrency: QUEUE_CONFIGS.email.concurrency,
|
||||||
},
|
},
|
||||||
);
|
);
|
||||||
|
|||||||
@@ -1,4 +1,5 @@
|
|||||||
import { Worker, type Job } from 'bullmq';
|
import { Worker, type Job } from 'bullmq';
|
||||||
|
import { env } from '@/lib/env';
|
||||||
|
|
||||||
import type { ConnectionOptions } from 'bullmq';
|
import type { ConnectionOptions } from 'bullmq';
|
||||||
import { logger } from '@/lib/logger';
|
import { logger } from '@/lib/logger';
|
||||||
@@ -26,7 +27,7 @@ export const exportWorker = new Worker(
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
connection: { url: process.env.REDIS_URL! } as ConnectionOptions,
|
connection: { url: env.REDIS_URL } as ConnectionOptions,
|
||||||
concurrency: QUEUE_CONFIGS.export.concurrency,
|
concurrency: QUEUE_CONFIGS.export.concurrency,
|
||||||
},
|
},
|
||||||
);
|
);
|
||||||
|
|||||||
@@ -1,4 +1,5 @@
|
|||||||
import { Worker, type Job } from 'bullmq';
|
import { Worker, type Job } from 'bullmq';
|
||||||
|
import { env } from '@/lib/env';
|
||||||
|
|
||||||
import type { ConnectionOptions } from 'bullmq';
|
import type { ConnectionOptions } from 'bullmq';
|
||||||
import { logger } from '@/lib/logger';
|
import { logger } from '@/lib/logger';
|
||||||
@@ -14,7 +15,7 @@ export const importWorker = new Worker(
|
|||||||
// - Note: maxAttempts=1 - imports are idempotent, user retries manually
|
// - Note: maxAttempts=1 - imports are idempotent, user retries manually
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
connection: { url: process.env.REDIS_URL! } as ConnectionOptions,
|
connection: { url: env.REDIS_URL } as ConnectionOptions,
|
||||||
concurrency: QUEUE_CONFIGS.import.concurrency,
|
concurrency: QUEUE_CONFIGS.import.concurrency,
|
||||||
},
|
},
|
||||||
);
|
);
|
||||||
|
|||||||
@@ -1,4 +1,5 @@
|
|||||||
import { Worker, type Job } from 'bullmq';
|
import { Worker, type Job } from 'bullmq';
|
||||||
|
import { env } from '@/lib/env';
|
||||||
import { and, eq, lt, isNotNull } from 'drizzle-orm';
|
import { and, eq, lt, isNotNull } from 'drizzle-orm';
|
||||||
|
|
||||||
import type { ConnectionOptions } from 'bullmq';
|
import type { ConnectionOptions } from 'bullmq';
|
||||||
@@ -6,12 +7,16 @@ import { db } from '@/lib/db';
|
|||||||
import { formSubmissions } from '@/lib/db/schema/documents';
|
import { formSubmissions } from '@/lib/db/schema/documents';
|
||||||
import { gdprExports } from '@/lib/db/schema/gdpr';
|
import { gdprExports } from '@/lib/db/schema/gdpr';
|
||||||
import { aiUsageLedger } from '@/lib/db/schema/ai-usage';
|
import { aiUsageLedger } from '@/lib/db/schema/ai-usage';
|
||||||
|
import { errorEvents } from '@/lib/db/schema/system';
|
||||||
import { logger } from '@/lib/logger';
|
import { logger } from '@/lib/logger';
|
||||||
import { getStorageBackend } from '@/lib/storage';
|
import { getStorageBackend } from '@/lib/storage';
|
||||||
import { QUEUE_CONFIGS } from '@/lib/queue';
|
import { QUEUE_CONFIGS } from '@/lib/queue';
|
||||||
|
|
||||||
/** AI usage rows older than this are deleted by the retention job. */
|
/** AI usage rows older than this are deleted by the retention job. */
|
||||||
const AI_USAGE_RETENTION_DAYS = 90;
|
const AI_USAGE_RETENTION_DAYS = 90;
|
||||||
|
/** error_events rows older than this are pruned. Migration 0040 declares
|
||||||
|
* this contract; the worker had no implementation until now. */
|
||||||
|
const ERROR_EVENTS_RETENTION_DAYS = 90;
|
||||||
|
|
||||||
export const maintenanceWorker = new Worker(
|
export const maintenanceWorker = new Worker(
|
||||||
'maintenance',
|
'maintenance',
|
||||||
@@ -113,12 +118,27 @@ export const maintenanceWorker = new Worker(
|
|||||||
);
|
);
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
|
case 'error-events-retention': {
|
||||||
|
// Honor the contract from migration 0040: error_events older than
|
||||||
|
// ERROR_EVENTS_RETENTION_DAYS get dropped. Otherwise the table
|
||||||
|
// grows unbounded and the admin error log becomes unusable.
|
||||||
|
const cutoff = new Date(Date.now() - ERROR_EVENTS_RETENTION_DAYS * 24 * 60 * 60 * 1000);
|
||||||
|
const result = await db
|
||||||
|
.delete(errorEvents)
|
||||||
|
.where(lt(errorEvents.createdAt, cutoff))
|
||||||
|
.returning({ requestId: errorEvents.requestId });
|
||||||
|
logger.info(
|
||||||
|
{ deleted: result.length, retentionDays: ERROR_EVENTS_RETENTION_DAYS },
|
||||||
|
'Error events retention sweep complete',
|
||||||
|
);
|
||||||
|
break;
|
||||||
|
}
|
||||||
default:
|
default:
|
||||||
logger.warn({ jobName: job.name }, 'Unknown maintenance job');
|
logger.warn({ jobName: job.name }, 'Unknown maintenance job');
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
connection: { url: process.env.REDIS_URL! } as ConnectionOptions,
|
connection: { url: env.REDIS_URL } as ConnectionOptions,
|
||||||
concurrency: QUEUE_CONFIGS.maintenance.concurrency,
|
concurrency: QUEUE_CONFIGS.maintenance.concurrency,
|
||||||
},
|
},
|
||||||
);
|
);
|
||||||
|
|||||||
@@ -1,4 +1,5 @@
|
|||||||
import { Worker, type Job } from 'bullmq';
|
import { Worker, type Job } from 'bullmq';
|
||||||
|
import { env } from '@/lib/env';
|
||||||
|
|
||||||
import type { ConnectionOptions } from 'bullmq';
|
import type { ConnectionOptions } from 'bullmq';
|
||||||
import { logger } from '@/lib/logger';
|
import { logger } from '@/lib/logger';
|
||||||
@@ -79,7 +80,7 @@ export const notificationsWorker = new Worker(
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
connection: { url: process.env.REDIS_URL! } as ConnectionOptions,
|
connection: { url: env.REDIS_URL } as ConnectionOptions,
|
||||||
concurrency: QUEUE_CONFIGS.notifications.concurrency,
|
concurrency: QUEUE_CONFIGS.notifications.concurrency,
|
||||||
},
|
},
|
||||||
);
|
);
|
||||||
|
|||||||
@@ -1,4 +1,5 @@
|
|||||||
import { Worker, type Job } from 'bullmq';
|
import { Worker, type Job } from 'bullmq';
|
||||||
|
import { env } from '@/lib/env';
|
||||||
|
|
||||||
import type { ConnectionOptions } from 'bullmq';
|
import type { ConnectionOptions } from 'bullmq';
|
||||||
import { logger } from '@/lib/logger';
|
import { logger } from '@/lib/logger';
|
||||||
@@ -21,10 +22,7 @@ export const reportsWorker = new Worker(
|
|||||||
.select()
|
.select()
|
||||||
.from(scheduledReports)
|
.from(scheduledReports)
|
||||||
.where(
|
.where(
|
||||||
and(
|
and(eq(scheduledReports.isActive, true), lte(scheduledReports.nextRunAt, new Date())),
|
||||||
eq(scheduledReports.isActive, true),
|
|
||||||
lte(scheduledReports.nextRunAt, new Date()),
|
|
||||||
),
|
|
||||||
);
|
);
|
||||||
|
|
||||||
for (const report of dueReports) {
|
for (const report of dueReports) {
|
||||||
@@ -64,7 +62,7 @@ export const reportsWorker = new Worker(
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
connection: { url: process.env.REDIS_URL! } as ConnectionOptions,
|
connection: { url: env.REDIS_URL } as ConnectionOptions,
|
||||||
concurrency: QUEUE_CONFIGS.reports.concurrency,
|
concurrency: QUEUE_CONFIGS.reports.concurrency,
|
||||||
},
|
},
|
||||||
);
|
);
|
||||||
|
|||||||
@@ -1,4 +1,5 @@
|
|||||||
import { Worker, type Job } from 'bullmq';
|
import { Worker, type Job } from 'bullmq';
|
||||||
|
import { env } from '@/lib/env';
|
||||||
import { createHmac } from 'node:crypto';
|
import { createHmac } from 'node:crypto';
|
||||||
import { lookup } from 'node:dns/promises';
|
import { lookup } from 'node:dns/promises';
|
||||||
|
|
||||||
@@ -277,7 +278,7 @@ export const webhooksWorker = new Worker(
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
connection: { url: process.env.REDIS_URL! } as ConnectionOptions,
|
connection: { url: env.REDIS_URL } as ConnectionOptions,
|
||||||
concurrency: QUEUE_CONFIGS.webhooks.concurrency,
|
concurrency: QUEUE_CONFIGS.webhooks.concurrency,
|
||||||
},
|
},
|
||||||
);
|
);
|
||||||
|
|||||||
@@ -1,8 +1,9 @@
|
|||||||
import Redis from 'ioredis';
|
import Redis from 'ioredis';
|
||||||
|
import { env } from '@/lib/env';
|
||||||
|
|
||||||
import { logger } from '@/lib/logger';
|
import { logger } from '@/lib/logger';
|
||||||
|
|
||||||
const redisUrl = process.env.REDIS_URL!;
|
const redisUrl = env.REDIS_URL;
|
||||||
|
|
||||||
export const redis = new Redis(redisUrl, {
|
export const redis = new Redis(redisUrl, {
|
||||||
maxRetriesPerRequest: 3,
|
maxRetriesPerRequest: 3,
|
||||||
|
|||||||
Reference in New Issue
Block a user