mirror of
https://github.com/xCyanGrizzly/DragonsStash.git
synced 2026-05-10 22:01:16 +00:00
Multipart archives where individual parts fit under Telegram's 2GB limit but total size exceeds 4GB were being silently skipped. These can now be processed up to 200GB total, with each part uploading directly. Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
19 lines
976 B
TypeScript
19 lines
976 B
TypeScript
export const config = {
|
|
databaseUrl: process.env.DATABASE_URL ?? "",
|
|
workerIntervalMinutes: parseInt(process.env.WORKER_INTERVAL_MINUTES ?? "60", 10),
|
|
tempDir: process.env.WORKER_TEMP_DIR ?? "/tmp/zips",
|
|
tdlibStateDir: process.env.TDLIB_STATE_DIR ?? "/data/tdlib",
|
|
maxZipSizeMB: parseInt(process.env.WORKER_MAX_ZIP_SIZE_MB ?? "204800", 10),
|
|
logLevel: (process.env.LOG_LEVEL ?? "info") as "debug" | "info" | "warn" | "error",
|
|
telegramApiId: parseInt(process.env.TELEGRAM_API_ID ?? "0", 10),
|
|
telegramApiHash: process.env.TELEGRAM_API_HASH ?? "",
|
|
/** Maximum jitter added to scheduler interval (in minutes) */
|
|
jitterMinutes: 5,
|
|
/** Maximum time span for multipart archive parts (in hours). 0 = no limit. */
|
|
multipartTimeoutHours: parseInt(process.env.MULTIPART_TIMEOUT_HOURS ?? "0", 10),
|
|
/** Delay between Telegram API calls (in ms) to avoid rate limits */
|
|
apiDelayMs: 1000,
|
|
/** Max retries for rate-limited requests */
|
|
maxRetries: 5,
|
|
} as const;
|