fix(server): hata yönetimini iyileştir
Zamanlayıcı ve qbit istemcisi bileşenlerinde hata işleme yeteneklerini güçlendirir. - loop.scheduler: qbit hatalarında sistem durumunu ve sağlık bilgisini güncelleme ekler. - qbit.client: geçici ağ hatalarını (EAI_AGAIN vb.) algılayarak oturum durumunu sıfırlar. - timer.worker: global hata yakalama ekleyerek işleyicinin çökmesini engeller ve hataları günlüğe kaydeder.
This commit is contained in:
@@ -1,7 +1,7 @@
|
|||||||
import { QbitClient } from "../qbit/qbit.client"
|
import { QbitClient } from "../qbit/qbit.client"
|
||||||
import { tickLoopJobs } from "./loop.engine"
|
import { tickLoopJobs } from "./loop.engine"
|
||||||
import { getStatusSnapshot, refreshJobsStatus, setTorrentsStatus } from "../status/status.service"
|
import { getStatusSnapshot, refreshJobsStatus, setQbitStatus, setTorrentsStatus } from "../status/status.service"
|
||||||
import { emitStatusUpdate } from "../realtime/emitter"
|
import { emitQbitHealth, emitStatusUpdate } from "../realtime/emitter"
|
||||||
import { logger } from "../utils/logger"
|
import { logger } from "../utils/logger"
|
||||||
|
|
||||||
export const startLoopScheduler = (qbit: QbitClient, intervalMs: number) => {
|
export const startLoopScheduler = (qbit: QbitClient, intervalMs: number) => {
|
||||||
@@ -20,7 +20,19 @@ export const startLoopScheduler = (qbit: QbitClient, intervalMs: number) => {
|
|||||||
jobs,
|
jobs,
|
||||||
});
|
});
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
|
const message = error instanceof Error ? error.message : "Unknown error";
|
||||||
logger.error({ error }, "Loop scheduler tick failed");
|
logger.error({ error }, "Loop scheduler tick failed");
|
||||||
|
setQbitStatus({ ok: false, lastError: message });
|
||||||
|
emitQbitHealth({ ok: false, lastError: message });
|
||||||
|
try {
|
||||||
|
const current = await getStatusSnapshot();
|
||||||
|
emitStatusUpdate({
|
||||||
|
...current,
|
||||||
|
qbit: { ...current.qbit, ok: false, lastError: message },
|
||||||
|
});
|
||||||
|
} catch {
|
||||||
|
// Swallow secondary status errors to keep scheduler alive.
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}, intervalMs);
|
}, intervalMs);
|
||||||
};
|
};
|
||||||
|
|||||||
@@ -48,6 +48,18 @@ export class QbitClient {
|
|||||||
}
|
}
|
||||||
return await fn();
|
return await fn();
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
|
if (axios.isAxiosError(error)) {
|
||||||
|
const code = error.code ?? "";
|
||||||
|
const transient =
|
||||||
|
code === "EAI_AGAIN" ||
|
||||||
|
code === "ENOTFOUND" ||
|
||||||
|
code === "ECONNREFUSED" ||
|
||||||
|
code === "ECONNRESET" ||
|
||||||
|
code === "ETIMEDOUT";
|
||||||
|
if (transient) {
|
||||||
|
this.loggedIn = false;
|
||||||
|
}
|
||||||
|
}
|
||||||
if (
|
if (
|
||||||
axios.isAxiosError(error) &&
|
axios.isAxiosError(error) &&
|
||||||
(error.response?.status === 401 || error.response?.status === 403)
|
(error.response?.status === 401 || error.response?.status === 403)
|
||||||
|
|||||||
@@ -4,6 +4,7 @@ import { readDb, writeDb } from "../storage/jsondb";
|
|||||||
import { TimerLog, TimerSummary } from "../types";
|
import { TimerLog, TimerSummary } from "../types";
|
||||||
import { emitTimerLog, emitTimerSummary } from "../realtime/emitter";
|
import { emitTimerLog, emitTimerSummary } from "../realtime/emitter";
|
||||||
import { nowIso } from "../utils/time";
|
import { nowIso } from "../utils/time";
|
||||||
|
import { logger } from "../utils/logger";
|
||||||
|
|
||||||
const MAX_LOGS = 2000;
|
const MAX_LOGS = 2000;
|
||||||
|
|
||||||
@@ -17,76 +18,80 @@ const normalizeTags = (tags?: string, category?: string) => {
|
|||||||
|
|
||||||
export const startTimerWorker = (qbit: QbitClient, intervalMs: number) => {
|
export const startTimerWorker = (qbit: QbitClient, intervalMs: number) => {
|
||||||
setInterval(async () => {
|
setInterval(async () => {
|
||||||
const db = await readDb();
|
try {
|
||||||
const rules = db.timerRules ?? [];
|
const db = await readDb();
|
||||||
if (rules.length === 0) {
|
const rules = db.timerRules ?? [];
|
||||||
return;
|
if (rules.length === 0) {
|
||||||
}
|
return;
|
||||||
const torrents = await qbit.getTorrentsInfo();
|
|
||||||
let summary: TimerSummary =
|
|
||||||
db.timerSummary ?? {
|
|
||||||
totalDeleted: 0,
|
|
||||||
totalSeededSeconds: 0,
|
|
||||||
totalUploadedBytes: 0,
|
|
||||||
updatedAt: nowIso(),
|
|
||||||
};
|
|
||||||
|
|
||||||
const logs: TimerLog[] = [];
|
|
||||||
|
|
||||||
for (const torrent of torrents) {
|
|
||||||
const tags = normalizeTags(torrent.tags, torrent.category);
|
|
||||||
const matchingRules = rules.filter((rule) => {
|
|
||||||
return rule.tags.some((tag) => tags.includes(tag.toLowerCase()));
|
|
||||||
});
|
|
||||||
if (matchingRules.length === 0) {
|
|
||||||
continue;
|
|
||||||
}
|
}
|
||||||
const matchingRule = matchingRules.reduce((best, current) =>
|
const torrents = await qbit.getTorrentsInfo();
|
||||||
current.seedLimitSeconds < best.seedLimitSeconds ? current : best
|
let summary: TimerSummary =
|
||||||
);
|
db.timerSummary ?? {
|
||||||
const addedOnMs = Number(torrent.added_on ?? 0) * 1000;
|
totalDeleted: 0,
|
||||||
const elapsedSeconds =
|
totalSeededSeconds: 0,
|
||||||
addedOnMs > 0
|
totalUploadedBytes: 0,
|
||||||
? Math.max(0, (Date.now() - addedOnMs) / 1000)
|
updatedAt: nowIso(),
|
||||||
: Number(torrent.seeding_time ?? 0);
|
};
|
||||||
const seedingSeconds = Number(torrent.seeding_time ?? 0);
|
|
||||||
if (elapsedSeconds < matchingRule.seedLimitSeconds) {
|
const logs: TimerLog[] = [];
|
||||||
continue;
|
|
||||||
|
for (const torrent of torrents) {
|
||||||
|
const tags = normalizeTags(torrent.tags, torrent.category);
|
||||||
|
const matchingRules = rules.filter((rule) => {
|
||||||
|
return rule.tags.some((tag) => tags.includes(tag.toLowerCase()));
|
||||||
|
});
|
||||||
|
if (matchingRules.length === 0) {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
const matchingRule = matchingRules.reduce((best, current) =>
|
||||||
|
current.seedLimitSeconds < best.seedLimitSeconds ? current : best
|
||||||
|
);
|
||||||
|
const addedOnMs = Number(torrent.added_on ?? 0) * 1000;
|
||||||
|
const elapsedSeconds =
|
||||||
|
addedOnMs > 0
|
||||||
|
? Math.max(0, (Date.now() - addedOnMs) / 1000)
|
||||||
|
: Number(torrent.seeding_time ?? 0);
|
||||||
|
const seedingSeconds = Number(torrent.seeding_time ?? 0);
|
||||||
|
if (elapsedSeconds < matchingRule.seedLimitSeconds) {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
try {
|
||||||
|
await qbit.deleteTorrent(torrent.hash, matchingRule.deleteFiles ?? true);
|
||||||
|
} catch (error) {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
const logEntry: TimerLog = {
|
||||||
|
id: randomUUID(),
|
||||||
|
hash: torrent.hash,
|
||||||
|
name: torrent.name,
|
||||||
|
sizeBytes: torrent.size,
|
||||||
|
tracker: torrent.tracker,
|
||||||
|
tags,
|
||||||
|
category: torrent.category,
|
||||||
|
seedingTimeSeconds: seedingSeconds,
|
||||||
|
uploadedBytes: torrent.uploaded ?? 0,
|
||||||
|
deletedAt: nowIso(),
|
||||||
|
};
|
||||||
|
logs.push(logEntry);
|
||||||
|
summary = {
|
||||||
|
totalDeleted: summary.totalDeleted + 1,
|
||||||
|
totalSeededSeconds: summary.totalSeededSeconds + seedingSeconds,
|
||||||
|
totalUploadedBytes: summary.totalUploadedBytes + (torrent.uploaded ?? 0),
|
||||||
|
updatedAt: nowIso(),
|
||||||
|
};
|
||||||
|
emitTimerLog(logEntry);
|
||||||
|
emitTimerSummary(summary);
|
||||||
}
|
}
|
||||||
|
|
||||||
try {
|
if (logs.length > 0) {
|
||||||
await qbit.deleteTorrent(torrent.hash, matchingRule.deleteFiles ?? true);
|
db.timerLogs = [...(db.timerLogs ?? []), ...logs].slice(-MAX_LOGS);
|
||||||
} catch (error) {
|
db.timerSummary = summary;
|
||||||
continue;
|
await writeDb(db);
|
||||||
}
|
}
|
||||||
|
} catch (error) {
|
||||||
const logEntry: TimerLog = {
|
logger.error({ error }, "Timer worker tick failed");
|
||||||
id: randomUUID(),
|
|
||||||
hash: torrent.hash,
|
|
||||||
name: torrent.name,
|
|
||||||
sizeBytes: torrent.size,
|
|
||||||
tracker: torrent.tracker,
|
|
||||||
tags,
|
|
||||||
category: torrent.category,
|
|
||||||
seedingTimeSeconds: seedingSeconds,
|
|
||||||
uploadedBytes: torrent.uploaded ?? 0,
|
|
||||||
deletedAt: nowIso(),
|
|
||||||
};
|
|
||||||
logs.push(logEntry);
|
|
||||||
summary = {
|
|
||||||
totalDeleted: summary.totalDeleted + 1,
|
|
||||||
totalSeededSeconds: summary.totalSeededSeconds + seedingSeconds,
|
|
||||||
totalUploadedBytes: summary.totalUploadedBytes + (torrent.uploaded ?? 0),
|
|
||||||
updatedAt: nowIso(),
|
|
||||||
};
|
|
||||||
emitTimerLog(logEntry);
|
|
||||||
emitTimerSummary(summary);
|
|
||||||
}
|
|
||||||
|
|
||||||
if (logs.length > 0) {
|
|
||||||
db.timerLogs = [...(db.timerLogs ?? []), ...logs].slice(-MAX_LOGS);
|
|
||||||
db.timerSummary = summary;
|
|
||||||
await writeDb(db);
|
|
||||||
}
|
}
|
||||||
}, intervalMs);
|
}, intervalMs);
|
||||||
};
|
};
|
||||||
|
|||||||
Reference in New Issue
Block a user