From 9d89f987ba0748bbf978d17b815040d316b19620 Mon Sep 17 00:00:00 2001 From: MohamedBassem Date: Sun, 26 May 2024 10:18:42 +0000 Subject: fix(crawler): Only update the database if full page archival is enabled --- apps/workers/crawlerWorker.ts | 38 +++++++++++++++++++------------------- 1 file changed, 19 insertions(+), 19 deletions(-) (limited to 'apps') diff --git a/apps/workers/crawlerWorker.ts b/apps/workers/crawlerWorker.ts index 87632019..f2e0e4a8 100644 --- a/apps/workers/crawlerWorker.ts +++ b/apps/workers/crawlerWorker.ts @@ -388,9 +388,6 @@ async function archiveWebpage( userId: string, jobId: string, ) { - if (!serverConfig.crawler.fullPageArchive) { - return; - } logger.info(`[Crawler][${jobId}] Will attempt to archive page ...`); const urlParsed = new URL(url); const baseUrl = `${urlParsed.protocol}//${urlParsed.host}`; @@ -499,22 +496,25 @@ async function runCrawler(job: Job) { }); // Do the archival as a separate last step as it has the potential for failure - const fullPageArchiveAssetId = await archiveWebpage( - htmlContent, - browserUrl, - userId, - jobId, - ); - await db - .update(bookmarkLinks) - .set({ - fullPageArchiveAssetId, - }) - .where(eq(bookmarkLinks.id, bookmarkId)); - - if (oldFullPageArchiveAssetId) { - deleteAsset({ userId, assetId: oldFullPageArchiveAssetId }).catch( - () => ({}), + if (serverConfig.crawler.fullPageArchive) { + const fullPageArchiveAssetId = await archiveWebpage( + htmlContent, + browserUrl, + userId, + jobId, ); + + await db + .update(bookmarkLinks) + .set({ + fullPageArchiveAssetId, + }) + .where(eq(bookmarkLinks.id, bookmarkId)); + + if (oldFullPageArchiveAssetId) { + deleteAsset({ userId, assetId: oldFullPageArchiveAssetId }).catch( + () => ({}), + ); + } } } -- cgit v1.2.3-70-g09d2