From 70d6e9dcae2c5088abb555da1a3273734e6fca6d Mon Sep 17 00:00:00 2001 From: Hongbo Wu Date: Thu, 29 Aug 2024 18:18:17 +0800 Subject: [PATCH] reduce logs --- packages/api/src/jobs/trigger_rule.ts | 1 - packages/api/src/queue-processor.ts | 20 ----------------- packages/api/src/resolvers/article/index.ts | 13 ----------- packages/api/src/services/features.ts | 4 ---- packages/api/src/services/library_item.ts | 15 ------------- packages/api/src/services/user.ts | 4 ---- .../api/src/services/user_personalization.ts | 4 ---- packages/api/src/utils/rate_limit.ts | 2 -- packages/content-fetch/src/analytics.ts | 2 -- packages/content-fetch/src/request_handler.ts | 7 +----- packages/content-fetch/src/worker.ts | 22 +------------------ .../content-handler/src/content-handler.ts | 2 -- packages/import-handler/src/csv.ts | 2 -- packages/inbound-email-handler/src/index.ts | 2 -- packages/pdf-handler/src/index.ts | 3 --- packages/text-to-speech/src/index.ts | 3 --- packages/thumbnail-handler/src/index.ts | 3 --- 17 files changed, 2 insertions(+), 107 deletions(-) diff --git a/packages/api/src/jobs/trigger_rule.ts b/packages/api/src/jobs/trigger_rule.ts index 4ea469706..1033c7e47 100644 --- a/packages/api/src/jobs/trigger_rule.ts +++ b/packages/api/src/jobs/trigger_rule.ts @@ -289,7 +289,6 @@ export const triggerRule = async (jobData: TriggerRuleJobData) => { // get rules by calling api const rules = await findEnabledRules(userId, ruleEventType) if (rules.length === 0) { - console.log('No rules found') return false } diff --git a/packages/api/src/queue-processor.ts b/packages/api/src/queue-processor.ts index cdc7e9817..556809a64 100644 --- a/packages/api/src/queue-processor.ts +++ b/packages/api/src/queue-processor.ts @@ -362,26 +362,6 @@ const main = async () => { await setupCronJobs() - const queueEvents = new QueueEvents(BACKEND_QUEUE_NAME, { - connection: workerRedisClient, - }) - - queueEvents.on('added', async (job) => { - console.log('added job: ', job.jobId, job.name) - }) - - queueEvents.on('removed', async (job) => { - console.log('removed job: ', job.jobId) - }) - - queueEvents.on('completed', async (job) => { - console.log('completed job: ', job.jobId) - }) - - queueEvents.on('failed', async (job) => { - console.log('failed job: ', job.jobId) - }) - workerRedisClient.on('error', (error) => { console.trace('[queue-processor]: redis worker error', { error }) }) diff --git a/packages/api/src/resolvers/article/index.ts b/packages/api/src/resolvers/article/index.ts index 037f3958e..adde391a4 100644 --- a/packages/api/src/resolvers/article/index.ts +++ b/packages/api/src/resolvers/article/index.ts @@ -318,13 +318,6 @@ export const createArticleResolver = authorized< savedAt, }) - log.info('New article saving', { - parsedArticle: Object.assign({}, libraryItemToSave, { - readableContent: undefined, - originalContent: undefined, - }), - }) - if (uploadFileId) { const uploadFileData = await setFileUploadComplete(uploadFileId) if (!uploadFileData || !uploadFileData.id || !uploadFileData.fileName) { @@ -461,12 +454,6 @@ export const setBookmarkArticleResolver = authorized< }, }) - log.info('Article unbookmarked', { - item: Object.assign({}, deletedLibraryItem, { - readableContent: undefined, - originalContent: undefined, - }), - }) // Make sure article.id instead of userArticle.id has passed. We use it for cache updates return { bookmarkedArticle: deletedLibraryItem, diff --git a/packages/api/src/services/features.ts b/packages/api/src/services/features.ts index 307d3cc59..8c6de236f 100644 --- a/packages/api/src/services/features.ts +++ b/packages/api/src/services/features.ts @@ -206,8 +206,6 @@ export const userDigestEligible = async (uid: string): Promise => { const featuresCacheKey = (userId: string) => `cache:features:${userId}` export const getFeaturesCache = async (userId: string) => { - logger.debug('getFeaturesCache', { userId }) - const cachedFeatures = await redisDataSource.redisClient?.get( featuresCacheKey(userId) ) @@ -221,8 +219,6 @@ export const getFeaturesCache = async (userId: string) => { export const setFeaturesCache = async (userId: string, features: Feature[]) => { const value = JSON.stringify(features) - logger.debug('setFeaturesCache', { userId, value }) - return redisDataSource.redisClient?.set( featuresCacheKey(userId), value, diff --git a/packages/api/src/services/library_item.ts b/packages/api/src/services/library_item.ts index 9a945a3a4..cd24966ce 100644 --- a/packages/api/src/services/library_item.ts +++ b/packages/api/src/services/library_item.ts @@ -1749,11 +1749,6 @@ const totalCountCacheKey = (userId: string, args: SearchArgs) => { } export const getCachedTotalCount = async (userId: string, args: SearchArgs) => { - logger.debug('Getting cached total count:', { - userId, - args, - }) - const cacheKey = totalCountCacheKey(userId, args) const cachedCount = await redisDataSource.redisClient?.get(cacheKey) if (!cachedCount) { @@ -1770,11 +1765,6 @@ export const setCachedTotalCount = async ( ) => { const cacheKey = totalCountCacheKey(userId, args) - logger.debug('Setting cached total count:', { - cacheKey, - count, - }) - await redisDataSource.redisClient?.set(cacheKey, count, 'EX', 600) } @@ -1785,10 +1775,5 @@ export const deleteCachedTotalCount = async (userId: string) => { return } - logger.debug('Deleting keys:', { - keys, - userId, - }) - await redisDataSource.redisClient?.del(keys) } diff --git a/packages/api/src/services/user.ts b/packages/api/src/services/user.ts index c5dc336b9..3f7ed7d6c 100644 --- a/packages/api/src/services/user.ts +++ b/packages/api/src/services/user.ts @@ -161,8 +161,6 @@ export const findUserAndPersonalization = async (id: string) => { const userCacheKey = (id: string) => `cache:user:${id}` export const getCachedUser = async (id: string) => { - logger.debug(`Getting user from cache: ${id}`) - const user = await redisDataSource.redisClient?.get(userCacheKey(id)) if (!user) { return undefined @@ -172,8 +170,6 @@ export const getCachedUser = async (id: string) => { } export const cacheUser = async (user: User) => { - logger.debug(`Caching user: ${user.id}`) - await redisDataSource.redisClient?.set( userCacheKey(user.id), JSON.stringify(user), diff --git a/packages/api/src/services/user_personalization.ts b/packages/api/src/services/user_personalization.ts index b6c0bfe2f..c38d2f51c 100644 --- a/packages/api/src/services/user_personalization.ts +++ b/packages/api/src/services/user_personalization.ts @@ -177,8 +177,6 @@ const userDefaultShortcuts = async (userId: string): Promise => { const shortcutsCacheKey = (userId: string) => `cache:shortcuts:${userId}` export const getShortcutsCache = async (userId: string) => { - logger.debug(`Getting shortcuts from cache: ${userId}`) - const cachedShortcuts = await redisDataSource.redisClient?.get( shortcutsCacheKey(userId) ) @@ -189,8 +187,6 @@ export const getShortcutsCache = async (userId: string) => { } export const cacheShortcuts = async (userId: string, shortcuts: Shortcut[]) => { - logger.debug(`Caching shortcuts: ${userId}`) - await redisDataSource.redisClient?.set( shortcutsCacheKey(userId), JSON.stringify(shortcuts), diff --git a/packages/api/src/utils/rate_limit.ts b/packages/api/src/utils/rate_limit.ts index 27675dc6e..b185209b6 100644 --- a/packages/api/src/utils/rate_limit.ts +++ b/packages/api/src/utils/rate_limit.ts @@ -33,7 +33,6 @@ export const apiLimiter = rateLimit({ const claims = await getClaimsByToken(token) return claims ? 60 : 15 } catch (e) { - console.log('non-authenticated request') return 15 } }, @@ -53,7 +52,6 @@ export const apiHourLimiter = rateLimit({ const claims = await getClaimsByToken(token) return claims ? 600 : 150 } catch (e) { - console.log('non-authenticated request') return 150 } }, diff --git a/packages/content-fetch/src/analytics.ts b/packages/content-fetch/src/analytics.ts index 6b422f446..1c6c16269 100644 --- a/packages/content-fetch/src/analytics.ts +++ b/packages/content-fetch/src/analytics.ts @@ -31,8 +31,6 @@ class PostHogClient implements AnalyticClient { return } - - console.log('analytics', { userIds, result, properties }) } } diff --git a/packages/content-fetch/src/request_handler.ts b/packages/content-fetch/src/request_handler.ts index e5ff9bc97..8ef69acbe 100644 --- a/packages/content-fetch/src/request_handler.ts +++ b/packages/content-fetch/src/request_handler.ts @@ -299,12 +299,7 @@ export const processFetchContentJob = async ( } if (fetchResult.content && !NO_CACHE_URLS.includes(url)) { - const cacheResult = await cacheFetchResult( - redisDataSource, - key, - fetchResult - ) - console.log('cache result', cacheResult) + await cacheFetchResult(redisDataSource, key, fetchResult) } } diff --git a/packages/content-fetch/src/worker.ts b/packages/content-fetch/src/worker.ts index 1a3cef64e..f209b279b 100644 --- a/packages/content-fetch/src/worker.ts +++ b/packages/content-fetch/src/worker.ts @@ -1,5 +1,5 @@ import { RedisDataSource } from '@omnivore/utils' -import { Job, Queue, QueueEvents, RedisClient, Worker } from 'bullmq' +import { Job, Queue, RedisClient, Worker } from 'bullmq' import { JobData, processFetchContentJob } from './request_handler' export const QUEUE = 'omnivore-content-fetch-queue' @@ -53,25 +53,5 @@ export const createWorker = ( console.error('worker error:', err) }) - const queueEvents = new QueueEvents(queueName, { - connection: redisDataSource.queueRedisClient, - }) - - queueEvents.on('added', (job) => { - console.log('added job:', job.jobId, job.name) - }) - - queueEvents.on('removed', (job) => { - console.log('removed job:', job.jobId) - }) - - queueEvents.on('completed', (job) => { - console.log('completed job:', job.jobId) - }) - - queueEvents.on('failed', (job) => { - console.log('failed job:', job.jobId) - }) - return worker } diff --git a/packages/content-handler/src/content-handler.ts b/packages/content-handler/src/content-handler.ts index 9c935b5fb..c5049cead 100644 --- a/packages/content-handler/src/content-handler.ts +++ b/packages/content-handler/src/content-handler.ts @@ -159,8 +159,6 @@ export abstract class ContentHandler { html, headers, }: NewsletterInput): Promise { - console.log('handleNewsletter', from, to, subject, headers) - if (!from || !html || !subject || !to) { console.log('invalid newsletter email') throw new Error('invalid newsletter email') diff --git a/packages/import-handler/src/csv.ts b/packages/import-handler/src/csv.ts index 7acdff4ce..e76eac60e 100644 --- a/packages/import-handler/src/csv.ts +++ b/packages/import-handler/src/csv.ts @@ -13,8 +13,6 @@ const parseLabels = (labels: string): string[] => { // labels follows format: "[""label1"",""label2""]" return JSON.parse(labels) as string[] } catch (error) { - console.debug('invalid labels format', labels) - // labels follows format: "[label1,label2]" return labels .slice(1, -1) diff --git a/packages/inbound-email-handler/src/index.ts b/packages/inbound-email-handler/src/index.ts index 5b7f4d33f..e523baa1b 100644 --- a/packages/inbound-email-handler/src/index.ts +++ b/packages/inbound-email-handler/src/index.ts @@ -55,10 +55,8 @@ export const inboundEmailHandler = Sentry.GCPFunction.wrapHttpFunction( console.log('no data or name for ', part) } } - console.log('parsed', parsed) const headers = parseHeaders(parsed.headers) - console.log('headers', headers) // original sender email address const from = parsed['from'] diff --git a/packages/pdf-handler/src/index.ts b/packages/pdf-handler/src/index.ts index 7e4ab43c6..b36acf185 100644 --- a/packages/pdf-handler/src/index.ts +++ b/packages/pdf-handler/src/index.ts @@ -45,7 +45,6 @@ const getDocumentUrl = async ( const [url] = await file.getSignedUrl(options) return new URL(url) } catch (e) { - console.debug('error getting signed url', e) return undefined } } @@ -101,8 +100,6 @@ export const pdfHandler = Sentry.GCPFunction.wrapHttpFunction( return res.send('ok') } - console.log('handling pdf data', data) - let content, title, author, diff --git a/packages/text-to-speech/src/index.ts b/packages/text-to-speech/src/index.ts index a933a940e..211bc1e00 100644 --- a/packages/text-to-speech/src/index.ts +++ b/packages/text-to-speech/src/index.ts @@ -214,7 +214,6 @@ export const textToSpeechHandler = Sentry.GCPFunction.wrapHttpFunction( export const textToSpeechStreamingHandler = Sentry.GCPFunction.wrapHttpFunction( async (req, res) => { - console.log('Text to speech steaming request body:', req.body) if (!process.env.JWT_SECRET) { console.error('JWT_SECRET not exists') return res.status(500).send({ errorCodes: 'JWT_SECRET_NOT_EXISTS' }) @@ -314,7 +313,6 @@ export const textToSpeechStreamingHandler = Sentry.GCPFunction.wrapHttpFunction( // check if audio file already exists const [exists] = await audioFile.exists() if (exists) { - console.debug('Audio file already exists') ;[audioData] = await audioFile.download() const [speechMarksExists] = await speechMarksFile.exists() if (speechMarksExists) { @@ -341,7 +339,6 @@ export const textToSpeechStreamingHandler = Sentry.GCPFunction.wrapHttpFunction( }) } - console.debug('saving audio file') // upload audio data to GCS await audioFile.save(audioData) // upload speech marks to GCS diff --git a/packages/thumbnail-handler/src/index.ts b/packages/thumbnail-handler/src/index.ts index ec4d69d29..8950d10c1 100644 --- a/packages/thumbnail-handler/src/index.ts +++ b/packages/thumbnail-handler/src/index.ts @@ -290,7 +290,6 @@ export const thumbnailHandler = Sentry.GCPFunction.wrapHttpFunction( const token = req.headers.cookie?.split('auth=')[1] if (!token) { - console.debug('no token') return res.status(401).send('UNAUTHORIZED') } let uid = '' @@ -300,12 +299,10 @@ export const thumbnailHandler = Sentry.GCPFunction.wrapHttpFunction( } uid = decoded.uid } catch (e) { - console.debug(e) return res.status(401).send('UNAUTHORIZED') } if (!isThumbnailRequest(req.body)) { - console.debug('bad request') return res.status(400).send('BAD_REQUEST') }