diff --git a/packages/cron/src/jobs/pins-backup.js b/packages/cron/src/jobs/pins-backup.js index 9c591ac324..146132c221 100644 --- a/packages/cron/src/jobs/pins-backup.js +++ b/packages/cron/src/jobs/pins-backup.js @@ -32,7 +32,7 @@ export default class Backup { this.BLOCK_TIMEOUT = 1000 * 30 // timeout if we don't receive a block after 30s this.REPORT_INTERVAL = 1000 * 60 // log download progress every minute this.MAX_UPLOAD_DAG_SIZE = 1024 * 1024 * 1024 * 32 // We don't limit in psa pin transfers in this case, but we still want to log if we have larger pin requests. - this.log = debug('backup:pins') + this.log = console.log this.env = env this.LIMIT = env.QUERY_LIMIT !== undefined ? env.QUERY_LIMIT : 10000 this.GET_PINNED_PINS_QUERY = ` diff --git a/packages/cron/test/pins-backup.spec.js b/packages/cron/test/pins-backup.spec.js index bf863cb433..2b74fcc917 100644 --- a/packages/cron/test/pins-backup.spec.js +++ b/packages/cron/test/pins-backup.spec.js @@ -12,6 +12,8 @@ import { CID } from 'multiformats' import { sha256 } from 'multiformats/hashes/sha2' import * as pb from '@ipld/dag-pb' +process.env.DEBUG = 'backup:pins' + const env = { ...process.env, DEBUG: 'backup:pins', @@ -72,7 +74,7 @@ async function createDagsAndRequests (files, authKey, cluster, dbClient, { return fileContentMap } -describe('cron - pins backup', () => { +describe.only('cron - pins backup', () => { let user, userId, authKey, dbClient, cluster, backup, rwPg, roPg before(async () => {