diff --git a/lib/manager/npm/extract/locked-versions.js b/lib/manager/npm/extract/locked-versions.js index 6665492d2f17fdbf8fc02ad8acdcfad138b01062..e064b705565ac3bf1417316538c7cc05c78d812c 100644 --- a/lib/manager/npm/extract/locked-versions.js +++ b/lib/manager/npm/extract/locked-versions.js @@ -11,7 +11,7 @@ async function getLockedVersions(packageFiles) { for (const packageFile of packageFiles) { const { yarnLock, npmLock, pnpmShrinkwrap } = packageFile; if (yarnLock) { - logger.debug('Found yarnLock'); + logger.trace('Found yarnLock'); if (!lockFileCache[yarnLock]) { logger.debug('Retrieving/parsing ' + yarnLock); lockFileCache[yarnLock] = await getYarnLock(yarnLock); diff --git a/lib/manager/npm/post-update/index.js b/lib/manager/npm/post-update/index.js index 0bde47c6a0ea7249d06baf638eecb26f610e6b91..c4f7e67d22b8be8c3cb574f57a95802463da7af8 100644 --- a/lib/manager/npm/post-update/index.js +++ b/lib/manager/npm/post-update/index.js @@ -63,7 +63,7 @@ function determineLockFileDirs(config, packageFiles) { } for (const p of config.updatedPackageFiles) { - logger.debug(`Checking ${p.name} for lock files`); + logger.trace(`Checking ${p.name} for lock files`); const packageFile = getPackageFile(p.name); // lerna first if (packageFile.lernaDir && (packageFile.npmLock || packageFile.yarnLock)) { @@ -507,25 +507,25 @@ async function getAdditionalFiles(config, packageFiles) { for (const packageFile of packageFiles.npm) { const baseDir = path.dirname(packageFile.packageFile); const filename = upath.join(baseDir, lockFile); - logger.debug('Checking for ' + filename); + logger.trace('Checking for ' + filename); const existingContent = await platform.getFile( filename, config.parentBranch ); if (existingContent) { - logger.debug('Found lock file'); + logger.trace('Found lock file'); const lockFilePath = upath.join(config.tmpDir.path, filename); - logger.debug('Checking against ' + lockFilePath); + logger.trace('Checking against ' + lockFilePath); try { const newContent = await fs.readFile(lockFilePath, 'utf8'); if (newContent !== existingContent) { - logger.debug('File is updated'); + logger.debug('File is updated: ' + lockFilePath); updatedLockFiles.push({ name: filename, contents: newContent, }); } else { - logger.debug('File is unchanged'); + logger.trace('File is unchanged'); } } catch (err) { logger.warn( @@ -534,7 +534,7 @@ async function getAdditionalFiles(config, packageFiles) { ); } } else { - logger.debug('No lock file found'); + logger.trace('No lock file found'); } } } diff --git a/lib/workers/repository/process/extract-update.js b/lib/workers/repository/process/extract-update.js index 474a7c45567dd94f488175e4b6c21d881507e69b..342bbbee3cebc9872244369c35b8e15095d8e818 100644 --- a/lib/workers/repository/process/extract-update.js +++ b/lib/workers/repository/process/extract-update.js @@ -11,7 +11,7 @@ module.exports = { async function extractAndUpdate(config) { logger.debug('extractAndUpdate()'); const packageFiles = await extractAllDependencies(config); - logger.debug({ config: packageFiles }, 'packageFiles'); + logger.trace({ config: packageFiles }, 'packageFiles'); await fetchUpdates(config, packageFiles); logger.debug({ config: packageFiles }, 'packageFiles with updates'); const { branches, branchList } = branchifyUpgrades(config, packageFiles); diff --git a/lib/workers/repository/process/fetch.js b/lib/workers/repository/process/fetch.js index 33276a4716e4e1e1cebad6f4e45a719ffc0fc3a3..06b2e7d2e0c33d3a30ad7ea8acbeb17dcff4d0a3 100644 --- a/lib/workers/repository/process/fetch.js +++ b/lib/workers/repository/process/fetch.js @@ -33,7 +33,7 @@ async function fetchDepUpdates(packageFileConfig, dep) { dep.skipReason = 'disabled'; } else { dep.updates = await getPackageUpdates(manager, depConfig); - logger.debug({ + logger.trace({ packageFile, manager, depName,