Fixed Evil Angel upcoming and actor association issues. Moving from console.log to logger.
This commit is contained in:
@@ -4,6 +4,7 @@ const Promise = require('bluebird');
|
||||
const UrlPattern = require('url-pattern');
|
||||
const moment = require('moment');
|
||||
|
||||
const logger = require('./logger');
|
||||
const knex = require('./knex');
|
||||
const argv = require('./argv');
|
||||
const scrapers = require('./scrapers/scrapers');
|
||||
@@ -252,7 +253,7 @@ async function storeActor(actor, scraped = false, scrapeSuccess = false) {
|
||||
|
||||
await storeSocialLinks(actor.social, actorEntry.id);
|
||||
|
||||
console.log(`Added new entry for actor '${actor.name}'`);
|
||||
logger.info(`Added new entry for actor '${actor.name}'`);
|
||||
|
||||
return actorEntry;
|
||||
}
|
||||
@@ -267,7 +268,7 @@ async function updateActor(actor, scraped = false, scrapeSuccess = false) {
|
||||
|
||||
await storeSocialLinks(actor.social, actor.id);
|
||||
|
||||
console.log(`Updated entry for actor '${actor.name}'`);
|
||||
logger.info(`Updated entry for actor '${actor.name}'`);
|
||||
|
||||
return actorEntry;
|
||||
}
|
||||
@@ -338,7 +339,7 @@ async function scrapeActors(actorNames) {
|
||||
const profile = await mergeProfiles(profiles, actorEntry);
|
||||
|
||||
if (profile === null) {
|
||||
console.log(`Could not find profile for actor '${actorName}'`);
|
||||
logger.warn(`Could not find profile for actor '${actorName}'`);
|
||||
|
||||
if (argv.save) {
|
||||
await updateActor(profile, true, false);
|
||||
@@ -380,7 +381,7 @@ async function scrapeActors(actorNames) {
|
||||
}, newActorEntry.name);
|
||||
}
|
||||
} catch (error) {
|
||||
console.warn(actorName, error);
|
||||
logger.warn(actorName, error);
|
||||
}
|
||||
}, {
|
||||
concurrency: 3,
|
||||
@@ -404,7 +405,7 @@ async function associateActors(mappedActors, releases) {
|
||||
knex('releases_actors').whereIn('release_id', releases.map(release => release.id)),
|
||||
]);
|
||||
|
||||
console.log(actorNames, actorSlugs, existingActorEntries.map(actor => actor.name));
|
||||
console.log(mappedActors);
|
||||
|
||||
const associations = await Promise.map(Object.entries(mappedActors), async ([actorName, releaseIds]) => {
|
||||
try {
|
||||
@@ -421,7 +422,7 @@ async function associateActors(mappedActors, releases) {
|
||||
.some(associationEntry => associationEntry.actor_id === association.actor_id
|
||||
&& associationEntry.release_id === association.release_id));
|
||||
} catch (error) {
|
||||
console.error(actorName, error);
|
||||
logger.error(actorName, error);
|
||||
return null;
|
||||
}
|
||||
});
|
||||
|
||||
@@ -3,6 +3,7 @@
|
||||
const Promise = require('bluebird');
|
||||
const moment = require('moment');
|
||||
|
||||
const logger = require('./logger');
|
||||
const knex = require('./knex');
|
||||
const argv = require('./argv');
|
||||
const whereOr = require('./utils/where-or');
|
||||
@@ -319,8 +320,6 @@ async function storeReleaseAssets(releases) {
|
||||
|
||||
await createMediaDirectory('releases', subpath);
|
||||
|
||||
console.log(release.id);
|
||||
|
||||
try {
|
||||
// don't use Promise.all to prevent concurrency issues with duplicate detection
|
||||
if (release.poster) {
|
||||
@@ -379,7 +378,7 @@ async function storeRelease(release) {
|
||||
|
||||
if (updatedRelease) {
|
||||
await associateTags(release, updatedRelease.id);
|
||||
console.log(`Updated release "${release.title}" (${existingRelease.id}, ${release.site.name})`);
|
||||
logger.info(`Updated release "${release.title}" (${existingRelease.id}, ${release.site.name})`);
|
||||
}
|
||||
|
||||
await associateTags(release, existingRelease.id);
|
||||
@@ -393,7 +392,7 @@ async function storeRelease(release) {
|
||||
|
||||
await associateTags(release, releaseEntry.id);
|
||||
|
||||
console.log(`Stored release "${release.title}" (${releaseEntry.id}, ${release.site.name})`);
|
||||
logger.info(`Stored release "${release.title}" (${releaseEntry.id}, ${release.site.name})`);
|
||||
|
||||
return releaseEntry.id;
|
||||
}
|
||||
@@ -410,7 +409,7 @@ async function storeReleases(releases) {
|
||||
...releaseWithChannelSite,
|
||||
};
|
||||
} catch (error) {
|
||||
console.error(error);
|
||||
logger.error(error);
|
||||
|
||||
return null;
|
||||
}
|
||||
|
||||
@@ -4,6 +4,7 @@ const Promise = require('bluebird');
|
||||
const moment = require('moment');
|
||||
|
||||
const argv = require('./argv');
|
||||
const logger = require('./logger');
|
||||
const knex = require('./knex');
|
||||
const { fetchIncludedSites } = require('./sites');
|
||||
const scrapers = require('./scrapers/scrapers');
|
||||
@@ -42,7 +43,7 @@ async function scrapeUniqueReleases(scraper, site, afterDate = getAfterDate(), a
|
||||
.filter(release => !duplicateReleaseIds.has(String(release.entryId)) // release is already in database
|
||||
&& moment(release.date).isAfter(afterDate)); // release is older than specified date limit
|
||||
|
||||
console.log(`\x1b[90m${site.name}: Scraped page ${page}, ${uniqueReleases.length} unique recent releases\x1b[0m`);
|
||||
logger.info(`\x1b[90m${site.name}: Scraped page ${page}, ${uniqueReleases.length} unique recent releases\x1b[0m`);
|
||||
|
||||
if (
|
||||
uniqueReleases.length > 0
|
||||
@@ -78,7 +79,7 @@ async function deepFetchReleases(baseReleases) {
|
||||
deep: true,
|
||||
};
|
||||
} catch (error) {
|
||||
console.error(error);
|
||||
logger.error(error);
|
||||
|
||||
return {
|
||||
...release,
|
||||
@@ -99,7 +100,7 @@ async function scrapeSiteReleases(scraper, site) {
|
||||
scrapeUpcomingReleases(scraper, site), // fetch basic release info from upcoming overview
|
||||
]);
|
||||
|
||||
console.log(`${site.name}: Found ${newReleases.length} recent releases, ${upcomingReleases.length} upcoming releases`);
|
||||
logger.info(`${site.name}: Found ${newReleases.length} recent releases, ${upcomingReleases.length} upcoming releases`);
|
||||
|
||||
const baseReleases = [...newReleases, ...upcomingReleases];
|
||||
|
||||
@@ -118,7 +119,7 @@ async function scrapeSites() {
|
||||
const scraper = scrapers.releases[site.slug] || scrapers.releases[site.network.slug];
|
||||
|
||||
if (!scraper) {
|
||||
console.warn(`No scraper found for '${site.name}' (${site.slug})`);
|
||||
logger.warn(`No scraper found for '${site.name}' (${site.slug})`);
|
||||
return [];
|
||||
}
|
||||
|
||||
@@ -126,10 +127,10 @@ async function scrapeSites() {
|
||||
return await scrapeSiteReleases(scraper, site);
|
||||
} catch (error) {
|
||||
if (argv.debug) {
|
||||
console.error(`${site.name}: Failed to scrape releases`, error);
|
||||
logger.error(`${site.name}: Failed to scrape releases`, error);
|
||||
}
|
||||
|
||||
console.warn(`${site.id}: Failed to scrape releases`);
|
||||
logger.warn(`${site.id}: Failed to scrape releases`);
|
||||
|
||||
return [];
|
||||
}
|
||||
|
||||
@@ -112,7 +112,7 @@ async function fetchLatest(site, page = 1, upcoming = false) {
|
||||
requests: [
|
||||
{
|
||||
indexName: 'all_scenes',
|
||||
params: `query=&hitsPerPage=36&maxValuesPerFacet=100&page=${page - 1}&facetFilters=[["lesbian:"],["bisex:"],["shemale:"],["upcoming:${upcoming ? '' : 0}"]]`,
|
||||
params: `query=&hitsPerPage=36&maxValuesPerFacet=100&page=${page - 1}&facetFilters=[["lesbian:"],["bisex:"],["shemale:"],["upcoming:${upcoming ? 1 : 0}"]]`,
|
||||
},
|
||||
],
|
||||
}, {
|
||||
|
||||
@@ -1,6 +1,7 @@
|
||||
'use strict';
|
||||
|
||||
const knex = require('../knex');
|
||||
const logger = require('../logger');
|
||||
|
||||
async function upsert(table, items, identifier = ['id'], _knex) {
|
||||
const identifiers = Array.isArray(identifier) ? identifier : [identifier];
|
||||
@@ -28,8 +29,8 @@ async function upsert(table, items, identifier = ['id'], _knex) {
|
||||
});
|
||||
|
||||
if (knex) {
|
||||
console.log(`${table}: Inserting ${insert.length}`);
|
||||
console.log(`${table}: Updating ${update.length}`);
|
||||
logger.debug(`${table}: Inserting ${insert.length}`);
|
||||
logger.debug(`${table}: Updating ${update.length}`);
|
||||
|
||||
const [inserted, updated] = await Promise.all([
|
||||
knex(table).returning('*').insert(insert),
|
||||
|
||||
Reference in New Issue
Block a user