Added preflight method to scrapers. Added Assylum.

This commit is contained in:
ThePendulum 2020-02-12 01:54:54 +01:00
parent bec26ee072
commit 2f70de8e11
14 changed files with 185 additions and 19 deletions

Binary file not shown.

After

Width:  |  Height:  |  Size: 41 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 1.6 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 16 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 96 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 41 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 81 KiB

View File

@ -26,6 +26,12 @@ const networks = [
url: 'https://www.adulttime.com',
description: 'Adult Time is a premium streaming service for adults! Watch adult movies, series, and channels from the top names in the industry.',
},
{
slug: 'assylum',
name: 'Assylum',
url: 'https://www.assylum.com',
description: 'At Assylum, submissive girls get dominated with rough anal sex, ass to mouth, hard BDSM, and sexual humiliation and degradation.',
},
{
slug: 'babes',
name: 'Babes',

View File

@ -294,6 +294,27 @@ const sites = [
description: 'LadyGonzo.com is a new Adult Time porn series featuring Joanna Angel shooting hardcore sex and gonzo porn movies the way she\'d like to see it!',
network: 'adulttime',
},
// ASSYLUM
{
slug: 'assylum',
name: 'Assylum',
url: 'https://www.assylum.com',
description: 'At Assylum, submissive girls get dominated with rough anal sex, ass to mouth, hard BDSM, and sexual humiliation and degradation.',
network: 'assylum',
parameters: {
a: 68,
},
},
{
slug: 'slavemouth',
name: 'Slave Mouth',
url: 'https://www.slavemouth.com',
description: 'Submissive girls get their mouths punished hard by Dr. Mercies, with facefucking, gagging, frozen cum bukkake, face bondage, ass eating, and sexual degradation.',
network: 'assylum',
parameters: {
a: 183,
},
},
// BABES
{
name: 'Babes',
@ -1255,8 +1276,8 @@ const sites = [
network: 'digitalplayground',
},
{
slug: 'rawcuts',
name: 'Raw Cuts',
slug: 'rawcut',
name: 'Raw Cut',
url: 'https://www.digitalplayground.com/scenes?site=208',
description: '',
network: 'digitalplayground',

View File

@ -362,7 +362,9 @@ async function scrapeActors(actorNames) {
const actorEntry = await knex('actors').where({ slug: actorSlug }).first();
const sources = argv.sources || config.profiles || Object.keys(scrapers.actors);
const profiles = await Promise.map(sources, async (source) => {
const finalSources = argv.withReleases ? sources.flat() : sources; // ignore race-to-success grouping when scenes are requested
const profiles = await Promise.map(finalSources, async (source) => {
// const [scraperSlug, scraper] = source;
const profileScrapers = [].concat(source).map(slug => ({ scraperSlug: slug, scraper: scrapers.actors[slug] }));

View File

@ -33,7 +33,7 @@ async function findSite(url, release) {
return null;
}
async function scrapeRelease(source, basicRelease = null, type = 'scene') {
async function scrapeRelease(source, basicRelease = null, type = 'scene', preflight) {
// profile scraper may return either URLs or pre-scraped scenes
const sourceIsUrl = typeof source === 'string';
const url = sourceIsUrl ? source : source?.url;
@ -64,8 +64,8 @@ async function scrapeRelease(source, basicRelease = null, type = 'scene') {
}
const scrapedRelease = type === 'scene'
? await scraper.fetchScene(url, site, release)
: await scraper.fetchMovie(url, site, release);
? await scraper.fetchScene(url, site, release, preflight)
: await scraper.fetchMovie(url, site, release, preflight);
return {
...release,
@ -77,8 +77,8 @@ async function scrapeRelease(source, basicRelease = null, type = 'scene') {
};
}
async function scrapeReleases(sources, release = null, type = 'scene') {
const scrapedReleases = await Promise.map(sources, async source => scrapeRelease(source, release, type), {
async function scrapeReleases(sources, release = null, type = 'scene', preflight = null) {
const scrapedReleases = await Promise.map(sources, async source => scrapeRelease(source, release, type, preflight), {
concurrency: 5,
});

View File

@ -37,12 +37,12 @@ async function findDuplicateReleaseIds(latestReleases, accReleases) {
.concat(accReleases.map(release => String(release.entryId))));
}
async function scrapeUniqueReleases(scraper, site, afterDate = getAfterDate(), accReleases = [], page = argv.page) {
async function scrapeUniqueReleases(scraper, site, preflight, afterDate = getAfterDate(), accReleases = [], page = argv.page) {
if (!argv.latest || !scraper.fetchLatest) {
return [];
}
const latestReleases = await scraper.fetchLatest(site, page);
const latestReleases = await scraper.fetchLatest(site, page, preflight);
if (latestReleases.length === 0) {
return accReleases;
@ -68,7 +68,7 @@ async function scrapeUniqueReleases(scraper, site, afterDate = getAfterDate(), a
|| (argv.last && accReleases.length + uniqueReleases.length < argv.last))
) {
// oldest release on page is newer that specified date range, or latest count has not yet been met, fetch next page
return scrapeUniqueReleases(scraper, site, afterDate, accReleases.concat(uniqueReleasesWithSite), page + 1);
return scrapeUniqueReleases(scraper, site, preflight, afterDate, accReleases.concat(uniqueReleasesWithSite), page + 1);
}
if (argv.latest && uniqueReleases.length >= argv.latest) {
@ -82,9 +82,9 @@ async function scrapeUniqueReleases(scraper, site, afterDate = getAfterDate(), a
return accReleases.concat(uniqueReleasesWithSite).slice(0, argv.nullDateLimit);
}
async function scrapeUpcomingReleases(scraper, site) {
async function scrapeUpcomingReleases(scraper, site, preflight) {
if (argv.upcoming && scraper.fetchUpcoming) {
const upcomingReleases = await scraper.fetchUpcoming(site);
const upcomingReleases = await scraper.fetchUpcoming(site, 1, preflight);
return upcomingReleases
? upcomingReleases.map(release => ({ ...release, site, upcoming: true }))
@ -94,11 +94,11 @@ async function scrapeUpcomingReleases(scraper, site) {
return [];
}
async function deepFetchReleases(baseReleases) {
async function deepFetchReleases(baseReleases, preflight) {
return Promise.map(baseReleases, async (release) => {
if (release.url || (release.path && release.site)) {
try {
const fullRelease = await scrapeRelease(release.url || release.path, release, 'scene');
const fullRelease = await scrapeRelease(release.url || release.path, release, 'scene', preflight);
if (fullRelease) {
return {
@ -128,9 +128,11 @@ async function deepFetchReleases(baseReleases) {
}
async function scrapeSiteReleases(scraper, site) {
const preflight = await scraper.preflight?.(site);
const [newReleases, upcomingReleases] = await Promise.all([
scrapeUniqueReleases(scraper, site), // fetch basic release info from scene overview
scrapeUpcomingReleases(scraper, site), // fetch basic release info from upcoming overview
scrapeUniqueReleases(scraper, site, preflight), // fetch basic release info from scene overview
scrapeUpcomingReleases(scraper, site, preflight), // fetch basic release info from upcoming overview
]);
if (argv.upcoming) {
@ -141,7 +143,7 @@ async function scrapeSiteReleases(scraper, site) {
if (argv.deep) {
// follow URL for every release
return deepFetchReleases(baseReleases);
return deepFetchReleases(baseReleases, preflight);
}
return baseReleases;

131
src/scrapers/assylum.js Normal file
View File

@ -0,0 +1,131 @@
'use strict';
const { get, geta, ctxa } = require('../utils/q');
function extractActors(actorString) {
return actorString
?.replace(/.*:|\(.*\)|\d+(-|\s)year(-|\s)old|nurses?|tangled/ig, '') // remove Patient:, (date) and other nonsense
.split(/\band\b|\bvs\b|\/|,|&/ig)
.map(actor => actor.trim())
.filter(actor => !!actor && !/\banal\b|\bschool\b|\bgamer\b|\breturn\b|\bfor\b|\bare\b|\bpart\b|realdoll|bimbo|p\d+/ig.test(actor))
|| [];
}
function matchActors(actorString, models) {
return models
.filter(model => new RegExp(model.name, 'i')
.test(actorString));
}
function scrapeLatest(scenes, site, models) {
return scenes.map(({ q, qd, qu, qi }) => {
const release = {};
const pathname = qu('a.itemimg').slice(1);
[release.entryId] = pathname.split('/').slice(-1);
release.url = `${site.url}${pathname}`;
release.title = q('.itemimg img', 'alt') || q('h4 a', true);
release.description = q('.mas_longdescription', true);
release.date = qd('.movie_info2', 'MM/DD/YY', /\d{2}\/\d{2}\/\d{2}/);
const actorString = q('.mas_description', true);
const actors = matchActors(actorString, models);
if (actors.length > 0) release.actors = actors;
else release.actors = extractActors(actorString);
const posterPath = qi('.itemimg img');
release.poster = `${site.url}/${posterPath}`;
return release;
});
}
function scrapeScene({ html, q, qa, qd, qis }, url, site, models) {
const release = { url };
[release.entryId] = url.split('/').slice(-1);
release.title = q('.mas_title', true);
release.description = q('.mas_longdescription', true);
release.date = qd('.mas_description', 'MMMM DD, YYYY', /\w+ \d{1,2}, \d{4}/);
const actorString = q('.mas_description', true).replace(/\w+ \d{1,2}, \d{4}/, '');
const actors = matchActors(actorString, models);
if (actors.length > 0) release.actors = actors;
else release.actors = extractActors(actorString);
release.tags = qa('.tags a', true);
release.photos = qis('.stills img').map(photoPath => `${site.url}/${photoPath}`);
const posterIndex = 'splash:';
const poster = html.slice(html.indexOf('faceimages/', posterIndex), html.indexOf('.jpg', posterIndex) + 4);
if (poster) release.poster = `${site.url}/${poster}`;
const trailerIndex = html.indexOf('video/mp4');
const trailer = html.slice(html.indexOf('/content', trailerIndex), html.indexOf('.mp4', trailerIndex) + 4);
if (trailer) release.trailer = { src: `${site.url}${trailer}` };
return release;
}
function extractModels({ el }, site) {
const models = ctxa(el, '.item');
return models.map(({ q, qu }) => {
const actor = { gender: 'female' };
const avatar = q('.itemimg img');
actor.avatar = `${site.url}/${avatar.src}`;
actor.name = avatar.alt
.split(':').slice(-1)[0]
.replace(/xtreme girl|nurse/ig, '')
.trim();
const actorPath = qu('.itemimg');
actor.url = `${site.url}${actorPath.slice(1)}`;
return actor;
});
}
async function fetchModels(site, page = 1, accModels = []) {
const url = `${site.url}/?models/${page}`;
const qModels = await get(url);
if (qModels) {
const models = extractModels(qModels, site);
const nextPage = qModels.qa('.pagenumbers', true)
.map(pageX => Number(pageX))
.filter(Boolean) // remove << and >>
.includes(page + 1);
if (nextPage) {
return fetchModels(site, page + 1, accModels.concat(models));
}
return accModels.concat(models, { name: 'Dr. Gray' });
}
return [];
}
async function fetchLatest(site, page = 1, models) {
const url = `${site.url}/show.php?a=${site.parameters.a}_${page}`;
const qLatest = await geta(url, '.item');
return qLatest && scrapeLatest(qLatest, site, models);
}
async function fetchScene(url, site, release, preflight) {
const models = preflight || await fetchModels(site);
const qScene = await get(url);
return qScene && scrapeScene(qScene, url, site, models);
}
module.exports = {
fetchLatest,
fetchScene,
preflight: fetchModels,
};

View File

@ -1,6 +1,7 @@
'use strict';
const adulttime = require('./adulttime');
const assylum = require('./assylum');
const babes = require('./babes');
const bang = require('./bang');
const bangbros = require('./bangbros');
@ -52,10 +53,11 @@ const xempire = require('./xempire');
module.exports = {
releases: {
adulttime,
'21naturals': naturals,
'21sextreme': sextreme,
'21sextury': sextury,
adulttime,
assylum,
babes,
bang,
bangbros,

View File

@ -149,6 +149,7 @@ const funcs = {
qurl,
qurls,
qa: qall,
qs: qall,
qd: qdate,
qi: qimage,
qis: qimages,
@ -175,6 +176,7 @@ function init(element, window) {
return {
element,
el: element,
html: element.outerHTML || element.body.outerHTML,
...(window && {
window,