Compare commits

..

No commits in common. "2e0fba3de986edf7c25455cbc9b1eaf1d5e0d35a" and "f5939c81d37c5bff57367616f3f3507350e73a69" have entirely different histories.

14 changed files with 11 additions and 90 deletions

View File

@ -171,10 +171,6 @@ module.exports = {
'gloryholesecrets', 'gloryholesecrets',
'aziani', 'aziani',
'legalporno', 'legalporno',
[
'firstanalquest',
'doubleviewcasting',
],
[ [
'silverstonedvd', 'silverstonedvd',
'silviasaint', 'silviasaint',

Binary file not shown.

Before

Width:  |  Height:  |  Size: 15 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 31 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 14 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 15 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 103 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 22 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 96 KiB

View File

@ -795,7 +795,7 @@ const tagPhotos = [
['facefucking', 3, 'Adriana Chechik in "Performing Magic Butt Tricks With Jules Jordan. What Will Disappear In Her Ass?" for Jules Jordan'], ['facefucking', 3, 'Adriana Chechik in "Performing Magic Butt Tricks With Jules Jordan. What Will Disappear In Her Ass?" for Jules Jordan'],
['fake-boobs', 14, 'Rikki Six for Dream Dolls'], ['fake-boobs', 14, 'Rikki Six for Dream Dolls'],
['fake-boobs', 2, 'Gia Milana in "Hot Anal Latina" for HardX'], ['fake-boobs', 2, 'Gia Milana in "Hot Anal Latina" for HardX'],
['fake-boobs', 17, 'Felina in "With Flowers On The Floor" for LouisDeMirabert'], ['fake-boobs', 17, 'Felina in "With Flors On The Floor" for LouisDeMirabert'],
['fake-boobs', 18, 'Ebony Godess for Action Girls'], ['fake-boobs', 18, 'Ebony Godess for Action Girls'],
['fake-boobs', 1, 'Lela Star in "Thick" for Jules Jordan'], ['fake-boobs', 1, 'Lela Star in "Thick" for Jules Jordan'],
['fake-boobs', 16, 'Marsha May in "Once You Go Black 7" for Jules Jordan'], ['fake-boobs', 16, 'Marsha May in "Once You Go Black 7" for Jules Jordan'],

View File

@ -1,7 +1,6 @@
'use strict'; 'use strict';
const config = require('config'); const config = require('config');
const util = require('util');
const Promise = require('bluebird'); const Promise = require('bluebird');
const moment = require('moment'); const moment = require('moment');
const blake2 = require('blake2'); const blake2 = require('blake2');
@ -564,10 +563,8 @@ async function scrapeProfiles(actor, sources, entitiesBySlug, existingProfilesBy
// config may group sources to try until success // config may group sources to try until success
return await [].concat(source).reduce(async (outcome, scraperSlug) => outcome.catch(async () => { return await [].concat(source).reduce(async (outcome, scraperSlug) => outcome.catch(async () => {
try { try {
const entity = entitiesBySlug[scraperSlug] || null;
const scraper = scrapers[scraperSlug]; const scraper = scrapers[scraperSlug];
const layoutScraper = scraper?.[entity.parameters?.layout] || scraper; const entity = entitiesBySlug[scraperSlug] || null;
const context = { const context = {
...entity, ...entity,
@ -580,7 +577,7 @@ async function scrapeProfiles(actor, sources, entitiesBySlug, existingProfilesBy
const label = context.entity?.name; const label = context.entity?.name;
if (!layoutScraper?.fetchProfile) { if (!scraper?.fetchProfile) {
logger.warn(`No profile profile scraper available for ${scraperSlug}`); logger.warn(`No profile profile scraper available for ${scraperSlug}`);
throw new Error(`No profile profile scraper available for ${scraperSlug}`); throw new Error(`No profile profile scraper available for ${scraperSlug}`);
} }
@ -600,7 +597,7 @@ async function scrapeProfiles(actor, sources, entitiesBySlug, existingProfilesBy
logger.verbose(`Searching profile for '${actor.name}' on '${label}'`); logger.verbose(`Searching profile for '${actor.name}' on '${label}'`);
const profile = await layoutScraper.fetchProfile(curateActor({ const profile = await scraper.fetchProfile(curateActor({
...existingProfile, ...existingProfile,
...actor, ...actor,
}), context, include); }), context, include);
@ -690,7 +687,7 @@ async function scrapeActors(argNames) {
.whereNull('alias_for'), .whereNull('alias_for'),
]); ]);
const entitiesBySlug = entities.reduce((acc, entity) => ({ ...acc, [entity.slug]: acc[entity.slug] || entity }), {}); const entitiesBySlug = entities.reduce((acc, entity) => ({ ...acc, [entity.slug]: entity }), {});
const existingActorEntriesBySlugAndEntryId = existingActorEntries.reduce((acc, actorEntry) => ({ const existingActorEntriesBySlugAndEntryId = existingActorEntries.reduce((acc, actorEntry) => ({
...acc, ...acc,
@ -734,8 +731,8 @@ async function scrapeActors(argNames) {
logger.info(`Scraped ${profiles.length} profiles`); logger.info(`Scraped ${profiles.length} profiles`);
if (argv.report) { if (argv.inspect) {
console.log(util.inspect(profiles, { depth: Infinity, colors: true })); console.log(profiles);
} }
if (argv.save) { if (argv.save) {

View File

@ -66,7 +66,7 @@ async function init() {
const movieScenes = argv.movieScenes ? deepMovies.map(movie => movie.scenes?.map(scene => ({ ...scene, entity: movie.entity }))).flat().filter(Boolean) : []; const movieScenes = argv.movieScenes ? deepMovies.map(movie => movie.scenes?.map(scene => ({ ...scene, entity: movie.entity }))).flat().filter(Boolean) : [];
const deepMovieScenes = argv.deep ? await fetchScenes(movieScenes) : movieScenes; const deepMovieScenes = argv.deep ? await fetchScenes(movieScenes) : movieScenes;
if (argv.report) { if (argv.inspect) {
console.log(util.inspect(deepScenes, { depth: Infinity, colors: true })); console.log(util.inspect(deepScenes, { depth: Infinity, colors: true }));
console.log(util.inspect(deepMovies, { depth: Infinity, colors: true })); console.log(util.inspect(deepMovies, { depth: Infinity, colors: true }));
} }

View File

@ -223,7 +223,7 @@ const { argv } = yargs
type: 'boolean', type: 'boolean',
default: true, default: true,
}) })
.option('report', { .option('inspect', {
describe: 'Show data in console.', describe: 'Show data in console.',
type: 'boolean', type: 'boolean',
default: false, default: false,

View File

@ -1,7 +1,6 @@
'use strict'; 'use strict';
const qu = require('../utils/qu'); const qu = require('../utils/qu');
const slugify = require('../utils/slugify');
function scrapeAllA(scenes, channel) { function scrapeAllA(scenes, channel) {
return scenes.map(({ query }) => { return scenes.map(({ query }) => {
@ -30,6 +29,7 @@ function scrapeAllA(scenes, channel) {
release.tags = query.cnts('.tags a'); release.tags = query.cnts('.tags a');
release.rating = query.number('.thumb-rating'); release.rating = query.number('.thumb-rating');
console.log(release);
return release; return release;
}); });
} }
@ -63,56 +63,10 @@ function scrapeSceneA({ query }, url, channel) {
release.trailer = query.url('a[href*="get_file/"], .download a'); release.trailer = query.url('a[href*="get_file/"], .download a');
console.log(release);
return release; return release;
} }
function scrapeProfileA({ query, el }, entity) {
const profile = {};
const bio = query.all('.list-model-info li, .profile-info li').reduce((acc, bioEl) => ({
...acc,
[slugify(query.cnt(bioEl, '.title, span'), '_')]: query.cnt(bioEl, ':nth-child(2)') || query.q(bioEl, ':nth-child(2)', 'title') || query.text(bioEl),
}), {});
profile.dateOfBirth = qu.parseDate(bio.birth_date || bio.date_of_birth, 'DD MMMM, YYYY');
profile.birthPlace = bio.nationality || bio.place_of_birth || null;
profile.weight = Number(bio.weight?.match(/\d+/)?.[0]);
profile.height = Number(bio.height?.match(/\d+/)?.[0]);
profile.eyes = bio.eye_color;
profile.hairColor = bio.hair || bio.hair_color;
profile.aliases = query.text('.sub-title')?.replace(/:\s*/, '').split(/,\s*/);
if (bio.measurements || bio.body_shape_dimensions) {
const [, bust, cup, waist, hip] = (bio.measurements || bio.body_shape_dimensions).match(/(\d+)(\w+)-(\d+)-(\d+)/);
profile.bust = Number(bust);
profile.cup = cup;
profile.waist = Number(waist);
profile.hip = Number(hip);
}
const description = query.cnt('.model-biography p');
const avatar = query.img('.model-box img, .profile-model-photo', 'src', { origin: entity.url });
if (!/there is no description/.test(description)) {
profile.description = description;
}
if (avatar) {
profile.avatar = [
avatar,
avatar.replace('s2_', 's1_'),
];
}
profile.scenes = scrapeAllA(qu.initAll(el, '.list-thumbs .thumb, .main-thumbs > li'), entity);
return profile;
}
async function fetchLatestA(channel, page) { async function fetchLatestA(channel, page) {
const url = channel.parameters?.latest const url = channel.parameters?.latest
? `${channel.parameters.latest}/${page}` ? `${channel.parameters.latest}/${page}`
@ -137,33 +91,9 @@ async function fetchSceneA(url, channel) {
return res.status; return res.status;
} }
async function fetchProfileA({ name, slug }, { entity }) {
const searchRes = await qu.getAll(`${entity.url}/models/search/?q=${name}`, '.thumb-modal, .big-thumb');
if (!searchRes.ok) {
return searchRes.status;
}
const actor = searchRes.items.find(({ query }) => slugify(query.cnt('.thumb-title a, .title')) === slug);
if (!actor) {
return null;
}
const actorUrl = actor.query.url('a', 'href', { origin: entity.url });
const actorRes = await qu.get(actorUrl);
if (actorRes.ok) {
return scrapeProfileA(actorRes.item, entity);
}
return null;
}
module.exports = { module.exports = {
a: { a: {
fetchLatest: fetchLatestA, fetchLatest: fetchLatestA,
fetchScene: fetchSceneA, fetchScene: fetchSceneA,
fetchProfile: fetchProfileA,
}, },
}; };

View File

@ -198,13 +198,11 @@ const scrapers = {
digitalplayground, digitalplayground,
dtfsluts: fullpornnetwork, dtfsluts: fullpornnetwork,
dorcelclub: dorcel, dorcelclub: dorcel,
doubleviewcasting: firstanalquest,
elegantangel, elegantangel,
evilangel, evilangel,
eyeontheguy: hush, eyeontheguy: hush,
fakehub, fakehub,
exploitedcollegegirls: fcuk, exploitedcollegegirls: fcuk,
firstanalquest,
forbondage: porndoe, forbondage: porndoe,
freeones, freeones,
gangbangcreampie: aziani, gangbangcreampie: aziani,