Compare commits

..

No commits in common. "6e2527e5c5686a81dd4a40757980c06642830e93" and "acc496be47120e1e885972df852b3553341d7fdd" have entirely different histories.

10 changed files with 32 additions and 80 deletions

View File

@ -90,7 +90,6 @@ export default {
} }
.movie { .movie {
height: 16rem;
display: flex; display: flex;
} }
@ -100,7 +99,7 @@ export default {
} }
.cover { .cover {
height: 100%; height: 16rem;
box-shadow: 0 0 3px var(--darken-weak); box-shadow: 0 0 3px var(--darken-weak);
img { img {
@ -114,8 +113,6 @@ export default {
} }
.info { .info {
display: flex;
flex-direction: column;
flex-grow: 1; flex-grow: 1;
overflow: hidden; overflow: hidden;
} }
@ -131,11 +128,9 @@ export default {
} }
.actors { .actors {
height: 0;
flex-grow: 1;
padding: 0 1rem; padding: 0 1rem;
margin: 0 0 1rem 0;
line-height: 1.5; line-height: 1.5;
overflow: hidden;
} }
.actor:not(:last-child)::after { .actor:not(:last-child)::after {
@ -155,7 +150,7 @@ export default {
} }
.tags { .tags {
padding: .25rem 1rem; padding: .2rem 1rem 0 1rem;
height: 1.75rem; height: 1.75rem;
line-height: 2; line-height: 2;
overflow: hidden; overflow: hidden;
@ -180,13 +175,8 @@ export default {
} }
@media(max-width: $breakpoint) { @media(max-width: $breakpoint) {
.movie { .cover {
height: 12rem; height: 12rem;
} }
/* ensure no half actor names show */
.actors {
margin: 0 0 1rem 0;
}
} }
</style> </style>

2
package-lock.json generated
View File

@ -1,6 +1,6 @@
{ {
"name": "traxxx", "name": "traxxx",
"version": "1.168.9", "version": "1.168.7",
"lockfileVersion": 1, "lockfileVersion": 1,
"requires": true, "requires": true,
"dependencies": { "dependencies": {

View File

@ -1,6 +1,6 @@
{ {
"name": "traxxx", "name": "traxxx",
"version": "1.168.9", "version": "1.168.7",
"description": "All the latest porn releases in one place", "description": "All the latest porn releases in one place",
"main": "src/app.js", "main": "src/app.js",
"scripts": { "scripts": {

View File

@ -17,10 +17,6 @@ const grandParentNetworks = [
name: 'Mind Geek', name: 'Mind Geek',
url: 'https://www.mindgeek.com', url: 'https://www.mindgeek.com',
description: '', description: '',
parameters: {
interval: 1000,
concurrency: 1,
},
}, },
{ {
slug: 'whalemember', slug: 'whalemember',

View File

@ -30,8 +30,6 @@ const { deleteScenes } = require('./releases');
const slugify = require('./utils/slugify'); const slugify = require('./utils/slugify');
const capitalize = require('./utils/capitalize'); const capitalize = require('./utils/capitalize');
const resolvePlace = require('./utils/resolve-place'); const resolvePlace = require('./utils/resolve-place');
const { resolveLayoutScraper } = require('./scrapers/resolve');
const getRecursiveParameters = require('./utils/get-recursive-parameters');
const hairColors = { const hairColors = {
'jet-black': 'black', 'jet-black': 'black',
@ -639,7 +637,10 @@ async function scrapeProfiles(actor, sources, entitiesBySlug, existingProfilesBy
const entity = entitiesBySlug[scraperSlug] || null; const entity = entitiesBySlug[scraperSlug] || null;
const scraper = scrapers[scraperSlug]; const scraper = scrapers[scraperSlug];
const layoutScraper = resolveLayoutScraper(entity, scraper); const layoutScraper = scraper?.[entity.parameters?.layout]
|| scraper?.[entity.parent?.parameters?.layout]
|| scraper?.[entity.parent?.parent?.parameters?.layout]
|| scraper;
const context = { const context = {
...entity, ...entity,
@ -648,7 +649,11 @@ async function scrapeProfiles(actor, sources, entitiesBySlug, existingProfilesBy
network: entity?.parent, network: entity?.parent,
entity, entity,
scraper: scraperSlug, scraper: scraperSlug,
parameters: getRecursiveParameters(entity), parameters: {
...entity?.parent?.parent?.parameters,
...entity?.parent?.parameters,
...entity?.parameters,
},
}; };
const label = context.entity?.name; const label = context.entity?.name;

View File

@ -9,7 +9,6 @@ const { resolveScraper, resolveLayoutScraper } = require('./scrapers/resolve');
const { fetchReleaseEntities, urlToSiteSlug } = require('./entities'); const { fetchReleaseEntities, urlToSiteSlug } = require('./entities');
const logger = require('./logger')(__filename); const logger = require('./logger')(__filename);
const qu = require('./utils/qu'); const qu = require('./utils/qu');
const getRecursiveParameters = require('./utils/get-recursive-parameters');
function toBaseReleases(baseReleasesOrUrls, entity = null) { function toBaseReleases(baseReleasesOrUrls, entity = null) {
if (!baseReleasesOrUrls) { if (!baseReleasesOrUrls) {
@ -107,19 +106,9 @@ async function scrapeRelease(baseRelease, entitiesBySlug, type = 'scene') {
try { try {
logger.verbose(`Fetching ${type} ${baseRelease.url}`); logger.verbose(`Fetching ${type} ${baseRelease.url}`);
const options = {
...include,
parameters: getRecursiveParameters(entity),
};
const scrapedRelease = type === 'scene' const scrapedRelease = type === 'scene'
? await fetchScene(layoutScraper, baseRelease.url, entity, baseRelease, options, null) ? await fetchScene(layoutScraper, baseRelease.url, entity, baseRelease, include, null)
: await layoutScraper.fetchMovie(baseRelease.url, entity, baseRelease, options, null); : await layoutScraper.fetchMovie(baseRelease.url, entity, baseRelease, include, null);
if (typeof scrapedRelease !== 'object' || Array.isArray(scrapedRelease)) {
// scraper is unable to fetch the releases and returned a HTTP code or null
throw new Error(`Scraper returned ${scrapedRelease} when fetching latest from '${entity.name}' (${entity.parent?.name})`);
}
// object-merge-advance will use null as explicit false on hard merged keys, even when null as explicit falls is disabled // object-merge-advance will use null as explicit false on hard merged keys, even when null as explicit falls is disabled
// filter out keys with null values to ensure original base value is used instead // filter out keys with null values to ensure original base value is used instead

View File

@ -168,6 +168,8 @@ function sortBaseTrailersByQuality(sources, role) {
return 0; return 0;
}); });
console.log(sortedSources);
return sortedSources; return sortedSources;
} }

View File

@ -68,7 +68,7 @@ function scrapeLatestX(data, site, filterChannel) {
|| (site.parameters?.native && `${site.url}/scene`) || (site.parameters?.native && `${site.url}/scene`)
|| `${site.parent.url}/scene`; || `${site.parent.url}/scene`;
release.url = `${basepath}/${release.entryId}/${slugify(release.title)}`; release.url = `${basepath}/${release.entryId}/`;
release.date = new Date(data.dateReleased); release.date = new Date(data.dateReleased);
release.actors = data.actors.map(actor => ({ name: actor.name, gender: actor.gender })); release.actors = data.actors.map(actor => ({ name: actor.name, gender: actor.gender }));
@ -143,7 +143,7 @@ function getUrl(site) {
throw new Error(`Mind Geek site '${site.name}' (${site.url}) not supported`); throw new Error(`Mind Geek site '${site.name}' (${site.url}) not supported`);
} }
async function getSession(site, parameters) { async function getSession(site) {
const cookieJar = new CookieJar(); const cookieJar = new CookieJar();
const session = http.session({ cookieJar }); const session = http.session({ cookieJar });
@ -152,11 +152,7 @@ async function getSession(site, parameters) {
? site.parent.url ? site.parent.url
: site.url; : site.url;
const res = await http.get(sessionUrl, { const res = await http.get(sessionUrl, { session });
session,
interval: parameters?.interval,
concurrency: parameters?.concurrency,
});
if (res.statusCode === 200) { if (res.statusCode === 200) {
const cookieString = await cookieJar.getCookieStringAsync(sessionUrl); const cookieString = await cookieJar.getCookieStringAsync(sessionUrl);
@ -216,12 +212,12 @@ function scrapeProfile(data, html, releases = [], networkName) {
return profile; return profile;
} }
async function fetchLatest(site, page = 1, options) { async function fetchLatest(site, page = 1) {
const url = getUrl(site); const url = getUrl(site);
const { searchParams } = new URL(url); const { searchParams } = new URL(url);
const siteId = searchParams.get('site'); const siteId = searchParams.get('site');
const { session, instanceToken } = await getSession(site, options.parameters); const { session, instanceToken } = await getSession(site);
const beforeDate = moment().add('1', 'day').format('YYYY-MM-DD'); const beforeDate = moment().add('1', 'day').format('YYYY-MM-DD');
const limit = 10; const limit = 10;
@ -231,8 +227,6 @@ async function fetchLatest(site, page = 1, options) {
const res = await http.get(apiUrl, { const res = await http.get(apiUrl, {
session, session,
interval: options.parameters.interval,
concurrency: options.parameters.concurrency,
headers: { headers: {
Instance: instanceToken, Instance: instanceToken,
Origin: site.url, Origin: site.url,
@ -247,16 +241,14 @@ async function fetchLatest(site, page = 1, options) {
return null; return null;
} }
async function fetchUpcoming(site, page, options) { async function fetchUpcoming(site) {
const url = getUrl(site); const url = getUrl(site);
const { session, instanceToken } = await getSession(site, options.parameters); const { session, instanceToken } = await getSession(site);
const apiUrl = 'https://site-api.project1service.com/v2/upcoming-releases'; const apiUrl = 'https://site-api.project1service.com/v2/upcoming-releases';
const res = await http.get(apiUrl, { const res = await http.get(apiUrl, {
session, session,
interval: options.parameters.interval,
concurrency: options.parameters.concurrency,
headers: { headers: {
Instance: instanceToken, Instance: instanceToken,
Origin: site.url, Origin: site.url,
@ -271,19 +263,17 @@ async function fetchUpcoming(site, page, options) {
return null; return null;
} }
async function fetchScene(url, site, baseScene, options) { async function fetchScene(url, site, baseScene) {
if (baseScene?.entryId) { if (baseScene?.entryId) {
// overview and deep data is the same, don't hit server unnecessarily // overview and deep data is the same, don't hit server unnecessarily
return baseScene; return baseScene;
} }
const entryId = url.match(/\d+/)[0]; const entryId = url.match(/\d+/)[0];
const { session, instanceToken } = await getSession(site, options.parameters); const { session, instanceToken } = await getSession(site);
const res = await http.get(`https://site-api.project1service.com/v2/releases/${entryId}`, { const res = await http.get(`https://site-api.project1service.com/v2/releases/${entryId}`, {
session, session,
interval: options.parameters.interval,
concurrency: options.parameters.concurrency,
headers: { headers: {
Instance: instanceToken, Instance: instanceToken,
}, },
@ -296,14 +286,12 @@ async function fetchScene(url, site, baseScene, options) {
return null; return null;
} }
async function fetchProfile({ name: actorName, slug: actorSlug }, { entity, parameters }) { async function fetchProfile({ name: actorName }, { entity }) {
// const url = `https://www.${networkOrNetworkSlug.slug || networkOrNetworkSlug}.com`; // const url = `https://www.${networkOrNetworkSlug.slug || networkOrNetworkSlug}.com`;
const { session, instanceToken } = await getSession(entity, parameters); const { session, instanceToken } = await getSession(entity);
const res = await http.get(`https://site-api.project1service.com/v1/actors/?search=${encodeURI(actorName)}`, { const res = await http.get(`https://site-api.project1service.com/v1/actors/?search=${encodeURI(actorName)}`, {
session, session,
interval: parameters.interval,
concurrency: parameters.concurrency,
headers: { headers: {
Instance: instanceToken, Instance: instanceToken,
}, },
@ -313,18 +301,13 @@ async function fetchProfile({ name: actorName, slug: actorSlug }, { entity, para
const actorData = res.body.result.find(actor => actor.name.toLowerCase() === actorName.toLowerCase()); const actorData = res.body.result.find(actor => actor.name.toLowerCase() === actorName.toLowerCase());
if (actorData) { if (actorData) {
const actorUrl = `https://www.${entity.slug}.com/${entity.parameters?.actorPath || 'model'}/${actorData.id}/${actorSlug}`; const actorUrl = `https://www.${entity.slug}.com/${entity.parameters?.actorPath || 'model'}/${actorData.id}/`;
const actorReleasesUrl = `https://site-api.project1service.com/v2/releases?actorId=${actorData.id}&limit=100&offset=0&orderBy=-dateReleased&type=scene`; const actorReleasesUrl = `https://site-api.project1service.com/v2/releases?actorId=${actorData.id}&limit=100&offset=0&orderBy=-dateReleased&type=scene`;
const [actorRes, actorReleasesRes] = await Promise.all([ const [actorRes, actorReleasesRes] = await Promise.all([
http.get(actorUrl, { http.get(actorUrl),
interval: parameters.interval,
concurrency: parameters.concurrency,
}),
http.get(actorReleasesUrl, { http.get(actorReleasesUrl, {
session, session,
interval: parameters.interval,
concurrency: parameters.concurrency,
headers: { headers: {
Instance: instanceToken, Instance: instanceToken,
}, },

View File

@ -11,7 +11,6 @@ const { curateRelease } = require('./releases');
const include = require('./utils/argv-include')(argv); const include = require('./utils/argv-include')(argv);
const { resolveScraper, resolveLayoutScraper } = require('./scrapers/resolve'); const { resolveScraper, resolveLayoutScraper } = require('./scrapers/resolve');
const { fetchIncludedEntities } = require('./entities'); const { fetchIncludedEntities } = require('./entities');
const getRecursiveParameters = require('./utils/get-recursive-parameters');
const emptyReleases = { uniqueReleases: [], duplicateReleases: [] }; const emptyReleases = { uniqueReleases: [], duplicateReleases: [] };
@ -98,7 +97,6 @@ async function scrapeReleases(scraper, entity, preData, isUpcoming) {
const options = { const options = {
...config.options[scraper.slug], ...config.options[scraper.slug],
...include, ...include,
parameters: getRecursiveParameters(entity),
}; };
const pageReleases = isUpcoming const pageReleases = isUpcoming

View File

@ -1,11 +0,0 @@
'use strict';
function getRecursiveParameters(entity, parameters) {
if (entity.parent) {
return getRecursiveParameters(entity.parent, { ...parameters, ...entity.parameters });
}
return { ...parameters, ...entity.parameters };
}
module.exports = getRecursiveParameters;