Reduced media concurrency to assess effect on memory. Moved qu context removal to deep scrape runner. Updated movie graphql queries.
This commit is contained in:
@@ -377,6 +377,7 @@ async function writeLazy(image, lazypath) {
|
||||
|
||||
async function storeImageFile(media, hashDir, hashSubDir, filename, filedir, filepath, options) {
|
||||
logger.silly(`Storing permanent media files for ${media.id} from ${media.src} at ${filepath}`);
|
||||
logger.debug(`Memory usage at image storage: ${process.memoryUsage.rss() / 1000000} MB (${media.src})`);
|
||||
|
||||
try {
|
||||
const thumbdir = config.s3.enabled ? path.join(media.role, 'thumbs') : path.join(media.role, 'thumbs', hashDir, hashSubDir);
|
||||
@@ -746,7 +747,8 @@ async function storeMedias(baseMedias, options) {
|
||||
const fetchedMedias = await Promise.map(
|
||||
baseMedias,
|
||||
async (baseMedia) => fetchMedia(baseMedia, { existingSourceMediaByUrl, existingExtractMediaByUrl }),
|
||||
{ concurrency: 100 }, // don't overload disk (or network, although this has its own throttling)
|
||||
// { concurrency: 100 }, // don't overload disk (or network, although this has its own throttling)
|
||||
{ concurrency: 10 }, // don't overload disk (or network, although this has its own throttling)
|
||||
);
|
||||
|
||||
const { uniqueHashMedias, existingHashMedias } = await findHashDuplicates(fetchedMedias);
|
||||
|
||||
Reference in New Issue
Block a user