ripunzel/src/app.js

133 lines
3.7 KiB
JavaScript
Raw Normal View History

2018-04-09 22:26:30 +00:00
'use strict';
const config = require('config');
const Snoowrap = require('snoowrap');
const fs = require('fs-extra');
2018-07-02 01:45:20 +00:00
const Promise = require('bluebird');
const exiftool = require('node-exiftool');
const exiftoolBin = require('dist-exiftool');
const cron = require('node-cron');
const { format } = require('date-fns');
require('array.prototype.flatten').shim();
const reddit = new Snoowrap(config.reddit.api);
const args = require('./cli.js')();
const dissectLink = require('./dissectLink.js');
const curatePosts = require('./curate/posts.js');
const { attachContentInfo, getInfo } = require('./fetch/info.js');
const { fetchSaveUserContent, fetchSaveDirectContent } = require('./fetch/content.js');
2018-05-05 00:27:15 +00:00
const getPosts = require('./sources/getPosts.js')(reddit, args);
const getUserPosts = require('./sources/getUserPosts.js')(reddit, args);
2018-04-18 02:04:39 +00:00
async function getFileContents(location, label) {
try {
const fileContents = await fs.readFile(location, 'utf8');
return fileContents.split('\n').filter(entry => entry);
} catch (error) {
console.log('\x1b[31m%s\x1b[0m', `Could not read ${label} file '${location}': ${error}.`);
return [];
}
}
async function getCompletePosts() {
let userPosts = {};
let ignoreIds = [];
let usernames = args.users || [];
let postIds = args.posts || [];
if (args.fileUsers) {
usernames = usernames.concat(await getFileContents(args.fileUsers, 'username'));
}
if (args.filePosts) {
postIds = postIds.concat(await getFileContents(args.filePosts, 'post ID'));
}
if (!usernames.length && !postIds.length) {
throw new Error('Could not retrieve any posts. Did you supply --users, --posts, --file-users or --file-posts?');
}
if (usernames.length) {
userPosts = await getUserPosts(usernames);
}
if (postIds.length) {
userPosts = await getPosts(postIds, userPosts);
}
if (args.fileIgnore) {
ignoreIds = await getFileContents(args.fileIgnore, 'ignore');
}
const curatedUserPosts = curatePosts(userPosts, ignoreIds, args);
return attachContentInfo(curatedUserPosts);
}
async function getDirectContent(links, ep) {
return Promise.map(links, async (link) => {
const host = dissectLink(link);
const info = await getInfo(host);
return fetchSaveDirectContent(info, host, ep);
}, {
concurrency: 5,
});
}
async function getCompleteContents(ep) {
if (args.fetch) {
return getDirectContent([args.fetch], ep);
}
if (args.fileDirect) {
return getDirectContent(await getFileContents(args.fileDirect, 'direct'), ep);
}
return null;
}
function fetchSavePosts(userPosts, ep) {
2018-07-02 01:45:20 +00:00
// don't map to apply concurrency limit and reduce network stress
return Promise.reduce(Object.values(userPosts), (acc, user) => fetchSaveUserContent(user, ep, args), null);
}
async function initApp() {
try {
const ep = new exiftool.ExiftoolProcess(exiftoolBin);
await ep.open();
if (args.fetch || args.fileDirect) {
await getCompleteContents(ep);
return;
}
const userPosts = await getCompletePosts();
await fetchSavePosts(userPosts, ep);
await ep.close();
if (args.watch) {
console.log(`[${format(new Date(), 'YYYY-MM-DD HH:mm:ss')}] Watch-mode enabled, checking again for new posts according to crontab '${config.fetch.watch.schedule}'.`);
}
} catch (error) {
if (args.debug) {
console.log('\x1b[31m%s\x1b[0m', error.stack);
} else {
console.log('\x1b[31m%s\x1b[0m', error.message);
}
}
}
initApp();
if (args.watch) {
cron.schedule(config.fetch.watch.schedule, initApp);
}