diff options
| author | syuilo <syuilotan@yahoo.co.jp> | 2019-02-06 20:56:48 +0900 |
|---|---|---|
| committer | syuilo <syuilotan@yahoo.co.jp> | 2019-02-06 20:56:48 +0900 |
| commit | 27d16c6a120d05e69eb069035a851c48c44548c5 (patch) | |
| tree | 678e3a68df78837652047f8f4da4306e22377cd6 /src/queue/processors | |
| parent | Merge branch 'develop' of https://github.com/syuilo/misskey into develop (diff) | |
| download | sharkey-27d16c6a120d05e69eb069035a851c48c44548c5.tar.gz sharkey-27d16c6a120d05e69eb069035a851c48c44548c5.tar.bz2 sharkey-27d16c6a120d05e69eb069035a851c48c44548c5.zip | |
Resolve #4151
Diffstat (limited to 'src/queue/processors')
| -rw-r--r-- | src/queue/processors/export-blocking.ts | 90 | ||||
| -rw-r--r-- | src/queue/processors/export-following.ts | 90 | ||||
| -rw-r--r-- | src/queue/processors/export-mute.ts | 90 | ||||
| -rw-r--r-- | src/queue/processors/index.ts | 6 |
4 files changed, 276 insertions, 0 deletions
diff --git a/src/queue/processors/export-blocking.ts b/src/queue/processors/export-blocking.ts new file mode 100644 index 0000000000..b7a2d3122f --- /dev/null +++ b/src/queue/processors/export-blocking.ts @@ -0,0 +1,90 @@ + +import * as bq from 'bee-queue'; +import * as tmp from 'tmp'; +import * as fs from 'fs'; +import * as mongo from 'mongodb'; + +import { queueLogger } from '../logger'; +import addFile from '../../services/drive/add-file'; +import User from '../../models/user'; +import dateFormat = require('dateformat'); +import Blocking from '../../models/blocking'; +import config from '../../config'; + +const logger = queueLogger.createSubLogger('export-blocking'); + +export async function exportBlocking(job: bq.Job, done: any): Promise<void> { + logger.info(`Exporting blocking of ${job.data.user._id} ...`); + + const user = await User.findOne({ + _id: new mongo.ObjectID(job.data.user._id.toString()) + }); + + // Create temp file + const [path, cleanup] = await new Promise<[string, any]>((res, rej) => { + tmp.file((e, path, fd, cleanup) => { + if (e) return rej(e); + res([path, cleanup]); + }); + }); + + logger.info(`Temp file is ${path}`); + + const stream = fs.createWriteStream(path, { flags: 'a' }); + + let exportedCount = 0; + let ended = false; + let cursor: any = null; + + while (!ended) { + const blockings = await Blocking.find({ + blockerId: user._id, + ...(cursor ? { _id: { $gt: cursor } } : {}) + }, { + limit: 100, + sort: { + _id: 1 + } + }); + + if (blockings.length === 0) { + ended = true; + job.reportProgress(100); + break; + } + + cursor = blockings[blockings.length - 1]._id; + + for (const block of blockings) { + const u = await User.findOne({ _id: block.blockeeId }, { fields: { username: true, host: true } }); + const content = u.host ? `${u.username}@${u.host}` : `${u.username}@${config.host}`; + await new Promise((res, rej) => { + stream.write(content + '\n', err => { + if (err) { + logger.error(err); + rej(err); + } else { + res(); + } + }); + }); + exportedCount++; + } + + const total = await Blocking.count({ + blockerId: user._id, + }); + + job.reportProgress(exportedCount / total); + } + + stream.end(); + logger.succ(`Exported to: ${path}`); + + const fileName = dateFormat(new Date(), 'blocking-yyyy-mm-dd-HH-MM-ss') + '.json'; + const driveFile = await addFile(user, path, fileName); + + logger.succ(`Exported to: ${driveFile._id}`); + cleanup(); + done(); +} diff --git a/src/queue/processors/export-following.ts b/src/queue/processors/export-following.ts new file mode 100644 index 0000000000..7515926361 --- /dev/null +++ b/src/queue/processors/export-following.ts @@ -0,0 +1,90 @@ + +import * as bq from 'bee-queue'; +import * as tmp from 'tmp'; +import * as fs from 'fs'; +import * as mongo from 'mongodb'; + +import { queueLogger } from '../logger'; +import addFile from '../../services/drive/add-file'; +import User from '../../models/user'; +import dateFormat = require('dateformat'); +import Following from '../../models/following'; +import config from '../../config'; + +const logger = queueLogger.createSubLogger('export-following'); + +export async function exportFollowing(job: bq.Job, done: any): Promise<void> { + logger.info(`Exporting following of ${job.data.user._id} ...`); + + const user = await User.findOne({ + _id: new mongo.ObjectID(job.data.user._id.toString()) + }); + + // Create temp file + const [path, cleanup] = await new Promise<[string, any]>((res, rej) => { + tmp.file((e, path, fd, cleanup) => { + if (e) return rej(e); + res([path, cleanup]); + }); + }); + + logger.info(`Temp file is ${path}`); + + const stream = fs.createWriteStream(path, { flags: 'a' }); + + let exportedCount = 0; + let ended = false; + let cursor: any = null; + + while (!ended) { + const followings = await Following.find({ + followerId: user._id, + ...(cursor ? { _id: { $gt: cursor } } : {}) + }, { + limit: 100, + sort: { + _id: 1 + } + }); + + if (followings.length === 0) { + ended = true; + job.reportProgress(100); + break; + } + + cursor = followings[followings.length - 1]._id; + + for (const following of followings) { + const u = await User.findOne({ _id: following.followeeId }, { fields: { username: true, host: true } }); + const content = u.host ? `${u.username}@${u.host}` : `${u.username}@${config.host}`; + await new Promise((res, rej) => { + stream.write(content + '\n', err => { + if (err) { + logger.error(err); + rej(err); + } else { + res(); + } + }); + }); + exportedCount++; + } + + const total = await Following.count({ + followerId: user._id, + }); + + job.reportProgress(exportedCount / total); + } + + stream.end(); + logger.succ(`Exported to: ${path}`); + + const fileName = dateFormat(new Date(), 'following-yyyy-mm-dd-HH-MM-ss') + '.json'; + const driveFile = await addFile(user, path, fileName); + + logger.succ(`Exported to: ${driveFile._id}`); + cleanup(); + done(); +} diff --git a/src/queue/processors/export-mute.ts b/src/queue/processors/export-mute.ts new file mode 100644 index 0000000000..e526aa06f6 --- /dev/null +++ b/src/queue/processors/export-mute.ts @@ -0,0 +1,90 @@ + +import * as bq from 'bee-queue'; +import * as tmp from 'tmp'; +import * as fs from 'fs'; +import * as mongo from 'mongodb'; + +import { queueLogger } from '../logger'; +import addFile from '../../services/drive/add-file'; +import User from '../../models/user'; +import dateFormat = require('dateformat'); +import Mute from '../../models/mute'; +import config from '../../config'; + +const logger = queueLogger.createSubLogger('export-mute'); + +export async function exportMute(job: bq.Job, done: any): Promise<void> { + logger.info(`Exporting mute of ${job.data.user._id} ...`); + + const user = await User.findOne({ + _id: new mongo.ObjectID(job.data.user._id.toString()) + }); + + // Create temp file + const [path, cleanup] = await new Promise<[string, any]>((res, rej) => { + tmp.file((e, path, fd, cleanup) => { + if (e) return rej(e); + res([path, cleanup]); + }); + }); + + logger.info(`Temp file is ${path}`); + + const stream = fs.createWriteStream(path, { flags: 'a' }); + + let exportedCount = 0; + let ended = false; + let cursor: any = null; + + while (!ended) { + const mutes = await Mute.find({ + muterId: user._id, + ...(cursor ? { _id: { $gt: cursor } } : {}) + }, { + limit: 100, + sort: { + _id: 1 + } + }); + + if (mutes.length === 0) { + ended = true; + job.reportProgress(100); + break; + } + + cursor = mutes[mutes.length - 1]._id; + + for (const mute of mutes) { + const u = await User.findOne({ _id: mute.muteeId }, { fields: { username: true, host: true } }); + const content = u.host ? `${u.username}@${u.host}` : `${u.username}@${config.host}`; + await new Promise((res, rej) => { + stream.write(content + '\n', err => { + if (err) { + logger.error(err); + rej(err); + } else { + res(); + } + }); + }); + exportedCount++; + } + + const total = await Mute.count({ + muterId: user._id, + }); + + job.reportProgress(exportedCount / total); + } + + stream.end(); + logger.succ(`Exported to: ${path}`); + + const fileName = dateFormat(new Date(), 'mute-yyyy-mm-dd-HH-MM-ss') + '.json'; + const driveFile = await addFile(user, path, fileName); + + logger.succ(`Exported to: ${driveFile._id}`); + cleanup(); + done(); +} diff --git a/src/queue/processors/index.ts b/src/queue/processors/index.ts index 3f08fe29fb..6869983205 100644 --- a/src/queue/processors/index.ts +++ b/src/queue/processors/index.ts @@ -1,12 +1,18 @@ import deliver from './http/deliver'; import processInbox from './http/process-inbox'; import { exportNotes } from './export-notes'; +import { exportFollowing } from './export-following'; +import { exportMute } from './export-mute'; +import { exportBlocking } from './export-blocking'; import { queueLogger } from '../logger'; const handlers: any = { deliver, processInbox, exportNotes, + exportFollowing, + exportMute, + exportBlocking, }; export default (job: any, done: any) => { |