mirror of
https://github.com/BobbyWibowo/lolisafe.git
synced 2024-12-14 16:36:21 +00:00
d5c24165cf
Updated Cloudfalre's cache purge utility. It will now split URLs into chunks of 30 URLs which then will be purged one chunk at a time. I just found out Cloudflare's API have a limit of 30 URLs for the API.
782 lines
26 KiB
JavaScript
782 lines
26 KiB
JavaScript
const config = require('./../config')
|
|
const crypto = require('crypto')
|
|
const db = require('knex')(config.database)
|
|
const fetch = require('node-fetch')
|
|
const fs = require('fs')
|
|
const multer = require('multer')
|
|
const path = require('path')
|
|
const perms = require('./permissionController')
|
|
const randomstring = require('randomstring')
|
|
const utils = require('./utilsController')
|
|
|
|
const uploadsController = {}
|
|
|
|
const maxTries = config.uploads.maxTries || 1
|
|
const uploadsDir = path.join(__dirname, '..', config.uploads.folder)
|
|
const chunkedUploads = Boolean(config.uploads.chunkSize)
|
|
const chunksDir = path.join(uploadsDir, 'chunks')
|
|
const maxSize = config.uploads.maxSize
|
|
const maxSizeBytes = parseInt(maxSize) * 1000 * 1000
|
|
const urlMaxSizeBytes = parseInt(config.uploads.urlMaxSize) * 1000 * 1000
|
|
|
|
const storage = multer.diskStorage({
|
|
destination (req, file, cb) {
|
|
// If chunked uploads is disabled or the uploaded file is not a chunk
|
|
if (!chunkedUploads || (req.body.uuid === undefined && req.body.chunkindex === undefined))
|
|
return cb(null, uploadsDir)
|
|
|
|
const uuidDir = path.join(chunksDir, req.body.uuid)
|
|
fs.access(uuidDir, error => {
|
|
if (!error) return cb(null, uuidDir)
|
|
fs.mkdir(uuidDir, error => {
|
|
if (!error) return cb(null, uuidDir)
|
|
console.error(error)
|
|
// eslint-disable-next-line standard/no-callback-literal
|
|
return cb('Could not process the chunked upload. Try again?')
|
|
})
|
|
})
|
|
},
|
|
filename (req, file, cb) {
|
|
// If chunked uploads is disabled or the uploaded file is not a chunk
|
|
if (!chunkedUploads || (req.body.uuid === undefined && req.body.chunkindex === undefined)) {
|
|
const extension = utils.extname(file.originalname)
|
|
const length = uploadsController.getFileNameLength(req)
|
|
return uploadsController.getUniqueRandomName(length, extension, req.app.get('uploads-set'))
|
|
.then(name => cb(null, name))
|
|
.catch(error => cb(error))
|
|
}
|
|
|
|
// index.extension (e.i. 0, 1, ..., n - will prepend zeros depending on the amount of chunks)
|
|
const digits = req.body.totalchunkcount !== undefined ? `${req.body.totalchunkcount - 1}`.length : 1
|
|
const zeros = new Array(digits + 1).join('0')
|
|
const name = (zeros + req.body.chunkindex).slice(-digits)
|
|
return cb(null, name)
|
|
}
|
|
})
|
|
|
|
const upload = multer({
|
|
storage,
|
|
limits: {
|
|
fileSize: maxSizeBytes
|
|
},
|
|
fileFilter (req, file, cb) {
|
|
const extname = utils.extname(file.originalname)
|
|
if (uploadsController.isExtensionFiltered(extname))
|
|
// eslint-disable-next-line standard/no-callback-literal
|
|
return cb(`${extname ? `${extname.substr(1).toUpperCase()} files` : 'Files with no extension'} are not permitted.`)
|
|
|
|
// Re-map Dropzone keys so people can manually use the API without prepending 'dz'
|
|
for (const key in req.body) {
|
|
if (!/^dz/.test(key)) continue
|
|
req.body[key.replace(/^dz/, '')] = req.body[key]
|
|
delete req.body[key]
|
|
}
|
|
|
|
if (req.body.chunkindex) {
|
|
if (!chunkedUploads)
|
|
// eslint-disable-next-line standard/no-callback-literal
|
|
return cb('Chunked uploads are disabled at the moment.')
|
|
|
|
const totalfilesize = parseInt(req.body.totalfilesize)
|
|
if (!isNaN(totalfilesize)) {
|
|
if (config.filterEmptyFile && totalfilesize === 0)
|
|
// eslint-disable-next-line standard/no-callback-literal
|
|
return cb('Empty files are not allowed.')
|
|
if (totalfilesize > maxSizeBytes)
|
|
// eslint-disable-next-line standard/no-callback-literal
|
|
return cb('Chunk error occurred. Total file size is larger than the maximum file size.')
|
|
}
|
|
}
|
|
|
|
return cb(null, true)
|
|
}
|
|
}).array('files[]')
|
|
|
|
uploadsController.isExtensionFiltered = extname => {
|
|
// If empty extension needs to be filtered
|
|
if (!extname && config.filterNoExtension) return true
|
|
// If there are extensions that have to be filtered
|
|
if (extname && Array.isArray(config.extensionsFilter) && config.extensionsFilter.length) {
|
|
const match = config.extensionsFilter.some(extension => extname === extension.toLowerCase())
|
|
const whitelist = config.extensionsFilterMode === 'whitelist'
|
|
if ((!whitelist && match) || (whitelist && !match)) return true
|
|
}
|
|
return false
|
|
}
|
|
|
|
uploadsController.getFileNameLength = req => {
|
|
// If the user has a preferred file length, make sure it is within the allowed range
|
|
if (req.headers.filelength)
|
|
return Math.min(Math.max(req.headers.filelength, config.uploads.fileLength.min), config.uploads.fileLength.max)
|
|
|
|
// Let's default it to 32 characters when config key is falsy
|
|
return config.uploads.fileLength.default || 32
|
|
}
|
|
|
|
uploadsController.getUniqueRandomName = (length, extension, set) => {
|
|
return new Promise((resolve, reject) => {
|
|
const access = i => {
|
|
const identifier = randomstring.generate(length)
|
|
if (config.uploads.cacheFileIdentifiers) {
|
|
// Check whether the identifier is already used in cache
|
|
if (set.has(identifier)) {
|
|
console.log(`Identifier ${identifier} is already in use (${++i}/${maxTries}).`)
|
|
if (i < maxTries) return access(i)
|
|
// eslint-disable-next-line prefer-promise-reject-errors
|
|
return reject('Sorry, we could not allocate a unique random name. Try again?')
|
|
}
|
|
set.add(identifier)
|
|
// console.log(`Added ${identifier} to identifiers cache`)
|
|
return resolve(identifier + extension)
|
|
} else {
|
|
// Less stricter collision check, as in the same identifier
|
|
// can be used by multiple different extensions
|
|
const name = identifier + extension
|
|
fs.access(path.join(uploadsDir, name), error => {
|
|
if (error) return resolve(name)
|
|
console.log(`A file named ${name} already exists (${++i}/${maxTries}).`)
|
|
if (i < maxTries) return access(i)
|
|
// eslint-disable-next-line prefer-promise-reject-errors
|
|
return reject('Sorry, we could not allocate a unique random name. Try again?')
|
|
})
|
|
}
|
|
}
|
|
access(0)
|
|
})
|
|
}
|
|
|
|
uploadsController.upload = async (req, res, next) => {
|
|
let user
|
|
if (config.private === true) {
|
|
user = await utils.authorize(req, res)
|
|
if (!user) return
|
|
} else if (req.headers.token) {
|
|
user = await db.table('users').where('token', req.headers.token).first()
|
|
}
|
|
|
|
if (user && (user.enabled === false || user.enabled === 0))
|
|
return res.json({ success: false, description: 'This account has been disabled.' })
|
|
|
|
if (user && user.fileLength && !req.headers.filelength)
|
|
req.headers.filelength = user.fileLength
|
|
|
|
let albumid = parseInt(req.headers.albumid || req.params.albumid)
|
|
if (isNaN(albumid)) albumid = null
|
|
|
|
if (req.body.urls)
|
|
return uploadsController.actuallyUploadByUrl(req, res, user, albumid)
|
|
else
|
|
return uploadsController.actuallyUpload(req, res, user, albumid)
|
|
}
|
|
|
|
uploadsController.actuallyUpload = async (req, res, user, albumid) => {
|
|
const erred = error => {
|
|
const isError = error instanceof Error
|
|
if (isError) console.error(error)
|
|
res.status(400).json({
|
|
success: false,
|
|
description: isError ? error.toString() : error
|
|
})
|
|
}
|
|
|
|
upload(req, res, async error => {
|
|
if (error) {
|
|
const expected = [
|
|
'LIMIT_FILE_SIZE',
|
|
'LIMIT_UNEXPECTED_FILE'
|
|
]
|
|
if (expected.includes(error.code)) return erred(error.toString())
|
|
return erred(error)
|
|
}
|
|
|
|
if (!req.files || !req.files.length) return erred('No files.')
|
|
|
|
// If chunked uploads is enabled and the uploaded file is a chunk, then just say that it was a success
|
|
if (chunkedUploads && req.body.uuid) return res.json({ success: true })
|
|
|
|
const infoMap = req.files.map(file => {
|
|
file.albumid = albumid
|
|
return {
|
|
path: path.join(__dirname, '..', config.uploads.folder, file.filename),
|
|
data: file
|
|
}
|
|
})
|
|
|
|
if (config.filterEmptyFile && infoMap.some(file => file.data.size === 0)) {
|
|
infoMap.forEach(file => {
|
|
utils.deleteFile(file.data.filename, req.app.get('uploads-set')).catch(console.error)
|
|
})
|
|
return erred('Empty files are not allowed.')
|
|
}
|
|
|
|
if (config.uploads.scan && config.uploads.scan.enabled) {
|
|
const scan = await uploadsController.scanFiles(req, infoMap)
|
|
if (scan) return erred(scan)
|
|
}
|
|
|
|
const result = await uploadsController.formatInfoMap(req, res, user, infoMap)
|
|
.catch(erred)
|
|
if (!result) return
|
|
|
|
uploadsController.processFilesForDisplay(req, res, result.files, result.existingFiles)
|
|
})
|
|
}
|
|
|
|
uploadsController.actuallyUploadByUrl = async (req, res, user, albumid) => {
|
|
const erred = error => {
|
|
const isError = error instanceof Error
|
|
if (isError) console.error(error)
|
|
res.status(400).json({
|
|
success: false,
|
|
description: isError ? error.toString() : error
|
|
})
|
|
}
|
|
|
|
if (!config.uploads.urlMaxSize) return erred('Upload by URLs is disabled at the moment.')
|
|
|
|
const urls = req.body.urls
|
|
if (!urls || !(urls instanceof Array)) return erred('Missing "urls" property (Array).')
|
|
|
|
let iteration = 0
|
|
const infoMap = []
|
|
for (let url of urls) {
|
|
const original = path.basename(url).split(/[?#]/)[0]
|
|
const extname = utils.extname(original)
|
|
|
|
// Extensions filter
|
|
let filtered = false
|
|
if (['blacklist', 'whitelist'].includes(config.uploads.urlExtensionsFilterMode))
|
|
if (Array.isArray(config.uploads.urlExtensionsFilter) && config.uploads.urlExtensionsFilter.length) {
|
|
const match = config.uploads.urlExtensionsFilter.some(extension => extname === extension.toLowerCase())
|
|
const whitelist = config.uploads.urlExtensionsFilterMode === 'whitelist'
|
|
filtered = ((!whitelist && match) || (whitelist && !match))
|
|
} else {
|
|
return erred('config.uploads.urlExtensionsFilter is not an array or is an empty array, please contact site owner.')
|
|
}
|
|
else filtered = uploadsController.isExtensionFiltered(extname)
|
|
|
|
if (filtered)
|
|
return erred(`${extname ? `${extname.substr(1).toUpperCase()} files` : 'Files with no extension'} are not permitted due to security reasons.`)
|
|
|
|
if (config.uploads.urlProxy)
|
|
url = config.uploads.urlProxy
|
|
.replace(/{url}/g, encodeURIComponent(url))
|
|
.replace(/{url-noprot}/g, encodeURIComponent(url.replace(/^https?:\/\//, '')))
|
|
|
|
try {
|
|
const fetchHead = await fetch(url, { method: 'HEAD' })
|
|
if (fetchHead.status !== 200)
|
|
return erred(`${fetchHead.status} ${fetchHead.statusText}`)
|
|
|
|
const headers = fetchHead.headers
|
|
const size = parseInt(headers.get('content-length'))
|
|
if (isNaN(size))
|
|
return erred('URLs with missing Content-Length HTTP header are not supported.')
|
|
|
|
if (size > urlMaxSizeBytes)
|
|
return erred('File too large.')
|
|
|
|
if (config.filterEmptyFile && size === 0)
|
|
return erred('Empty files are not allowed.')
|
|
|
|
// Limit max response body size with the size reported by Content-Length
|
|
const fetchFile = await fetch(url, { size })
|
|
if (fetchFile.status !== 200)
|
|
return erred(`${fetchHead.status} ${fetchHead.statusText}`)
|
|
|
|
const file = await fetchFile.buffer()
|
|
|
|
const length = uploadsController.getFileNameLength(req)
|
|
const name = await uploadsController.getUniqueRandomName(length, extname, req.app.get('uploads-set'))
|
|
|
|
const destination = path.join(uploadsDir, name)
|
|
fs.writeFile(destination, file, async error => {
|
|
if (error) return erred(error)
|
|
|
|
const data = {
|
|
filename: name,
|
|
originalname: original,
|
|
mimetype: headers.get('content-type').split(';')[0] || '',
|
|
size,
|
|
albumid
|
|
}
|
|
|
|
infoMap.push({
|
|
path: destination,
|
|
data
|
|
})
|
|
|
|
iteration++
|
|
if (iteration === urls.length) {
|
|
if (config.uploads.scan && config.uploads.scan.enabled) {
|
|
const scan = await uploadsController.scanFiles(req, infoMap)
|
|
if (scan) return erred(scan)
|
|
}
|
|
|
|
const result = await uploadsController.formatInfoMap(req, res, user, infoMap)
|
|
.catch(erred)
|
|
if (!result) return
|
|
|
|
uploadsController.processFilesForDisplay(req, res, result.files, result.existingFiles)
|
|
}
|
|
})
|
|
} catch (error) {
|
|
erred(error)
|
|
}
|
|
}
|
|
}
|
|
|
|
uploadsController.finishChunks = async (req, res, next) => {
|
|
if (!chunkedUploads)
|
|
return res.json({ success: false, description: 'Chunked upload is disabled at the moment.' })
|
|
|
|
let user
|
|
if (config.private === true) {
|
|
user = await utils.authorize(req, res)
|
|
if (!user) return
|
|
} else if (req.headers.token) {
|
|
user = await db.table('users').where('token', req.headers.token).first()
|
|
}
|
|
|
|
if (user && (user.enabled === false || user.enabled === 0))
|
|
return res.json({ success: false, description: 'This account has been disabled.' })
|
|
|
|
if (user && user.fileLength && !req.headers.filelength)
|
|
req.headers.filelength = user.fileLength
|
|
|
|
let albumid = parseInt(req.headers.albumid || req.params.albumid)
|
|
if (isNaN(albumid)) albumid = null
|
|
|
|
return uploadsController.actuallyFinishChunks(req, res, user, albumid)
|
|
}
|
|
|
|
uploadsController.actuallyFinishChunks = async (req, res, user, albumid) => {
|
|
const erred = error => {
|
|
const isError = error instanceof Error
|
|
if (isError) console.error(error)
|
|
res.status(400).json({
|
|
success: false,
|
|
description: isError ? error.toString() : error
|
|
})
|
|
}
|
|
|
|
const files = req.body.files
|
|
if (!files || !(files instanceof Array) || !files.length) return erred('Invalid "files" property (Array).')
|
|
|
|
let iteration = 0
|
|
const infoMap = []
|
|
for (const file of files) {
|
|
if (!file.uuid || typeof file.uuid !== 'string') return erred('Invalid "uuid" property (string).')
|
|
if (typeof file.count !== 'number' || file.count < 1) return erred('Invalid "count" property (number).')
|
|
|
|
const uuidDir = path.join(chunksDir, file.uuid)
|
|
fs.readdir(uuidDir, async (error, chunkNames) => {
|
|
if (error) {
|
|
if (error.code === 'ENOENT') return erred('UUID is not being used.')
|
|
return erred(error)
|
|
}
|
|
if (file.count < chunkNames.length) return erred('Chunks count mismatch.')
|
|
|
|
const extname = typeof file.original === 'string' ? utils.extname(file.original) : ''
|
|
if (uploadsController.isExtensionFiltered(extname))
|
|
return erred(`${extname ? `${extname.substr(1).toUpperCase()} files` : 'Files with no extension'} are not permitted due to security reasons.`)
|
|
|
|
const length = uploadsController.getFileNameLength(req)
|
|
const name = await uploadsController.getUniqueRandomName(length, extname, req.app.get('uploads-set'))
|
|
.catch(erred)
|
|
if (!name) return
|
|
|
|
const destination = path.join(uploadsDir, name)
|
|
|
|
// Sort chunk names
|
|
chunkNames.sort()
|
|
|
|
// Get total chunks size
|
|
const chunksTotalSize = await uploadsController.getTotalSize(uuidDir, chunkNames)
|
|
.catch(erred)
|
|
if (typeof chunksTotalSize !== 'number') return
|
|
|
|
const isEmpty = config.filterEmptyFile && (chunksTotalSize === 0)
|
|
const isBigger = chunksTotalSize > maxSizeBytes
|
|
if (isEmpty || isBigger) {
|
|
// Delete all chunks and remove chunks dir
|
|
const chunksCleaned = await uploadsController.cleanUpChunks(uuidDir, chunkNames)
|
|
.catch(erred)
|
|
if (!chunksCleaned) return
|
|
|
|
if (isEmpty)
|
|
return erred('Empty files are not allowed.')
|
|
else
|
|
return erred(`Total chunks size is bigger than ${maxSize}.`)
|
|
}
|
|
|
|
// Append all chunks
|
|
const destFileStream = fs.createWriteStream(destination, { flags: 'a' })
|
|
const chunksAppended = await uploadsController.appendToStream(destFileStream, uuidDir, chunkNames)
|
|
.catch(erred)
|
|
if (!chunksAppended) return
|
|
|
|
// Delete all chunks and remove chunks dir
|
|
const chunksCleaned = await uploadsController.cleanUpChunks(uuidDir, chunkNames)
|
|
.catch(erred)
|
|
if (!chunksCleaned) return
|
|
|
|
const data = {
|
|
filename: name,
|
|
originalname: file.original || '',
|
|
mimetype: file.type || '',
|
|
size: file.size || 0
|
|
}
|
|
|
|
data.albumid = parseInt(file.albumid)
|
|
if (isNaN(data.albumid)) data.albumid = albumid
|
|
|
|
infoMap.push({
|
|
path: destination,
|
|
data
|
|
})
|
|
|
|
iteration++
|
|
if (iteration === files.length) {
|
|
if (config.uploads.scan && config.uploads.scan.enabled) {
|
|
const scan = await uploadsController.scanFiles(req, infoMap)
|
|
if (scan) return erred(scan)
|
|
}
|
|
|
|
const result = await uploadsController.formatInfoMap(req, res, user, infoMap)
|
|
.catch(erred)
|
|
if (!result) return
|
|
|
|
uploadsController.processFilesForDisplay(req, res, result.files, result.existingFiles)
|
|
}
|
|
})
|
|
}
|
|
}
|
|
|
|
uploadsController.getTotalSize = (uuidDir, chunkNames) => {
|
|
return new Promise((resolve, reject) => {
|
|
let size = 0
|
|
const stat = i => {
|
|
if (i === chunkNames.length) return resolve(size)
|
|
fs.stat(path.join(uuidDir, chunkNames[i]), (error, stats) => {
|
|
if (error) return reject(error)
|
|
size += stats.size
|
|
stat(i + 1)
|
|
})
|
|
}
|
|
stat(0)
|
|
})
|
|
}
|
|
|
|
uploadsController.appendToStream = (destFileStream, uuidDr, chunkNames) => {
|
|
return new Promise((resolve, reject) => {
|
|
const append = i => {
|
|
if (i === chunkNames.length) {
|
|
destFileStream.end()
|
|
return resolve(true)
|
|
}
|
|
fs.createReadStream(path.join(uuidDr, chunkNames[i]))
|
|
.on('end', () => {
|
|
append(i + 1)
|
|
})
|
|
.on('error', error => {
|
|
console.error(error)
|
|
destFileStream.end()
|
|
return reject(error)
|
|
})
|
|
.pipe(destFileStream, { end: false })
|
|
}
|
|
append(0)
|
|
})
|
|
}
|
|
|
|
uploadsController.cleanUpChunks = (uuidDir, chunkNames) => {
|
|
return new Promise(async (resolve, reject) => {
|
|
await Promise.all(chunkNames.map(chunkName => {
|
|
return new Promise((resolve, reject) => {
|
|
const chunkPath = path.join(uuidDir, chunkName)
|
|
fs.unlink(chunkPath, error => {
|
|
if (error && error.code !== 'ENOENT') {
|
|
console.error(error)
|
|
return reject(error)
|
|
}
|
|
resolve()
|
|
})
|
|
})
|
|
})).catch(reject)
|
|
fs.rmdir(uuidDir, error => {
|
|
if (error) return reject(error)
|
|
resolve(true)
|
|
})
|
|
})
|
|
}
|
|
|
|
uploadsController.formatInfoMap = (req, res, user, infoMap) => {
|
|
return new Promise(async resolve => {
|
|
let iteration = 0
|
|
const files = []
|
|
const existingFiles = []
|
|
const albumsAuthorized = {}
|
|
|
|
for (const info of infoMap) {
|
|
// Check if the file exists by checking hash and size
|
|
const hash = crypto.createHash('md5')
|
|
const stream = fs.createReadStream(info.path)
|
|
|
|
stream.on('data', data => {
|
|
hash.update(data, 'utf8')
|
|
})
|
|
|
|
stream.on('end', async () => {
|
|
const fileHash = hash.digest('hex')
|
|
const dbFile = await db.table('files')
|
|
.where(function () {
|
|
if (user === undefined)
|
|
this.whereNull('userid')
|
|
else
|
|
this.where('userid', user.id)
|
|
})
|
|
.where({
|
|
hash: fileHash,
|
|
size: info.data.size
|
|
})
|
|
.first()
|
|
|
|
if (!dbFile) {
|
|
if (info.data.albumid && albumsAuthorized[info.data.albumid] === undefined) {
|
|
const authorized = await db.table('albums')
|
|
.where({
|
|
id: info.data.albumid,
|
|
userid: user.id
|
|
})
|
|
.first()
|
|
albumsAuthorized[info.data.albumid] = Boolean(authorized)
|
|
}
|
|
|
|
files.push({
|
|
name: info.data.filename,
|
|
original: info.data.originalname,
|
|
type: info.data.mimetype,
|
|
size: info.data.size,
|
|
hash: fileHash,
|
|
ip: req.ip,
|
|
albumid: albumsAuthorized[info.data.albumid] ? info.data.albumid : null,
|
|
userid: user !== undefined ? user.id : null,
|
|
timestamp: Math.floor(Date.now() / 1000)
|
|
})
|
|
} else {
|
|
utils.deleteFile(info.data.filename, req.app.get('uploads-set')).catch(console.error)
|
|
existingFiles.push(dbFile)
|
|
}
|
|
|
|
iteration++
|
|
if (iteration === infoMap.length)
|
|
resolve({ files, existingFiles })
|
|
})
|
|
}
|
|
})
|
|
}
|
|
|
|
uploadsController.scanFiles = (req, infoMap) => {
|
|
return new Promise(async (resolve, reject) => {
|
|
const scanner = req.app.get('clam-scanner')
|
|
let iteration = 0
|
|
for (const info of infoMap)
|
|
scanner.scanFile(info.path).then(reply => {
|
|
if (!reply.includes('OK') || reply.includes('FOUND')) {
|
|
// eslint-disable-next-line no-control-regex
|
|
const virus = reply.replace(/^stream: /, '').replace(/ FOUND\u0000$/, '')
|
|
console.log(`ClamAV: ${info.data.filename}: ${virus} FOUND.`)
|
|
return resolve(virus)
|
|
}
|
|
|
|
iteration++
|
|
if (iteration === infoMap.length)
|
|
resolve(null)
|
|
}).catch(reject)
|
|
}).then(virus => {
|
|
if (!virus) return false
|
|
// If there is at least one dirty file, then delete all files
|
|
const set = req.app.get('uploads-set')
|
|
infoMap.forEach(info => {
|
|
utils.deleteFile(info.data.filename).catch(console.error)
|
|
if (set) {
|
|
const identifier = info.data.filename.split('.')[0]
|
|
set.delete(identifier)
|
|
// console.log(`Removed ${identifier} from identifiers cache (formatInfoMap)`)
|
|
}
|
|
})
|
|
// Unfortunately, we will only be returning name of the first virus
|
|
// even if the current session was made up by multiple virus types
|
|
return `Virus detected: ${virus}.`
|
|
}).catch(error => {
|
|
console.error(`ClamAV: ${error.toString()}.`)
|
|
return `ClamAV: ${error.code}, please contact site owner.`
|
|
})
|
|
}
|
|
|
|
uploadsController.processFilesForDisplay = async (req, res, files, existingFiles) => {
|
|
const responseFiles = []
|
|
|
|
if (files.length) {
|
|
// Insert new files to DB
|
|
await db.table('files').insert(files)
|
|
|
|
for (const file of files)
|
|
responseFiles.push(file)
|
|
}
|
|
|
|
if (existingFiles.length)
|
|
for (const file of existingFiles)
|
|
responseFiles.push(file)
|
|
|
|
// We send response first before generating thumbnails and updating album timestamps
|
|
const nojs = req.path === '/nojs'
|
|
res.json({
|
|
success: true,
|
|
files: responseFiles.map(file => {
|
|
const result = {
|
|
name: file.name,
|
|
size: file.size,
|
|
url: `${config.domain}/${file.name}`
|
|
}
|
|
// Add original name if it's /nojs route
|
|
if (nojs) result.original = file.original
|
|
return result
|
|
})
|
|
})
|
|
|
|
const albumids = []
|
|
for (const file of files) {
|
|
if (file.albumid && !albumids.includes(file.albumid))
|
|
albumids.push(file.albumid)
|
|
|
|
if (utils.mayGenerateThumb(utils.extname(file.name)))
|
|
utils.generateThumbs(file.name)
|
|
}
|
|
|
|
if (albumids.length)
|
|
db.table('albums')
|
|
.whereIn('id', albumids)
|
|
.update('editedAt', Math.floor(Date.now() / 1000))
|
|
.catch(console.error)
|
|
}
|
|
|
|
uploadsController.delete = async (req, res) => {
|
|
const id = parseInt(req.body.id)
|
|
const body = {
|
|
field: 'id',
|
|
values: isNaN(id) ? undefined : [id]
|
|
}
|
|
req.body = body
|
|
return uploadsController.bulkDelete(req, res)
|
|
}
|
|
|
|
uploadsController.bulkDelete = async (req, res) => {
|
|
const user = await utils.authorize(req, res)
|
|
if (!user) return
|
|
|
|
const field = req.body.field || 'id'
|
|
const values = req.body.values
|
|
|
|
if (!values || !Array.isArray(values) || !values.length)
|
|
return res.json({ success: false, description: 'No array of files specified.' })
|
|
|
|
const failed = await utils.bulkDeleteFiles(field, values, user, req.app.get('uploads-set'))
|
|
if (failed.length < values.length)
|
|
return res.json({ success: true, failed })
|
|
|
|
return res.json({ success: false, description: 'Could not delete any files.' })
|
|
}
|
|
|
|
uploadsController.list = async (req, res) => {
|
|
const user = await utils.authorize(req, res)
|
|
if (!user) return
|
|
|
|
// Headers is string-only, this seem to be the safest and lightest
|
|
const all = req.headers.all === '1'
|
|
const uploader = req.headers.uploader
|
|
const ismoderator = perms.is(user, 'moderator')
|
|
if ((all || uploader) && !ismoderator) return res.status(403).end()
|
|
|
|
let uploaderID = null
|
|
if (uploader)
|
|
uploaderID = await db.table('users')
|
|
.where('username', uploader)
|
|
.select('id')
|
|
.first()
|
|
.then(row => row ? row.id : null)
|
|
|
|
function filter () {
|
|
if (req.params.id === undefined)
|
|
this.where('id', '<>', '')
|
|
else
|
|
this.where('albumid', req.params.id)
|
|
if (!ismoderator || !all)
|
|
this.where('userid', user.id)
|
|
else if (uploaderID)
|
|
this.where('userid', uploaderID)
|
|
}
|
|
|
|
const count = await db.table('files')
|
|
.where(filter)
|
|
.count('id as count')
|
|
.then(rows => rows[0].count)
|
|
if (!count) return res.json({ success: true, files: [], count })
|
|
|
|
let offset = req.params.page
|
|
if (offset === undefined) offset = 0
|
|
|
|
const files = await db.table('files')
|
|
.where(filter)
|
|
.orderBy('id', 'DESC')
|
|
.limit(25)
|
|
.offset(25 * offset)
|
|
.select('id', 'albumid', 'timestamp', 'name', 'userid', 'size')
|
|
|
|
const albums = await db.table('albums')
|
|
.where(function () {
|
|
this.where('enabled', 1)
|
|
if (!all || !ismoderator)
|
|
this.where('userid', user.id)
|
|
})
|
|
|
|
const basedomain = config.domain
|
|
const userids = []
|
|
|
|
for (const file of files) {
|
|
file.file = `${basedomain}/${file.name}`
|
|
|
|
file.album = ''
|
|
if (file.albumid !== undefined)
|
|
for (const album of albums)
|
|
if (file.albumid === album.id)
|
|
file.album = album.name
|
|
|
|
// Only push usernames if we are a moderator
|
|
if (all && ismoderator)
|
|
if (file.userid !== undefined && file.userid !== null && file.userid !== '')
|
|
userids.push(file.userid)
|
|
|
|
file.extname = utils.extname(file.name)
|
|
if (utils.mayGenerateThumb(file.extname))
|
|
file.thumb = `${basedomain}/thumbs/${file.name.slice(0, -file.extname.length)}.png`
|
|
}
|
|
|
|
// If we are a normal user, send response
|
|
if (!ismoderator) return res.json({ success: true, files, count })
|
|
|
|
// If we are a moderator but there are no uploads attached to a user, send response
|
|
if (userids.length === 0) return res.json({ success: true, files, count })
|
|
|
|
const users = await db.table('users').whereIn('id', userids)
|
|
for (const dbUser of users)
|
|
for (const file of files)
|
|
if (file.userid === dbUser.id)
|
|
file.username = dbUser.username
|
|
|
|
return res.json({ success: true, files, count })
|
|
}
|
|
|
|
module.exports = uploadsController
|