fs.copyFile() for chunks data on non-default path

Closes #314
This commit is contained in:
Bobby Wibowo 2020-11-21 06:31:36 +07:00
parent 95247daa80
commit 51ab9a6fc5
No known key found for this signature in database
GPG Key ID: 51C3A1E1E22D26CF
2 changed files with 11 additions and 3 deletions

View File

@ -9,6 +9,7 @@ const self = {}
// Promisify these fs functions
const fsFuncs = [
'access',
'copyFile',
'lstat',
'mkdir',
'readdir',

View File

@ -35,6 +35,8 @@ const chunkedUploadsTimeout = config.uploads.chunkSize.timeout || 1800000
const chunksData = {}
// Hard-coded min chunk size of 1 MB (e.g. 50 MB = max 50 chunks)
const maxChunksCount = maxSize
// Use fs.copyFile() instead of fs.rename() if chunks dir is NOT inside uploads dir
const chunksCopyFile = !paths.chunks.startsWith(paths.uploads)
const extensionsFilter = Array.isArray(config.extensionsFilter) &&
config.extensionsFilter.length
@ -553,8 +555,13 @@ self.actuallyFinishChunks = async (req, res, user) => {
const name = await self.getUniqueRandomName(length, file.extname)
// Move tmp file to final destination
// For fs.copyFile(), tmpfile will eventually be unlinked by self.cleanUpChunks()
const destination = path.join(paths.uploads, name)
if (chunksCopyFile) {
await paths.copyFile(tmpfile, destination)
} else {
await paths.rename(tmpfile, destination)
}
const hash = chunksData[file.uuid].hasher.digest('hex')
// Continue even when encountering errors
@ -595,7 +602,7 @@ self.actuallyFinishChunks = async (req, res, user) => {
if (chunksData[file.uuid].hasher) {
chunksData[file.uuid].hasher.dispose()
}
} catch (error) {}
} catch (_) {}
self.cleanUpChunks(file.uuid).catch(logger.error)
})