//...
archive.on("error", function (err) {
throw err;
});
archive.pipe(output);
archive.glob("**/*", {
cwd: PATHS.EXTENSION,
ignore: ["*.zip"], //ignore the output .zip file
});
//...
const fs = require('fs');
const archiver = require('archiver');
const output = fs.createWriteStream(`glitch_release_${+new Date()}.zip`);
const archive = archiver('zip', {
zlib: { level: 9 } // Sets the compression level.
});
output.on('close', () => {
// eslint-disable-next-line no-console
console.log(`${archive.pointer()} total bytes`);
// eslint-disable-next-line no-console
console.log('archiver has been finalized and the output file descriptor has closed.');
});
//...
async function(command, callback) {
//...
if (isConfigOk) {
//...
const archive = archiver('zip', {
zlib: { level: 9 }
});
archive.pipe(output);
archive.append(fs.createReadStream(lambdaPath), { name: 'index.js' });
archive.finalize();
this.log(formatLog(`Temporary Lambda function archive created at ${lambdaArchivePath}`, 'ok'));
//...
}
}
// require modules
const fs = require('fs');
const archiver = require('archiver');
// create a file to stream archive data to.
const output = fs.createWriteStream(__dirname + '/example.zip');
const archive = archiver('zip', {
zlib: { level: 9 } // Sets the compression level.
});
// listen for all archive data to be written
// 'close' event is fired only when a file descriptor is involved
output.on('close', function() {
console.log(archive.pointer() + ' total bytes');
console.log('archiver has been finalized and the output file descriptor has closed.');
});
// This event is fired when the data source is drained no matter what was the data source.
// It is not part of this library but rather from the NodeJS Stream API.
// @see: https://nodejs.org/api/stream.html#stream_event_end
output.on('end', function() {
console.log('Data has been drained');
});
// good practice to catch warnings (ie stat failures and other non-blocking errors)
archive.on('warning', function(err) {
if (err.code === 'ENOENT') {
// log warning
} else {
// throw error
throw err;
}
});
// good practice to catch this error explicitly
archive.on('error', function(err) {
throw err;
});
// pipe archive data to the file
archive.pipe(output);
// append a file from stream
const file1 = __dirname + '/file1.txt';
archive.append(fs.createReadStream(file1), { name: 'file1.txt' });
// append a file from string
archive.append('string cheese!', { name: 'file2.txt' });
// append a file from buffer
const buffer3 = Buffer.from('buff it!');
archive.append(buffer3, { name: 'file3.txt' });
// append a file
archive.file('file1.txt', { name: 'file4.txt' });
// append files from a sub-directory and naming it `new-subdir` within the archive
archive.directory('subdir/', 'new-subdir');
// append files from a sub-directory, putting its contents at the root of archive
archive.directory('subdir/', false);
// append files from a glob pattern
archive.glob('file*.txt', {cwd:__dirname});
// finalize the archive (ie we are done appending files but streams have to finish yet)
// 'close', 'end' or 'finish' may be fired right after calling this method so register to them beforehand
archive.finalize();
function compressFolder(folder, fileName) {
return new Promise((resolve, reject) => {
const output = fs.createWriteStream(fileName);
const archive = archiver('zip');
output.on('close', resolve);
archive.on('error', reject);
archive.pipe(output);
archive.directory(folder, false);
archive.finalize();
});
}