Compare commits
5 Commits
developmen
...
feature/ZE
Author | SHA1 | Date |
---|---|---|
Dora Korpar | 48d766a313 | |
Dora Korpar | 72144d77c2 | |
Dora Korpar | 5415bcffd1 | |
Dora Korpar | e73de18b21 | |
Dora Korpar | 8e0617f946 |
|
@ -1,242 +0,0 @@
|
||||||
const { errors } = require('arsenal');
|
|
||||||
const crypto = require('crypto');
|
|
||||||
const constants = require('../../../../constants');
|
|
||||||
|
|
||||||
/**
|
|
||||||
* createAggregateETag - creates ETag from concatenated MPU part ETags to
|
|
||||||
* mimic AWS
|
|
||||||
* @param {string} concatETags - string of concatenated MPU part ETags
|
|
||||||
* @param {array} partList - list of parts to complete MPU with
|
|
||||||
* @return {string} aggregateETag - final complete MPU obj ETag
|
|
||||||
*/
|
|
||||||
function createAggregateETag(concatETags, partList) {
|
|
||||||
// AWS documentation is unclear on what the MD5 is that it returns
|
|
||||||
// in the response for a complete multipart upload request.
|
|
||||||
// The docs state that they might or might not
|
|
||||||
// return the MD5 of the complete object. It appears
|
|
||||||
// they are returning the MD5 of the parts' MD5s so that is
|
|
||||||
// what we have done here. We:
|
|
||||||
// 1) concatenate the hex version of the
|
|
||||||
// individual ETags
|
|
||||||
// 2) convert the concatenated hex to binary
|
|
||||||
// 3) take the md5 of the binary
|
|
||||||
// 4) create the hex digest of the md5
|
|
||||||
// 5) add '-' plus the number of parts at the end
|
|
||||||
|
|
||||||
// Convert the concatenated hex ETags to binary
|
|
||||||
const bufferedHex = Buffer.from(concatETags, 'hex');
|
|
||||||
// Convert the buffer to a binary string
|
|
||||||
const binaryString = bufferedHex.toString('binary');
|
|
||||||
// Get the md5 of the binary string
|
|
||||||
const md5Hash = crypto.createHash('md5');
|
|
||||||
md5Hash.update(binaryString, 'binary');
|
|
||||||
// Get the hex digest of the md5
|
|
||||||
let aggregateETag = md5Hash.digest('hex');
|
|
||||||
// Add the number of parts at the end
|
|
||||||
aggregateETag = `${aggregateETag}-${partList.length}`;
|
|
||||||
|
|
||||||
return aggregateETag;
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* generateMpuPartStorageInfo - generates info needed for storage of
|
|
||||||
* completed MPU object
|
|
||||||
* @param {array} filteredPartList - list of parts filtered from metadata
|
|
||||||
* @return {object} partsInfo - contains three keys: aggregateETag,
|
|
||||||
* dataLocations, and calculatedSize
|
|
||||||
*/
|
|
||||||
function generateMpuPartStorageInfo(filteredPartList) {
|
|
||||||
// Assemble array of part locations, aggregate size
|
|
||||||
// and build string to create aggregate ETag
|
|
||||||
let calculatedSize = 0;
|
|
||||||
const dataLocations = [];
|
|
||||||
let concatETags = '';
|
|
||||||
const partsInfo = {};
|
|
||||||
|
|
||||||
filteredPartList.forEach((storedPart, index) => {
|
|
||||||
const partETagWithoutQuotes =
|
|
||||||
storedPart.ETag.slice(1, -1);
|
|
||||||
const dataStoreETag = `${index + 1}:${partETagWithoutQuotes}`;
|
|
||||||
concatETags += partETagWithoutQuotes;
|
|
||||||
|
|
||||||
// If part was put by a regular put part rather than a
|
|
||||||
// copy it is always one location. With a put part
|
|
||||||
// copy, could be multiple locations so loop over array
|
|
||||||
// of locations.
|
|
||||||
for (let j = 0; j < storedPart.locations.length; j++) {
|
|
||||||
// If the piece has parts (was a put part object
|
|
||||||
// copy) each piece will have a size attribute.
|
|
||||||
// Otherwise, the piece was put by a regular put
|
|
||||||
// part and the size the of the piece is the full
|
|
||||||
// part size.
|
|
||||||
const location = storedPart.locations[j];
|
|
||||||
// If there is no location, move on
|
|
||||||
if (!location || typeof location !== 'object') {
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
let pieceSize = Number.parseInt(storedPart.size, 10);
|
|
||||||
if (location.size) {
|
|
||||||
pieceSize = Number.parseInt(location.size, 10);
|
|
||||||
}
|
|
||||||
const pieceRetrievalInfo = {
|
|
||||||
key: location.key,
|
|
||||||
size: pieceSize,
|
|
||||||
start: calculatedSize,
|
|
||||||
dataStoreName: location.dataStoreName,
|
|
||||||
dataStoreETag,
|
|
||||||
cryptoScheme: location.sseCryptoScheme,
|
|
||||||
cipheredDataKey: location.sseCipheredDataKey,
|
|
||||||
};
|
|
||||||
dataLocations.push(pieceRetrievalInfo);
|
|
||||||
// eslint-disable-next-line no-param-reassign
|
|
||||||
calculatedSize += pieceSize;
|
|
||||||
}
|
|
||||||
});
|
|
||||||
|
|
||||||
partsInfo.aggregateETag =
|
|
||||||
createAggregateETag(concatETags, filteredPartList);
|
|
||||||
partsInfo.dataLocations = dataLocations;
|
|
||||||
partsInfo.calculatedSize = calculatedSize;
|
|
||||||
return partsInfo;
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* validateAndFilterMpuParts - validates part list sent by user and filters
|
|
||||||
* parts stored in metadata against user part list
|
|
||||||
* @param {array} storedParts - array of parts stored in metadata
|
|
||||||
* @param {array} jsonList - array of parts sent by user for completion
|
|
||||||
* @param {string} mpuOverviewKey - metadata mpu key
|
|
||||||
* @param {string} splitter - mpu key divider
|
|
||||||
* @param {object} log - Werelogs instance
|
|
||||||
* @return {object} filtersPartsObj - contains 3 keys: partList, keysToDelete,
|
|
||||||
* and extraPartLocations
|
|
||||||
*/
|
|
||||||
function validateAndFilterMpuParts(storedParts, jsonList, mpuOverviewKey,
|
|
||||||
splitter, log) {
|
|
||||||
let storedPartsCopy = [];
|
|
||||||
const filteredPartsObj = {};
|
|
||||||
filteredPartsObj.partList = [];
|
|
||||||
|
|
||||||
const keysToDelete = [];
|
|
||||||
storedParts.forEach(item => {
|
|
||||||
keysToDelete.push(item.key);
|
|
||||||
storedPartsCopy.push({
|
|
||||||
// In order to delete the part listing in the shadow
|
|
||||||
// bucket, need the full key
|
|
||||||
key: item.key,
|
|
||||||
ETag: `"${item.value.ETag}"`,
|
|
||||||
size: item.value.Size,
|
|
||||||
locations: Array.isArray(item.value.partLocations) ?
|
|
||||||
item.value.partLocations : [item.value.partLocations],
|
|
||||||
});
|
|
||||||
});
|
|
||||||
keysToDelete.push(mpuOverviewKey);
|
|
||||||
|
|
||||||
// Check list sent to make sure valid
|
|
||||||
const partLength = jsonList.Part.length;
|
|
||||||
// A user can put more parts than they end up including
|
|
||||||
// in the completed MPU but there cannot be more
|
|
||||||
// parts in the complete message than were already put
|
|
||||||
if (partLength > storedPartsCopy.length) {
|
|
||||||
filteredPartsObj.error = errors.InvalidPart;
|
|
||||||
return filteredPartsObj;
|
|
||||||
}
|
|
||||||
|
|
||||||
let extraParts = [];
|
|
||||||
const extraPartLocations = [];
|
|
||||||
|
|
||||||
for (let i = 0; i < partLength; i++) {
|
|
||||||
const part = jsonList.Part[i];
|
|
||||||
const partNumber = Number.parseInt(part.PartNumber[0], 10);
|
|
||||||
// If the complete list of parts sent with
|
|
||||||
// the complete multipart upload request is not
|
|
||||||
// in ascending order return an error
|
|
||||||
if (i > 0) {
|
|
||||||
const previousPartNumber =
|
|
||||||
Number.parseInt(jsonList.Part[i - 1].PartNumber[0], 10);
|
|
||||||
if (partNumber <= previousPartNumber) {
|
|
||||||
filteredPartsObj.error = errors.InvalidPartOrder;
|
|
||||||
return filteredPartsObj;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
let isPartUploaded = false;
|
|
||||||
while (storedPartsCopy.length > 0 && !isPartUploaded) {
|
|
||||||
const storedPart = storedPartsCopy[0];
|
|
||||||
const storedPartNumber =
|
|
||||||
Number.parseInt(storedPart.key.split(splitter)[1], 10);
|
|
||||||
|
|
||||||
if (storedPartNumber === partNumber) {
|
|
||||||
isPartUploaded = true;
|
|
||||||
filteredPartsObj.partList.push(storedPart);
|
|
||||||
|
|
||||||
let partETag = part.ETag[0].replace(/['"]/g, '');
|
|
||||||
// some clients send base64, convert to hex
|
|
||||||
// 32 chars = 16 bytes(2 chars-per-byte) = 128 bits of
|
|
||||||
// MD5 hex
|
|
||||||
if (partETag.length !== 32) {
|
|
||||||
const buffered = Buffer.from(part.ETag[0], 'base64')
|
|
||||||
.toString('hex');
|
|
||||||
partETag = `${buffered}`;
|
|
||||||
}
|
|
||||||
partETag = `"${partETag}"`;
|
|
||||||
// If list of parts sent with complete mpu request contains
|
|
||||||
// a part ETag that does not match the ETag for the part
|
|
||||||
// stored in metadata, return an error
|
|
||||||
if (partETag !== storedPart.ETag) {
|
|
||||||
filteredPartsObj.error = errors.InvalidPart;
|
|
||||||
return filteredPartsObj;
|
|
||||||
}
|
|
||||||
|
|
||||||
// If any part other than the last part is less than
|
|
||||||
// 5MB, return an error
|
|
||||||
const storedPartSize =
|
|
||||||
Number.parseInt(storedPart.size, 10);
|
|
||||||
// allow smaller parts for testing
|
|
||||||
if (process.env.MPU_TESTING) {
|
|
||||||
log.info('MPU_TESTING env variable setting',
|
|
||||||
{ setting: process.env.MPU_TESTING });
|
|
||||||
}
|
|
||||||
if (process.env.MPU_TESTING !== 'yes' &&
|
|
||||||
i < jsonList.Part.length - 1 &&
|
|
||||||
storedPartSize < constants.minimumAllowedPartSize) {
|
|
||||||
log.debug('part too small on complete mpu');
|
|
||||||
filteredPartsObj.error = errors.EntityTooSmall;
|
|
||||||
return filteredPartsObj;
|
|
||||||
}
|
|
||||||
|
|
||||||
storedPartsCopy = storedPartsCopy.splice(1);
|
|
||||||
} else {
|
|
||||||
extraParts.push(storedPart);
|
|
||||||
storedPartsCopy = storedPartsCopy.splice(1);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if (!isPartUploaded) {
|
|
||||||
filteredPartsObj.error = errors.InvalidPart;
|
|
||||||
return filteredPartsObj;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
extraParts = extraParts.concat(storedPartsCopy);
|
|
||||||
// if extra parts, need to delete the data when done with completing
|
|
||||||
// mpu so extract the info to delete here
|
|
||||||
if (extraParts.length > 0) {
|
|
||||||
extraParts.forEach(part => {
|
|
||||||
const locations = part.locations;
|
|
||||||
locations.forEach(location => {
|
|
||||||
if (!location || typeof location !== 'object') {
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
extraPartLocations.push(location);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
}
|
|
||||||
filteredPartsObj.keysToDelete = keysToDelete;
|
|
||||||
filteredPartsObj.extraPartLocations = extraPartLocations;
|
|
||||||
return filteredPartsObj;
|
|
||||||
}
|
|
||||||
|
|
||||||
module.exports = {
|
|
||||||
generateMpuPartStorageInfo,
|
|
||||||
validateAndFilterMpuParts,
|
|
||||||
createAggregateETag,
|
|
||||||
};
|
|
|
@ -1,12 +1,11 @@
|
||||||
const async = require('async');
|
const async = require('async');
|
||||||
|
const arsenal = require('arsenal');
|
||||||
const { parseString } = require('xml2js');
|
const { parseString } = require('xml2js');
|
||||||
const { errors, versioning, s3middleware } = require('arsenal');
|
const { errors, versioning, s3middleware } = require('arsenal');
|
||||||
|
|
||||||
const convertToXml = s3middleware.convertToXml;
|
const convertToXml = s3middleware.convertToXml;
|
||||||
const { pushMetric } = require('../utapi/utilities');
|
const { pushMetric } = require('../utapi/utilities');
|
||||||
const getReplicationInfo = require('./apiUtils/object/getReplicationInfo');
|
const getReplicationInfo = require('./apiUtils/object/getReplicationInfo');
|
||||||
const { validateAndFilterMpuParts, generateMpuPartStorageInfo } =
|
|
||||||
require('./apiUtils/object/processMpuParts');
|
|
||||||
const { config } = require('../Config');
|
const { config } = require('../Config');
|
||||||
const multipleBackendGateway = require('../data/multipleBackendGateway');
|
const multipleBackendGateway = require('../data/multipleBackendGateway');
|
||||||
|
|
||||||
|
@ -18,9 +17,11 @@ const { versioningPreprocessing, checkQueryVersionId }
|
||||||
const metadata = require('../metadata/wrapper');
|
const metadata = require('../metadata/wrapper');
|
||||||
const services = require('../services');
|
const services = require('../services');
|
||||||
const { metadataValidateBucketAndObj } = require('../metadata/metadataUtils');
|
const { metadataValidateBucketAndObj } = require('../metadata/metadataUtils');
|
||||||
const { skipMpuPartProcessing } = require('../data/external/utils');
|
const { skipMpuPartProcessing } = arsenal.storage.data.external.backendUtils;
|
||||||
const locationConstraintCheck
|
const locationConstraintCheck
|
||||||
= require('./apiUtils/object/locationConstraintCheck');
|
= require('./apiUtils/object/locationConstraintCheck');
|
||||||
|
const { validateAndFilterMpuParts, generateMpuPartStorageInfo } =
|
||||||
|
s3middleware.processMpuParts;
|
||||||
|
|
||||||
const logger = require('../utilities/logger');
|
const logger = require('../utilities/logger');
|
||||||
|
|
||||||
|
|
|
@ -1,621 +0,0 @@
|
||||||
const AWS = require('aws-sdk');
|
|
||||||
const { errors, s3middleware } = require('arsenal');
|
|
||||||
const werelogs = require('werelogs');
|
|
||||||
const MD5Sum = s3middleware.MD5Sum;
|
|
||||||
const getMetaHeaders = s3middleware.userMetadata.getMetaHeaders;
|
|
||||||
const createLogger = require('../multipleBackendLogger');
|
|
||||||
const { prepareStream } = require('../../api/apiUtils/object/prepareStream');
|
|
||||||
const { logHelper, removeQuotes, trimXMetaPrefix } = require('./utils');
|
|
||||||
const { config } = require('../../Config');
|
|
||||||
|
|
||||||
const missingVerIdInternalError = errors.InternalError.customizeDescription(
|
|
||||||
'Invalid state. Please ensure versioning is enabled ' +
|
|
||||||
'in AWS for the location constraint and try again.'
|
|
||||||
);
|
|
||||||
|
|
||||||
class AwsClient {
|
|
||||||
constructor(config) {
|
|
||||||
this.clientType = 'aws_s3';
|
|
||||||
this.type = 'AWS';
|
|
||||||
this._s3Params = config.s3Params;
|
|
||||||
this._awsBucketName = config.bucketName;
|
|
||||||
this._bucketMatch = config.bucketMatch;
|
|
||||||
this._dataStoreName = config.dataStoreName;
|
|
||||||
this._serverSideEncryption = config.serverSideEncryption;
|
|
||||||
this._supportsVersioning = config.supportsVersioning;
|
|
||||||
this._client = new AWS.S3(this._s3Params);
|
|
||||||
this._logger = new werelogs.Logger('AwsClient');
|
|
||||||
}
|
|
||||||
|
|
||||||
setup(cb) {
|
|
||||||
// this request implicitly updates the endpoint for the location
|
|
||||||
// the following code explcitly sets it to avoid surprises
|
|
||||||
this._client.getBucketLocation({ Bucket: this._awsBucketName },
|
|
||||||
(err, res) => {
|
|
||||||
if (err && err.code !== 'AuthorizationHeaderMalformed') {
|
|
||||||
this._logger.error('error during setup', {
|
|
||||||
error: err,
|
|
||||||
method: 'AwsClient.setup',
|
|
||||||
});
|
|
||||||
return cb(err);
|
|
||||||
}
|
|
||||||
let region;
|
|
||||||
if (err && err.code === 'AuthorizationHeaderMalformed') {
|
|
||||||
// set regional endpoint
|
|
||||||
region = err.region;
|
|
||||||
} else if (res) {
|
|
||||||
region = res.LocationConstraint;
|
|
||||||
}
|
|
||||||
const isAWS = this._s3Params.endpoint.endsWith('amazonaws.com');
|
|
||||||
if (region && isAWS) {
|
|
||||||
const endpoint = `s3.${region}.amazonaws.com`;
|
|
||||||
this._logger.debug('setting regional endpoint', {
|
|
||||||
method: 'AwsClient.setup',
|
|
||||||
region,
|
|
||||||
endpoint,
|
|
||||||
});
|
|
||||||
this._client.endpoint = new AWS.Endpoint(endpoint);
|
|
||||||
}
|
|
||||||
return cb();
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
_createAwsKey(requestBucketName, requestObjectKey,
|
|
||||||
bucketMatch) {
|
|
||||||
if (bucketMatch) {
|
|
||||||
return requestObjectKey;
|
|
||||||
}
|
|
||||||
return `${requestBucketName}/${requestObjectKey}`;
|
|
||||||
}
|
|
||||||
|
|
||||||
toObjectGetInfo(objectKey, bucketName) {
|
|
||||||
return {
|
|
||||||
key: this._createAwsKey(bucketName, objectKey, this._bucketMatch),
|
|
||||||
dataStoreName: this._dataStoreName,
|
|
||||||
};
|
|
||||||
}
|
|
||||||
|
|
||||||
put(stream, size, keyContext, reqUids, callback) {
|
|
||||||
const awsKey = this._createAwsKey(keyContext.bucketName,
|
|
||||||
keyContext.objectKey, this._bucketMatch);
|
|
||||||
const metaHeaders = trimXMetaPrefix(keyContext.metaHeaders);
|
|
||||||
const log = createLogger(reqUids);
|
|
||||||
|
|
||||||
const putCb = (err, data) => {
|
|
||||||
if (err) {
|
|
||||||
logHelper(log, 'error', 'err from data backend',
|
|
||||||
err, this._dataStoreName, this.clientType);
|
|
||||||
return callback(errors.ServiceUnavailable
|
|
||||||
.customizeDescription('Error returned from ' +
|
|
||||||
`${this.type}: ${err.message}`)
|
|
||||||
);
|
|
||||||
}
|
|
||||||
if (!data.VersionId && this._supportsVersioning) {
|
|
||||||
logHelper(log, 'error', 'missing version id for data ' +
|
|
||||||
'backend object', missingVerIdInternalError,
|
|
||||||
this._dataStoreName, this.clientType);
|
|
||||||
return callback(missingVerIdInternalError);
|
|
||||||
}
|
|
||||||
const dataStoreVersionId = data.VersionId;
|
|
||||||
return callback(null, awsKey, dataStoreVersionId);
|
|
||||||
};
|
|
||||||
|
|
||||||
const params = {
|
|
||||||
Bucket: this._awsBucketName,
|
|
||||||
Key: awsKey,
|
|
||||||
};
|
|
||||||
// we call data.put to create a delete marker, but it's actually a
|
|
||||||
// delete request in call to AWS
|
|
||||||
if (keyContext.isDeleteMarker) {
|
|
||||||
return this._client.deleteObject(params, putCb);
|
|
||||||
}
|
|
||||||
const uploadParams = params;
|
|
||||||
uploadParams.Metadata = metaHeaders;
|
|
||||||
uploadParams.ContentLength = size;
|
|
||||||
if (this._serverSideEncryption) {
|
|
||||||
uploadParams.ServerSideEncryption = 'AES256';
|
|
||||||
}
|
|
||||||
if (keyContext.tagging) {
|
|
||||||
uploadParams.Tagging = keyContext.tagging;
|
|
||||||
}
|
|
||||||
if (keyContext.contentType !== undefined) {
|
|
||||||
uploadParams.ContentType = keyContext.contentType;
|
|
||||||
}
|
|
||||||
if (keyContext.cacheControl !== undefined) {
|
|
||||||
uploadParams.CacheControl = keyContext.cacheControl;
|
|
||||||
}
|
|
||||||
if (keyContext.contentDisposition !== undefined) {
|
|
||||||
uploadParams.ContentDisposition = keyContext.contentDisposition;
|
|
||||||
}
|
|
||||||
if (keyContext.contentEncoding !== undefined) {
|
|
||||||
uploadParams.ContentEncoding = keyContext.contentEncoding;
|
|
||||||
}
|
|
||||||
if (!stream) {
|
|
||||||
return this._client.putObject(uploadParams, putCb);
|
|
||||||
}
|
|
||||||
|
|
||||||
uploadParams.Body = stream;
|
|
||||||
return this._client.upload(uploadParams, putCb);
|
|
||||||
}
|
|
||||||
head(objectGetInfo, reqUids, callback) {
|
|
||||||
const log = createLogger(reqUids);
|
|
||||||
const { key, dataStoreVersionId } = objectGetInfo;
|
|
||||||
return this._client.headObject({
|
|
||||||
Bucket: this._awsBucketName,
|
|
||||||
Key: key,
|
|
||||||
VersionId: dataStoreVersionId,
|
|
||||||
}, err => {
|
|
||||||
if (err) {
|
|
||||||
logHelper(log, 'error', 'error heading object ' +
|
|
||||||
'from datastore', err, this._dataStoreName);
|
|
||||||
if (err.code === 'NotFound') {
|
|
||||||
const error = errors.ServiceUnavailable
|
|
||||||
.customizeDescription(
|
|
||||||
`Unexpected error from ${this.type}: ` +
|
|
||||||
`"NotFound". Data on ${this.type} ` +
|
|
||||||
'may have been altered outside of CloudServer.'
|
|
||||||
);
|
|
||||||
return callback(error);
|
|
||||||
}
|
|
||||||
return callback(errors.ServiceUnavailable
|
|
||||||
.customizeDescription('Error returned from ' +
|
|
||||||
`${this.type}: ${err.message}`)
|
|
||||||
);
|
|
||||||
}
|
|
||||||
return callback();
|
|
||||||
});
|
|
||||||
}
|
|
||||||
get(objectGetInfo, range, reqUids, callback) {
|
|
||||||
const log = createLogger(reqUids);
|
|
||||||
const { key, dataStoreVersionId } = objectGetInfo;
|
|
||||||
const request = this._client.getObject({
|
|
||||||
Bucket: this._awsBucketName,
|
|
||||||
Key: key,
|
|
||||||
VersionId: dataStoreVersionId,
|
|
||||||
Range: range ? `bytes=${range[0]}-${range[1]}` : null,
|
|
||||||
}).on('success', response => {
|
|
||||||
log.trace(`${this.type} GET request response headers`,
|
|
||||||
{ responseHeaders: response.httpResponse.headers,
|
|
||||||
backendType: this.clientType });
|
|
||||||
});
|
|
||||||
const stream = request.createReadStream().on('error', err => {
|
|
||||||
logHelper(log, 'error',
|
|
||||||
`error streaming data from ${this.type}`,
|
|
||||||
err, this._dataStoreName, this.clientType);
|
|
||||||
return callback(err);
|
|
||||||
});
|
|
||||||
return callback(null, stream);
|
|
||||||
}
|
|
||||||
delete(objectGetInfo, reqUids, callback) {
|
|
||||||
const { key, dataStoreVersionId, deleteVersion } = objectGetInfo;
|
|
||||||
const log = createLogger(reqUids);
|
|
||||||
const params = {
|
|
||||||
Bucket: this._awsBucketName,
|
|
||||||
Key: key,
|
|
||||||
};
|
|
||||||
if (deleteVersion) {
|
|
||||||
params.VersionId = dataStoreVersionId;
|
|
||||||
}
|
|
||||||
return this._client.deleteObject(params, err => {
|
|
||||||
if (err) {
|
|
||||||
logHelper(log, 'error', 'error deleting object from ' +
|
|
||||||
'datastore', err, this._dataStoreName, this.clientType);
|
|
||||||
if (err.code === 'NoSuchVersion' || err.code === 'NoSuchKey') {
|
|
||||||
// data may have been deleted directly from the AWS backend
|
|
||||||
// don't want to retry the delete and errors are not
|
|
||||||
// sent back to client anyway, so no need to return err
|
|
||||||
return callback();
|
|
||||||
}
|
|
||||||
return callback(errors.ServiceUnavailable
|
|
||||||
.customizeDescription('Error returned from ' +
|
|
||||||
`${this.type}: ${err.message}`)
|
|
||||||
);
|
|
||||||
}
|
|
||||||
return callback();
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
healthcheck(location, callback) {
|
|
||||||
const awsResp = {};
|
|
||||||
this._client.headBucket({ Bucket: this._awsBucketName },
|
|
||||||
err => {
|
|
||||||
/* eslint-disable no-param-reassign */
|
|
||||||
if (err) {
|
|
||||||
awsResp[location] = { error: err, external: true };
|
|
||||||
return callback(null, awsResp);
|
|
||||||
}
|
|
||||||
if (!this._supportsVersioning) {
|
|
||||||
awsResp[location] = {
|
|
||||||
message: 'Congrats! You own the bucket',
|
|
||||||
};
|
|
||||||
return callback(null, awsResp);
|
|
||||||
}
|
|
||||||
return this._client.getBucketVersioning({
|
|
||||||
Bucket: this._awsBucketName },
|
|
||||||
(err, data) => {
|
|
||||||
if (err) {
|
|
||||||
awsResp[location] = { error: err, external: true };
|
|
||||||
} else if (!data.Status ||
|
|
||||||
data.Status === 'Suspended') {
|
|
||||||
awsResp[location] = {
|
|
||||||
versioningStatus: data.Status,
|
|
||||||
error: 'Versioning must be enabled',
|
|
||||||
external: true,
|
|
||||||
};
|
|
||||||
} else {
|
|
||||||
awsResp[location] = {
|
|
||||||
versioningStatus: data.Status,
|
|
||||||
message: 'Congrats! You own the bucket',
|
|
||||||
};
|
|
||||||
}
|
|
||||||
return callback(null, awsResp);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
createMPU(key, metaHeaders, bucketName, websiteRedirectHeader, contentType,
|
|
||||||
cacheControl, contentDisposition, contentEncoding, log, callback) {
|
|
||||||
const metaHeadersTrimmed = {};
|
|
||||||
Object.keys(metaHeaders).forEach(header => {
|
|
||||||
if (header.startsWith('x-amz-meta-')) {
|
|
||||||
const headerKey = header.substring(11);
|
|
||||||
metaHeadersTrimmed[headerKey] = metaHeaders[header];
|
|
||||||
}
|
|
||||||
});
|
|
||||||
const awsBucket = this._awsBucketName;
|
|
||||||
const awsKey = this._createAwsKey(bucketName, key, this._bucketMatch);
|
|
||||||
const params = {
|
|
||||||
Bucket: awsBucket,
|
|
||||||
Key: awsKey,
|
|
||||||
WebsiteRedirectLocation: websiteRedirectHeader,
|
|
||||||
Metadata: metaHeadersTrimmed,
|
|
||||||
ContentType: contentType,
|
|
||||||
CacheControl: cacheControl,
|
|
||||||
ContentDisposition: contentDisposition,
|
|
||||||
ContentEncoding: contentEncoding,
|
|
||||||
};
|
|
||||||
return this._client.createMultipartUpload(params, (err, mpuResObj) => {
|
|
||||||
if (err) {
|
|
||||||
logHelper(log, 'error', 'err from data backend',
|
|
||||||
err, this._dataStoreName, this.clientType);
|
|
||||||
return callback(errors.ServiceUnavailable
|
|
||||||
.customizeDescription('Error returned from ' +
|
|
||||||
`${this.type}: ${err.message}`)
|
|
||||||
);
|
|
||||||
}
|
|
||||||
return callback(null, mpuResObj);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
uploadPart(request, streamingV4Params, stream, size, key, uploadId,
|
|
||||||
partNumber, bucketName, log, callback) {
|
|
||||||
let hashedStream = stream;
|
|
||||||
if (request) {
|
|
||||||
const partStream = prepareStream(request, streamingV4Params,
|
|
||||||
log, callback);
|
|
||||||
hashedStream = new MD5Sum();
|
|
||||||
partStream.pipe(hashedStream);
|
|
||||||
}
|
|
||||||
|
|
||||||
const awsBucket = this._awsBucketName;
|
|
||||||
const awsKey = this._createAwsKey(bucketName, key, this._bucketMatch);
|
|
||||||
const params = { Bucket: awsBucket, Key: awsKey, UploadId: uploadId,
|
|
||||||
Body: hashedStream, ContentLength: size,
|
|
||||||
PartNumber: partNumber };
|
|
||||||
return this._client.uploadPart(params, (err, partResObj) => {
|
|
||||||
if (err) {
|
|
||||||
logHelper(log, 'error', 'err from data backend ' +
|
|
||||||
'on uploadPart', err, this._dataStoreName, this.clientType);
|
|
||||||
return callback(errors.ServiceUnavailable
|
|
||||||
.customizeDescription('Error returned from ' +
|
|
||||||
`${this.type}: ${err.message}`)
|
|
||||||
);
|
|
||||||
}
|
|
||||||
// Because we manually add quotes to ETag later, remove quotes here
|
|
||||||
const noQuotesETag =
|
|
||||||
partResObj.ETag.substring(1, partResObj.ETag.length - 1);
|
|
||||||
const dataRetrievalInfo = {
|
|
||||||
key: awsKey,
|
|
||||||
dataStoreType: 'aws_s3',
|
|
||||||
dataStoreName: this._dataStoreName,
|
|
||||||
dataStoreETag: noQuotesETag,
|
|
||||||
};
|
|
||||||
return callback(null, dataRetrievalInfo);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
listParts(key, uploadId, bucketName, partNumberMarker, maxParts, log,
|
|
||||||
callback) {
|
|
||||||
const awsBucket = this._awsBucketName;
|
|
||||||
const awsKey = this._createAwsKey(bucketName, key, this._bucketMatch);
|
|
||||||
const params = { Bucket: awsBucket, Key: awsKey, UploadId: uploadId,
|
|
||||||
PartNumberMarker: partNumberMarker, MaxParts: maxParts };
|
|
||||||
return this._client.listParts(params, (err, partList) => {
|
|
||||||
if (err) {
|
|
||||||
logHelper(log, 'error', 'err from data backend on listPart',
|
|
||||||
err, this._dataStoreName, this.clientType);
|
|
||||||
return callback(errors.ServiceUnavailable
|
|
||||||
.customizeDescription('Error returned from ' +
|
|
||||||
`${this.type}: ${err.message}`)
|
|
||||||
);
|
|
||||||
}
|
|
||||||
// build storedParts object to mimic Scality S3 backend returns
|
|
||||||
const storedParts = {};
|
|
||||||
storedParts.IsTruncated = partList.IsTruncated;
|
|
||||||
storedParts.Contents = [];
|
|
||||||
storedParts.Contents = partList.Parts.map(item => {
|
|
||||||
// We manually add quotes to ETag later, so remove quotes here
|
|
||||||
const noQuotesETag =
|
|
||||||
item.ETag.substring(1, item.ETag.length - 1);
|
|
||||||
return {
|
|
||||||
partNumber: item.PartNumber,
|
|
||||||
value: {
|
|
||||||
Size: item.Size,
|
|
||||||
ETag: noQuotesETag,
|
|
||||||
LastModified: item.LastModified,
|
|
||||||
},
|
|
||||||
};
|
|
||||||
});
|
|
||||||
return callback(null, storedParts);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* completeMPU - complete multipart upload on AWS backend
|
|
||||||
* @param {object} jsonList - user-sent list of parts to include in
|
|
||||||
* final mpu object
|
|
||||||
* @param {object} mdInfo - object containing 3 keys: storedParts,
|
|
||||||
* mpuOverviewKey, and splitter
|
|
||||||
* @param {string} key - object key
|
|
||||||
* @param {string} uploadId - multipart upload id string
|
|
||||||
* @param {string} bucketName - name of bucket
|
|
||||||
* @param {RequestLogger} log - logger instance
|
|
||||||
* @param {function} callback - callback function
|
|
||||||
* @return {(Error|object)} - return Error if complete MPU fails, otherwise
|
|
||||||
* object containing completed object key, eTag, and contentLength
|
|
||||||
*/
|
|
||||||
completeMPU(jsonList, mdInfo, key, uploadId, bucketName, log, callback) {
|
|
||||||
const awsBucket = this._awsBucketName;
|
|
||||||
const awsKey = this._createAwsKey(bucketName, key, this._bucketMatch);
|
|
||||||
const mpuError = {
|
|
||||||
InvalidPart: true,
|
|
||||||
InvalidPartOrder: true,
|
|
||||||
EntityTooSmall: true,
|
|
||||||
};
|
|
||||||
const partArray = [];
|
|
||||||
const partList = jsonList.Part;
|
|
||||||
partList.forEach(partObj => {
|
|
||||||
const partParams = { PartNumber: partObj.PartNumber[0],
|
|
||||||
ETag: partObj.ETag[0] };
|
|
||||||
partArray.push(partParams);
|
|
||||||
});
|
|
||||||
const mpuParams = {
|
|
||||||
Bucket: awsBucket, Key: awsKey, UploadId: uploadId,
|
|
||||||
MultipartUpload: {
|
|
||||||
Parts: partArray,
|
|
||||||
},
|
|
||||||
};
|
|
||||||
const completeObjData = { key: awsKey };
|
|
||||||
return this._client.completeMultipartUpload(mpuParams,
|
|
||||||
(err, completeMpuRes) => {
|
|
||||||
if (err) {
|
|
||||||
if (mpuError[err.code]) {
|
|
||||||
logHelper(log, 'trace', 'err from data backend on ' +
|
|
||||||
'completeMPU', err, this._dataStoreName, this.clientType);
|
|
||||||
return callback(errors[err.code]);
|
|
||||||
}
|
|
||||||
logHelper(log, 'error', 'err from data backend on ' +
|
|
||||||
'completeMPU', err, this._dataStoreName, this.clientType);
|
|
||||||
return callback(errors.ServiceUnavailable
|
|
||||||
.customizeDescription('Error returned from ' +
|
|
||||||
`${this.type}: ${err.message}`)
|
|
||||||
);
|
|
||||||
}
|
|
||||||
if (!completeMpuRes.VersionId && this._supportsVersioning) {
|
|
||||||
logHelper(log, 'error', 'missing version id for data ' +
|
|
||||||
'backend object', missingVerIdInternalError,
|
|
||||||
this._dataStoreName, this.clientType);
|
|
||||||
return callback(missingVerIdInternalError);
|
|
||||||
}
|
|
||||||
// need to get content length of new object to store
|
|
||||||
// in our metadata
|
|
||||||
return this._client.headObject({ Bucket: awsBucket, Key: awsKey },
|
|
||||||
(err, objHeaders) => {
|
|
||||||
if (err) {
|
|
||||||
logHelper(log, 'trace', 'err from data backend on ' +
|
|
||||||
'headObject', err, this._dataStoreName, this.clientType);
|
|
||||||
return callback(errors.ServiceUnavailable
|
|
||||||
.customizeDescription('Error returned from ' +
|
|
||||||
`${this.type}: ${err.message}`)
|
|
||||||
);
|
|
||||||
}
|
|
||||||
// remove quotes from eTag because they're added later
|
|
||||||
completeObjData.eTag = completeMpuRes.ETag
|
|
||||||
.substring(1, completeMpuRes.ETag.length - 1);
|
|
||||||
completeObjData.dataStoreVersionId = completeMpuRes.VersionId;
|
|
||||||
completeObjData.contentLength =
|
|
||||||
Number.parseInt(objHeaders.ContentLength, 10);
|
|
||||||
return callback(null, completeObjData);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
abortMPU(key, uploadId, bucketName, log, callback) {
|
|
||||||
const awsBucket = this._awsBucketName;
|
|
||||||
const awsKey = this._createAwsKey(bucketName, key, this._bucketMatch);
|
|
||||||
const abortParams = {
|
|
||||||
Bucket: awsBucket, Key: awsKey, UploadId: uploadId,
|
|
||||||
};
|
|
||||||
return this._client.abortMultipartUpload(abortParams, err => {
|
|
||||||
if (err) {
|
|
||||||
logHelper(log, 'error', 'There was an error aborting ' +
|
|
||||||
'the MPU on AWS S3. You should abort directly on AWS S3 ' +
|
|
||||||
'using the same uploadId.', err, this._dataStoreName,
|
|
||||||
this.clientType);
|
|
||||||
return callback(errors.ServiceUnavailable
|
|
||||||
.customizeDescription('Error returned from ' +
|
|
||||||
`${this.type}: ${err.message}`)
|
|
||||||
);
|
|
||||||
}
|
|
||||||
return callback();
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
objectPutTagging(key, bucket, objectMD, log, callback) {
|
|
||||||
const awsBucket = this._awsBucketName;
|
|
||||||
const awsKey = this._createAwsKey(bucket, key, this._bucketMatch);
|
|
||||||
const dataStoreVersionId = objectMD.location[0].dataStoreVersionId;
|
|
||||||
const tagParams = {
|
|
||||||
Bucket: awsBucket,
|
|
||||||
Key: awsKey,
|
|
||||||
VersionId: dataStoreVersionId,
|
|
||||||
};
|
|
||||||
const keyArray = Object.keys(objectMD.tags);
|
|
||||||
tagParams.Tagging = {};
|
|
||||||
tagParams.Tagging.TagSet = keyArray.map(key => {
|
|
||||||
const value = objectMD.tags[key];
|
|
||||||
return { Key: key, Value: value };
|
|
||||||
});
|
|
||||||
return this._client.putObjectTagging(tagParams, err => {
|
|
||||||
if (err) {
|
|
||||||
logHelper(log, 'error', 'error from data backend on ' +
|
|
||||||
'putObjectTagging', err,
|
|
||||||
this._dataStoreName, this.clientType);
|
|
||||||
return callback(errors.ServiceUnavailable
|
|
||||||
.customizeDescription('Error returned from ' +
|
|
||||||
`${this.type}: ${err.message}`)
|
|
||||||
);
|
|
||||||
}
|
|
||||||
return callback();
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
objectDeleteTagging(key, bucket, objectMD, log, callback) {
|
|
||||||
const awsBucket = this._awsBucketName;
|
|
||||||
const awsKey = this._createAwsKey(bucket, key, this._bucketMatch);
|
|
||||||
const dataStoreVersionId = objectMD.location[0].dataStoreVersionId;
|
|
||||||
const tagParams = {
|
|
||||||
Bucket: awsBucket,
|
|
||||||
Key: awsKey,
|
|
||||||
VersionId: dataStoreVersionId,
|
|
||||||
};
|
|
||||||
return this._client.deleteObjectTagging(tagParams, err => {
|
|
||||||
if (err) {
|
|
||||||
logHelper(log, 'error', 'error from data backend on ' +
|
|
||||||
'deleteObjectTagging', err,
|
|
||||||
this._dataStoreName, this.clientType);
|
|
||||||
return callback(errors.ServiceUnavailable
|
|
||||||
.customizeDescription('Error returned from ' +
|
|
||||||
`${this.type}: ${err.message}`)
|
|
||||||
);
|
|
||||||
}
|
|
||||||
return callback();
|
|
||||||
});
|
|
||||||
}
|
|
||||||
copyObject(request, destLocationConstraintName, sourceKey,
|
|
||||||
sourceLocationConstraintName, storeMetadataParams, log, callback) {
|
|
||||||
const destBucketName = request.bucketName;
|
|
||||||
const destObjectKey = request.objectKey;
|
|
||||||
const destAwsKey = this._createAwsKey(destBucketName, destObjectKey,
|
|
||||||
this._bucketMatch);
|
|
||||||
|
|
||||||
const sourceAwsBucketName =
|
|
||||||
config.getAwsBucketName(sourceLocationConstraintName);
|
|
||||||
|
|
||||||
const metadataDirective = request.headers['x-amz-metadata-directive'];
|
|
||||||
const metaHeaders = trimXMetaPrefix(getMetaHeaders(request.headers));
|
|
||||||
const awsParams = {
|
|
||||||
Bucket: this._awsBucketName,
|
|
||||||
Key: destAwsKey,
|
|
||||||
CopySource: `${sourceAwsBucketName}/${sourceKey}`,
|
|
||||||
Metadata: metaHeaders,
|
|
||||||
MetadataDirective: metadataDirective,
|
|
||||||
};
|
|
||||||
if (destLocationConstraintName &&
|
|
||||||
config.isAWSServerSideEncrytion(destLocationConstraintName)) {
|
|
||||||
awsParams.ServerSideEncryption = 'AES256';
|
|
||||||
}
|
|
||||||
this._client.copyObject(awsParams, (err, copyResult) => {
|
|
||||||
if (err) {
|
|
||||||
if (err.code === 'AccessDenied') {
|
|
||||||
logHelper(log, 'error', 'Unable to access ' +
|
|
||||||
`${sourceAwsBucketName} ${this.type} bucket`, err,
|
|
||||||
this._dataStoreName, this.clientType);
|
|
||||||
return callback(errors.AccessDenied
|
|
||||||
.customizeDescription('Error: Unable to access ' +
|
|
||||||
`${sourceAwsBucketName} ${this.type} bucket`)
|
|
||||||
);
|
|
||||||
}
|
|
||||||
logHelper(log, 'error', 'error from data backend on ' +
|
|
||||||
'copyObject', err, this._dataStoreName, this.clientType);
|
|
||||||
return callback(errors.ServiceUnavailable
|
|
||||||
.customizeDescription('Error returned from ' +
|
|
||||||
`${this.type}: ${err.message}`)
|
|
||||||
);
|
|
||||||
}
|
|
||||||
if (!copyResult.VersionId && this._supportsVersioning) {
|
|
||||||
this._logger.debug('No VersionId found in response, ' +
|
|
||||||
'calling headObject to resolve');
|
|
||||||
return this._client.headObject({
|
|
||||||
Bucket: this._awsBucketName,
|
|
||||||
Key: destAwsKey,
|
|
||||||
}, (err, data) => {
|
|
||||||
if (err || !data.VersionId) {
|
|
||||||
logHelper(log, 'error', 'missing version id for data ' +
|
|
||||||
'backend object', missingVerIdInternalError,
|
|
||||||
this._dataStoreName, this.clientType);
|
|
||||||
return callback(missingVerIdInternalError);
|
|
||||||
}
|
|
||||||
return callback(null, destAwsKey, data.VersionId);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
return callback(null, destAwsKey, copyResult.VersionId);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
uploadPartCopy(request, awsSourceKey, sourceLocationConstraintName,
|
|
||||||
log, callback) {
|
|
||||||
const destBucketName = request.bucketName;
|
|
||||||
const destObjectKey = request.objectKey;
|
|
||||||
const destAwsKey = this._createAwsKey(destBucketName, destObjectKey,
|
|
||||||
this._bucketMatch);
|
|
||||||
|
|
||||||
const sourceAwsBucketName =
|
|
||||||
config.getAwsBucketName(sourceLocationConstraintName);
|
|
||||||
|
|
||||||
const uploadId = request.query.uploadId;
|
|
||||||
const partNumber = request.query.partNumber;
|
|
||||||
const copySourceRange = request.headers['x-amz-copy-source-range'];
|
|
||||||
|
|
||||||
const params = {
|
|
||||||
Bucket: this._awsBucketName,
|
|
||||||
CopySource: `${sourceAwsBucketName}/${awsSourceKey}`,
|
|
||||||
CopySourceRange: copySourceRange,
|
|
||||||
Key: destAwsKey,
|
|
||||||
PartNumber: partNumber,
|
|
||||||
UploadId: uploadId,
|
|
||||||
};
|
|
||||||
return this._client.uploadPartCopy(params, (err, res) => {
|
|
||||||
if (err) {
|
|
||||||
if (err.code === 'AccessDenied') {
|
|
||||||
logHelper(log, 'error', 'Unable to access ' +
|
|
||||||
`${sourceAwsBucketName} AWS bucket`, err,
|
|
||||||
this._dataStoreName, this.clientType);
|
|
||||||
return callback(errors.AccessDenied
|
|
||||||
.customizeDescription('Error: Unable to access ' +
|
|
||||||
`${sourceAwsBucketName} AWS bucket`)
|
|
||||||
);
|
|
||||||
}
|
|
||||||
logHelper(log, 'error', 'error from data backend on ' +
|
|
||||||
'uploadPartCopy', err, this._dataStoreName, this.clientType);
|
|
||||||
return callback(errors.ServiceUnavailable
|
|
||||||
.customizeDescription('Error returned from ' +
|
|
||||||
`${this.type}: ${err.message}`)
|
|
||||||
);
|
|
||||||
}
|
|
||||||
const eTag = removeQuotes(res.CopyPartResult.ETag);
|
|
||||||
return callback(null, eTag);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
module.exports = AwsClient;
|
|
|
@ -1,449 +0,0 @@
|
||||||
const url = require('url');
|
|
||||||
|
|
||||||
const { errors, s3middleware } = require('arsenal');
|
|
||||||
const azure = require('azure-storage');
|
|
||||||
const createLogger = require('../multipleBackendLogger');
|
|
||||||
const { logHelper, translateAzureMetaHeaders } = require('./utils');
|
|
||||||
const { config } = require('../../Config');
|
|
||||||
const { validateAndFilterMpuParts } =
|
|
||||||
require('../../api/apiUtils/object/processMpuParts');
|
|
||||||
const constants = require('../../../constants');
|
|
||||||
const metadata = require('../../metadata/wrapper');
|
|
||||||
const packageVersion = require('../../../package.json').version;
|
|
||||||
const azureMpuUtils = s3middleware.azureHelper.mpuUtils;
|
|
||||||
|
|
||||||
azure.Constants.USER_AGENT_PRODUCT_NAME = constants.productName;
|
|
||||||
azure.Constants.USER_AGENT_PRODUCT_VERSION = packageVersion;
|
|
||||||
|
|
||||||
class AzureClient {
|
|
||||||
constructor(config) {
|
|
||||||
this._azureStorageEndpoint = config.azureStorageEndpoint;
|
|
||||||
this._azureStorageCredentials = config.azureStorageCredentials;
|
|
||||||
this._azureContainerName = config.azureContainerName;
|
|
||||||
this._client = azure.createBlobService(
|
|
||||||
this._azureStorageCredentials.storageAccountName,
|
|
||||||
this._azureStorageCredentials.storageAccessKey,
|
|
||||||
this._azureStorageEndpoint);
|
|
||||||
this._dataStoreName = config.dataStoreName;
|
|
||||||
this._bucketMatch = config.bucketMatch;
|
|
||||||
if (config.proxy && config.proxy.url) {
|
|
||||||
const parsedUrl = url.parse(config.proxy.url);
|
|
||||||
if (!parsedUrl.port) {
|
|
||||||
parsedUrl.port = 80;
|
|
||||||
}
|
|
||||||
const proxyParams = parsedUrl;
|
|
||||||
if (config.proxy.certs) {
|
|
||||||
Object.assign(proxyParams, config.proxy.certs);
|
|
||||||
}
|
|
||||||
this._client.setProxy(proxyParams);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
_errorWrapper(s3Method, azureMethod, args, log, cb) {
|
|
||||||
if (log) {
|
|
||||||
log.info(`calling azure ${azureMethod}`);
|
|
||||||
}
|
|
||||||
try {
|
|
||||||
this._client[azureMethod].apply(this._client, args);
|
|
||||||
} catch (err) {
|
|
||||||
const error = errors.ServiceUnavailable;
|
|
||||||
if (log) {
|
|
||||||
log.error('error thrown by Azure Storage Client Library',
|
|
||||||
{ error: err.message, stack: err.stack, s3Method,
|
|
||||||
azureMethod, dataStoreName: this._dataStoreName });
|
|
||||||
}
|
|
||||||
cb(error.customizeDescription('Error from Azure ' +
|
|
||||||
`method: ${azureMethod} on ${s3Method} S3 call: ` +
|
|
||||||
`${err.message}`));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
_createAzureKey(requestBucketName, requestObjectKey,
|
|
||||||
bucketMatch) {
|
|
||||||
if (bucketMatch) {
|
|
||||||
return requestObjectKey;
|
|
||||||
}
|
|
||||||
return `${requestBucketName}/${requestObjectKey}`;
|
|
||||||
}
|
|
||||||
|
|
||||||
_getMetaHeaders(objectMD) {
|
|
||||||
const metaHeaders = {};
|
|
||||||
Object.keys(objectMD).forEach(mdKey => {
|
|
||||||
const isMetaHeader = mdKey.startsWith('x-amz-meta-');
|
|
||||||
if (isMetaHeader) {
|
|
||||||
metaHeaders[mdKey] = objectMD[mdKey];
|
|
||||||
}
|
|
||||||
});
|
|
||||||
return translateAzureMetaHeaders(metaHeaders);
|
|
||||||
}
|
|
||||||
|
|
||||||
// Before putting or deleting object on Azure, check if MPU exists with
|
|
||||||
// same key name. If it does, do not allow put or delete because Azure
|
|
||||||
// will delete all blocks with same key name
|
|
||||||
protectAzureBlocks(bucketName, objectKey, dataStoreName, log, cb) {
|
|
||||||
const mpuBucketName = `${constants.mpuBucketPrefix}${bucketName}`;
|
|
||||||
const splitter = constants.splitter;
|
|
||||||
const listingParams = {
|
|
||||||
prefix: `overview${splitter}${objectKey}`,
|
|
||||||
listingType: 'MPU',
|
|
||||||
splitter,
|
|
||||||
maxKeys: 1,
|
|
||||||
};
|
|
||||||
|
|
||||||
return metadata.listMultipartUploads(mpuBucketName, listingParams,
|
|
||||||
log, (err, mpuList) => {
|
|
||||||
if (err && !err.NoSuchBucket) {
|
|
||||||
log.error('Error listing MPUs for Azure delete',
|
|
||||||
{ error: err, dataStoreName });
|
|
||||||
return cb(errors.ServiceUnavailable);
|
|
||||||
}
|
|
||||||
if (mpuList && mpuList.Uploads && mpuList.Uploads.length > 0) {
|
|
||||||
const error = errors.MPUinProgress;
|
|
||||||
log.error('Error: cannot put/delete object to Azure with ' +
|
|
||||||
'same key name as ongoing MPU on Azure',
|
|
||||||
{ error, dataStoreName });
|
|
||||||
return cb(error);
|
|
||||||
}
|
|
||||||
// If listMultipartUploads returns a NoSuchBucket error or the
|
|
||||||
// mpu list is empty, there are no conflicting MPUs, so continue
|
|
||||||
return cb();
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
toObjectGetInfo(objectKey, bucketName) {
|
|
||||||
return {
|
|
||||||
key: this._createAzureKey(bucketName, objectKey, this._bucketMatch),
|
|
||||||
dataStoreName: this._dataStoreName,
|
|
||||||
};
|
|
||||||
}
|
|
||||||
|
|
||||||
put(stream, size, keyContext, reqUids, callback) {
|
|
||||||
const log = createLogger(reqUids);
|
|
||||||
// before blob is put, make sure there is no ongoing MPU with same key
|
|
||||||
this.protectAzureBlocks(keyContext.bucketName,
|
|
||||||
keyContext.objectKey, this._dataStoreName, log, err => {
|
|
||||||
// if error returned, there is ongoing MPU, so do not put
|
|
||||||
if (err) {
|
|
||||||
return callback(err.customizeDescription(
|
|
||||||
`Error putting object to Azure: ${err.message}`));
|
|
||||||
}
|
|
||||||
const azureKey = this._createAzureKey(keyContext.bucketName,
|
|
||||||
keyContext.objectKey, this._bucketMatch);
|
|
||||||
const options = {
|
|
||||||
metadata: translateAzureMetaHeaders(keyContext.metaHeaders,
|
|
||||||
keyContext.tagging),
|
|
||||||
contentSettings: {
|
|
||||||
contentType: keyContext.contentType || undefined,
|
|
||||||
cacheControl: keyContext.cacheControl || undefined,
|
|
||||||
contentDisposition: keyContext.contentDisposition ||
|
|
||||||
undefined,
|
|
||||||
contentEncoding: keyContext.contentEncoding || undefined,
|
|
||||||
},
|
|
||||||
};
|
|
||||||
if (size === 0) {
|
|
||||||
return this._errorWrapper('put', 'createBlockBlobFromText',
|
|
||||||
[this._azureContainerName, azureKey, '', options,
|
|
||||||
err => {
|
|
||||||
if (err) {
|
|
||||||
logHelper(log, 'error', 'err from Azure PUT data ' +
|
|
||||||
'backend', err, this._dataStoreName);
|
|
||||||
return callback(errors.ServiceUnavailable
|
|
||||||
.customizeDescription('Error returned from ' +
|
|
||||||
`Azure: ${err.message}`));
|
|
||||||
}
|
|
||||||
return callback(null, azureKey);
|
|
||||||
}], log, callback);
|
|
||||||
}
|
|
||||||
return this._errorWrapper('put', 'createBlockBlobFromStream',
|
|
||||||
[this._azureContainerName, azureKey, stream, size, options,
|
|
||||||
err => {
|
|
||||||
if (err) {
|
|
||||||
logHelper(log, 'error', 'err from Azure PUT data ' +
|
|
||||||
'backend', err, this._dataStoreName);
|
|
||||||
return callback(errors.ServiceUnavailable
|
|
||||||
.customizeDescription('Error returned from ' +
|
|
||||||
`Azure: ${err.message}`));
|
|
||||||
}
|
|
||||||
return callback(null, azureKey);
|
|
||||||
}], log, callback);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
head(objectGetInfo, reqUids, callback) {
|
|
||||||
const log = createLogger(reqUids);
|
|
||||||
const { key, azureStreamingOptions } = objectGetInfo;
|
|
||||||
return this._errorWrapper('head', 'getBlobProperties',
|
|
||||||
[this._azureContainerName, key, azureStreamingOptions,
|
|
||||||
err => {
|
|
||||||
if (err) {
|
|
||||||
logHelper(log, 'error', 'err from Azure HEAD data backend',
|
|
||||||
err, this._dataStoreName);
|
|
||||||
if (err.code === 'NotFound') {
|
|
||||||
const error = errors.ServiceUnavailable
|
|
||||||
.customizeDescription(
|
|
||||||
'Unexpected error from Azure: "NotFound". Data ' +
|
|
||||||
'on Azure may have been altered outside of ' +
|
|
||||||
'CloudServer.');
|
|
||||||
return callback(error);
|
|
||||||
}
|
|
||||||
return callback(errors.ServiceUnavailable
|
|
||||||
.customizeDescription('Error returned from ' +
|
|
||||||
`Azure: ${err.message}`));
|
|
||||||
}
|
|
||||||
return callback();
|
|
||||||
}], log, callback);
|
|
||||||
}
|
|
||||||
|
|
||||||
get(objectGetInfo, range, reqUids, callback) {
|
|
||||||
const log = createLogger(reqUids);
|
|
||||||
// for backwards compatibility
|
|
||||||
const { key, response, azureStreamingOptions } = objectGetInfo;
|
|
||||||
let streamingOptions;
|
|
||||||
if (azureStreamingOptions) {
|
|
||||||
// option coming from api.get()
|
|
||||||
streamingOptions = azureStreamingOptions;
|
|
||||||
} else if (range) {
|
|
||||||
// option coming from multipleBackend.upload()
|
|
||||||
const rangeStart = range[0] ? range[0].toString() : undefined;
|
|
||||||
const rangeEnd = range[1] ? range[1].toString() : undefined;
|
|
||||||
streamingOptions = { rangeStart, rangeEnd };
|
|
||||||
}
|
|
||||||
this._errorWrapper('get', 'getBlobToStream',
|
|
||||||
[this._azureContainerName, key, response, streamingOptions,
|
|
||||||
err => {
|
|
||||||
if (err) {
|
|
||||||
logHelper(log, 'error', 'err from Azure GET data backend',
|
|
||||||
err, this._dataStoreName);
|
|
||||||
return callback(errors.ServiceUnavailable);
|
|
||||||
}
|
|
||||||
return callback(null, response);
|
|
||||||
}], log, callback);
|
|
||||||
}
|
|
||||||
|
|
||||||
delete(objectGetInfo, reqUids, callback) {
|
|
||||||
const log = createLogger(reqUids);
|
|
||||||
// for backwards compatibility
|
|
||||||
const key = typeof objectGetInfo === 'string' ? objectGetInfo :
|
|
||||||
objectGetInfo.key;
|
|
||||||
return this._errorWrapper('delete', 'deleteBlobIfExists',
|
|
||||||
[this._azureContainerName, key,
|
|
||||||
err => {
|
|
||||||
if (err) {
|
|
||||||
const log = createLogger(reqUids);
|
|
||||||
logHelper(log, 'error', 'error deleting object from ' +
|
|
||||||
'Azure datastore', err, this._dataStoreName);
|
|
||||||
return callback(errors.ServiceUnavailable
|
|
||||||
.customizeDescription('Error returned from ' +
|
|
||||||
`Azure: ${err.message}`));
|
|
||||||
}
|
|
||||||
return callback();
|
|
||||||
}], log, callback);
|
|
||||||
}
|
|
||||||
|
|
||||||
healthcheck(location, callback, flightCheckOnStartUp) {
|
|
||||||
const azureResp = {};
|
|
||||||
const healthCheckAction = flightCheckOnStartUp ?
|
|
||||||
'createContainerIfNotExists' : 'doesContainerExist';
|
|
||||||
this._errorWrapper('checkAzureHealth', healthCheckAction,
|
|
||||||
[this._azureContainerName, err => {
|
|
||||||
/* eslint-disable no-param-reassign */
|
|
||||||
if (err) {
|
|
||||||
azureResp[location] = { error: err.message,
|
|
||||||
external: true };
|
|
||||||
return callback(null, azureResp);
|
|
||||||
}
|
|
||||||
azureResp[location] = {
|
|
||||||
message:
|
|
||||||
'Congrats! You can access the Azure storage account',
|
|
||||||
};
|
|
||||||
return callback(null, azureResp);
|
|
||||||
}], null, callback);
|
|
||||||
}
|
|
||||||
|
|
||||||
uploadPart(request, streamingV4Params, partStream, size, key, uploadId,
|
|
||||||
partNumber, bucket, log, callback) {
|
|
||||||
const azureKey = this._createAzureKey(bucket, key, this._bucketMatch);
|
|
||||||
const params = { bucketName: this._azureContainerName,
|
|
||||||
partNumber, size, objectKey: azureKey, uploadId };
|
|
||||||
const stream = request || partStream;
|
|
||||||
|
|
||||||
if (request && request.headers && request.headers['content-md5']) {
|
|
||||||
params.contentMD5 = request.headers['content-md5'];
|
|
||||||
}
|
|
||||||
const dataRetrievalInfo = {
|
|
||||||
key: azureKey,
|
|
||||||
partNumber,
|
|
||||||
dataStoreName: this._dataStoreName,
|
|
||||||
dataStoreType: 'azure',
|
|
||||||
numberSubParts: azureMpuUtils.getSubPartInfo(size)
|
|
||||||
.expectedNumberSubParts,
|
|
||||||
};
|
|
||||||
|
|
||||||
if (size === 0) {
|
|
||||||
log.debug('0-byte part does not store data',
|
|
||||||
{ method: 'uploadPart' });
|
|
||||||
dataRetrievalInfo.dataStoreETag = azureMpuUtils.zeroByteETag;
|
|
||||||
dataRetrievalInfo.numberSubParts = 0;
|
|
||||||
return callback(null, dataRetrievalInfo);
|
|
||||||
}
|
|
||||||
if (size <= azureMpuUtils.maxSubPartSize) {
|
|
||||||
const errorWrapperFn = this._errorWrapper.bind(this);
|
|
||||||
return azureMpuUtils.putSinglePart(errorWrapperFn,
|
|
||||||
stream, params, this._dataStoreName, log, (err, dataStoreETag) => {
|
|
||||||
if (err) {
|
|
||||||
return callback(err);
|
|
||||||
}
|
|
||||||
dataRetrievalInfo.dataStoreETag = dataStoreETag;
|
|
||||||
return callback(null, dataRetrievalInfo);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
const errorWrapperFn = this._errorWrapper.bind(this);
|
|
||||||
return azureMpuUtils.putSubParts(errorWrapperFn, stream,
|
|
||||||
params, this._dataStoreName, log, (err, dataStoreETag) => {
|
|
||||||
if (err) {
|
|
||||||
return callback(err);
|
|
||||||
}
|
|
||||||
dataRetrievalInfo.dataStoreETag = dataStoreETag;
|
|
||||||
return callback(null, dataRetrievalInfo);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
completeMPU(jsonList, mdInfo, key, uploadId, bucket, metaHeaders,
|
|
||||||
contentSettings, log, callback) {
|
|
||||||
const azureKey = this._createAzureKey(bucket, key, this._bucketMatch);
|
|
||||||
const commitList = {
|
|
||||||
UncommittedBlocks: jsonList.uncommittedBlocks || [],
|
|
||||||
};
|
|
||||||
let filteredPartsObj;
|
|
||||||
if (!jsonList.uncommittedBlocks) {
|
|
||||||
const { storedParts, mpuOverviewKey, splitter } = mdInfo;
|
|
||||||
filteredPartsObj = validateAndFilterMpuParts(storedParts, jsonList,
|
|
||||||
mpuOverviewKey, splitter, log);
|
|
||||||
|
|
||||||
filteredPartsObj.partList.forEach(part => {
|
|
||||||
// part.locations is always array of 1, which contains data info
|
|
||||||
const subPartIds =
|
|
||||||
azureMpuUtils.getSubPartIds(part.locations[0], uploadId);
|
|
||||||
commitList.UncommittedBlocks.push(...subPartIds);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
const options = {
|
|
||||||
contentSettings,
|
|
||||||
metadata: metaHeaders ? translateAzureMetaHeaders(metaHeaders) :
|
|
||||||
undefined,
|
|
||||||
};
|
|
||||||
this._errorWrapper('completeMPU', 'commitBlocks',
|
|
||||||
[this._azureContainerName, azureKey, commitList, options,
|
|
||||||
err => {
|
|
||||||
if (err) {
|
|
||||||
logHelper(log, 'error', 'err completing MPU on Azure ' +
|
|
||||||
'datastore', err, this._dataStoreName);
|
|
||||||
return callback(errors.ServiceUnavailable
|
|
||||||
.customizeDescription('Error returned from ' +
|
|
||||||
`Azure: ${err.message}`));
|
|
||||||
}
|
|
||||||
const completeObjData = {
|
|
||||||
key: azureKey,
|
|
||||||
filteredPartsObj,
|
|
||||||
};
|
|
||||||
return callback(null, completeObjData);
|
|
||||||
}], log, callback);
|
|
||||||
}
|
|
||||||
|
|
||||||
objectPutTagging(key, bucket, objectMD, log, callback) {
|
|
||||||
const azureKey = this._createAzureKey(bucket, key, this._bucketMatch);
|
|
||||||
const azureMD = this._getMetaHeaders(objectMD);
|
|
||||||
azureMD.tags = JSON.stringify(objectMD.tags);
|
|
||||||
this._errorWrapper('objectPutTagging', 'setBlobMetadata',
|
|
||||||
[this._azureContainerName, azureKey, azureMD,
|
|
||||||
err => {
|
|
||||||
if (err) {
|
|
||||||
logHelper(log, 'error', 'err putting object tags to ' +
|
|
||||||
'Azure backend', err, this._dataStoreName);
|
|
||||||
return callback(errors.ServiceUnavailable);
|
|
||||||
}
|
|
||||||
return callback();
|
|
||||||
}], log, callback);
|
|
||||||
}
|
|
||||||
|
|
||||||
objectDeleteTagging(key, bucket, objectMD, log, callback) {
|
|
||||||
const azureKey = this._createAzureKey(bucket, key, this._bucketMatch);
|
|
||||||
const azureMD = this._getMetaHeaders(objectMD);
|
|
||||||
this._errorWrapper('objectDeleteTagging', 'setBlobMetadata',
|
|
||||||
[this._azureContainerName, azureKey, azureMD,
|
|
||||||
err => {
|
|
||||||
if (err) {
|
|
||||||
logHelper(log, 'error', 'err putting object tags to ' +
|
|
||||||
'Azure backend', err, this._dataStoreName);
|
|
||||||
return callback(errors.ServiceUnavailable);
|
|
||||||
}
|
|
||||||
return callback();
|
|
||||||
}], log, callback);
|
|
||||||
}
|
|
||||||
|
|
||||||
copyObject(request, destLocationConstraintName, sourceKey,
|
|
||||||
sourceLocationConstraintName, storeMetadataParams, log, callback) {
|
|
||||||
const destContainerName = request.bucketName;
|
|
||||||
const destObjectKey = request.objectKey;
|
|
||||||
|
|
||||||
const destAzureKey = this._createAzureKey(destContainerName,
|
|
||||||
destObjectKey, this._bucketMatch);
|
|
||||||
|
|
||||||
const sourceContainerName =
|
|
||||||
config.locationConstraints[sourceLocationConstraintName]
|
|
||||||
.details.azureContainerName;
|
|
||||||
|
|
||||||
let options;
|
|
||||||
if (storeMetadataParams.metaHeaders) {
|
|
||||||
options = { metadata:
|
|
||||||
translateAzureMetaHeaders(storeMetadataParams.metaHeaders) };
|
|
||||||
}
|
|
||||||
|
|
||||||
this._errorWrapper('copyObject', 'startCopyBlob',
|
|
||||||
[`${this._azureStorageEndpoint}` +
|
|
||||||
`${sourceContainerName}/${sourceKey}`,
|
|
||||||
this._azureContainerName, destAzureKey, options,
|
|
||||||
(err, res) => {
|
|
||||||
if (err) {
|
|
||||||
if (err.code === 'CannotVerifyCopySource') {
|
|
||||||
logHelper(log, 'error', 'Unable to access ' +
|
|
||||||
`${sourceContainerName} Azure Container`, err,
|
|
||||||
this._dataStoreName);
|
|
||||||
return callback(errors.AccessDenied
|
|
||||||
.customizeDescription('Error: Unable to access ' +
|
|
||||||
`${sourceContainerName} Azure Container`)
|
|
||||||
);
|
|
||||||
}
|
|
||||||
logHelper(log, 'error', 'error from data backend on ' +
|
|
||||||
'copyObject', err, this._dataStoreName);
|
|
||||||
return callback(errors.ServiceUnavailable
|
|
||||||
.customizeDescription('Error returned from ' +
|
|
||||||
`AWS: ${err.message}`)
|
|
||||||
);
|
|
||||||
}
|
|
||||||
if (res.copy.status === 'pending') {
|
|
||||||
logHelper(log, 'error', 'Azure copy status is pending',
|
|
||||||
err, this._dataStoreName);
|
|
||||||
const copyId = res.copy.id;
|
|
||||||
this._client.abortCopyBlob(this._azureContainerName,
|
|
||||||
destAzureKey, copyId, err => {
|
|
||||||
if (err) {
|
|
||||||
logHelper(log, 'error', 'error from data backend ' +
|
|
||||||
'on abortCopyBlob', err, this._dataStoreName);
|
|
||||||
return callback(errors.ServiceUnavailable
|
|
||||||
.customizeDescription('Error returned from ' +
|
|
||||||
`AWS on abortCopyBlob: ${err.message}`)
|
|
||||||
);
|
|
||||||
}
|
|
||||||
return callback(errors.InvalidObjectState
|
|
||||||
.customizeDescription('Error: Azure copy status was ' +
|
|
||||||
'pending. It has been aborted successfully')
|
|
||||||
);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
return callback(null, destAzureKey);
|
|
||||||
}], log, callback);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
module.exports = AzureClient;
|
|
|
@ -1,33 +0,0 @@
|
||||||
const { errors } = require('arsenal');
|
|
||||||
const MpuHelper = require('./mpuHelper');
|
|
||||||
const { createMpuKey, logger } = require('../GcpUtils');
|
|
||||||
const { logHelper } = require('../../utils');
|
|
||||||
|
|
||||||
/**
|
|
||||||
* abortMPU - remove all objects of a GCP Multipart Upload
|
|
||||||
* @param {object} params - abortMPU params
|
|
||||||
* @param {string} params.Bucket - bucket name
|
|
||||||
* @param {string} params.MPU - mpu bucket name
|
|
||||||
* @param {string} params.Key - object key
|
|
||||||
* @param {number} params.UploadId - MPU upload id
|
|
||||||
* @param {function} callback - callback function to call
|
|
||||||
* @return {undefined}
|
|
||||||
*/
|
|
||||||
function abortMPU(params, callback) {
|
|
||||||
if (!params || !params.Key || !params.UploadId ||
|
|
||||||
!params.Bucket || !params.MPU) {
|
|
||||||
const error = errors.InvalidRequest
|
|
||||||
.customizeDescription('Missing required parameter');
|
|
||||||
logHelper(logger, 'error', 'error in abortMultipartUpload', error);
|
|
||||||
return callback(error);
|
|
||||||
}
|
|
||||||
const mpuHelper = new MpuHelper(this);
|
|
||||||
const delParams = {
|
|
||||||
Bucket: params.Bucket,
|
|
||||||
MPU: params.MPU,
|
|
||||||
Prefix: createMpuKey(params.Key, params.UploadId),
|
|
||||||
};
|
|
||||||
return mpuHelper.removeParts(delParams, callback);
|
|
||||||
}
|
|
||||||
|
|
||||||
module.exports = abortMPU;
|
|
|
@ -1,69 +0,0 @@
|
||||||
const async = require('async');
|
|
||||||
const { errors } = require('arsenal');
|
|
||||||
const MpuHelper = require('./mpuHelper');
|
|
||||||
const { createMpuKey, logger } = require('../GcpUtils');
|
|
||||||
const { logHelper } = require('../../utils');
|
|
||||||
|
|
||||||
/**
|
|
||||||
* completeMPU - merges a list of parts into a single object
|
|
||||||
* @param {object} params - completeMPU params
|
|
||||||
* @param {string} params.Bucket - bucket name
|
|
||||||
* @param {string} params.MPU - mpu bucket name
|
|
||||||
* @param {string} params.Key - object key
|
|
||||||
* @param {number} params.UploadId - MPU upload id
|
|
||||||
* @param {Object} params.MultipartUpload - MPU upload object
|
|
||||||
* @param {Object[]} param.MultipartUpload.Parts - a list of parts to merge
|
|
||||||
* @param {function} callback - callback function to call with MPU result
|
|
||||||
* @return {undefined}
|
|
||||||
*/
|
|
||||||
function completeMPU(params, callback) {
|
|
||||||
if (!params || !params.MultipartUpload ||
|
|
||||||
!params.MultipartUpload.Parts || !params.UploadId ||
|
|
||||||
!params.Bucket || !params.Key) {
|
|
||||||
const error = errors.InvalidRequest
|
|
||||||
.customizeDescription('Missing required parameter');
|
|
||||||
logHelper(logger, 'error', 'error in completeMultipartUpload', error);
|
|
||||||
return callback(error);
|
|
||||||
}
|
|
||||||
const partList = params.MultipartUpload.Parts;
|
|
||||||
// verify that the part list is in order
|
|
||||||
if (params.MultipartUpload.Parts.length === 0) {
|
|
||||||
const error = errors.InvalidRequest
|
|
||||||
.customizeDescription('You must specify at least one part');
|
|
||||||
logHelper(logger, 'error', 'error in completeMultipartUpload', error);
|
|
||||||
return callback(error);
|
|
||||||
}
|
|
||||||
for (let ind = 1; ind < partList.length; ++ind) {
|
|
||||||
if (partList[ind - 1].PartNumber >= partList[ind].PartNumber) {
|
|
||||||
logHelper(logger, 'error', 'error in completeMultipartUpload',
|
|
||||||
errors.InvalidPartOrder);
|
|
||||||
return callback(errors.InvalidPartOrder);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
const mpuHelper = new MpuHelper(this); // this === GcpClient
|
|
||||||
return async.waterfall([
|
|
||||||
next => {
|
|
||||||
// first compose: in mpu bucket
|
|
||||||
// max 10,000 => 313 parts
|
|
||||||
// max component count per object 32
|
|
||||||
logger.trace('completeMultipartUpload: compose',
|
|
||||||
{ partCount: partList.length });
|
|
||||||
mpuHelper.splitMerge(params, partList, 'compose', next);
|
|
||||||
},
|
|
||||||
(numParts, next) => mpuHelper.composeFinal(numParts, params, next),
|
|
||||||
(result, next) => mpuHelper.generateMpuResult(result, partList, next),
|
|
||||||
(result, aggregateETag, next) =>
|
|
||||||
mpuHelper.copyToMain(result, aggregateETag, params, next),
|
|
||||||
(mpuResult, next) => {
|
|
||||||
const delParams = {
|
|
||||||
Bucket: params.Bucket,
|
|
||||||
MPU: params.MPU,
|
|
||||||
Prefix: createMpuKey(params.Key, params.UploadId),
|
|
||||||
};
|
|
||||||
return mpuHelper.removeParts(delParams,
|
|
||||||
err => next(err, mpuResult));
|
|
||||||
},
|
|
||||||
], callback);
|
|
||||||
}
|
|
||||||
|
|
||||||
module.exports = completeMPU;
|
|
|
@ -1,51 +0,0 @@
|
||||||
const uuid = require('uuid/v4');
|
|
||||||
const { errors } = require('arsenal');
|
|
||||||
const { createMpuKey, logger, getPutTagsMetadata } = require('../GcpUtils');
|
|
||||||
const { logHelper } = require('../../utils');
|
|
||||||
|
|
||||||
/**
|
|
||||||
* createMPU - creates a MPU upload on GCP (sets a 0-byte object placeholder
|
|
||||||
* with for the final composed object)
|
|
||||||
* @param {object} params - createMPU param
|
|
||||||
* @param {string} params.Bucket - bucket name
|
|
||||||
* @param {string} params.Key - object key
|
|
||||||
* @param {string} params.Metadata - object Metadata
|
|
||||||
* @param {string} params.ContentType - Content-Type header
|
|
||||||
* @param {string} params.CacheControl - Cache-Control header
|
|
||||||
* @param {string} params.ContentDisposition - Content-Disposition header
|
|
||||||
* @param {string} params.ContentEncoding - Content-Encoding header
|
|
||||||
* @param {function} callback - callback function to call with the generated
|
|
||||||
* upload-id for MPU operations
|
|
||||||
* @return {undefined}
|
|
||||||
*/
|
|
||||||
function createMPU(params, callback) {
|
|
||||||
// As google cloud does not have a create MPU function,
|
|
||||||
// create an empty 'init' object that will temporarily store the
|
|
||||||
// object metadata and return an upload ID to mimic an AWS MPU
|
|
||||||
if (!params || !params.Bucket || !params.Key) {
|
|
||||||
const error = errors.InvalidRequest
|
|
||||||
.customizeDescription('Missing required parameter');
|
|
||||||
logHelper(logger, 'error', 'error in createMultipartUpload', error);
|
|
||||||
return callback(error);
|
|
||||||
}
|
|
||||||
const uploadId = uuid().replace(/-/g, '');
|
|
||||||
const mpuParams = {
|
|
||||||
Bucket: params.Bucket,
|
|
||||||
Key: createMpuKey(params.Key, uploadId, 'init'),
|
|
||||||
Metadata: params.Metadata,
|
|
||||||
ContentType: params.ContentType,
|
|
||||||
CacheControl: params.CacheControl,
|
|
||||||
ContentDisposition: params.ContentDisposition,
|
|
||||||
ContentEncoding: params.ContentEncoding,
|
|
||||||
};
|
|
||||||
mpuParams.Metadata = getPutTagsMetadata(mpuParams.Metadata, params.Tagging);
|
|
||||||
return this.putObject(mpuParams, err => {
|
|
||||||
if (err) {
|
|
||||||
logHelper(logger, 'error', 'error in createMPU - putObject', err);
|
|
||||||
return callback(err);
|
|
||||||
}
|
|
||||||
return callback(null, { UploadId: uploadId });
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
module.exports = createMPU;
|
|
|
@ -1,24 +0,0 @@
|
||||||
const async = require('async');
|
|
||||||
|
|
||||||
const { stripTags } = require('../GcpUtils');
|
|
||||||
|
|
||||||
function deleteObjectTagging(params, callback) {
|
|
||||||
return async.waterfall([
|
|
||||||
next => this.headObject({
|
|
||||||
Bucket: params.Bucket,
|
|
||||||
Key: params.Key,
|
|
||||||
VersionId: params.VersionId,
|
|
||||||
}, next),
|
|
||||||
(resObj, next) => {
|
|
||||||
const completeMD = stripTags(resObj.Metadata);
|
|
||||||
this.copyObject({
|
|
||||||
Bucket: params.Bucket,
|
|
||||||
Key: params.Key,
|
|
||||||
CopySource: `${params.Bucket}/${params.Key}`,
|
|
||||||
Metadata: completeMD,
|
|
||||||
MetadataDirective: 'REPLACE',
|
|
||||||
}, next);
|
|
||||||
},
|
|
||||||
], callback);
|
|
||||||
}
|
|
||||||
module.exports = deleteObjectTagging;
|
|
|
@ -1,19 +0,0 @@
|
||||||
const { retrieveTags } = require('../GcpUtils');
|
|
||||||
|
|
||||||
function getObjectTagging(params, callback) {
|
|
||||||
const headParams = {
|
|
||||||
Bucket: params.Bucket,
|
|
||||||
Key: params.Key,
|
|
||||||
VersionId: params.VersionId,
|
|
||||||
};
|
|
||||||
this.headObject(headParams, (err, res) => {
|
|
||||||
const TagSet = retrieveTags(res.Metadata);
|
|
||||||
const retObj = {
|
|
||||||
VersionId: res.VersionId,
|
|
||||||
TagSet,
|
|
||||||
};
|
|
||||||
return callback(null, retObj);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
module.exports = getObjectTagging;
|
|
|
@ -1,14 +0,0 @@
|
||||||
module.exports = {
|
|
||||||
// mpu functions
|
|
||||||
abortMultipartUpload: require('./abortMPU'),
|
|
||||||
completeMultipartUpload: require('./completeMPU'),
|
|
||||||
createMultipartUpload: require('./createMPU'),
|
|
||||||
listParts: require('./listParts'),
|
|
||||||
uploadPart: require('./uploadPart'),
|
|
||||||
uploadPartCopy: require('./uploadPartCopy'),
|
|
||||||
// object tagging
|
|
||||||
putObject: require('./putObject'),
|
|
||||||
putObjectTagging: require('./putTagging'),
|
|
||||||
getObjectTagging: require('./getTagging'),
|
|
||||||
deleteObjectTagging: require('./deleteTagging'),
|
|
||||||
};
|
|
|
@ -1,42 +0,0 @@
|
||||||
const { errors } = require('arsenal');
|
|
||||||
const { createMpuKey, logger } = require('../GcpUtils');
|
|
||||||
const { logHelper } = require('../../utils');
|
|
||||||
|
|
||||||
/**
|
|
||||||
* listParts - list uploaded MPU parts
|
|
||||||
* @param {object} params - listParts param
|
|
||||||
* @param {string} params.Bucket - bucket name
|
|
||||||
* @param {string} params.Key - object key
|
|
||||||
* @param {string} params.UploadId - MPU upload id
|
|
||||||
* @param {function} callback - callback function to call with the list of parts
|
|
||||||
* @return {undefined}
|
|
||||||
*/
|
|
||||||
function listParts(params, callback) {
|
|
||||||
if (!params || !params.UploadId || !params.Bucket || !params.Key) {
|
|
||||||
const error = errors.InvalidRequest
|
|
||||||
.customizeDescription('Missing required parameter');
|
|
||||||
logHelper(logger, 'error', 'error in listParts', error);
|
|
||||||
return callback(error);
|
|
||||||
}
|
|
||||||
if (params.PartNumberMarker && params.PartNumberMarker < 0) {
|
|
||||||
return callback(errors.InvalidArgument
|
|
||||||
.customizeDescription('The request specified an invalid marker'));
|
|
||||||
}
|
|
||||||
const mpuParams = {
|
|
||||||
Bucket: params.Bucket,
|
|
||||||
Prefix: createMpuKey(params.Key, params.UploadId, 'parts'),
|
|
||||||
Marker: createMpuKey(params.Key, params.UploadId,
|
|
||||||
params.PartNumberMarker, 'parts'),
|
|
||||||
MaxKeys: params.MaxParts,
|
|
||||||
};
|
|
||||||
return this.listObjects(mpuParams, (err, res) => {
|
|
||||||
if (err) {
|
|
||||||
logHelper(logger, 'error',
|
|
||||||
'error in listParts - listObjects', err);
|
|
||||||
return callback(err);
|
|
||||||
}
|
|
||||||
return callback(null, res);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
module.exports = listParts;
|
|
|
@ -1,316 +0,0 @@
|
||||||
const async = require('async');
|
|
||||||
const Backoff = require('backo');
|
|
||||||
const { eachSlice, createMpuKey, createMpuList, logger } =
|
|
||||||
require('../GcpUtils');
|
|
||||||
const { logHelper } = require('../../utils');
|
|
||||||
const { createAggregateETag } =
|
|
||||||
require('../../../../api/apiUtils/object/processMpuParts');
|
|
||||||
|
|
||||||
const BACKOFF_PARAMS = { min: 1000, max: 300000, jitter: 0.1, factor: 1.5 };
|
|
||||||
|
|
||||||
class MpuHelper {
|
|
||||||
constructor(service, options = {}) {
|
|
||||||
this.service = service;
|
|
||||||
this.backoffParams = {
|
|
||||||
min: options.min || BACKOFF_PARAMS.min,
|
|
||||||
max: options.max || BACKOFF_PARAMS.max,
|
|
||||||
jitter: options.jitter || BACKOFF_PARAMS.jitter,
|
|
||||||
factor: options.factor || BACKOFF_PARAMS.factor,
|
|
||||||
};
|
|
||||||
}
|
|
||||||
|
|
||||||
_retry(fnName, params, callback) {
|
|
||||||
const backoff = new Backoff(this.backoffParams);
|
|
||||||
const handleFunc = (fnName, params, retry, callback) => {
|
|
||||||
const timeout = backoff.duration();
|
|
||||||
return setTimeout((params, cb) =>
|
|
||||||
this.service[fnName](params, cb), timeout, params,
|
|
||||||
(err, res) => {
|
|
||||||
if (err) {
|
|
||||||
if (err.statusCode === 429 || err.code === 429) {
|
|
||||||
if (fnName === 'composeObject') {
|
|
||||||
logger.trace('composeObject: slow down request',
|
|
||||||
{ retryCount: retry, timeout });
|
|
||||||
} else if (fnName === 'copyObject') {
|
|
||||||
logger.trace('copyObject: slow down request',
|
|
||||||
{ retryCount: retry, timeout });
|
|
||||||
}
|
|
||||||
return handleFunc(
|
|
||||||
fnName, params, retry + 1, callback);
|
|
||||||
}
|
|
||||||
logHelper(logger, 'error', `${fnName} failed`, err);
|
|
||||||
return callback(err);
|
|
||||||
}
|
|
||||||
backoff.reset();
|
|
||||||
return callback(null, res);
|
|
||||||
});
|
|
||||||
};
|
|
||||||
handleFunc(fnName, params, 0, callback);
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* retryCompose - exponential backoff retry implementation for the compose
|
|
||||||
* operation
|
|
||||||
* @param {object} params - compose object params
|
|
||||||
* @param {function} callback - callback function to call with the result
|
|
||||||
* of the compose operation
|
|
||||||
* @return {undefined}
|
|
||||||
*/
|
|
||||||
retryCompose(params, callback) {
|
|
||||||
this._retry('composeObject', params, callback);
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
|
||||||
* retryCopy - exponential backoff retry implementation for the copy
|
|
||||||
* operation
|
|
||||||
* @param {object} params - copy object params
|
|
||||||
* @param {function} callback - callback function to call with the result
|
|
||||||
* of the copy operation
|
|
||||||
* @return {undefined}
|
|
||||||
*/
|
|
||||||
retryCopy(params, callback) {
|
|
||||||
this._retry('copyObject', params, callback);
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* splitMerge - breaks down the MPU list of parts to be compose on GCP;
|
|
||||||
* splits partList into chunks of 32 objects, the limit of each compose
|
|
||||||
* operation.
|
|
||||||
* @param {object} params - complete MPU params
|
|
||||||
* @param {string} params.Bucket - bucket name
|
|
||||||
* @param {string} params.MPU - mpu bucket name
|
|
||||||
* @param {string} params.Key - object key
|
|
||||||
* @param {string} params.UploadId - MPU upload id
|
|
||||||
* @param {object[]} partList - list of parts for complete multipart upload
|
|
||||||
* @param {string} level - the phase name of the MPU process
|
|
||||||
* @param {function} callback - the callback function to call
|
|
||||||
* @return {undefined}
|
|
||||||
*/
|
|
||||||
splitMerge(params, partList, level, callback) {
|
|
||||||
// create composition of slices from the partList array
|
|
||||||
return async.mapLimit(eachSlice.call(partList, 32),
|
|
||||||
this.service._maxConcurrent,
|
|
||||||
(infoParts, cb) => {
|
|
||||||
const mpuPartList = infoParts.Parts.map(item =>
|
|
||||||
({ PartName: item.PartName }));
|
|
||||||
const partNumber = infoParts.PartNumber;
|
|
||||||
const tmpKey =
|
|
||||||
createMpuKey(params.Key, params.UploadId, partNumber, level);
|
|
||||||
const mergedObject = { PartName: tmpKey };
|
|
||||||
if (mpuPartList.length < 2) {
|
|
||||||
logger.trace(
|
|
||||||
'splitMerge: parts are fewer than 2, copy instead');
|
|
||||||
// else just perform a copy
|
|
||||||
const copyParams = {
|
|
||||||
Bucket: params.MPU,
|
|
||||||
Key: tmpKey,
|
|
||||||
CopySource: `${params.MPU}/${mpuPartList[0].PartName}`,
|
|
||||||
};
|
|
||||||
return this.service.copyObject(copyParams, (err, res) => {
|
|
||||||
if (err) {
|
|
||||||
logHelper(logger, 'error',
|
|
||||||
'error in splitMerge - copyObject', err);
|
|
||||||
return cb(err);
|
|
||||||
}
|
|
||||||
mergedObject.VersionId = res.VersionId;
|
|
||||||
mergedObject.ETag = res.ETag;
|
|
||||||
return cb(null, mergedObject);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
const composeParams = {
|
|
||||||
Bucket: params.MPU,
|
|
||||||
Key: tmpKey,
|
|
||||||
MultipartUpload: { Parts: mpuPartList },
|
|
||||||
};
|
|
||||||
return this.retryCompose(composeParams, (err, res) => {
|
|
||||||
if (err) {
|
|
||||||
return cb(err);
|
|
||||||
}
|
|
||||||
mergedObject.VersionId = res.VersionId;
|
|
||||||
mergedObject.ETag = res.ETag;
|
|
||||||
return cb(null, mergedObject);
|
|
||||||
});
|
|
||||||
}, (err, res) => {
|
|
||||||
if (err) {
|
|
||||||
return callback(err);
|
|
||||||
}
|
|
||||||
return callback(null, res.length);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* removeParts - remove all objects created to perform a multipart upload
|
|
||||||
* @param {object} params - remove parts params
|
|
||||||
* @param {string} params.Bucket - bucket name
|
|
||||||
* @param {string} params.MPU - mpu bucket name
|
|
||||||
* @param {string} params.Key - object key
|
|
||||||
* @param {string} params.UploadId - MPU upload id
|
|
||||||
* @param {function} callback - callback function to call
|
|
||||||
* @return {undefined}
|
|
||||||
*/
|
|
||||||
removeParts(params, callback) {
|
|
||||||
const _getObjectVersions = callback => {
|
|
||||||
logger.trace('remove all parts from mpu bucket');
|
|
||||||
let partsList = [];
|
|
||||||
let isTruncated = true;
|
|
||||||
let nextMarker;
|
|
||||||
return async.whilst(() => isTruncated, next => {
|
|
||||||
const listParams = {
|
|
||||||
Bucket: params.MPU,
|
|
||||||
Prefix: params.Prefix,
|
|
||||||
Marker: nextMarker,
|
|
||||||
};
|
|
||||||
return this.service.listVersions(listParams, (err, res) => {
|
|
||||||
if (err) {
|
|
||||||
logHelper(logger, 'error', 'error in ' +
|
|
||||||
'removeParts - listVersions', err);
|
|
||||||
return next(err);
|
|
||||||
}
|
|
||||||
nextMarker = res.NextMarker;
|
|
||||||
isTruncated = res.IsTruncated;
|
|
||||||
partsList = partsList.concat(res.Versions);
|
|
||||||
return next();
|
|
||||||
});
|
|
||||||
}, err => callback(err, partsList));
|
|
||||||
};
|
|
||||||
|
|
||||||
const _deleteObjects = (partsList, callback) => {
|
|
||||||
logger.trace('successfully listed mpu parts', {
|
|
||||||
objectCount: partsList.length,
|
|
||||||
});
|
|
||||||
return async.eachLimit(partsList, 10, (obj, next) => {
|
|
||||||
const delParams = {
|
|
||||||
Bucket: params.MPU,
|
|
||||||
Key: obj.Key,
|
|
||||||
VersionId: obj.VersionId,
|
|
||||||
};
|
|
||||||
this.service.deleteObject(delParams, err => {
|
|
||||||
if (err) {
|
|
||||||
logHelper(logger, 'error',
|
|
||||||
'error deleting object', err);
|
|
||||||
return next(err);
|
|
||||||
}
|
|
||||||
return next();
|
|
||||||
});
|
|
||||||
}, err => callback(err));
|
|
||||||
};
|
|
||||||
|
|
||||||
return async.waterfall([
|
|
||||||
_getObjectVersions,
|
|
||||||
_deleteObjects,
|
|
||||||
], err => callback(err));
|
|
||||||
}
|
|
||||||
|
|
||||||
composeFinal(numParts, params, callback) {
|
|
||||||
// final compose:
|
|
||||||
// number of parts to compose <= 10
|
|
||||||
// perform final compose in mpu bucket
|
|
||||||
logger.trace('completeMultipartUpload: final compose');
|
|
||||||
const parts = createMpuList(params, 'compose', numParts);
|
|
||||||
const partList = parts.map(item => (
|
|
||||||
{ PartName: item.PartName }));
|
|
||||||
if (partList.length < 2) {
|
|
||||||
logger.trace(
|
|
||||||
'fewer than 2 parts, skip to copy phase');
|
|
||||||
return callback(null, partList[0].PartName);
|
|
||||||
}
|
|
||||||
const composeParams = {
|
|
||||||
Bucket: params.MPU,
|
|
||||||
Key: createMpuKey(params.Key, params.UploadId, 'final'),
|
|
||||||
MultipartUpload: { Parts: partList },
|
|
||||||
};
|
|
||||||
return this.retryCompose(composeParams, err => {
|
|
||||||
if (err) {
|
|
||||||
return callback(err);
|
|
||||||
}
|
|
||||||
return callback(null, null);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Create MPU Aggregate ETag
|
|
||||||
*/
|
|
||||||
generateMpuResult(res, partList, callback) {
|
|
||||||
const concatETag = partList.reduce((prev, curr) =>
|
|
||||||
prev + curr.ETag.substring(1, curr.ETag.length - 1), '');
|
|
||||||
const aggregateETag = createAggregateETag(concatETag, partList);
|
|
||||||
return callback(null, res, aggregateETag);
|
|
||||||
}
|
|
||||||
|
|
||||||
copyToMain(res, aggregateETag, params, callback) {
|
|
||||||
// move object from mpu bucket into the main bucket
|
|
||||||
// retrieve initial metadata then compose the object
|
|
||||||
const copySource = res ||
|
|
||||||
createMpuKey(params.Key, params.UploadId, 'final');
|
|
||||||
return async.waterfall([
|
|
||||||
next => {
|
|
||||||
// retrieve metadata from init object in mpu bucket
|
|
||||||
const headParams = {
|
|
||||||
Bucket: params.MPU,
|
|
||||||
Key: createMpuKey(params.Key, params.UploadId,
|
|
||||||
'init'),
|
|
||||||
};
|
|
||||||
logger.trace('retrieving object metadata');
|
|
||||||
return this.service.headObject(headParams, (err, res) => {
|
|
||||||
if (err) {
|
|
||||||
logHelper(logger, 'error',
|
|
||||||
'error in createMultipartUpload - headObject',
|
|
||||||
err);
|
|
||||||
return next(err);
|
|
||||||
}
|
|
||||||
return next(null, res);
|
|
||||||
});
|
|
||||||
},
|
|
||||||
(res, next) => {
|
|
||||||
const metadata = res.Metadata;
|
|
||||||
// copy the final object into the main bucket
|
|
||||||
const copyMetadata = Object.assign({}, metadata);
|
|
||||||
copyMetadata['scal-etag'] = aggregateETag;
|
|
||||||
const copyParams = {
|
|
||||||
Bucket: params.Bucket,
|
|
||||||
Key: params.Key,
|
|
||||||
Metadata: copyMetadata,
|
|
||||||
MetadataDirective: 'REPLACE',
|
|
||||||
CopySource: `${params.MPU}/${copySource}`,
|
|
||||||
ContentType: res.ContentType,
|
|
||||||
CacheControl: res.CacheControl,
|
|
||||||
ContentEncoding: res.ContentEncoding,
|
|
||||||
ContentDisposition: res.ContentDisposition,
|
|
||||||
ContentLanguage: res.ContentLanguage,
|
|
||||||
};
|
|
||||||
logger.trace('copyParams', { copyParams });
|
|
||||||
this.retryCopy(copyParams, (err, res) => {
|
|
||||||
if (err) {
|
|
||||||
logHelper(logger, 'error', 'error in ' +
|
|
||||||
'createMultipartUpload - final copyObject',
|
|
||||||
err);
|
|
||||||
return next(err);
|
|
||||||
}
|
|
||||||
const mpuResult = {
|
|
||||||
Bucket: params.Bucket,
|
|
||||||
Key: params.Key,
|
|
||||||
VersionId: res.VersionId,
|
|
||||||
ETag: `"${aggregateETag}"`,
|
|
||||||
};
|
|
||||||
return this.service.headObject({
|
|
||||||
Bucket: params.Bucket,
|
|
||||||
Key: params.Key,
|
|
||||||
VersionId: res.VersionId,
|
|
||||||
}, (err, res) => {
|
|
||||||
if (err) {
|
|
||||||
logHelper(logger, 'error', 'error in ' +
|
|
||||||
'createMultipartUpload - final head object',
|
|
||||||
err);
|
|
||||||
return next(err);
|
|
||||||
}
|
|
||||||
mpuResult.ContentLength = res.ContentLength;
|
|
||||||
return next(null, mpuResult);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
},
|
|
||||||
], callback);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
module.exports = MpuHelper;
|
|
|
@ -1,11 +0,0 @@
|
||||||
const { getPutTagsMetadata } = require('../GcpUtils');
|
|
||||||
|
|
||||||
function putObject(params, callback) {
|
|
||||||
const putParams = Object.assign({}, params);
|
|
||||||
putParams.Metadata = getPutTagsMetadata(putParams.Metadata, params.Tagging);
|
|
||||||
delete putParams.Tagging;
|
|
||||||
// error handling will be by the actual putObject request
|
|
||||||
return this.putObjectReq(putParams, callback);
|
|
||||||
}
|
|
||||||
|
|
||||||
module.exports = putObject;
|
|
|
@ -1,33 +0,0 @@
|
||||||
const async = require('async');
|
|
||||||
const { errors } = require('arsenal');
|
|
||||||
|
|
||||||
const { processTagSet } = require('../GcpUtils');
|
|
||||||
|
|
||||||
function putObjectTagging(params, callback) {
|
|
||||||
if (!params.Tagging || !params.Tagging.TagSet) {
|
|
||||||
return callback(errors.MissingParameter);
|
|
||||||
}
|
|
||||||
const tagRes = processTagSet(params.Tagging.TagSet);
|
|
||||||
if (tagRes instanceof Error) {
|
|
||||||
return callback(tagRes);
|
|
||||||
}
|
|
||||||
return async.waterfall([
|
|
||||||
next => this.headObject({
|
|
||||||
Bucket: params.Bucket,
|
|
||||||
Key: params.Key,
|
|
||||||
VersionId: params.VersionId,
|
|
||||||
}, next),
|
|
||||||
(resObj, next) => {
|
|
||||||
const completeMD = Object.assign({}, resObj.Metadata, tagRes);
|
|
||||||
this.copyObject({
|
|
||||||
Bucket: params.Bucket,
|
|
||||||
Key: params.Key,
|
|
||||||
CopySource: `${params.Bucket}/${params.Key}`,
|
|
||||||
Metadata: completeMD,
|
|
||||||
MetadataDirective: 'REPLACE',
|
|
||||||
}, next);
|
|
||||||
},
|
|
||||||
], callback);
|
|
||||||
}
|
|
||||||
|
|
||||||
module.exports = putObjectTagging;
|
|
|
@ -1,43 +0,0 @@
|
||||||
const { errors } = require('arsenal');
|
|
||||||
const { getPartNumber, createMpuKey, logger } = require('../GcpUtils');
|
|
||||||
const { logHelper } = require('../../utils');
|
|
||||||
|
|
||||||
/**
|
|
||||||
* uploadPart - upload part
|
|
||||||
* @param {object} params - upload part params
|
|
||||||
* @param {string} params.Bucket - bucket name
|
|
||||||
* @param {string} params.Key - object key
|
|
||||||
* @param {function} callback - callback function to call
|
|
||||||
* @return {undefined}
|
|
||||||
*/
|
|
||||||
function uploadPart(params, callback) {
|
|
||||||
if (!params || !params.UploadId || !params.Bucket || !params.Key) {
|
|
||||||
const error = errors.InvalidRequest
|
|
||||||
.customizeDescription('Missing required parameter');
|
|
||||||
logHelper(logger, 'error', 'error in uploadPart', error);
|
|
||||||
return callback(error);
|
|
||||||
}
|
|
||||||
const partNumber = getPartNumber(params.PartNumber);
|
|
||||||
if (!partNumber) {
|
|
||||||
const error = errors.InvalidArgument
|
|
||||||
.customizeDescription('PartNumber is invalid');
|
|
||||||
logHelper(logger, 'debug', 'error in uploadPart', error);
|
|
||||||
return callback(error);
|
|
||||||
}
|
|
||||||
const mpuParams = {
|
|
||||||
Bucket: params.Bucket,
|
|
||||||
Key: createMpuKey(params.Key, params.UploadId, partNumber),
|
|
||||||
Body: params.Body,
|
|
||||||
ContentLength: params.ContentLength,
|
|
||||||
};
|
|
||||||
return this.putObjectReq(mpuParams, (err, res) => {
|
|
||||||
if (err) {
|
|
||||||
logHelper(logger, 'error',
|
|
||||||
'error in uploadPart - putObject', err);
|
|
||||||
return callback(err);
|
|
||||||
}
|
|
||||||
return callback(null, res);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
module.exports = uploadPart;
|
|
|
@ -1,37 +0,0 @@
|
||||||
const { errors } = require('arsenal');
|
|
||||||
const { getPartNumber, createMpuKey, logger } = require('../GcpUtils');
|
|
||||||
const { logHelper } = require('../../utils');
|
|
||||||
|
|
||||||
/**
|
|
||||||
* uploadPartCopy - upload part copy
|
|
||||||
* @param {object} params - upload part copy params
|
|
||||||
* @param {string} params.Bucket - bucket name
|
|
||||||
* @param {string} params.Key - object key
|
|
||||||
* @param {string} params.CopySource - source object to copy
|
|
||||||
* @param {function} callback - callback function to call
|
|
||||||
* @return {undefined}
|
|
||||||
*/
|
|
||||||
function uploadPartCopy(params, callback) {
|
|
||||||
if (!params || !params.UploadId || !params.Bucket || !params.Key ||
|
|
||||||
!params.CopySource) {
|
|
||||||
const error = errors.InvalidRequest
|
|
||||||
.customizeDescription('Missing required parameter');
|
|
||||||
logHelper(logger, 'error', 'error in uploadPartCopy', error);
|
|
||||||
return callback(error);
|
|
||||||
}
|
|
||||||
const partNumber = getPartNumber(params.PartNumber);
|
|
||||||
if (!partNumber) {
|
|
||||||
const error = errors.InvalidArgument
|
|
||||||
.customizeDescription('PartNumber is not a number');
|
|
||||||
logHelper(logger, 'debug', 'error in uploadPartCopy', error);
|
|
||||||
return callback(error);
|
|
||||||
}
|
|
||||||
const mpuParams = {
|
|
||||||
Bucket: params.Bucket,
|
|
||||||
Key: createMpuKey(params.Key, params.UploadId, partNumber),
|
|
||||||
CopySource: params.CopySource,
|
|
||||||
};
|
|
||||||
return this.copyObject(mpuParams, callback);
|
|
||||||
}
|
|
||||||
|
|
||||||
module.exports = uploadPartCopy;
|
|
|
@ -1,402 +0,0 @@
|
||||||
const async = require('async');
|
|
||||||
const assert = require('assert');
|
|
||||||
const stream = require('stream');
|
|
||||||
const { errors } = require('arsenal');
|
|
||||||
const { minimumAllowedPartSize, gcpMaximumAllowedPartCount } =
|
|
||||||
require('../../../../constants');
|
|
||||||
const { createMpuList, logger } = require('./GcpUtils');
|
|
||||||
const { logHelper } = require('../utils');
|
|
||||||
|
|
||||||
|
|
||||||
function sliceFn(body, size) {
|
|
||||||
const array = [];
|
|
||||||
let partNumber = 1;
|
|
||||||
for (let ind = 0; ind < body.length; ind += size) {
|
|
||||||
array.push({
|
|
||||||
Body: body.slice(ind, ind + size),
|
|
||||||
PartNumber: partNumber++,
|
|
||||||
});
|
|
||||||
}
|
|
||||||
return array;
|
|
||||||
}
|
|
||||||
|
|
||||||
class GcpManagedUpload {
|
|
||||||
/**
|
|
||||||
* GcpMangedUpload - class to mimic the upload method in AWS-SDK
|
|
||||||
* To-Do: implement retry on failure like S3's upload
|
|
||||||
* @param {GcpService} service - client object
|
|
||||||
* @param {object} params - upload params
|
|
||||||
* @param {string} params.Bucket - bucket name
|
|
||||||
* @param {string} params.MPU - mpu bucket name
|
|
||||||
* @param {string} params.Key - object key
|
|
||||||
* @param {object} options - config setting for GcpManagedUpload object
|
|
||||||
* @param {number} options.partSize - set object chunk size
|
|
||||||
* @param {number} options.queueSize - set the number of concurrent upload
|
|
||||||
* @return {object} - return an GcpManagedUpload object
|
|
||||||
*/
|
|
||||||
constructor(service, params, options = {}) {
|
|
||||||
this.service = service;
|
|
||||||
this.params = params;
|
|
||||||
this.mainBucket =
|
|
||||||
this.params.Bucket || this.service.config.mainBucket;
|
|
||||||
this.mpuBucket =
|
|
||||||
this.params.MPU || this.service.config.mpuBucket;
|
|
||||||
this.partSize = minimumAllowedPartSize;
|
|
||||||
this.queueSize = options.queueSize || 4;
|
|
||||||
this.validateBody();
|
|
||||||
this.setPartSize();
|
|
||||||
// multipart information
|
|
||||||
this.parts = {};
|
|
||||||
this.uploadedParts = 0;
|
|
||||||
this.activeParts = 0;
|
|
||||||
this.partBuffers = [];
|
|
||||||
this.partQueue = [];
|
|
||||||
this.partBufferLength = 0;
|
|
||||||
this.totalChunkedBytes = 0;
|
|
||||||
this.partNumber = 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* validateBody - validate that body contains data to upload. If body is not
|
|
||||||
* of type stream, it must then be of either string or buffer. If string,
|
|
||||||
* convert to a Buffer type and split into chunks if body is large enough
|
|
||||||
* @return {undefined}
|
|
||||||
*/
|
|
||||||
validateBody() {
|
|
||||||
this.body = this.params.Body;
|
|
||||||
assert(this.body, errors.MissingRequestBodyError.customizeDescription(
|
|
||||||
'Missing request body'));
|
|
||||||
this.totalBytes = this.params.ContentLength;
|
|
||||||
if (this.body instanceof stream) {
|
|
||||||
assert.strictEqual(typeof this.totalBytes, 'number',
|
|
||||||
errors.MissingContentLength.customizeDescription(
|
|
||||||
'If body is a stream, ContentLength must be provided'));
|
|
||||||
} else {
|
|
||||||
if (typeof this.body === 'string') {
|
|
||||||
this.body = Buffer.from(this.body);
|
|
||||||
}
|
|
||||||
this.totalBytes = this.body.byteLength;
|
|
||||||
assert(this.totalBytes, errors.InternalError.customizeDescription(
|
|
||||||
'Unable to perform upload'));
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
setPartSize() {
|
|
||||||
const newPartSize =
|
|
||||||
Math.ceil(this.totalBytes / gcpMaximumAllowedPartCount);
|
|
||||||
if (newPartSize > this.partSize) this.partSize = newPartSize;
|
|
||||||
this.totalParts = Math.ceil(this.totalBytes / this.partSize);
|
|
||||||
if (this.body instanceof Buffer && this.totalParts > 1) {
|
|
||||||
this.slicedParts = sliceFn(this.body, this.partSize);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* cleanUp - function that is called if GcpManagedUpload fails at any point,
|
|
||||||
* perform clean up of used resources. Ends the request by calling an
|
|
||||||
* internal callback function
|
|
||||||
* @param {Error} err - Error object
|
|
||||||
* @return {undefined}
|
|
||||||
*/
|
|
||||||
cleanUp(err) {
|
|
||||||
// is only called when an error happens
|
|
||||||
if (this.failed || this.completed) {
|
|
||||||
return undefined;
|
|
||||||
}
|
|
||||||
this.failed = true;
|
|
||||||
if (this.uploadId) {
|
|
||||||
// if MPU was successfuly created
|
|
||||||
return this.abortMPU(mpuErr => {
|
|
||||||
if (mpuErr) {
|
|
||||||
logHelper(logger, 'error',
|
|
||||||
'GcpMangedUpload: abortMPU failed in cleanup', mpuErr);
|
|
||||||
}
|
|
||||||
return this.callback(err);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
return this.callback(err);
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* abortMPU - function that is called to remove a multipart upload
|
|
||||||
* @param {function} callback - callback function to call to complete the
|
|
||||||
* upload
|
|
||||||
* @return {undefined}
|
|
||||||
*/
|
|
||||||
abortMPU(callback) {
|
|
||||||
const params = {
|
|
||||||
Bucket: this.mainBucket,
|
|
||||||
MPU: this.mpuBucket,
|
|
||||||
UploadId: this.uploadId,
|
|
||||||
Key: this.params.Key,
|
|
||||||
};
|
|
||||||
this.service.abortMultipartUpload(params, callback);
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* completeUpload - function that is called to to complete a multipart
|
|
||||||
* upload
|
|
||||||
* @param {function} callback - callback function to call to complete the
|
|
||||||
* upload
|
|
||||||
* @return {undefined}
|
|
||||||
*/
|
|
||||||
completeUpload() {
|
|
||||||
if (this.failed || this.completed) {
|
|
||||||
return undefined;
|
|
||||||
}
|
|
||||||
const params = {
|
|
||||||
Bucket: this.mainBucket,
|
|
||||||
MPU: this.mpuBucket,
|
|
||||||
Key: this.params.Key,
|
|
||||||
UploadId: this.uploadId,
|
|
||||||
MultipartUpload: {},
|
|
||||||
};
|
|
||||||
params.MultipartUpload.Parts =
|
|
||||||
createMpuList(params, 'parts', this.uploadedParts)
|
|
||||||
.map(item =>
|
|
||||||
Object.assign(item, { ETag: this.parts[item.PartNumber] }));
|
|
||||||
return this.service.completeMultipartUpload(params,
|
|
||||||
(err, res) => {
|
|
||||||
if (err) {
|
|
||||||
return this.cleanUp(err);
|
|
||||||
}
|
|
||||||
this.completed = true;
|
|
||||||
return this.callback(null, res);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* send - function that is called to execute the method request
|
|
||||||
* @param {function} callback - callback function to be called and stored
|
|
||||||
* at the completion of the method
|
|
||||||
* @return {undefined}
|
|
||||||
*/
|
|
||||||
send(callback) {
|
|
||||||
if (this.called || this.callback) {
|
|
||||||
return undefined;
|
|
||||||
}
|
|
||||||
this.failed = false;
|
|
||||||
this.called = true;
|
|
||||||
this.callback = callback;
|
|
||||||
if (this.totalBytes <= this.partSize) {
|
|
||||||
return this.uploadSingle();
|
|
||||||
}
|
|
||||||
if (this.slicedParts) {
|
|
||||||
return this.uploadBufferSlices();
|
|
||||||
}
|
|
||||||
if (this.body instanceof stream) {
|
|
||||||
// stream type
|
|
||||||
this.body.on('error', err => this.cleanUp(err))
|
|
||||||
.on('readable', () => this.chunkStream())
|
|
||||||
.on('end', () => {
|
|
||||||
this.isDoneChunking = true;
|
|
||||||
this.chunkStream();
|
|
||||||
|
|
||||||
if (this.isDoneChunking && this.uploadedParts >= 1 &&
|
|
||||||
this.uploadedParts === this.totalParts) {
|
|
||||||
this.completeUpload();
|
|
||||||
}
|
|
||||||
});
|
|
||||||
}
|
|
||||||
return undefined;
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* uploadSingle - perform a regular put object upload if the object is
|
|
||||||
* small enough
|
|
||||||
* @return {undefined}
|
|
||||||
*/
|
|
||||||
uploadSingle() {
|
|
||||||
if (this.failed || this.completed) {
|
|
||||||
return undefined;
|
|
||||||
}
|
|
||||||
// use putObject to upload the single part object
|
|
||||||
const params = Object.assign({}, this.params);
|
|
||||||
params.Bucket = this.mainBucket;
|
|
||||||
delete params.MPU;
|
|
||||||
return this.service.putObject(params, (err, res) => {
|
|
||||||
if (err) {
|
|
||||||
return this.cleanUp(err);
|
|
||||||
}
|
|
||||||
// return results from a putObject request
|
|
||||||
this.completed = true;
|
|
||||||
return this.callback(null, res);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* uploadBufferSlices - perform a multipart upload for body of type string
|
|
||||||
* or Buffer.
|
|
||||||
* @return {undefined}
|
|
||||||
*/
|
|
||||||
uploadBufferSlices() {
|
|
||||||
if (this.failed || this.completed) {
|
|
||||||
return undefined;
|
|
||||||
}
|
|
||||||
if (this.slicedParts.length <= 1 && this.totalParts) {
|
|
||||||
// there is only one part
|
|
||||||
return this.uploadSingle();
|
|
||||||
}
|
|
||||||
// multiple slices
|
|
||||||
return async.series([
|
|
||||||
// createMultipartUpload
|
|
||||||
next => {
|
|
||||||
const params = this.params;
|
|
||||||
params.Bucket = this.mpuBucket;
|
|
||||||
this.service.createMultipartUpload(params, (err, res) => {
|
|
||||||
if (!err) {
|
|
||||||
this.uploadId = res.UploadId;
|
|
||||||
}
|
|
||||||
return next(err);
|
|
||||||
});
|
|
||||||
},
|
|
||||||
next => async.eachLimit(this.slicedParts, this.queueSize,
|
|
||||||
(uploadPart, done) => {
|
|
||||||
const params = {
|
|
||||||
Bucket: this.mpuBucket,
|
|
||||||
Key: this.params.Key,
|
|
||||||
UploadId: this.uploadId,
|
|
||||||
Body: uploadPart.Body,
|
|
||||||
PartNumber: uploadPart.PartNumber,
|
|
||||||
};
|
|
||||||
this.service.uploadPart(params, (err, res) => {
|
|
||||||
if (!err) {
|
|
||||||
this.parts[uploadPart.PartNumber] = res.ETag;
|
|
||||||
this.uploadedParts++;
|
|
||||||
}
|
|
||||||
return done(err);
|
|
||||||
});
|
|
||||||
}, next),
|
|
||||||
], err => {
|
|
||||||
if (err) {
|
|
||||||
return this.cleanUp(new Error(
|
|
||||||
'GcpManagedUpload: unable to complete upload'));
|
|
||||||
}
|
|
||||||
return this.completeUpload();
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* chunkStream - read stream up until the max chunk size then call an
|
|
||||||
* uploadPart method on that chunk. If more than chunk size has be read,
|
|
||||||
* move the extra bytes into a queue for the next read.
|
|
||||||
* @return {undefined}
|
|
||||||
*/
|
|
||||||
chunkStream() {
|
|
||||||
const buf = this.body.read(this.partSize - this.partBufferLength) ||
|
|
||||||
this.body.read();
|
|
||||||
|
|
||||||
if (buf) {
|
|
||||||
this.partBuffers.push(buf);
|
|
||||||
this.partBufferLength += buf.length;
|
|
||||||
this.totalChunkedBytes += buf.length;
|
|
||||||
}
|
|
||||||
|
|
||||||
let pbuf;
|
|
||||||
if (this.partBufferLength >= this.partSize) {
|
|
||||||
pbuf = Buffer.concat(this.partBuffers);
|
|
||||||
this.partBuffers = [];
|
|
||||||
this.partBufferLength = 0;
|
|
||||||
|
|
||||||
if (pbuf.length > this.partSize) {
|
|
||||||
const rest = pbuf.slice(this.partSize);
|
|
||||||
this.partBuffers.push(rest);
|
|
||||||
this.partBufferLength += rest.length;
|
|
||||||
pbuf = pbuf.slice(0, this.partSize);
|
|
||||||
}
|
|
||||||
this.processChunk(pbuf);
|
|
||||||
}
|
|
||||||
|
|
||||||
// when chunking the last part
|
|
||||||
if (this.isDoneChunking && !this.completeChunking) {
|
|
||||||
this.completeChunking = true;
|
|
||||||
pbuf = Buffer.concat(this.partBuffers);
|
|
||||||
this.partBuffers = [];
|
|
||||||
this.partBufferLength = 0;
|
|
||||||
if (pbuf.length > 0) {
|
|
||||||
this.processChunk(pbuf);
|
|
||||||
} else {
|
|
||||||
if (this.uploadedParts === 0) {
|
|
||||||
// this is a 0-byte object
|
|
||||||
this.uploadSingle();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
this.body.read(0);
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* processChunk - create a multipart request if one does not exist;
|
|
||||||
* otherwise, call uploadChunk to upload a chunk
|
|
||||||
* @param {Buffer} chunk - bytes to be uploaded
|
|
||||||
* @return {undefined}
|
|
||||||
*/
|
|
||||||
processChunk(chunk) {
|
|
||||||
const partNumber = ++this.partNumber;
|
|
||||||
if (!this.uploadId) {
|
|
||||||
// if multipart upload does not exist
|
|
||||||
if (!this.multipartReq) {
|
|
||||||
const params = this.params;
|
|
||||||
params.Bucket = this.mpuBucket;
|
|
||||||
this.multipartReq =
|
|
||||||
this.service.createMultipartUpload(params, (err, res) => {
|
|
||||||
if (err) {
|
|
||||||
return this.cleanUp(err);
|
|
||||||
}
|
|
||||||
this.uploadId = res.UploadId;
|
|
||||||
this.uploadChunk(chunk, partNumber);
|
|
||||||
if (this.partQueue.length > 0) {
|
|
||||||
this.partQueue.forEach(
|
|
||||||
part => this.uploadChunk(...part));
|
|
||||||
}
|
|
||||||
return undefined;
|
|
||||||
});
|
|
||||||
} else {
|
|
||||||
this.partQueue.push([chunk, partNumber]);
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
// queues chunks for upload
|
|
||||||
this.uploadChunk(chunk, partNumber);
|
|
||||||
this.activeParts++;
|
|
||||||
if (this.activeParts < this.queueSize) {
|
|
||||||
this.chunkStream();
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* uploadChunk - perform the partUpload
|
|
||||||
* @param {Buffer} chunk - bytes to be uploaded
|
|
||||||
* @param {number} partNumber - upload object part number
|
|
||||||
* @return {undefined}
|
|
||||||
*/
|
|
||||||
uploadChunk(chunk, partNumber) {
|
|
||||||
if (this.failed || this.completed) {
|
|
||||||
return undefined;
|
|
||||||
}
|
|
||||||
const params = {
|
|
||||||
Bucket: this.mpuBucket,
|
|
||||||
Key: this.params.Key,
|
|
||||||
UploadId: this.uploadId,
|
|
||||||
PartNumber: partNumber,
|
|
||||||
Body: chunk,
|
|
||||||
ContentLength: chunk.length,
|
|
||||||
};
|
|
||||||
return this.service.uploadPart(params, (err, res) => {
|
|
||||||
if (err) {
|
|
||||||
return this.cleanUp(err);
|
|
||||||
}
|
|
||||||
this.parts[partNumber] = res.ETag;
|
|
||||||
this.uploadedParts++;
|
|
||||||
this.activeParts--;
|
|
||||||
if (this.totalParts === this.uploadedParts &&
|
|
||||||
this.isDoneChunking) {
|
|
||||||
return this.completeUpload();
|
|
||||||
}
|
|
||||||
return this.chunkStream();
|
|
||||||
});
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
module.exports = GcpManagedUpload;
|
|
|
@ -1,141 +0,0 @@
|
||||||
const AWS = require('aws-sdk');
|
|
||||||
const { errors } = require('arsenal');
|
|
||||||
const Service = AWS.Service;
|
|
||||||
|
|
||||||
const GcpApis = require('./GcpApis');
|
|
||||||
const GcpServiceSetup = require('./GcpServiceSetup');
|
|
||||||
const GcpManagedUpload = require('./GcpManagedUpload');
|
|
||||||
|
|
||||||
AWS.apiLoader.services.gcp = {};
|
|
||||||
const GCP = Service.defineService('gcp', ['2017-11-01']);
|
|
||||||
|
|
||||||
Object.assign(GCP.prototype, GcpServiceSetup, {
|
|
||||||
_maxConcurrent: 5,
|
|
||||||
|
|
||||||
// Implemented APIs
|
|
||||||
// Bucket API
|
|
||||||
getBucket(params, callback) {
|
|
||||||
return this.listObjects(params, callback);
|
|
||||||
},
|
|
||||||
|
|
||||||
// Object APIs
|
|
||||||
upload(params, callback) {
|
|
||||||
try {
|
|
||||||
const uploader = new GcpManagedUpload(this, params);
|
|
||||||
return uploader.send(callback);
|
|
||||||
} catch (err) {
|
|
||||||
return callback(err);
|
|
||||||
}
|
|
||||||
},
|
|
||||||
|
|
||||||
putObjectCopy(params, callback) {
|
|
||||||
return this.copyObject(params, callback);
|
|
||||||
},
|
|
||||||
|
|
||||||
// TO-DO: Implement the following APIs
|
|
||||||
// Service API
|
|
||||||
listBuckets(params, callback) {
|
|
||||||
return callback(errors.NotImplemented
|
|
||||||
.customizeDescription('GCP: listBuckets not implemented'));
|
|
||||||
},
|
|
||||||
|
|
||||||
// Bucket APIs
|
|
||||||
getBucketLocation(params, callback) {
|
|
||||||
return callback(errors.NotImplemented
|
|
||||||
.customizeDescription('GCP: getBucketLocation not implemented'));
|
|
||||||
},
|
|
||||||
|
|
||||||
deleteBucket(params, callback) {
|
|
||||||
return callback(errors.NotImplemented
|
|
||||||
.customizeDescription('GCP: deleteBucket not implemented'));
|
|
||||||
},
|
|
||||||
|
|
||||||
listObjectVersions(params, callback) {
|
|
||||||
return callback(errors.NotImplemented
|
|
||||||
.customizeDescription('GCP: listObjectVersions not implemented'));
|
|
||||||
},
|
|
||||||
|
|
||||||
createBucket(params, callback) {
|
|
||||||
return callback(errors.NotImplemented
|
|
||||||
.customizeDescription('GCP: createBucket not implemented'));
|
|
||||||
},
|
|
||||||
|
|
||||||
putBucket(params, callback) {
|
|
||||||
return callback(errors.NotImplemented
|
|
||||||
.customizeDescription('GCP: putBucket not implemented'));
|
|
||||||
},
|
|
||||||
|
|
||||||
getBucketAcl(params, callback) {
|
|
||||||
return callback(errors.NotImplemented
|
|
||||||
.customizeDescription('GCP: getBucketAcl not implemented'));
|
|
||||||
},
|
|
||||||
|
|
||||||
putBucketAcl(params, callback) {
|
|
||||||
return callback(errors.NotImplemented
|
|
||||||
.customizeDescription('GCP: putBucketAcl not implemented'));
|
|
||||||
},
|
|
||||||
|
|
||||||
putBucketWebsite(params, callback) {
|
|
||||||
return callback(errors.NotImplemented
|
|
||||||
.customizeDescription('GCP: putBucketWebsite not implemented'));
|
|
||||||
},
|
|
||||||
|
|
||||||
getBucketWebsite(params, callback) {
|
|
||||||
return callback(errors.NotImplemented
|
|
||||||
.customizeDescription('GCP: getBucketWebsite not implemented'));
|
|
||||||
},
|
|
||||||
|
|
||||||
deleteBucketWebsite(params, callback) {
|
|
||||||
return callback(errors.NotImplemented
|
|
||||||
.customizeDescription('GCP: deleteBucketWebsite not implemented'));
|
|
||||||
},
|
|
||||||
|
|
||||||
putBucketCors(params, callback) {
|
|
||||||
return callback(errors.NotImplemented
|
|
||||||
.customizeDescription('GCP: putBucketCors not implemented'));
|
|
||||||
},
|
|
||||||
|
|
||||||
getBucketCors(params, callback) {
|
|
||||||
return callback(errors.NotImplemented
|
|
||||||
.customizeDescription('GCP: getBucketCors not implemented'));
|
|
||||||
},
|
|
||||||
|
|
||||||
deleteBucketCors(params, callback) {
|
|
||||||
return callback(errors.NotImplemented
|
|
||||||
.customizeDescription('GCP: deleteBucketCors not implemented'));
|
|
||||||
},
|
|
||||||
|
|
||||||
// Object APIs
|
|
||||||
putObjectTagging(params, callback) {
|
|
||||||
return callback(errors.NotImplemented
|
|
||||||
.customizeDescription('GCP: putObjectTagging not implemented'));
|
|
||||||
},
|
|
||||||
|
|
||||||
deleteObjectTagging(params, callback) {
|
|
||||||
return callback(errors.NotImplemented
|
|
||||||
.customizeDescription('GCP: deleteObjectTagging not implemented'));
|
|
||||||
},
|
|
||||||
|
|
||||||
putObjectAcl(params, callback) {
|
|
||||||
return callback(errors.NotImplemented
|
|
||||||
.customizeDescription('GCP: putObjectAcl not implemented'));
|
|
||||||
},
|
|
||||||
|
|
||||||
getObjectAcl(params, callback) {
|
|
||||||
return callback(errors.NotImplemented
|
|
||||||
.customizeDescription('GCP: getObjectAcl not implemented'));
|
|
||||||
},
|
|
||||||
});
|
|
||||||
|
|
||||||
Object.assign(GCP.prototype, GcpApis);
|
|
||||||
|
|
||||||
Object.defineProperty(AWS.apiLoader.services.gcp, '2017-11-01', {
|
|
||||||
get: function get() {
|
|
||||||
const model = require('./gcp-2017-11-01.api.json');
|
|
||||||
return model;
|
|
||||||
},
|
|
||||||
enumerable: true,
|
|
||||||
configurable: true,
|
|
||||||
});
|
|
||||||
|
|
||||||
module.exports = GCP;
|
|
|
@ -1,246 +0,0 @@
|
||||||
/**
|
|
||||||
* This file contains the private methods for the GCP service to form/sign
|
|
||||||
* Google Cloud Storage requests.
|
|
||||||
*
|
|
||||||
* It uses a modified version of the S3 service private methods to form/sign
|
|
||||||
* requests compatible with the Google Cloud Storage XML api.
|
|
||||||
*/
|
|
||||||
|
|
||||||
const AWS = require('aws-sdk');
|
|
||||||
const GcpSigner = require('./GcpSigner');
|
|
||||||
|
|
||||||
module.exports = {
|
|
||||||
/**
|
|
||||||
* GCP compatible request signer
|
|
||||||
* @return {AWS.Signer} AWS Signer
|
|
||||||
*/
|
|
||||||
getSignerClass() {
|
|
||||||
return GcpSigner;
|
|
||||||
},
|
|
||||||
|
|
||||||
/**
|
|
||||||
* set service region
|
|
||||||
* @return {undefined}
|
|
||||||
*/
|
|
||||||
validateService() {
|
|
||||||
if (!this.config.region) {
|
|
||||||
this.config.region = 'us-east-1';
|
|
||||||
}
|
|
||||||
},
|
|
||||||
|
|
||||||
/**
|
|
||||||
* setup listeners for building requests
|
|
||||||
* @param {AWS.Request} request - AWS request object
|
|
||||||
* @return {undefined}
|
|
||||||
*/
|
|
||||||
setupRequestListeners(request) {
|
|
||||||
request.addListener('validate', this.validateBucketEndpoint);
|
|
||||||
request.addListener('build', this.addContentType);
|
|
||||||
request.addListener('build', this.populateURI);
|
|
||||||
request.addListener('build', this.computeContentMd5);
|
|
||||||
request.addListener('extractError', this.extractRequestIds);
|
|
||||||
},
|
|
||||||
|
|
||||||
/**
|
|
||||||
* validate that when bucket endpoitn flag is set, root level apis are
|
|
||||||
* inaccessible
|
|
||||||
* @param {AWS.Request} req - AWS request object
|
|
||||||
* @returns {undefined}
|
|
||||||
* @api private
|
|
||||||
*/
|
|
||||||
validateBucketEndpoint(req) {
|
|
||||||
if (!req.params.Bucket && req.service.config.s3BucketEndpoint) {
|
|
||||||
const msg =
|
|
||||||
'Cannot send requests to root API with `s3BucketEndpoint` set.';
|
|
||||||
throw AWS.util.error(new Error(),
|
|
||||||
{ code: 'ConfigError', message: msg });
|
|
||||||
}
|
|
||||||
},
|
|
||||||
|
|
||||||
/**
|
|
||||||
* S3 prefers dns-compatible bucket names to be moved from the uri path
|
|
||||||
* to the hostname as a sub-domain. This is not possible, even for
|
|
||||||
* dns-compat buckets when using SSL and the bucket name contains a dot.
|
|
||||||
* The ssl wildcard certificate is only 1-level deep.
|
|
||||||
* @param {AWS.Request} req - AWS request object
|
|
||||||
* @returns {undefined}
|
|
||||||
*
|
|
||||||
* @api private
|
|
||||||
*/
|
|
||||||
populateURI(req) {
|
|
||||||
const httpRequest = req.httpRequest;
|
|
||||||
const b = req.params.Bucket;
|
|
||||||
const service = req.service;
|
|
||||||
const endpoint = httpRequest.endpoint;
|
|
||||||
|
|
||||||
if (b) {
|
|
||||||
if (!service.pathStyleBucketName(b)) {
|
|
||||||
if (!service.config.s3BucketEndpoint) {
|
|
||||||
endpoint.hostname = `${b}.${endpoint.hostname}`;
|
|
||||||
}
|
|
||||||
const port = endpoint.port;
|
|
||||||
if (port !== 80 && port !== 443) {
|
|
||||||
endpoint.host = `${endpoint.hostname}:${endpoint.port}`;
|
|
||||||
} else {
|
|
||||||
endpoint.host = endpoint.hostname;
|
|
||||||
}
|
|
||||||
// needed for signing the request
|
|
||||||
httpRequest.virtualHostedBucket = b;
|
|
||||||
service.removeVirtualHostedBucketFromPath(req);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Takes the bucket name out of the path if bucket is virtual-hosted
|
|
||||||
* @param {AWS.Request} req - AWS request object
|
|
||||||
* @returns {undefined}
|
|
||||||
*
|
|
||||||
* @api private
|
|
||||||
*/
|
|
||||||
removeVirtualHostedBucketFromPath(req) {
|
|
||||||
const httpRequest = req.httpRequest;
|
|
||||||
const bucket = httpRequest.virtualHostedBucket;
|
|
||||||
if (bucket && httpRequest.path) {
|
|
||||||
httpRequest.path =
|
|
||||||
httpRequest.path.replace(new RegExp(`/${bucket}`), '');
|
|
||||||
if (httpRequest.path[0] !== '/') {
|
|
||||||
httpRequest.path = `/${httpRequest.path}`;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Adds a default content type if none is supplied.
|
|
||||||
* @param {AWS.Request} req - AWS request object
|
|
||||||
* @returns {undefined}
|
|
||||||
*
|
|
||||||
* @api private
|
|
||||||
*/
|
|
||||||
addContentType(req) {
|
|
||||||
const httpRequest = req.httpRequest;
|
|
||||||
if (httpRequest.method === 'GET' || httpRequest.method === 'HEAD') {
|
|
||||||
// Content-Type is not set in GET/HEAD requests
|
|
||||||
delete httpRequest.headers['Content-Type'];
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
// always have a Content-Type
|
|
||||||
if (!httpRequest.headers['Content-Type']) {
|
|
||||||
httpRequest.headers['Content-Type'] = 'application/octet-stream';
|
|
||||||
}
|
|
||||||
|
|
||||||
const contentType = httpRequest.headers['Content-Type'];
|
|
||||||
if (AWS.util.isBrowser()) {
|
|
||||||
if (typeof httpRequest.body === 'string' &&
|
|
||||||
!contentType.match(/;\s*charset=/)) {
|
|
||||||
const charset = '; charset=UTF-8';
|
|
||||||
httpRequest.headers['Content-Type'] += charset;
|
|
||||||
} else {
|
|
||||||
const replaceFn = (_, prefix, charsetName) =>
|
|
||||||
prefix + charsetName.toUpperCase();
|
|
||||||
|
|
||||||
httpRequest.headers['Content-Type'] =
|
|
||||||
contentType.replace(/(;\s*charset=)(.+)$/, replaceFn);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
|
||||||
|
|
||||||
computableChecksumOperations: {
|
|
||||||
putBucketCors: true,
|
|
||||||
putBucketLifecycle: true,
|
|
||||||
putBucketLifecycleConfiguration: true,
|
|
||||||
putBucketTagging: true,
|
|
||||||
deleteObjects: true,
|
|
||||||
putBucketReplication: true,
|
|
||||||
},
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Checks whether checksums should be computed for the request.
|
|
||||||
* If the request requires checksums to be computed, this will always
|
|
||||||
* return true, otherwise it depends on whether
|
|
||||||
* {AWS.Config.computeChecksums} is set.
|
|
||||||
* @param {AWS.Request} req - the request to check against
|
|
||||||
* @return {Boolean} whether to compute checksums for a request.
|
|
||||||
*
|
|
||||||
* @api private
|
|
||||||
*/
|
|
||||||
willComputeChecksums(req) {
|
|
||||||
if (this.computableChecksumOperations[req.operation]) return true;
|
|
||||||
if (!this.config.computeChecksums) return false;
|
|
||||||
|
|
||||||
// TODO: compute checksums for Stream objects
|
|
||||||
if (!AWS.util.Buffer.isBuffer(req.httpRequest.body) &&
|
|
||||||
typeof req.httpRequest.body !== 'string') {
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
return false;
|
|
||||||
},
|
|
||||||
|
|
||||||
/**
|
|
||||||
* A listener that computes the Content-MD5 and sets it in the header.
|
|
||||||
* @param {AWS.Request} req - AWS request object
|
|
||||||
* @returns {undefined}
|
|
||||||
*
|
|
||||||
* @see AWS.S3.willComputeChecksums
|
|
||||||
* @api private
|
|
||||||
*/
|
|
||||||
computeContentMd5(req) {
|
|
||||||
if (req.service.willComputeChecksums(req)) {
|
|
||||||
const md5 = AWS.util.crypto.md5(req.httpRequest.body, 'base64');
|
|
||||||
// eslint-disable-next-line no-param-reassign
|
|
||||||
req.httpRequest.headers['Content-MD5'] = md5;
|
|
||||||
}
|
|
||||||
},
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Returns true if the bucket name should be left in the URI path for
|
|
||||||
* a request to S3. This function takes into account the current
|
|
||||||
* endpoint protocol (e.g. http or https).
|
|
||||||
* @param {string} bucketName - bucket name
|
|
||||||
* @returns {Boolean} whether request should use path style
|
|
||||||
*
|
|
||||||
* @api private
|
|
||||||
*/
|
|
||||||
pathStyleBucketName(bucketName) {
|
|
||||||
// user can force path style requests via the configuration
|
|
||||||
if (this.config.s3ForcePathStyle) return true;
|
|
||||||
if (this.config.s3BucketEndpoint) return false;
|
|
||||||
|
|
||||||
if (this.dnsCompatibleBucketName(bucketName)) {
|
|
||||||
return this.config.sslEnabled && bucketName.match(/\./);
|
|
||||||
}
|
|
||||||
return true; // not dns compatible names must always use path style
|
|
||||||
},
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Returns true if the bucket name is DNS compatible. Buckets created
|
|
||||||
* outside of the classic region MUST be DNS compatible.
|
|
||||||
* @param {string} bucketName - bucket name
|
|
||||||
* @returns {Boolean} whether bucket name is dns compatible
|
|
||||||
*
|
|
||||||
* @api private
|
|
||||||
*/
|
|
||||||
dnsCompatibleBucketName(bucketName) {
|
|
||||||
const b = bucketName;
|
|
||||||
const domain = new RegExp(/^[a-z0-9][a-z0-9\.\-]{1,61}[a-z0-9]$/);
|
|
||||||
const ipAddress = new RegExp(/(\d+\.){3}\d+/);
|
|
||||||
const dots = new RegExp(/\.\./);
|
|
||||||
return b.match(domain) && !b.match(ipAddress) && !b.match(dots);
|
|
||||||
},
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Extracts GCP specific request ids from the http response.
|
|
||||||
* @param {object} resp - response object
|
|
||||||
* @returns {undefined}
|
|
||||||
* @api private
|
|
||||||
*/
|
|
||||||
extractRequestIds(resp) {
|
|
||||||
const requestId = resp.httpResponse.headers ?
|
|
||||||
resp.httpResponse.headers['x-guploader-uploadid'] : null;
|
|
||||||
if (resp.error) {
|
|
||||||
// eslint-disable-next-line no-param-reassign
|
|
||||||
resp.error.requestId = resp.requestId || requestId;
|
|
||||||
}
|
|
||||||
},
|
|
||||||
};
|
|
|
@ -1,51 +0,0 @@
|
||||||
const url = require('url');
|
|
||||||
const qs = require('querystring');
|
|
||||||
const AWS = require('aws-sdk');
|
|
||||||
const werelogs = require('werelogs');
|
|
||||||
const { constructStringToSignV2 } = require('arsenal').auth.client;
|
|
||||||
|
|
||||||
const logger = new werelogs.Logger('GcpSigner');
|
|
||||||
|
|
||||||
function genQueryObject(uri) {
|
|
||||||
const queryString = url.parse(uri).query;
|
|
||||||
return qs.parse(queryString);
|
|
||||||
}
|
|
||||||
|
|
||||||
const GcpSigner = AWS.util.inherit(AWS.Signers.RequestSigner, {
|
|
||||||
constructor: function GcpSigner(request) {
|
|
||||||
AWS.Signers.RequestSigner.call(this, request);
|
|
||||||
},
|
|
||||||
|
|
||||||
addAuthorization: function addAuthorization(credentials, date) {
|
|
||||||
if (!this.request.headers['presigned-expires']) {
|
|
||||||
this.request.headers['x-goog-date'] = AWS.util.date.rfc822(date);
|
|
||||||
}
|
|
||||||
|
|
||||||
const signature =
|
|
||||||
this.sign(credentials.secretAccessKey, this.stringToSign());
|
|
||||||
const auth = `GOOG1 ${credentials.accessKeyId}: ${signature}`;
|
|
||||||
|
|
||||||
this.request.headers.Authorization = auth;
|
|
||||||
},
|
|
||||||
|
|
||||||
stringToSign: function stringToSign() {
|
|
||||||
const requestObject = {
|
|
||||||
url: this.request.path,
|
|
||||||
method: this.request.method,
|
|
||||||
host: this.request.endpoint.host,
|
|
||||||
headers: this.request.headers,
|
|
||||||
bucketName: this.request.virtualHostedBucket,
|
|
||||||
query: genQueryObject(this.request.path) || {},
|
|
||||||
};
|
|
||||||
requestObject.gotBucketNameFromHost =
|
|
||||||
requestObject.host.indexOf(this.request.virtualHostedBucket) >= 0;
|
|
||||||
const data = Object.assign({}, this.request.headers);
|
|
||||||
return constructStringToSignV2(requestObject, data, logger, 'GCP');
|
|
||||||
},
|
|
||||||
|
|
||||||
sign: function sign(secret, string) {
|
|
||||||
return AWS.util.crypto.hmac(secret, string, 'base64', 'sha1');
|
|
||||||
},
|
|
||||||
});
|
|
||||||
|
|
||||||
module.exports = GcpSigner;
|
|
|
@ -1,169 +0,0 @@
|
||||||
const werelogs = require('werelogs');
|
|
||||||
const { errors, s3middleware } = require('arsenal');
|
|
||||||
|
|
||||||
const _config = require('../../../Config').config;
|
|
||||||
const { gcpTaggingPrefix } = require('../../../../constants');
|
|
||||||
|
|
||||||
werelogs.configure({
|
|
||||||
level: _config.log.logLevel,
|
|
||||||
dump: _config.log.dumpLevel,
|
|
||||||
});
|
|
||||||
|
|
||||||
const logger = new werelogs.Logger('gcpUtil');
|
|
||||||
|
|
||||||
function eachSlice(size) {
|
|
||||||
this.array = [];
|
|
||||||
let partNumber = 1;
|
|
||||||
for (let ind = 0; ind < this.length; ind += size) {
|
|
||||||
this.array.push({
|
|
||||||
Parts: this.slice(ind, ind + size),
|
|
||||||
PartNumber: partNumber++,
|
|
||||||
});
|
|
||||||
}
|
|
||||||
return this.array;
|
|
||||||
}
|
|
||||||
|
|
||||||
function getSourceInfo(CopySource) {
|
|
||||||
const source =
|
|
||||||
CopySource.startsWith('/') ? CopySource.slice(1) : CopySource;
|
|
||||||
const sourceArray = source.split(/\/(.+)/);
|
|
||||||
const sourceBucket = sourceArray[0];
|
|
||||||
const sourceObject = sourceArray[1];
|
|
||||||
return { sourceBucket, sourceObject };
|
|
||||||
}
|
|
||||||
|
|
||||||
function getPaddedPartNumber(number) {
|
|
||||||
return `000000${number}`.substr(-5);
|
|
||||||
}
|
|
||||||
|
|
||||||
function getPartNumber(number) {
|
|
||||||
if (isNaN(number)) {
|
|
||||||
return undefined;
|
|
||||||
}
|
|
||||||
if (typeof number === 'string') {
|
|
||||||
return parseInt(number, 10);
|
|
||||||
}
|
|
||||||
return number;
|
|
||||||
}
|
|
||||||
|
|
||||||
function createMpuKey(key, uploadId, partNumberArg, fileNameArg) {
|
|
||||||
let partNumber = partNumberArg;
|
|
||||||
let fileName = fileNameArg;
|
|
||||||
|
|
||||||
if (typeof partNumber === 'string' && fileName === undefined) {
|
|
||||||
fileName = partNumber;
|
|
||||||
partNumber = null;
|
|
||||||
}
|
|
||||||
const paddedNumber = getPaddedPartNumber(partNumber);
|
|
||||||
if (fileName && typeof fileName === 'string') {
|
|
||||||
// if partNumber is given, return a "full file path"
|
|
||||||
// else return a "directory path"
|
|
||||||
return partNumber ? `${key}-${uploadId}/${fileName}/${paddedNumber}` :
|
|
||||||
`${key}-${uploadId}/${fileName}`;
|
|
||||||
}
|
|
||||||
if (partNumber && typeof partNumber === 'number') {
|
|
||||||
// filename wasn't passed as an argument. Create default
|
|
||||||
return `${key}-${uploadId}/parts/${paddedNumber}`;
|
|
||||||
}
|
|
||||||
// returns a "directory path"
|
|
||||||
return `${key}-${uploadId}/`;
|
|
||||||
}
|
|
||||||
|
|
||||||
function createMpuList(params, level, size) {
|
|
||||||
// populate and return a parts list for compose
|
|
||||||
const retList = [];
|
|
||||||
for (let i = 1; i <= size; ++i) {
|
|
||||||
retList.push({
|
|
||||||
PartName: createMpuKey(params.Key, params.UploadId, i, level),
|
|
||||||
PartNumber: i,
|
|
||||||
});
|
|
||||||
}
|
|
||||||
return retList;
|
|
||||||
}
|
|
||||||
|
|
||||||
function processTagSet(tagSet = []) {
|
|
||||||
if (tagSet.length > 10) {
|
|
||||||
return errors.BadRequest
|
|
||||||
.customizeDescription('Object tags cannot be greater than 10');
|
|
||||||
}
|
|
||||||
let error = undefined;
|
|
||||||
const tagAsMeta = {};
|
|
||||||
const taggingDict = {};
|
|
||||||
tagSet.every(tag => {
|
|
||||||
const { Key: key, Value: value } = tag;
|
|
||||||
if (key.length > 128) {
|
|
||||||
error = errors.InvalidTag
|
|
||||||
.customizeDescription(
|
|
||||||
`The TagKey provided is too long, ${key.length}`);
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
if (value.length > 256) {
|
|
||||||
error = errors.InvalidTag
|
|
||||||
.customizeDescription(
|
|
||||||
`The TagValue provided is too long, ${value.length}`);
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
if (taggingDict[key]) {
|
|
||||||
error = errors.InvalidTag
|
|
||||||
.customizeDescription(
|
|
||||||
'Cannot provide multiple Tags with the same key');
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
tagAsMeta[`${gcpTaggingPrefix}${key}`] = value;
|
|
||||||
taggingDict[key] = true;
|
|
||||||
return true;
|
|
||||||
});
|
|
||||||
if (error) {
|
|
||||||
return error;
|
|
||||||
}
|
|
||||||
return tagAsMeta;
|
|
||||||
}
|
|
||||||
|
|
||||||
function stripTags(metadata = {}) {
|
|
||||||
const retMD = Object.assign({}, metadata);
|
|
||||||
Object.keys(retMD).forEach(key => {
|
|
||||||
if (key.startsWith(gcpTaggingPrefix)) {
|
|
||||||
delete retMD[key];
|
|
||||||
}
|
|
||||||
});
|
|
||||||
return retMD;
|
|
||||||
}
|
|
||||||
|
|
||||||
function retrieveTags(metadata = {}) {
|
|
||||||
const retTagSet = [];
|
|
||||||
Object.keys(metadata).forEach(key => {
|
|
||||||
if (key.startsWith(gcpTaggingPrefix)) {
|
|
||||||
retTagSet.push({
|
|
||||||
Key: key.slice(gcpTaggingPrefix.length),
|
|
||||||
Value: metadata[key],
|
|
||||||
});
|
|
||||||
}
|
|
||||||
});
|
|
||||||
return retTagSet;
|
|
||||||
}
|
|
||||||
|
|
||||||
function getPutTagsMetadata(metadata, tagging = '') {
|
|
||||||
let retMetadata = metadata || {};
|
|
||||||
retMetadata = stripTags(retMetadata);
|
|
||||||
const tagObj = s3middleware.tagging.parseTagFromQuery(tagging);
|
|
||||||
Object.keys(tagObj).forEach(header => {
|
|
||||||
const prefixed = `${gcpTaggingPrefix}${header}`.toLowerCase();
|
|
||||||
retMetadata[prefixed] = tagObj[header];
|
|
||||||
});
|
|
||||||
return retMetadata;
|
|
||||||
}
|
|
||||||
|
|
||||||
module.exports = {
|
|
||||||
// functions
|
|
||||||
eachSlice,
|
|
||||||
createMpuKey,
|
|
||||||
createMpuList,
|
|
||||||
getSourceInfo,
|
|
||||||
processTagSet,
|
|
||||||
stripTags,
|
|
||||||
retrieveTags,
|
|
||||||
getPutTagsMetadata,
|
|
||||||
getPartNumber,
|
|
||||||
// util objects
|
|
||||||
logger,
|
|
||||||
};
|
|
|
@ -1,932 +0,0 @@
|
||||||
{
|
|
||||||
"version": "1.0",
|
|
||||||
"metadata": {
|
|
||||||
"apiVersion": "2017-11-01",
|
|
||||||
"checksumFormat": "md5",
|
|
||||||
"endpointPrefix": "storage",
|
|
||||||
"globalEndpoint": "storage.googleapi.com",
|
|
||||||
"protocol": "rest-xml",
|
|
||||||
"serviceAbbreviation": "GCP",
|
|
||||||
"serviceFullName": "Google Cloud Storage",
|
|
||||||
"signatureVersion": "s3",
|
|
||||||
"timestampFormat": "rfc822",
|
|
||||||
"uid": "gcp-2017-11-01"
|
|
||||||
},
|
|
||||||
"operations": {
|
|
||||||
"HeadBucket": {
|
|
||||||
"http": {
|
|
||||||
"method": "HEAD",
|
|
||||||
"requestUri": "/{Bucket}"
|
|
||||||
},
|
|
||||||
"input": {
|
|
||||||
"type": "structure",
|
|
||||||
"required": [
|
|
||||||
"Bucket"
|
|
||||||
],
|
|
||||||
"members": {
|
|
||||||
"Bucket": {
|
|
||||||
"location": "uri",
|
|
||||||
"locationName": "Bucket"
|
|
||||||
},
|
|
||||||
"ProjectId": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-project-id"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"output": {
|
|
||||||
"type": "structure",
|
|
||||||
"members": {
|
|
||||||
"MetaVersionId": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-metageneration"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"listObjects": {
|
|
||||||
"http": {
|
|
||||||
"method": "GET",
|
|
||||||
"requestUri": "/{Bucket}"
|
|
||||||
},
|
|
||||||
"input": {
|
|
||||||
"type": "structure",
|
|
||||||
"required": [
|
|
||||||
"Bucket"
|
|
||||||
],
|
|
||||||
"members": {
|
|
||||||
"Bucket": {
|
|
||||||
"location": "uri",
|
|
||||||
"locationName": "Bucket"
|
|
||||||
},
|
|
||||||
"Delimiter": {
|
|
||||||
"location": "querystring",
|
|
||||||
"locationName": "delimiter"
|
|
||||||
},
|
|
||||||
"Marker": {
|
|
||||||
"location": "querystring",
|
|
||||||
"locationName": "marker"
|
|
||||||
},
|
|
||||||
"MaxKeys": {
|
|
||||||
"location": "querystring",
|
|
||||||
"locationName": "max-keys",
|
|
||||||
"type": "integer"
|
|
||||||
},
|
|
||||||
"Prefix": {
|
|
||||||
"location": "querystring",
|
|
||||||
"locationName": "prefix"
|
|
||||||
},
|
|
||||||
"ProjectId": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-project-id"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"output": {
|
|
||||||
"type": "structure",
|
|
||||||
"members": {
|
|
||||||
"IsTruncated": {
|
|
||||||
"type": "boolean"
|
|
||||||
},
|
|
||||||
"Marker": {},
|
|
||||||
"NextMarker": {},
|
|
||||||
"Contents": {
|
|
||||||
"shape": "ContentsShape"
|
|
||||||
},
|
|
||||||
"Name": {},
|
|
||||||
"Prefix": {},
|
|
||||||
"Delimiter": {},
|
|
||||||
"MaxKeys": {
|
|
||||||
"type": "integer"
|
|
||||||
},
|
|
||||||
"CommonPrefixes": {
|
|
||||||
"shape": "CommonPrefixShape"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"listVersions": {
|
|
||||||
"http": {
|
|
||||||
"method": "GET",
|
|
||||||
"requestUri": "/{Bucket}?versions"
|
|
||||||
},
|
|
||||||
"input": {
|
|
||||||
"type": "structure",
|
|
||||||
"required": [
|
|
||||||
"Bucket"
|
|
||||||
],
|
|
||||||
"members": {
|
|
||||||
"Bucket": {
|
|
||||||
"location": "uri",
|
|
||||||
"locationName": "Bucket"
|
|
||||||
},
|
|
||||||
"Delimiter": {
|
|
||||||
"location": "querystring",
|
|
||||||
"locationName": "delimiter"
|
|
||||||
},
|
|
||||||
"Marker": {
|
|
||||||
"location": "querystring",
|
|
||||||
"locationName": "marker"
|
|
||||||
},
|
|
||||||
"MaxKeys": {
|
|
||||||
"location": "querystring",
|
|
||||||
"locationName": "max-keys",
|
|
||||||
"type": "integer"
|
|
||||||
},
|
|
||||||
"Prefix": {
|
|
||||||
"location": "querystring",
|
|
||||||
"locationName": "prefix"
|
|
||||||
},
|
|
||||||
"ProjectId": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-project-id"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"output": {
|
|
||||||
"type": "structure",
|
|
||||||
"members": {
|
|
||||||
"IsTruncated": {
|
|
||||||
"type": "boolean"
|
|
||||||
},
|
|
||||||
"Marker": {},
|
|
||||||
"NextMarker": {},
|
|
||||||
"Versions": {
|
|
||||||
"locationName": "Version",
|
|
||||||
"shape": "ContentsShape"
|
|
||||||
},
|
|
||||||
"Name": {},
|
|
||||||
"Prefix": {},
|
|
||||||
"Delimiter": {},
|
|
||||||
"MaxKeys": {
|
|
||||||
"type": "integer"
|
|
||||||
},
|
|
||||||
"CommonPrefixes": {
|
|
||||||
"shape": "CommonPrefixShape"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"PutBucketVersioning": {
|
|
||||||
"http": {
|
|
||||||
"method": "PUT",
|
|
||||||
"requestUri": "/{Bucket}?versioning"
|
|
||||||
},
|
|
||||||
"input": {
|
|
||||||
"type": "structure",
|
|
||||||
"required": [
|
|
||||||
"Bucket",
|
|
||||||
"VersioningConfiguration"
|
|
||||||
],
|
|
||||||
"members": {
|
|
||||||
"Bucket": {
|
|
||||||
"location": "uri",
|
|
||||||
"locationName": "Bucket"
|
|
||||||
},
|
|
||||||
"ContentMD5": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Content-MD5"
|
|
||||||
},
|
|
||||||
"VersioningConfiguration": {
|
|
||||||
"locationName": "VersioningConfiguration",
|
|
||||||
"type": "structure",
|
|
||||||
"members": {
|
|
||||||
"Status": {}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"payload": "VersioningConfiguration"
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"GetBucketVersioning": {
|
|
||||||
"http": {
|
|
||||||
"method": "GET",
|
|
||||||
"requestUri": "/{Bucket}?versioning"
|
|
||||||
},
|
|
||||||
"input": {
|
|
||||||
"type": "structure",
|
|
||||||
"required": [
|
|
||||||
"Bucket"
|
|
||||||
],
|
|
||||||
"members": {
|
|
||||||
"Bucket": {
|
|
||||||
"location": "uri",
|
|
||||||
"locationName": "Bucket"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"output": {
|
|
||||||
"type": "structure",
|
|
||||||
"members": {
|
|
||||||
"Status": {}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"HeadObject": {
|
|
||||||
"http": {
|
|
||||||
"method": "HEAD",
|
|
||||||
"requestUri": "/{Bucket}/{Key+}"
|
|
||||||
},
|
|
||||||
"input": {
|
|
||||||
"type": "structure",
|
|
||||||
"required": [
|
|
||||||
"Bucket",
|
|
||||||
"Key"
|
|
||||||
],
|
|
||||||
"members": {
|
|
||||||
"Date": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Date",
|
|
||||||
"type": "timestamp"
|
|
||||||
},
|
|
||||||
"Bucket": {
|
|
||||||
"location": "uri",
|
|
||||||
"locationName": "Bucket"
|
|
||||||
},
|
|
||||||
"IfMatch": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "If-Match"
|
|
||||||
},
|
|
||||||
"IfModifiedSince": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "If-Modified-Since",
|
|
||||||
"type": "timestamp"
|
|
||||||
},
|
|
||||||
"IfNoneMatch": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "If-None-Match"
|
|
||||||
},
|
|
||||||
"IfUnmodifiedSince": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "If-Unmodified-Since",
|
|
||||||
"type": "timestamp"
|
|
||||||
},
|
|
||||||
"Range": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Range"
|
|
||||||
},
|
|
||||||
"Key": {
|
|
||||||
"location": "uri",
|
|
||||||
"locationName": "Key"
|
|
||||||
},
|
|
||||||
"Range": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Range"
|
|
||||||
},
|
|
||||||
"VersionId": {
|
|
||||||
"location": "querystring",
|
|
||||||
"locationName": "generation"
|
|
||||||
},
|
|
||||||
"ProjectId": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-project-id"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"output": {
|
|
||||||
"type": "structure",
|
|
||||||
"members": {
|
|
||||||
"Date": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Date",
|
|
||||||
"type": "timestamp"
|
|
||||||
},
|
|
||||||
"AcceptRanges": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "accept-ranges"
|
|
||||||
},
|
|
||||||
"Expiration": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-expiration"
|
|
||||||
},
|
|
||||||
"LastModified": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Last-Modified",
|
|
||||||
"type": "timestamp"
|
|
||||||
},
|
|
||||||
"ContentLength": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Content-Length",
|
|
||||||
"type": "long"
|
|
||||||
},
|
|
||||||
"ContentHash": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-hash"
|
|
||||||
},
|
|
||||||
"ETag": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "ETag"
|
|
||||||
},
|
|
||||||
"VersionId": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-generation"
|
|
||||||
},
|
|
||||||
"MetaVersionId": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-metageneration"
|
|
||||||
},
|
|
||||||
"CacheControl": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Cache-Control"
|
|
||||||
},
|
|
||||||
"ContentDisposition": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Content-Disposition"
|
|
||||||
},
|
|
||||||
"ContentEncoding": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Content-Encoding"
|
|
||||||
},
|
|
||||||
"ContentLanguage": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Content-Language"
|
|
||||||
},
|
|
||||||
"ContentType": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Content-Type"
|
|
||||||
},
|
|
||||||
"Expires": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Expires",
|
|
||||||
"type": "timestamp"
|
|
||||||
},
|
|
||||||
"Metadata": {
|
|
||||||
"shape": "MetadataShape",
|
|
||||||
"location": "headers",
|
|
||||||
"locationName": "x-goog-meta-"
|
|
||||||
},
|
|
||||||
"StorageClass": {
|
|
||||||
"location": "headers",
|
|
||||||
"locationName": "x-goog-storage-class"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"PutObjectReq": {
|
|
||||||
"http": {
|
|
||||||
"method": "PUT",
|
|
||||||
"requestUri": "/{Bucket}/{Key+}"
|
|
||||||
},
|
|
||||||
"input": {
|
|
||||||
"type": "structure",
|
|
||||||
"required": [
|
|
||||||
"Bucket",
|
|
||||||
"Key"
|
|
||||||
],
|
|
||||||
"members": {
|
|
||||||
"Date": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Date",
|
|
||||||
"type": "timestamp"
|
|
||||||
},
|
|
||||||
"ACL": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-acl"
|
|
||||||
},
|
|
||||||
"Body": {
|
|
||||||
"streaming": true,
|
|
||||||
"type": "blob"
|
|
||||||
},
|
|
||||||
"Bucket": {
|
|
||||||
"location": "uri",
|
|
||||||
"locationName": "Bucket"
|
|
||||||
},
|
|
||||||
"CacheControl": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Cache-Control"
|
|
||||||
},
|
|
||||||
"ContentDisposition": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Content-Disposition"
|
|
||||||
},
|
|
||||||
"ContentEncoding": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Content-Encoding"
|
|
||||||
},
|
|
||||||
"ContentLanguage": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Content-Language"
|
|
||||||
},
|
|
||||||
"ContentLength": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Content-Length",
|
|
||||||
"type": "long"
|
|
||||||
},
|
|
||||||
"ContentMD5": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Content-MD5"
|
|
||||||
},
|
|
||||||
"ContentType": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Content-Type"
|
|
||||||
},
|
|
||||||
"Expires": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Expires",
|
|
||||||
"type": "timestamp"
|
|
||||||
},
|
|
||||||
"Key": {
|
|
||||||
"location": "uri",
|
|
||||||
"locationName": "Key"
|
|
||||||
},
|
|
||||||
"Metadata": {
|
|
||||||
"shape": "MetadataShape",
|
|
||||||
"location": "headers",
|
|
||||||
"locationName": "x-goog-meta-"
|
|
||||||
},
|
|
||||||
"ProjectId": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-project-id"
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"payload": "Body"
|
|
||||||
},
|
|
||||||
"output": {
|
|
||||||
"type": "structure",
|
|
||||||
"members": {
|
|
||||||
"Expiration": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-expiration"
|
|
||||||
},
|
|
||||||
"ETag": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "ETag"
|
|
||||||
},
|
|
||||||
"ContentHash": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-hash"
|
|
||||||
},
|
|
||||||
"VersionId": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-generation"
|
|
||||||
},
|
|
||||||
"MetaVersionId": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-metageneration"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"GetObject": {
|
|
||||||
"http": {
|
|
||||||
"method": "GET",
|
|
||||||
"requestUri": "/{Bucket}/{Key+}"
|
|
||||||
},
|
|
||||||
"input": {
|
|
||||||
"type": "structure",
|
|
||||||
"required": [
|
|
||||||
"Bucket",
|
|
||||||
"Key"
|
|
||||||
],
|
|
||||||
"members": {
|
|
||||||
"Bucket": {
|
|
||||||
"location": "uri",
|
|
||||||
"locationName": "Bucket"
|
|
||||||
},
|
|
||||||
"IfMatch": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "If-Match"
|
|
||||||
},
|
|
||||||
"IfModifiedSince": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "If-Modified-Since",
|
|
||||||
"type": "timestamp"
|
|
||||||
},
|
|
||||||
"IfNoneMatch": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "If-None-Match"
|
|
||||||
},
|
|
||||||
"IfUnmodifiedSince": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "If-Unmodified-Since",
|
|
||||||
"type": "timestamp"
|
|
||||||
},
|
|
||||||
"Key": {
|
|
||||||
"location": "uri",
|
|
||||||
"locationName": "Key"
|
|
||||||
},
|
|
||||||
"Range": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Range"
|
|
||||||
},
|
|
||||||
"ResponseCacheControl": {
|
|
||||||
"location": "querystring",
|
|
||||||
"locationName": "response-cache-control"
|
|
||||||
},
|
|
||||||
"ResponseContentDisposition": {
|
|
||||||
"location": "querystring",
|
|
||||||
"locationName": "response-content-disposition"
|
|
||||||
},
|
|
||||||
"ResponseContentEncoding": {
|
|
||||||
"location": "querystring",
|
|
||||||
"locationName": "response-content-encoding"
|
|
||||||
},
|
|
||||||
"ResponseContentLanguage": {
|
|
||||||
"location": "querystring",
|
|
||||||
"locationName": "response-content-language"
|
|
||||||
},
|
|
||||||
"ResponseContentType": {
|
|
||||||
"location": "querystring",
|
|
||||||
"locationName": "response-content-type"
|
|
||||||
},
|
|
||||||
"ResponseExpires": {
|
|
||||||
"location": "querystring",
|
|
||||||
"locationName": "response-expires",
|
|
||||||
"type": "timestamp"
|
|
||||||
},
|
|
||||||
"VersionId": {
|
|
||||||
"location": "querystring",
|
|
||||||
"locationName": "generation"
|
|
||||||
},
|
|
||||||
"ProjectId": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-project-id"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"output": {
|
|
||||||
"type": "structure",
|
|
||||||
"members": {
|
|
||||||
"Body": {
|
|
||||||
"streaming": true,
|
|
||||||
"type": "blob"
|
|
||||||
},
|
|
||||||
"AcceptRanges": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "accept-ranges"
|
|
||||||
},
|
|
||||||
"Expiration": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-expiration"
|
|
||||||
},
|
|
||||||
"LastModified": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Last-Modified",
|
|
||||||
"type": "timestamp"
|
|
||||||
},
|
|
||||||
"ContentLength": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Content-Length",
|
|
||||||
"type": "long"
|
|
||||||
},
|
|
||||||
"ETag": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "ETag"
|
|
||||||
},
|
|
||||||
"VersionId": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-generation"
|
|
||||||
},
|
|
||||||
"MetaVersionId": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-metageneration"
|
|
||||||
},
|
|
||||||
"CacheControl": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Cache-Control"
|
|
||||||
},
|
|
||||||
"ContentDisposition": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Content-Disposition"
|
|
||||||
},
|
|
||||||
"ContentEncoding": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Content-Encoding"
|
|
||||||
},
|
|
||||||
"ContentLanguage": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Content-Language"
|
|
||||||
},
|
|
||||||
"ContentRange": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Content-Range"
|
|
||||||
},
|
|
||||||
"ContentType": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Content-Type"
|
|
||||||
},
|
|
||||||
"ContentHash": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-hash"
|
|
||||||
},
|
|
||||||
"Expires": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Expires",
|
|
||||||
"type": "timestamp"
|
|
||||||
},
|
|
||||||
"WebsiteRedirectLocation": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-website-redirect-location"
|
|
||||||
},
|
|
||||||
"ServerSideEncryption": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-server-side-encryption"
|
|
||||||
},
|
|
||||||
"Metadata": {
|
|
||||||
"shape": "MetadataShape",
|
|
||||||
"location": "headers",
|
|
||||||
"locationName": "x-goog-meta-"
|
|
||||||
},
|
|
||||||
"StorageClass": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-storage-class"
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"payload": "Body"
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"DeleteObject": {
|
|
||||||
"http": {
|
|
||||||
"method": "DELETE",
|
|
||||||
"requestUri": "/{Bucket}/{Key+}"
|
|
||||||
},
|
|
||||||
"input": {
|
|
||||||
"type": "structure",
|
|
||||||
"required": [
|
|
||||||
"Bucket",
|
|
||||||
"Key"
|
|
||||||
],
|
|
||||||
"members": {
|
|
||||||
"Bucket": {
|
|
||||||
"location": "uri",
|
|
||||||
"locationName": "Bucket"
|
|
||||||
},
|
|
||||||
"Key": {
|
|
||||||
"location": "uri",
|
|
||||||
"locationName": "Key"
|
|
||||||
},
|
|
||||||
"VersionId": {
|
|
||||||
"location": "querystring",
|
|
||||||
"locationName": "generation"
|
|
||||||
},
|
|
||||||
"ProjectId": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-project-id"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"output": {
|
|
||||||
"type": "structure",
|
|
||||||
"members": {
|
|
||||||
"VersionId": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-generation"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"ComposeObject": {
|
|
||||||
"http": {
|
|
||||||
"method": "PUT",
|
|
||||||
"requestUri": "/{Bucket}/{Key+}?compose"
|
|
||||||
},
|
|
||||||
"input": {
|
|
||||||
"type": "structure",
|
|
||||||
"required": [
|
|
||||||
"Bucket",
|
|
||||||
"Key"
|
|
||||||
],
|
|
||||||
"members": {
|
|
||||||
"Bucket": {
|
|
||||||
"location": "uri",
|
|
||||||
"locationName": "Bucket"
|
|
||||||
},
|
|
||||||
"Source": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-copy-source"
|
|
||||||
},
|
|
||||||
"Key": {
|
|
||||||
"location": "uri",
|
|
||||||
"locationName": "Key"
|
|
||||||
},
|
|
||||||
"MetadataDirective": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-metadata-directive"
|
|
||||||
},
|
|
||||||
"ContentDisposition": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Content-Disposition"
|
|
||||||
},
|
|
||||||
"Content-Encoding": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Content-Encoding"
|
|
||||||
},
|
|
||||||
"MultipartUpload": {
|
|
||||||
"locationName": "ComposeRequest",
|
|
||||||
"type": "structure",
|
|
||||||
"members": {
|
|
||||||
"Parts": {
|
|
||||||
"locationName": "Component",
|
|
||||||
"type": "list",
|
|
||||||
"member": {
|
|
||||||
"type": "structure",
|
|
||||||
"members": {
|
|
||||||
"PartName": {
|
|
||||||
"locationName": "Name"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"flattened": true
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"Metadata": {
|
|
||||||
"shape": "MetadataShape",
|
|
||||||
"location": "headers",
|
|
||||||
"locationName": "x-goog-meta-"
|
|
||||||
},
|
|
||||||
"ProjectId": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-project-id"
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"payload": "MultipartUpload"
|
|
||||||
},
|
|
||||||
"output": {
|
|
||||||
"type": "structure",
|
|
||||||
"members": {
|
|
||||||
"Expiration": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-expiration"
|
|
||||||
},
|
|
||||||
"ETag": {},
|
|
||||||
"VersioId": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-generation"
|
|
||||||
},
|
|
||||||
"MetaVersionId": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-metageneration"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"CopyObject": {
|
|
||||||
"http": {
|
|
||||||
"method": "PUT",
|
|
||||||
"requestUri": "/{Bucket}/{Key+}"
|
|
||||||
},
|
|
||||||
"input": {
|
|
||||||
"type": "structure",
|
|
||||||
"required": [
|
|
||||||
"Bucket",
|
|
||||||
"CopySource",
|
|
||||||
"Key"
|
|
||||||
],
|
|
||||||
"members": {
|
|
||||||
"ACL": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-acl"
|
|
||||||
},
|
|
||||||
"Bucket": {
|
|
||||||
"location": "uri",
|
|
||||||
"locationName": "Bucket"
|
|
||||||
},
|
|
||||||
"CacheControl": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Cache-Control"
|
|
||||||
},
|
|
||||||
"ContentDisposition": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Content-Disposition"
|
|
||||||
},
|
|
||||||
"ContentEncoding": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Content-Encoding"
|
|
||||||
},
|
|
||||||
"ContentLanguage": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Content-Language"
|
|
||||||
},
|
|
||||||
"ContentType": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Content-Type"
|
|
||||||
},
|
|
||||||
"CopySource": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-copy-source"
|
|
||||||
},
|
|
||||||
"CopySourceIfMatch": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-copy-source-if-match"
|
|
||||||
},
|
|
||||||
"CopySourceIfModifiedSince": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-copy-source-if-modified-since"
|
|
||||||
},
|
|
||||||
"CopySourceIfNoneMatch": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-copy-source-if-none-match"
|
|
||||||
},
|
|
||||||
"CopySourceIfUnmodifiedSince": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-copy-source-if-unmodified-since",
|
|
||||||
"type": "timestamp"
|
|
||||||
},
|
|
||||||
"Expires": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "Expires",
|
|
||||||
"type": "timestamp"
|
|
||||||
},
|
|
||||||
"Key": {
|
|
||||||
"location": "uri",
|
|
||||||
"locationName": "Key"
|
|
||||||
},
|
|
||||||
"Metadata": {
|
|
||||||
"shape": "MetadataShape",
|
|
||||||
"location": "headers",
|
|
||||||
"locationName": "x-goog-meta-"
|
|
||||||
},
|
|
||||||
"MetadataDirective": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-metadata-directive"
|
|
||||||
},
|
|
||||||
"ProjectId": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-project-id"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"output": {
|
|
||||||
"type": "structure",
|
|
||||||
"members": {
|
|
||||||
"CopyObjectResult": {
|
|
||||||
"type": "structure",
|
|
||||||
"members": {
|
|
||||||
"ETag": {},
|
|
||||||
"LastModified": {
|
|
||||||
"type": "timestamp"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"Expiration": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-expiration"
|
|
||||||
},
|
|
||||||
"ContentHash": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-hash"
|
|
||||||
},
|
|
||||||
"VersionId": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-generation"
|
|
||||||
},
|
|
||||||
"MetaVersionId": {
|
|
||||||
"location": "header",
|
|
||||||
"locationName": "x-goog-metageneration"
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"payload": "CopyObjectResult"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"shapes": {
|
|
||||||
"MetadataShape": {
|
|
||||||
"type": "map",
|
|
||||||
"key": {},
|
|
||||||
"value": {}
|
|
||||||
},
|
|
||||||
"OwnerShape": {
|
|
||||||
"locationName": "Owner",
|
|
||||||
"type": "structure",
|
|
||||||
"members": {
|
|
||||||
"ID": {},
|
|
||||||
"Name": {}
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"ContentsShape": {
|
|
||||||
"type": "list",
|
|
||||||
"member": {
|
|
||||||
"type": "structure",
|
|
||||||
"members": {
|
|
||||||
"Key": {},
|
|
||||||
"LastModified": {
|
|
||||||
"type": "timestamp"
|
|
||||||
},
|
|
||||||
"ETag": {},
|
|
||||||
"Size": {
|
|
||||||
"type": "integer"
|
|
||||||
},
|
|
||||||
"StorageClass": {},
|
|
||||||
"Owner": {
|
|
||||||
"shape": "OwnerShape"
|
|
||||||
},
|
|
||||||
"VersionId": {
|
|
||||||
"locationName": "Generation"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"flattened": true
|
|
||||||
},
|
|
||||||
"CommonPrefixShape": {
|
|
||||||
"type": "list",
|
|
||||||
"member": {
|
|
||||||
"type": "structure",
|
|
||||||
"members": {
|
|
||||||
"Prefix": {}
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"flattened": true
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
|
@ -1,5 +0,0 @@
|
||||||
module.exports = {
|
|
||||||
GCP: require('./GcpService'),
|
|
||||||
GcpSigner: require('./GcpSigner'),
|
|
||||||
GcpUtils: require('./GcpUtils'),
|
|
||||||
};
|
|
|
@ -1,302 +0,0 @@
|
||||||
const async = require('async');
|
|
||||||
const { errors, s3middleware } = require('arsenal');
|
|
||||||
const MD5Sum = s3middleware.MD5Sum;
|
|
||||||
|
|
||||||
const { GCP, GcpUtils } = require('./GCP');
|
|
||||||
const { createMpuKey } = GcpUtils;
|
|
||||||
const AwsClient = require('./AwsClient');
|
|
||||||
const { prepareStream } = require('../../api/apiUtils/object/prepareStream');
|
|
||||||
const { logHelper, removeQuotes } = require('./utils');
|
|
||||||
const { config } = require('../../Config');
|
|
||||||
|
|
||||||
const missingVerIdInternalError = errors.InternalError.customizeDescription(
|
|
||||||
'Invalid state. Please ensure versioning is enabled ' +
|
|
||||||
'in GCP for the location constraint and try again.');
|
|
||||||
|
|
||||||
/**
|
|
||||||
* Class representing a Google Cloud Storage backend object
|
|
||||||
* @extends AwsClient
|
|
||||||
*/
|
|
||||||
class GcpClient extends AwsClient {
|
|
||||||
/**
|
|
||||||
* constructor - creates a Gcp backend client object (inherits )
|
|
||||||
* @param {object} config - configuration object for Gcp Backend up
|
|
||||||
* @param {object} config.s3params - S3 configuration
|
|
||||||
* @param {string} config.bucketName - GCP bucket name
|
|
||||||
* @param {string} config.mpuBucket - GCP mpu bucket name
|
|
||||||
* @param {boolean} config.bucketMatch - bucket match flag
|
|
||||||
* @param {object} config.authParams - GCP service credentials
|
|
||||||
* @param {string} config.dataStoreName - locationConstraint name
|
|
||||||
* @param {booblean} config.serverSideEncryption - server side encryption
|
|
||||||
* flag
|
|
||||||
* @return {object} - returns a GcpClient object
|
|
||||||
*/
|
|
||||||
constructor(config) {
|
|
||||||
super(config);
|
|
||||||
this.clientType = 'gcp';
|
|
||||||
this.type = 'GCP';
|
|
||||||
this._gcpBucketName = config.bucketName;
|
|
||||||
this._mpuBucketName = config.mpuBucket;
|
|
||||||
this._createGcpKey = this._createAwsKey;
|
|
||||||
this._gcpParams = Object.assign(this._s3Params, {
|
|
||||||
mainBucket: this._gcpBucketName,
|
|
||||||
mpuBucket: this._mpuBucketName,
|
|
||||||
});
|
|
||||||
this._client = new GCP(this._gcpParams);
|
|
||||||
// reassign inherited list parts method from AWS to trigger
|
|
||||||
// listing using S3 metadata part list instead of request to GCP
|
|
||||||
this.listParts = undefined;
|
|
||||||
}
|
|
||||||
|
|
||||||
toObjectGetInfo(objectKey, bucketName) {
|
|
||||||
return {
|
|
||||||
key: this._createGcpKey(bucketName, objectKey, this._bucketMatch),
|
|
||||||
dataStoreName: this._dataStoreName,
|
|
||||||
};
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* healthcheck - the gcp health requires checking multiple buckets:
|
|
||||||
* main and mpu buckets
|
|
||||||
* @param {string} location - location name
|
|
||||||
* @param {function} callback - callback function to call with the bucket
|
|
||||||
* statuses
|
|
||||||
* @return {undefined}
|
|
||||||
*/
|
|
||||||
healthcheck(location, callback) {
|
|
||||||
const checkBucketHealth = (bucket, cb) => {
|
|
||||||
let bucketResp;
|
|
||||||
this._client.headBucket({ Bucket: bucket }, err => {
|
|
||||||
if (err) {
|
|
||||||
bucketResp = {
|
|
||||||
gcpBucket: bucket,
|
|
||||||
error: err };
|
|
||||||
} else {
|
|
||||||
bucketResp = {
|
|
||||||
gcpBucket: bucket,
|
|
||||||
message: 'Congrats! You own the bucket',
|
|
||||||
};
|
|
||||||
}
|
|
||||||
return cb(null, bucketResp);
|
|
||||||
});
|
|
||||||
};
|
|
||||||
const gcpResp = {};
|
|
||||||
async.parallel({
|
|
||||||
main: done => checkBucketHealth(this._gcpBucketName, done),
|
|
||||||
mpu: done => checkBucketHealth(this._mpuBucketName, done),
|
|
||||||
}, (err, result) => {
|
|
||||||
if (err) {
|
|
||||||
return callback(errors.InternalFailure
|
|
||||||
.customizeDescription('Unable to perform health check'));
|
|
||||||
}
|
|
||||||
|
|
||||||
gcpResp[location] = result.main.error || result.mpu.error ?
|
|
||||||
{ error: true, external: true } : {};
|
|
||||||
Object.assign(gcpResp[location], result);
|
|
||||||
return callback(null, gcpResp);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
createMPU(key, metaHeaders, bucketName, websiteRedirectHeader, contentType,
|
|
||||||
cacheControl, contentDisposition, contentEncoding, log, callback) {
|
|
||||||
const metaHeadersTrimmed = {};
|
|
||||||
Object.keys(metaHeaders).forEach(header => {
|
|
||||||
if (header.startsWith('x-amz-meta-')) {
|
|
||||||
const headerKey = header.substring(11);
|
|
||||||
metaHeadersTrimmed[headerKey] = metaHeaders[header];
|
|
||||||
}
|
|
||||||
});
|
|
||||||
const gcpKey = this._createGcpKey(bucketName, key, this._bucketMatch);
|
|
||||||
const params = {
|
|
||||||
Bucket: this._mpuBucketName,
|
|
||||||
Key: gcpKey,
|
|
||||||
Metadata: metaHeadersTrimmed,
|
|
||||||
ContentType: contentType,
|
|
||||||
CacheControl: cacheControl,
|
|
||||||
ContentDisposition: contentDisposition,
|
|
||||||
ContentEncoding: contentEncoding,
|
|
||||||
};
|
|
||||||
return this._client.createMultipartUpload(params, (err, mpuResObj) => {
|
|
||||||
if (err) {
|
|
||||||
logHelper(log, 'error', 'err from data backend',
|
|
||||||
err, this._dataStoreName, this.clientType);
|
|
||||||
return callback(errors.ServiceUnavailable
|
|
||||||
.customizeDescription('Error returned from ' +
|
|
||||||
`GCP: ${err.message}`)
|
|
||||||
);
|
|
||||||
}
|
|
||||||
return callback(null, mpuResObj);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
completeMPU(jsonList, mdInfo, key, uploadId, bucketName, log, callback) {
|
|
||||||
const gcpKey = this._createGcpKey(bucketName, key, this._bucketMatch);
|
|
||||||
const partArray = [];
|
|
||||||
const partList = jsonList.Part;
|
|
||||||
for (let i = 0; i < partList.length; ++i) {
|
|
||||||
const partObj = partList[i];
|
|
||||||
if (!partObj.PartNumber || !partObj.ETag) {
|
|
||||||
return callback(errors.MalformedXML);
|
|
||||||
}
|
|
||||||
const number = partObj.PartNumber[0];
|
|
||||||
// check if the partNumber is an actual number throw an error
|
|
||||||
// otherwise
|
|
||||||
if (isNaN(number)) {
|
|
||||||
return callback(errors.MalformedXML);
|
|
||||||
}
|
|
||||||
const partNumber = parseInt(number, 10);
|
|
||||||
const partParams = {
|
|
||||||
PartName: createMpuKey(gcpKey, uploadId, partNumber),
|
|
||||||
PartNumber: partNumber,
|
|
||||||
ETag: partObj.ETag[0],
|
|
||||||
};
|
|
||||||
partArray.push(partParams);
|
|
||||||
}
|
|
||||||
const mpuParams = {
|
|
||||||
Bucket: this._gcpBucketName,
|
|
||||||
MPU: this._mpuBucketName,
|
|
||||||
Key: gcpKey,
|
|
||||||
UploadId: uploadId,
|
|
||||||
MultipartUpload: { Parts: partArray },
|
|
||||||
};
|
|
||||||
const completeObjData = { key: gcpKey };
|
|
||||||
return this._client.completeMultipartUpload(mpuParams,
|
|
||||||
(err, completeMpuRes) => {
|
|
||||||
if (err) {
|
|
||||||
logHelper(log, 'error', 'err from data backend on ' +
|
|
||||||
'completeMPU', err, this._dataStoreName, this.clientType);
|
|
||||||
return callback(errors.ServiceUnavailable
|
|
||||||
.customizeDescription('Error returned from ' +
|
|
||||||
`GCP: ${err.message}`)
|
|
||||||
);
|
|
||||||
}
|
|
||||||
if (!completeMpuRes.VersionId) {
|
|
||||||
logHelper(log, 'error', 'missing version id for data ' +
|
|
||||||
'backend object', missingVerIdInternalError,
|
|
||||||
this._dataStoreName, this.clientType);
|
|
||||||
return callback(missingVerIdInternalError);
|
|
||||||
}
|
|
||||||
// remove quotes from eTag because they're added later
|
|
||||||
completeObjData.eTag = removeQuotes(completeMpuRes.ETag);
|
|
||||||
completeObjData.dataStoreVersionId = completeMpuRes.VersionId;
|
|
||||||
completeObjData.contentLength =
|
|
||||||
Number.parseInt(completeMpuRes.ContentLength, 10);
|
|
||||||
return callback(null, completeObjData);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
uploadPart(request, streamingV4Params, stream, size, key, uploadId,
|
|
||||||
partNumber, bucketName, log, callback) {
|
|
||||||
let hashedStream = stream;
|
|
||||||
if (request) {
|
|
||||||
const partStream = prepareStream(request, streamingV4Params,
|
|
||||||
log, callback);
|
|
||||||
hashedStream = new MD5Sum();
|
|
||||||
partStream.pipe(hashedStream);
|
|
||||||
}
|
|
||||||
|
|
||||||
const gcpKey = this._createGcpKey(bucketName, key, this._bucketMatch);
|
|
||||||
const params = {
|
|
||||||
Bucket: this._mpuBucketName,
|
|
||||||
Key: gcpKey,
|
|
||||||
UploadId: uploadId,
|
|
||||||
Body: hashedStream,
|
|
||||||
ContentLength: size,
|
|
||||||
PartNumber: partNumber };
|
|
||||||
return this._client.uploadPart(params, (err, partResObj) => {
|
|
||||||
if (err) {
|
|
||||||
logHelper(log, 'error', 'err from data backend ' +
|
|
||||||
'on uploadPart', err, this._dataStoreName, this.clientType);
|
|
||||||
return callback(errors.ServiceUnavailable
|
|
||||||
.customizeDescription('Error returned from ' +
|
|
||||||
`GCP: ${err.message}`)
|
|
||||||
);
|
|
||||||
}
|
|
||||||
// remove quotes from eTag because they're added later
|
|
||||||
const noQuotesETag = removeQuotes(partResObj.ETag);
|
|
||||||
const dataRetrievalInfo = {
|
|
||||||
key: gcpKey,
|
|
||||||
dataStoreType: 'gcp',
|
|
||||||
dataStoreName: this._dataStoreName,
|
|
||||||
dataStoreETag: noQuotesETag,
|
|
||||||
};
|
|
||||||
return callback(null, dataRetrievalInfo);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
uploadPartCopy(request, gcpSourceKey, sourceLocationConstraintName, log,
|
|
||||||
callback) {
|
|
||||||
const destBucketName = request.bucketName;
|
|
||||||
const destObjectKey = request.objectKey;
|
|
||||||
const destGcpKey = this._createGcpKey(destBucketName, destObjectKey,
|
|
||||||
this._bucketMatch);
|
|
||||||
|
|
||||||
const sourceGcpBucketName =
|
|
||||||
config.getGcpBucketNames(sourceLocationConstraintName).bucketName;
|
|
||||||
|
|
||||||
const uploadId = request.query.uploadId;
|
|
||||||
const partNumber = request.query.partNumber;
|
|
||||||
const copySourceRange = request.headers['x-amz-copy-source-range'];
|
|
||||||
|
|
||||||
if (copySourceRange) {
|
|
||||||
return callback(errors.NotImplemented
|
|
||||||
.customizeDescription('Error returned from ' +
|
|
||||||
`${this.clientType}: copySourceRange not implemented`)
|
|
||||||
);
|
|
||||||
}
|
|
||||||
|
|
||||||
const params = {
|
|
||||||
Bucket: this._mpuBucketName,
|
|
||||||
CopySource: `${sourceGcpBucketName}/${gcpSourceKey}`,
|
|
||||||
Key: destGcpKey,
|
|
||||||
UploadId: uploadId,
|
|
||||||
PartNumber: partNumber,
|
|
||||||
};
|
|
||||||
return this._client.uploadPartCopy(params, (err, res) => {
|
|
||||||
if (err) {
|
|
||||||
if (err.code === 'AccesssDenied') {
|
|
||||||
logHelper(log, 'error', 'Unable to access ' +
|
|
||||||
`${sourceGcpBucketName} GCP bucket`, err,
|
|
||||||
this._dataStoreName, this.clientType);
|
|
||||||
return callback(errors.AccessDenied
|
|
||||||
.customizeDescription('Error: Unable to access ' +
|
|
||||||
`${sourceGcpBucketName} GCP bucket`)
|
|
||||||
);
|
|
||||||
}
|
|
||||||
logHelper(log, 'error', 'error from data backend on ' +
|
|
||||||
'uploadPartCopy', err, this._dataStoreName);
|
|
||||||
return callback(errors.ServiceUnavailable
|
|
||||||
.customizeDescription('Error returned from ' +
|
|
||||||
`GCP: ${err.message}`)
|
|
||||||
);
|
|
||||||
}
|
|
||||||
// remove quotes from eTag because they're added later
|
|
||||||
const eTag = removeQuotes(res.CopyObjectResult.ETag);
|
|
||||||
return callback(null, eTag);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
abortMPU(key, uploadId, bucketName, log, callback) {
|
|
||||||
const gcpKey = this._createGcpKey(bucketName, key, this._bucketMatch);
|
|
||||||
const getParams = {
|
|
||||||
Bucket: this._gcpBucketName,
|
|
||||||
MPU: this._mpuBucketName,
|
|
||||||
Key: gcpKey,
|
|
||||||
UploadId: uploadId,
|
|
||||||
};
|
|
||||||
return this._client.abortMultipartUpload(getParams, err => {
|
|
||||||
if (err) {
|
|
||||||
logHelper(log, 'error', 'err from data backend ' +
|
|
||||||
'on abortMPU', err, this._dataStoreName, this.clientType);
|
|
||||||
return callback(errors.ServiceUnavailable
|
|
||||||
.customizeDescription('Error returned from ' +
|
|
||||||
`GCP: ${err.message}`)
|
|
||||||
);
|
|
||||||
}
|
|
||||||
return callback();
|
|
||||||
});
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
module.exports = GcpClient;
|
|
||||||
|
|
|
@ -1,149 +0,0 @@
|
||||||
const arsenal = require('arsenal');
|
|
||||||
const errors = arsenal.errors;
|
|
||||||
const createLogger = require('../multipleBackendLogger');
|
|
||||||
const { logHelper } = require('./utils');
|
|
||||||
|
|
||||||
class PfsClient {
|
|
||||||
constructor(config) {
|
|
||||||
const { host, port } = config.endpoint;
|
|
||||||
|
|
||||||
this.clientType = 'pfs';
|
|
||||||
this._bucketMatch = config.bucketMatch;
|
|
||||||
this._dataStoreName = config.dataStoreName;
|
|
||||||
this._restClient = new arsenal.network.rest.RESTClient({
|
|
||||||
host,
|
|
||||||
port,
|
|
||||||
isPassthrough: true,
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
setup(cb) {
|
|
||||||
return cb();
|
|
||||||
}
|
|
||||||
|
|
||||||
_createFsKey(requestBucketName, requestObjectKey, bucketMatch) {
|
|
||||||
if (bucketMatch) {
|
|
||||||
return requestObjectKey;
|
|
||||||
}
|
|
||||||
return `${requestBucketName}/${requestObjectKey}`;
|
|
||||||
}
|
|
||||||
|
|
||||||
toObjectGetInfo(objectKey, bucketName) {
|
|
||||||
return {
|
|
||||||
key: this._createFsKey(bucketName, objectKey, this._bucketMatch),
|
|
||||||
dataStoreName: this._dataStoreName,
|
|
||||||
};
|
|
||||||
}
|
|
||||||
|
|
||||||
put(stream, size, keyContext, reqUids, callback) {
|
|
||||||
const log = createLogger(reqUids);
|
|
||||||
if (keyContext.metaHeaders['x-amz-meta-mdonly'] === 'true') {
|
|
||||||
const b64 = keyContext.metaHeaders['x-amz-meta-md5chksum'];
|
|
||||||
let md5 = null;
|
|
||||||
if (b64) {
|
|
||||||
md5 = new Buffer(b64, 'base64').toString('hex');
|
|
||||||
}
|
|
||||||
return callback(null, keyContext.objectKey, '',
|
|
||||||
keyContext.metaHeaders['x-amz-meta-size'],
|
|
||||||
md5
|
|
||||||
);
|
|
||||||
}
|
|
||||||
logHelper(log, 'error', 'Not implemented', errors.NotImplemented,
|
|
||||||
this._dataStoreName, this.clientType);
|
|
||||||
return callback(errors.NotImplemented);
|
|
||||||
}
|
|
||||||
|
|
||||||
get(objectGetInfo, range, reqUids, callback) {
|
|
||||||
const log = createLogger(reqUids);
|
|
||||||
this._restClient.get(objectGetInfo.key, range, reqUids, (err, rs) => {
|
|
||||||
if (err) {
|
|
||||||
logHelper(log, 'error', 'err from data backend', err,
|
|
||||||
this._dataStoreName, this.clientType);
|
|
||||||
return callback(err, null);
|
|
||||||
}
|
|
||||||
return callback(null, rs);
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
delete(objectGetInfo, reqUids, callback) {
|
|
||||||
const log = createLogger(reqUids);
|
|
||||||
const key = typeof objectGetInfo === 'string' ? objectGetInfo :
|
|
||||||
objectGetInfo.key;
|
|
||||||
this._restClient.delete(key, reqUids, err => {
|
|
||||||
if (err) {
|
|
||||||
logHelper(log, 'error', 'err from data backend', err,
|
|
||||||
this._dataStoreName, this.clientType);
|
|
||||||
return callback(err);
|
|
||||||
}
|
|
||||||
return callback();
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
// TODO: Implement a healthcheck
|
|
||||||
healthcheck(location, callback) {
|
|
||||||
const fsResp = {};
|
|
||||||
return callback(null, fsResp);
|
|
||||||
}
|
|
||||||
|
|
||||||
createMPU(key, metaHeaders, bucketName, websiteRedirectHeader, contentType,
|
|
||||||
cacheControl, contentDisposition, contentEncoding, log, callback) {
|
|
||||||
logHelper(log, 'error', 'Not implemented', errors.NotImplemented,
|
|
||||||
this._dataStoreName, this.clientType);
|
|
||||||
return callback(errors.NotImplemented);
|
|
||||||
}
|
|
||||||
|
|
||||||
uploadPart(request, streamingV4Params, stream, size, key, uploadId,
|
|
||||||
partNumber, bucketName, log, callback) {
|
|
||||||
logHelper(log, 'error', 'Not implemented', errors.NotImplemented,
|
|
||||||
this._dataStoreName, this.clientType);
|
|
||||||
return callback(errors.NotImplemented);
|
|
||||||
}
|
|
||||||
|
|
||||||
listParts(key, uploadId, bucketName, partNumberMarker, maxParts, log,
|
|
||||||
callback) {
|
|
||||||
logHelper(log, 'error', 'Not implemented', errors.NotImplemented,
|
|
||||||
this._dataStoreName, this.clientType);
|
|
||||||
return callback(errors.NotImplemented);
|
|
||||||
}
|
|
||||||
|
|
||||||
completeMPU(jsonList, mdInfo, key, uploadId, bucketName, log, callback) {
|
|
||||||
logHelper(log, 'error', 'Not implemented', errors.NotImplemented,
|
|
||||||
this._dataStoreName, this.clientType);
|
|
||||||
return callback(errors.NotImplemented);
|
|
||||||
}
|
|
||||||
|
|
||||||
abortMPU(key, uploadId, bucketName, log, callback) {
|
|
||||||
logHelper(log, 'error', 'Not implemented', errors.NotImplemented,
|
|
||||||
this._dataStoreName, this.clientType);
|
|
||||||
return callback(errors.NotImplemented);
|
|
||||||
}
|
|
||||||
|
|
||||||
objectPutTagging(key, bucket, objectMD, log, callback) {
|
|
||||||
logHelper(log, 'error', 'Not implemented', errors.NotImplemented,
|
|
||||||
this._dataStoreName, this.clientType);
|
|
||||||
return callback(errors.NotImplemented);
|
|
||||||
}
|
|
||||||
|
|
||||||
objectDeleteTagging(key, bucket, objectMD, log, callback) {
|
|
||||||
logHelper(log, 'error', 'Not implemented', errors.NotImplemented,
|
|
||||||
this._dataStoreName, this.clientType);
|
|
||||||
return callback(errors.NotImplemented);
|
|
||||||
}
|
|
||||||
|
|
||||||
copyObject(request, destLocationConstraintName, sourceKey,
|
|
||||||
sourceLocationConstraintName, storeMetadataParams, log, callback) {
|
|
||||||
logHelper(log, 'error', 'Not implemented', errors.NotImplemented,
|
|
||||||
this._dataStoreName, this.clientType);
|
|
||||||
return callback(errors.NotImplemented);
|
|
||||||
}
|
|
||||||
|
|
||||||
uploadPartCopy(request, awsSourceKey, sourceLocationConstraintName,
|
|
||||||
log, callback) {
|
|
||||||
logHelper(log, 'error', 'Not implemented', errors.NotImplemented,
|
|
||||||
this._dataStoreName, this.clientType);
|
|
||||||
return callback(errors.NotImplemented);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
module.exports = PfsClient;
|
|
||||||
|
|
|
@ -1,164 +0,0 @@
|
||||||
const async = require('async');
|
|
||||||
const constants = require('../../../constants');
|
|
||||||
const { config } = require('../../../lib/Config');
|
|
||||||
|
|
||||||
/* eslint-disable camelcase */
|
|
||||||
const backendHealth = {
|
|
||||||
aws_s3: {
|
|
||||||
response: undefined,
|
|
||||||
time: 0,
|
|
||||||
},
|
|
||||||
azure: {
|
|
||||||
response: undefined,
|
|
||||||
time: 0,
|
|
||||||
},
|
|
||||||
gcp: {
|
|
||||||
reponse: undefined,
|
|
||||||
time: 0,
|
|
||||||
},
|
|
||||||
};
|
|
||||||
/* eslint-enable camelcase */
|
|
||||||
|
|
||||||
const utils = {
|
|
||||||
logHelper(log, level, description, error, dataStoreName, backendType) {
|
|
||||||
const { message, name, requestId, extendedRequestId } = error;
|
|
||||||
log[level](description, {
|
|
||||||
error: message,
|
|
||||||
errorName: name,
|
|
||||||
dataStoreName,
|
|
||||||
backendType,
|
|
||||||
extRequestId: requestId,
|
|
||||||
extExtendedRequestId: extendedRequestId,
|
|
||||||
});
|
|
||||||
},
|
|
||||||
// take off the 'x-amz-meta-'
|
|
||||||
trimXMetaPrefix(obj) {
|
|
||||||
const newObj = {};
|
|
||||||
const metaObj = obj || {};
|
|
||||||
Object.keys(metaObj).forEach(key => {
|
|
||||||
const newKey = key.substring(11);
|
|
||||||
newObj[newKey] = metaObj[key];
|
|
||||||
});
|
|
||||||
return newObj;
|
|
||||||
},
|
|
||||||
removeQuotes(word) {
|
|
||||||
return word.slice(1, -1);
|
|
||||||
},
|
|
||||||
skipMpuPartProcessing(completeObjData) {
|
|
||||||
const backendType = completeObjData.dataStoreType;
|
|
||||||
if (constants.mpuMDStoredExternallyBackend[backendType]) {
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
return false;
|
|
||||||
},
|
|
||||||
|
|
||||||
/**
|
|
||||||
* checkAzureBackendMatch - checks that the external backend location for
|
|
||||||
* two data objects is the same and is Azure
|
|
||||||
* @param {array} objectDataOne - data of first object to compare
|
|
||||||
* @param {object} objectDataTwo - data of second object to compare
|
|
||||||
* @return {boolean} - true if both data backends are Azure, false if not
|
|
||||||
*/
|
|
||||||
checkAzureBackendMatch(objectDataOne, objectDataTwo) {
|
|
||||||
if (objectDataOne.dataStoreType === 'azure' &&
|
|
||||||
objectDataTwo.dataStoreType === 'azure') {
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
return false;
|
|
||||||
},
|
|
||||||
|
|
||||||
/**
|
|
||||||
* externalBackendCopy - Server side copy should only be allowed:
|
|
||||||
* 1) if source object and destination object are both on aws, both
|
|
||||||
* on azure, or both on gcp
|
|
||||||
* 2) if azure to azure, must be the same storage account since Azure
|
|
||||||
* copy outside of an account is async
|
|
||||||
* 3) if the source bucket is not an encrypted bucket and the
|
|
||||||
* destination bucket is not an encrypted bucket (unless the copy
|
|
||||||
* is all within the same bucket).
|
|
||||||
* @param {string} locationConstraintSrc - location constraint of the source
|
|
||||||
* @param {string} locationConstraintDest - location constraint of the
|
|
||||||
* destination
|
|
||||||
* @param {object} sourceBucketMD - metadata of the source bucket
|
|
||||||
* @param {object} destBucketMD - metadata of the destination bucket
|
|
||||||
* @return {boolean} - true if copying object from one
|
|
||||||
* externalbackend to the same external backend and for Azure if it is the
|
|
||||||
* same account since Azure copy outside of an account is async
|
|
||||||
*/
|
|
||||||
externalBackendCopy(locationConstraintSrc, locationConstraintDest,
|
|
||||||
sourceBucketMD, destBucketMD) {
|
|
||||||
const sourceBucketName = sourceBucketMD.getName();
|
|
||||||
const destBucketName = destBucketMD.getName();
|
|
||||||
const isSameBucket = sourceBucketName === destBucketName;
|
|
||||||
const bucketsNotEncrypted = destBucketMD.getServerSideEncryption()
|
|
||||||
=== sourceBucketMD.getServerSideEncryption() &&
|
|
||||||
sourceBucketMD.getServerSideEncryption() === null;
|
|
||||||
const sourceLocationConstraintType =
|
|
||||||
config.getLocationConstraintType(locationConstraintSrc);
|
|
||||||
const locationTypeMatch =
|
|
||||||
config.getLocationConstraintType(locationConstraintSrc) ===
|
|
||||||
config.getLocationConstraintType(locationConstraintDest);
|
|
||||||
return locationTypeMatch && (isSameBucket || bucketsNotEncrypted) &&
|
|
||||||
(sourceLocationConstraintType === 'aws_s3' ||
|
|
||||||
sourceLocationConstraintType === 'gcp' ||
|
|
||||||
(sourceLocationConstraintType === 'azure' &&
|
|
||||||
config.isSameAzureAccount(locationConstraintSrc,
|
|
||||||
locationConstraintDest)));
|
|
||||||
},
|
|
||||||
|
|
||||||
checkExternalBackend(clients, locations, type, flightCheckOnStartUp,
|
|
||||||
externalBackendHealthCheckInterval, cb) {
|
|
||||||
const checkStatus = backendHealth[type] || {};
|
|
||||||
if (locations.length === 0) {
|
|
||||||
return process.nextTick(cb, null, []);
|
|
||||||
}
|
|
||||||
if (!flightCheckOnStartUp && checkStatus.response &&
|
|
||||||
Date.now() - checkStatus.time
|
|
||||||
< externalBackendHealthCheckInterval) {
|
|
||||||
return process.nextTick(cb, null, checkStatus.response);
|
|
||||||
}
|
|
||||||
let locationsToCheck;
|
|
||||||
if (flightCheckOnStartUp) {
|
|
||||||
// check all locations if flight check on start up
|
|
||||||
locationsToCheck = locations;
|
|
||||||
} else {
|
|
||||||
const randomLocation = locations[Math.floor(Math.random() *
|
|
||||||
locations.length)];
|
|
||||||
locationsToCheck = [randomLocation];
|
|
||||||
}
|
|
||||||
return async.mapLimit(locationsToCheck, 5, (location, next) => {
|
|
||||||
const client = clients[location];
|
|
||||||
client.healthcheck(location, next, flightCheckOnStartUp);
|
|
||||||
}, (err, results) => {
|
|
||||||
if (err) {
|
|
||||||
return cb(err);
|
|
||||||
}
|
|
||||||
if (!flightCheckOnStartUp) {
|
|
||||||
checkStatus.response = results;
|
|
||||||
checkStatus.time = Date.now();
|
|
||||||
}
|
|
||||||
return cb(null, results);
|
|
||||||
});
|
|
||||||
},
|
|
||||||
translateAzureMetaHeaders(metaHeaders, tags) {
|
|
||||||
const translatedMetaHeaders = {};
|
|
||||||
if (tags) {
|
|
||||||
// tags are passed as string of format 'key1=value1&key2=value2'
|
|
||||||
const tagObj = {};
|
|
||||||
const tagArr = tags.split('&');
|
|
||||||
tagArr.forEach(keypair => {
|
|
||||||
const equalIndex = keypair.indexOf('=');
|
|
||||||
const key = keypair.substring(0, equalIndex);
|
|
||||||
tagObj[key] = keypair.substring(equalIndex + 1);
|
|
||||||
});
|
|
||||||
translatedMetaHeaders.tags = JSON.stringify(tagObj);
|
|
||||||
}
|
|
||||||
Object.keys(metaHeaders).forEach(headerName => {
|
|
||||||
const translated = headerName.substring(11).replace(/-/g, '_');
|
|
||||||
translatedMetaHeaders[translated] = metaHeaders[headerName];
|
|
||||||
});
|
|
||||||
return translatedMetaHeaders;
|
|
||||||
},
|
|
||||||
};
|
|
||||||
|
|
||||||
module.exports = utils;
|
|
|
@ -3,16 +3,16 @@ const http = require('http');
|
||||||
const url = require('url');
|
const url = require('url');
|
||||||
const AWS = require('aws-sdk');
|
const AWS = require('aws-sdk');
|
||||||
const Sproxy = require('sproxydclient');
|
const Sproxy = require('sproxydclient');
|
||||||
|
const externalBackends = require('arsenal').storage.data.external;
|
||||||
const HttpsProxyAgent = require('https-proxy-agent');
|
const HttpsProxyAgent = require('https-proxy-agent');
|
||||||
|
|
||||||
const DataFileBackend = require('./file/backend');
|
const DataFileBackend = require('./file/backend');
|
||||||
const inMemory = require('./in_memory/backend').backend;
|
const inMemory = require('./in_memory/backend').backend;
|
||||||
const AwsClient = require('./external/AwsClient');
|
const AwsClient = externalBackends.AwsClient;
|
||||||
const GcpClient = require('./external/GcpClient');
|
const GcpClient = externalBackends.GcpClient;
|
||||||
const AzureClient = require('./external/AzureClient');
|
const AzureClient = externalBackends.AzureClient;
|
||||||
const PfsClient = require('./external/PfsClient');
|
const PfsClient = externalBackends.PfsClient;
|
||||||
const proxyCompareUrl = require('./proxyCompareUrl');
|
const { proxyCompareUrl } = externalBackends.backendUtils;
|
||||||
const constants = require('../../constants');
|
const constants = require('../../constants');
|
||||||
|
|
||||||
const { config } = require('../Config');
|
const { config } = require('../Config');
|
||||||
|
|
|
@ -1,17 +1,17 @@
|
||||||
const { errors, s3middleware } = require('arsenal');
|
const { errors, s3middleware, storage } = require('arsenal');
|
||||||
const { parseTagFromQuery } = s3middleware.tagging;
|
const { parseTagFromQuery } = s3middleware.tagging;
|
||||||
|
|
||||||
const createLogger = require('./multipleBackendLogger');
|
|
||||||
const async = require('async');
|
const async = require('async');
|
||||||
|
|
||||||
const { config } = require('../Config');
|
const { config } = require('../Config');
|
||||||
const parseLC = require('./locationConstraintParser');
|
const parseLC = require('./locationConstraintParser');
|
||||||
const DataFileBackend = require('./file/backend');
|
const DataFileBackend = require('./file/backend');
|
||||||
|
const metadata = require('../metadata/wrapper');
|
||||||
const { checkExternalBackend } = require('./external/utils');
|
|
||||||
const { externalBackendHealthCheckInterval } = require('../../constants');
|
const { externalBackendHealthCheckInterval } = require('../../constants');
|
||||||
const locationStorageCheck =
|
const locationStorageCheck =
|
||||||
require('../api/apiUtils/object/locationStorageCheck');
|
require('../api/apiUtils/object/locationStorageCheck');
|
||||||
|
const { createLogger, checkExternalBackend } =
|
||||||
|
storage.data.external.backendUtils;
|
||||||
|
|
||||||
let clients = parseLC(config);
|
let clients = parseLC(config);
|
||||||
config.on('location-constraints-update', () => {
|
config.on('location-constraints-update', () => {
|
||||||
|
@ -78,7 +78,8 @@ const multipleBackendGateway = {
|
||||||
dataStoreMD5,
|
dataStoreMD5,
|
||||||
};
|
};
|
||||||
return callback(null, dataRetrievalInfo);
|
return callback(null, dataRetrievalInfo);
|
||||||
});
|
// send metadata as param for AzureClient in Arsenal
|
||||||
|
}, metadata);
|
||||||
},
|
},
|
||||||
|
|
||||||
head: (objectGetInfoArr, reqUids, callback) => {
|
head: (objectGetInfoArr, reqUids, callback) => {
|
||||||
|
@ -204,7 +205,7 @@ const multipleBackendGateway = {
|
||||||
return cb(err);
|
return cb(err);
|
||||||
}
|
}
|
||||||
return client.uploadPart(request, streamingV4Params, stream,
|
return client.uploadPart(request, streamingV4Params, stream,
|
||||||
size, key, uploadId, partNumber, bucketName, log,
|
size, key, uploadId, partNumber, bucketName, config, log,
|
||||||
(err, partInfo) => {
|
(err, partInfo) => {
|
||||||
if (err) {
|
if (err) {
|
||||||
// if error putting part, counter should be decremented
|
// if error putting part, counter should be decremented
|
||||||
|
@ -330,7 +331,7 @@ const multipleBackendGateway = {
|
||||||
const client = clients[location];
|
const client = clients[location];
|
||||||
if (client.uploadPartCopy) {
|
if (client.uploadPartCopy) {
|
||||||
return client.uploadPartCopy(request, awsSourceKey,
|
return client.uploadPartCopy(request, awsSourceKey,
|
||||||
sourceLocationConstraintName,
|
sourceLocationConstraintName, config,
|
||||||
log, cb);
|
log, cb);
|
||||||
}
|
}
|
||||||
return cb(errors.NotImplemented.customizeDescription(
|
return cb(errors.NotImplemented.customizeDescription(
|
||||||
|
|
|
@ -1,11 +0,0 @@
|
||||||
const werelogs = require('werelogs');
|
|
||||||
|
|
||||||
const logger = new werelogs.Logger('MultipleBackendGateway');
|
|
||||||
|
|
||||||
function createLogger(reqUids) {
|
|
||||||
return reqUids ?
|
|
||||||
logger.newRequestLoggerFromSerializedUids(reqUids) :
|
|
||||||
logger.newRequestLogger();
|
|
||||||
}
|
|
||||||
|
|
||||||
module.exports = createLogger;
|
|
|
@ -1,53 +0,0 @@
|
||||||
/**
|
|
||||||
* proxyCompareUrl - compares request endpoint to urls in NO_PROXY env var
|
|
||||||
* @param {string} endpoint - url of request
|
|
||||||
* @return {bool} true if request endpoint matches no proxy, false if not
|
|
||||||
*/
|
|
||||||
function proxyCompareUrl(endpoint) {
|
|
||||||
const noProxy = process.env.NO_PROXY || process.env.no_proxy;
|
|
||||||
if (!noProxy) {
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
// noProxy env var is a comma separated list of urls not to proxy
|
|
||||||
const noProxyList = noProxy.split(',');
|
|
||||||
if (noProxyList.includes(endpoint)) {
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
const epArr = endpoint.split('.');
|
|
||||||
// reverse array to make comparison easier
|
|
||||||
epArr.reverse();
|
|
||||||
let match = false;
|
|
||||||
for (let j = 0; j < noProxyList.length; j++) {
|
|
||||||
const urlArr = noProxyList[j].split('.');
|
|
||||||
urlArr.reverse();
|
|
||||||
for (let i = 0; i < epArr.length; i++) {
|
|
||||||
if (epArr[i] === urlArr[i]) {
|
|
||||||
match = true;
|
|
||||||
} else if (urlArr[i] === '*' && i === (urlArr.length - 1)) {
|
|
||||||
// if first character of url is '*', remaining endpoint matches
|
|
||||||
match = true;
|
|
||||||
break;
|
|
||||||
} else if (urlArr[i] === '' && i === (urlArr.length - 1)) {
|
|
||||||
// if first character of url is '.', it is treated as wildcard
|
|
||||||
match = true;
|
|
||||||
break;
|
|
||||||
} else if (urlArr[i] === '*') {
|
|
||||||
match = true;
|
|
||||||
} else if (epArr[i] !== urlArr[i]) {
|
|
||||||
match = false;
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
// if endpoint matches noProxy element, stop checking
|
|
||||||
if (match) {
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
// if endpoint matches, request should not be proxied
|
|
||||||
if (match) {
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
|
|
||||||
module.exports = proxyCompareUrl;
|
|
|
@ -1,22 +1,23 @@
|
||||||
const async = require('async');
|
const async = require('async');
|
||||||
const { errors, s3middleware } = require('arsenal');
|
const { errors, s3middleware, storage } = require('arsenal');
|
||||||
const PassThrough = require('stream').PassThrough;
|
const PassThrough = require('stream').PassThrough;
|
||||||
|
const assert = require('assert');
|
||||||
|
|
||||||
const DataFileInterface = require('./file/backend');
|
const DataFileInterface = require('./file/backend');
|
||||||
const inMemory = require('./in_memory/backend').backend;
|
const inMemory = require('./in_memory/backend').backend;
|
||||||
const multipleBackendGateway = require('./multipleBackendGateway');
|
const multipleBackendGateway = require('./multipleBackendGateway');
|
||||||
const utils = require('./external/utils');
|
|
||||||
const { config } = require('../Config');
|
const { config } = require('../Config');
|
||||||
const MD5Sum = s3middleware.MD5Sum;
|
|
||||||
const NullStream = s3middleware.NullStream;
|
|
||||||
const assert = require('assert');
|
|
||||||
const kms = require('../kms/wrapper');
|
const kms = require('../kms/wrapper');
|
||||||
const externalBackends = require('../../constants').externalBackends;
|
|
||||||
const constants = require('../../constants');
|
const constants = require('../../constants');
|
||||||
const { BackendInfo } = require('../api/apiUtils/object/BackendInfo');
|
const { BackendInfo } = require('../api/apiUtils/object/BackendInfo');
|
||||||
const locationStorageCheck =
|
const locationStorageCheck =
|
||||||
require('../api/apiUtils/object/locationStorageCheck');
|
require('../api/apiUtils/object/locationStorageCheck');
|
||||||
const RelayMD5Sum = require('../utilities/RelayMD5Sum');
|
const RelayMD5Sum = require('../utilities/RelayMD5Sum');
|
||||||
|
const externalBackends = constants.externalBackends;
|
||||||
|
const MD5Sum = s3middleware.MD5Sum;
|
||||||
|
const NullStream = s3middleware.NullStream;
|
||||||
|
const { backendUtils } = storage.data.external;
|
||||||
|
|
||||||
const skipError = new Error('skip');
|
const skipError = new Error('skip');
|
||||||
|
|
||||||
let CdmiData;
|
let CdmiData;
|
||||||
|
@ -478,7 +479,7 @@ const data = {
|
||||||
cb) => {
|
cb) => {
|
||||||
const serverSideEncryption = destBucketMD.getServerSideEncryption();
|
const serverSideEncryption = destBucketMD.getServerSideEncryption();
|
||||||
if (config.backends.data === 'multiple' &&
|
if (config.backends.data === 'multiple' &&
|
||||||
utils.externalBackendCopy(sourceLocationConstraintName,
|
backendUtils.externalBackendCopy(config, sourceLocationConstraintName,
|
||||||
storeMetadataParams.dataStoreName, sourceBucketMD, destBucketMD)) {
|
storeMetadataParams.dataStoreName, sourceBucketMD, destBucketMD)) {
|
||||||
const destLocationConstraintName =
|
const destLocationConstraintName =
|
||||||
storeMetadataParams.dataStoreName;
|
storeMetadataParams.dataStoreName;
|
||||||
|
@ -486,7 +487,7 @@ const data = {
|
||||||
const externalSourceKey = objectGetInfo.key;
|
const externalSourceKey = objectGetInfo.key;
|
||||||
return client.copyObject(request, destLocationConstraintName,
|
return client.copyObject(request, destLocationConstraintName,
|
||||||
externalSourceKey, sourceLocationConstraintName,
|
externalSourceKey, sourceLocationConstraintName,
|
||||||
storeMetadataParams, log, (error, objectRetrievalInfo) => {
|
storeMetadataParams, config, log, (error, objectRetrievalInfo) => {
|
||||||
if (error) {
|
if (error) {
|
||||||
return cb(error);
|
return cb(error);
|
||||||
}
|
}
|
||||||
|
|
|
@ -19,11 +19,10 @@
|
||||||
},
|
},
|
||||||
"homepage": "https://github.com/scality/S3#readme",
|
"homepage": "https://github.com/scality/S3#readme",
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"arsenal": "github:scality/arsenal#84bf7bd",
|
"arsenal": "github:scality/arsenal#d7ca4c17",
|
||||||
"async": "~2.5.0",
|
"async": "~2.5.0",
|
||||||
"aws-sdk": "2.28.0",
|
"aws-sdk": "2.28.0",
|
||||||
"azure-storage": "^2.1.0",
|
"azure-storage": "^2.1.0",
|
||||||
"backo": "^1.1.0",
|
|
||||||
"bucketclient": "scality/bucketclient#5aa99d7",
|
"bucketclient": "scality/bucketclient#5aa99d7",
|
||||||
"bufferutil": "^3.0.5",
|
"bufferutil": "^3.0.5",
|
||||||
"commander": "^2.9.0",
|
"commander": "^2.9.0",
|
||||||
|
|
|
@ -1,12 +1,12 @@
|
||||||
const async = require('async');
|
const async = require('async');
|
||||||
const assert = require('assert');
|
const assert = require('assert');
|
||||||
|
const arsenal = require('arsenal');
|
||||||
|
|
||||||
const withV4 = require('../../support/withV4');
|
const withV4 = require('../../support/withV4');
|
||||||
const BucketUtility = require('../../../lib/utility/bucket-util');
|
const BucketUtility = require('../../../lib/utility/bucket-util');
|
||||||
const { describeSkipIfNotMultipleOrCeph, gcpClient, gcpBucketMPU, gcpLocation,
|
const { describeSkipIfNotMultipleOrCeph, gcpClient, gcpBucketMPU, gcpLocation,
|
||||||
genUniqID } = require('../utils');
|
genUniqID } = require('../utils');
|
||||||
const { createMpuKey } =
|
const { createMpuKey } = arsenal.storage.data.external.GcpUtils;
|
||||||
require('../../../../../../lib/data/external/GCP').GcpUtils;
|
|
||||||
|
|
||||||
const bucket = `initmpugcp${genUniqID()}`;
|
const bucket = `initmpugcp${genUniqID()}`;
|
||||||
const keyName = `somekey-${genUniqID()}`;
|
const keyName = `somekey-${genUniqID()}`;
|
||||||
|
|
|
@ -1,13 +1,13 @@
|
||||||
const assert = require('assert');
|
const assert = require('assert');
|
||||||
const async = require('async');
|
const async = require('async');
|
||||||
|
const arsenal = require('arsenal');
|
||||||
|
|
||||||
const withV4 = require('../../support/withV4');
|
const withV4 = require('../../support/withV4');
|
||||||
const BucketUtility = require('../../../lib/utility/bucket-util');
|
const BucketUtility = require('../../../lib/utility/bucket-util');
|
||||||
const { describeSkipIfNotMultipleOrCeph, gcpClient, gcpBucket, gcpBucketMPU,
|
const { describeSkipIfNotMultipleOrCeph, gcpClient, gcpBucket, gcpBucketMPU,
|
||||||
gcpLocation, gcpLocationMismatch, uniqName, genUniqID }
|
gcpLocation, gcpLocationMismatch, uniqName, genUniqID }
|
||||||
= require('../utils');
|
= require('../utils');
|
||||||
const { createMpuKey } =
|
const { createMpuKey } = arsenal.storage.data.external.GcpUtils;
|
||||||
require('../../../../../../lib/data/external/GCP').GcpUtils;
|
|
||||||
|
|
||||||
const keyObject = 'putgcp';
|
const keyObject = 'putgcp';
|
||||||
const bucket = `putpartgcp${genUniqID()}`;
|
const bucket = `putpartgcp${genUniqID()}`;
|
||||||
|
|
|
@ -108,7 +108,7 @@ callback) {
|
||||||
|
|
||||||
describeSkipIfNotMultiple('MultipleBackend object copy: AWS',
|
describeSkipIfNotMultiple('MultipleBackend object copy: AWS',
|
||||||
function testSuite() {
|
function testSuite() {
|
||||||
this.timeout(250000);
|
this.timeout(300000);
|
||||||
withV4(sigCfg => {
|
withV4(sigCfg => {
|
||||||
beforeEach(() => {
|
beforeEach(() => {
|
||||||
bucketUtil = new BucketUtility('default', sigCfg);
|
bucketUtil = new BucketUtility('default', sigCfg);
|
||||||
|
@ -142,8 +142,14 @@ function testSuite() {
|
||||||
afterEach(() => {
|
afterEach(() => {
|
||||||
process.stdout.write('Emptying bucket\n');
|
process.stdout.write('Emptying bucket\n');
|
||||||
return bucketUtil.empty(bucket)
|
return bucketUtil.empty(bucket)
|
||||||
.then(() => bucketUtil.empty(bucketAws))
|
.then(() => {
|
||||||
.then(() => bucketUtil.empty(awsServerSideEncryptionbucket))
|
process.stdout.write('Emptying aws bucket\n');
|
||||||
|
bucketUtil.empty(bucketAws);
|
||||||
|
})
|
||||||
|
.then(() => {
|
||||||
|
process.stdout.write('Empty aws encryption bucket\n');
|
||||||
|
bucketUtil.empty(awsServerSideEncryptionbucket);
|
||||||
|
})
|
||||||
.then(() => {
|
.then(() => {
|
||||||
process.stdout.write(`Deleting bucket ${bucket}\n`);
|
process.stdout.write(`Deleting bucket ${bucket}\n`);
|
||||||
return bucketUtil.deleteOne(bucket);
|
return bucketUtil.deleteOne(bucket);
|
||||||
|
|
|
@ -1,13 +1,13 @@
|
||||||
const assert = require('assert');
|
const assert = require('assert');
|
||||||
const crypto = require('crypto');
|
const crypto = require('crypto');
|
||||||
const { errors } = require('arsenal');
|
const { errors, storage } = require('arsenal');
|
||||||
const AWS = require('aws-sdk');
|
const AWS = require('aws-sdk');
|
||||||
const uuid = require('uuid/v4');
|
const uuid = require('uuid/v4');
|
||||||
|
|
||||||
const async = require('async');
|
const async = require('async');
|
||||||
const azure = require('azure-storage');
|
const azure = require('azure-storage');
|
||||||
|
|
||||||
const { GCP } = require('../../../../../lib/data/external/GCP');
|
const { GCP } = storage.data.external;
|
||||||
|
|
||||||
const { getRealAwsConfig } = require('../support/awsConfig');
|
const { getRealAwsConfig } = require('../support/awsConfig');
|
||||||
const { config } = require('../../../../../lib/Config');
|
const { config } = require('../../../../../lib/Config');
|
||||||
|
|
|
@ -1,6 +1,7 @@
|
||||||
const assert = require('assert');
|
const assert = require('assert');
|
||||||
const async = require('async');
|
const async = require('async');
|
||||||
const { GCP } = require('../../../../../../lib/data/external/GCP');
|
const arsenal = require('arsenal');
|
||||||
|
const { GCP } = arsenal.storage.data.external;
|
||||||
const { makeGcpRequest } = require('../../../utils/makeRequest');
|
const { makeGcpRequest } = require('../../../utils/makeRequest');
|
||||||
const { gcpRequestRetry, genUniqID } = require('../../../utils/gcpUtils');
|
const { gcpRequestRetry, genUniqID } = require('../../../utils/gcpUtils');
|
||||||
const { getRealAwsConfig } =
|
const { getRealAwsConfig } =
|
||||||
|
|
|
@ -1,6 +1,7 @@
|
||||||
const assert = require('assert');
|
const assert = require('assert');
|
||||||
const async = require('async');
|
const async = require('async');
|
||||||
const { GCP } = require('../../../../../../lib/data/external/GCP');
|
const arsenal = require('arsenal');
|
||||||
|
const { GCP } = arsenal.storage.data.external;
|
||||||
const { makeGcpRequest } = require('../../../utils/makeRequest');
|
const { makeGcpRequest } = require('../../../utils/makeRequest');
|
||||||
const { gcpRequestRetry, genUniqID } = require('../../../utils/gcpUtils');
|
const { gcpRequestRetry, genUniqID } = require('../../../utils/gcpUtils');
|
||||||
const { getRealAwsConfig } =
|
const { getRealAwsConfig } =
|
||||||
|
|
|
@ -1,5 +1,6 @@
|
||||||
const assert = require('assert');
|
const assert = require('assert');
|
||||||
const { GCP } = require('../../../../../../lib/data/external/GCP');
|
const arsenal = require('arsenal');
|
||||||
|
const { GCP } = arsenal.storage.data.external;
|
||||||
const { gcpRequestRetry, genUniqID } = require('../../../utils/gcpUtils');
|
const { gcpRequestRetry, genUniqID } = require('../../../utils/gcpUtils');
|
||||||
const { getRealAwsConfig } =
|
const { getRealAwsConfig } =
|
||||||
require('../../../../aws-node-sdk/test/support/awsConfig');
|
require('../../../../aws-node-sdk/test/support/awsConfig');
|
||||||
|
|
|
@ -1,7 +1,8 @@
|
||||||
const assert = require('assert');
|
const assert = require('assert');
|
||||||
const async = require('async');
|
const async = require('async');
|
||||||
|
const arsenal = require('arsenal');
|
||||||
const xml2js = require('xml2js');
|
const xml2js = require('xml2js');
|
||||||
const { GCP } = require('../../../../../../lib/data/external/GCP');
|
const { GCP } = arsenal.storage.data.external;
|
||||||
const { makeGcpRequest } = require('../../../utils/makeRequest');
|
const { makeGcpRequest } = require('../../../utils/makeRequest');
|
||||||
const { gcpRequestRetry, genUniqID } = require('../../../utils/gcpUtils');
|
const { gcpRequestRetry, genUniqID } = require('../../../utils/gcpUtils');
|
||||||
const { getRealAwsConfig } =
|
const { getRealAwsConfig } =
|
||||||
|
|
|
@ -1,6 +1,7 @@
|
||||||
const assert = require('assert');
|
const assert = require('assert');
|
||||||
const async = require('async');
|
const async = require('async');
|
||||||
const { GCP, GcpUtils } = require('../../../../../../lib/data/external/GCP');
|
const arsenal = require('arsenal');
|
||||||
|
const { GCP, GcpUtils } = arsenal.storage.data.external;
|
||||||
const { gcpRequestRetry, setBucketClass, gcpMpuSetup, genUniqID } =
|
const { gcpRequestRetry, setBucketClass, gcpMpuSetup, genUniqID } =
|
||||||
require('../../../utils/gcpUtils');
|
require('../../../utils/gcpUtils');
|
||||||
const { getRealAwsConfig } =
|
const { getRealAwsConfig } =
|
||||||
|
|
|
@ -1,6 +1,7 @@
|
||||||
const assert = require('assert');
|
const assert = require('assert');
|
||||||
const async = require('async');
|
const async = require('async');
|
||||||
const { GCP } = require('../../../../../../lib/data/external/GCP');
|
const arsenal = require('arsenal');
|
||||||
|
const { GCP } = arsenal.storage.data.external;
|
||||||
const { makeGcpRequest } = require('../../../utils/makeRequest');
|
const { makeGcpRequest } = require('../../../utils/makeRequest');
|
||||||
const { gcpRequestRetry, genUniqID } = require('../../../utils/gcpUtils');
|
const { gcpRequestRetry, genUniqID } = require('../../../utils/gcpUtils');
|
||||||
const { getRealAwsConfig } =
|
const { getRealAwsConfig } =
|
||||||
|
|
|
@ -1,6 +1,7 @@
|
||||||
const assert = require('assert');
|
const assert = require('assert');
|
||||||
const async = require('async');
|
const async = require('async');
|
||||||
const { GCP } = require('../../../../../../lib/data/external/GCP');
|
const arsenal = require('arsenal');
|
||||||
|
const { GCP } = arsenal.storage.data.external;
|
||||||
const { makeGcpRequest } = require('../../../utils/makeRequest');
|
const { makeGcpRequest } = require('../../../utils/makeRequest');
|
||||||
const { gcpRequestRetry, genUniqID } = require('../../../utils/gcpUtils');
|
const { gcpRequestRetry, genUniqID } = require('../../../utils/gcpUtils');
|
||||||
const { getRealAwsConfig } =
|
const { getRealAwsConfig } =
|
||||||
|
|
|
@ -1,6 +1,7 @@
|
||||||
const assert = require('assert');
|
const assert = require('assert');
|
||||||
const async = require('async');
|
const async = require('async');
|
||||||
const { GCP } = require('../../../../../../lib/data/external/GCP');
|
const arsenal = require('arsenal');
|
||||||
|
const { GCP } = arsenal.storage.data.external;
|
||||||
const { gcpRequestRetry, setBucketClass, gcpMpuSetup, genUniqID } =
|
const { gcpRequestRetry, setBucketClass, gcpMpuSetup, genUniqID } =
|
||||||
require('../../../utils/gcpUtils');
|
require('../../../utils/gcpUtils');
|
||||||
const { getRealAwsConfig } =
|
const { getRealAwsConfig } =
|
||||||
|
|
|
@ -1,6 +1,7 @@
|
||||||
const assert = require('assert');
|
const assert = require('assert');
|
||||||
const async = require('async');
|
const async = require('async');
|
||||||
const { GCP } = require('../../../../../../lib/data/external/GCP');
|
const arsenal = require('arsenal');
|
||||||
|
const { GCP } = arsenal.storage.data.external;
|
||||||
const { makeGcpRequest } = require('../../../utils/makeRequest');
|
const { makeGcpRequest } = require('../../../utils/makeRequest');
|
||||||
const { gcpRequestRetry, genDelTagObj, genUniqID } =
|
const { gcpRequestRetry, genDelTagObj, genUniqID } =
|
||||||
require('../../../utils/gcpUtils');
|
require('../../../utils/gcpUtils');
|
||||||
|
|
|
@ -1,5 +1,6 @@
|
||||||
const assert = require('assert');
|
const assert = require('assert');
|
||||||
const { GCP } = require('../../../../../../lib/data/external/GCP');
|
const arsenal = require('arsenal');
|
||||||
|
const { GCP } = arsenal.storage.data.external;
|
||||||
const { makeGcpRequest } = require('../../../utils/makeRequest');
|
const { makeGcpRequest } = require('../../../utils/makeRequest');
|
||||||
const { gcpRequestRetry, genUniqID } = require('../../../utils/gcpUtils');
|
const { gcpRequestRetry, genUniqID } = require('../../../utils/gcpUtils');
|
||||||
const { getRealAwsConfig } =
|
const { getRealAwsConfig } =
|
||||||
|
|
|
@ -1,5 +1,6 @@
|
||||||
const assert = require('assert');
|
const assert = require('assert');
|
||||||
const { GCP } = require('../../../../../../lib/data/external/GCP');
|
const arsenal = require('arsenal');
|
||||||
|
const { GCP } = arsenal.storage.data.external;
|
||||||
const { makeGcpRequest } = require('../../../utils/makeRequest');
|
const { makeGcpRequest } = require('../../../utils/makeRequest');
|
||||||
const { gcpRequestRetry, genGetTagObj, genUniqID } =
|
const { gcpRequestRetry, genGetTagObj, genUniqID } =
|
||||||
require('../../../utils/gcpUtils');
|
require('../../../utils/gcpUtils');
|
||||||
|
|
|
@ -1,5 +1,6 @@
|
||||||
const assert = require('assert');
|
const assert = require('assert');
|
||||||
const { GCP } = require('../../../../../../lib/data/external/GCP');
|
const arsenal = require('arsenal');
|
||||||
|
const { GCP } = arsenal.storage.data.external;
|
||||||
const { makeGcpRequest } = require('../../../utils/makeRequest');
|
const { makeGcpRequest } = require('../../../utils/makeRequest');
|
||||||
const { gcpRequestRetry, genUniqID } = require('../../../utils/gcpUtils');
|
const { gcpRequestRetry, genUniqID } = require('../../../utils/gcpUtils');
|
||||||
const { getRealAwsConfig } =
|
const { getRealAwsConfig } =
|
||||||
|
|
|
@ -1,6 +1,7 @@
|
||||||
const assert = require('assert');
|
const assert = require('assert');
|
||||||
const async = require('async');
|
const async = require('async');
|
||||||
const { GCP } = require('../../../../../../lib/data/external/GCP');
|
const arsenal = require('arsenal');
|
||||||
|
const { GCP } = arsenal.storage.data.external;
|
||||||
const { makeGcpRequest } = require('../../../utils/makeRequest');
|
const { makeGcpRequest } = require('../../../utils/makeRequest');
|
||||||
const { gcpRequestRetry, setBucketClass, genUniqID } =
|
const { gcpRequestRetry, setBucketClass, genUniqID } =
|
||||||
require('../../../utils/gcpUtils');
|
require('../../../utils/gcpUtils');
|
||||||
|
|
|
@ -1,5 +1,6 @@
|
||||||
const assert = require('assert');
|
const assert = require('assert');
|
||||||
const { GCP } = require('../../../../../../lib/data/external/GCP');
|
const arsenal = require('arsenal');
|
||||||
|
const { GCP } = arsenal.storage.data.external;
|
||||||
const { makeGcpRequest } = require('../../../utils/makeRequest');
|
const { makeGcpRequest } = require('../../../utils/makeRequest');
|
||||||
const { gcpRequestRetry, genUniqID } = require('../../../utils/gcpUtils');
|
const { gcpRequestRetry, genUniqID } = require('../../../utils/gcpUtils');
|
||||||
const { getRealAwsConfig } =
|
const { getRealAwsConfig } =
|
||||||
|
|
|
@ -1,6 +1,7 @@
|
||||||
const assert = require('assert');
|
const assert = require('assert');
|
||||||
const async = require('async');
|
const async = require('async');
|
||||||
const { GCP } = require('../../../../../../lib/data/external/GCP');
|
const arsenal = require('arsenal');
|
||||||
|
const { GCP } = arsenal.storage.data.external;
|
||||||
const { makeGcpRequest } = require('../../../utils/makeRequest');
|
const { makeGcpRequest } = require('../../../utils/makeRequest');
|
||||||
const { gcpRequestRetry, genPutTagObj, genUniqID } =
|
const { gcpRequestRetry, genPutTagObj, genUniqID } =
|
||||||
require('../../../utils/gcpUtils');
|
require('../../../utils/gcpUtils');
|
||||||
|
|
|
@ -1,6 +1,7 @@
|
||||||
const assert = require('assert');
|
const assert = require('assert');
|
||||||
const async = require('async');
|
const async = require('async');
|
||||||
const { GCP } = require('../../../../../../lib/data/external/GCP');
|
const arsenal = require('arsenal');
|
||||||
|
const { GCP } = arsenal.storage.data.external;
|
||||||
const { gcpRequestRetry, setBucketClass, genUniqID } =
|
const { gcpRequestRetry, setBucketClass, genUniqID } =
|
||||||
require('../../../utils/gcpUtils');
|
require('../../../utils/gcpUtils');
|
||||||
const { getRealAwsConfig } =
|
const { getRealAwsConfig } =
|
||||||
|
|
|
@ -1,11 +1,11 @@
|
||||||
const { auth } = require('arsenal');
|
const { auth, storage } = require('arsenal');
|
||||||
|
|
||||||
const http = require('http');
|
const http = require('http');
|
||||||
const https = require('https');
|
const https = require('https');
|
||||||
const querystring = require('querystring');
|
const querystring = require('querystring');
|
||||||
|
|
||||||
const conf = require('../../../../lib/Config').config;
|
const conf = require('../../../../lib/Config').config;
|
||||||
const { GcpSigner } = require('../../../../lib/data/external/GCP');
|
const { GcpSigner } = storage.data.external;
|
||||||
|
|
||||||
const transport = conf.https ? https : http;
|
const transport = conf.https ? https : http;
|
||||||
const ipAddress = process.env.IP ? process.env.IP : '127.0.0.1';
|
const ipAddress = process.env.IP ? process.env.IP : '127.0.0.1';
|
||||||
|
|
|
@ -1,106 +0,0 @@
|
||||||
const assert = require('assert');
|
|
||||||
|
|
||||||
const AwsClient = require('../../../lib/data/external/AwsClient');
|
|
||||||
const GcpClient = require('../../../lib/data/external/GcpClient');
|
|
||||||
const AzureClient = require('../../../lib/data/external/AzureClient');
|
|
||||||
const DummyService = require('../DummyService');
|
|
||||||
const { DummyRequestLogger } = require('../helpers');
|
|
||||||
|
|
||||||
const backendClients = [
|
|
||||||
{
|
|
||||||
Class: AwsClient,
|
|
||||||
name: 'AwsClient',
|
|
||||||
config: {
|
|
||||||
s3Params: {},
|
|
||||||
bucketName: 'awsTestBucketName',
|
|
||||||
dataStoreName: 'awsDataStore',
|
|
||||||
serverSideEncryption: false,
|
|
||||||
type: 'aws',
|
|
||||||
},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Class: GcpClient,
|
|
||||||
name: 'GcpClient',
|
|
||||||
config: {
|
|
||||||
s3Params: {},
|
|
||||||
bucketName: 'gcpTestBucketName',
|
|
||||||
mpuBucket: 'gcpTestMpuBucketName',
|
|
||||||
dataStoreName: 'gcpDataStore',
|
|
||||||
type: 'gcp',
|
|
||||||
},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Class: AzureClient,
|
|
||||||
name: 'AzureClient',
|
|
||||||
config: {
|
|
||||||
azureStorageEndpoint: '',
|
|
||||||
azureStorageCredentials: {
|
|
||||||
storageAccountName: 'scality',
|
|
||||||
storageAccessKey: 'Zm9vCg==',
|
|
||||||
},
|
|
||||||
azureContainerName: 'azureTestBucketName',
|
|
||||||
dataStoreName: 'azureDataStore',
|
|
||||||
type: 'azure',
|
|
||||||
},
|
|
||||||
},
|
|
||||||
];
|
|
||||||
const log = new DummyRequestLogger();
|
|
||||||
|
|
||||||
describe('external backend clients', () => {
|
|
||||||
backendClients.forEach(backend => {
|
|
||||||
let testClient;
|
|
||||||
|
|
||||||
before(() => {
|
|
||||||
testClient = new backend.Class(backend.config);
|
|
||||||
testClient._client = new DummyService({ versioning: true });
|
|
||||||
});
|
|
||||||
|
|
||||||
if (backend.config.type !== 'azure') {
|
|
||||||
it(`${backend.name} completeMPU should return correctly ` +
|
|
||||||
'typed mpu results', done => {
|
|
||||||
const jsonList = {
|
|
||||||
Part: [
|
|
||||||
{
|
|
||||||
PartNumber: [1],
|
|
||||||
ETag: ['testpart0001etag'],
|
|
||||||
},
|
|
||||||
{
|
|
||||||
PartNumber: [2],
|
|
||||||
ETag: ['testpart0002etag'],
|
|
||||||
},
|
|
||||||
{
|
|
||||||
PartNumber: [3],
|
|
||||||
ETag: ['testpart0003etag'],
|
|
||||||
},
|
|
||||||
],
|
|
||||||
};
|
|
||||||
const key = 'externalBackendTestKey';
|
|
||||||
const bucketName = 'externalBackendTestBucket';
|
|
||||||
const uploadId = 'externalBackendTestUploadId';
|
|
||||||
testClient.completeMPU(jsonList, null, key,
|
|
||||||
uploadId, bucketName, log, (err, res) => {
|
|
||||||
assert.strictEqual(typeof res.key, 'string');
|
|
||||||
assert.strictEqual(typeof res.eTag, 'string');
|
|
||||||
assert.strictEqual(typeof res.dataStoreVersionId,
|
|
||||||
'string');
|
|
||||||
assert.strictEqual(typeof res.contentLength, 'number');
|
|
||||||
return done();
|
|
||||||
});
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
it(`${backend.name} toObjectGetInfo should return correct ` +
|
|
||||||
'objectGetInfo object', () => {
|
|
||||||
const key = 'externalBackendTestKey';
|
|
||||||
const bucketName = 'externalBackendTestBucket';
|
|
||||||
const objectGetInfo = testClient.toObjectGetInfo(key, bucketName);
|
|
||||||
assert.deepStrictEqual(objectGetInfo, {
|
|
||||||
// bucketMatch === false => expect bucket name to be
|
|
||||||
// prefixed to the backend key
|
|
||||||
key: 'externalBackendTestBucket/externalBackendTestKey',
|
|
||||||
dataStoreName: backend.config.dataStoreName,
|
|
||||||
});
|
|
||||||
});
|
|
||||||
// To-Do: test the other external client methods
|
|
||||||
});
|
|
||||||
});
|
|
|
@ -1,194 +0,0 @@
|
||||||
const assert = require('assert');
|
|
||||||
const http = require('http');
|
|
||||||
const { GCP } = require('../../../lib/data/external/GCP');
|
|
||||||
|
|
||||||
const httpPort = 8888;
|
|
||||||
|
|
||||||
// test values
|
|
||||||
const host = 'localhost:8888';
|
|
||||||
const Bucket = 'testrequestbucket';
|
|
||||||
const Key = 'testRequestKey';
|
|
||||||
const MultipartUpload = { Parts: [{ PartName: 'part' }] };
|
|
||||||
const CopySource = 'copyBucket/copyKey';
|
|
||||||
const accessKeyId = 'accesskey';
|
|
||||||
const secretAccessKey = 'secretaccesskey';
|
|
||||||
|
|
||||||
function handler(isPathStyle) {
|
|
||||||
return (req, res) => {
|
|
||||||
if (isPathStyle) {
|
|
||||||
assert(req.headers.host, host);
|
|
||||||
assert(req.url.includes(Bucket));
|
|
||||||
} else {
|
|
||||||
assert(req.headers.host, `${Bucket}.${host}`);
|
|
||||||
assert(!req.url.includes(Bucket));
|
|
||||||
}
|
|
||||||
res.end();
|
|
||||||
};
|
|
||||||
}
|
|
||||||
|
|
||||||
const invalidBucketNames = [
|
|
||||||
'..',
|
|
||||||
'.bucketname',
|
|
||||||
'bucketname.',
|
|
||||||
'bucketName.',
|
|
||||||
'aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa',
|
|
||||||
'256.256.256.256',
|
|
||||||
'',
|
|
||||||
];
|
|
||||||
|
|
||||||
function badBucketNameHandler(req, res) {
|
|
||||||
assert(req.headers.host, host);
|
|
||||||
const bucketFromUrl = req.url.split('/')[1];
|
|
||||||
assert.strictEqual(typeof bucketFromUrl, 'string');
|
|
||||||
assert(invalidBucketNames.includes(bucketFromUrl));
|
|
||||||
res.end();
|
|
||||||
}
|
|
||||||
|
|
||||||
const operations = [
|
|
||||||
{
|
|
||||||
op: 'headBucket',
|
|
||||||
params: { Bucket },
|
|
||||||
},
|
|
||||||
{
|
|
||||||
op: 'listObjects',
|
|
||||||
params: { Bucket },
|
|
||||||
},
|
|
||||||
{
|
|
||||||
op: 'listVersions',
|
|
||||||
params: { Bucket },
|
|
||||||
},
|
|
||||||
{
|
|
||||||
op: 'getBucketVersioning',
|
|
||||||
params: { Bucket },
|
|
||||||
},
|
|
||||||
{
|
|
||||||
op: 'headObject',
|
|
||||||
params: { Bucket, Key },
|
|
||||||
},
|
|
||||||
{
|
|
||||||
op: 'putObject',
|
|
||||||
params: { Bucket, Key },
|
|
||||||
},
|
|
||||||
{
|
|
||||||
op: 'getObject',
|
|
||||||
params: { Bucket, Key },
|
|
||||||
},
|
|
||||||
{
|
|
||||||
op: 'deleteObject',
|
|
||||||
params: { Bucket, Key },
|
|
||||||
},
|
|
||||||
{
|
|
||||||
op: 'composeObject',
|
|
||||||
params: { Bucket, Key, MultipartUpload },
|
|
||||||
},
|
|
||||||
{
|
|
||||||
op: 'copyObject',
|
|
||||||
params: { Bucket, Key, CopySource },
|
|
||||||
},
|
|
||||||
];
|
|
||||||
|
|
||||||
describe('GcpService request behavior', function testSuite() {
|
|
||||||
this.timeout(120000);
|
|
||||||
let httpServer;
|
|
||||||
let client;
|
|
||||||
|
|
||||||
before(done => {
|
|
||||||
client = new GCP({
|
|
||||||
endpoint: `http://${host}`,
|
|
||||||
maxRetries: 0,
|
|
||||||
s3ForcePathStyle: false,
|
|
||||||
accessKeyId,
|
|
||||||
secretAccessKey,
|
|
||||||
});
|
|
||||||
httpServer =
|
|
||||||
http.createServer(badBucketNameHandler).listen(httpPort);
|
|
||||||
httpServer.on('listening', done);
|
|
||||||
httpServer.on('error', err => {
|
|
||||||
process.stdout.write(`https server: ${err.stack}\n`);
|
|
||||||
process.exit(1);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
after('Terminating Server', () => {
|
|
||||||
httpServer.close();
|
|
||||||
});
|
|
||||||
|
|
||||||
|
|
||||||
invalidBucketNames.forEach(bucket => {
|
|
||||||
it(`should not use dns-style if bucket isn't dns compatible: ${bucket}`,
|
|
||||||
done => {
|
|
||||||
client.headBucket({ Bucket: bucket }, err => {
|
|
||||||
assert.ifError(err);
|
|
||||||
done();
|
|
||||||
});
|
|
||||||
});
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
describe('GcpService pathStyle tests', function testSuite() {
|
|
||||||
this.timeout(120000);
|
|
||||||
let httpServer;
|
|
||||||
let client;
|
|
||||||
|
|
||||||
before(done => {
|
|
||||||
client = new GCP({
|
|
||||||
endpoint: `http://${host}`,
|
|
||||||
maxRetries: 0,
|
|
||||||
s3ForcePathStyle: true,
|
|
||||||
accessKeyId,
|
|
||||||
secretAccessKey,
|
|
||||||
});
|
|
||||||
httpServer =
|
|
||||||
http.createServer(handler(true)).listen(httpPort);
|
|
||||||
httpServer.on('listening', done);
|
|
||||||
httpServer.on('error', err => {
|
|
||||||
process.stdout.write(`https server: ${err.stack}\n`);
|
|
||||||
process.exit(1);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
after('Terminating Server', () => {
|
|
||||||
httpServer.close();
|
|
||||||
});
|
|
||||||
|
|
||||||
operations.forEach(test => it(`GCP::${test.op}`, done => {
|
|
||||||
client[test.op](test.params, err => {
|
|
||||||
assert.ifError(err);
|
|
||||||
done();
|
|
||||||
});
|
|
||||||
}));
|
|
||||||
});
|
|
||||||
|
|
||||||
describe('GcpService dnsStyle tests', function testSuite() {
|
|
||||||
this.timeout(120000);
|
|
||||||
let httpServer;
|
|
||||||
let client;
|
|
||||||
|
|
||||||
before(done => {
|
|
||||||
client = new GCP({
|
|
||||||
endpoint: `http://localhost:${httpPort}`,
|
|
||||||
maxRetries: 0,
|
|
||||||
s3ForcePathStyle: false,
|
|
||||||
accessKeyId,
|
|
||||||
secretAccessKey,
|
|
||||||
});
|
|
||||||
httpServer =
|
|
||||||
http.createServer(handler(false)).listen(httpPort);
|
|
||||||
httpServer.on('listening', done);
|
|
||||||
httpServer.on('error', err => {
|
|
||||||
process.stdout.write(`https server: ${err.stack}\n`);
|
|
||||||
process.exit(1);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
after('Terminating Server', () => {
|
|
||||||
httpServer.close();
|
|
||||||
});
|
|
||||||
|
|
||||||
operations.forEach(test => it(`GCP::${test.op}`, done => {
|
|
||||||
client[test.op](test.params, err => {
|
|
||||||
assert.ifError(err);
|
|
||||||
done();
|
|
||||||
});
|
|
||||||
}));
|
|
||||||
});
|
|
|
@ -1,7 +1,8 @@
|
||||||
const assert = require('assert');
|
const assert = require('assert');
|
||||||
const { errors } = require('arsenal');
|
const { errors, storage } = require('arsenal');
|
||||||
|
|
||||||
const AwsClient = require('../../../lib/data/external/AwsClient');
|
const AwsClient = storage.data.external.AwsClient;
|
||||||
|
const { config } = require('../../../lib/Config');
|
||||||
const DummyService = require('../DummyService');
|
const DummyService = require('../DummyService');
|
||||||
const { DummyRequestLogger } = require('../helpers');
|
const { DummyRequestLogger } = require('../helpers');
|
||||||
|
|
||||||
|
@ -105,7 +106,8 @@ describe('AwsClient::copyObject', () => {
|
||||||
testClient._supportsVersioning = test.input.supportsVersioning;
|
testClient._supportsVersioning = test.input.supportsVersioning;
|
||||||
testClient._client.versioning = test.input.enableMockVersioning;
|
testClient._client.versioning = test.input.enableMockVersioning;
|
||||||
testClient.copyObject(copyObjectRequest, null, key,
|
testClient.copyObject(copyObjectRequest, null, key,
|
||||||
sourceLocationConstraint, null, log, err => test.callback(err, done));
|
sourceLocationConstraint, null, config, log,
|
||||||
|
err => test.callback(err, done));
|
||||||
}));
|
}));
|
||||||
});
|
});
|
||||||
|
|
||||||
|
|
|
@ -1,6 +1,6 @@
|
||||||
const assert = require('assert');
|
const assert = require('assert');
|
||||||
|
const AwsClient = require('arsenal').storage.data.external.AwsClient;
|
||||||
const parseLC = require('../../../lib/data/locationConstraintParser');
|
const parseLC = require('../../../lib/data/locationConstraintParser');
|
||||||
const AwsClient = require('../../../lib/data/external/AwsClient');
|
|
||||||
const inMemory = require('../../../lib/data/in_memory/backend').backend;
|
const inMemory = require('../../../lib/data/in_memory/backend').backend;
|
||||||
const DataFileInterface = require('../../../lib/data/file/backend');
|
const DataFileInterface = require('../../../lib/data/file/backend');
|
||||||
|
|
||||||
|
|
|
@ -1,7 +1,10 @@
|
||||||
const assert = require('assert');
|
const assert = require('assert');
|
||||||
const utils = require('../../../lib/data/external/utils');
|
const arsenal = require('arsenal');
|
||||||
|
const utils = arsenal.storage.data.external.backendUtils;
|
||||||
const BucketInfo = require('arsenal').models.BucketInfo;
|
const BucketInfo = require('arsenal').models.BucketInfo;
|
||||||
|
|
||||||
|
const { config } = require('../../../lib/Config');
|
||||||
|
|
||||||
const userBucketOwner = 'Bart';
|
const userBucketOwner = 'Bart';
|
||||||
const creationDate = new Date().toJSON();
|
const creationDate = new Date().toJSON();
|
||||||
const serverSideEncryption = { cryptoScheme: 123, algorithm: 'algo',
|
const serverSideEncryption = { cryptoScheme: 123, algorithm: 'algo',
|
||||||
|
@ -136,7 +139,7 @@ describe('Testing Config.js function: ', () => {
|
||||||
'and destination location constraint ===' +
|
'and destination location constraint ===' +
|
||||||
` ${result.destLocationConstraintName} and ${result.description}`,
|
` ${result.destLocationConstraintName} and ${result.description}`,
|
||||||
done => {
|
done => {
|
||||||
const isCopy = utils.externalBackendCopy(
|
const isCopy = utils.externalBackendCopy(config,
|
||||||
result.sourceLocationConstraintName,
|
result.sourceLocationConstraintName,
|
||||||
result.destLocationConstraintName, result.sourceBucketMD,
|
result.destLocationConstraintName, result.sourceBucketMD,
|
||||||
result.destBucketMD);
|
result.destBucketMD);
|
||||||
|
|
|
@ -1,65 +0,0 @@
|
||||||
const assert = require('assert');
|
|
||||||
const uuid = require('uuid/v4');
|
|
||||||
const { createMpuKey, createMpuList } =
|
|
||||||
require('../../../lib/data/external/GCP').GcpUtils;
|
|
||||||
|
|
||||||
const key = `somekey${Date.now()}`;
|
|
||||||
const uploadId = uuid().replace(/-/g, '');
|
|
||||||
const phase = 'createMpulist';
|
|
||||||
const size = 2;
|
|
||||||
const correctMpuList = [
|
|
||||||
{ PartName: `${key}-${uploadId}/${phase}/00001`, PartNumber: 1 },
|
|
||||||
{ PartName: `${key}-${uploadId}/${phase}/00002`, PartNumber: 2 },
|
|
||||||
];
|
|
||||||
|
|
||||||
describe('GcpUtils MPU Helper Functions:', () => {
|
|
||||||
describe('createMpuKey', () => {
|
|
||||||
const tests = [
|
|
||||||
{
|
|
||||||
it: 'if phase and part number are given',
|
|
||||||
input: { phase: 'test', partNumber: 1 },
|
|
||||||
output: `${key}-${uploadId}/test/00001`,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
it: 'if only phase is given',
|
|
||||||
input: { phase: 'test' },
|
|
||||||
output: `${key}-${uploadId}/test`,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
it: 'if part number is given',
|
|
||||||
input: { partNumber: 1 },
|
|
||||||
output: `${key}-${uploadId}/parts/00001`,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
it: 'if phase and part number aren not given',
|
|
||||||
input: {},
|
|
||||||
output: `${key}-${uploadId}/`,
|
|
||||||
},
|
|
||||||
];
|
|
||||||
tests.forEach(test => {
|
|
||||||
it(test.it, () => {
|
|
||||||
const { partNumber, phase } = test.input;
|
|
||||||
assert.strictEqual(createMpuKey(
|
|
||||||
key, uploadId, partNumber, phase), test.output);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
describe('createMpuList', () => {
|
|
||||||
const tests = [
|
|
||||||
{
|
|
||||||
it: 'should create valid mpu list',
|
|
||||||
input: { phase, size },
|
|
||||||
output: correctMpuList,
|
|
||||||
},
|
|
||||||
];
|
|
||||||
tests.forEach(test => {
|
|
||||||
it(test.it, () => {
|
|
||||||
const { phase, size } = test.input;
|
|
||||||
assert.deepStrictEqual(createMpuList(
|
|
||||||
{ Key: key, UploadId: uploadId }, phase, size),
|
|
||||||
test.output);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
});
|
|
||||||
});
|
|
|
@ -1,156 +0,0 @@
|
||||||
const assert = require('assert');
|
|
||||||
const { errors } = require('arsenal');
|
|
||||||
const { gcpTaggingPrefix } = require('../../../constants');
|
|
||||||
const { genPutTagObj } =
|
|
||||||
require('../../../tests/functional/raw-node/utils/gcpUtils');
|
|
||||||
const { processTagSet, stripTags, retrieveTags, getPutTagsMetadata } =
|
|
||||||
require('../../../lib/data/external/GCP').GcpUtils;
|
|
||||||
|
|
||||||
const maxTagSize = 10;
|
|
||||||
const validTagSet = genPutTagObj(2);
|
|
||||||
const validTagObj = {};
|
|
||||||
validTagObj[`${gcpTaggingPrefix}key0`] = 'Value0';
|
|
||||||
validTagObj[`${gcpTaggingPrefix}key1`] = 'Value1';
|
|
||||||
const tagQuery = 'key0=Value0&key1=Value1';
|
|
||||||
const invalidSizeTagSet = genPutTagObj(maxTagSize + 1);
|
|
||||||
const invalidDuplicateTagSet = genPutTagObj(maxTagSize, true);
|
|
||||||
const invalidKeyTagSet = [{ Key: Buffer.alloc(129, 'a'), Value: 'value' }];
|
|
||||||
const invalidValueTagSet = [{ Key: 'key', Value: Buffer.alloc(257, 'a') }];
|
|
||||||
const onlyMetadata = {
|
|
||||||
metadata1: 'metadatavalue1',
|
|
||||||
metadata2: 'metadatavalue2',
|
|
||||||
};
|
|
||||||
const tagMetadata = Object.assign({}, validTagObj, onlyMetadata);
|
|
||||||
const oldTagMetadata = {};
|
|
||||||
oldTagMetadata[`${gcpTaggingPrefix}Old`] = 'OldValue0';
|
|
||||||
const withPriorTags = Object.assign({}, onlyMetadata, oldTagMetadata);
|
|
||||||
|
|
||||||
describe('GcpUtils Tagging Helper Functions:', () => {
|
|
||||||
describe('processTagSet', () => {
|
|
||||||
const tests = [
|
|
||||||
{
|
|
||||||
it: 'should return tag object as metadata for valid tag set',
|
|
||||||
input: validTagSet,
|
|
||||||
output: validTagObj,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
it: 'should return error for invalid tag set size',
|
|
||||||
input: invalidSizeTagSet,
|
|
||||||
output: errors.BadRequest.customizeDescription(
|
|
||||||
'Object tags cannot be greater than 10'),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
it: 'should return error for duplicate tag keys',
|
|
||||||
input: invalidDuplicateTagSet,
|
|
||||||
output: errors.InvalidTag.customizeDescription(
|
|
||||||
'Cannot provide multiple Tags with the same key'),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
it: 'should return error for invalid "key" value',
|
|
||||||
input: invalidKeyTagSet,
|
|
||||||
output: errors.InvalidTag.customizeDescription(
|
|
||||||
'The TagKey provided is too long, 129'),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
it: 'should return error for invalid "value" value',
|
|
||||||
input: invalidValueTagSet,
|
|
||||||
output: errors.InvalidTag.customizeDescription(
|
|
||||||
'The TagValue provided is too long, 257'),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
it: 'should return empty tag object when input is undefined',
|
|
||||||
input: undefined,
|
|
||||||
output: {},
|
|
||||||
},
|
|
||||||
];
|
|
||||||
tests.forEach(test => {
|
|
||||||
it(test.it, () => {
|
|
||||||
assert.deepStrictEqual(processTagSet(test.input), test.output);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
describe('stripTags', () => {
|
|
||||||
const tests = [
|
|
||||||
{
|
|
||||||
it: 'should return metadata without tag',
|
|
||||||
input: tagMetadata,
|
|
||||||
output: onlyMetadata,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
it: 'should return empty object if metadata only has tags',
|
|
||||||
input: validTagObj,
|
|
||||||
output: {},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
it: 'should return empty object if input is undefined',
|
|
||||||
input: undefined,
|
|
||||||
output: {},
|
|
||||||
},
|
|
||||||
];
|
|
||||||
tests.forEach(test => {
|
|
||||||
it(test.it, () => {
|
|
||||||
assert.deepStrictEqual(stripTags(test.input), test.output);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
describe('retrieveTags', () => {
|
|
||||||
const tests = [
|
|
||||||
{
|
|
||||||
it: 'should return tagSet from given input metadata',
|
|
||||||
input: tagMetadata,
|
|
||||||
output: validTagSet,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
it: 'should return empty when metadata does not have tags',
|
|
||||||
input: onlyMetadata,
|
|
||||||
output: [],
|
|
||||||
},
|
|
||||||
{
|
|
||||||
it: 'should return empty if input is undefined',
|
|
||||||
input: undefined,
|
|
||||||
output: [],
|
|
||||||
},
|
|
||||||
];
|
|
||||||
tests.forEach(test => {
|
|
||||||
it(test.it, () => {
|
|
||||||
assert.deepStrictEqual(retrieveTags(test.input), test.output);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
describe('getPutTagsMetadata', () => {
|
|
||||||
const tests = [
|
|
||||||
{
|
|
||||||
it: 'should return correct object when' +
|
|
||||||
' given a tag query string and a metadata obj',
|
|
||||||
input: { metadata: Object.assign({}, onlyMetadata), tagQuery },
|
|
||||||
output: tagMetadata,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
it: 'should return correct object when given only query string',
|
|
||||||
input: { tagQuery },
|
|
||||||
output: validTagObj,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
it: 'should return correct object when only metadata is given',
|
|
||||||
input: { metadata: onlyMetadata },
|
|
||||||
output: onlyMetadata,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
it: 'should return metadata with correct tag properties ' +
|
|
||||||
'if given a metdata with prior tags and query string',
|
|
||||||
input: { metadata: Object.assign({}, withPriorTags), tagQuery },
|
|
||||||
output: tagMetadata,
|
|
||||||
},
|
|
||||||
];
|
|
||||||
tests.forEach(test => {
|
|
||||||
it(test.it, () => {
|
|
||||||
const { metadata, tagQuery } = test.input;
|
|
||||||
assert.deepStrictEqual(
|
|
||||||
getPutTagsMetadata(metadata, tagQuery), test.output);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
});
|
|
||||||
});
|
|
|
@ -1,63 +0,0 @@
|
||||||
const assert = require('assert');
|
|
||||||
const { checkExternalBackend } = require('../../../lib/data/external/utils');
|
|
||||||
const awsLocations = [
|
|
||||||
'awsbackend',
|
|
||||||
];
|
|
||||||
|
|
||||||
const statusSuccess = {
|
|
||||||
versioningStatus: 'Enabled',
|
|
||||||
message: 'Congrats! You own the bucket',
|
|
||||||
};
|
|
||||||
|
|
||||||
const statusFailure = {
|
|
||||||
versioningStatus: 'Suspended',
|
|
||||||
error: 'Versioning must be enabled',
|
|
||||||
external: true,
|
|
||||||
};
|
|
||||||
|
|
||||||
const externalBackendHealthCheckInterval = 10000;
|
|
||||||
|
|
||||||
function getClients(isSuccess) {
|
|
||||||
const status = isSuccess ? statusSuccess : statusFailure;
|
|
||||||
return {
|
|
||||||
awsbackend: {
|
|
||||||
healthcheck: (location, cb) => cb(null, { awsbackend: status }),
|
|
||||||
},
|
|
||||||
};
|
|
||||||
}
|
|
||||||
|
|
||||||
describe('Testing _checkExternalBackend', function describeF() {
|
|
||||||
this.timeout(50000);
|
|
||||||
beforeEach(done => {
|
|
||||||
const clients = getClients(true);
|
|
||||||
return checkExternalBackend(clients, awsLocations, 'aws_s3', false,
|
|
||||||
externalBackendHealthCheckInterval, done);
|
|
||||||
});
|
|
||||||
it('should not refresh response before externalBackendHealthCheckInterval',
|
|
||||||
done => {
|
|
||||||
const clients = getClients(false);
|
|
||||||
return checkExternalBackend(clients, awsLocations, 'aws_s3',
|
|
||||||
false, externalBackendHealthCheckInterval, (err, res) => {
|
|
||||||
if (err) {
|
|
||||||
return done(err);
|
|
||||||
}
|
|
||||||
assert.strictEqual(res[0].awsbackend, statusSuccess);
|
|
||||||
return done();
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
it('should refresh response after externalBackendHealthCheckInterval',
|
|
||||||
done => {
|
|
||||||
const clients = getClients(false);
|
|
||||||
setTimeout(() => {
|
|
||||||
checkExternalBackend(clients, awsLocations, 'aws_s3',
|
|
||||||
false, externalBackendHealthCheckInterval, (err, res) => {
|
|
||||||
if (err) {
|
|
||||||
return done(err);
|
|
||||||
}
|
|
||||||
assert.strictEqual(res[0].awsbackend, statusFailure);
|
|
||||||
return done();
|
|
||||||
});
|
|
||||||
}, externalBackendHealthCheckInterval + 1);
|
|
||||||
});
|
|
||||||
});
|
|
|
@ -1,55 +0,0 @@
|
||||||
const assert = require('assert');
|
|
||||||
const proxyCompareUrl = require('../../../lib/data/proxyCompareUrl');
|
|
||||||
|
|
||||||
const testCases = [
|
|
||||||
{
|
|
||||||
endpoint: 'test.scality.com',
|
|
||||||
noProxy: '',
|
|
||||||
expRes: false,
|
|
||||||
desc: 'no NO_PROXY env var set',
|
|
||||||
},
|
|
||||||
{
|
|
||||||
endpoint: 'test.scality.com',
|
|
||||||
noProxy: 'test.*.com',
|
|
||||||
expRes: true,
|
|
||||||
desc: 'NO_PROXY matches with middle wildcard',
|
|
||||||
},
|
|
||||||
{
|
|
||||||
endpoint: 'test.scality.com',
|
|
||||||
noProxy: '*.com',
|
|
||||||
expRes: true,
|
|
||||||
desc: 'NO_PROXY matches with beginning wildcard',
|
|
||||||
},
|
|
||||||
{
|
|
||||||
endpoint: 'test.scality.com',
|
|
||||||
noProxy: '.scality.com',
|
|
||||||
expRes: true,
|
|
||||||
desc: 'NO_PROXY matches with beginning period',
|
|
||||||
},
|
|
||||||
{
|
|
||||||
endpoint: 'test.scality.com',
|
|
||||||
noProxy: 'test.nomatch,test.scality.*',
|
|
||||||
expRes: true,
|
|
||||||
desc: 'match with wildcard',
|
|
||||||
},
|
|
||||||
{
|
|
||||||
endpoint: 'test.scality.com',
|
|
||||||
noProxy: 'test.nomatch,no.scality.no,no.*.com,scality.com',
|
|
||||||
expRes: false,
|
|
||||||
desc: 'no match',
|
|
||||||
},
|
|
||||||
];
|
|
||||||
|
|
||||||
describe('proxyCompareURL util function', () => {
|
|
||||||
testCases.forEach(test => {
|
|
||||||
it(`should return ${test.expRes} if ${test.desc}`, () => {
|
|
||||||
process.env.NO_PROXY = test.noProxy;
|
|
||||||
const proxyMatch = proxyCompareUrl(test.endpoint);
|
|
||||||
assert.strictEqual(test.expRes, proxyMatch);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
after(() => {
|
|
||||||
process.env.NO_PROXY = '';
|
|
||||||
});
|
|
||||||
});
|
|
Loading…
Reference in New Issue