Compare commits

...

28 Commits

Author SHA1 Message Date
Anurag Mittal 16973fcbd9
updated acc route 2024-07-04 13:11:10 +02:00
Anurag Mittal 368283c1f2
updated api.js 2024-07-04 11:01:56 +02:00
Anurag Mittal 4e6a2eac6f
fixed crashes 2024-07-04 01:42:53 +02:00
Anurag Mittal e157222ea7
added events for active span 2024-07-01 14:23:59 +02:00
Anurag Mittal 716c982b36
updated arsenal 2024-07-01 14:19:21 +02:00
Anurag Mittal 785b893305
added more details to authV4 2024-07-01 13:53:45 +02:00
Anurag Mittal f943a5ffa7
added auth span details 2024-06-28 18:09:57 +02:00
Anurag Mittal 7466d488ea
updated sampling to 5 % 2024-06-28 09:40:24 +02:00
Anurag Mittal d311934fe7
updated api names and added ratio based instrumentation 2024-06-28 09:08:06 +02:00
Anurag Mittal 308b2176da
updated server name 2024-06-26 01:53:28 +02:00
Anurag Mittal 4e2c8a01cb
updated get object 2024-06-26 01:24:56 +02:00
Anurag Mittal 51880c4cf4
updated span names 2024-06-26 00:38:21 +02:00
Anurag Mittal 9e6394d25a
updated spans in arsenal 2024-06-26 00:01:34 +02:00
Anurag Mittal 2e7ae3a84b
updated span names and events for callApiMethod 2024-06-25 23:40:47 +02:00
Anurag Mittal ec67fd9fdd
S3c-8896 done with DeleteObject 2024-06-19 03:02:23 +02:00
Anurag Mittal 396665daa8
fixed all apis 2024-06-19 02:50:42 +02:00
Anurag Mittal 91d45c2fdf
S3C-8896: headobject 2024-06-19 01:56:32 +02:00
Anurag Mittal b962782e9f
fixup for list buckets 2024-06-19 00:28:24 +02:00
Anurag Mittal 9fcdd567c8
S3C-8896: PutObject complete 2024-06-19 00:27:44 +02:00
Anurag Mittal e02e70a33e
debugging internal MD handlers 2024-06-18 17:18:50 +02:00
Anurag Mittal 8af82991cd
done with callApiMethod 2024-06-18 15:22:13 +02:00
Anurag Mittal 1b5afd9ce7
fixup for callApiMethod 2024-06-18 09:59:16 +02:00
Anurag Mittal 4f677a1797
S3C-8896: added instrumentation to callApiMethod 2024-06-18 02:54:10 +02:00
Anurag Mittal a3d54d4b61
S3C-8896: added-auto-instrumentation-for-utapi-reindex-and-replay 2024-06-17 20:47:55 +02:00
Anurag Mittal 785fbb3ee4
S3C-8896-GetObject-instrumention-with-data-backend 2024-06-17 20:44:20 +02:00
Anurag Mittal 7d8529de75
fixup 2024-06-17 20:17:40 +02:00
Anurag Mittal cddfb8d84b
S3C-8896: added access to tracer and span in server for each request 2024-06-17 18:12:33 +02:00
Anurag Mittal 4ba88274c4
S3C-8896-added-OTel-basic-instrumentation 2024-06-17 17:21:59 +02:00
35 changed files with 3503 additions and 1254 deletions

28
customSampler.js Normal file
View File

@ -0,0 +1,28 @@
const { ParentBasedSampler, TraceIdRatioBasedSampler, SamplingDecision } = require('@opentelemetry/sdk-trace-base');
class HealthcheckExcludingSampler {
constructor(sampler, excludeHealthcheck) {
this._sampler = sampler;
this._excludeHealthcheck = excludeHealthcheck;
}
shouldSample(context, traceId, spanName, spanKind, attributes, links) {
const url = attributes['http.url'] || '';
if (this._excludeHealthcheck && (url.includes('healthcheck') || url.includes('metrics'))) {
return { decision: SamplingDecision.NOT_RECORD };
}
return this._sampler.shouldSample(context, traceId, spanName, spanKind, attributes, links);
}
toString() {
return `HealthcheckExcludingSampler{${this._sampler.toString()}}`;
}
}
function createSampler(samplingRatio, excludeHealthcheck) {
return new ParentBasedSampler({
root: new HealthcheckExcludingSampler(new TraceIdRatioBasedSampler(samplingRatio), excludeHealthcheck),
});
}
module.exports = { createSampler };

69
instrumentation.js Normal file
View File

@ -0,0 +1,69 @@
const opentelemetry = require('@opentelemetry/sdk-node');
const { Resource } = require('@opentelemetry/resources');
const { PeriodicExportingMetricReader } = require('@opentelemetry/sdk-metrics');
const { getNodeAutoInstrumentations } = require('@opentelemetry/auto-instrumentations-node');
const { OTLPMetricExporter } = require('@opentelemetry/exporter-metrics-otlp-proto');
const { OTLPTraceExporter } = require('@opentelemetry/exporter-trace-otlp-proto');
const { IORedisInstrumentation } = require('@opentelemetry/instrumentation-ioredis');
const { createSampler } = require('./customSampler');
const {
SEMRESATTRS_SERVICE_NAME,
SEMRESATTRS_SERVICE_VERSION,
} = require('@opentelemetry/semantic-conventions');
// Define resource with service name and version
const resource = new Resource({
[SEMRESATTRS_SERVICE_NAME]: process.env.OTEL_SERVICE_NAME || 'cisco-s3-cloudserver',
[SEMRESATTRS_SERVICE_VERSION]: '7.70.47',
});
const collectorHost = process.env.OTEL_COLLECTOR_HOST || 'localhost';
const collectorPort = process.env.OTEL_COLLECTOR_PORT || 4318;
const samplingRatio = parseFloat(process.env.OTEL_SAMPLING_RATIO) || 0.05;
const excludeHealthcheck = process.env.OTEL_EXCLUDE_HEALTHCHECK || 'true';
// OTLP Trace Exporter configuration
const traceExporter = new OTLPTraceExporter({
url: `http://${collectorHost}:${collectorPort}/v1/traces`,
headers: {},
});
// Metric Reader configuration
const metricReader = new PeriodicExportingMetricReader({
exporter: new OTLPMetricExporter({
url: `http://${collectorHost}:${collectorPort}/v1/metrics`,
headers: {},
concurrencyLimit: 1,
}),
});
// Node SDK configuration
const sdk = new opentelemetry.NodeSDK({
traceExporter,
resource,
metricReader,
// sampler: new TraceIdRatioBasedSampler(samplingRatio),
sampler: createSampler(samplingRatio, excludeHealthcheck),
instrumentations: [
getNodeAutoInstrumentations({
'@opentelemetry/instrumentation-fs': {
enabled: false,
},
'@opentelemetry/instrumentation-http': {
responseHook: (span, operations) => {
span.updateName(
`${operations.req.protocol} ${operations.req.method} ${operations.req.path.split('&')[0]}`);
},
},
}),
new IORedisInstrumentation({
requestHook: (span, { cmdName, cmdArgs }) => {
span.updateName(`Redis:: ${cmdName.toUpperCase()} cache operation for ${cmdArgs[0].split(':')[0]}`);
},
}),
],
});
// Start the Node SDK
sdk.start();

View File

@ -1,5 +1,6 @@
const { auth, errors, policies } = require('arsenal');
const async = require('async');
const opentelemetry = require('@opentelemetry/api');
const bucketDelete = require('./bucketDelete');
const bucketDeleteCors = require('./bucketDeleteCors');
@ -75,215 +76,297 @@ auth.setHandler(vault);
/* eslint-disable no-param-reassign */
const api = {
callApiMethod(apiMethod, request, response, log, callback) {
// Attach the apiMethod method to the request, so it can used by monitoring in the server
// eslint-disable-next-line no-param-reassign
request.apiMethod = apiMethod;
const actionLog = monitoringMap[apiMethod];
if (!actionLog &&
apiMethod !== 'websiteGet' &&
apiMethod !== 'websiteHead' &&
apiMethod !== 'corsPreflight') {
log.error('callApiMethod(): No actionLog for this api method', {
apiMethod,
callApiMethod(apiMethod, request, response, log, callback, oTel) {
const action = monitoringMap[apiMethod];
const {
cloudserverApiSpan,
activeSpan,
activeTracerContext,
tracer,
} = oTel;
activeSpan.addEvent(`Initiating Cloudserver API Method ${apiMethod}`);
// cloudserverApiSpan.updateName(`Performing operations to satisfy ${action} API request. Cloudserver API method: ${apiMethod}()`);
cloudserverApiSpan.updateName(`Executing ${action} API Method`);
// TODO: Try out activeTracerContext instead of opentelemetry.context.active()
const cloudserverApiSpanContext = opentelemetry.trace.setSpan(
activeTracerContext,
cloudserverApiSpan,
);
return tracer.startActiveSpan(' Setting Up Request Headers and Context for Authentication and Authorization', undefined, cloudserverApiSpanContext, callApiMethod => {
callApiMethod.setAttributes({
'code.function': 'callApiMethod()',
'code.filename': 'lib/api/api.js',
'code.lineno': 79,
'rpc.method': action,
});
}
log.addDefaultFields({
service: 's3',
action: actionLog,
bucketName: request.bucketName,
});
if (request.objectKey) {
// Attach the apiMethod method to the request, so it can used by monitoring in the server
// eslint-disable-next-line no-param-reassign
request.apiMethod = apiMethod;
const actionLog = monitoringMap[apiMethod];
if (!actionLog &&
apiMethod !== 'websiteGet' &&
apiMethod !== 'websiteHead' &&
apiMethod !== 'corsPreflight') {
log.error('callApiMethod(): No actionLog for this api method', {
apiMethod,
});
}
log.addDefaultFields({
objectKey: request.objectKey,
service: 's3',
action: actionLog,
bucketName: request.bucketName,
});
}
let returnTagCount = true;
const validationRes = validateQueryAndHeaders(request, log);
if (validationRes.error) {
log.debug('request query / header validation failed', {
error: validationRes.error,
method: 'api.callApiMethod',
});
return process.nextTick(callback, validationRes.error);
}
// no need to check auth on website or cors preflight requests
if (apiMethod === 'websiteGet' || apiMethod === 'websiteHead' ||
apiMethod === 'corsPreflight') {
request.actionImplicitDenies = false;
return this[apiMethod](request, log, callback);
}
const { sourceBucket, sourceObject, sourceVersionId, parsingError } =
parseCopySource(apiMethod, request.headers['x-amz-copy-source']);
if (parsingError) {
log.debug('error parsing copy source', {
error: parsingError,
});
return process.nextTick(callback, parsingError);
}
const { httpHeadersSizeError } = checkHttpHeadersSize(request.headers);
if (httpHeadersSizeError) {
log.debug('http header size limit exceeded', {
error: httpHeadersSizeError,
});
return process.nextTick(callback, httpHeadersSizeError);
}
const requestContexts = prepareRequestContexts(apiMethod, request,
sourceBucket, sourceObject, sourceVersionId);
// Extract all the _apiMethods and store them in an array
const apiMethods = requestContexts ? requestContexts.map(context => context._apiMethod) : [];
// Attach the names to the current request
// eslint-disable-next-line no-param-reassign
request.apiMethods = apiMethods;
function checkAuthResults(authResults) {
if (request.objectKey) {
log.addDefaultFields({
objectKey: request.objectKey,
});
}
let returnTagCount = true;
const isImplicitDeny = {};
let isOnlyImplicitDeny = true;
if (apiMethod === 'objectGet') {
// first item checks s3:GetObject(Version) action
if (!authResults[0].isAllowed && !authResults[0].isImplicit) {
log.trace('get object authorization denial from Vault');
return errors.AccessDenied;
}
// TODO add support for returnTagCount in the bucket policy
// checks
isImplicitDeny[authResults[0].action] = authResults[0].isImplicit;
// second item checks s3:GetObject(Version)Tagging action
if (!authResults[1].isAllowed) {
log.trace('get tagging authorization denial ' +
'from Vault');
returnTagCount = false;
}
} else {
for (let i = 0; i < authResults.length; i++) {
isImplicitDeny[authResults[i].action] = true;
if (!authResults[i].isAllowed && !authResults[i].isImplicit) {
// Any explicit deny rejects the current API call
log.trace('authorization denial from Vault');
const validationRes = validateQueryAndHeaders(request, log);
if (validationRes.error) {
activeSpan.recordException(validationRes.error);
callApiMethod.end();
log.debug('request query / header validation failed', {
error: validationRes.error,
method: 'api.callApiMethod',
});
return process.nextTick(callback, validationRes.error);
}
// no need to check auth on website or cors preflight requests
if (apiMethod === 'websiteGet' || apiMethod === 'websiteHead' ||
apiMethod === 'corsPreflight') {
callApiMethod.end();
activeSpan.addEvent(`Forwarding Request to ${apiMethod} Handler`);
request.actionImplicitDenies = false;
return this[apiMethod](request, log, callback);
}
const { sourceBucket, sourceObject, sourceVersionId, parsingError } =
parseCopySource(apiMethod, request.headers['x-amz-copy-source']);
if (parsingError) {
activeSpan.recordException(parsingError);
callApiMethod.end();
log.debug('error parsing copy source', {
error: parsingError,
});
return process.nextTick(callback, parsingError);
}
const { httpHeadersSizeError } = checkHttpHeadersSize(request.headers);
if (httpHeadersSizeError) {
activeSpan.recordException(httpHeadersSizeError);
callApiMethod.end();
log.debug('http header size limit exceeded', {
error: httpHeadersSizeError,
});
return process.nextTick(callback, httpHeadersSizeError);
}
activeSpan.addEvent('Initializing Authentication and Authorization Request Contexts');
const requestContexts = prepareRequestContexts(apiMethod, request,
sourceBucket, sourceObject, sourceVersionId);
activeSpan.addEvent('Authentication and Authorization Request Contexts Initialized');
// Extract all the _apiMethods and store them in an array
const apiMethods = requestContexts ? requestContexts.map(context => context._apiMethod) : [];
// Attach the names to the current request
// eslint-disable-next-line no-param-reassign
request.apiMethods = apiMethods;
function checkAuthResults(authResults) {
let returnTagCount = true;
const isImplicitDeny = {};
let isOnlyImplicitDeny = true;
if (apiMethod === 'objectGet') {
// first item checks s3:GetObject(Version) action
if (!authResults[0].isAllowed && !authResults[0].isImplicit) {
activeSpan.recordException(errors.AccessDenied);
log.trace('get object authorization denial from Vault');
return errors.AccessDenied;
}
if (authResults[i].isAllowed) {
// If the action is allowed, the result is not implicit
// Deny.
isImplicitDeny[authResults[i].action] = false;
isOnlyImplicitDeny = false;
// TODO add support for returnTagCount in the bucket policy
// checks
isImplicitDeny[authResults[0].action] = authResults[0].isImplicit;
// second item checks s3:GetObject(Version)Tagging action
if (!authResults[1].isAllowed) {
log.trace('get tagging authorization denial ' +
'from Vault');
returnTagCount = false;
}
} else {
for (let i = 0; i < authResults.length; i++) {
isImplicitDeny[authResults[i].action] = true;
if (!authResults[i].isAllowed && !authResults[i].isImplicit) {
activeSpan.recordException(errors.AccessDenied);
// Any explicit deny rejects the current API call
log.trace('authorization denial from Vault');
return errors.AccessDenied;
}
if (authResults[i].isAllowed) {
// If the action is allowed, the result is not implicit
// Deny.
isImplicitDeny[authResults[i].action] = false;
isOnlyImplicitDeny = false;
}
}
}
// These two APIs cannot use ACLs or Bucket Policies, hence, any
// implicit deny from vault must be treated as an explicit deny.
if ((apiMethod === 'bucketPut' || apiMethod === 'serviceGet') && isOnlyImplicitDeny) {
activeSpan.recordException(errors.AccessDenied);
return errors.AccessDenied;
}
return { returnTagCount, isImplicitDeny };
}
// These two APIs cannot use ACLs or Bucket Policies, hence, any
// implicit deny from vault must be treated as an explicit deny.
if ((apiMethod === 'bucketPut' || apiMethod === 'serviceGet') && isOnlyImplicitDeny) {
return errors.AccessDenied;
}
return { returnTagCount, isImplicitDeny };
}
return async.waterfall([
next => auth.server.doAuth(
request, log, (err, userInfo, authorizationResults, streamingV4Params) => {
if (err) {
log.trace('authentication error', { error: err });
return next(err);
}
return next(null, userInfo, authorizationResults, streamingV4Params);
}, 's3', requestContexts),
(userInfo, authorizationResults, streamingV4Params, next) => {
const authNames = { accountName: userInfo.getAccountDisplayName() };
if (userInfo.isRequesterAnIAMUser()) {
authNames.userName = userInfo.getIAMdisplayName();
}
log.addDefaultFields(authNames);
if (apiMethod === 'objectPut' || apiMethod === 'objectPutPart') {
return next(null, userInfo, authorizationResults, streamingV4Params);
}
// issue 100 Continue to the client
writeContinue(request, response);
const MAX_POST_LENGTH = request.method === 'POST' ?
1024 * 1024 : 1024 * 1024 / 2; // 1 MB or 512 KB
const post = [];
let postLength = 0;
request.on('data', chunk => {
postLength += chunk.length;
// Sanity check on post length
if (postLength <= MAX_POST_LENGTH) {
post.push(chunk);
}
});
request.on('error', err => {
log.trace('error receiving request', {
error: err,
callApiMethod.end();
activeSpan.addEvent('Starting User Authentication and Authorization');
return async.waterfall([
next => tracer.startActiveSpan('Authenticating and Authorizing the Request with Vault', undefined, cloudserverApiSpanContext, iamAuthSpan => {
const iamAuthSpanContext = opentelemetry.trace.setSpan(
cloudserverApiSpanContext,
iamAuthSpan,
);
activeSpan.addEvent('Starting Vault Authentication and Authorization');
iamAuthSpan.setAttributes({
'code.function': 'callApiMethod()',
'code.filename': 'lib/api/api.js',
'code.lineno': 220,
'rpc.method': 'AuthV4',
});
return auth.server.doAuth(
request, log, (err, userInfo, authorizationResults, streamingV4Params) => {
activeSpan.addEvent('Received Response from Vault');
iamAuthSpan.end();
if (err) {
activeSpan.recordException(err);
log.trace('authentication error', { error: err });
return next(err);
}
return next(null, userInfo, authorizationResults, streamingV4Params);
}, 's3', requestContexts, { activeSpan, activeTracerContext: iamAuthSpanContext, tracer });
}),
(userInfo, authorizationResults, streamingV4Params, next) => tracer.startActiveSpan('Process Vault Response', undefined, cloudserverApiSpanContext, currentSpan => {
currentSpan.setAttributes({
'code.function': 'callApiMethod()',
'code.filename': 'lib/api/api.js',
'code.lineno': 220,
});
activeSpan.addEvent('Processing Vault Response');
const authNames = { accountName: userInfo.getAccountDisplayName() };
if (userInfo.isRequesterAnIAMUser()) {
authNames.userName = userInfo.getIAMdisplayName();
}
log.addDefaultFields(authNames);
if (apiMethod === 'objectPut' || apiMethod === 'objectPutPart') {
currentSpan.end();
return next(null, userInfo, authorizationResults, streamingV4Params);
}
// issue 100 Continue to the client
writeContinue(request, response);
const MAX_POST_LENGTH = request.method === 'POST' ?
1024 * 1024 : 1024 * 1024 / 2; // 1 MB or 512 KB
const post = [];
let postLength = 0;
request.on('data', chunk => {
postLength += chunk.length;
// Sanity check on post length
if (postLength <= MAX_POST_LENGTH) {
post.push(chunk);
}
});
return next(errors.InternalError);
});
request.on('end', () => {
if (postLength > MAX_POST_LENGTH) {
log.error('body length is too long for request type',
{ postLength });
return next(errors.InvalidRequest);
}
// Convert array of post buffers into one string
request.post = Buffer.concat(post, postLength).toString();
return next(null, userInfo, authorizationResults, streamingV4Params);
});
return undefined;
},
// Tag condition keys require information from CloudServer for evaluation
(userInfo, authorizationResults, streamingV4Params, next) => tagConditionKeyAuth(
authorizationResults,
request,
requestContexts,
apiMethod,
log,
(err, authResultsWithTags) => {
if (err) {
log.trace('tag authentication error', { error: err });
return next(err);
}
return next(null, userInfo, authResultsWithTags, streamingV4Params);
},
),
], (err, userInfo, authorizationResults, streamingV4Params) => {
if (err) {
return callback(err);
}
if (authorizationResults) {
const checkedResults = checkAuthResults(authorizationResults);
if (checkedResults instanceof Error) {
return callback(checkedResults);
request.on('error', err => {
activeSpan.recordException(err);
currentSpan.end();
log.trace('error receiving request', {
error: err,
});
return next(errors.InternalError);
});
request.on('end', () => {
if (postLength > MAX_POST_LENGTH) {
activeSpan.recordException(errors.InvalidRequest);
currentSpan.end();
log.error('body length is too long for request type',
{ postLength });
return next(errors.InvalidRequest);
}
// Convert array of post buffers into one string
request.post = Buffer.concat(post, postLength).toString();
activeSpan.addEvent('Vault Response Processed');
return next(null, userInfo, authorizationResults, streamingV4Params);
});
return undefined;
}),
// Tag condition keys require information from CloudServer for evaluation
(userInfo, authorizationResults, streamingV4Params, next) => tracer.startActiveSpan('Validate Tag condition keys for Authorization', undefined, cloudserverApiSpanContext, currentSpan => {
activeSpan.addEvent('Validating Tag Conditions in Authorization');
return tagConditionKeyAuth(
authorizationResults,
request,
requestContexts,
apiMethod,
log,
(err, authResultsWithTags) => {
currentSpan.end();
if (err) {
activeSpan.recordException(err);
log.trace('tag authentication error', { error: err });
return next(err);
}
return next(null, userInfo, authResultsWithTags, streamingV4Params);
},
);
}),
], (err, userInfo, authorizationResults, streamingV4Params) => {
activeSpan.addEvent('Vault Authentication and Authorization Completed');
if (err) {
activeSpan.recordException(err);
return callback(err);
}
returnTagCount = checkedResults.returnTagCount;
request.actionImplicitDenies = checkedResults.isImplicitDeny;
} else {
// create an object of keys apiMethods with all values to false:
// for backward compatibility, all apiMethods are allowed by default
// thus it is explicitly allowed, so implicit deny is false
request.actionImplicitDenies = apiMethods.reduce((acc, curr) => {
acc[curr] = false;
return acc;
}, {});
}
if (apiMethod === 'objectPut' || apiMethod === 'objectPutPart') {
request._response = response;
return this[apiMethod](userInfo, request, streamingV4Params,
log, callback, authorizationResults);
}
if (apiMethod === 'objectCopy' || apiMethod === 'objectPutCopyPart') {
return this[apiMethod](userInfo, request, sourceBucket,
sourceObject, sourceVersionId, log, callback);
}
if (apiMethod === 'objectGet') {
return this[apiMethod](userInfo, request, returnTagCount, log, callback);
}
return this[apiMethod](userInfo, request, log, callback);
if (authorizationResults) {
activeSpan.addEvent('Verifying Authorization Results');
const checkedResults = checkAuthResults(authorizationResults);
if (checkedResults instanceof Error) {
activeSpan.recordException(checkedResults);
return callback(checkedResults);
}
activeSpan.addEvent('Authorization Results Verified');
returnTagCount = checkedResults.returnTagCount;
request.actionImplicitDenies = checkedResults.isImplicitDeny;
} else {
// create an object of keys apiMethods with all values to false:
// for backward compatibility, all apiMethods are allowed by default
// thus it is explicitly allowed, so implicit deny is false
request.actionImplicitDenies = apiMethods.reduce((acc, curr) => {
acc[curr] = false;
return acc;
}, {});
}
cloudserverApiSpan.setAttributes({
'code.lineno': 95,
'code.filename': 'lib/api/api.js',
'code.function': `callApiMethod().${apiMethod}()`,
});
activeSpan.setAttributes({
'aws.s3.apiMethod': apiMethod,
});
activeSpan.addEvent(`Forwarding Request to ${apiMethod} Handler`);
if (apiMethod === 'objectPut' || apiMethod === 'objectPutPart') {
request._response = response;
return this[apiMethod](userInfo, request, streamingV4Params,
log, callback, authorizationResults, oTel);
}
if (apiMethod === 'objectCopy' || apiMethod === 'objectPutCopyPart') {
return this[apiMethod](userInfo, request, sourceBucket,
sourceObject, sourceVersionId, log, callback, oTel);
}
if (apiMethod === 'objectGet') {
return this[apiMethod](userInfo, request, returnTagCount, log, callback, oTel);
}
return this[apiMethod](userInfo, request, log, callback, oTel);
});
});
},
bucketDelete,

View File

@ -8,7 +8,7 @@ const { standardMetadataValidateBucketAndObj } =
const services = require('../../../services');
function abortMultipartUpload(authInfo, bucketName, objectKey, uploadId, log,
callback, request) {
callback, request, oTel) {
const metadataValMPUparams = {
authInfo,
bucketName,
@ -42,7 +42,7 @@ function abortMultipartUpload(authInfo, bucketName, objectKey, uploadId, log,
'bucketPolicyGoAhead';
}
return next(null, destinationBucket);
});
}, oTel);
},
function checkMPUval(destBucket, next) {
metadataValParams.log = log;

View File

@ -1,5 +1,6 @@
const async = require('async');
const { errors, s3middleware } = require('arsenal');
const opentelemetry = require('@opentelemetry/api');
const getMetaHeaders = s3middleware.userMetadata.getMetaHeaders;
const constants = require('../../../../constants');
@ -46,20 +47,62 @@ function _checkAndApplyZenkoMD(metaHeaders, request, isDeleteMarker) {
}
function _storeInMDandDeleteData(bucketName, dataGetInfo, cipherBundle,
metadataStoreParams, dataToDelete, log, requestMethod, callback) {
services.metadataStoreObject(bucketName, dataGetInfo,
cipherBundle, metadataStoreParams, (err, result) => {
if (err) {
return callback(err);
}
metadataStoreParams, dataToDelete, log, requestMethod, callback, oTel) {
const { activeSpan, activeTracerContext, tracer } = oTel;
return async.waterfall([
next => tracer.startActiveSpan('Update Metadata for the stored object', undefined, activeTracerContext, currentSpan => {
currentSpan.setAttributes({
'code.function': '_storeInMDandDeleteData.storeMetadata()',
'code.filename': 'lib/api/apiUtils/object/createAndStoreObject.js',
'code.lineno': 62,
});
return next(null, currentSpan);
}),
(currentSpan, next) => {
const context = opentelemetry.trace.setSpan(
activeTracerContext,
currentSpan,
);
services.metadataStoreObject(bucketName, dataGetInfo, cipherBundle, metadataStoreParams, (err, result) => {
if (err) {
activeSpan.recordException(err);
currentSpan.end();
return next(err);
}
return next(null, result, currentSpan);
}, { activeSpan, activeTracerContext: context, tracer });
},
(result, currentSpan, next) => {
activeSpan.addEvent('Stored Metadata');
currentSpan.end();
return next(null, result);
},
(result, next) => tracer.startActiveSpan('Batch delete Old Data if Necessary', undefined, activeTracerContext, currentSpan => {
currentSpan.setAttributes({
'code.function': '_storeInMDandDeleteData.deleteOldData()',
'code.filename': 'lib/api/apiUtils/object/createAndStoreObject.js',
'code.lineno': 80,
});
return next(null, result, currentSpan);
}),
(result, currentSpan, next) => {
if (dataToDelete) {
const newDataStoreName = Array.isArray(dataGetInfo) ?
dataGetInfo[0].dataStoreName : null;
return data.batchDelete(dataToDelete, requestMethod,
newDataStoreName, log, err => callback(err, result));
activeSpan.addEvent('Batch delete Old Data');
const newDataStoreName = Array.isArray(dataGetInfo) ? dataGetInfo[0].dataStoreName : null;
data.batchDelete(dataToDelete, requestMethod, newDataStoreName, log, err => {
next(err, result, currentSpan);
});
} else {
activeSpan.addEvent('No Old Data to Delete');
next(null, result, currentSpan);
}
return callback(null, result);
});
},
(result, currentSpan, next) => {
activeSpan.addEvent('Deleted Old Data');
currentSpan.end();
return next(null, result);
},
], callback);
}
/** createAndStoreObject - store data, store metadata, and delete old data
@ -78,14 +121,23 @@ function _storeInMDandDeleteData(bucketName, dataGetInfo, cipherBundle,
* credentialScope (to be used for streaming v4 auth if applicable)
* @param {(object|null)} overheadField - fields to be included in metadata overhead
* @param {RequestLogger} log - logger instance
* @param {function} callback - callback function
* @param {function} callback - callback function]
* @param {object} oTel - OpenTelemetry methods
* @return {undefined} and call callback with (err, result) -
* result.contentMD5 - content md5 of new object or version
* result.versionId - unencrypted versionId returned by metadata
*/
function createAndStoreObject(bucketName, bucketMD, objectKey, objMD, authInfo,
canonicalID, cipherBundle, request, isDeleteMarker, streamingV4Params,
overheadField, log, callback) {
overheadField, log, callback, oTel) {
const { activeSpan, activeTracerContext, tracer } = oTel;
const objectStorageParamSpan = tracer.startSpan('Validating Storage Request Parameters', undefined, activeTracerContext);
objectStorageParamSpan.setAttributes({
'code.function': 'createAndStoreObject()',
'code.filename': 'lib/api/apiUtils/object/createAndStoreObject.js',
'code.lineno': 87,
});
activeSpan.addEvent('Entered createAndStoreObject()');
const size = isDeleteMarker ? 0 : request.parsedContentLength;
// although the request method may actually be 'DELETE' if creating a
// delete marker, for our purposes we consider this to be a 'PUT'
@ -97,8 +149,11 @@ function createAndStoreObject(bucketName, bucketMD, objectKey, objMD, authInfo,
const err = errors.InvalidRedirectLocation;
log.debug('invalid x-amz-website-redirect-location' +
`value ${websiteRedirectHeader}`, { error: err });
activeSpan.recordException(err);
objectStorageParamSpan.end();
return callback(err);
}
activeSpan.addEvent('Validated x-amz-website-redirect-location Header');
const metaHeaders = isDeleteMarker ? [] : getMetaHeaders(request.headers);
if (metaHeaders instanceof Error) {
@ -106,11 +161,15 @@ function createAndStoreObject(bucketName, bucketMD, objectKey, objMD, authInfo,
error: metaHeaders,
method: 'createAndStoreObject',
});
activeSpan.recordException(metaHeaders);
objectStorageParamSpan.end();
return process.nextTick(() => callback(metaHeaders));
}
activeSpan.addEvent('Extracted User Provided Metadata Headers');
// if receiving a request from Zenko for a delete marker, we place a
// user-metadata field on the object
_checkAndApplyZenkoMD(metaHeaders, request, isDeleteMarker);
activeSpan.addEvent('Applied Zenko Metadata If Request From Zenko');
log.trace('meta headers', { metaHeaders, method: 'objectPut' });
const objectKeyContext = {
@ -140,6 +199,7 @@ function createAndStoreObject(bucketName, bucketMD, objectKey, objMD, authInfo,
overheadField,
log,
};
activeSpan.addEvent('Set Metadata Store Parameters And Set Object Key Context');
if (!isDeleteMarker) {
metadataStoreParams.contentType = request.headers['content-type'];
@ -157,6 +217,7 @@ function createAndStoreObject(bucketName, bucketMD, objectKey, objMD, authInfo,
metadataStoreParams.defaultRetention
= defaultObjectLockConfiguration;
}
activeSpan.addEvent('DeleteMarker Not True, Set Metadata Store Request Headers');
}
// if creating new delete marker and there is an existing object, copy
@ -166,21 +227,27 @@ function createAndStoreObject(bucketName, bucketMD, objectKey, objMD, authInfo,
request.headers[constants.objectLocationConstraintHeader] =
objMD[constants.objectLocationConstraintHeader];
metadataStoreParams.originOp = 's3:ObjectRemoved:DeleteMarkerCreated';
activeSpan.addEvent('DeleteMarker True And Object Metadata Preset, Set Metadata Store Request Headers');
}
const backendInfoObj =
locationConstraintCheck(request, null, bucketMD, log);
if (backendInfoObj.err) {
return process.nextTick(() => {
activeSpan.recordException(backendInfoObj.err);
objectStorageParamSpan.end();
callback(backendInfoObj.err);
});
}
activeSpan.addEvent('Checked Location Constraint');
const backendInfo = backendInfoObj.backendInfo;
const location = backendInfo.getControllingLocationConstraint();
const locationType = backendInfoObj.defaultedToDataBackend ? location :
config.getLocationConstraintType(location);
metadataStoreParams.dataStoreName = location;
activeSpan.addEvent('Set Backend Info and Location');
if (versioningNotImplBackends[locationType]) {
const vcfg = bucketMD.getVersioningConfiguration();
@ -190,32 +257,66 @@ function createAndStoreObject(bucketName, bucketMD, objectKey, objMD, authInfo,
log.debug(externalVersioningErrorMessage,
{ method: 'createAndStoreObject', error: errors.NotImplemented });
return process.nextTick(() => {
activeSpan.recordException(backendInfoObj.err);
objectStorageParamSpan.end(errors.NotImplemented.customizeDescription(
externalVersioningErrorMessage
));
callback(errors.NotImplemented.customizeDescription(
externalVersioningErrorMessage));
});
}
}
activeSpan.addEvent('Checked Versioning');
if (objMD && objMD.uploadId) {
metadataStoreParams.oldReplayId = objMD.uploadId;
}
activeSpan.addEvent('Checked for Upload ID');
/* eslint-disable camelcase */
const dontSkipBackend = externalBackends;
/* eslint-enable camelcase */
objectStorageParamSpan.end();
return async.waterfall([
function storeData(next) {
next => tracer.startActiveSpan('Updating object Data in RING Using sproxyd', undefined, activeTracerContext, currentSpan => {
currentSpan.setAttributes({
'code.function': 'createAndStoreObject.storeData()',
'code.filename': 'lib/api/objectPut.js',
'code.lineno': 246,
});
return next(null, currentSpan);
}),
function storeData(currentSpan, next) {
if (size === 0 && !dontSkipBackend[locationType]) {
metadataStoreParams.contentMD5 = constants.emptyFileMd5;
return next(null, null, null);
return next(null, null, null, currentSpan);
}
return dataStore(objectKeyContext, cipherBundle, request, size,
streamingV4Params, backendInfo, log, next);
streamingV4Params, backendInfo, log, (err, dataGetInfo, calculatedHash) => {
if (err) {
activeSpan.recordException(err);
currentSpan.end();
return next(err);
}
return next(null, dataGetInfo, calculatedHash, currentSpan);
});
},
function processDataResult(dataGetInfo, calculatedHash, next) {
(dataGetInfo, calculatedHash, currentSpan, next) => {
activeSpan.addEvent('Stored Data');
currentSpan.end();
return next(null, dataGetInfo, calculatedHash);
},
(dataGetInfo, calculatedHash, next) => tracer.startActiveSpan('Processing Data Storage Result', undefined, activeTracerContext, currentSpan => {
currentSpan.setAttributes({
'code.function': 'createAndStoreObject.processDataResult()',
'code.filename': 'lib/api/objectPut.js',
'code.lineno': 273,
});
return next(null, dataGetInfo, calculatedHash, currentSpan);
}),
function processDataResult(dataGetInfo, calculatedHash, currentSpan, next) {
if (dataGetInfo === null || dataGetInfo === undefined) {
return next(null, null);
return next(null, null, currentSpan);
}
// So that data retrieval information for MPU's and
// regular puts are stored in the same data structure,
@ -234,9 +335,22 @@ function createAndStoreObject(bucketName, bucketMD, objectKey, objMD, authInfo,
cipherBundle.cipheredDataKey;
}
metadataStoreParams.contentMD5 = calculatedHash;
return next(null, dataGetInfoArr);
return next(null, dataGetInfoArr, currentSpan);
},
function getVersioningInfo(infoArr, next) {
(infoArr, currentSpan, next) => {
activeSpan.addEvent('Processed Data Result');
currentSpan.end();
return next(null, infoArr);
},
(infoArr, next) => tracer.startActiveSpan('Fetching Versioning Information', undefined, activeTracerContext, currentSpan => {
currentSpan.setAttributes({
'code.function': 'createAndStoreObject.getVersioningInfo()',
'code.filename': 'lib/api/objectPut.js',
'code.lineno': 310,
});
return next(null, infoArr, currentSpan);
}),
function getVersioningInfo(infoArr, currentSpan, next) {
return versioningPreprocessing(bucketName, bucketMD,
metadataStoreParams.objectKey, objMD, log, (err, options) => {
if (err) {
@ -249,10 +363,27 @@ function createAndStoreObject(bucketName, bucketMD, objectKey, objMD, authInfo,
method: 'versioningPreprocessing',
});
}
return next(err, options, infoArr);
return next(err, options, infoArr, currentSpan);
});
},
function storeMDAndDeleteData(options, infoArr, next) {
(options, infoArr, currentSpan, next) => {
activeSpan.addEvent('Got Versioning Info');
currentSpan.end();
return next(null, options, infoArr);
},
(options, infoArr, next) => tracer.startActiveSpan('Updating Metadata and Deleting Old Data', undefined, activeTracerContext, currentSpan => {
currentSpan.setAttributes({
'code.function': 'createAndStoreObject.storeMDAndDeleteData()',
'code.filename': 'lib/api/objectPut.js',
'code.lineno': 339,
});
return next(null, options, infoArr, currentSpan);
}),
function storeMDAndDeleteData(options, infoArr, currentSpan, next) {
const context = opentelemetry.trace.setSpan(
activeTracerContext,
currentSpan,
);
metadataStoreParams.versionId = options.versionId;
metadataStoreParams.versioning = options.versioning;
metadataStoreParams.isNull = options.isNull;
@ -262,7 +393,20 @@ function createAndStoreObject(bucketName, bucketMD, objectKey, objMD, authInfo,
}
return _storeInMDandDeleteData(bucketName, infoArr,
cipherBundle, metadataStoreParams,
options.dataToDelete, log, requestMethod, next);
options.dataToDelete, log, requestMethod,
(err, result) => {
if (err) {
activeSpan.recordException(err);
currentSpan.end();
return next(err);
}
return next(null, result, currentSpan);
}, { activeSpan, activeTracerContext: context, tracer });
},
(result, currentSpan, next) => {
activeSpan.addEvent('Stored Metadata and Deleted Old Data');
currentSpan.end();
return next(null, result);
},
], callback);
}

View File

@ -273,13 +273,13 @@ function hasGovernanceBypassHeader(headers) {
* @param {function} cb - callback returns errors.AccessDenied if the authorization fails
* @returns {undefined} -
*/
function checkUserGovernanceBypass(request, authInfo, bucketMD, objectKey, log, cb) {
function checkUserGovernanceBypass(request, authInfo, bucketMD, objectKey, log, cb, oTel) {
log.trace(
'object in GOVERNANCE mode and is user, checking for attached policies',
{ method: 'checkUserPolicyGovernanceBypass' },
);
const authParams = auth.server.extractParams(request, log, 's3', request.query);
const authParams = auth.server.extractParams(request, log, 's3', request.query, oTel);
const ip = policies.requestUtils.getClientIp(request, config);
const requestContextParams = {
constantParams: {

View File

@ -1,5 +1,6 @@
const querystring = require('querystring');
const { errors, versioning, s3middleware } = require('arsenal');
const opentelemetry = require('@opentelemetry/api');
const constants = require('../../constants');
const services = require('../services');
@ -288,108 +289,145 @@ function handleResult(listParams, requestMaxKeys, encoding, authInfo,
* with either error code or xml response body
* @return {undefined}
*/
function bucketGet(authInfo, request, log, callback) {
const params = request.query;
const bucketName = request.bucketName;
const v2 = params['list-type'];
if (v2 !== undefined && Number.parseInt(v2, 10) !== 2) {
return callback(errors.InvalidArgument.customizeDescription('Invalid ' +
'List Type specified in Request'));
}
if (v2) {
log.addDefaultFields({
action: 'ListObjectsV2',
function bucketGet(authInfo, request, log, callback, oTel) {
const {
cloudserverApiSpan,
activeSpan,
activeTracerContext,
tracer,
} = oTel;
activeSpan.addEvent('bucketGet: bucketGet called');
const cloudserverApiSpanContext = opentelemetry.trace.setSpan(
activeTracerContext,
cloudserverApiSpan,
);
return tracer.startActiveSpan('ListObjectsOrVersions - Processing Bucket Get Request', undefined, cloudserverApiSpanContext, listObjectsSpan => {
activeSpan.addEvent('bucketGet: processing request');
listObjectsSpan.setAttributes({
'code.filepath': 'lib/api/bucketGet.js',
'code.function': 'bucketGet',
'code.lineno': 304,
});
} else if (params.versions !== undefined) {
log.addDefaultFields({
action: 'ListObjectVersions',
});
}
log.debug('processing request', { method: 'bucketGet' });
const encoding = params['encoding-type'];
if (encoding !== undefined && encoding !== 'url') {
monitoring.promMetrics(
'GET', bucketName, 400, 'listBucket');
return callback(errors.InvalidArgument.customizeDescription('Invalid ' +
'Encoding Method specified in Request'));
}
const requestMaxKeys = params['max-keys'] ?
Number.parseInt(params['max-keys'], 10) : 1000;
if (Number.isNaN(requestMaxKeys) || requestMaxKeys < 0) {
monitoring.promMetrics(
'GET', bucketName, 400, 'listBucket');
return callback(errors.InvalidArgument);
}
// AWS only returns 1000 keys even if max keys are greater.
// Max keys stated in response xml can be greater than actual
// keys returned.
const actualMaxKeys = Math.min(constants.listingHardLimit, requestMaxKeys);
const metadataValParams = {
authInfo,
bucketName,
requestType: request.apiMethods || 'bucketGet',
request,
};
const listParams = {
listingType: 'DelimiterMaster',
maxKeys: actualMaxKeys,
prefix: params.prefix,
};
if (params.delimiter) {
listParams.delimiter = params.delimiter;
}
if (v2) {
listParams.v2 = true;
listParams.startAfter = params['start-after'];
listParams.continuationToken =
decryptToken(params['continuation-token']);
listParams.fetchOwner = params['fetch-owner'] === 'true';
} else {
listParams.marker = params.marker;
}
standardMetadataValidateBucket(metadataValParams, request.actionImplicitDenies, log, (err, bucket) => {
const corsHeaders = collectCorsHeaders(request.headers.origin,
request.method, bucket);
if (err) {
log.debug('error processing request', { error: err });
const listObjectsSpanContext = opentelemetry.trace.setSpan(
cloudserverApiSpanContext,
listObjectsSpan,
);
const params = request.query;
const bucketName = request.bucketName;
const v2 = params['list-type'];
if (v2 !== undefined && Number.parseInt(v2, 10) !== 2) {
activeSpan.recordException(errors.InvalidArgument);
return callback(errors.InvalidArgument.customizeDescription('Invalid ' +
'List Type specified in Request'));
}
if (v2) {
activeSpan.setAttributes({
'aws.s3.apiMethod': 'bucketGet - ListObjectsV2',
});
log.addDefaultFields({
action: 'ListObjectsV2',
});
} else if (params.versions !== undefined) {
log.addDefaultFields({
action: 'ListObjectVersions',
});
activeSpan.setAttributes({
'aws.s3.apiMethod': 'bucketGet - ListObjectVersions',
});
}
log.debug('processing request', { method: 'bucketGet' });
const encoding = params['encoding-type'];
if (encoding !== undefined && encoding !== 'url') {
monitoring.promMetrics(
'GET', bucketName, err.code, 'listBucket');
return callback(err, null, corsHeaders);
'GET', bucketName, 400, 'listBucket');
activeSpan.recordException(errors.InvalidArgument);
return callback(errors.InvalidArgument.customizeDescription('Invalid ' +
'Encoding Method specified in Request'));
}
if (params.versions !== undefined) {
listParams.listingType = 'DelimiterVersions';
delete listParams.marker;
listParams.keyMarker = params['key-marker'];
listParams.versionIdMarker = params['version-id-marker'] ?
versionIdUtils.decode(params['version-id-marker']) : undefined;
const requestMaxKeys = params['max-keys'] ?
Number.parseInt(params['max-keys'], 10) : 1000;
if (Number.isNaN(requestMaxKeys) || requestMaxKeys < 0) {
monitoring.promMetrics(
'GET', bucketName, 400, 'listBucket');
activeSpan.recordException(errors.InvalidArgument);
return callback(errors.InvalidArgument);
}
if (!requestMaxKeys) {
const emptyList = {
CommonPrefixes: [],
Contents: [],
Versions: [],
IsTruncated: false,
};
return handleResult(listParams, requestMaxKeys, encoding, authInfo,
bucketName, emptyList, corsHeaders, log, callback);
// AWS only returns 1000 keys even if max keys are greater.
// Max keys stated in response xml can be greater than actual
// keys returned.
const actualMaxKeys = Math.min(constants.listingHardLimit, requestMaxKeys);
const metadataValParams = {
authInfo,
bucketName,
requestType: request.apiMethods || 'bucketGet',
request,
};
const listParams = {
listingType: 'DelimiterMaster',
maxKeys: actualMaxKeys,
prefix: params.prefix,
};
if (params.delimiter) {
listParams.delimiter = params.delimiter;
}
return services.getObjectListing(bucketName, listParams, log,
(err, list) => {
if (v2) {
listParams.v2 = true;
listParams.startAfter = params['start-after'];
listParams.continuationToken =
decryptToken(params['continuation-token']);
listParams.fetchOwner = params['fetch-owner'] === 'true';
} else {
listParams.marker = params.marker;
}
activeSpan.addEvent('bucketGet: processed request');
return standardMetadataValidateBucket(metadataValParams, request.actionImplicitDenies, log, (err, bucket) => {
const corsHeaders = collectCorsHeaders(request.headers.origin,
request.method, bucket);
if (err) {
log.debug('error processing request', { error: err });
monitoring.promMetrics(
'GET', bucketName, err.code, 'listBucket');
return callback(err, null, corsHeaders);
}
return handleResult(listParams, requestMaxKeys, encoding, authInfo,
bucketName, list, corsHeaders, log, callback);
});
if (params.versions !== undefined) {
listParams.listingType = 'DelimiterVersions';
delete listParams.marker;
listParams.keyMarker = params['key-marker'];
listParams.versionIdMarker = params['version-id-marker'] ?
versionIdUtils.decode(params['version-id-marker']) : undefined;
}
if (!requestMaxKeys) {
const emptyList = {
CommonPrefixes: [],
Contents: [],
Versions: [],
IsTruncated: false,
};
listObjectsSpan.end();
return handleResult(listParams, requestMaxKeys, encoding, authInfo,
bucketName, emptyList, corsHeaders, log, callback);
}
return services.getObjectListing(bucketName, listParams, log,
(err, list) => {
listObjectsSpan.end();
if (err) {
log.debug('error processing request', { error: err });
monitoring.promMetrics(
'GET', bucketName, err.code, 'listBucket');
return callback(err, null, corsHeaders);
}
activeSpan.setAttributes({
'aws.s3.objects.count': list.Contents ? list.Contents.length : 0,
// 'aws.s3.action': 'bucketGet',
});
return handleResult(listParams, requestMaxKeys, encoding, authInfo,
bucketName, list, corsHeaders, log, callback);
});
}, { activeSpan, activeTracerContext: listObjectsSpanContext, tracer });
});
return undefined;
}
module.exports = bucketGet;

View File

@ -197,7 +197,7 @@ function authBucketPut(authParams, bucketName, locationConstraint, request, auth
* @param {function} callback - callback to server
* @return {undefined}
*/
function bucketPut(authInfo, request, log, callback) {
function bucketPut(authInfo, request, log, callback, oTel) {
log.debug('processing request', { method: 'bucketPut' });
if (authInfo.isRequesterPublicUser()) {
@ -230,7 +230,7 @@ function bucketPut(authInfo, request, log, callback) {
return next(null, locationConstraint);
}
const authParams = auth.server.extractParams(request, log, 's3', request.query);
const authParams = auth.server.extractParams(request, log, 's3', request.query, oTel);
const requestConstantParams = authBucketPut(
authParams, bucketName, locationConstraint, request, authInfo
);

View File

@ -67,7 +67,7 @@ const REPLICATION_ACTION = 'MPU';
* @param {function} callback - callback to server
* @return {undefined}
*/
function completeMultipartUpload(authInfo, request, log, callback) {
function completeMultipartUpload(authInfo, request, log, callback, oTel) {
log.debug('processing request', { method: 'completeMultipartUpload' });
const bucketName = request.bucketName;
const objectKey = request.objectKey;
@ -119,7 +119,7 @@ function completeMultipartUpload(authInfo, request, log, callback) {
requestType: request.apiMethods || 'completeMultipartUpload',
request,
};
standardMetadataValidateBucketAndObj(metadataValParams, request.actionImplicitDenies, log, next);
standardMetadataValidateBucketAndObj(metadataValParams, request.actionImplicitDenies, log, next, oTel);
},
function validateMultipart(destBucket, objMD, next) {
if (objMD) {
@ -434,7 +434,7 @@ function completeMultipartUpload(authInfo, request, log, callback) {
return next(null, mpuBucket, keysToDelete, aggregateETag,
extraPartLocations, destinationBucket,
generatedVersionId);
});
}, oTel);
},
function deletePartsMetadata(mpuBucket, keysToDelete, aggregateETag,
extraPartLocations, destinationBucket, generatedVersionId, next) {

View File

@ -44,7 +44,7 @@ Sample xml response:
* @return {undefined} calls callback from router
* with err and result as arguments
*/
function initiateMultipartUpload(authInfo, request, log, callback) {
function initiateMultipartUpload(authInfo, request, log, callback, oTel) {
log.debug('processing request', { method: 'initiateMultipartUpload' });
const bucketName = request.bucketName;
const objectKey = request.objectKey;
@ -275,7 +275,7 @@ function initiateMultipartUpload(authInfo, request, log, callback) {
return next(error, corsHeaders);
}
return next(null, corsHeaders, destinationBucket);
}),
}, oTel),
(corsHeaders, destinationBucket, next) => {
if (destinationBucket.hasDeletedFlag() && accountCanonicalID !== destinationBucket.getOwner()) {
log.trace('deleted flag on bucket and request from non-owner account');

View File

@ -73,7 +73,7 @@ function buildXML(xmlParams, xml, encodingFn) {
* @param {function} callback - callback to server
* @return {undefined}
*/
function listParts(authInfo, request, log, callback) {
function listParts(authInfo, request, log, callback, oTel) {
log.debug('processing request', { method: 'listParts' });
const bucketName = request.bucketName;
@ -131,7 +131,7 @@ function listParts(authInfo, request, log, callback) {
'bucketPolicyGoAhead';
}
return next(null, destinationBucket);
});
}, oTel);
},
function waterfall2(destBucket, next) {
metadataValMPUparams.log = log;

View File

@ -1,4 +1,5 @@
const crypto = require('crypto');
const opentelemetry = require('@opentelemetry/api');
const async = require('async');
const { parseString } = require('xml2js');
@ -246,202 +247,222 @@ function initializeMultiObjectDeleteWithBatchingSupport(bucketName, inPlay, log,
* successfullyDeleted, totalContentLengthDeleted)
*/
function getObjMetadataAndDelete(authInfo, canonicalID, request,
bucketName, bucket, quietSetting, errorResults, inPlay, log, next) {
const successfullyDeleted = [];
let totalContentLengthDeleted = 0;
let numOfObjectsRemoved = 0;
const skipError = new Error('skip');
const objectLockedError = new Error('object locked');
let deleteFromStorage = [];
bucketName, bucket, quietSetting, errorResults, inPlay, log, next, oTel) {
const {
currentSpan,
activeSpan,
activeTracerContext,
tracer,
} = oTel;
return tracer.startActiveSpan('Get Object Metadata and Delete', undefined, activeTracerContext, getObjMetadataAndDeleteSpan => {
getObjMetadataAndDeleteSpan.setAttributes({
'code.function': 'getObjMetadataAndDelete()',
'code.filename': 'lib/api/multiObjectDelete.js',
'code.lineno': 221,
});
activeSpan.addEvent('Getting Object Metadata and Deleting Objects');
const successfullyDeleted = [];
let totalContentLengthDeleted = 0;
let numOfObjectsRemoved = 0;
const skipError = new Error('skip');
const objectLockedError = new Error('object locked');
let deleteFromStorage = [];
return async.waterfall([
callback => initializeMultiObjectDeleteWithBatchingSupport(bucketName, inPlay, log, callback),
(cache, callback) => async.forEachLimit(inPlay, config.multiObjectDeleteConcurrency, (entry, moveOn) => {
async.waterfall([
callback => callback(...decodeObjectVersion(entry, bucketName)),
// for obj deletes, no need to check acl's at object level
// (authority is at the bucket level for obj deletes)
(versionId, callback) => metadataUtils.metadataGetObject(bucketName, entry.key,
versionId, cache, log, (err, objMD) => callback(err, objMD, versionId)),
(objMD, versionId, callback) => {
if (!objMD) {
const verCfg = bucket.getVersioningConfiguration();
// To adhere to AWS behavior, create a delete marker
// if trying to delete an object that does not exist
// when versioning has been configured
if (verCfg && !entry.versionId) {
log.debug('trying to delete specific version ' +
'that does not exist');
return async.waterfall([
callback => initializeMultiObjectDeleteWithBatchingSupport(bucketName, inPlay, log, callback),
(cache, callback) => async.forEachLimit(inPlay, config.multiObjectDeleteConcurrency, (entry, moveOn) => {
async.waterfall([
callback => callback(...decodeObjectVersion(entry, bucketName)),
// for obj deletes, no need to check acl's at object level
// (authority is at the bucket level for obj deletes)
(versionId, callback) => metadataUtils.metadataGetObject(bucketName, entry.key,
versionId, cache, log, (err, objMD) => callback(err, objMD, versionId)),
(objMD, versionId, callback) => {
if (!objMD) {
const verCfg = bucket.getVersioningConfiguration();
// To adhere to AWS behavior, create a delete marker
// if trying to delete an object that does not exist
// when versioning has been configured
if (verCfg && !entry.versionId) {
log.debug('trying to delete specific version ' +
'that does not exist');
return callback(null, objMD, versionId);
}
// otherwise if particular key does not exist, AWS
// returns success for key so add to successfullyDeleted
// list and move on
successfullyDeleted.push({ entry });
return callback(skipError);
}
if (versionId && objMD.location &&
Array.isArray(objMD.location) && objMD.location[0]) {
// we need this information for data deletes to AWS
// eslint-disable-next-line no-param-reassign
objMD.location[0].deleteVersion = true;
}
return callback(null, objMD, versionId);
},
(objMD, versionId, callback) => {
// AWS only returns an object lock error if a version id
// is specified, else continue to create a delete marker
if (!versionId || !bucket.isObjectLockEnabled()) {
return callback(null, null, objMD, versionId);
}
const hasGovernanceBypass = hasGovernanceBypassHeader(request.headers);
if (hasGovernanceBypass && authInfo.isRequesterAnIAMUser()) {
return checkUserGovernanceBypass(request, authInfo, bucket, entry.key, log, error => {
if (error && error.is.AccessDenied) {
log.debug('user does not have BypassGovernanceRetention and object is locked',
{ error });
return callback(objectLockedError);
}
if (error) {
return callback(error);
}
return callback(null, hasGovernanceBypass, objMD, versionId);
});
}
return callback(null, hasGovernanceBypass, objMD, versionId);
},
(hasGovernanceBypass, objMD, versionId, callback) => {
// AWS only returns an object lock error if a version id
// is specified, else continue to create a delete marker
if (!versionId || !bucket.isObjectLockEnabled()) {
return callback(null, objMD, versionId);
}
// otherwise if particular key does not exist, AWS
// returns success for key so add to successfullyDeleted
// list and move on
successfullyDeleted.push({ entry });
return callback(skipError);
}
if (versionId && objMD.location &&
Array.isArray(objMD.location) && objMD.location[0]) {
// we need this information for data deletes to AWS
// eslint-disable-next-line no-param-reassign
objMD.location[0].deleteVersion = true;
}
return callback(null, objMD, versionId);
},
(objMD, versionId, callback) => {
// AWS only returns an object lock error if a version id
// is specified, else continue to create a delete marker
if (!versionId || !bucket.isObjectLockEnabled()) {
return callback(null, null, objMD, versionId);
}
const hasGovernanceBypass = hasGovernanceBypassHeader(request.headers);
if (hasGovernanceBypass && authInfo.isRequesterAnIAMUser()) {
return checkUserGovernanceBypass(request, authInfo, bucket, entry.key, log, error => {
if (error && error.is.AccessDenied) {
log.debug('user does not have BypassGovernanceRetention and object is locked',
{ error });
return callback(objectLockedError);
}
if (error) {
return callback(error);
}
return callback(null, hasGovernanceBypass, objMD, versionId);
const objLockInfo = new ObjectLockInfo({
mode: objMD.retentionMode,
date: objMD.retentionDate,
legalHold: objMD.legalHold || false,
});
}
return callback(null, hasGovernanceBypass, objMD, versionId);
},
(hasGovernanceBypass, objMD, versionId, callback) => {
// AWS only returns an object lock error if a version id
// is specified, else continue to create a delete marker
if (!versionId || !bucket.isObjectLockEnabled()) {
// If the object can not be deleted raise an error
if (!objLockInfo.canModifyObject(hasGovernanceBypass)) {
log.debug('trying to delete locked object');
return callback(objectLockedError);
}
return callback(null, objMD, versionId);
},
(objMD, versionId, callback) => {
const options = preprocessingVersioningDelete(
bucketName, bucket, objMD, versionId, config.nullVersionCompatMode);
const deleteInfo = {};
if (options && options.deleteData) {
options.overheadField = overheadField;
deleteInfo.deleted = true;
if (!_bucketRequiresOplogUpdate(bucket)) {
options.doesNotNeedOpogUpdate = true;
}
if (objMD.uploadId) {
// eslint-disable-next-line
options.replayId = objMD.uploadId;
}
return services.deleteObject(bucketName, objMD,
entry.key, options, config.multiObjectDeleteEnableOptimizations, log, (err, toDelete) => {
if (err) {
return callback(err);
}
if (toDelete) {
deleteFromStorage = deleteFromStorage.concat(toDelete);
}
return callback(null, objMD, deleteInfo);
});
}
deleteInfo.newDeleteMarker = true;
// This call will create a delete-marker
return createAndStoreObject(bucketName, bucket, entry.key,
objMD, authInfo, canonicalID, null, request,
deleteInfo.newDeleteMarker, null, overheadField, log, (err, result) =>
callback(err, objMD, deleteInfo, result.versionId), oTel);
},
], (err, objMD, deleteInfo, versionId) => {
if (err === skipError) {
return moveOn();
} else if (err === objectLockedError) {
errorResults.push({ entry, error: errors.AccessDenied, objectLocked: true });
return moveOn();
} else if (err) {
log.error('error deleting object', { error: err, entry });
errorResults.push({ entry, error: err });
return moveOn();
}
const objLockInfo = new ObjectLockInfo({
mode: objMD.retentionMode,
date: objMD.retentionDate,
legalHold: objMD.legalHold || false,
if (deleteInfo.deleted && objMD['content-length']) {
numOfObjectsRemoved++;
totalContentLengthDeleted += objMD['content-length'];
}
let isDeleteMarker;
let deleteMarkerVersionId;
// - If trying to delete an object that does not exist (if a new
// delete marker was created)
// - Or if an object exists but no version was specified
// return DeleteMarkerVersionId equals the versionID of the marker
// you just generated and DeleteMarker tag equals true
if (deleteInfo.newDeleteMarker) {
isDeleteMarker = true;
deleteMarkerVersionId = versionIdUtils.encode(
versionId, config.versionIdEncodingType);
// In this case we are putting a new object (i.e., the delete
// marker), so we decrement the numOfObjectsRemoved value.
numOfObjectsRemoved--;
// If trying to delete a delete marker, DeleteMarkerVersionId equals
// deleteMarker's versionID and DeleteMarker equals true
} else if (objMD && objMD.isDeleteMarker) {
isDeleteMarker = true;
deleteMarkerVersionId = entry.versionId;
}
successfullyDeleted.push({
entry, isDeleteMarker,
deleteMarkerVersionId,
});
// If the object can not be deleted raise an error
if (!objLockInfo.canModifyObject(hasGovernanceBypass)) {
log.debug('trying to delete locked object');
return callback(objectLockedError);
}
return callback(null, objMD, versionId);
},
(objMD, versionId, callback) => {
const options = preprocessingVersioningDelete(
bucketName, bucket, objMD, versionId, config.nullVersionCompatMode);
const deleteInfo = {};
if (options && options.deleteData) {
options.overheadField = overheadField;
deleteInfo.deleted = true;
if (!_bucketRequiresOplogUpdate(bucket)) {
options.doesNotNeedOpogUpdate = true;
}
if (objMD.uploadId) {
// eslint-disable-next-line
options.replayId = objMD.uploadId;
}
return services.deleteObject(bucketName, objMD,
entry.key, options, config.multiObjectDeleteEnableOptimizations, log, (err, toDelete) => {
if (err) {
return callback(err);
}
if (toDelete) {
deleteFromStorage = deleteFromStorage.concat(toDelete);
}
return callback(null, objMD, deleteInfo);
});
}
deleteInfo.newDeleteMarker = true;
// This call will create a delete-marker
return createAndStoreObject(bucketName, bucket, entry.key,
objMD, authInfo, canonicalID, null, request,
deleteInfo.newDeleteMarker, null, overheadField, log, (err, result) =>
callback(err, objMD, deleteInfo, result.versionId));
},
], (err, objMD, deleteInfo, versionId) => {
if (err === skipError) {
return moveOn();
} else if (err === objectLockedError) {
errorResults.push({ entry, error: errors.AccessDenied, objectLocked: true });
return moveOn();
} else if (err) {
log.error('error deleting object', { error: err, entry });
errorResults.push({ entry, error: err });
return moveOn();
}
if (deleteInfo.deleted && objMD['content-length']) {
numOfObjectsRemoved++;
totalContentLengthDeleted += objMD['content-length'];
}
let isDeleteMarker;
let deleteMarkerVersionId;
// - If trying to delete an object that does not exist (if a new
// delete marker was created)
// - Or if an object exists but no version was specified
// return DeleteMarkerVersionId equals the versionID of the marker
// you just generated and DeleteMarker tag equals true
if (deleteInfo.newDeleteMarker) {
isDeleteMarker = true;
deleteMarkerVersionId = versionIdUtils.encode(
versionId, config.versionIdEncodingType);
// In this case we are putting a new object (i.e., the delete
// marker), so we decrement the numOfObjectsRemoved value.
numOfObjectsRemoved--;
// If trying to delete a delete marker, DeleteMarkerVersionId equals
// deleteMarker's versionID and DeleteMarker equals true
} else if (objMD && objMD.isDeleteMarker) {
isDeleteMarker = true;
deleteMarkerVersionId = entry.versionId;
}
successfullyDeleted.push({
entry, isDeleteMarker,
deleteMarkerVersionId,
});
return moveOn();
});
},
// end of forEach func
err => {
// Batch delete all objects
const onDone = () => callback(err, quietSetting, errorResults, numOfObjectsRemoved,
successfullyDeleted, totalContentLengthDeleted, bucket);
},
// end of forEach func
err => {
// Batch delete all objects
const onDone = () => {
activeSpan.addEvent('Got Object Metadata and Deleted Objects');
currentSpan.end();
return callback(err, quietSetting, errorResults, numOfObjectsRemoved,
successfullyDeleted, totalContentLengthDeleted, bucket)
};
if (err && deleteFromStorage.length === 0) {
log.trace('no objects to delete from data backend');
return onDone();
}
// If error but we have objects in the list, delete them to ensure
// consistent state.
log.trace('deleting objects from data backend');
// Split the array into chunks
const chunks = [];
while (deleteFromStorage.length > 0) {
chunks.push(deleteFromStorage.splice(0, config.multiObjectDeleteConcurrency));
}
return async.each(chunks, (chunk, done) => data.batchDelete(chunk, null, null,
logger.newRequestLoggerFromSerializedUids(log.getSerializedUids()), done),
err => {
if (err) {
log.error('error deleting objects from data backend', { error: err });
return onDone(err);
}
if (err && deleteFromStorage.length === 0) {
log.trace('no objects to delete from data backend');
return onDone();
});
}),
], (err, ...results) => {
// if general error from metadata return error
if (err) {
monitoring.promMetrics('DELETE', bucketName, err.code,
'multiObjectDelete');
return next(err);
}
return next(null, ...results);
}
// If error but we have objects in the list, delete them to ensure
// consistent state.
log.trace('deleting objects from data backend');
// Split the array into chunks
const chunks = [];
while (deleteFromStorage.length > 0) {
chunks.push(deleteFromStorage.splice(0, config.multiObjectDeleteConcurrency));
}
return async.each(chunks, (chunk, done) => data.batchDelete(chunk, null, null,
logger.newRequestLoggerFromSerializedUids(log.getSerializedUids()), done),
err => {
if (err) {
log.error('error deleting objects from data backend', { error: err });
return onDone(err);
}
return onDone();
});
}),
], (err, ...results) => {
activeSpan.addEvent('Got Object Metadata and Deleted Objects');
getObjMetadataAndDeleteSpan.end();
// if general error from metadata return error
if (err) {
monitoring.promMetrics('DELETE', bucketName, err.code,
'multiObjectDelete');
return next(err);
}
return next(null, ...results);
});
});
}
@ -460,224 +481,318 @@ function getObjMetadataAndDelete(authInfo, canonicalID, request,
* @param {function} callback - callback to server
* @return {undefined}
*/
function multiObjectDelete(authInfo, request, log, callback) {
log.debug('processing request', { method: 'multiObjectDelete' });
if (!request.post) {
monitoring.promMetrics('DELETE', request.bucketName, 400,
'multiObjectDelete');
return callback(errors.MissingRequestBodyError);
}
const md5 = crypto.createHash('md5')
.update(request.post, 'utf8').digest('base64');
if (md5 !== request.headers['content-md5']) {
monitoring.promMetrics('DELETE', request.bucketName, 400,
'multiObjectDelete');
return callback(errors.BadDigest);
}
const bucketName = request.bucketName;
const canonicalID = authInfo.getCanonicalID();
return async.waterfall([
function parseXML(next) {
return _parseXml(request.post,
(err, quietSetting, objects) => {
if (err || objects.length < 1 || objects.length > 1000) {
return next(errors.MalformedXML);
}
return next(null, quietSetting, objects);
});
},
function checkBucketMetadata(quietSetting, objects, next) {
const errorResults = [];
return metadata.getBucket(bucketName, log, (err, bucketMD) => {
if (err) {
log.trace('error retrieving bucket metadata',
{ error: err });
return next(err);
}
// check whether bucket has transient or deleted flag
if (bucketShield(bucketMD, 'objectDelete')) {
return next(errors.NoSuchBucket);
}
if (!isBucketAuthorized(bucketMD, 'objectDelete', canonicalID, authInfo, log, request,
request.actionImplicitDenies)) {
log.trace("access denied due to bucket acl's");
// if access denied at the bucket level, no access for
// any of the objects so all results will be error results
objects.forEach(entry => {
errorResults.push({
entry,
error: errors.AccessDenied,
});
});
// by sending an empty array as the objects array
// async.forEachLimit below will not actually
// make any calls to metadata or data but will continue on
// to the next step to build xml
return next(null, quietSetting, errorResults, [], bucketMD);
}
return next(null, quietSetting, errorResults, objects, bucketMD);
});
},
function checkPolicies(quietSetting, errorResults, objects, bucketMD, next) {
// track keys that are still on track to be deleted
const inPlay = [];
// if request from account, no need to check policies
// all objects are inPlay so send array of object keys
// as inPlay argument
if (!authInfo.isRequesterAnIAMUser()) {
return next(null, quietSetting, errorResults, objects, bucketMD);
}
// TODO: once arsenal's extractParams is separated from doAuth
// function, refactor so only extract once and send
// params on to this api
const authParams = auth.server.extractParams(request, log,
's3', request.query);
const ip = requestUtils.getClientIp(request, config);
const requestContextParams = {
constantParams: {
headers: request.headers,
query: request.query,
generalResource: request.bucketName,
requesterIp: ip,
sslEnabled: request.connection.encrypted,
apiMethod: 'objectDelete',
awsService: 's3',
locationConstraint: null,
requesterInfo: authInfo,
signatureVersion: authParams.params.data.authType,
authType: authParams.params.data.signatureVersion,
signatureAge: authParams.params.data.signatureAge,
},
parameterize: {
// eslint-disable-next-line
specificResource: objects.map(entry => {
return {
key: entry.key,
versionId: entry.versionId,
};
}),
},
};
return vault.checkPolicies(requestContextParams, authInfo.getArn(),
log, (err, authorizationResults) => {
// there were no policies so received a blanket AccessDenied
if (err && err.is.AccessDenied) {
objects.forEach(entry => {
errorResults.push({
entry,
error: errors.AccessDenied });
});
// send empty array for inPlay
return next(null, quietSetting, errorResults, [], bucketMD);
}
if (err) {
log.trace('error checking policies', {
error: err,
method: 'multiObjectDelete.checkPolicies',
});
return next(err);
}
if (objects.length !== authorizationResults.length) {
log.error('vault did not return correct number of ' +
'authorization results', {
authorizationResultsLength:
authorizationResults.length,
objectsLength: objects.length,
});
return next(errors.InternalError);
}
// Convert authorization results into an easier to handle format
const actionImplicitDenies = authorizationResults.reduce((acc, curr, idx) => {
const apiMethod = authorizationResults[idx].action;
// eslint-disable-next-line no-param-reassign
acc[apiMethod] = curr.isImplicit;
return acc;
}, {});
for (let i = 0; i < authorizationResults.length; i++) {
const result = authorizationResults[i];
// result is { isAllowed: true,
// arn: arn:aws:s3:::bucket/object,
// versionId: sampleversionId } unless not allowed
// in which case no isAllowed key will be present
const slashIndex = result.arn.indexOf('/');
if (slashIndex === -1) {
log.error('wrong arn format from vault');
return next(errors.InternalError);
}
const entry = {
key: result.arn.slice(slashIndex + 1),
versionId: result.versionId,
};
// Deny immediately if there is an explicit deny
if (!result.isImplicit && !result.isAllowed) {
errorResults.push({
entry,
error: errors.AccessDenied,
});
continue;
}
// Evaluate against the bucket policies
const areAllActionsAllowed = evaluateBucketPolicyWithIAM(
bucketMD,
Object.keys(actionImplicitDenies),
canonicalID,
authInfo,
actionImplicitDenies,
log,
request);
if (areAllActionsAllowed) {
inPlay.push(entry);
} else {
errorResults.push({
entry,
error: errors.AccessDenied,
});
}
}
return next(null, quietSetting, errorResults, inPlay, bucketMD);
});
},
function getObjMetadataAndDeleteStep(quietSetting, errorResults, inPlay,
bucket, next) {
return getObjMetadataAndDelete(authInfo, canonicalID, request,
bucketName, bucket, quietSetting, errorResults, inPlay,
log, next);
},
], (err, quietSetting, errorResults, numOfObjectsRemoved,
successfullyDeleted, totalContentLengthDeleted, bucket) => {
const corsHeaders = collectCorsHeaders(request.headers.origin,
request.method, bucket);
if (err) {
monitoring.promMetrics('DELETE', bucketName, err.code,
'multiObjectDelete');
return callback(err, null, corsHeaders);
}
const xml = _formatXML(quietSetting, errorResults,
successfullyDeleted);
const deletedKeys = successfullyDeleted.map(item => item.key);
const removedDeleteMarkers = successfullyDeleted
.filter(item => item.isDeleteMarker && item.entry && item.entry.versionId)
.length;
pushMetric('multiObjectDelete', log, {
authInfo,
canonicalID: bucket ? bucket.getOwner() : '',
bucket: bucketName,
keys: deletedKeys,
byteLength: Number.parseInt(totalContentLengthDeleted, 10),
numberOfObjects: numOfObjectsRemoved,
removedDeleteMarkers,
isDelete: true,
function multiObjectDelete(authInfo, request, log, callback, oTel) {
const {
cloudserverApiSpan,
activeSpan,
activeTracerContext,
tracer,
} = oTel;
activeSpan.addEvent('Entered objectPut() Function');
const cloudserverApiSpanContext = opentelemetry.trace.setSpan(
activeTracerContext,
cloudserverApiSpan,
);
return tracer.startActiveSpan('Validating and Deleting Objects from S3', undefined, cloudserverApiSpanContext, multiObjectDeleteSpan => {
const multiObjectDeleteSpanContext = opentelemetry.trace.setSpan(
activeTracerContext,
multiObjectDeleteSpan,
);
multiObjectDeleteSpan.setAttributes({
'code.function': 'multiObjectDelete()',
'code.filename': 'lib/api/multiObjectDelete.js',
'code.lineno': 464,
});
activeSpan.addEvent('Processing Delete Objects (multi) Request');
log.debug('processing request', { method: 'multiObjectDelete' });
if (!request.post) {
monitoring.promMetrics('DELETE', request.bucketName, 400,
'multiObjectDelete');
activeSpan.recordException(errors.MissingRequestBodyError);
multiObjectDeleteSpan.end();
return callback(errors.MissingRequestBodyError);
}
const md5 = crypto.createHash('md5')
.update(request.post, 'utf8').digest('base64');
if (md5 !== request.headers['content-md5']) {
monitoring.promMetrics('DELETE', request.bucketName, 400,
'multiObjectDelete');
activeSpan.recordException(errors.BadDigest);
multiObjectDeleteSpan.end();
return callback(errors.BadDigest);
}
const bucketName = request.bucketName;
const canonicalID = authInfo.getCanonicalID();
return async.waterfall([
function parseXML(next) {
activeSpan.addEvent('Parsing XML Request');
return tracer.startActiveSpan('Bucket check for Server side configuration - SSE', undefined, multiObjectDeleteSpanContext, currentSpan => {
currentSpan.setAttributes({
'code.function': 'multiObjectDelete()',
'code.filename': 'lib/api/multiObjectDelete.js',
'code.lineno': 505,
});
return _parseXml(request.post,
(err, quietSetting, objects) => {
if (err || objects.length < 1 || objects.length > 1000) {
activeSpan.recordException(errors.MalformedXML);
multiObjectDeleteSpan.end();
activeSpan.recordException(errors.MalformedXML);
currentSpan.end();
multiObjectDeleteSpan.end();
return next(errors.MalformedXML);
}
activeSpan.setAttributes({
'aws.s3.objects.count': objects.length,
});
activeSpan.addEvent('Parsed XML Request');
currentSpan.end();
return next(null, quietSetting, objects);
});
});
},
function checkBucketMetadata(quietSetting, objects, next) {
activeSpan.addEvent('Retrieve Bucket Metadata');
return tracer.startActiveSpan('Checking Bucket Metadata and check if user is authorized', undefined, multiObjectDeleteSpanContext, currentSpan => {
const errorResults = [];
return metadata.getBucket(bucketName, log, (err, bucketMD) => {
if (err) {
log.trace('error retrieving bucket metadata',
{ error: err });
return next(err);
}
activeSpan.addEvent('Retrieved Bucket Metadata');
// check whether bucket has transient or deleted flag
if (bucketShield(bucketMD, 'objectDelete')) {
return next(errors.NoSuchBucket);
}
activeSpan.addEvent('Checking Bucket Authorization');
if (!isBucketAuthorized(bucketMD, 'objectDelete', canonicalID, authInfo, log, request,
request.actionImplicitDenies)) {
log.trace("access denied due to bucket acl's");
// if access denied at the bucket level, no access for
// any of the objects so all results will be error results
objects.forEach(entry => {
errorResults.push({
entry,
error: errors.AccessDenied,
});
});
// by sending an empty array as the objects array
// async.forEachLimit below will not actually
// make any calls to metadata or data but will continue on
// to the next step to build xml
activeSpan.recordException(errors.AccessDenied);
currentSpan.end();
return next(null, quietSetting, errorResults, [], bucketMD);
}
activeSpan.addEvent('User is authorized to perform object delete on the bucket for all resources');
currentSpan.end();
return next(null, quietSetting, errorResults, objects, bucketMD);
});
});
},
function checkPolicies(quietSetting, errorResults, objects, bucketMD, next) {
return tracer.startActiveSpan('Starting Vault Authorization for each object', undefined, multiObjectDeleteSpanContext, currentSpan => {
const currentSpanContext = opentelemetry.trace.setSpan(
multiObjectDeleteSpanContext,
currentSpan,
);
activeSpan.addEvent('Preparing reqeust context for all objects');
// track keys that are still on track to be deleted
const inPlay = [];
// if request from account, no need to check policies
// all objects are inPlay so send array of object keys
// as inPlay argument
if (!authInfo.isRequesterAnIAMUser()) {
activeSpan.addEvent('Account user no policies need to be checked');
currentSpan.end();
return next(null, quietSetting, errorResults, objects, bucketMD);
}
// TODO: once arsenal's extractParams is separated from doAuth
// function, refactor so only extract once and send
// params on to this api
const authParams = auth.server.extractParams(request, log,
's3', request.query, { activeSpan, activeTracerContext: currentSpanContext, tracer });
const ip = requestUtils.getClientIp(request, config);
const requestContextParams = {
constantParams: {
headers: request.headers,
query: request.query,
generalResource: request.bucketName,
requesterIp: ip,
sslEnabled: request.connection.encrypted,
apiMethod: 'objectDelete',
awsService: 's3',
locationConstraint: null,
requesterInfo: authInfo,
signatureVersion: authParams.params.data.authType,
authType: authParams.params.data.signatureVersion,
signatureAge: authParams.params.data.signatureAge,
},
parameterize: {
// eslint-disable-next-line
specificResource: objects.map(entry => {
return {
key: entry.key,
versionId: entry.versionId,
};
}),
},
};
activeSpan.addEvent('Request Context prepared for all objects');
activeSpan.addEvent('Starting Vault Authorization for all objects');
return vault.checkPolicies(requestContextParams, authInfo.getArn(),
log, (err, authorizationResults) => {
activeSpan.addEvent('Vault Authorization completed for all objects');
// there were no policies so received a blanket AccessDenied
if (err && err.is.AccessDenied) {
objects.forEach(entry => {
errorResults.push({
entry,
error: errors.AccessDenied });
});
activeSpan.recordException(errors.AccessDenied);
currentSpan.end();
// send empty array for inPlay
return next(null, quietSetting, errorResults, [], bucketMD);
}
if (err) {
log.trace('error checking policies', {
error: err,
method: 'multiObjectDelete.checkPolicies',
});
activeSpan.recordException(err);
currentSpan.end();
multiObjectDeleteSpan.end();
return next(err);
}
if (objects.length !== authorizationResults.length) {
log.error('vault did not return correct number of ' +
'authorization results', {
authorizationResultsLength:
authorizationResults.length,
objectsLength: objects.length,
});
activeSpan.recordException(errors.InternalError);
currentSpan.end();
multiObjectDeleteSpan.end();
return next(errors.InternalError);
}
// Convert authorization results into an easier to handle format
const actionImplicitDenies = authorizationResults.reduce((acc, curr, idx) => {
const apiMethod = authorizationResults[idx].action;
// eslint-disable-next-line no-param-reassign
acc[apiMethod] = curr.isImplicit;
return acc;
}, {});
activeSpan.addEvent('Checking authorization results for all objects')
for (let i = 0; i < authorizationResults.length; i++) {
const result = authorizationResults[i];
// result is { isAllowed: true,
// arn: arn:aws:s3:::bucket/object,
// versionId: sampleversionId } unless not allowed
// in which case no isAllowed key will be present
const slashIndex = result.arn.indexOf('/');
if (slashIndex === -1) {
log.error('wrong arn format from vault');
return next(errors.InternalError);
}
const entry = {
key: result.arn.slice(slashIndex + 1),
versionId: result.versionId,
};
// Deny immediately if there is an explicit deny
if (!result.isImplicit && !result.isAllowed) {
errorResults.push({
entry,
error: errors.AccessDenied,
});
continue;
}
// Evaluate against the bucket policies
const areAllActionsAllowed = evaluateBucketPolicyWithIAM(
bucketMD,
Object.keys(actionImplicitDenies),
canonicalID,
authInfo,
actionImplicitDenies,
log,
request);
if (areAllActionsAllowed) {
inPlay.push(entry);
} else {
errorResults.push({
entry,
error: errors.AccessDenied,
});
}
}
activeSpan.addEvent('Authorization results checked for all objects');
currentSpan.end();
return next(null, quietSetting, errorResults, inPlay, bucketMD);
});
});
},
function getObjMetadataAndDeleteStep(quietSetting, errorResults, inPlay,
bucket, next) {
return tracer.startActiveSpan('Getting Object Metadata and Deleting Objects', undefined, multiObjectDeleteSpanContext, currentSpan => {
const currentSpanContext = opentelemetry.trace.setSpan(
multiObjectDeleteSpanContext,
currentSpan,
);
return getObjMetadataAndDelete(authInfo, canonicalID, request,
bucketName, bucket, quietSetting, errorResults, inPlay,
log, next, { currentSpan, activeSpan, activeTracerContext: currentSpanContext, tracer });
});
},
], (err, quietSetting, errorResults, numOfObjectsRemoved,
successfullyDeleted, totalContentLengthDeleted, bucket) => {
activeSpan.addEvent('Processing Delete Objects (multi) Request; objects deleted');
const corsHeaders = collectCorsHeaders(request.headers.origin,
request.method, bucket);
if (err) {
monitoring.promMetrics('DELETE', bucketName, err.code,
'multiObjectDelete');
activeSpan.recordException(err);
multiObjectDeleteSpan.end();
return callback(err, null, corsHeaders);
}
const xml = _formatXML(quietSetting, errorResults,
successfullyDeleted);
const deletedKeys = successfullyDeleted.map(item => item.key);
const removedDeleteMarkers = successfullyDeleted
.filter(item => item.isDeleteMarker && item.entry && item.entry.versionId)
.length;
pushMetric('multiObjectDelete', log, {
authInfo,
canonicalID: bucket ? bucket.getOwner() : '',
bucket: bucketName,
keys: deletedKeys,
byteLength: Number.parseInt(totalContentLengthDeleted, 10),
numberOfObjects: numOfObjectsRemoved,
removedDeleteMarkers,
isDelete: true,
});
activeSpan.setAttributes({
'aws.s3.objects.size': totalContentLengthDeleted,
'aws.s3.objects.deleted.count': numOfObjectsRemoved,
'aws.s3.object.size': totalContentLengthDeleted / numOfObjectsRemoved,
});
monitoring.promMetrics('DELETE', bucketName, '200',
'multiObjectDelete',
Number.parseInt(totalContentLengthDeleted, 10), null, null,
numOfObjectsRemoved);
activeSpan.addEvent('Returning Response');
multiObjectDeleteSpan.end();
// cloudserverApiSpan.end();
return callback(null, xml, corsHeaders);
});
monitoring.promMetrics('DELETE', bucketName, '200',
'multiObjectDelete',
Number.parseInt(totalContentLengthDeleted, 10), null, null,
numOfObjectsRemoved);
return callback(null, xml, corsHeaders);
});
}

View File

@ -202,7 +202,7 @@ function _prepMetadata(request, sourceObjMD, headers, sourceIsDestination,
* @return {undefined}
*/
function objectCopy(authInfo, request, sourceBucket,
sourceObject, sourceVersionId, log, callback) {
sourceObject, sourceVersionId, log, callback, oTel) {
log.debug('processing request', { method: 'objectCopy' });
const destBucketName = request.bucketName;
const destObjectKey = request.objectKey;
@ -263,7 +263,7 @@ function objectCopy(authInfo, request, sourceBucket,
return next(errors.NoSuchBucket);
}
return next(null, destBucketMD, destObjMD);
});
}, oTel);
},
function checkSourceAuthorization(destBucketMD, destObjMD, next) {
return standardMetadataValidateBucketAndObj(valGetParams, request.actionImplicitDenies, log,
@ -340,7 +340,7 @@ function objectCopy(authInfo, request, sourceBucket,
return next(null, storeMetadataParams, dataLocator,
sourceBucketMD, destBucketMD, destObjMD,
sourceLocationConstraintName, backendInfoDest);
});
}, oTel);
},
function getSSEConfiguration(storeMetadataParams, dataLocator, sourceBucketMD,
destBucketMD, destObjMD, sourceLocationConstraintName,
@ -487,8 +487,8 @@ function objectCopy(authInfo, request, sourceBucket,
return next(null, dataToDelete, result, destBucketMD,
storeMetadataParams, serverSideEncryption,
sourceObjSize, destObjPrevSize);
});
},
}, oTel);
},
function deleteExistingData(dataToDelete, storingNewMdResult,
destBucketMD, storeMetadataParams, serverSideEncryption,
sourceObjSize, destObjPrevSize, next) {

View File

@ -26,9 +26,10 @@ const { overheadField } = require('../../constants');
* includes normalized headers
* @param {Logger} log - werelogs request instance
* @param {function} cb - final cb to call with the result and response headers
* @param {object} oTel - tracing information
* @return {undefined}
*/
function objectDelete(authInfo, request, log, cb) {
function objectDelete(authInfo, request, log, cb, oTel) {
log.debug('processing request', { method: 'objectDelete' });
if (authInfo.isRequesterPublicUser()) {
log.debug('operation not available for public user');
@ -97,7 +98,7 @@ function objectDelete(authInfo, request, log, cb) {
});
}
return next(null, bucketMD, objMD);
});
}, oTel);
},
function checkGovernanceBypassHeader(bucketMD, objectMD, next) {
// AWS only returns an object lock error if a version id
@ -113,7 +114,7 @@ function objectDelete(authInfo, request, log, cb) {
return next(err, bucketMD);
}
return next(null, hasGovernanceBypass, bucketMD, objectMD);
});
}, oTel);
}
return next(null, hasGovernanceBypass, bucketMD, objectMD);
},
@ -172,7 +173,7 @@ function objectDelete(authInfo, request, log, cb) {
objectKey, objectMD, authInfo, canonicalID, null, request,
deleteInfo.newDeleteMarker, null, overheadField, log, (err, newDelMarkerRes) => {
next(err, bucketMD, objectMD, newDelMarkerRes, deleteInfo);
});
}, oTel);
},
], (err, bucketMD, objectMD, result, deleteInfo) => {
const resHeaders = collectCorsHeaders(request.headers.origin,
@ -254,6 +255,11 @@ function objectDelete(authInfo, request, log, cb) {
location: objectMD.dataStoreName,
isDelete: true,
});
const { activeSpan } = oTel;
activeSpan.setAttributes({
'aws.s3.object.size': objectMD['content-length'],
'aws.s3.region': objectMD.dataStoreName,
});
monitoring.promMetrics('DELETE', bucketName, '200', 'deleteObject',
Number.parseInt(objectMD['content-length'], 10));
}

View File

@ -22,7 +22,7 @@ const REPLICATION_ACTION = 'DELETE_TAGGING';
* @param {function} callback - callback to server
* @return {undefined}
*/
function objectDeleteTagging(authInfo, request, log, callback) {
function objectDeleteTagging(authInfo, request, log, callback, oTel) {
log.debug('processing request', { method: 'objectDeleteTagging' });
const bucketName = request.bucketName;
@ -71,7 +71,7 @@ function objectDeleteTagging(authInfo, request, log, callback) {
return next(errors.MethodNotAllowed, bucket);
}
return next(null, bucket, objectMD);
}),
}, oTel),
(bucket, objectMD, next) => {
// eslint-disable-next-line no-param-reassign
objectMD.tags = {};

View File

@ -1,5 +1,6 @@
const { errors, s3middleware } = require('arsenal');
const { parseRange } = require('arsenal').network.http.utils;
const opentelemetry = require('@opentelemetry/api');
const { data } = require('../data/wrapper');
@ -25,236 +26,301 @@ const validateHeaders = s3middleware.validateConditionalHeaders;
* @param {function} callback - callback to function in route
* @return {undefined}
*/
function objectGet(authInfo, request, returnTagCount, log, callback) {
log.debug('processing request', { method: 'objectGet' });
const bucketName = request.bucketName;
const objectKey = request.objectKey;
const decodedVidResult = decodeVersionId(request.query);
if (decodedVidResult instanceof Error) {
log.trace('invalid versionId query', {
versionId: request.query.versionId,
error: decodedVidResult,
function objectGet(authInfo, request, returnTagCount, log, callback, oTel) {
const {
cloudserverApiSpan,
activeSpan,
activeTracerContext,
tracer,
} = oTel;
activeSpan.addEvent('Entered objectGet() Function');
const cloudserverApiSpanContext = opentelemetry.trace.setSpan(
activeTracerContext,
cloudserverApiSpan,
);
return tracer.startActiveSpan('Getting Object MD from S3', undefined, cloudserverApiSpanContext, objectGetSpan => {
const objectGetSpanContext = opentelemetry.trace.setSpan(
activeTracerContext,
objectGetSpan,
);
objectGetSpan.setAttributes({
'code.function': 'objectGet()',
'code.filename': 'lib/api/objectGet.js',
'code.lineno': 29,
});
return callback(decodedVidResult);
}
const versionId = decodedVidResult;
activeSpan.addEvent('Processing Object Get Request');
log.debug('processing request', { method: 'objectGet' });
const bucketName = request.bucketName;
const objectKey = request.objectKey;
const mdValParams = {
authInfo,
bucketName,
objectKey,
versionId,
getDeleteMarker: true,
requestType: request.apiMethods || 'objectGet',
request,
};
return standardMetadataValidateBucketAndObj(mdValParams, request.actionImplicitDenies, log,
(err, bucket, objMD) => {
const corsHeaders = collectCorsHeaders(request.headers.origin,
request.method, bucket);
if (err) {
log.debug('error processing request', {
error: err,
method: 'metadataValidateBucketAndObj',
const decodedVidResult = decodeVersionId(request.query);
if (decodedVidResult instanceof Error) {
activeSpan.recordException(decodedVidResult);
objectGetSpan.end();
log.trace('invalid versionId query', {
versionId: request.query.versionId,
error: decodedVidResult,
});
monitoring.promMetrics(
'GET', bucketName, err.code, 'getObject');
return callback(err, null, corsHeaders);
return callback(decodedVidResult);
}
if (!objMD) {
const err = versionId ? errors.NoSuchVersion : errors.NoSuchKey;
monitoring.promMetrics(
'GET', bucketName, err.code, 'getObject');
return callback(err, null, corsHeaders);
}
const verCfg = bucket.getVersioningConfiguration();
if (objMD.isDeleteMarker) {
const responseMetaHeaders = Object.assign({},
{ 'x-amz-delete-marker': true }, corsHeaders);
if (!versionId) {
monitoring.promMetrics(
'GET', bucketName, 404, 'getObject');
return callback(errors.NoSuchKey, null, responseMetaHeaders);
}
// return MethodNotAllowed if requesting a specific
// version that has a delete marker
responseMetaHeaders['x-amz-version-id'] =
getVersionIdResHeader(verCfg, objMD);
monitoring.promMetrics(
'GET', bucketName, 405, 'getObject');
return callback(errors.MethodNotAllowed, null,
responseMetaHeaders);
}
const headerValResult = validateHeaders(request.headers,
objMD['last-modified'], objMD['content-md5']);
if (headerValResult.error) {
return callback(headerValResult.error, null, corsHeaders);
}
const responseMetaHeaders = collectResponseHeaders(objMD,
corsHeaders, verCfg, returnTagCount);
const versionId = decodedVidResult;
setExpirationHeaders(responseMetaHeaders, {
lifecycleConfig: bucket.getLifecycleConfiguration(),
objectParams: {
key: objectKey,
tags: objMD.tags,
date: objMD['last-modified'],
},
isVersionedReq: !!versionId,
});
const objLength = (objMD.location === null ?
0 : parseInt(objMD['content-length'], 10));
let byteRange;
const streamingParams = {};
if (request.headers.range) {
const { range, error } = parseRange(request.headers.range,
objLength);
if (error) {
monitoring.promMetrics(
'GET', bucketName, 400, 'getObject');
return callback(error, null, corsHeaders);
}
responseMetaHeaders['Accept-Ranges'] = 'bytes';
if (range) {
byteRange = range;
// End of range should be included so + 1
responseMetaHeaders['Content-Length'] =
range[1] - range[0] + 1;
responseMetaHeaders['Content-Range'] =
`bytes ${range[0]}-${range[1]}/${objLength}`;
streamingParams.rangeStart = (range[0] || typeof range[0] === 'number') ?
range[0].toString() : undefined;
streamingParams.rangeEnd = range[1] ?
range[1].toString() : undefined;
}
}
let dataLocator = null;
if (objMD.location !== null) {
// To provide for backwards compatibility before
// md-model-version 2, need to handle cases where
// objMD.location is just a string
dataLocator = Array.isArray(objMD.location) ?
objMD.location : [{ key: objMD.location }];
// if the data backend is azure, there will only ever be at
// most one item in the dataLocator array
if (dataLocator[0] && dataLocator[0].dataStoreType === 'azure') {
dataLocator[0].azureStreamingOptions = streamingParams;
}
let partNumber = null;
if (request.query && request.query.partNumber !== undefined) {
if (byteRange) {
const error = errors.InvalidRequest
.customizeDescription('Cannot specify both Range ' +
'header and partNumber query parameter.');
monitoring.promMetrics(
'GET', bucketName, 400, 'getObject');
return callback(error, null, corsHeaders);
}
partNumber = Number.parseInt(request.query.partNumber, 10);
if (Number.isNaN(partNumber)) {
const error = errors.InvalidArgument
.customizeDescription('Part number must be a number.');
monitoring.promMetrics(
'GET', bucketName, 400, 'getObject');
return callback(error, null, corsHeaders);
}
if (partNumber < 1 || partNumber > 10000) {
const error = errors.InvalidArgument
.customizeDescription('Part number must be an ' +
'integer between 1 and 10000, inclusive.');
monitoring.promMetrics(
'GET', bucketName, 400, 'getObject');
return callback(error, null, corsHeaders);
}
}
// If have a data model before version 2, cannot support
// get range for objects with multiple parts
if (byteRange && dataLocator.length > 1 &&
dataLocator[0].start === undefined) {
monitoring.promMetrics(
'GET', bucketName, 501, 'getObject');
return callback(errors.NotImplemented, null, corsHeaders);
}
if (objMD['x-amz-server-side-encryption']) {
for (let i = 0; i < dataLocator.length; i++) {
dataLocator[i].masterKeyId =
objMD['x-amz-server-side-encryption-aws-kms-key-id'];
dataLocator[i].algorithm =
objMD['x-amz-server-side-encryption'];
}
}
if (partNumber) {
const locations = [];
let locationPartNumber;
for (let i = 0; i < objMD.location.length; i++) {
const { dataStoreETag } = objMD.location[i];
if (dataStoreETag) {
locationPartNumber =
Number.parseInt(dataStoreETag.split(':')[0], 10);
} else {
/**
* Location objects prior to GA7.1 do not include the
* dataStoreETag field so we cannot find the part range,
* the objects are treated as if they only have 1 part
*/
locationPartNumber = 1;
}
// Get all parts that belong to the requested part number
if (partNumber === locationPartNumber) {
locations.push(objMD.location[i]);
} else if (locationPartNumber > partNumber) {
break;
}
}
if (locations.length === 0) {
monitoring.promMetrics(
'GET', bucketName, 400, 'getObject');
return callback(errors.InvalidPartNumber, null,
corsHeaders);
}
const { start } = locations[0];
const endLocation = locations[locations.length - 1];
const end = endLocation.start + endLocation.size - 1;
responseMetaHeaders['Content-Length'] = end - start + 1;
const partByteRange = [start, end];
dataLocator = setPartRanges(dataLocator, partByteRange);
const partsCount = getPartCountFromMd5(objMD);
if (partsCount) {
responseMetaHeaders['x-amz-mp-parts-count'] =
partsCount;
}
} else {
dataLocator = setPartRanges(dataLocator, byteRange);
}
}
return data.head(dataLocator, log, err => {
const mdValParams = {
authInfo,
bucketName,
objectKey,
versionId,
getDeleteMarker: true,
requestType: request.apiMethods || 'objectGet',
request,
};
const mdSpan = tracer.startSpan('Validating Bucket and Object Metadata', undefined, objectGetSpanContext);
const mdSpanContext = opentelemetry.trace.setSpan(
objectGetSpanContext,
mdSpan,
);
return standardMetadataValidateBucketAndObj(mdValParams, request.actionImplicitDenies, log,
(err, bucket, objMD) => {
mdSpan.end();
activeSpan.addEvent('Bucket and Object Metadata Validation Complete');
activeSpan.addEvent('Collecting CORS Headers');
const corsHeaders = collectCorsHeaders(request.headers.origin,
request.method, bucket);
activeSpan.addEvent('CORS Headers Collected');
if (err) {
log.error('error from external backend checking for ' +
'object existence', { error: err });
log.debug('error processing request', {
error: err,
method: 'metadataValidateBucketAndObj',
});
monitoring.promMetrics(
'GET', bucketName, err.code, 'getObject');
return callback(err);
activeSpan.recordException(err);
objectGetSpan.end();
return callback(err, null, corsHeaders);
}
pushMetric('getObject', log, {
authInfo,
bucket: bucketName,
keys: [objectKey],
newByteLength:
Number.parseInt(responseMetaHeaders['Content-Length'], 10),
versionId: objMD.versionId,
location: objMD.dataStoreName,
activeSpan.addEvent('Bucket Policy Validation Passed');
if (!objMD) {
const err = versionId ? errors.NoSuchVersion : errors.NoSuchKey;
monitoring.promMetrics(
'GET', bucketName, err.code, 'getObject');
activeSpan.recordException(errors.NoSuchBucket);
objectGetSpan.end();
return callback(err, null, corsHeaders);
}
const verCfg = bucket.getVersioningConfiguration();
activeSpan.addEvent('Got Versioning Configuration');
if (objMD.isDeleteMarker) {
const responseMetaHeaders = Object.assign({},
{ 'x-amz-delete-marker': true }, corsHeaders);
if (!versionId) {
monitoring.promMetrics(
'GET', bucketName, 404, 'getObject');
activeSpan.recordException(errors.NoSuchKey);
objectGetSpan.end();
return callback(errors.NoSuchKey, null, responseMetaHeaders);
}
// return MethodNotAllowed if requesting a specific
// version that has a delete marker
responseMetaHeaders['x-amz-version-id'] =
getVersionIdResHeader(verCfg, objMD);
monitoring.promMetrics(
'GET', bucketName, 405, 'getObject');
activeSpan.recordException(errors.MethodNotAllowed);
objectGetSpan.end();
return callback(errors.MethodNotAllowed, null,
responseMetaHeaders);
}
const headerValResult = validateHeaders(request.headers,
objMD['last-modified'], objMD['content-md5']);
if (headerValResult.error) {
activeSpan.recordException(headerValResult.error);
objectGetSpan.end();
return callback(headerValResult.error, null, corsHeaders);
}
activeSpan.addEvent('Validated Request Headers against last-modified date of object and/or ETag');
const responseMetaHeaders = collectResponseHeaders(objMD,
corsHeaders, verCfg, returnTagCount);
activeSpan.addEvent('Collected Response Headers from Object Metadata');
setExpirationHeaders(responseMetaHeaders, {
lifecycleConfig: bucket.getLifecycleConfiguration(),
objectParams: {
key: objectKey,
tags: objMD.tags,
date: objMD['last-modified'],
},
isVersionedReq: !!versionId,
});
monitoring.promMetrics('GET', bucketName, '200', 'getObject',
Number.parseInt(responseMetaHeaders['Content-Length'], 10));
return callback(null, dataLocator, responseMetaHeaders,
byteRange);
});
activeSpan.addEvent('Expiration Headers Set');
const objLength = (objMD.location === null ?
0 : parseInt(objMD['content-length'], 10));
let byteRange;
const streamingParams = {};
if (request.headers.range) {
const { range, error } = parseRange(request.headers.range,
objLength);
if (error) {
monitoring.promMetrics(
'GET', bucketName, 400, 'getObject');
activeSpan.recordException(error);
objectGetSpan.end();
return callback(error, null, corsHeaders);
}
responseMetaHeaders['Accept-Ranges'] = 'bytes';
if (range) {
byteRange = range;
// End of range should be included so + 1
responseMetaHeaders['Content-Length'] =
range[1] - range[0] + 1;
responseMetaHeaders['Content-Range'] =
`bytes ${range[0]}-${range[1]}/${objLength}`;
streamingParams.rangeStart = (range[0] || typeof range[0] === 'number') ?
range[0].toString() : undefined;
streamingParams.rangeEnd = range[1] ?
range[1].toString() : undefined;
}
}
let dataLocator = null;
if (objMD.location !== null) {
// To provide for backwards compatibility before
// md-model-version 2, need to handle cases where
// objMD.location is just a string
dataLocator = Array.isArray(objMD.location) ?
objMD.location : [{ key: objMD.location }];
// if the data backend is azure, there will only ever be at
// most one item in the dataLocator array
if (dataLocator[0] && dataLocator[0].dataStoreType === 'azure') {
dataLocator[0].azureStreamingOptions = streamingParams;
}
activeSpan.addEvent('Data Locator Set');
let partNumber = null;
if (request.query && request.query.partNumber !== undefined) {
if (byteRange) {
const error = errors.InvalidRequest
.customizeDescription('Cannot specify both Range ' +
'header and partNumber query parameter.');
monitoring.promMetrics(
'GET', bucketName, 400, 'getObject');
return callback(error, null, corsHeaders);
}
partNumber = Number.parseInt(request.query.partNumber, 10);
if (Number.isNaN(partNumber)) {
const error = errors.InvalidArgument
.customizeDescription('Part number must be a number.');
monitoring.promMetrics(
'GET', bucketName, 400, 'getObject');
return callback(error, null, corsHeaders);
}
if (partNumber < 1 || partNumber > 10000) {
const error = errors.InvalidArgument
.customizeDescription('Part number must be an ' +
'integer between 1 and 10000, inclusive.');
monitoring.promMetrics(
'GET', bucketName, 400, 'getObject');
return callback(error, null, corsHeaders);
}
}
// If have a data model before version 2, cannot support
// get range for objects with multiple parts
if (byteRange && dataLocator.length > 1 &&
dataLocator[0].start === undefined) {
monitoring.promMetrics(
'GET', bucketName, 501, 'getObject');
return callback(errors.NotImplemented, null, corsHeaders);
}
if (objMD['x-amz-server-side-encryption']) {
for (let i = 0; i < dataLocator.length; i++) {
dataLocator[i].masterKeyId =
objMD['x-amz-server-side-encryption-aws-kms-key-id'];
dataLocator[i].algorithm =
objMD['x-amz-server-side-encryption'];
}
}
activeSpan.addEvent('Retrieved Server-Side Encryption Configuration')
if (partNumber) {
const locations = [];
let locationPartNumber;
for (let i = 0; i < objMD.location.length; i++) {
const { dataStoreETag } = objMD.location[i];
if (dataStoreETag) {
locationPartNumber =
Number.parseInt(dataStoreETag.split(':')[0], 10);
} else {
/**
* Location objects prior to GA7.1 do not include the
* dataStoreETag field so we cannot find the part range,
* the objects are treated as if they only have 1 part
*/
locationPartNumber = 1;
}
// Get all parts that belong to the requested part number
if (partNumber === locationPartNumber) {
locations.push(objMD.location[i]);
} else if (locationPartNumber > partNumber) {
break;
}
}
if (locations.length === 0) {
monitoring.promMetrics(
'GET', bucketName, 400, 'getObject');
return callback(errors.InvalidPartNumber, null,
corsHeaders);
}
const { start } = locations[0];
const endLocation = locations[locations.length - 1];
const end = endLocation.start + endLocation.size - 1;
responseMetaHeaders['Content-Length'] = end - start + 1;
const partByteRange = [start, end];
dataLocator = setPartRanges(dataLocator, partByteRange);
const partsCount = getPartCountFromMd5(objMD);
if (partsCount) {
responseMetaHeaders['x-amz-mp-parts-count'] =
partsCount;
}
} else {
dataLocator = setPartRanges(dataLocator, byteRange);
}
}
return tracer.startActiveSpan('Performing head request to locate data', undefined, objectGetSpanContext, dataLocatorSpan => {
dataLocatorSpan.setAttributes({
'code.function': 'objectGet()',
'code.filename': 'lib/api/objectGet.js',
'code.lineno': 289,
});
return data.head(dataLocator, log, err => {
dataLocatorSpan.end();
if (err) {
log.error('error from external backend checking for ' +
'object existence', { error: err });
monitoring.promMetrics(
'GET', bucketName, err.code, 'getObject');
return callback(err);
}
pushMetric('getObject', log, {
authInfo,
bucket: bucketName,
keys: [objectKey],
newByteLength:
Number.parseInt(responseMetaHeaders['Content-Length'], 10),
versionId: objMD.versionId,
location: objMD.dataStoreName,
});
activeSpan.setAttributes({
'aws.s3.object.size': Number.parseInt(responseMetaHeaders['Content-Length'], 10),
'aws.s3.region': objMD.dataStoreName,
'aws.s3.object.versionId': objMD.versionId,
});
monitoring.promMetrics('GET', bucketName, '200', 'getObject',
Number.parseInt(responseMetaHeaders['Content-Length'], 10));
activeSpan.addEvent('Exiting objectGet() Function');
objectGetSpan.end();
return callback(null, dataLocator, responseMetaHeaders,
byteRange);
});
});
}, { activeSpan, activeTracerContext: mdSpanContext, tracer });
});
}

View File

@ -39,7 +39,7 @@ const monitoring = require('../utilities/metrics');
* @param {function} callback - callback to respond to http request
* @return {undefined}
*/
function objectGetACL(authInfo, request, log, callback) {
function objectGetACL(authInfo, request, log, callback, oTel) {
log.debug('processing request', { method: 'objectGetACL' });
const bucketName = request.bucketName;
const objectKey = request.objectKey;
@ -103,7 +103,7 @@ function objectGetACL(authInfo, request, log, callback) {
return next(errors.NoSuchKey);
}
return next(null, bucket, objectMD);
});
}, oTel);
},
function gatherACLs(bucket, objectMD, next) {
const verCfg = bucket.getVersioningConfiguration();

View File

@ -18,7 +18,7 @@ const { convertToXml } = s3middleware.objectLegalHold;
* @param {function} callback - callback to server
* @return {undefined}
*/
function objectGetLegalHold(authInfo, request, log, callback) {
function objectGetLegalHold(authInfo, request, log, callback, oTel) {
log.debug('processing request', { method: 'objectGetLegalHold' });
const { bucketName, objectKey, query } = request;
@ -80,7 +80,7 @@ function objectGetLegalHold(authInfo, request, log, callback) {
'Bucket is missing Object Lock Configuration'));
}
return next(null, bucket, objectMD);
}),
}, oTel),
(bucket, objectMD, next) => {
const { legalHold } = objectMD;
const xml = convertToXml(legalHold);

View File

@ -18,7 +18,7 @@ const { convertToXml } = s3middleware.retention;
* @param {function} callback - callback to server
* @return {undefined}
*/
function objectGetRetention(authInfo, request, log, callback) {
function objectGetRetention(authInfo, request, log, callback, oTel) {
log.debug('processing request', { method: 'objectGetRetention' });
const { bucketName, objectKey } = request;
@ -80,7 +80,7 @@ function objectGetRetention(authInfo, request, log, callback) {
'Bucket is missing Object Lock Configuration'));
}
return next(null, bucket, objectMD);
}),
}, oTel),
(bucket, objectMD, next) => {
const { retentionMode, retentionDate } = objectMD;
if (!retentionMode || !retentionDate) {

View File

@ -18,7 +18,7 @@ const monitoring = require('../utilities/metrics');
* @param {function} callback - callback to server
* @return {undefined}
*/
function objectGetTagging(authInfo, request, log, callback) {
function objectGetTagging(authInfo, request, log, callback, oTel) {
log.debug('processing request', { method: 'objectGetTagging' });
const bucketName = request.bucketName;
@ -75,7 +75,7 @@ function objectGetTagging(authInfo, request, log, callback) {
return next(errors.NoSuchKey);
}
return next(null, bucket, objectMD);
}),
}, oTel),
(bucket, objectMD, next) => {
const tags = objectMD.tags;
const xml = convertToXml(tags);

View File

@ -22,10 +22,11 @@ const { setExpirationHeaders } = require('./apiUtils/object/expirationHeaders');
* @param {object} request - normalized request object
* @param {object} log - Werelogs logger
* @param {function} callback - callback to function in route
* @param {object} oTel - OpenTelemetry activeSpan, context and Tracer object
* @return {undefined}
*
*/
function objectHead(authInfo, request, log, callback) {
function objectHead(authInfo, request, log, callback, oTel) {
log.debug('processing request', { method: 'objectHead' });
const bucketName = request.bucketName;
const objectKey = request.objectKey;
@ -158,9 +159,14 @@ function objectHead(authInfo, request, log, callback) {
versionId: objMD ? objMD.versionId : undefined,
location: objMD ? objMD.dataStoreName : undefined,
});
const { activeSpan } = oTel;
activeSpan.setAttributes({
// 'aws.s3.action': 'objectHead',
'aws.s3.region': objMD ? objMD.dataStoreName : undefined,
});
monitoring.promMetrics('HEAD', bucketName, '200', 'headObject');
return callback(null, responseHeaders);
});
}, oTel);
}
module.exports = objectHead;

View File

@ -1,5 +1,6 @@
const async = require('async');
const { errors, versioning } = require('arsenal');
const opentelemetry = require('@opentelemetry/api');
const aclUtils = require('../utilities/aclUtils');
const { cleanUpBucket } = require('./apiUtils/bucket/bucketCreation');
@ -39,171 +40,304 @@ const versionIdUtils = versioning.VersionID;
* (to be used for streaming v4 auth if applicable)
* @param {object} log - the log request
* @param {Function} callback - final callback to call with the result
* @param {object} authorizationResults - authorization results from
* @param {object} oTel - OpenTelemetry methods
* @return {undefined}
*/
function objectPut(authInfo, request, streamingV4Params, log, callback) {
log.debug('processing request', { method: 'objectPut' });
function objectPut(authInfo, request, streamingV4Params, log, callback, authorizationResults, oTel) {
const {
bucketName,
headers,
method,
objectKey,
parsedContentLength,
query,
} = request;
if (!aclUtils.checkGrantHeaderValidity(headers)) {
log.trace('invalid acl header');
monitoring.promMetrics('PUT', request.bucketName, 400,
'putObject');
return callback(errors.InvalidArgument);
}
const queryContainsVersionId = checkQueryVersionId(query);
if (queryContainsVersionId instanceof Error) {
return callback(queryContainsVersionId);
}
const invalidSSEError = errors.InvalidArgument.customizeDescription(
'The encryption method specified is not supported');
const requestType = request.apiMethods || 'objectPut';
const valParams = { authInfo, bucketName, objectKey, requestType, request };
const canonicalID = authInfo.getCanonicalID();
if (hasNonPrintables(objectKey)) {
return callback(errors.InvalidInput.customizeDescription(
'object keys cannot contain non-printable characters',
));
}
const checksumHeaderErr = validateChecksumHeaders(headers);
if (checksumHeaderErr) {
return callback(checksumHeaderErr);
}
log.trace('owner canonicalID to send to data', { canonicalID });
return standardMetadataValidateBucketAndObj(valParams, request.actionImplicitDenies, log,
(err, bucket, objMD) => {
const responseHeaders = collectCorsHeaders(headers.origin,
method, bucket);
if (err) {
log.trace('error processing request', {
error: err,
method: 'metadataValidateBucketAndObj',
});
monitoring.promMetrics('PUT', bucketName, err.code, 'putObject');
return callback(err, responseHeaders);
cloudserverApiSpan,
activeSpan,
activeTracerContext,
tracer,
} = oTel;
activeSpan.addEvent('Entered objectPut() Function');
const cloudserverApiSpanContext = opentelemetry.trace.setSpan(
activeTracerContext,
cloudserverApiSpan,
);
return tracer.startActiveSpan('Validating and Updating Object MD in S3', undefined, cloudserverApiSpanContext, objectPutSpan => {
const objectPutSpanContext = opentelemetry.trace.setSpan(
activeTracerContext,
objectPutSpan,
);
objectPutSpan.setAttributes({
'code.function': 'objectPut()',
'code.filename': 'lib/api/objectPut.js',
'code.lineno': 45,
});
log.debug('processing request', { method: 'objectPut' });
activeSpan.addEvent('Processing Object Put Request');
const {
bucketName,
headers,
method,
objectKey,
parsedContentLength,
query,
} = request;
if (!aclUtils.checkGrantHeaderValidity(headers)) {
activeSpan.recordException(errors.InvalidArgument);
objectPutSpan.end();
log.trace('invalid acl header');
monitoring.promMetrics('PUT', request.bucketName, 400,
'putObject');
return callback(errors.InvalidArgument);
}
if (bucket.hasDeletedFlag() && canonicalID !== bucket.getOwner()) {
log.trace('deleted flag on bucket and request ' +
'from non-owner account');
monitoring.promMetrics('PUT', bucketName, 404, 'putObject');
return callback(errors.NoSuchBucket);
const queryContainsVersionId = checkQueryVersionId(query);
if (queryContainsVersionId instanceof Error) {
activeSpan.recordException(queryContainsVersionId);
objectPutSpan.end();
return callback(queryContainsVersionId);
}
const invalidSSEError = errors.InvalidArgument.customizeDescription(
'The encryption method specified is not supported');
const requestType = request.apiMethods || 'objectPut';
const valParams = { authInfo, bucketName, objectKey, requestType, request };
const canonicalID = authInfo.getCanonicalID();
if (hasNonPrintables(objectKey)) {
activeSpan.recordException(errors.InvalidInput);
objectPutSpan.end();
return callback(errors.InvalidInput.customizeDescription(
'object keys cannot contain non-printable characters',
));
}
return async.waterfall([
function handleTransientOrDeleteBuckets(next) {
if (bucket.hasTransientFlag() || bucket.hasDeletedFlag()) {
return cleanUpBucket(bucket, canonicalID, log, next);
}
return next();
},
function getSSEConfig(next) {
return getObjectSSEConfiguration(headers, bucket, log,
(err, sseConfig) => {
if (err) {
log.error('error getting server side encryption config', { err });
return next(invalidSSEError);
}
return next(null, sseConfig);
}
);
},
function createCipherBundle(serverSideEncryptionConfig, next) {
if (serverSideEncryptionConfig) {
return kms.createCipherBundle(
serverSideEncryptionConfig, log, next);
}
return next(null, null);
},
function objectCreateAndStore(cipherBundle, next) {
const objectLockValidationError
= validateHeaders(bucket, headers, log);
if (objectLockValidationError) {
return next(objectLockValidationError);
}
writeContinue(request, request._response);
return createAndStoreObject(bucketName,
bucket, objectKey, objMD, authInfo, canonicalID, cipherBundle,
request, false, streamingV4Params, overheadField, log, next);
},
], (err, storingResult) => {
const checksumHeaderErr = validateChecksumHeaders(headers);
if (checksumHeaderErr) {
activeSpan.recordException(checksumHeaderErr);
objectPutSpan.end();
return callback(checksumHeaderErr);
}
log.trace('owner canonicalID to send to data', { canonicalID });
const mdSpan = tracer.startSpan('Validating Bucket and Object Metadata', undefined, objectPutSpanContext);
const mdSpanContext = opentelemetry.trace.setSpan(
objectPutSpanContext,
mdSpan,
);
return standardMetadataValidateBucketAndObj(valParams, request.actionImplicitDenies, log,
(err, bucket, objMD) => {
mdSpan.end();
activeSpan.addEvent('Bucket and Object Metadata Validation Complete');
activeSpan.addEvent('Collecting CORS Headers');
const responseHeaders = collectCorsHeaders(headers.origin,
method, bucket);
activeSpan.addEvent('CORS Headers Collected');
if (err) {
monitoring.promMetrics('PUT', bucketName, err.code,
'putObject');
log.trace('error processing request', {
error: err,
method: 'metadataValidateBucketAndObj',
});
monitoring.promMetrics('PUT', bucketName, err.code, 'putObject');
activeSpan.recordException(err);
objectPutSpan.end();
return callback(err, responseHeaders);
}
// ingestSize assumes that these custom headers indicate
// an ingestion PUT which is a metadata only operation.
// Since these headers can be modified client side, they
// should be used with caution if needed for precise
// metrics.
const ingestSize = (request.headers['x-amz-meta-mdonly']
&& !Number.isNaN(request.headers['x-amz-meta-size']))
? Number.parseInt(request.headers['x-amz-meta-size'], 10) : null;
const newByteLength = parsedContentLength;
activeSpan.addEvent('Bucket Policy Validation Passed');
if (bucket.hasDeletedFlag() && canonicalID !== bucket.getOwner()) {
log.trace('deleted flag on bucket and request ' +
'from non-owner account');
monitoring.promMetrics('PUT', bucketName, 404, 'putObject');
activeSpan.recordException(errors.NoSuchBucket);
objectPutSpan.end();
return callback(errors.NoSuchBucket);
}
setExpirationHeaders(responseHeaders, {
lifecycleConfig: bucket.getLifecycleConfiguration(),
objectParams: {
key: objectKey,
date: storingResult.lastModified,
tags: storingResult.tags,
return async.waterfall([
function handleTransientOrDeleteBuckets(next) {
if (bucket.hasTransientFlag() || bucket.hasDeletedFlag()) {
activeSpan.addEvent('Bucket in Transient or Deleted State');
return tracer.startActiveSpan('Checking Server-Side Encryption Configuration', undefined, objectPutSpanContext, currentSpan => {
currentSpan.setAttributes({
'code.function': 'objectPut()',
'code.filename': 'lib/api/objectPut.js',
'code.lineno': 168,
});
return cleanUpBucket(bucket, log, err => {
activeSpan.addEvent('Bucket Cleanup Complete');
currentSpan.end();
if (err) {
activeSpan.recordException(err);
objectPutSpan.end();
return next(err);
}
return next();
});
});
}
return next(null);
},
});
// Utapi expects null or a number for oldByteLength:
// * null - new object
// * 0 or > 0 - existing object with content-length 0 or > 0
// objMD here is the master version that we would
// have overwritten if there was an existing version or object
//
// TODO: Handle utapi metrics for null version overwrites.
const oldByteLength = objMD && objMD['content-length']
!== undefined ? objMD['content-length'] : null;
if (storingResult) {
// ETag's hex should always be enclosed in quotes
responseHeaders.ETag = `"${storingResult.contentMD5}"`;
}
const vcfg = bucket.getVersioningConfiguration();
const isVersionedObj = vcfg && vcfg.Status === 'Enabled';
if (isVersionedObj) {
if (storingResult && storingResult.versionId) {
responseHeaders['x-amz-version-id'] =
versionIdUtils.encode(storingResult.versionId,
config.versionIdEncodingType);
next => tracer.startActiveSpan('Bucket check for Server side configuration - SSE', undefined, objectPutSpanContext, currentSpan => {
currentSpan.setAttributes({
'code.function': 'objectPut()',
'code.filename': 'lib/api/objectPut.js',
'code.lineno': 178,
});
return next(null, currentSpan);
}),
function getSSEConfig(currentSpan, next) {
activeSpan.addEvent('Retrieving Server-Side Encryption Configuration');
return getObjectSSEConfiguration(headers, bucket, log,
(err, sseConfig) => {
if (err) {
log.error('error getting server side encryption config', { err });
activeSpan.recordException(invalidSSEError);
currentSpan.end();
objectPutSpan.end();
return next(invalidSSEError);
}
return next(null, sseConfig, currentSpan);
}
);
},
(sseConfig, currentSpan, next) => {
activeSpan.addEvent('Retrieved Server-Side Encryption Configuration');
currentSpan.end();
return next(null, sseConfig);
},
(sseConfig, next) => tracer.startActiveSpan('Creating Cipher Bundle from SSE Configuration', undefined, objectPutSpanContext, currentSpan => {
currentSpan.setAttributes({
'code.function': 'objectPut()',
'code.filename': 'lib/api/objectPut.js',
'code.lineno': 205,
});
return next(null, sseConfig, currentSpan);
}),
function createCipherBundle(serverSideEncryptionConfig, currentSpan, next) {
if (serverSideEncryptionConfig) {
activeSpan.addEvent('Creating KMS Cipher Bundle');
currentSpan.end();
objectPutSpan.end();
return kms.createCipherBundle(
serverSideEncryptionConfig, log, next);
}
activeSpan.addEvent('No Server-Side Encryption Configuration Found');
return next(null, currentSpan);
},
(currentSpan, next) => {
activeSpan.addEvent('Got Server-Side Encryption Configuration');
currentSpan.end();
return next();
},
(next) => tracer.startActiveSpan('Validating and Updating Object Data in RING', undefined, objectPutSpanContext, currentSpan => {
currentSpan.setAttributes({
'code.function': 'objectPut()',
'code.filename': 'lib/api/objectPut.js',
'code.lineno': 229,
});
return next(null, null, currentSpan);
}),
function objectCreateAndStore(cipherBundle, currentSpan, next) {
const currentSpanContext = opentelemetry.trace.setSpan(
objectPutSpanContext,
currentSpan,
);
activeSpan.addEvent('Started Object Creation and Storage');
const objectLockValidationError
= validateHeaders(bucket, headers, log);
if (objectLockValidationError) {
activeSpan.recordException(objectLockValidationError);
currentSpan.end();
objectPutSpan.end();
return next(objectLockValidationError);
}
writeContinue(request, request._response);
return createAndStoreObject(bucketName,
bucket, objectKey, objMD, authInfo, canonicalID, cipherBundle,
request, false, streamingV4Params, overheadField, log, (err, storingResult) => {
if (err) {
activeSpan.recordException(err);
currentSpan.end();
objectPutSpan.end();
return next(err);
}
return next(null, storingResult, currentSpan);
}, { activeSpan, activeTracerContext: currentSpanContext, tracer });
},
], (err, storingResult, currentSpan) => {
currentSpan.end();
if (err) {
monitoring.promMetrics('PUT', bucketName, err.code,
'putObject');
activeSpan.recordException(err);
objectPutSpan.end();
return callback(err, responseHeaders);
}
}
activeSpan.addEvent('Completed Object Creation and Storage');
// ingestSize assumes that these custom headers indicate
// an ingestion PUT which is a metadata only operation.
// Since these headers can be modified client side, they
// should be used with caution if needed for precise
// metrics.
const ingestSize = (request.headers['x-amz-meta-mdonly']
&& !Number.isNaN(request.headers['x-amz-meta-size']))
? Number.parseInt(request.headers['x-amz-meta-size'], 10) : null;
const newByteLength = parsedContentLength;
// Only pre-existing non-versioned objects get 0 all others use 1
const numberOfObjects = !isVersionedObj && oldByteLength !== null ? 0 : 1;
setExpirationHeaders(responseHeaders, {
lifecycleConfig: bucket.getLifecycleConfiguration(),
objectParams: {
key: objectKey,
date: storingResult.lastModified,
tags: storingResult.tags,
},
});
// only the bucket owner's metrics should be updated, regardless of
// who the requester is
pushMetric('putObject', log, {
authInfo,
canonicalID: bucket.getOwner(),
bucket: bucketName,
keys: [objectKey],
newByteLength,
oldByteLength: isVersionedObj ? null : oldByteLength,
versionId: isVersionedObj && storingResult ? storingResult.versionId : undefined,
location: bucket.getLocationConstraint(),
numberOfObjects,
// Utapi expects null or a number for oldByteLength:
// * null - new object
// * 0 or > 0 - existing object with content-length 0 or > 0
// objMD here is the master version that we would
// have overwritten if there was an existing version or object
//
// TODO: Handle utapi metrics for null version overwrites.
const oldByteLength = objMD && objMD['content-length']
!== undefined ? objMD['content-length'] : null;
if (storingResult) {
// ETag's hex should always be enclosed in quotes
responseHeaders.ETag = `"${storingResult.contentMD5}"`;
}
const vcfg = bucket.getVersioningConfiguration();
const isVersionedObj = vcfg && vcfg.Status === 'Enabled';
if (isVersionedObj) {
if (storingResult && storingResult.versionId) {
responseHeaders['x-amz-version-id'] =
versionIdUtils.encode(storingResult.versionId,
config.versionIdEncodingType);
}
}
// Only pre-existing non-versioned objects get 0 all others use 1
const numberOfObjects = !isVersionedObj && oldByteLength !== null ? 0 : 1;
// only the bucket owner's metrics should be updated, regardless of
// who the requester is
pushMetric('putObject', log, {
authInfo,
canonicalID: bucket.getOwner(),
bucket: bucketName,
keys: [objectKey],
newByteLength,
oldByteLength: isVersionedObj ? null : oldByteLength,
versionId: isVersionedObj && storingResult ? storingResult.versionId : undefined,
location: bucket.getLocationConstraint(),
numberOfObjects,
});
activeSpan.setAttributes({
'aws.s3.object.size': newByteLength,
'aws.s3.region': bucket.getLocationConstraint(),
'aws.s3.objects.count': numberOfObjects,
'aws.s3.upload_id': storingResult ? storingResult.uploadId : undefined,
});
monitoring.promMetrics('PUT', bucketName, '200',
'putObject', newByteLength, oldByteLength, isVersionedObj,
null, ingestSize);
activeSpan.addEvent('Exiting objectPut() Function');
objectPutSpan.end();
return callback(null, responseHeaders);
});
monitoring.promMetrics('PUT', bucketName, '200',
'putObject', newByteLength, oldByteLength, isVersionedObj,
null, ingestSize);
return callback(null, responseHeaders);
});
}, { activeSpan, activeTracerContext: mdSpanContext, tracer });
});
}

View File

@ -43,7 +43,7 @@ const { config } = require('../Config');
* @param {function} cb - cb to server
* @return {undefined}
*/
function objectPutACL(authInfo, request, log, cb) {
function objectPutACL(authInfo, request, log, cb, oTel) {
log.debug('processing request', { method: 'objectPutACL' });
const { bucketName, objectKey } = request;
const newCannedACL = request.headers['x-amz-acl'];
@ -130,7 +130,7 @@ function objectPutACL(authInfo, request, log, cb) {
return next(errors.MethodNotAllowed, bucket);
}
return next(null, bucket, objectMD);
});
}, oTel);
},
function parseAclFromXml(bucket, objectMD, next) {
// If not setting acl through headers, parse body

View File

@ -33,7 +33,7 @@ const skipError = new Error('skip');
* @return {undefined}
*/
function objectPutCopyPart(authInfo, request, sourceBucket,
sourceObject, reqVersionId, log, callback) {
sourceObject, reqVersionId, log, callback, oTel) {
log.debug('processing request', { method: 'objectPutCopyPart' });
const destBucketName = request.bucketName;
const destObjectKey = request.objectKey;
@ -108,7 +108,7 @@ function objectPutCopyPart(authInfo, request, sourceBucket,
return next(errors.NoSuchBucket);
}
return next(null, destBucketMD);
});
}, oTel);
},
function checkSourceAuthorization(destBucketMD, next) {
return standardMetadataValidateBucketAndObj(valGetParams, request.actionImplicitDenies, log,
@ -179,7 +179,7 @@ function objectPutCopyPart(authInfo, request, sourceBucket,
return next(null, copyLocator.dataLocator, destBucketMD,
copyLocator.copyObjectSize, sourceVerId,
sourceLocationConstraintName);
});
}, oTel);
},
// get MPU shadow bucket to get splitter based on MD version
function getMpuShadowBucket(dataLocator, destBucketMD,

View File

@ -22,7 +22,7 @@ const REPLICATION_ACTION = 'PUT_LEGAL_HOLD';
* @param {function} callback - callback to server
* @return {undefined}
*/
function objectPutLegalHold(authInfo, request, log, callback) {
function objectPutLegalHold(authInfo, request, log, callback, oTel) {
log.debug('processing request', { method: 'objectPutLegalHold' });
const { bucketName, objectKey } = request;
@ -77,7 +77,7 @@ function objectPutLegalHold(authInfo, request, log, callback) {
), bucket);
}
return next(null, bucket, objectMD);
}),
}, oTel),
(bucket, objectMD, next) => {
log.trace('parsing legal hold');
parseLegalHoldXml(request.post, log, (err, res) =>

View File

@ -51,10 +51,12 @@ function _getPartKey(uploadId, splitter, paddedPartNumber) {
* (to be used for streaming v4 auth if applicable)
* @param {object} log - Werelogs logger
* @param {function} cb - final callback to call with the result
* @param {object} authorizationResults authrorization results
* @param {object} oTel - OpenTelemetry activeSpan, context and Tracer object
* @return {undefined}
*/
function objectPutPart(authInfo, request, streamingV4Params, log,
cb) {
cb, authorizationResults, oTel) {
log.debug('processing request', { method: 'objectPutPart' });
const size = request.parsedContentLength;

View File

@ -23,7 +23,7 @@ const REPLICATION_ACTION = 'PUT_RETENTION';
* @param {function} callback - callback to server
* @return {undefined}
*/
function objectPutRetention(authInfo, request, log, callback) {
function objectPutRetention(authInfo, request, log, callback, oTel) {
log.debug('processing request', { method: 'objectPutRetention' });
const { bucketName, objectKey } = request;
@ -91,7 +91,7 @@ function objectPutRetention(authInfo, request, log, callback) {
), bucket);
}
return next(null, bucket, retentionInfo, objectMD);
}),
}, oTel),
(bucket, retentionInfo, objectMD, next) => {
const hasGovernanceBypass = hasGovernanceBypassHeader(request.headers);
if (hasGovernanceBypass && authInfo.isRequesterAnIAMUser()) {
@ -103,7 +103,7 @@ function objectPutRetention(authInfo, request, log, callback) {
return next(err, bucket);
}
return next(null, bucket, retentionInfo, hasGovernanceBypass, objectMD);
});
}, oTel);
}
return next(null, bucket, retentionInfo, hasGovernanceBypass, objectMD);
},

View File

@ -23,7 +23,7 @@ const REPLICATION_ACTION = 'PUT_TAGGING';
* @param {function} callback - callback to server
* @return {undefined}
*/
function objectPutTagging(authInfo, request, log, callback) {
function objectPutTagging(authInfo, request, log, callback, oTel) {
log.debug('processing request', { method: 'objectPutTagging' });
const { bucketName, objectKey } = request;
@ -71,7 +71,7 @@ function objectPutTagging(authInfo, request, log, callback) {
return next(errors.MethodNotAllowed, bucket);
}
return next(null, bucket, objectMD);
}),
}, oTel),
(bucket, objectMD, next) => {
log.trace('parsing tag(s)');
parseTagXml(request.post, log, (err, tags) =>

View File

@ -1,4 +1,5 @@
const { errors } = require('arsenal');
const opentelemetry = require('@opentelemetry/api');
const constants = require('../../constants');
const services = require('../services');
@ -49,45 +50,74 @@ function generateXml(xml, owner, userBuckets, splitter) {
* @param {object} request - normalized request object
* @param {object} log - Werelogs logger
* @param {function} callback - callback
* @param {object} oTel - OpenTelemetry methods
* @return {undefined}
*/
function serviceGet(authInfo, request, log, callback) {
log.debug('processing request', { method: 'serviceGet' });
if (authInfo.isRequesterPublicUser()) {
log.debug('operation not available for public user');
monitoring.promMetrics(
'GET', request.bucketName, 403, 'getService');
return callback(errors.AccessDenied);
}
const xml = [];
const canonicalId = authInfo.getCanonicalID();
xml.push(
'<?xml version="1.0" encoding="UTF-8"?>',
'<ListAllMyBucketsResult xmlns="http://s3.amazonaws.com/doc/' +
'2006-03-01/">',
'<Owner>',
`<ID>${canonicalId}</ID>`,
`<DisplayName>${authInfo.getAccountDisplayName()}` +
'</DisplayName>',
'</Owner>',
'<Buckets>'
function serviceGet(authInfo, request, log, callback, oTel) {
const {
cloudserverApiSpan,
activeSpan,
activeTracerContext,
tracer,
} = oTel;
activeSpan.addEvent('Entered serviceGet()');
const cloudserverApiSpanContext = opentelemetry.trace.setSpan(
activeTracerContext,
cloudserverApiSpan,
);
return services.getService(authInfo, request, log, constants.splitter,
(err, userBuckets, splitter) => {
if (err) {
monitoring.promMetrics(
'GET', userBuckets, err.code, 'getService');
return callback(err);
}
// TODO push metric for serviceGet
// pushMetric('getService', log, {
// bucket: bucketName,
// });
monitoring.promMetrics('GET', userBuckets, '200', 'getService');
return callback(null, generateXml(xml, canonicalId, userBuckets,
splitter));
return tracer.startActiveSpan('ListBuckets API:: Checking bucket policies, and listing buckets', undefined, cloudserverApiSpanContext, serviceGetSpan => {
serviceGetSpan.setAttributes({
'code.function': 'serviceGet()',
'code.filename': 'lib/api/serviceGet.js',
'code.lineno': 56,
});
const ctx = opentelemetry.trace.setSpan(
activeTracerContext,
serviceGetSpan,
);
log.debug('processing request', { method: 'serviceGet' });
if (authInfo.isRequesterPublicUser()) {
log.debug('operation not available for public user');
monitoring.promMetrics(
'GET', request.bucketName, 403, 'getService');
activeSpan.recordException(errors.AccessDenied);
serviceGetSpan.end();
return callback(errors.AccessDenied);
}
const xml = [];
const canonicalId = authInfo.getCanonicalID();
xml.push(
'<?xml version="1.0" encoding="UTF-8"?>',
'<ListAllMyBucketsResult xmlns="http://s3.amazonaws.com/doc/' +
'2006-03-01/">',
'<Owner>',
`<ID>${canonicalId}</ID>`,
`<DisplayName>${authInfo.getAccountDisplayName()}` +
'</DisplayName>',
'</Owner>',
'<Buckets>'
);
return services.getService(authInfo, request, log, constants.splitter,
(err, userBuckets, splitter) => {
if (err) {
monitoring.promMetrics(
'GET', userBuckets, err.code, 'getService');
activeSpan.recordException(err);
serviceGetSpan.end();
return callback(err);
}
// TODO push metric for serviceGet
// pushMetric('getService', log, {
// bucket: bucketName,
// });
monitoring.promMetrics('GET', userBuckets, '200', 'getService');
activeSpan.addEvent('Leaving serviceGet()');
serviceGetSpan.end();
return callback(null, generateXml(xml, canonicalId, userBuckets,
splitter));
}, undefined, { activeSpan, activeTracerContext: ctx, tracer });
});
}
module.exports = serviceGet;

View File

@ -169,6 +169,7 @@ function validateBucket(bucket, params, log, actionImplicitDenies = {}) {
}
return null;
}
/** standardMetadataValidateBucketAndObj - retrieve bucket and object md from metadata
* and check if user is authorized to access them.
* @param {object} params - function parameters
@ -183,59 +184,101 @@ function validateBucket(bucket, params, log, actionImplicitDenies = {}) {
* @param {function} callback - callback
* @return {undefined} - and call callback with params err, bucket md
*/
function standardMetadataValidateBucketAndObj(params, actionImplicitDenies, log, callback) {
function standardMetadataValidateBucketAndObj(params, actionImplicitDenies, log, callback, oTel) {
const {
activeSpan,
activeTracerContext,
tracer,
} = oTel;
activeSpan.addEvent('Entered standardMetadataValidateBucketAndObj()');
const { authInfo, bucketName, objectKey, versionId, getDeleteMarker, request } = params;
let requestType = params.requestType;
if (!Array.isArray(requestType)) {
requestType = [requestType];
}
async.waterfall([
next => {
// versionId may be 'null', which asks metadata to fetch the null key specifically
next => tracer.startActiveSpan('Fetching Metadata for Bucket and Object', undefined, activeTracerContext, getBucketAndObjectMDSpan => {
getBucketAndObjectMDSpan.setAttributes({
'code.function': 'standardMetadataValidateBucketAndObj()',
'code.filename': 'lib/metadata/metadataUtils.js',
'code.lineno': 213,
});
const getOptions = { versionId };
if (getDeleteMarker) {
getOptions.getDeleteMarker = true;
}
activeSpan.addEvent('Initiating Metadata Fetch for Bucket and Object');
// if some implicit iamAuthzResults, return AccessDenied
// before leaking any state information
return metadata.getBucketAndObjectMD(bucketName, objectKey, getOptions, log, (err, getResult) => {
getBucketAndObjectMDSpan.end();
if (err) {
// if some implicit iamAuthzResults, return AccessDenied
// before leaking any state information
if (actionImplicitDenies && Object.values(actionImplicitDenies).some(v => v === true)) {
activeSpan.recordException(errors.AccessDenied);
return next(errors.AccessDenied);
}
activeSpan.recordException(err);
return next(err);
}
activeSpan.addEvent('Completed Metadata Fetch for Bucket and Object');
return next(null, getResult);
});
},
}),
(getResult, next) => {
const bucket = getResult.bucket ?
BucketInfo.deSerialize(getResult.bucket) : undefined;
if (!bucket) {
log.debug('bucketAttrs is undefined', {
bucket: bucketName,
method: 'metadataValidateBucketAndObj',
return tracer.startActiveSpan('Validating Bucket and Object Metadata and Bucket Policies', undefined, activeTracerContext, validateBucketAndObjectMDSpan => {
validateBucketAndObjectMDSpan.setAttributes({
'code.function': 'standardMetadataValidateBucketAndObj()',
'code.filename': 'lib/metadata/metadataUtils.js',
'code.lineno': 237,
});
return next(errors.NoSuchBucket);
}
const validationError = validateBucket(bucket, params, log, actionImplicitDenies);
if (validationError) {
return next(validationError, bucket);
}
const objMD = getResult.obj ? JSON.parse(getResult.obj) : undefined;
if (!objMD && versionId === 'null') {
return getNullVersionFromMaster(bucketName, objectKey, log,
(err, nullVer) => next(err, bucket, nullVer));
}
return next(null, bucket, objMD);
const bucket = getResult.bucket ?
BucketInfo.deSerialize(getResult.bucket) : undefined;
if (!bucket) {
log.debug('bucketAttrs is undefined', {
bucket: bucketName,
method: 'metadataValidateBucketAndObj',
});
activeSpan.recordException(errors.NoSuchBucket);
validateBucketAndObjectMDSpan.end();
return next(errors.NoSuchBucket);
}
const validationError = validateBucket(bucket, params, log, actionImplicitDenies);
if (validationError) {
activeSpan.recordException(validationError);
validateBucketAndObjectMDSpan.end();
return next(validationError, bucket);
}
const objMD = getResult.obj ? JSON.parse(getResult.obj) : undefined;
if (!objMD && versionId === 'null') {
activeSpan.addEvent('Fetching Null Version from Master');
return getNullVersionFromMaster(bucketName, objectKey, log,
(err, nullVer) => {
activeSpan.addEvent('Completed Null Version Fetch from Master');
validateBucketAndObjectMDSpan.end();
if (err) {
activeSpan.recordException(err);
validateBucketAndObjectMDSpan.end();
return next(err, bucket);
}
activeSpan.addEvent('Completed Bucket and Object Metadata Validation');
return next(null, bucket, nullVer);
});
}
validateBucketAndObjectMDSpan.end();
activeSpan.addEvent('Completed Bucket and Object Metadata Validation');
return next(null, bucket, objMD);
});
},
(bucket, objMD, next) => {
activeSpan.addEvent('Initiating Bucket Policy Authorization Checks');
const canonicalID = authInfo.getCanonicalID();
if (!isObjAuthorized(bucket, objMD, requestType, canonicalID, authInfo, log, request,
actionImplicitDenies)) {
log.debug('access denied for user on object', { requestType });
activeSpan.recordException(errors.AccessDenied);
return next(errors.AccessDenied, bucket);
}
activeSpan.addEvent('Completed Bucket Policy Authorization Checks');
return next(null, bucket, objMD);
},
], (err, bucket, objMD) => {
@ -243,9 +286,11 @@ function standardMetadataValidateBucketAndObj(params, actionImplicitDenies, log,
// still return bucket for cors headers
return callback(err, bucket);
}
activeSpan.addEvent('Leaving standardMetadataValidateBucketAndObj()');
return callback(null, bucket, objMD);
});
}
/** standardMetadataValidateBucket - retrieve bucket from metadata and check if user
* is authorized to access it
* @param {object} params - function parameters
@ -258,7 +303,34 @@ function standardMetadataValidateBucketAndObj(params, actionImplicitDenies, log,
* @param {function} callback - callback
* @return {undefined} - and call callback with params err, bucket md
*/
function standardMetadataValidateBucket(params, actionImplicitDenies, log, callback) {
function standardMetadataValidateBucket(params, actionImplicitDenies, log, callback, oTel) {
if (oTel) {
const {
activeSpan,
activeTracerContext,
tracer,
} = oTel;
activeSpan.addEvent('Entered standardMetadataValidateBucket()');
return tracer.startActiveSpan('Fetching Metadata for Bucket', undefined, activeTracerContext, getBucketMDSpan => {
const { bucketName } = params;
return metadata.getBucket(bucketName, log, (err, bucket) => {
getBucketMDSpan.end();
if (err) {
// if some implicit actionImplicitDenies, return AccessDenied before
// leaking any state information
if (actionImplicitDenies && Object.values(actionImplicitDenies).some(v => v === true)) {
return callback(errors.AccessDenied);
}
log.debug('metadata getbucket failed', { error: err });
activeSpan.recordException(err);
return callback(err);
}
const validationError = validateBucket(bucket, params, log, actionImplicitDenies);
activeSpan.addEvent('Completed Bucket Metadata Fetch');
return callback(validationError, bucket);
});
});
}
const { bucketName } = params;
return metadata.getBucket(bucketName, log, (err, bucket) => {
if (err) {

View File

@ -2,6 +2,7 @@ const url = require('url');
const async = require('async');
const httpProxy = require('http-proxy');
const querystring = require('querystring');
const opentelemetry = require('@opentelemetry/api');
const backbeatProxy = httpProxy.createProxyServer({
ignorePath: true,
@ -1213,6 +1214,31 @@ const backbeatRoutes = {
};
function routeBackbeat(clientIP, request, response, log) {
const tracer = opentelemetry.trace.getTracer(process.env.OTEL_SERVICE_NAME || 'cloudserver-service');
let activeSpan = opentelemetry.trace.getActiveSpan(opentelemetry.context.active());
if (!activeSpan) {
// generic name, name will be updates later when we extract more context
activeSpan = tracer.startSpan('routeRequest');
}
const activeTracerContext = opentelemetry.trace.setSpan(
opentelemetry.context.active(),
activeSpan,
);
activeSpan.addEvent('New request received by NodeJS Cloudserver server');
// const traceId = activeSpan.spanContext().traceId;
// const spanId = activeSpan.spanContext().spanId;
// const exemplarLabels = { traceId, spanId };
activeSpan.setAttributes({
'code.filepath': 'lib/server.js',
'code.function': 'routeRequest()',
'code.lineno': 98,
'rpc.service': 'S3',
});
const oTel = {
activeSpan,
activeTracerContext,
tracer,
};
// Attach the apiMethod method to the request, so it can used by monitoring in the server
// eslint-disable-next-line no-param-reassign
request.apiMethod = 'routeBackbeat';
@ -1341,7 +1367,7 @@ function routeBackbeat(clientIP, request, response, log) {
requestType: request.apiMethods || 'ReplicateObject',
request,
};
return standardMetadataValidateBucketAndObj(mdValParams, request.actionImplicitDenies, log, next);
return standardMetadataValidateBucketAndObj(mdValParams, request.actionImplicitDenies, log, next, oTel);
},
(bucketInfo, objMd, next) => {
if (useMultipleBackend) {

View File

@ -2,6 +2,7 @@ const http = require('http');
const https = require('https');
const cluster = require('cluster');
const { series } = require('async');
const opentelemetry = require('@opentelemetry/api');
const arsenal = require('arsenal');
const { RedisClient, StatsClient } = arsenal.metrics;
const monitoringClient = require('./utilities/monitoringHandler');
@ -89,6 +90,30 @@ class S3Server {
* @returns {void}
*/
routeRequest(req, res) {
const tracer = opentelemetry.trace.getTracer(process.env.OTEL_SERVICE_NAME || 'cloudserver-service');
let activeSpan = opentelemetry.trace.getActiveSpan(opentelemetry.context.active());
if (!activeSpan) {
// generic name, name will be updates later when we extract more context
activeSpan = tracer.startSpan('routeRequest');
}
const activeTracerContext = opentelemetry.trace.setSpan(
opentelemetry.context.active(),
activeSpan,
);
activeSpan.addEvent('New request received by NodeJS Cloudserver server');
// const traceId = activeSpan.spanContext().traceId;
// const spanId = activeSpan.spanContext().spanId;
// const exemplarLabels = { traceId, spanId };
activeSpan.setAttributes({
'code.filepath': 'lib/server.js',
'code.function': 'routeRequest()',
'code.lineno': 98,
'rpc.service': 'S3',
'rpc.system': 'aws-api',
'code.url': 'https://github.com/scality/cloudserver/blob/development/7.70/lib/server.js#L91',
'server.address': process.env.HOSTNAME || 'anurag-local',
'server.port': 8000,
});
metrics.httpActiveRequests.inc();
const requestStartTime = process.hrtime.bigint();
@ -109,6 +134,8 @@ class S3Server {
if (req.apiMethod) {
labels.action = req.apiMethod;
}
activeSpan.addEvent('Response sent to client');
activeSpan.end();
metrics.httpRequestsTotal.labels(labels).inc();
metrics.httpRequestDurationSeconds
.labels(labels)
@ -138,6 +165,11 @@ class S3Server {
metadata,
locStorageCheckFn: locationStorageCheck,
vault,
oTel: {
tracer,
activeSpan,
activeTracerContext,
},
},
};
routes(req, res, params, logger, _config);

View File

@ -1,4 +1,5 @@
const assert = require('assert');
const opentelemetry = require('@opentelemetry/api');
const async = require('async');
const { errors, s3middleware } = require('arsenal');
@ -29,48 +30,67 @@ function setLastModifiedFromHeader(md, metaHeaders) {
}
const services = {
getService(authInfo, request, log, splitter, cb, overrideUserbucket) {
const canonicalID = authInfo.getCanonicalID();
assert.strictEqual(typeof splitter, 'string');
const prefix = `${canonicalID}${splitter}`;
const bucketUsers = overrideUserbucket || usersBucket;
// Note: we are limiting max keys on a bucket listing to 10000
// AWS does not limit but they only allow 100 buckets
// (without special increase)
// TODO: Consider implementing pagination like object listing
// with respect to bucket listing so can go beyond 10000
metadata.listObject(bucketUsers, { prefix, maxKeys: 10000 }, log,
(err, listResponse) => {
// If MD responds with NoSuchBucket, this means the
// hidden usersBucket has not yet been created for
// the domain. If this is the case, it means
// that no buckets in this domain have been created so
// it follows that this particular user has no buckets.
// So, the get service listing should not have any
// buckets to list. By returning an empty array, the
// getService API will just respond with the user info
// without listing any buckets.
if (err && err.NoSuchBucket) {
log.trace('no buckets found');
// If we checked the old user bucket, that means we
// already checked the new user bucket. If neither the
// old user bucket or the new user bucket exist, no buckets
// have yet been created in the namespace so an empty
// listing should be returned
if (overrideUserbucket) {
return cb(null, [], splitter);
}
// Since there were no results from checking the
// new users bucket, we check the old users bucket
return this.getService(authInfo, request, log,
constants.oldSplitter, cb, oldUsersBucket);
}
if (err) {
log.error('error from metadata', { error: err });
return cb(err);
}
return cb(null, listResponse.Contents, splitter);
getService(authInfo, request, log, splitter, cb, overrideUserbucket, oTel) {
const {
activeSpan,
activeTracerContext,
tracer,
} = oTel;
return tracer.startActiveSpan(`Metadata request for ${overrideUserbucket ? `${overrideUserbucket}` : `${usersBucket}`}`, undefined, activeTracerContext, serviceGetServiceSpan => {
serviceGetServiceSpan.setAttributes({
'code.function': 'services.getService()',
'code.filename': 'lib/services.js',
'code.lineno': 33,
});
const canonicalID = authInfo.getCanonicalID();
assert.strictEqual(typeof splitter, 'string');
const prefix = `${canonicalID}${splitter}`;
const bucketUsers = overrideUserbucket || usersBucket;
// Note: we are limiting max keys on a bucket listing to 10000
// AWS does not limit but they only allow 100 buckets
// (without special increase)
// TODO: Consider implementing pagination like object listing
// with respect to bucket listing so can go beyond 10000
metadata.listObject(bucketUsers, { prefix, maxKeys: 10000 }, log,
(err, listResponse) => {
// If MD responds with NoSuchBucket, this means the
// hidden usersBucket has not yet been created for
// the domain. If this is the case, it means
// that no buckets in this domain have been created so
// it follows that this particular user has no buckets.
// So, the get service listing should not have any
// buckets to list. By returning an empty array, the
// getService API will just respond with the user info
// without listing any buckets.
if (err && err.NoSuchBucket) {
serviceGetServiceSpan.end();
log.trace('no buckets found');
// If we checked the old user bucket, that means we
// already checked the new user bucket. If neither the
// old user bucket or the new user bucket exist, no buckets
// have yet been created in the namespace so an empty
// listing should be returned
if (overrideUserbucket) {
return cb(null, [], splitter);
}
serviceGetServiceSpan.end();
activeSpan.addEvent('Checking metadata for old user buckets');
// Since there were no results from checking the
// new users bucket, we check the old users bucket
return this.getService(authInfo, request, log,
constants.oldSplitter, cb, oldUsersBucket);
}
if (err) {
log.error('error from metadata', { error: err });
activeSpan.recordException(err);
serviceGetServiceSpan.end();
return cb(err);
}
serviceGetServiceSpan.end();
activeSpan.addEvent('Got metadata, exiting serviceGet()');
return cb(null, listResponse.Contents, splitter);
});
});
},
/**
@ -98,9 +118,21 @@ const services = {
* @param {object} cipherBundle - server side encryption information
* @param {object} params - custom built object containing resource details.
* @param {function} cb - callback containing result for the next task
* @param {function} oTel - opentelemetry methods
* @return {function} executes callback with err or ETag as arguments
*/
metadataStoreObject(bucketName, dataGetInfo, cipherBundle, params, cb) {
metadataStoreObject(bucketName, dataGetInfo, cipherBundle, params, cb, oTel) {
const {
activeSpan,
activeTracerContext,
tracer,
} = oTel;
const objectStorageParamSpan = tracer.startSpan('Extracting Parameters For Metadata Storage', undefined, activeTracerContext);
objectStorageParamSpan.setAttributes({
'code.function': 'createAndStoreObject()',
'code.filename': 'lib/api/apiUtils/object/createAndStoreObject.js',
'code.lineno': 87,
});
const { objectKey, authInfo, size, contentMD5, metaHeaders,
contentType, cacheControl, contentDisposition, contentEncoding,
expires, multipart, headers, overrideMetadata, log,
@ -215,7 +247,11 @@ const services = {
error: validationTagRes,
method: 'metadataStoreObject',
});
return process.nextTick(() => cb(validationTagRes));
return process.nextTick(() => {
activeSpan.recordException(validationTagRes);
objectStorageParamSpan.end();
return cb(validationTagRes);
});
}
md.setTags(validationTagRes);
}
@ -242,13 +278,22 @@ const services = {
}
log.trace('object metadata', { omVal: md.getValue() });
objectStorageParamSpan.end();
// If this is not the completion of a multipart upload or
// the creation of a delete marker, parse the headers to
// get the ACL's if any
return async.waterfall([
callback => {
next => tracer.startActiveSpan('Parsed And Processed ACLs If Needed', undefined, activeTracerContext, currentSpan => {
currentSpan.setAttributes({
'code.function': 'metadataStoreObject()',
'code.filename': 'lib/services.js',
'code.lineno': 294,
});
return next(null, currentSpan);
}),
(currentSpan, callback) => {
if (multipart || md.getIsDeleteMarker()) {
return callback();
return callback(null, currentSpan);
}
const parseAclParams = {
headers,
@ -260,17 +305,42 @@ const services = {
acl.parseAclFromHeaders(parseAclParams, (err, parsedACL) => {
if (err) {
log.debug('error parsing acl', { error: err });
activeSpan.recordException(err);
currentSpan.end();
return callback(err);
}
md.setAcl(parsedACL);
return callback();
return callback(null, currentSpan);
});
return null;
},
callback => metadata.putObjectMD(bucketName, objectKey, md,
options, log, callback),
], (err, data) => {
(currentSpan, callback) => {
activeSpan.addEvent('Parsed and processed ACLs if needed');
currentSpan.end();
return callback();
},
next => tracer.startActiveSpan('Update object metadata - PUT', undefined, activeTracerContext, currentSpan => {
currentSpan.setAttributes({
'code.function': 'metadataStoreObject()',
'code.filename': 'lib/services.js',
'code.lineno': 329,
});
return next(null, currentSpan);
}),
(currentSpan, callback) => metadata.putObjectMD(bucketName, objectKey, md,
options, log, (err, data) => {
if (err) {
log.error('error from metadata', { error: err });
activeSpan.recordException(err);
currentSpan.end();
return callback(err);
}
return callback(null, data, currentSpan);
}),
], (err, data, currentSpan) => {
currentSpan.end();
if (err) {
activeSpan.recordException(err);
log.error('error from metadata', { error: err });
return cb(err);
}

View File

@ -20,7 +20,17 @@
"homepage": "https://github.com/scality/S3#readme",
"dependencies": {
"@hapi/joi": "^17.1.0",
"arsenal": "git+https://github.com/scality/arsenal#7.70.29",
"@opentelemetry/api": "^1.9.0",
"@opentelemetry/auto-instrumentations-node": "^0.47.1",
"@opentelemetry/exporter-metrics-otlp-proto": "^0.52.0",
"@opentelemetry/exporter-trace-otlp-proto": "^0.52.0",
"@opentelemetry/instrumentation-ioredis": "^0.41.0",
"@opentelemetry/resources": "^1.25.0",
"@opentelemetry/sdk-metrics": "^1.25.0",
"@opentelemetry/sdk-trace-base": "^1.25.0",
"@opentelemetry/sdk-trace-web": "^1.25.0",
"@opentelemetry/semantic-conventions": "^1.25.0",
"arsenal": "git+https://github.com/scality/Arsenal#d90bc4f741446012885640e93d45052fad4f7ea6",
"async": "~2.5.0",
"aws-sdk": "2.905.0",
"azure-storage": "^2.1.0",
@ -82,13 +92,13 @@
"lint_md": "mdlint $(git ls-files '*.md')",
"mem_backend": "S3BACKEND=mem node index.js",
"start": "npm-run-all --parallel start_dmd start_s3server",
"start_mdserver": "node mdserver.js",
"start_dataserver": "node dataserver.js",
"start_s3server": "node index.js",
"start_mdserver": "OTEL_SERVICE_NAME=cisco-s3-mdserver node --require ./instrumentation.js mdserver.js",
"start_dataserver": "OTEL_SERVICE_NAME=cisco-s3-dataserver node --require ./instrumentation.js dataserver.js",
"start_s3server": "OTEL_SERVICE_NAME=cisco-s3-cloudserver node --require ./instrumentation.js index.js",
"start_dmd": "npm-run-all --parallel start_mdserver start_dataserver",
"start_utapi": "node lib/utapi/utapi.js",
"utapi_replay": "node lib/utapi/utapiReplay.js",
"utapi_reindex": "node lib/utapi/utapiReindex.js",
"start_utapi": "OTEL_SERVICE_NAME=cisco-s3-utapi node --require ./instrumentation.js lib/utapi/utapi.js",
"utapi_replay": "OTEL_SERVICE_NAME=cisco-s3-utapi-replay node --require ./instrumentation.js lib/utapi/utapiReplay.js",
"utapi_reindex": "OTEL_SERVICE_NAME=cisco-s3-utapi-reindex node --require ./instrumentation.js lib/utapi/utapiReindex.js",
"test": "CI=true S3BACKEND=mem mocha --reporter mocha-multi-reporters --reporter-options configFile=$INIT_CWD/tests/reporter-config.json --recursive tests/unit",
"test_versionid_base62": "VERSION_ID_ENCODING_TYPE=base62 CI=true S3BACKEND=mem mocha --reporter mocha-multi-reporters --reporter-options configFile=$INIT_CWD/tests/reporter-config.json --recursive tests/unit/api",
"test_legacy_location": "CI=true S3_LOCATION_FILE=tests/locationConfig/locationConfigLegacy.json S3BACKEND=mem mocha --reporter mocha-multi-reporters --reporter-options configFile=$INIT_CWD/tests/reporter-config.json --recursive tests/unit",

1330
yarn.lock

File diff suppressed because it is too large Load Diff