mirror of
https://github.com/aws-actions/configure-aws-credentials.git
synced 2026-03-12 18:07:10 -04:00
chore: Update dist
This commit is contained in:
2
dist/869.index.js
generated
vendored
2
dist/869.index.js
generated
vendored
@@ -28,7 +28,7 @@ const resolveCredentialSource = (credentialSource, profileName, logger) => {
|
|||||||
},
|
},
|
||||||
Environment: async (options) => {
|
Environment: async (options) => {
|
||||||
logger?.debug("@aws-sdk/credential-provider-ini - credential_source is Environment");
|
logger?.debug("@aws-sdk/credential-provider-ini - credential_source is Environment");
|
||||||
const { fromEnv } = await Promise.resolve(/* import() */).then(__webpack_require__.t.bind(__webpack_require__, 5606, 19));
|
const { fromEnv } = await __webpack_require__.e(/* import() */ 121).then(__webpack_require__.t.bind(__webpack_require__, 7121, 19));
|
||||||
return async () => fromEnv(options)().then(setNamedProvider);
|
return async () => fromEnv(options)().then(setNamedProvider);
|
||||||
},
|
},
|
||||||
};
|
};
|
||||||
|
|||||||
2
dist/cleanup/869.index.js
generated
vendored
2
dist/cleanup/869.index.js
generated
vendored
@@ -28,7 +28,7 @@ const resolveCredentialSource = (credentialSource, profileName, logger) => {
|
|||||||
},
|
},
|
||||||
Environment: async (options) => {
|
Environment: async (options) => {
|
||||||
logger?.debug("@aws-sdk/credential-provider-ini - credential_source is Environment");
|
logger?.debug("@aws-sdk/credential-provider-ini - credential_source is Environment");
|
||||||
const { fromEnv } = await Promise.resolve(/* import() */).then(__webpack_require__.t.bind(__webpack_require__, 5606, 19));
|
const { fromEnv } = await __webpack_require__.e(/* import() */ 121).then(__webpack_require__.t.bind(__webpack_require__, 7121, 19));
|
||||||
return async () => fromEnv(options)().then(setNamedProvider);
|
return async () => fromEnv(options)().then(setNamedProvider);
|
||||||
},
|
},
|
||||||
};
|
};
|
||||||
|
|||||||
199
dist/cleanup/index.js
generated
vendored
199
dist/cleanup/index.js
generated
vendored
@@ -8881,55 +8881,6 @@ exports.parseXmlBody = parseXmlBody;
|
|||||||
exports.parseXmlErrorBody = parseXmlErrorBody;
|
exports.parseXmlErrorBody = parseXmlErrorBody;
|
||||||
|
|
||||||
|
|
||||||
/***/ }),
|
|
||||||
|
|
||||||
/***/ 5606:
|
|
||||||
/***/ ((__unused_webpack_module, exports, __nccwpck_require__) => {
|
|
||||||
|
|
||||||
"use strict";
|
|
||||||
|
|
||||||
|
|
||||||
var client = __nccwpck_require__(5152);
|
|
||||||
var propertyProvider = __nccwpck_require__(1238);
|
|
||||||
|
|
||||||
const ENV_KEY = "AWS_ACCESS_KEY_ID";
|
|
||||||
const ENV_SECRET = "AWS_SECRET_ACCESS_KEY";
|
|
||||||
const ENV_SESSION = "AWS_SESSION_TOKEN";
|
|
||||||
const ENV_EXPIRATION = "AWS_CREDENTIAL_EXPIRATION";
|
|
||||||
const ENV_CREDENTIAL_SCOPE = "AWS_CREDENTIAL_SCOPE";
|
|
||||||
const ENV_ACCOUNT_ID = "AWS_ACCOUNT_ID";
|
|
||||||
const fromEnv = (init) => async () => {
|
|
||||||
init?.logger?.debug("@aws-sdk/credential-provider-env - fromEnv");
|
|
||||||
const accessKeyId = process.env[ENV_KEY];
|
|
||||||
const secretAccessKey = process.env[ENV_SECRET];
|
|
||||||
const sessionToken = process.env[ENV_SESSION];
|
|
||||||
const expiry = process.env[ENV_EXPIRATION];
|
|
||||||
const credentialScope = process.env[ENV_CREDENTIAL_SCOPE];
|
|
||||||
const accountId = process.env[ENV_ACCOUNT_ID];
|
|
||||||
if (accessKeyId && secretAccessKey) {
|
|
||||||
const credentials = {
|
|
||||||
accessKeyId,
|
|
||||||
secretAccessKey,
|
|
||||||
...(sessionToken && { sessionToken }),
|
|
||||||
...(expiry && { expiration: new Date(expiry) }),
|
|
||||||
...(credentialScope && { credentialScope }),
|
|
||||||
...(accountId && { accountId }),
|
|
||||||
};
|
|
||||||
client.setCredentialFeature(credentials, "CREDENTIALS_ENV_VARS", "g");
|
|
||||||
return credentials;
|
|
||||||
}
|
|
||||||
throw new propertyProvider.CredentialsProviderError("Unable to find environment variable credentials.", { logger: init?.logger });
|
|
||||||
};
|
|
||||||
|
|
||||||
exports.ENV_ACCOUNT_ID = ENV_ACCOUNT_ID;
|
|
||||||
exports.ENV_CREDENTIAL_SCOPE = ENV_CREDENTIAL_SCOPE;
|
|
||||||
exports.ENV_EXPIRATION = ENV_EXPIRATION;
|
|
||||||
exports.ENV_KEY = ENV_KEY;
|
|
||||||
exports.ENV_SECRET = ENV_SECRET;
|
|
||||||
exports.ENV_SESSION = ENV_SESSION;
|
|
||||||
exports.fromEnv = fromEnv;
|
|
||||||
|
|
||||||
|
|
||||||
/***/ }),
|
/***/ }),
|
||||||
|
|
||||||
/***/ 5861:
|
/***/ 5861:
|
||||||
@@ -8938,7 +8889,7 @@ exports.fromEnv = fromEnv;
|
|||||||
"use strict";
|
"use strict";
|
||||||
|
|
||||||
|
|
||||||
var credentialProviderEnv = __nccwpck_require__(5606);
|
var credentialProviderEnv = __nccwpck_require__(6153);
|
||||||
var propertyProvider = __nccwpck_require__(1238);
|
var propertyProvider = __nccwpck_require__(1238);
|
||||||
var sharedIniFileLoader = __nccwpck_require__(4964);
|
var sharedIniFileLoader = __nccwpck_require__(4964);
|
||||||
|
|
||||||
@@ -9088,6 +9039,55 @@ exports.credentialsWillNeedRefresh = credentialsWillNeedRefresh;
|
|||||||
exports.defaultProvider = defaultProvider;
|
exports.defaultProvider = defaultProvider;
|
||||||
|
|
||||||
|
|
||||||
|
/***/ }),
|
||||||
|
|
||||||
|
/***/ 6153:
|
||||||
|
/***/ ((__unused_webpack_module, exports, __nccwpck_require__) => {
|
||||||
|
|
||||||
|
"use strict";
|
||||||
|
|
||||||
|
|
||||||
|
var client = __nccwpck_require__(5152);
|
||||||
|
var propertyProvider = __nccwpck_require__(1238);
|
||||||
|
|
||||||
|
const ENV_KEY = "AWS_ACCESS_KEY_ID";
|
||||||
|
const ENV_SECRET = "AWS_SECRET_ACCESS_KEY";
|
||||||
|
const ENV_SESSION = "AWS_SESSION_TOKEN";
|
||||||
|
const ENV_EXPIRATION = "AWS_CREDENTIAL_EXPIRATION";
|
||||||
|
const ENV_CREDENTIAL_SCOPE = "AWS_CREDENTIAL_SCOPE";
|
||||||
|
const ENV_ACCOUNT_ID = "AWS_ACCOUNT_ID";
|
||||||
|
const fromEnv = (init) => async () => {
|
||||||
|
init?.logger?.debug("@aws-sdk/credential-provider-env - fromEnv");
|
||||||
|
const accessKeyId = process.env[ENV_KEY];
|
||||||
|
const secretAccessKey = process.env[ENV_SECRET];
|
||||||
|
const sessionToken = process.env[ENV_SESSION];
|
||||||
|
const expiry = process.env[ENV_EXPIRATION];
|
||||||
|
const credentialScope = process.env[ENV_CREDENTIAL_SCOPE];
|
||||||
|
const accountId = process.env[ENV_ACCOUNT_ID];
|
||||||
|
if (accessKeyId && secretAccessKey) {
|
||||||
|
const credentials = {
|
||||||
|
accessKeyId,
|
||||||
|
secretAccessKey,
|
||||||
|
...(sessionToken && { sessionToken }),
|
||||||
|
...(expiry && { expiration: new Date(expiry) }),
|
||||||
|
...(credentialScope && { credentialScope }),
|
||||||
|
...(accountId && { accountId }),
|
||||||
|
};
|
||||||
|
client.setCredentialFeature(credentials, "CREDENTIALS_ENV_VARS", "g");
|
||||||
|
return credentials;
|
||||||
|
}
|
||||||
|
throw new propertyProvider.CredentialsProviderError("Unable to find environment variable credentials.", { logger: init?.logger });
|
||||||
|
};
|
||||||
|
|
||||||
|
exports.ENV_ACCOUNT_ID = ENV_ACCOUNT_ID;
|
||||||
|
exports.ENV_CREDENTIAL_SCOPE = ENV_CREDENTIAL_SCOPE;
|
||||||
|
exports.ENV_EXPIRATION = ENV_EXPIRATION;
|
||||||
|
exports.ENV_KEY = ENV_KEY;
|
||||||
|
exports.ENV_SECRET = ENV_SECRET;
|
||||||
|
exports.ENV_SESSION = ENV_SESSION;
|
||||||
|
exports.fromEnv = fromEnv;
|
||||||
|
|
||||||
|
|
||||||
/***/ }),
|
/***/ }),
|
||||||
|
|
||||||
/***/ 2590:
|
/***/ 2590:
|
||||||
@@ -11743,6 +11743,13 @@ class CborShapeSerializer extends protocols.SerdeContext {
|
|||||||
const [k, v] = sourceObject.$unknown;
|
const [k, v] = sourceObject.$unknown;
|
||||||
newObject[k] = v;
|
newObject[k] = v;
|
||||||
}
|
}
|
||||||
|
else if (typeof sourceObject.__type === "string") {
|
||||||
|
for (const [k, v] of Object.entries(sourceObject)) {
|
||||||
|
if (!(k in newObject)) {
|
||||||
|
newObject[k] = this.serialize(15, v);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
else if (ns.isDocumentSchema()) {
|
else if (ns.isDocumentSchema()) {
|
||||||
for (const key of Object.keys(sourceObject)) {
|
for (const key of Object.keys(sourceObject)) {
|
||||||
@@ -11847,6 +11854,13 @@ class CborShapeDeserializer extends protocols.SerdeContext {
|
|||||||
const k = keys.values().next().value;
|
const k = keys.values().next().value;
|
||||||
newObject.$unknown = [k, value[k]];
|
newObject.$unknown = [k, value[k]];
|
||||||
}
|
}
|
||||||
|
else if (typeof value.__type === "string") {
|
||||||
|
for (const [k, v] of Object.entries(value)) {
|
||||||
|
if (!(k in newObject)) {
|
||||||
|
newObject[k] = v;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
else if (value instanceof serde.NumericValue) {
|
else if (value instanceof serde.NumericValue) {
|
||||||
return value;
|
return value;
|
||||||
@@ -12167,6 +12181,9 @@ class HttpBindingProtocol extends HttpProtocol {
|
|||||||
const memberTraits = memberNs.getMergedTraits() ?? {};
|
const memberTraits = memberNs.getMergedTraits() ?? {};
|
||||||
const inputMemberValue = input[memberName];
|
const inputMemberValue = input[memberName];
|
||||||
if (inputMemberValue == null && !memberNs.isIdempotencyToken()) {
|
if (inputMemberValue == null && !memberNs.isIdempotencyToken()) {
|
||||||
|
if (memberTraits.httpLabel) {
|
||||||
|
throw new Error(`No value provided for input HTTP label: ${memberName}.`);
|
||||||
|
}
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
if (memberTraits.httpPayload) {
|
if (memberTraits.httpPayload) {
|
||||||
@@ -19967,6 +19984,45 @@ function modeOf(chunk, allowBuffer = true) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/***/ }),
|
||||||
|
|
||||||
|
/***/ 3492:
|
||||||
|
/***/ ((__unused_webpack_module, exports) => {
|
||||||
|
|
||||||
|
"use strict";
|
||||||
|
|
||||||
|
Object.defineProperty(exports, "__esModule", ({ value: true }));
|
||||||
|
exports.getAwsChunkedEncodingStream = void 0;
|
||||||
|
const getAwsChunkedEncodingStream = (readableStream, options) => {
|
||||||
|
const { base64Encoder, bodyLengthChecker, checksumAlgorithmFn, checksumLocationName, streamHasher } = options;
|
||||||
|
const checksumRequired = base64Encoder !== undefined &&
|
||||||
|
bodyLengthChecker !== undefined &&
|
||||||
|
checksumAlgorithmFn !== undefined &&
|
||||||
|
checksumLocationName !== undefined &&
|
||||||
|
streamHasher !== undefined;
|
||||||
|
const digest = checksumRequired ? streamHasher(checksumAlgorithmFn, readableStream) : undefined;
|
||||||
|
const reader = readableStream.getReader();
|
||||||
|
return new ReadableStream({
|
||||||
|
async pull(controller) {
|
||||||
|
const { value, done } = await reader.read();
|
||||||
|
if (done) {
|
||||||
|
controller.enqueue(`0\r\n`);
|
||||||
|
if (checksumRequired) {
|
||||||
|
const checksum = base64Encoder(await digest);
|
||||||
|
controller.enqueue(`${checksumLocationName}:${checksum}\r\n`);
|
||||||
|
controller.enqueue(`\r\n`);
|
||||||
|
}
|
||||||
|
controller.close();
|
||||||
|
}
|
||||||
|
else {
|
||||||
|
controller.enqueue(`${(bodyLengthChecker(value) || 0).toString(16)}\r\n${value}\r\n`);
|
||||||
|
}
|
||||||
|
},
|
||||||
|
});
|
||||||
|
};
|
||||||
|
exports.getAwsChunkedEncodingStream = getAwsChunkedEncodingStream;
|
||||||
|
|
||||||
|
|
||||||
/***/ }),
|
/***/ }),
|
||||||
|
|
||||||
/***/ 6522:
|
/***/ 6522:
|
||||||
@@ -19975,23 +20031,35 @@ function modeOf(chunk, allowBuffer = true) {
|
|||||||
"use strict";
|
"use strict";
|
||||||
|
|
||||||
Object.defineProperty(exports, "__esModule", ({ value: true }));
|
Object.defineProperty(exports, "__esModule", ({ value: true }));
|
||||||
exports.getAwsChunkedEncodingStream = void 0;
|
exports.getAwsChunkedEncodingStream = getAwsChunkedEncodingStream;
|
||||||
const stream_1 = __nccwpck_require__(2203);
|
const node_stream_1 = __nccwpck_require__(7075);
|
||||||
const getAwsChunkedEncodingStream = (readableStream, options) => {
|
const getAwsChunkedEncodingStream_browser_1 = __nccwpck_require__(3492);
|
||||||
|
const stream_type_check_1 = __nccwpck_require__(4414);
|
||||||
|
function getAwsChunkedEncodingStream(stream, options) {
|
||||||
|
const readable = stream;
|
||||||
|
const readableStream = stream;
|
||||||
|
if ((0, stream_type_check_1.isReadableStream)(readableStream)) {
|
||||||
|
return (0, getAwsChunkedEncodingStream_browser_1.getAwsChunkedEncodingStream)(readableStream, options);
|
||||||
|
}
|
||||||
const { base64Encoder, bodyLengthChecker, checksumAlgorithmFn, checksumLocationName, streamHasher } = options;
|
const { base64Encoder, bodyLengthChecker, checksumAlgorithmFn, checksumLocationName, streamHasher } = options;
|
||||||
const checksumRequired = base64Encoder !== undefined &&
|
const checksumRequired = base64Encoder !== undefined &&
|
||||||
checksumAlgorithmFn !== undefined &&
|
checksumAlgorithmFn !== undefined &&
|
||||||
checksumLocationName !== undefined &&
|
checksumLocationName !== undefined &&
|
||||||
streamHasher !== undefined;
|
streamHasher !== undefined;
|
||||||
const digest = checksumRequired ? streamHasher(checksumAlgorithmFn, readableStream) : undefined;
|
const digest = checksumRequired ? streamHasher(checksumAlgorithmFn, readable) : undefined;
|
||||||
const awsChunkedEncodingStream = new stream_1.Readable({ read: () => { } });
|
const awsChunkedEncodingStream = new node_stream_1.Readable({
|
||||||
readableStream.on("data", (data) => {
|
read: () => { },
|
||||||
|
});
|
||||||
|
readable.on("data", (data) => {
|
||||||
const length = bodyLengthChecker(data) || 0;
|
const length = bodyLengthChecker(data) || 0;
|
||||||
|
if (length === 0) {
|
||||||
|
return;
|
||||||
|
}
|
||||||
awsChunkedEncodingStream.push(`${length.toString(16)}\r\n`);
|
awsChunkedEncodingStream.push(`${length.toString(16)}\r\n`);
|
||||||
awsChunkedEncodingStream.push(data);
|
awsChunkedEncodingStream.push(data);
|
||||||
awsChunkedEncodingStream.push("\r\n");
|
awsChunkedEncodingStream.push("\r\n");
|
||||||
});
|
});
|
||||||
readableStream.on("end", async () => {
|
readable.on("end", async () => {
|
||||||
awsChunkedEncodingStream.push(`0\r\n`);
|
awsChunkedEncodingStream.push(`0\r\n`);
|
||||||
if (checksumRequired) {
|
if (checksumRequired) {
|
||||||
const checksum = base64Encoder(await digest);
|
const checksum = base64Encoder(await digest);
|
||||||
@@ -20001,8 +20069,7 @@ const getAwsChunkedEncodingStream = (readableStream, options) => {
|
|||||||
awsChunkedEncodingStream.push(null);
|
awsChunkedEncodingStream.push(null);
|
||||||
});
|
});
|
||||||
return awsChunkedEncodingStream;
|
return awsChunkedEncodingStream;
|
||||||
};
|
}
|
||||||
exports.getAwsChunkedEncodingStream = getAwsChunkedEncodingStream;
|
|
||||||
|
|
||||||
|
|
||||||
/***/ }),
|
/***/ }),
|
||||||
@@ -20138,6 +20205,14 @@ class Uint8ArrayBlobAdapter extends Uint8Array {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
Object.defineProperty(exports, "isBlob", ({
|
||||||
|
enumerable: true,
|
||||||
|
get: function () { return streamTypeCheck.isBlob; }
|
||||||
|
}));
|
||||||
|
Object.defineProperty(exports, "isReadableStream", ({
|
||||||
|
enumerable: true,
|
||||||
|
get: function () { return streamTypeCheck.isReadableStream; }
|
||||||
|
}));
|
||||||
exports.Uint8ArrayBlobAdapter = Uint8ArrayBlobAdapter;
|
exports.Uint8ArrayBlobAdapter = Uint8ArrayBlobAdapter;
|
||||||
Object.keys(ChecksumStream).forEach(function (k) {
|
Object.keys(ChecksumStream).forEach(function (k) {
|
||||||
if (k !== 'default' && !Object.prototype.hasOwnProperty.call(exports, k)) Object.defineProperty(exports, k, {
|
if (k !== 'default' && !Object.prototype.hasOwnProperty.call(exports, k)) Object.defineProperty(exports, k, {
|
||||||
@@ -20181,12 +20256,6 @@ Object.keys(splitStream).forEach(function (k) {
|
|||||||
get: function () { return splitStream[k]; }
|
get: function () { return splitStream[k]; }
|
||||||
});
|
});
|
||||||
});
|
});
|
||||||
Object.keys(streamTypeCheck).forEach(function (k) {
|
|
||||||
if (k !== 'default' && !Object.prototype.hasOwnProperty.call(exports, k)) Object.defineProperty(exports, k, {
|
|
||||||
enumerable: true,
|
|
||||||
get: function () { return streamTypeCheck[k]; }
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
|
|
||||||
/***/ }),
|
/***/ }),
|
||||||
|
|||||||
199
dist/index.js
generated
vendored
199
dist/index.js
generated
vendored
@@ -9742,55 +9742,6 @@ exports.parseXmlBody = parseXmlBody;
|
|||||||
exports.parseXmlErrorBody = parseXmlErrorBody;
|
exports.parseXmlErrorBody = parseXmlErrorBody;
|
||||||
|
|
||||||
|
|
||||||
/***/ }),
|
|
||||||
|
|
||||||
/***/ 5606:
|
|
||||||
/***/ ((__unused_webpack_module, exports, __nccwpck_require__) => {
|
|
||||||
|
|
||||||
"use strict";
|
|
||||||
|
|
||||||
|
|
||||||
var client = __nccwpck_require__(5152);
|
|
||||||
var propertyProvider = __nccwpck_require__(1238);
|
|
||||||
|
|
||||||
const ENV_KEY = "AWS_ACCESS_KEY_ID";
|
|
||||||
const ENV_SECRET = "AWS_SECRET_ACCESS_KEY";
|
|
||||||
const ENV_SESSION = "AWS_SESSION_TOKEN";
|
|
||||||
const ENV_EXPIRATION = "AWS_CREDENTIAL_EXPIRATION";
|
|
||||||
const ENV_CREDENTIAL_SCOPE = "AWS_CREDENTIAL_SCOPE";
|
|
||||||
const ENV_ACCOUNT_ID = "AWS_ACCOUNT_ID";
|
|
||||||
const fromEnv = (init) => async () => {
|
|
||||||
init?.logger?.debug("@aws-sdk/credential-provider-env - fromEnv");
|
|
||||||
const accessKeyId = process.env[ENV_KEY];
|
|
||||||
const secretAccessKey = process.env[ENV_SECRET];
|
|
||||||
const sessionToken = process.env[ENV_SESSION];
|
|
||||||
const expiry = process.env[ENV_EXPIRATION];
|
|
||||||
const credentialScope = process.env[ENV_CREDENTIAL_SCOPE];
|
|
||||||
const accountId = process.env[ENV_ACCOUNT_ID];
|
|
||||||
if (accessKeyId && secretAccessKey) {
|
|
||||||
const credentials = {
|
|
||||||
accessKeyId,
|
|
||||||
secretAccessKey,
|
|
||||||
...(sessionToken && { sessionToken }),
|
|
||||||
...(expiry && { expiration: new Date(expiry) }),
|
|
||||||
...(credentialScope && { credentialScope }),
|
|
||||||
...(accountId && { accountId }),
|
|
||||||
};
|
|
||||||
client.setCredentialFeature(credentials, "CREDENTIALS_ENV_VARS", "g");
|
|
||||||
return credentials;
|
|
||||||
}
|
|
||||||
throw new propertyProvider.CredentialsProviderError("Unable to find environment variable credentials.", { logger: init?.logger });
|
|
||||||
};
|
|
||||||
|
|
||||||
exports.ENV_ACCOUNT_ID = ENV_ACCOUNT_ID;
|
|
||||||
exports.ENV_CREDENTIAL_SCOPE = ENV_CREDENTIAL_SCOPE;
|
|
||||||
exports.ENV_EXPIRATION = ENV_EXPIRATION;
|
|
||||||
exports.ENV_KEY = ENV_KEY;
|
|
||||||
exports.ENV_SECRET = ENV_SECRET;
|
|
||||||
exports.ENV_SESSION = ENV_SESSION;
|
|
||||||
exports.fromEnv = fromEnv;
|
|
||||||
|
|
||||||
|
|
||||||
/***/ }),
|
/***/ }),
|
||||||
|
|
||||||
/***/ 5861:
|
/***/ 5861:
|
||||||
@@ -9799,7 +9750,7 @@ exports.fromEnv = fromEnv;
|
|||||||
"use strict";
|
"use strict";
|
||||||
|
|
||||||
|
|
||||||
var credentialProviderEnv = __nccwpck_require__(5606);
|
var credentialProviderEnv = __nccwpck_require__(6153);
|
||||||
var propertyProvider = __nccwpck_require__(1238);
|
var propertyProvider = __nccwpck_require__(1238);
|
||||||
var sharedIniFileLoader = __nccwpck_require__(4964);
|
var sharedIniFileLoader = __nccwpck_require__(4964);
|
||||||
|
|
||||||
@@ -9949,6 +9900,55 @@ exports.credentialsWillNeedRefresh = credentialsWillNeedRefresh;
|
|||||||
exports.defaultProvider = defaultProvider;
|
exports.defaultProvider = defaultProvider;
|
||||||
|
|
||||||
|
|
||||||
|
/***/ }),
|
||||||
|
|
||||||
|
/***/ 6153:
|
||||||
|
/***/ ((__unused_webpack_module, exports, __nccwpck_require__) => {
|
||||||
|
|
||||||
|
"use strict";
|
||||||
|
|
||||||
|
|
||||||
|
var client = __nccwpck_require__(5152);
|
||||||
|
var propertyProvider = __nccwpck_require__(1238);
|
||||||
|
|
||||||
|
const ENV_KEY = "AWS_ACCESS_KEY_ID";
|
||||||
|
const ENV_SECRET = "AWS_SECRET_ACCESS_KEY";
|
||||||
|
const ENV_SESSION = "AWS_SESSION_TOKEN";
|
||||||
|
const ENV_EXPIRATION = "AWS_CREDENTIAL_EXPIRATION";
|
||||||
|
const ENV_CREDENTIAL_SCOPE = "AWS_CREDENTIAL_SCOPE";
|
||||||
|
const ENV_ACCOUNT_ID = "AWS_ACCOUNT_ID";
|
||||||
|
const fromEnv = (init) => async () => {
|
||||||
|
init?.logger?.debug("@aws-sdk/credential-provider-env - fromEnv");
|
||||||
|
const accessKeyId = process.env[ENV_KEY];
|
||||||
|
const secretAccessKey = process.env[ENV_SECRET];
|
||||||
|
const sessionToken = process.env[ENV_SESSION];
|
||||||
|
const expiry = process.env[ENV_EXPIRATION];
|
||||||
|
const credentialScope = process.env[ENV_CREDENTIAL_SCOPE];
|
||||||
|
const accountId = process.env[ENV_ACCOUNT_ID];
|
||||||
|
if (accessKeyId && secretAccessKey) {
|
||||||
|
const credentials = {
|
||||||
|
accessKeyId,
|
||||||
|
secretAccessKey,
|
||||||
|
...(sessionToken && { sessionToken }),
|
||||||
|
...(expiry && { expiration: new Date(expiry) }),
|
||||||
|
...(credentialScope && { credentialScope }),
|
||||||
|
...(accountId && { accountId }),
|
||||||
|
};
|
||||||
|
client.setCredentialFeature(credentials, "CREDENTIALS_ENV_VARS", "g");
|
||||||
|
return credentials;
|
||||||
|
}
|
||||||
|
throw new propertyProvider.CredentialsProviderError("Unable to find environment variable credentials.", { logger: init?.logger });
|
||||||
|
};
|
||||||
|
|
||||||
|
exports.ENV_ACCOUNT_ID = ENV_ACCOUNT_ID;
|
||||||
|
exports.ENV_CREDENTIAL_SCOPE = ENV_CREDENTIAL_SCOPE;
|
||||||
|
exports.ENV_EXPIRATION = ENV_EXPIRATION;
|
||||||
|
exports.ENV_KEY = ENV_KEY;
|
||||||
|
exports.ENV_SECRET = ENV_SECRET;
|
||||||
|
exports.ENV_SESSION = ENV_SESSION;
|
||||||
|
exports.fromEnv = fromEnv;
|
||||||
|
|
||||||
|
|
||||||
/***/ }),
|
/***/ }),
|
||||||
|
|
||||||
/***/ 2590:
|
/***/ 2590:
|
||||||
@@ -12604,6 +12604,13 @@ class CborShapeSerializer extends protocols.SerdeContext {
|
|||||||
const [k, v] = sourceObject.$unknown;
|
const [k, v] = sourceObject.$unknown;
|
||||||
newObject[k] = v;
|
newObject[k] = v;
|
||||||
}
|
}
|
||||||
|
else if (typeof sourceObject.__type === "string") {
|
||||||
|
for (const [k, v] of Object.entries(sourceObject)) {
|
||||||
|
if (!(k in newObject)) {
|
||||||
|
newObject[k] = this.serialize(15, v);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
else if (ns.isDocumentSchema()) {
|
else if (ns.isDocumentSchema()) {
|
||||||
for (const key of Object.keys(sourceObject)) {
|
for (const key of Object.keys(sourceObject)) {
|
||||||
@@ -12708,6 +12715,13 @@ class CborShapeDeserializer extends protocols.SerdeContext {
|
|||||||
const k = keys.values().next().value;
|
const k = keys.values().next().value;
|
||||||
newObject.$unknown = [k, value[k]];
|
newObject.$unknown = [k, value[k]];
|
||||||
}
|
}
|
||||||
|
else if (typeof value.__type === "string") {
|
||||||
|
for (const [k, v] of Object.entries(value)) {
|
||||||
|
if (!(k in newObject)) {
|
||||||
|
newObject[k] = v;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
else if (value instanceof serde.NumericValue) {
|
else if (value instanceof serde.NumericValue) {
|
||||||
return value;
|
return value;
|
||||||
@@ -13028,6 +13042,9 @@ class HttpBindingProtocol extends HttpProtocol {
|
|||||||
const memberTraits = memberNs.getMergedTraits() ?? {};
|
const memberTraits = memberNs.getMergedTraits() ?? {};
|
||||||
const inputMemberValue = input[memberName];
|
const inputMemberValue = input[memberName];
|
||||||
if (inputMemberValue == null && !memberNs.isIdempotencyToken()) {
|
if (inputMemberValue == null && !memberNs.isIdempotencyToken()) {
|
||||||
|
if (memberTraits.httpLabel) {
|
||||||
|
throw new Error(`No value provided for input HTTP label: ${memberName}.`);
|
||||||
|
}
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
if (memberTraits.httpPayload) {
|
if (memberTraits.httpPayload) {
|
||||||
@@ -20828,6 +20845,45 @@ function modeOf(chunk, allowBuffer = true) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
/***/ }),
|
||||||
|
|
||||||
|
/***/ 3492:
|
||||||
|
/***/ ((__unused_webpack_module, exports) => {
|
||||||
|
|
||||||
|
"use strict";
|
||||||
|
|
||||||
|
Object.defineProperty(exports, "__esModule", ({ value: true }));
|
||||||
|
exports.getAwsChunkedEncodingStream = void 0;
|
||||||
|
const getAwsChunkedEncodingStream = (readableStream, options) => {
|
||||||
|
const { base64Encoder, bodyLengthChecker, checksumAlgorithmFn, checksumLocationName, streamHasher } = options;
|
||||||
|
const checksumRequired = base64Encoder !== undefined &&
|
||||||
|
bodyLengthChecker !== undefined &&
|
||||||
|
checksumAlgorithmFn !== undefined &&
|
||||||
|
checksumLocationName !== undefined &&
|
||||||
|
streamHasher !== undefined;
|
||||||
|
const digest = checksumRequired ? streamHasher(checksumAlgorithmFn, readableStream) : undefined;
|
||||||
|
const reader = readableStream.getReader();
|
||||||
|
return new ReadableStream({
|
||||||
|
async pull(controller) {
|
||||||
|
const { value, done } = await reader.read();
|
||||||
|
if (done) {
|
||||||
|
controller.enqueue(`0\r\n`);
|
||||||
|
if (checksumRequired) {
|
||||||
|
const checksum = base64Encoder(await digest);
|
||||||
|
controller.enqueue(`${checksumLocationName}:${checksum}\r\n`);
|
||||||
|
controller.enqueue(`\r\n`);
|
||||||
|
}
|
||||||
|
controller.close();
|
||||||
|
}
|
||||||
|
else {
|
||||||
|
controller.enqueue(`${(bodyLengthChecker(value) || 0).toString(16)}\r\n${value}\r\n`);
|
||||||
|
}
|
||||||
|
},
|
||||||
|
});
|
||||||
|
};
|
||||||
|
exports.getAwsChunkedEncodingStream = getAwsChunkedEncodingStream;
|
||||||
|
|
||||||
|
|
||||||
/***/ }),
|
/***/ }),
|
||||||
|
|
||||||
/***/ 6522:
|
/***/ 6522:
|
||||||
@@ -20836,23 +20892,35 @@ function modeOf(chunk, allowBuffer = true) {
|
|||||||
"use strict";
|
"use strict";
|
||||||
|
|
||||||
Object.defineProperty(exports, "__esModule", ({ value: true }));
|
Object.defineProperty(exports, "__esModule", ({ value: true }));
|
||||||
exports.getAwsChunkedEncodingStream = void 0;
|
exports.getAwsChunkedEncodingStream = getAwsChunkedEncodingStream;
|
||||||
const stream_1 = __nccwpck_require__(2203);
|
const node_stream_1 = __nccwpck_require__(7075);
|
||||||
const getAwsChunkedEncodingStream = (readableStream, options) => {
|
const getAwsChunkedEncodingStream_browser_1 = __nccwpck_require__(3492);
|
||||||
|
const stream_type_check_1 = __nccwpck_require__(4414);
|
||||||
|
function getAwsChunkedEncodingStream(stream, options) {
|
||||||
|
const readable = stream;
|
||||||
|
const readableStream = stream;
|
||||||
|
if ((0, stream_type_check_1.isReadableStream)(readableStream)) {
|
||||||
|
return (0, getAwsChunkedEncodingStream_browser_1.getAwsChunkedEncodingStream)(readableStream, options);
|
||||||
|
}
|
||||||
const { base64Encoder, bodyLengthChecker, checksumAlgorithmFn, checksumLocationName, streamHasher } = options;
|
const { base64Encoder, bodyLengthChecker, checksumAlgorithmFn, checksumLocationName, streamHasher } = options;
|
||||||
const checksumRequired = base64Encoder !== undefined &&
|
const checksumRequired = base64Encoder !== undefined &&
|
||||||
checksumAlgorithmFn !== undefined &&
|
checksumAlgorithmFn !== undefined &&
|
||||||
checksumLocationName !== undefined &&
|
checksumLocationName !== undefined &&
|
||||||
streamHasher !== undefined;
|
streamHasher !== undefined;
|
||||||
const digest = checksumRequired ? streamHasher(checksumAlgorithmFn, readableStream) : undefined;
|
const digest = checksumRequired ? streamHasher(checksumAlgorithmFn, readable) : undefined;
|
||||||
const awsChunkedEncodingStream = new stream_1.Readable({ read: () => { } });
|
const awsChunkedEncodingStream = new node_stream_1.Readable({
|
||||||
readableStream.on("data", (data) => {
|
read: () => { },
|
||||||
|
});
|
||||||
|
readable.on("data", (data) => {
|
||||||
const length = bodyLengthChecker(data) || 0;
|
const length = bodyLengthChecker(data) || 0;
|
||||||
|
if (length === 0) {
|
||||||
|
return;
|
||||||
|
}
|
||||||
awsChunkedEncodingStream.push(`${length.toString(16)}\r\n`);
|
awsChunkedEncodingStream.push(`${length.toString(16)}\r\n`);
|
||||||
awsChunkedEncodingStream.push(data);
|
awsChunkedEncodingStream.push(data);
|
||||||
awsChunkedEncodingStream.push("\r\n");
|
awsChunkedEncodingStream.push("\r\n");
|
||||||
});
|
});
|
||||||
readableStream.on("end", async () => {
|
readable.on("end", async () => {
|
||||||
awsChunkedEncodingStream.push(`0\r\n`);
|
awsChunkedEncodingStream.push(`0\r\n`);
|
||||||
if (checksumRequired) {
|
if (checksumRequired) {
|
||||||
const checksum = base64Encoder(await digest);
|
const checksum = base64Encoder(await digest);
|
||||||
@@ -20862,8 +20930,7 @@ const getAwsChunkedEncodingStream = (readableStream, options) => {
|
|||||||
awsChunkedEncodingStream.push(null);
|
awsChunkedEncodingStream.push(null);
|
||||||
});
|
});
|
||||||
return awsChunkedEncodingStream;
|
return awsChunkedEncodingStream;
|
||||||
};
|
}
|
||||||
exports.getAwsChunkedEncodingStream = getAwsChunkedEncodingStream;
|
|
||||||
|
|
||||||
|
|
||||||
/***/ }),
|
/***/ }),
|
||||||
@@ -20999,6 +21066,14 @@ class Uint8ArrayBlobAdapter extends Uint8Array {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
Object.defineProperty(exports, "isBlob", ({
|
||||||
|
enumerable: true,
|
||||||
|
get: function () { return streamTypeCheck.isBlob; }
|
||||||
|
}));
|
||||||
|
Object.defineProperty(exports, "isReadableStream", ({
|
||||||
|
enumerable: true,
|
||||||
|
get: function () { return streamTypeCheck.isReadableStream; }
|
||||||
|
}));
|
||||||
exports.Uint8ArrayBlobAdapter = Uint8ArrayBlobAdapter;
|
exports.Uint8ArrayBlobAdapter = Uint8ArrayBlobAdapter;
|
||||||
Object.keys(ChecksumStream).forEach(function (k) {
|
Object.keys(ChecksumStream).forEach(function (k) {
|
||||||
if (k !== 'default' && !Object.prototype.hasOwnProperty.call(exports, k)) Object.defineProperty(exports, k, {
|
if (k !== 'default' && !Object.prototype.hasOwnProperty.call(exports, k)) Object.defineProperty(exports, k, {
|
||||||
@@ -21042,12 +21117,6 @@ Object.keys(splitStream).forEach(function (k) {
|
|||||||
get: function () { return splitStream[k]; }
|
get: function () { return splitStream[k]; }
|
||||||
});
|
});
|
||||||
});
|
});
|
||||||
Object.keys(streamTypeCheck).forEach(function (k) {
|
|
||||||
if (k !== 'default' && !Object.prototype.hasOwnProperty.call(exports, k)) Object.defineProperty(exports, k, {
|
|
||||||
enumerable: true,
|
|
||||||
get: function () { return streamTypeCheck[k]; }
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
|
|
||||||
/***/ }),
|
/***/ }),
|
||||||
|
|||||||
Reference in New Issue
Block a user