mirror of
https://github.com/dependabot/fetch-metadata.git
synced 2026-03-12 18:07:12 -04:00
13630 lines
411 KiB
JavaScript
Generated
13630 lines
411 KiB
JavaScript
Generated
/******/ (() => { // webpackBootstrap
|
|
/******/ var __webpack_modules__ = ({
|
|
|
|
/***/ 7351:
|
|
/***/ (function(__unused_webpack_module, exports, __nccwpck_require__) {
|
|
|
|
"use strict";
|
|
|
|
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
if (k2 === undefined) k2 = k;
|
|
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
}) : (function(o, m, k, k2) {
|
|
if (k2 === undefined) k2 = k;
|
|
o[k2] = m[k];
|
|
}));
|
|
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
}) : function(o, v) {
|
|
o["default"] = v;
|
|
});
|
|
var __importStar = (this && this.__importStar) || function (mod) {
|
|
if (mod && mod.__esModule) return mod;
|
|
var result = {};
|
|
if (mod != null) for (var k in mod) if (k !== "default" && Object.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
__setModuleDefault(result, mod);
|
|
return result;
|
|
};
|
|
Object.defineProperty(exports, "__esModule", ({ value: true }));
|
|
exports.issue = exports.issueCommand = void 0;
|
|
const os = __importStar(__nccwpck_require__(2087));
|
|
const utils_1 = __nccwpck_require__(5278);
|
|
/**
|
|
* Commands
|
|
*
|
|
* Command Format:
|
|
* ::name key=value,key=value::message
|
|
*
|
|
* Examples:
|
|
* ::warning::This is the message
|
|
* ::set-env name=MY_VAR::some value
|
|
*/
|
|
function issueCommand(command, properties, message) {
|
|
const cmd = new Command(command, properties, message);
|
|
process.stdout.write(cmd.toString() + os.EOL);
|
|
}
|
|
exports.issueCommand = issueCommand;
|
|
function issue(name, message = '') {
|
|
issueCommand(name, {}, message);
|
|
}
|
|
exports.issue = issue;
|
|
const CMD_STRING = '::';
|
|
class Command {
|
|
constructor(command, properties, message) {
|
|
if (!command) {
|
|
command = 'missing.command';
|
|
}
|
|
this.command = command;
|
|
this.properties = properties;
|
|
this.message = message;
|
|
}
|
|
toString() {
|
|
let cmdStr = CMD_STRING + this.command;
|
|
if (this.properties && Object.keys(this.properties).length > 0) {
|
|
cmdStr += ' ';
|
|
let first = true;
|
|
for (const key in this.properties) {
|
|
if (this.properties.hasOwnProperty(key)) {
|
|
const val = this.properties[key];
|
|
if (val) {
|
|
if (first) {
|
|
first = false;
|
|
}
|
|
else {
|
|
cmdStr += ',';
|
|
}
|
|
cmdStr += `${key}=${escapeProperty(val)}`;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
cmdStr += `${CMD_STRING}${escapeData(this.message)}`;
|
|
return cmdStr;
|
|
}
|
|
}
|
|
function escapeData(s) {
|
|
return utils_1.toCommandValue(s)
|
|
.replace(/%/g, '%25')
|
|
.replace(/\r/g, '%0D')
|
|
.replace(/\n/g, '%0A');
|
|
}
|
|
function escapeProperty(s) {
|
|
return utils_1.toCommandValue(s)
|
|
.replace(/%/g, '%25')
|
|
.replace(/\r/g, '%0D')
|
|
.replace(/\n/g, '%0A')
|
|
.replace(/:/g, '%3A')
|
|
.replace(/,/g, '%2C');
|
|
}
|
|
//# sourceMappingURL=command.js.map
|
|
|
|
/***/ }),
|
|
|
|
/***/ 2186:
|
|
/***/ (function(__unused_webpack_module, exports, __nccwpck_require__) {
|
|
|
|
"use strict";
|
|
|
|
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
if (k2 === undefined) k2 = k;
|
|
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
}) : (function(o, m, k, k2) {
|
|
if (k2 === undefined) k2 = k;
|
|
o[k2] = m[k];
|
|
}));
|
|
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
}) : function(o, v) {
|
|
o["default"] = v;
|
|
});
|
|
var __importStar = (this && this.__importStar) || function (mod) {
|
|
if (mod && mod.__esModule) return mod;
|
|
var result = {};
|
|
if (mod != null) for (var k in mod) if (k !== "default" && Object.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
__setModuleDefault(result, mod);
|
|
return result;
|
|
};
|
|
var __awaiter = (this && this.__awaiter) || function (thisArg, _arguments, P, generator) {
|
|
function adopt(value) { return value instanceof P ? value : new P(function (resolve) { resolve(value); }); }
|
|
return new (P || (P = Promise))(function (resolve, reject) {
|
|
function fulfilled(value) { try { step(generator.next(value)); } catch (e) { reject(e); } }
|
|
function rejected(value) { try { step(generator["throw"](value)); } catch (e) { reject(e); } }
|
|
function step(result) { result.done ? resolve(result.value) : adopt(result.value).then(fulfilled, rejected); }
|
|
step((generator = generator.apply(thisArg, _arguments || [])).next());
|
|
});
|
|
};
|
|
Object.defineProperty(exports, "__esModule", ({ value: true }));
|
|
exports.getState = exports.saveState = exports.group = exports.endGroup = exports.startGroup = exports.info = exports.warning = exports.error = exports.debug = exports.isDebug = exports.setFailed = exports.setCommandEcho = exports.setOutput = exports.getBooleanInput = exports.getMultilineInput = exports.getInput = exports.addPath = exports.setSecret = exports.exportVariable = exports.ExitCode = void 0;
|
|
const command_1 = __nccwpck_require__(7351);
|
|
const file_command_1 = __nccwpck_require__(717);
|
|
const utils_1 = __nccwpck_require__(5278);
|
|
const os = __importStar(__nccwpck_require__(2087));
|
|
const path = __importStar(__nccwpck_require__(5622));
|
|
/**
|
|
* The code to exit an action
|
|
*/
|
|
var ExitCode;
|
|
(function (ExitCode) {
|
|
/**
|
|
* A code indicating that the action was successful
|
|
*/
|
|
ExitCode[ExitCode["Success"] = 0] = "Success";
|
|
/**
|
|
* A code indicating that the action was a failure
|
|
*/
|
|
ExitCode[ExitCode["Failure"] = 1] = "Failure";
|
|
})(ExitCode = exports.ExitCode || (exports.ExitCode = {}));
|
|
//-----------------------------------------------------------------------
|
|
// Variables
|
|
//-----------------------------------------------------------------------
|
|
/**
|
|
* Sets env variable for this action and future actions in the job
|
|
* @param name the name of the variable to set
|
|
* @param val the value of the variable. Non-string values will be converted to a string via JSON.stringify
|
|
*/
|
|
// eslint-disable-next-line @typescript-eslint/no-explicit-any
|
|
function exportVariable(name, val) {
|
|
const convertedVal = utils_1.toCommandValue(val);
|
|
process.env[name] = convertedVal;
|
|
const filePath = process.env['GITHUB_ENV'] || '';
|
|
if (filePath) {
|
|
const delimiter = '_GitHubActionsFileCommandDelimeter_';
|
|
const commandValue = `${name}<<${delimiter}${os.EOL}${convertedVal}${os.EOL}${delimiter}`;
|
|
file_command_1.issueCommand('ENV', commandValue);
|
|
}
|
|
else {
|
|
command_1.issueCommand('set-env', { name }, convertedVal);
|
|
}
|
|
}
|
|
exports.exportVariable = exportVariable;
|
|
/**
|
|
* Registers a secret which will get masked from logs
|
|
* @param secret value of the secret
|
|
*/
|
|
function setSecret(secret) {
|
|
command_1.issueCommand('add-mask', {}, secret);
|
|
}
|
|
exports.setSecret = setSecret;
|
|
/**
|
|
* Prepends inputPath to the PATH (for this action and future actions)
|
|
* @param inputPath
|
|
*/
|
|
function addPath(inputPath) {
|
|
const filePath = process.env['GITHUB_PATH'] || '';
|
|
if (filePath) {
|
|
file_command_1.issueCommand('PATH', inputPath);
|
|
}
|
|
else {
|
|
command_1.issueCommand('add-path', {}, inputPath);
|
|
}
|
|
process.env['PATH'] = `${inputPath}${path.delimiter}${process.env['PATH']}`;
|
|
}
|
|
exports.addPath = addPath;
|
|
/**
|
|
* Gets the value of an input.
|
|
* Unless trimWhitespace is set to false in InputOptions, the value is also trimmed.
|
|
* Returns an empty string if the value is not defined.
|
|
*
|
|
* @param name name of the input to get
|
|
* @param options optional. See InputOptions.
|
|
* @returns string
|
|
*/
|
|
function getInput(name, options) {
|
|
const val = process.env[`INPUT_${name.replace(/ /g, '_').toUpperCase()}`] || '';
|
|
if (options && options.required && !val) {
|
|
throw new Error(`Input required and not supplied: ${name}`);
|
|
}
|
|
if (options && options.trimWhitespace === false) {
|
|
return val;
|
|
}
|
|
return val.trim();
|
|
}
|
|
exports.getInput = getInput;
|
|
/**
|
|
* Gets the values of an multiline input. Each value is also trimmed.
|
|
*
|
|
* @param name name of the input to get
|
|
* @param options optional. See InputOptions.
|
|
* @returns string[]
|
|
*
|
|
*/
|
|
function getMultilineInput(name, options) {
|
|
const inputs = getInput(name, options)
|
|
.split('\n')
|
|
.filter(x => x !== '');
|
|
return inputs;
|
|
}
|
|
exports.getMultilineInput = getMultilineInput;
|
|
/**
|
|
* Gets the input value of the boolean type in the YAML 1.2 "core schema" specification.
|
|
* Support boolean input list: `true | True | TRUE | false | False | FALSE` .
|
|
* The return value is also in boolean type.
|
|
* ref: https://yaml.org/spec/1.2/spec.html#id2804923
|
|
*
|
|
* @param name name of the input to get
|
|
* @param options optional. See InputOptions.
|
|
* @returns boolean
|
|
*/
|
|
function getBooleanInput(name, options) {
|
|
const trueValue = ['true', 'True', 'TRUE'];
|
|
const falseValue = ['false', 'False', 'FALSE'];
|
|
const val = getInput(name, options);
|
|
if (trueValue.includes(val))
|
|
return true;
|
|
if (falseValue.includes(val))
|
|
return false;
|
|
throw new TypeError(`Input does not meet YAML 1.2 "Core Schema" specification: ${name}\n` +
|
|
`Support boolean input list: \`true | True | TRUE | false | False | FALSE\``);
|
|
}
|
|
exports.getBooleanInput = getBooleanInput;
|
|
/**
|
|
* Sets the value of an output.
|
|
*
|
|
* @param name name of the output to set
|
|
* @param value value to store. Non-string values will be converted to a string via JSON.stringify
|
|
*/
|
|
// eslint-disable-next-line @typescript-eslint/no-explicit-any
|
|
function setOutput(name, value) {
|
|
process.stdout.write(os.EOL);
|
|
command_1.issueCommand('set-output', { name }, value);
|
|
}
|
|
exports.setOutput = setOutput;
|
|
/**
|
|
* Enables or disables the echoing of commands into stdout for the rest of the step.
|
|
* Echoing is disabled by default if ACTIONS_STEP_DEBUG is not set.
|
|
*
|
|
*/
|
|
function setCommandEcho(enabled) {
|
|
command_1.issue('echo', enabled ? 'on' : 'off');
|
|
}
|
|
exports.setCommandEcho = setCommandEcho;
|
|
//-----------------------------------------------------------------------
|
|
// Results
|
|
//-----------------------------------------------------------------------
|
|
/**
|
|
* Sets the action status to failed.
|
|
* When the action exits it will be with an exit code of 1
|
|
* @param message add error issue message
|
|
*/
|
|
function setFailed(message) {
|
|
process.exitCode = ExitCode.Failure;
|
|
error(message);
|
|
}
|
|
exports.setFailed = setFailed;
|
|
//-----------------------------------------------------------------------
|
|
// Logging Commands
|
|
//-----------------------------------------------------------------------
|
|
/**
|
|
* Gets whether Actions Step Debug is on or not
|
|
*/
|
|
function isDebug() {
|
|
return process.env['RUNNER_DEBUG'] === '1';
|
|
}
|
|
exports.isDebug = isDebug;
|
|
/**
|
|
* Writes debug message to user log
|
|
* @param message debug message
|
|
*/
|
|
function debug(message) {
|
|
command_1.issueCommand('debug', {}, message);
|
|
}
|
|
exports.debug = debug;
|
|
/**
|
|
* Adds an error issue
|
|
* @param message error issue message. Errors will be converted to string via toString()
|
|
*/
|
|
function error(message) {
|
|
command_1.issue('error', message instanceof Error ? message.toString() : message);
|
|
}
|
|
exports.error = error;
|
|
/**
|
|
* Adds an warning issue
|
|
* @param message warning issue message. Errors will be converted to string via toString()
|
|
*/
|
|
function warning(message) {
|
|
command_1.issue('warning', message instanceof Error ? message.toString() : message);
|
|
}
|
|
exports.warning = warning;
|
|
/**
|
|
* Writes info to log with console.log.
|
|
* @param message info message
|
|
*/
|
|
function info(message) {
|
|
process.stdout.write(message + os.EOL);
|
|
}
|
|
exports.info = info;
|
|
/**
|
|
* Begin an output group.
|
|
*
|
|
* Output until the next `groupEnd` will be foldable in this group
|
|
*
|
|
* @param name The name of the output group
|
|
*/
|
|
function startGroup(name) {
|
|
command_1.issue('group', name);
|
|
}
|
|
exports.startGroup = startGroup;
|
|
/**
|
|
* End an output group.
|
|
*/
|
|
function endGroup() {
|
|
command_1.issue('endgroup');
|
|
}
|
|
exports.endGroup = endGroup;
|
|
/**
|
|
* Wrap an asynchronous function call in a group.
|
|
*
|
|
* Returns the same type as the function itself.
|
|
*
|
|
* @param name The name of the group
|
|
* @param fn The function to wrap in the group
|
|
*/
|
|
function group(name, fn) {
|
|
return __awaiter(this, void 0, void 0, function* () {
|
|
startGroup(name);
|
|
let result;
|
|
try {
|
|
result = yield fn();
|
|
}
|
|
finally {
|
|
endGroup();
|
|
}
|
|
return result;
|
|
});
|
|
}
|
|
exports.group = group;
|
|
//-----------------------------------------------------------------------
|
|
// Wrapper action state
|
|
//-----------------------------------------------------------------------
|
|
/**
|
|
* Saves state for current action, the state can only be retrieved by this action's post job execution.
|
|
*
|
|
* @param name name of the state to store
|
|
* @param value value to store. Non-string values will be converted to a string via JSON.stringify
|
|
*/
|
|
// eslint-disable-next-line @typescript-eslint/no-explicit-any
|
|
function saveState(name, value) {
|
|
command_1.issueCommand('save-state', { name }, value);
|
|
}
|
|
exports.saveState = saveState;
|
|
/**
|
|
* Gets the value of an state set by this action's main execution.
|
|
*
|
|
* @param name name of the state to get
|
|
* @returns string
|
|
*/
|
|
function getState(name) {
|
|
return process.env[`STATE_${name}`] || '';
|
|
}
|
|
exports.getState = getState;
|
|
//# sourceMappingURL=core.js.map
|
|
|
|
/***/ }),
|
|
|
|
/***/ 717:
|
|
/***/ (function(__unused_webpack_module, exports, __nccwpck_require__) {
|
|
|
|
"use strict";
|
|
|
|
// For internal use, subject to change.
|
|
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
if (k2 === undefined) k2 = k;
|
|
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
}) : (function(o, m, k, k2) {
|
|
if (k2 === undefined) k2 = k;
|
|
o[k2] = m[k];
|
|
}));
|
|
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
}) : function(o, v) {
|
|
o["default"] = v;
|
|
});
|
|
var __importStar = (this && this.__importStar) || function (mod) {
|
|
if (mod && mod.__esModule) return mod;
|
|
var result = {};
|
|
if (mod != null) for (var k in mod) if (k !== "default" && Object.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
__setModuleDefault(result, mod);
|
|
return result;
|
|
};
|
|
Object.defineProperty(exports, "__esModule", ({ value: true }));
|
|
exports.issueCommand = void 0;
|
|
// We use any as a valid input type
|
|
/* eslint-disable @typescript-eslint/no-explicit-any */
|
|
const fs = __importStar(__nccwpck_require__(5747));
|
|
const os = __importStar(__nccwpck_require__(2087));
|
|
const utils_1 = __nccwpck_require__(5278);
|
|
function issueCommand(command, message) {
|
|
const filePath = process.env[`GITHUB_${command}`];
|
|
if (!filePath) {
|
|
throw new Error(`Unable to find environment variable for file command ${command}`);
|
|
}
|
|
if (!fs.existsSync(filePath)) {
|
|
throw new Error(`Missing file at path: ${filePath}`);
|
|
}
|
|
fs.appendFileSync(filePath, `${utils_1.toCommandValue(message)}${os.EOL}`, {
|
|
encoding: 'utf8'
|
|
});
|
|
}
|
|
exports.issueCommand = issueCommand;
|
|
//# sourceMappingURL=file-command.js.map
|
|
|
|
/***/ }),
|
|
|
|
/***/ 5278:
|
|
/***/ ((__unused_webpack_module, exports) => {
|
|
|
|
"use strict";
|
|
|
|
// We use any as a valid input type
|
|
/* eslint-disable @typescript-eslint/no-explicit-any */
|
|
Object.defineProperty(exports, "__esModule", ({ value: true }));
|
|
exports.toCommandValue = void 0;
|
|
/**
|
|
* Sanitizes an input into a string so it can be passed into issueCommand safely
|
|
* @param input input to sanitize into a string
|
|
*/
|
|
function toCommandValue(input) {
|
|
if (input === null || input === undefined) {
|
|
return '';
|
|
}
|
|
else if (typeof input === 'string' || input instanceof String) {
|
|
return input;
|
|
}
|
|
return JSON.stringify(input);
|
|
}
|
|
exports.toCommandValue = toCommandValue;
|
|
//# sourceMappingURL=utils.js.map
|
|
|
|
/***/ }),
|
|
|
|
/***/ 4087:
|
|
/***/ ((__unused_webpack_module, exports, __nccwpck_require__) => {
|
|
|
|
"use strict";
|
|
|
|
Object.defineProperty(exports, "__esModule", ({ value: true }));
|
|
exports.Context = void 0;
|
|
const fs_1 = __nccwpck_require__(5747);
|
|
const os_1 = __nccwpck_require__(2087);
|
|
class Context {
|
|
/**
|
|
* Hydrate the context from the environment
|
|
*/
|
|
constructor() {
|
|
var _a, _b, _c;
|
|
this.payload = {};
|
|
if (process.env.GITHUB_EVENT_PATH) {
|
|
if (fs_1.existsSync(process.env.GITHUB_EVENT_PATH)) {
|
|
this.payload = JSON.parse(fs_1.readFileSync(process.env.GITHUB_EVENT_PATH, { encoding: 'utf8' }));
|
|
}
|
|
else {
|
|
const path = process.env.GITHUB_EVENT_PATH;
|
|
process.stdout.write(`GITHUB_EVENT_PATH ${path} does not exist${os_1.EOL}`);
|
|
}
|
|
}
|
|
this.eventName = process.env.GITHUB_EVENT_NAME;
|
|
this.sha = process.env.GITHUB_SHA;
|
|
this.ref = process.env.GITHUB_REF;
|
|
this.workflow = process.env.GITHUB_WORKFLOW;
|
|
this.action = process.env.GITHUB_ACTION;
|
|
this.actor = process.env.GITHUB_ACTOR;
|
|
this.job = process.env.GITHUB_JOB;
|
|
this.runNumber = parseInt(process.env.GITHUB_RUN_NUMBER, 10);
|
|
this.runId = parseInt(process.env.GITHUB_RUN_ID, 10);
|
|
this.apiUrl = (_a = process.env.GITHUB_API_URL) !== null && _a !== void 0 ? _a : `https://api.github.com`;
|
|
this.serverUrl = (_b = process.env.GITHUB_SERVER_URL) !== null && _b !== void 0 ? _b : `https://github.com`;
|
|
this.graphqlUrl = (_c = process.env.GITHUB_GRAPHQL_URL) !== null && _c !== void 0 ? _c : `https://api.github.com/graphql`;
|
|
}
|
|
get issue() {
|
|
const payload = this.payload;
|
|
return Object.assign(Object.assign({}, this.repo), { number: (payload.issue || payload.pull_request || payload).number });
|
|
}
|
|
get repo() {
|
|
if (process.env.GITHUB_REPOSITORY) {
|
|
const [owner, repo] = process.env.GITHUB_REPOSITORY.split('/');
|
|
return { owner, repo };
|
|
}
|
|
if (this.payload.repository) {
|
|
return {
|
|
owner: this.payload.repository.owner.login,
|
|
repo: this.payload.repository.name
|
|
};
|
|
}
|
|
throw new Error("context.repo requires a GITHUB_REPOSITORY environment variable like 'owner/repo'");
|
|
}
|
|
}
|
|
exports.Context = Context;
|
|
//# sourceMappingURL=context.js.map
|
|
|
|
/***/ }),
|
|
|
|
/***/ 5438:
|
|
/***/ (function(__unused_webpack_module, exports, __nccwpck_require__) {
|
|
|
|
"use strict";
|
|
|
|
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
if (k2 === undefined) k2 = k;
|
|
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
}) : (function(o, m, k, k2) {
|
|
if (k2 === undefined) k2 = k;
|
|
o[k2] = m[k];
|
|
}));
|
|
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
}) : function(o, v) {
|
|
o["default"] = v;
|
|
});
|
|
var __importStar = (this && this.__importStar) || function (mod) {
|
|
if (mod && mod.__esModule) return mod;
|
|
var result = {};
|
|
if (mod != null) for (var k in mod) if (k !== "default" && Object.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
__setModuleDefault(result, mod);
|
|
return result;
|
|
};
|
|
Object.defineProperty(exports, "__esModule", ({ value: true }));
|
|
exports.getOctokit = exports.context = void 0;
|
|
const Context = __importStar(__nccwpck_require__(4087));
|
|
const utils_1 = __nccwpck_require__(3030);
|
|
exports.context = new Context.Context();
|
|
/**
|
|
* Returns a hydrated octokit ready to use for GitHub Actions
|
|
*
|
|
* @param token the repo PAT or GITHUB_TOKEN
|
|
* @param options other options to set
|
|
*/
|
|
function getOctokit(token, options) {
|
|
return new utils_1.GitHub(utils_1.getOctokitOptions(token, options));
|
|
}
|
|
exports.getOctokit = getOctokit;
|
|
//# sourceMappingURL=github.js.map
|
|
|
|
/***/ }),
|
|
|
|
/***/ 7914:
|
|
/***/ (function(__unused_webpack_module, exports, __nccwpck_require__) {
|
|
|
|
"use strict";
|
|
|
|
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
if (k2 === undefined) k2 = k;
|
|
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
}) : (function(o, m, k, k2) {
|
|
if (k2 === undefined) k2 = k;
|
|
o[k2] = m[k];
|
|
}));
|
|
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
}) : function(o, v) {
|
|
o["default"] = v;
|
|
});
|
|
var __importStar = (this && this.__importStar) || function (mod) {
|
|
if (mod && mod.__esModule) return mod;
|
|
var result = {};
|
|
if (mod != null) for (var k in mod) if (k !== "default" && Object.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
__setModuleDefault(result, mod);
|
|
return result;
|
|
};
|
|
Object.defineProperty(exports, "__esModule", ({ value: true }));
|
|
exports.getApiBaseUrl = exports.getProxyAgent = exports.getAuthString = void 0;
|
|
const httpClient = __importStar(__nccwpck_require__(9925));
|
|
function getAuthString(token, options) {
|
|
if (!token && !options.auth) {
|
|
throw new Error('Parameter token or opts.auth is required');
|
|
}
|
|
else if (token && options.auth) {
|
|
throw new Error('Parameters token and opts.auth may not both be specified');
|
|
}
|
|
return typeof options.auth === 'string' ? options.auth : `token ${token}`;
|
|
}
|
|
exports.getAuthString = getAuthString;
|
|
function getProxyAgent(destinationUrl) {
|
|
const hc = new httpClient.HttpClient();
|
|
return hc.getAgent(destinationUrl);
|
|
}
|
|
exports.getProxyAgent = getProxyAgent;
|
|
function getApiBaseUrl() {
|
|
return process.env['GITHUB_API_URL'] || 'https://api.github.com';
|
|
}
|
|
exports.getApiBaseUrl = getApiBaseUrl;
|
|
//# sourceMappingURL=utils.js.map
|
|
|
|
/***/ }),
|
|
|
|
/***/ 3030:
|
|
/***/ (function(__unused_webpack_module, exports, __nccwpck_require__) {
|
|
|
|
"use strict";
|
|
|
|
var __createBinding = (this && this.__createBinding) || (Object.create ? (function(o, m, k, k2) {
|
|
if (k2 === undefined) k2 = k;
|
|
Object.defineProperty(o, k2, { enumerable: true, get: function() { return m[k]; } });
|
|
}) : (function(o, m, k, k2) {
|
|
if (k2 === undefined) k2 = k;
|
|
o[k2] = m[k];
|
|
}));
|
|
var __setModuleDefault = (this && this.__setModuleDefault) || (Object.create ? (function(o, v) {
|
|
Object.defineProperty(o, "default", { enumerable: true, value: v });
|
|
}) : function(o, v) {
|
|
o["default"] = v;
|
|
});
|
|
var __importStar = (this && this.__importStar) || function (mod) {
|
|
if (mod && mod.__esModule) return mod;
|
|
var result = {};
|
|
if (mod != null) for (var k in mod) if (k !== "default" && Object.hasOwnProperty.call(mod, k)) __createBinding(result, mod, k);
|
|
__setModuleDefault(result, mod);
|
|
return result;
|
|
};
|
|
Object.defineProperty(exports, "__esModule", ({ value: true }));
|
|
exports.getOctokitOptions = exports.GitHub = exports.context = void 0;
|
|
const Context = __importStar(__nccwpck_require__(4087));
|
|
const Utils = __importStar(__nccwpck_require__(7914));
|
|
// octokit + plugins
|
|
const core_1 = __nccwpck_require__(6762);
|
|
const plugin_rest_endpoint_methods_1 = __nccwpck_require__(3044);
|
|
const plugin_paginate_rest_1 = __nccwpck_require__(4193);
|
|
exports.context = new Context.Context();
|
|
const baseUrl = Utils.getApiBaseUrl();
|
|
const defaults = {
|
|
baseUrl,
|
|
request: {
|
|
agent: Utils.getProxyAgent(baseUrl)
|
|
}
|
|
};
|
|
exports.GitHub = core_1.Octokit.plugin(plugin_rest_endpoint_methods_1.restEndpointMethods, plugin_paginate_rest_1.paginateRest).defaults(defaults);
|
|
/**
|
|
* Convience function to correctly format Octokit Options to pass into the constructor.
|
|
*
|
|
* @param token the repo PAT or GITHUB_TOKEN
|
|
* @param options other options to set
|
|
*/
|
|
function getOctokitOptions(token, options) {
|
|
const opts = Object.assign({}, options || {}); // Shallow clone - don't mutate the object provided by the caller
|
|
// Auth
|
|
const auth = Utils.getAuthString(token, opts);
|
|
if (auth) {
|
|
opts.auth = auth;
|
|
}
|
|
return opts;
|
|
}
|
|
exports.getOctokitOptions = getOctokitOptions;
|
|
//# sourceMappingURL=utils.js.map
|
|
|
|
/***/ }),
|
|
|
|
/***/ 9925:
|
|
/***/ ((__unused_webpack_module, exports, __nccwpck_require__) => {
|
|
|
|
"use strict";
|
|
|
|
Object.defineProperty(exports, "__esModule", ({ value: true }));
|
|
const http = __nccwpck_require__(8605);
|
|
const https = __nccwpck_require__(7211);
|
|
const pm = __nccwpck_require__(6443);
|
|
let tunnel;
|
|
var HttpCodes;
|
|
(function (HttpCodes) {
|
|
HttpCodes[HttpCodes["OK"] = 200] = "OK";
|
|
HttpCodes[HttpCodes["MultipleChoices"] = 300] = "MultipleChoices";
|
|
HttpCodes[HttpCodes["MovedPermanently"] = 301] = "MovedPermanently";
|
|
HttpCodes[HttpCodes["ResourceMoved"] = 302] = "ResourceMoved";
|
|
HttpCodes[HttpCodes["SeeOther"] = 303] = "SeeOther";
|
|
HttpCodes[HttpCodes["NotModified"] = 304] = "NotModified";
|
|
HttpCodes[HttpCodes["UseProxy"] = 305] = "UseProxy";
|
|
HttpCodes[HttpCodes["SwitchProxy"] = 306] = "SwitchProxy";
|
|
HttpCodes[HttpCodes["TemporaryRedirect"] = 307] = "TemporaryRedirect";
|
|
HttpCodes[HttpCodes["PermanentRedirect"] = 308] = "PermanentRedirect";
|
|
HttpCodes[HttpCodes["BadRequest"] = 400] = "BadRequest";
|
|
HttpCodes[HttpCodes["Unauthorized"] = 401] = "Unauthorized";
|
|
HttpCodes[HttpCodes["PaymentRequired"] = 402] = "PaymentRequired";
|
|
HttpCodes[HttpCodes["Forbidden"] = 403] = "Forbidden";
|
|
HttpCodes[HttpCodes["NotFound"] = 404] = "NotFound";
|
|
HttpCodes[HttpCodes["MethodNotAllowed"] = 405] = "MethodNotAllowed";
|
|
HttpCodes[HttpCodes["NotAcceptable"] = 406] = "NotAcceptable";
|
|
HttpCodes[HttpCodes["ProxyAuthenticationRequired"] = 407] = "ProxyAuthenticationRequired";
|
|
HttpCodes[HttpCodes["RequestTimeout"] = 408] = "RequestTimeout";
|
|
HttpCodes[HttpCodes["Conflict"] = 409] = "Conflict";
|
|
HttpCodes[HttpCodes["Gone"] = 410] = "Gone";
|
|
HttpCodes[HttpCodes["TooManyRequests"] = 429] = "TooManyRequests";
|
|
HttpCodes[HttpCodes["InternalServerError"] = 500] = "InternalServerError";
|
|
HttpCodes[HttpCodes["NotImplemented"] = 501] = "NotImplemented";
|
|
HttpCodes[HttpCodes["BadGateway"] = 502] = "BadGateway";
|
|
HttpCodes[HttpCodes["ServiceUnavailable"] = 503] = "ServiceUnavailable";
|
|
HttpCodes[HttpCodes["GatewayTimeout"] = 504] = "GatewayTimeout";
|
|
})(HttpCodes = exports.HttpCodes || (exports.HttpCodes = {}));
|
|
var Headers;
|
|
(function (Headers) {
|
|
Headers["Accept"] = "accept";
|
|
Headers["ContentType"] = "content-type";
|
|
})(Headers = exports.Headers || (exports.Headers = {}));
|
|
var MediaTypes;
|
|
(function (MediaTypes) {
|
|
MediaTypes["ApplicationJson"] = "application/json";
|
|
})(MediaTypes = exports.MediaTypes || (exports.MediaTypes = {}));
|
|
/**
|
|
* Returns the proxy URL, depending upon the supplied url and proxy environment variables.
|
|
* @param serverUrl The server URL where the request will be sent. For example, https://api.github.com
|
|
*/
|
|
function getProxyUrl(serverUrl) {
|
|
let proxyUrl = pm.getProxyUrl(new URL(serverUrl));
|
|
return proxyUrl ? proxyUrl.href : '';
|
|
}
|
|
exports.getProxyUrl = getProxyUrl;
|
|
const HttpRedirectCodes = [
|
|
HttpCodes.MovedPermanently,
|
|
HttpCodes.ResourceMoved,
|
|
HttpCodes.SeeOther,
|
|
HttpCodes.TemporaryRedirect,
|
|
HttpCodes.PermanentRedirect
|
|
];
|
|
const HttpResponseRetryCodes = [
|
|
HttpCodes.BadGateway,
|
|
HttpCodes.ServiceUnavailable,
|
|
HttpCodes.GatewayTimeout
|
|
];
|
|
const RetryableHttpVerbs = ['OPTIONS', 'GET', 'DELETE', 'HEAD'];
|
|
const ExponentialBackoffCeiling = 10;
|
|
const ExponentialBackoffTimeSlice = 5;
|
|
class HttpClientError extends Error {
|
|
constructor(message, statusCode) {
|
|
super(message);
|
|
this.name = 'HttpClientError';
|
|
this.statusCode = statusCode;
|
|
Object.setPrototypeOf(this, HttpClientError.prototype);
|
|
}
|
|
}
|
|
exports.HttpClientError = HttpClientError;
|
|
class HttpClientResponse {
|
|
constructor(message) {
|
|
this.message = message;
|
|
}
|
|
readBody() {
|
|
return new Promise(async (resolve, reject) => {
|
|
let output = Buffer.alloc(0);
|
|
this.message.on('data', (chunk) => {
|
|
output = Buffer.concat([output, chunk]);
|
|
});
|
|
this.message.on('end', () => {
|
|
resolve(output.toString());
|
|
});
|
|
});
|
|
}
|
|
}
|
|
exports.HttpClientResponse = HttpClientResponse;
|
|
function isHttps(requestUrl) {
|
|
let parsedUrl = new URL(requestUrl);
|
|
return parsedUrl.protocol === 'https:';
|
|
}
|
|
exports.isHttps = isHttps;
|
|
class HttpClient {
|
|
constructor(userAgent, handlers, requestOptions) {
|
|
this._ignoreSslError = false;
|
|
this._allowRedirects = true;
|
|
this._allowRedirectDowngrade = false;
|
|
this._maxRedirects = 50;
|
|
this._allowRetries = false;
|
|
this._maxRetries = 1;
|
|
this._keepAlive = false;
|
|
this._disposed = false;
|
|
this.userAgent = userAgent;
|
|
this.handlers = handlers || [];
|
|
this.requestOptions = requestOptions;
|
|
if (requestOptions) {
|
|
if (requestOptions.ignoreSslError != null) {
|
|
this._ignoreSslError = requestOptions.ignoreSslError;
|
|
}
|
|
this._socketTimeout = requestOptions.socketTimeout;
|
|
if (requestOptions.allowRedirects != null) {
|
|
this._allowRedirects = requestOptions.allowRedirects;
|
|
}
|
|
if (requestOptions.allowRedirectDowngrade != null) {
|
|
this._allowRedirectDowngrade = requestOptions.allowRedirectDowngrade;
|
|
}
|
|
if (requestOptions.maxRedirects != null) {
|
|
this._maxRedirects = Math.max(requestOptions.maxRedirects, 0);
|
|
}
|
|
if (requestOptions.keepAlive != null) {
|
|
this._keepAlive = requestOptions.keepAlive;
|
|
}
|
|
if (requestOptions.allowRetries != null) {
|
|
this._allowRetries = requestOptions.allowRetries;
|
|
}
|
|
if (requestOptions.maxRetries != null) {
|
|
this._maxRetries = requestOptions.maxRetries;
|
|
}
|
|
}
|
|
}
|
|
options(requestUrl, additionalHeaders) {
|
|
return this.request('OPTIONS', requestUrl, null, additionalHeaders || {});
|
|
}
|
|
get(requestUrl, additionalHeaders) {
|
|
return this.request('GET', requestUrl, null, additionalHeaders || {});
|
|
}
|
|
del(requestUrl, additionalHeaders) {
|
|
return this.request('DELETE', requestUrl, null, additionalHeaders || {});
|
|
}
|
|
post(requestUrl, data, additionalHeaders) {
|
|
return this.request('POST', requestUrl, data, additionalHeaders || {});
|
|
}
|
|
patch(requestUrl, data, additionalHeaders) {
|
|
return this.request('PATCH', requestUrl, data, additionalHeaders || {});
|
|
}
|
|
put(requestUrl, data, additionalHeaders) {
|
|
return this.request('PUT', requestUrl, data, additionalHeaders || {});
|
|
}
|
|
head(requestUrl, additionalHeaders) {
|
|
return this.request('HEAD', requestUrl, null, additionalHeaders || {});
|
|
}
|
|
sendStream(verb, requestUrl, stream, additionalHeaders) {
|
|
return this.request(verb, requestUrl, stream, additionalHeaders);
|
|
}
|
|
/**
|
|
* Gets a typed object from an endpoint
|
|
* Be aware that not found returns a null. Other errors (4xx, 5xx) reject the promise
|
|
*/
|
|
async getJson(requestUrl, additionalHeaders = {}) {
|
|
additionalHeaders[Headers.Accept] = this._getExistingOrDefaultHeader(additionalHeaders, Headers.Accept, MediaTypes.ApplicationJson);
|
|
let res = await this.get(requestUrl, additionalHeaders);
|
|
return this._processResponse(res, this.requestOptions);
|
|
}
|
|
async postJson(requestUrl, obj, additionalHeaders = {}) {
|
|
let data = JSON.stringify(obj, null, 2);
|
|
additionalHeaders[Headers.Accept] = this._getExistingOrDefaultHeader(additionalHeaders, Headers.Accept, MediaTypes.ApplicationJson);
|
|
additionalHeaders[Headers.ContentType] = this._getExistingOrDefaultHeader(additionalHeaders, Headers.ContentType, MediaTypes.ApplicationJson);
|
|
let res = await this.post(requestUrl, data, additionalHeaders);
|
|
return this._processResponse(res, this.requestOptions);
|
|
}
|
|
async putJson(requestUrl, obj, additionalHeaders = {}) {
|
|
let data = JSON.stringify(obj, null, 2);
|
|
additionalHeaders[Headers.Accept] = this._getExistingOrDefaultHeader(additionalHeaders, Headers.Accept, MediaTypes.ApplicationJson);
|
|
additionalHeaders[Headers.ContentType] = this._getExistingOrDefaultHeader(additionalHeaders, Headers.ContentType, MediaTypes.ApplicationJson);
|
|
let res = await this.put(requestUrl, data, additionalHeaders);
|
|
return this._processResponse(res, this.requestOptions);
|
|
}
|
|
async patchJson(requestUrl, obj, additionalHeaders = {}) {
|
|
let data = JSON.stringify(obj, null, 2);
|
|
additionalHeaders[Headers.Accept] = this._getExistingOrDefaultHeader(additionalHeaders, Headers.Accept, MediaTypes.ApplicationJson);
|
|
additionalHeaders[Headers.ContentType] = this._getExistingOrDefaultHeader(additionalHeaders, Headers.ContentType, MediaTypes.ApplicationJson);
|
|
let res = await this.patch(requestUrl, data, additionalHeaders);
|
|
return this._processResponse(res, this.requestOptions);
|
|
}
|
|
/**
|
|
* Makes a raw http request.
|
|
* All other methods such as get, post, patch, and request ultimately call this.
|
|
* Prefer get, del, post and patch
|
|
*/
|
|
async request(verb, requestUrl, data, headers) {
|
|
if (this._disposed) {
|
|
throw new Error('Client has already been disposed.');
|
|
}
|
|
let parsedUrl = new URL(requestUrl);
|
|
let info = this._prepareRequest(verb, parsedUrl, headers);
|
|
// Only perform retries on reads since writes may not be idempotent.
|
|
let maxTries = this._allowRetries && RetryableHttpVerbs.indexOf(verb) != -1
|
|
? this._maxRetries + 1
|
|
: 1;
|
|
let numTries = 0;
|
|
let response;
|
|
while (numTries < maxTries) {
|
|
response = await this.requestRaw(info, data);
|
|
// Check if it's an authentication challenge
|
|
if (response &&
|
|
response.message &&
|
|
response.message.statusCode === HttpCodes.Unauthorized) {
|
|
let authenticationHandler;
|
|
for (let i = 0; i < this.handlers.length; i++) {
|
|
if (this.handlers[i].canHandleAuthentication(response)) {
|
|
authenticationHandler = this.handlers[i];
|
|
break;
|
|
}
|
|
}
|
|
if (authenticationHandler) {
|
|
return authenticationHandler.handleAuthentication(this, info, data);
|
|
}
|
|
else {
|
|
// We have received an unauthorized response but have no handlers to handle it.
|
|
// Let the response return to the caller.
|
|
return response;
|
|
}
|
|
}
|
|
let redirectsRemaining = this._maxRedirects;
|
|
while (HttpRedirectCodes.indexOf(response.message.statusCode) != -1 &&
|
|
this._allowRedirects &&
|
|
redirectsRemaining > 0) {
|
|
const redirectUrl = response.message.headers['location'];
|
|
if (!redirectUrl) {
|
|
// if there's no location to redirect to, we won't
|
|
break;
|
|
}
|
|
let parsedRedirectUrl = new URL(redirectUrl);
|
|
if (parsedUrl.protocol == 'https:' &&
|
|
parsedUrl.protocol != parsedRedirectUrl.protocol &&
|
|
!this._allowRedirectDowngrade) {
|
|
throw new Error('Redirect from HTTPS to HTTP protocol. This downgrade is not allowed for security reasons. If you want to allow this behavior, set the allowRedirectDowngrade option to true.');
|
|
}
|
|
// we need to finish reading the response before reassigning response
|
|
// which will leak the open socket.
|
|
await response.readBody();
|
|
// strip authorization header if redirected to a different hostname
|
|
if (parsedRedirectUrl.hostname !== parsedUrl.hostname) {
|
|
for (let header in headers) {
|
|
// header names are case insensitive
|
|
if (header.toLowerCase() === 'authorization') {
|
|
delete headers[header];
|
|
}
|
|
}
|
|
}
|
|
// let's make the request with the new redirectUrl
|
|
info = this._prepareRequest(verb, parsedRedirectUrl, headers);
|
|
response = await this.requestRaw(info, data);
|
|
redirectsRemaining--;
|
|
}
|
|
if (HttpResponseRetryCodes.indexOf(response.message.statusCode) == -1) {
|
|
// If not a retry code, return immediately instead of retrying
|
|
return response;
|
|
}
|
|
numTries += 1;
|
|
if (numTries < maxTries) {
|
|
await response.readBody();
|
|
await this._performExponentialBackoff(numTries);
|
|
}
|
|
}
|
|
return response;
|
|
}
|
|
/**
|
|
* Needs to be called if keepAlive is set to true in request options.
|
|
*/
|
|
dispose() {
|
|
if (this._agent) {
|
|
this._agent.destroy();
|
|
}
|
|
this._disposed = true;
|
|
}
|
|
/**
|
|
* Raw request.
|
|
* @param info
|
|
* @param data
|
|
*/
|
|
requestRaw(info, data) {
|
|
return new Promise((resolve, reject) => {
|
|
let callbackForResult = function (err, res) {
|
|
if (err) {
|
|
reject(err);
|
|
}
|
|
resolve(res);
|
|
};
|
|
this.requestRawWithCallback(info, data, callbackForResult);
|
|
});
|
|
}
|
|
/**
|
|
* Raw request with callback.
|
|
* @param info
|
|
* @param data
|
|
* @param onResult
|
|
*/
|
|
requestRawWithCallback(info, data, onResult) {
|
|
let socket;
|
|
if (typeof data === 'string') {
|
|
info.options.headers['Content-Length'] = Buffer.byteLength(data, 'utf8');
|
|
}
|
|
let callbackCalled = false;
|
|
let handleResult = (err, res) => {
|
|
if (!callbackCalled) {
|
|
callbackCalled = true;
|
|
onResult(err, res);
|
|
}
|
|
};
|
|
let req = info.httpModule.request(info.options, (msg) => {
|
|
let res = new HttpClientResponse(msg);
|
|
handleResult(null, res);
|
|
});
|
|
req.on('socket', sock => {
|
|
socket = sock;
|
|
});
|
|
// If we ever get disconnected, we want the socket to timeout eventually
|
|
req.setTimeout(this._socketTimeout || 3 * 60000, () => {
|
|
if (socket) {
|
|
socket.end();
|
|
}
|
|
handleResult(new Error('Request timeout: ' + info.options.path), null);
|
|
});
|
|
req.on('error', function (err) {
|
|
// err has statusCode property
|
|
// res should have headers
|
|
handleResult(err, null);
|
|
});
|
|
if (data && typeof data === 'string') {
|
|
req.write(data, 'utf8');
|
|
}
|
|
if (data && typeof data !== 'string') {
|
|
data.on('close', function () {
|
|
req.end();
|
|
});
|
|
data.pipe(req);
|
|
}
|
|
else {
|
|
req.end();
|
|
}
|
|
}
|
|
/**
|
|
* Gets an http agent. This function is useful when you need an http agent that handles
|
|
* routing through a proxy server - depending upon the url and proxy environment variables.
|
|
* @param serverUrl The server URL where the request will be sent. For example, https://api.github.com
|
|
*/
|
|
getAgent(serverUrl) {
|
|
let parsedUrl = new URL(serverUrl);
|
|
return this._getAgent(parsedUrl);
|
|
}
|
|
_prepareRequest(method, requestUrl, headers) {
|
|
const info = {};
|
|
info.parsedUrl = requestUrl;
|
|
const usingSsl = info.parsedUrl.protocol === 'https:';
|
|
info.httpModule = usingSsl ? https : http;
|
|
const defaultPort = usingSsl ? 443 : 80;
|
|
info.options = {};
|
|
info.options.host = info.parsedUrl.hostname;
|
|
info.options.port = info.parsedUrl.port
|
|
? parseInt(info.parsedUrl.port)
|
|
: defaultPort;
|
|
info.options.path =
|
|
(info.parsedUrl.pathname || '') + (info.parsedUrl.search || '');
|
|
info.options.method = method;
|
|
info.options.headers = this._mergeHeaders(headers);
|
|
if (this.userAgent != null) {
|
|
info.options.headers['user-agent'] = this.userAgent;
|
|
}
|
|
info.options.agent = this._getAgent(info.parsedUrl);
|
|
// gives handlers an opportunity to participate
|
|
if (this.handlers) {
|
|
this.handlers.forEach(handler => {
|
|
handler.prepareRequest(info.options);
|
|
});
|
|
}
|
|
return info;
|
|
}
|
|
_mergeHeaders(headers) {
|
|
const lowercaseKeys = obj => Object.keys(obj).reduce((c, k) => ((c[k.toLowerCase()] = obj[k]), c), {});
|
|
if (this.requestOptions && this.requestOptions.headers) {
|
|
return Object.assign({}, lowercaseKeys(this.requestOptions.headers), lowercaseKeys(headers));
|
|
}
|
|
return lowercaseKeys(headers || {});
|
|
}
|
|
_getExistingOrDefaultHeader(additionalHeaders, header, _default) {
|
|
const lowercaseKeys = obj => Object.keys(obj).reduce((c, k) => ((c[k.toLowerCase()] = obj[k]), c), {});
|
|
let clientHeader;
|
|
if (this.requestOptions && this.requestOptions.headers) {
|
|
clientHeader = lowercaseKeys(this.requestOptions.headers)[header];
|
|
}
|
|
return additionalHeaders[header] || clientHeader || _default;
|
|
}
|
|
_getAgent(parsedUrl) {
|
|
let agent;
|
|
let proxyUrl = pm.getProxyUrl(parsedUrl);
|
|
let useProxy = proxyUrl && proxyUrl.hostname;
|
|
if (this._keepAlive && useProxy) {
|
|
agent = this._proxyAgent;
|
|
}
|
|
if (this._keepAlive && !useProxy) {
|
|
agent = this._agent;
|
|
}
|
|
// if agent is already assigned use that agent.
|
|
if (!!agent) {
|
|
return agent;
|
|
}
|
|
const usingSsl = parsedUrl.protocol === 'https:';
|
|
let maxSockets = 100;
|
|
if (!!this.requestOptions) {
|
|
maxSockets = this.requestOptions.maxSockets || http.globalAgent.maxSockets;
|
|
}
|
|
if (useProxy) {
|
|
// If using proxy, need tunnel
|
|
if (!tunnel) {
|
|
tunnel = __nccwpck_require__(4294);
|
|
}
|
|
const agentOptions = {
|
|
maxSockets: maxSockets,
|
|
keepAlive: this._keepAlive,
|
|
proxy: {
|
|
...((proxyUrl.username || proxyUrl.password) && {
|
|
proxyAuth: `${proxyUrl.username}:${proxyUrl.password}`
|
|
}),
|
|
host: proxyUrl.hostname,
|
|
port: proxyUrl.port
|
|
}
|
|
};
|
|
let tunnelAgent;
|
|
const overHttps = proxyUrl.protocol === 'https:';
|
|
if (usingSsl) {
|
|
tunnelAgent = overHttps ? tunnel.httpsOverHttps : tunnel.httpsOverHttp;
|
|
}
|
|
else {
|
|
tunnelAgent = overHttps ? tunnel.httpOverHttps : tunnel.httpOverHttp;
|
|
}
|
|
agent = tunnelAgent(agentOptions);
|
|
this._proxyAgent = agent;
|
|
}
|
|
// if reusing agent across request and tunneling agent isn't assigned create a new agent
|
|
if (this._keepAlive && !agent) {
|
|
const options = { keepAlive: this._keepAlive, maxSockets: maxSockets };
|
|
agent = usingSsl ? new https.Agent(options) : new http.Agent(options);
|
|
this._agent = agent;
|
|
}
|
|
// if not using private agent and tunnel agent isn't setup then use global agent
|
|
if (!agent) {
|
|
agent = usingSsl ? https.globalAgent : http.globalAgent;
|
|
}
|
|
if (usingSsl && this._ignoreSslError) {
|
|
// we don't want to set NODE_TLS_REJECT_UNAUTHORIZED=0 since that will affect request for entire process
|
|
// http.RequestOptions doesn't expose a way to modify RequestOptions.agent.options
|
|
// we have to cast it to any and change it directly
|
|
agent.options = Object.assign(agent.options || {}, {
|
|
rejectUnauthorized: false
|
|
});
|
|
}
|
|
return agent;
|
|
}
|
|
_performExponentialBackoff(retryNumber) {
|
|
retryNumber = Math.min(ExponentialBackoffCeiling, retryNumber);
|
|
const ms = ExponentialBackoffTimeSlice * Math.pow(2, retryNumber);
|
|
return new Promise(resolve => setTimeout(() => resolve(), ms));
|
|
}
|
|
static dateTimeDeserializer(key, value) {
|
|
if (typeof value === 'string') {
|
|
let a = new Date(value);
|
|
if (!isNaN(a.valueOf())) {
|
|
return a;
|
|
}
|
|
}
|
|
return value;
|
|
}
|
|
async _processResponse(res, options) {
|
|
return new Promise(async (resolve, reject) => {
|
|
const statusCode = res.message.statusCode;
|
|
const response = {
|
|
statusCode: statusCode,
|
|
result: null,
|
|
headers: {}
|
|
};
|
|
// not found leads to null obj returned
|
|
if (statusCode == HttpCodes.NotFound) {
|
|
resolve(response);
|
|
}
|
|
let obj;
|
|
let contents;
|
|
// get the result from the body
|
|
try {
|
|
contents = await res.readBody();
|
|
if (contents && contents.length > 0) {
|
|
if (options && options.deserializeDates) {
|
|
obj = JSON.parse(contents, HttpClient.dateTimeDeserializer);
|
|
}
|
|
else {
|
|
obj = JSON.parse(contents);
|
|
}
|
|
response.result = obj;
|
|
}
|
|
response.headers = res.message.headers;
|
|
}
|
|
catch (err) {
|
|
// Invalid resource (contents not json); leaving result obj null
|
|
}
|
|
// note that 3xx redirects are handled by the http layer.
|
|
if (statusCode > 299) {
|
|
let msg;
|
|
// if exception/error in body, attempt to get better error
|
|
if (obj && obj.message) {
|
|
msg = obj.message;
|
|
}
|
|
else if (contents && contents.length > 0) {
|
|
// it may be the case that the exception is in the body message as string
|
|
msg = contents;
|
|
}
|
|
else {
|
|
msg = 'Failed request: (' + statusCode + ')';
|
|
}
|
|
let err = new HttpClientError(msg, statusCode);
|
|
err.result = response.result;
|
|
reject(err);
|
|
}
|
|
else {
|
|
resolve(response);
|
|
}
|
|
});
|
|
}
|
|
}
|
|
exports.HttpClient = HttpClient;
|
|
|
|
|
|
/***/ }),
|
|
|
|
/***/ 6443:
|
|
/***/ ((__unused_webpack_module, exports) => {
|
|
|
|
"use strict";
|
|
|
|
Object.defineProperty(exports, "__esModule", ({ value: true }));
|
|
function getProxyUrl(reqUrl) {
|
|
let usingSsl = reqUrl.protocol === 'https:';
|
|
let proxyUrl;
|
|
if (checkBypass(reqUrl)) {
|
|
return proxyUrl;
|
|
}
|
|
let proxyVar;
|
|
if (usingSsl) {
|
|
proxyVar = process.env['https_proxy'] || process.env['HTTPS_PROXY'];
|
|
}
|
|
else {
|
|
proxyVar = process.env['http_proxy'] || process.env['HTTP_PROXY'];
|
|
}
|
|
if (proxyVar) {
|
|
proxyUrl = new URL(proxyVar);
|
|
}
|
|
return proxyUrl;
|
|
}
|
|
exports.getProxyUrl = getProxyUrl;
|
|
function checkBypass(reqUrl) {
|
|
if (!reqUrl.hostname) {
|
|
return false;
|
|
}
|
|
let noProxy = process.env['no_proxy'] || process.env['NO_PROXY'] || '';
|
|
if (!noProxy) {
|
|
return false;
|
|
}
|
|
// Determine the request port
|
|
let reqPort;
|
|
if (reqUrl.port) {
|
|
reqPort = Number(reqUrl.port);
|
|
}
|
|
else if (reqUrl.protocol === 'http:') {
|
|
reqPort = 80;
|
|
}
|
|
else if (reqUrl.protocol === 'https:') {
|
|
reqPort = 443;
|
|
}
|
|
// Format the request hostname and hostname with port
|
|
let upperReqHosts = [reqUrl.hostname.toUpperCase()];
|
|
if (typeof reqPort === 'number') {
|
|
upperReqHosts.push(`${upperReqHosts[0]}:${reqPort}`);
|
|
}
|
|
// Compare request host against noproxy
|
|
for (let upperNoProxyItem of noProxy
|
|
.split(',')
|
|
.map(x => x.trim().toUpperCase())
|
|
.filter(x => x)) {
|
|
if (upperReqHosts.some(x => x === upperNoProxyItem)) {
|
|
return true;
|
|
}
|
|
}
|
|
return false;
|
|
}
|
|
exports.checkBypass = checkBypass;
|
|
|
|
|
|
/***/ }),
|
|
|
|
/***/ 334:
|
|
/***/ ((__unused_webpack_module, exports) => {
|
|
|
|
"use strict";
|
|
|
|
|
|
Object.defineProperty(exports, "__esModule", ({ value: true }));
|
|
|
|
async function auth(token) {
|
|
const tokenType = token.split(/\./).length === 3 ? "app" : /^v\d+\./.test(token) ? "installation" : "oauth";
|
|
return {
|
|
type: "token",
|
|
token: token,
|
|
tokenType
|
|
};
|
|
}
|
|
|
|
/**
|
|
* Prefix token for usage in the Authorization header
|
|
*
|
|
* @param token OAuth token or JSON Web Token
|
|
*/
|
|
function withAuthorizationPrefix(token) {
|
|
if (token.split(/\./).length === 3) {
|
|
return `bearer ${token}`;
|
|
}
|
|
|
|
return `token ${token}`;
|
|
}
|
|
|
|
async function hook(token, request, route, parameters) {
|
|
const endpoint = request.endpoint.merge(route, parameters);
|
|
endpoint.headers.authorization = withAuthorizationPrefix(token);
|
|
return request(endpoint);
|
|
}
|
|
|
|
const createTokenAuth = function createTokenAuth(token) {
|
|
if (!token) {
|
|
throw new Error("[@octokit/auth-token] No token passed to createTokenAuth");
|
|
}
|
|
|
|
if (typeof token !== "string") {
|
|
throw new Error("[@octokit/auth-token] Token passed to createTokenAuth is not a string");
|
|
}
|
|
|
|
token = token.replace(/^(token|bearer) +/i, "");
|
|
return Object.assign(auth.bind(null, token), {
|
|
hook: hook.bind(null, token)
|
|
});
|
|
};
|
|
|
|
exports.createTokenAuth = createTokenAuth;
|
|
//# sourceMappingURL=index.js.map
|
|
|
|
|
|
/***/ }),
|
|
|
|
/***/ 6762:
|
|
/***/ ((__unused_webpack_module, exports, __nccwpck_require__) => {
|
|
|
|
"use strict";
|
|
|
|
|
|
Object.defineProperty(exports, "__esModule", ({ value: true }));
|
|
|
|
var universalUserAgent = __nccwpck_require__(5030);
|
|
var beforeAfterHook = __nccwpck_require__(3682);
|
|
var request = __nccwpck_require__(6234);
|
|
var graphql = __nccwpck_require__(8467);
|
|
var authToken = __nccwpck_require__(334);
|
|
|
|
function _objectWithoutPropertiesLoose(source, excluded) {
|
|
if (source == null) return {};
|
|
var target = {};
|
|
var sourceKeys = Object.keys(source);
|
|
var key, i;
|
|
|
|
for (i = 0; i < sourceKeys.length; i++) {
|
|
key = sourceKeys[i];
|
|
if (excluded.indexOf(key) >= 0) continue;
|
|
target[key] = source[key];
|
|
}
|
|
|
|
return target;
|
|
}
|
|
|
|
function _objectWithoutProperties(source, excluded) {
|
|
if (source == null) return {};
|
|
|
|
var target = _objectWithoutPropertiesLoose(source, excluded);
|
|
|
|
var key, i;
|
|
|
|
if (Object.getOwnPropertySymbols) {
|
|
var sourceSymbolKeys = Object.getOwnPropertySymbols(source);
|
|
|
|
for (i = 0; i < sourceSymbolKeys.length; i++) {
|
|
key = sourceSymbolKeys[i];
|
|
if (excluded.indexOf(key) >= 0) continue;
|
|
if (!Object.prototype.propertyIsEnumerable.call(source, key)) continue;
|
|
target[key] = source[key];
|
|
}
|
|
}
|
|
|
|
return target;
|
|
}
|
|
|
|
const VERSION = "3.4.0";
|
|
|
|
class Octokit {
|
|
constructor(options = {}) {
|
|
const hook = new beforeAfterHook.Collection();
|
|
const requestDefaults = {
|
|
baseUrl: request.request.endpoint.DEFAULTS.baseUrl,
|
|
headers: {},
|
|
request: Object.assign({}, options.request, {
|
|
// @ts-ignore internal usage only, no need to type
|
|
hook: hook.bind(null, "request")
|
|
}),
|
|
mediaType: {
|
|
previews: [],
|
|
format: ""
|
|
}
|
|
}; // prepend default user agent with `options.userAgent` if set
|
|
|
|
requestDefaults.headers["user-agent"] = [options.userAgent, `octokit-core.js/${VERSION} ${universalUserAgent.getUserAgent()}`].filter(Boolean).join(" ");
|
|
|
|
if (options.baseUrl) {
|
|
requestDefaults.baseUrl = options.baseUrl;
|
|
}
|
|
|
|
if (options.previews) {
|
|
requestDefaults.mediaType.previews = options.previews;
|
|
}
|
|
|
|
if (options.timeZone) {
|
|
requestDefaults.headers["time-zone"] = options.timeZone;
|
|
}
|
|
|
|
this.request = request.request.defaults(requestDefaults);
|
|
this.graphql = graphql.withCustomRequest(this.request).defaults(requestDefaults);
|
|
this.log = Object.assign({
|
|
debug: () => {},
|
|
info: () => {},
|
|
warn: console.warn.bind(console),
|
|
error: console.error.bind(console)
|
|
}, options.log);
|
|
this.hook = hook; // (1) If neither `options.authStrategy` nor `options.auth` are set, the `octokit` instance
|
|
// is unauthenticated. The `this.auth()` method is a no-op and no request hook is registered.
|
|
// (2) If only `options.auth` is set, use the default token authentication strategy.
|
|
// (3) If `options.authStrategy` is set then use it and pass in `options.auth`. Always pass own request as many strategies accept a custom request instance.
|
|
// TODO: type `options.auth` based on `options.authStrategy`.
|
|
|
|
if (!options.authStrategy) {
|
|
if (!options.auth) {
|
|
// (1)
|
|
this.auth = async () => ({
|
|
type: "unauthenticated"
|
|
});
|
|
} else {
|
|
// (2)
|
|
const auth = authToken.createTokenAuth(options.auth); // @ts-ignore ¯\_(ツ)_/¯
|
|
|
|
hook.wrap("request", auth.hook);
|
|
this.auth = auth;
|
|
}
|
|
} else {
|
|
const {
|
|
authStrategy
|
|
} = options,
|
|
otherOptions = _objectWithoutProperties(options, ["authStrategy"]);
|
|
|
|
const auth = authStrategy(Object.assign({
|
|
request: this.request,
|
|
log: this.log,
|
|
// we pass the current octokit instance as well as its constructor options
|
|
// to allow for authentication strategies that return a new octokit instance
|
|
// that shares the same internal state as the current one. The original
|
|
// requirement for this was the "event-octokit" authentication strategy
|
|
// of https://github.com/probot/octokit-auth-probot.
|
|
octokit: this,
|
|
octokitOptions: otherOptions
|
|
}, options.auth)); // @ts-ignore ¯\_(ツ)_/¯
|
|
|
|
hook.wrap("request", auth.hook);
|
|
this.auth = auth;
|
|
} // apply plugins
|
|
// https://stackoverflow.com/a/16345172
|
|
|
|
|
|
const classConstructor = this.constructor;
|
|
classConstructor.plugins.forEach(plugin => {
|
|
Object.assign(this, plugin(this, options));
|
|
});
|
|
}
|
|
|
|
static defaults(defaults) {
|
|
const OctokitWithDefaults = class extends this {
|
|
constructor(...args) {
|
|
const options = args[0] || {};
|
|
|
|
if (typeof defaults === "function") {
|
|
super(defaults(options));
|
|
return;
|
|
}
|
|
|
|
super(Object.assign({}, defaults, options, options.userAgent && defaults.userAgent ? {
|
|
userAgent: `${options.userAgent} ${defaults.userAgent}`
|
|
} : null));
|
|
}
|
|
|
|
};
|
|
return OctokitWithDefaults;
|
|
}
|
|
/**
|
|
* Attach a plugin (or many) to your Octokit instance.
|
|
*
|
|
* @example
|
|
* const API = Octokit.plugin(plugin1, plugin2, plugin3, ...)
|
|
*/
|
|
|
|
|
|
static plugin(...newPlugins) {
|
|
var _a;
|
|
|
|
const currentPlugins = this.plugins;
|
|
const NewOctokit = (_a = class extends this {}, _a.plugins = currentPlugins.concat(newPlugins.filter(plugin => !currentPlugins.includes(plugin))), _a);
|
|
return NewOctokit;
|
|
}
|
|
|
|
}
|
|
Octokit.VERSION = VERSION;
|
|
Octokit.plugins = [];
|
|
|
|
exports.Octokit = Octokit;
|
|
//# sourceMappingURL=index.js.map
|
|
|
|
|
|
/***/ }),
|
|
|
|
/***/ 9440:
|
|
/***/ ((__unused_webpack_module, exports, __nccwpck_require__) => {
|
|
|
|
"use strict";
|
|
|
|
|
|
Object.defineProperty(exports, "__esModule", ({ value: true }));
|
|
|
|
var isPlainObject = __nccwpck_require__(3287);
|
|
var universalUserAgent = __nccwpck_require__(5030);
|
|
|
|
function lowercaseKeys(object) {
|
|
if (!object) {
|
|
return {};
|
|
}
|
|
|
|
return Object.keys(object).reduce((newObj, key) => {
|
|
newObj[key.toLowerCase()] = object[key];
|
|
return newObj;
|
|
}, {});
|
|
}
|
|
|
|
function mergeDeep(defaults, options) {
|
|
const result = Object.assign({}, defaults);
|
|
Object.keys(options).forEach(key => {
|
|
if (isPlainObject.isPlainObject(options[key])) {
|
|
if (!(key in defaults)) Object.assign(result, {
|
|
[key]: options[key]
|
|
});else result[key] = mergeDeep(defaults[key], options[key]);
|
|
} else {
|
|
Object.assign(result, {
|
|
[key]: options[key]
|
|
});
|
|
}
|
|
});
|
|
return result;
|
|
}
|
|
|
|
function removeUndefinedProperties(obj) {
|
|
for (const key in obj) {
|
|
if (obj[key] === undefined) {
|
|
delete obj[key];
|
|
}
|
|
}
|
|
|
|
return obj;
|
|
}
|
|
|
|
function merge(defaults, route, options) {
|
|
if (typeof route === "string") {
|
|
let [method, url] = route.split(" ");
|
|
options = Object.assign(url ? {
|
|
method,
|
|
url
|
|
} : {
|
|
url: method
|
|
}, options);
|
|
} else {
|
|
options = Object.assign({}, route);
|
|
} // lowercase header names before merging with defaults to avoid duplicates
|
|
|
|
|
|
options.headers = lowercaseKeys(options.headers); // remove properties with undefined values before merging
|
|
|
|
removeUndefinedProperties(options);
|
|
removeUndefinedProperties(options.headers);
|
|
const mergedOptions = mergeDeep(defaults || {}, options); // mediaType.previews arrays are merged, instead of overwritten
|
|
|
|
if (defaults && defaults.mediaType.previews.length) {
|
|
mergedOptions.mediaType.previews = defaults.mediaType.previews.filter(preview => !mergedOptions.mediaType.previews.includes(preview)).concat(mergedOptions.mediaType.previews);
|
|
}
|
|
|
|
mergedOptions.mediaType.previews = mergedOptions.mediaType.previews.map(preview => preview.replace(/-preview/, ""));
|
|
return mergedOptions;
|
|
}
|
|
|
|
function addQueryParameters(url, parameters) {
|
|
const separator = /\?/.test(url) ? "&" : "?";
|
|
const names = Object.keys(parameters);
|
|
|
|
if (names.length === 0) {
|
|
return url;
|
|
}
|
|
|
|
return url + separator + names.map(name => {
|
|
if (name === "q") {
|
|
return "q=" + parameters.q.split("+").map(encodeURIComponent).join("+");
|
|
}
|
|
|
|
return `${name}=${encodeURIComponent(parameters[name])}`;
|
|
}).join("&");
|
|
}
|
|
|
|
const urlVariableRegex = /\{[^}]+\}/g;
|
|
|
|
function removeNonChars(variableName) {
|
|
return variableName.replace(/^\W+|\W+$/g, "").split(/,/);
|
|
}
|
|
|
|
function extractUrlVariableNames(url) {
|
|
const matches = url.match(urlVariableRegex);
|
|
|
|
if (!matches) {
|
|
return [];
|
|
}
|
|
|
|
return matches.map(removeNonChars).reduce((a, b) => a.concat(b), []);
|
|
}
|
|
|
|
function omit(object, keysToOmit) {
|
|
return Object.keys(object).filter(option => !keysToOmit.includes(option)).reduce((obj, key) => {
|
|
obj[key] = object[key];
|
|
return obj;
|
|
}, {});
|
|
}
|
|
|
|
// Based on https://github.com/bramstein/url-template, licensed under BSD
|
|
// TODO: create separate package.
|
|
//
|
|
// Copyright (c) 2012-2014, Bram Stein
|
|
// All rights reserved.
|
|
// Redistribution and use in source and binary forms, with or without
|
|
// modification, are permitted provided that the following conditions
|
|
// are met:
|
|
// 1. Redistributions of source code must retain the above copyright
|
|
// notice, this list of conditions and the following disclaimer.
|
|
// 2. Redistributions in binary form must reproduce the above copyright
|
|
// notice, this list of conditions and the following disclaimer in the
|
|
// documentation and/or other materials provided with the distribution.
|
|
// 3. The name of the author may not be used to endorse or promote products
|
|
// derived from this software without specific prior written permission.
|
|
// THIS SOFTWARE IS PROVIDED BY THE AUTHOR "AS IS" AND ANY EXPRESS OR IMPLIED
|
|
// WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF
|
|
// MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO
|
|
// EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT,
|
|
// INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,
|
|
// BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY
|
|
// OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
|
|
// NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE,
|
|
// EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
/* istanbul ignore file */
|
|
function encodeReserved(str) {
|
|
return str.split(/(%[0-9A-Fa-f]{2})/g).map(function (part) {
|
|
if (!/%[0-9A-Fa-f]/.test(part)) {
|
|
part = encodeURI(part).replace(/%5B/g, "[").replace(/%5D/g, "]");
|
|
}
|
|
|
|
return part;
|
|
}).join("");
|
|
}
|
|
|
|
function encodeUnreserved(str) {
|
|
return encodeURIComponent(str).replace(/[!'()*]/g, function (c) {
|
|
return "%" + c.charCodeAt(0).toString(16).toUpperCase();
|
|
});
|
|
}
|
|
|
|
function encodeValue(operator, value, key) {
|
|
value = operator === "+" || operator === "#" ? encodeReserved(value) : encodeUnreserved(value);
|
|
|
|
if (key) {
|
|
return encodeUnreserved(key) + "=" + value;
|
|
} else {
|
|
return value;
|
|
}
|
|
}
|
|
|
|
function isDefined(value) {
|
|
return value !== undefined && value !== null;
|
|
}
|
|
|
|
function isKeyOperator(operator) {
|
|
return operator === ";" || operator === "&" || operator === "?";
|
|
}
|
|
|
|
function getValues(context, operator, key, modifier) {
|
|
var value = context[key],
|
|
result = [];
|
|
|
|
if (isDefined(value) && value !== "") {
|
|
if (typeof value === "string" || typeof value === "number" || typeof value === "boolean") {
|
|
value = value.toString();
|
|
|
|
if (modifier && modifier !== "*") {
|
|
value = value.substring(0, parseInt(modifier, 10));
|
|
}
|
|
|
|
result.push(encodeValue(operator, value, isKeyOperator(operator) ? key : ""));
|
|
} else {
|
|
if (modifier === "*") {
|
|
if (Array.isArray(value)) {
|
|
value.filter(isDefined).forEach(function (value) {
|
|
result.push(encodeValue(operator, value, isKeyOperator(operator) ? key : ""));
|
|
});
|
|
} else {
|
|
Object.keys(value).forEach(function (k) {
|
|
if (isDefined(value[k])) {
|
|
result.push(encodeValue(operator, value[k], k));
|
|
}
|
|
});
|
|
}
|
|
} else {
|
|
const tmp = [];
|
|
|
|
if (Array.isArray(value)) {
|
|
value.filter(isDefined).forEach(function (value) {
|
|
tmp.push(encodeValue(operator, value));
|
|
});
|
|
} else {
|
|
Object.keys(value).forEach(function (k) {
|
|
if (isDefined(value[k])) {
|
|
tmp.push(encodeUnreserved(k));
|
|
tmp.push(encodeValue(operator, value[k].toString()));
|
|
}
|
|
});
|
|
}
|
|
|
|
if (isKeyOperator(operator)) {
|
|
result.push(encodeUnreserved(key) + "=" + tmp.join(","));
|
|
} else if (tmp.length !== 0) {
|
|
result.push(tmp.join(","));
|
|
}
|
|
}
|
|
}
|
|
} else {
|
|
if (operator === ";") {
|
|
if (isDefined(value)) {
|
|
result.push(encodeUnreserved(key));
|
|
}
|
|
} else if (value === "" && (operator === "&" || operator === "?")) {
|
|
result.push(encodeUnreserved(key) + "=");
|
|
} else if (value === "") {
|
|
result.push("");
|
|
}
|
|
}
|
|
|
|
return result;
|
|
}
|
|
|
|
function parseUrl(template) {
|
|
return {
|
|
expand: expand.bind(null, template)
|
|
};
|
|
}
|
|
|
|
function expand(template, context) {
|
|
var operators = ["+", "#", ".", "/", ";", "?", "&"];
|
|
return template.replace(/\{([^\{\}]+)\}|([^\{\}]+)/g, function (_, expression, literal) {
|
|
if (expression) {
|
|
let operator = "";
|
|
const values = [];
|
|
|
|
if (operators.indexOf(expression.charAt(0)) !== -1) {
|
|
operator = expression.charAt(0);
|
|
expression = expression.substr(1);
|
|
}
|
|
|
|
expression.split(/,/g).forEach(function (variable) {
|
|
var tmp = /([^:\*]*)(?::(\d+)|(\*))?/.exec(variable);
|
|
values.push(getValues(context, operator, tmp[1], tmp[2] || tmp[3]));
|
|
});
|
|
|
|
if (operator && operator !== "+") {
|
|
var separator = ",";
|
|
|
|
if (operator === "?") {
|
|
separator = "&";
|
|
} else if (operator !== "#") {
|
|
separator = operator;
|
|
}
|
|
|
|
return (values.length !== 0 ? operator : "") + values.join(separator);
|
|
} else {
|
|
return values.join(",");
|
|
}
|
|
} else {
|
|
return encodeReserved(literal);
|
|
}
|
|
});
|
|
}
|
|
|
|
function parse(options) {
|
|
// https://fetch.spec.whatwg.org/#methods
|
|
let method = options.method.toUpperCase(); // replace :varname with {varname} to make it RFC 6570 compatible
|
|
|
|
let url = (options.url || "/").replace(/:([a-z]\w+)/g, "{$1}");
|
|
let headers = Object.assign({}, options.headers);
|
|
let body;
|
|
let parameters = omit(options, ["method", "baseUrl", "url", "headers", "request", "mediaType"]); // extract variable names from URL to calculate remaining variables later
|
|
|
|
const urlVariableNames = extractUrlVariableNames(url);
|
|
url = parseUrl(url).expand(parameters);
|
|
|
|
if (!/^http/.test(url)) {
|
|
url = options.baseUrl + url;
|
|
}
|
|
|
|
const omittedParameters = Object.keys(options).filter(option => urlVariableNames.includes(option)).concat("baseUrl");
|
|
const remainingParameters = omit(parameters, omittedParameters);
|
|
const isBinaryRequest = /application\/octet-stream/i.test(headers.accept);
|
|
|
|
if (!isBinaryRequest) {
|
|
if (options.mediaType.format) {
|
|
// e.g. application/vnd.github.v3+json => application/vnd.github.v3.raw
|
|
headers.accept = headers.accept.split(/,/).map(preview => preview.replace(/application\/vnd(\.\w+)(\.v3)?(\.\w+)?(\+json)?$/, `application/vnd$1$2.${options.mediaType.format}`)).join(",");
|
|
}
|
|
|
|
if (options.mediaType.previews.length) {
|
|
const previewsFromAcceptHeader = headers.accept.match(/[\w-]+(?=-preview)/g) || [];
|
|
headers.accept = previewsFromAcceptHeader.concat(options.mediaType.previews).map(preview => {
|
|
const format = options.mediaType.format ? `.${options.mediaType.format}` : "+json";
|
|
return `application/vnd.github.${preview}-preview${format}`;
|
|
}).join(",");
|
|
}
|
|
} // for GET/HEAD requests, set URL query parameters from remaining parameters
|
|
// for PATCH/POST/PUT/DELETE requests, set request body from remaining parameters
|
|
|
|
|
|
if (["GET", "HEAD"].includes(method)) {
|
|
url = addQueryParameters(url, remainingParameters);
|
|
} else {
|
|
if ("data" in remainingParameters) {
|
|
body = remainingParameters.data;
|
|
} else {
|
|
if (Object.keys(remainingParameters).length) {
|
|
body = remainingParameters;
|
|
} else {
|
|
headers["content-length"] = 0;
|
|
}
|
|
}
|
|
} // default content-type for JSON if body is set
|
|
|
|
|
|
if (!headers["content-type"] && typeof body !== "undefined") {
|
|
headers["content-type"] = "application/json; charset=utf-8";
|
|
} // GitHub expects 'content-length: 0' header for PUT/PATCH requests without body.
|
|
// fetch does not allow to set `content-length` header, but we can set body to an empty string
|
|
|
|
|
|
if (["PATCH", "PUT"].includes(method) && typeof body === "undefined") {
|
|
body = "";
|
|
} // Only return body/request keys if present
|
|
|
|
|
|
return Object.assign({
|
|
method,
|
|
url,
|
|
headers
|
|
}, typeof body !== "undefined" ? {
|
|
body
|
|
} : null, options.request ? {
|
|
request: options.request
|
|
} : null);
|
|
}
|
|
|
|
function endpointWithDefaults(defaults, route, options) {
|
|
return parse(merge(defaults, route, options));
|
|
}
|
|
|
|
function withDefaults(oldDefaults, newDefaults) {
|
|
const DEFAULTS = merge(oldDefaults, newDefaults);
|
|
const endpoint = endpointWithDefaults.bind(null, DEFAULTS);
|
|
return Object.assign(endpoint, {
|
|
DEFAULTS,
|
|
defaults: withDefaults.bind(null, DEFAULTS),
|
|
merge: merge.bind(null, DEFAULTS),
|
|
parse
|
|
});
|
|
}
|
|
|
|
const VERSION = "6.0.11";
|
|
|
|
const userAgent = `octokit-endpoint.js/${VERSION} ${universalUserAgent.getUserAgent()}`; // DEFAULTS has all properties set that EndpointOptions has, except url.
|
|
// So we use RequestParameters and add method as additional required property.
|
|
|
|
const DEFAULTS = {
|
|
method: "GET",
|
|
baseUrl: "https://api.github.com",
|
|
headers: {
|
|
accept: "application/vnd.github.v3+json",
|
|
"user-agent": userAgent
|
|
},
|
|
mediaType: {
|
|
format: "",
|
|
previews: []
|
|
}
|
|
};
|
|
|
|
const endpoint = withDefaults(null, DEFAULTS);
|
|
|
|
exports.endpoint = endpoint;
|
|
//# sourceMappingURL=index.js.map
|
|
|
|
|
|
/***/ }),
|
|
|
|
/***/ 8467:
|
|
/***/ ((__unused_webpack_module, exports, __nccwpck_require__) => {
|
|
|
|
"use strict";
|
|
|
|
|
|
Object.defineProperty(exports, "__esModule", ({ value: true }));
|
|
|
|
var request = __nccwpck_require__(6234);
|
|
var universalUserAgent = __nccwpck_require__(5030);
|
|
|
|
const VERSION = "4.6.2";
|
|
|
|
class GraphqlError extends Error {
|
|
constructor(request, response) {
|
|
const message = response.data.errors[0].message;
|
|
super(message);
|
|
Object.assign(this, response.data);
|
|
Object.assign(this, {
|
|
headers: response.headers
|
|
});
|
|
this.name = "GraphqlError";
|
|
this.request = request; // Maintains proper stack trace (only available on V8)
|
|
|
|
/* istanbul ignore next */
|
|
|
|
if (Error.captureStackTrace) {
|
|
Error.captureStackTrace(this, this.constructor);
|
|
}
|
|
}
|
|
|
|
}
|
|
|
|
const NON_VARIABLE_OPTIONS = ["method", "baseUrl", "url", "headers", "request", "query", "mediaType"];
|
|
const FORBIDDEN_VARIABLE_OPTIONS = ["query", "method", "url"];
|
|
const GHES_V3_SUFFIX_REGEX = /\/api\/v3\/?$/;
|
|
function graphql(request, query, options) {
|
|
if (options) {
|
|
if (typeof query === "string" && "query" in options) {
|
|
return Promise.reject(new Error(`[@octokit/graphql] "query" cannot be used as variable name`));
|
|
}
|
|
|
|
for (const key in options) {
|
|
if (!FORBIDDEN_VARIABLE_OPTIONS.includes(key)) continue;
|
|
return Promise.reject(new Error(`[@octokit/graphql] "${key}" cannot be used as variable name`));
|
|
}
|
|
}
|
|
|
|
const parsedOptions = typeof query === "string" ? Object.assign({
|
|
query
|
|
}, options) : query;
|
|
const requestOptions = Object.keys(parsedOptions).reduce((result, key) => {
|
|
if (NON_VARIABLE_OPTIONS.includes(key)) {
|
|
result[key] = parsedOptions[key];
|
|
return result;
|
|
}
|
|
|
|
if (!result.variables) {
|
|
result.variables = {};
|
|
}
|
|
|
|
result.variables[key] = parsedOptions[key];
|
|
return result;
|
|
}, {}); // workaround for GitHub Enterprise baseUrl set with /api/v3 suffix
|
|
// https://github.com/octokit/auth-app.js/issues/111#issuecomment-657610451
|
|
|
|
const baseUrl = parsedOptions.baseUrl || request.endpoint.DEFAULTS.baseUrl;
|
|
|
|
if (GHES_V3_SUFFIX_REGEX.test(baseUrl)) {
|
|
requestOptions.url = baseUrl.replace(GHES_V3_SUFFIX_REGEX, "/api/graphql");
|
|
}
|
|
|
|
return request(requestOptions).then(response => {
|
|
if (response.data.errors) {
|
|
const headers = {};
|
|
|
|
for (const key of Object.keys(response.headers)) {
|
|
headers[key] = response.headers[key];
|
|
}
|
|
|
|
throw new GraphqlError(requestOptions, {
|
|
headers,
|
|
data: response.data
|
|
});
|
|
}
|
|
|
|
return response.data.data;
|
|
});
|
|
}
|
|
|
|
function withDefaults(request$1, newDefaults) {
|
|
const newRequest = request$1.defaults(newDefaults);
|
|
|
|
const newApi = (query, options) => {
|
|
return graphql(newRequest, query, options);
|
|
};
|
|
|
|
return Object.assign(newApi, {
|
|
defaults: withDefaults.bind(null, newRequest),
|
|
endpoint: request.request.endpoint
|
|
});
|
|
}
|
|
|
|
const graphql$1 = withDefaults(request.request, {
|
|
headers: {
|
|
"user-agent": `octokit-graphql.js/${VERSION} ${universalUserAgent.getUserAgent()}`
|
|
},
|
|
method: "POST",
|
|
url: "/graphql"
|
|
});
|
|
function withCustomRequest(customRequest) {
|
|
return withDefaults(customRequest, {
|
|
method: "POST",
|
|
url: "/graphql"
|
|
});
|
|
}
|
|
|
|
exports.graphql = graphql$1;
|
|
exports.withCustomRequest = withCustomRequest;
|
|
//# sourceMappingURL=index.js.map
|
|
|
|
|
|
/***/ }),
|
|
|
|
/***/ 4193:
|
|
/***/ ((__unused_webpack_module, exports) => {
|
|
|
|
"use strict";
|
|
|
|
|
|
Object.defineProperty(exports, "__esModule", ({ value: true }));
|
|
|
|
const VERSION = "2.13.3";
|
|
|
|
/**
|
|
* Some “list” response that can be paginated have a different response structure
|
|
*
|
|
* They have a `total_count` key in the response (search also has `incomplete_results`,
|
|
* /installation/repositories also has `repository_selection`), as well as a key with
|
|
* the list of the items which name varies from endpoint to endpoint.
|
|
*
|
|
* Octokit normalizes these responses so that paginated results are always returned following
|
|
* the same structure. One challenge is that if the list response has only one page, no Link
|
|
* header is provided, so this header alone is not sufficient to check wether a response is
|
|
* paginated or not.
|
|
*
|
|
* We check if a "total_count" key is present in the response data, but also make sure that
|
|
* a "url" property is not, as the "Get the combined status for a specific ref" endpoint would
|
|
* otherwise match: https://developer.github.com/v3/repos/statuses/#get-the-combined-status-for-a-specific-ref
|
|
*/
|
|
function normalizePaginatedListResponse(response) {
|
|
const responseNeedsNormalization = "total_count" in response.data && !("url" in response.data);
|
|
if (!responseNeedsNormalization) return response; // keep the additional properties intact as there is currently no other way
|
|
// to retrieve the same information.
|
|
|
|
const incompleteResults = response.data.incomplete_results;
|
|
const repositorySelection = response.data.repository_selection;
|
|
const totalCount = response.data.total_count;
|
|
delete response.data.incomplete_results;
|
|
delete response.data.repository_selection;
|
|
delete response.data.total_count;
|
|
const namespaceKey = Object.keys(response.data)[0];
|
|
const data = response.data[namespaceKey];
|
|
response.data = data;
|
|
|
|
if (typeof incompleteResults !== "undefined") {
|
|
response.data.incomplete_results = incompleteResults;
|
|
}
|
|
|
|
if (typeof repositorySelection !== "undefined") {
|
|
response.data.repository_selection = repositorySelection;
|
|
}
|
|
|
|
response.data.total_count = totalCount;
|
|
return response;
|
|
}
|
|
|
|
function iterator(octokit, route, parameters) {
|
|
const options = typeof route === "function" ? route.endpoint(parameters) : octokit.request.endpoint(route, parameters);
|
|
const requestMethod = typeof route === "function" ? route : octokit.request;
|
|
const method = options.method;
|
|
const headers = options.headers;
|
|
let url = options.url;
|
|
return {
|
|
[Symbol.asyncIterator]: () => ({
|
|
async next() {
|
|
if (!url) return {
|
|
done: true
|
|
};
|
|
const response = await requestMethod({
|
|
method,
|
|
url,
|
|
headers
|
|
});
|
|
const normalizedResponse = normalizePaginatedListResponse(response); // `response.headers.link` format:
|
|
// '<https://api.github.com/users/aseemk/followers?page=2>; rel="next", <https://api.github.com/users/aseemk/followers?page=2>; rel="last"'
|
|
// sets `url` to undefined if "next" URL is not present or `link` header is not set
|
|
|
|
url = ((normalizedResponse.headers.link || "").match(/<([^>]+)>;\s*rel="next"/) || [])[1];
|
|
return {
|
|
value: normalizedResponse
|
|
};
|
|
}
|
|
|
|
})
|
|
};
|
|
}
|
|
|
|
function paginate(octokit, route, parameters, mapFn) {
|
|
if (typeof parameters === "function") {
|
|
mapFn = parameters;
|
|
parameters = undefined;
|
|
}
|
|
|
|
return gather(octokit, [], iterator(octokit, route, parameters)[Symbol.asyncIterator](), mapFn);
|
|
}
|
|
|
|
function gather(octokit, results, iterator, mapFn) {
|
|
return iterator.next().then(result => {
|
|
if (result.done) {
|
|
return results;
|
|
}
|
|
|
|
let earlyExit = false;
|
|
|
|
function done() {
|
|
earlyExit = true;
|
|
}
|
|
|
|
results = results.concat(mapFn ? mapFn(result.value, done) : result.value.data);
|
|
|
|
if (earlyExit) {
|
|
return results;
|
|
}
|
|
|
|
return gather(octokit, results, iterator, mapFn);
|
|
});
|
|
}
|
|
|
|
const composePaginateRest = Object.assign(paginate, {
|
|
iterator
|
|
});
|
|
|
|
const paginatingEndpoints = ["GET /app/installations", "GET /applications/grants", "GET /authorizations", "GET /enterprises/{enterprise}/actions/permissions/organizations", "GET /enterprises/{enterprise}/actions/runner-groups", "GET /enterprises/{enterprise}/actions/runner-groups/{runner_group_id}/organizations", "GET /enterprises/{enterprise}/actions/runner-groups/{runner_group_id}/runners", "GET /enterprises/{enterprise}/actions/runners", "GET /enterprises/{enterprise}/actions/runners/downloads", "GET /events", "GET /gists", "GET /gists/public", "GET /gists/starred", "GET /gists/{gist_id}/comments", "GET /gists/{gist_id}/commits", "GET /gists/{gist_id}/forks", "GET /installation/repositories", "GET /issues", "GET /marketplace_listing/plans", "GET /marketplace_listing/plans/{plan_id}/accounts", "GET /marketplace_listing/stubbed/plans", "GET /marketplace_listing/stubbed/plans/{plan_id}/accounts", "GET /networks/{owner}/{repo}/events", "GET /notifications", "GET /organizations", "GET /orgs/{org}/actions/permissions/repositories", "GET /orgs/{org}/actions/runner-groups", "GET /orgs/{org}/actions/runner-groups/{runner_group_id}/repositories", "GET /orgs/{org}/actions/runner-groups/{runner_group_id}/runners", "GET /orgs/{org}/actions/runners", "GET /orgs/{org}/actions/runners/downloads", "GET /orgs/{org}/actions/secrets", "GET /orgs/{org}/actions/secrets/{secret_name}/repositories", "GET /orgs/{org}/blocks", "GET /orgs/{org}/credential-authorizations", "GET /orgs/{org}/events", "GET /orgs/{org}/failed_invitations", "GET /orgs/{org}/hooks", "GET /orgs/{org}/installations", "GET /orgs/{org}/invitations", "GET /orgs/{org}/invitations/{invitation_id}/teams", "GET /orgs/{org}/issues", "GET /orgs/{org}/members", "GET /orgs/{org}/migrations", "GET /orgs/{org}/migrations/{migration_id}/repositories", "GET /orgs/{org}/outside_collaborators", "GET /orgs/{org}/projects", "GET /orgs/{org}/public_members", "GET /orgs/{org}/repos", "GET /orgs/{org}/team-sync/groups", "GET /orgs/{org}/teams", "GET /orgs/{org}/teams/{team_slug}/discussions", "GET /orgs/{org}/teams/{team_slug}/discussions/{discussion_number}/comments", "GET /orgs/{org}/teams/{team_slug}/discussions/{discussion_number}/comments/{comment_number}/reactions", "GET /orgs/{org}/teams/{team_slug}/discussions/{discussion_number}/reactions", "GET /orgs/{org}/teams/{team_slug}/invitations", "GET /orgs/{org}/teams/{team_slug}/members", "GET /orgs/{org}/teams/{team_slug}/projects", "GET /orgs/{org}/teams/{team_slug}/repos", "GET /orgs/{org}/teams/{team_slug}/team-sync/group-mappings", "GET /orgs/{org}/teams/{team_slug}/teams", "GET /projects/columns/{column_id}/cards", "GET /projects/{project_id}/collaborators", "GET /projects/{project_id}/columns", "GET /repos/{owner}/{repo}/actions/artifacts", "GET /repos/{owner}/{repo}/actions/runners", "GET /repos/{owner}/{repo}/actions/runners/downloads", "GET /repos/{owner}/{repo}/actions/runs", "GET /repos/{owner}/{repo}/actions/runs/{run_id}/artifacts", "GET /repos/{owner}/{repo}/actions/runs/{run_id}/jobs", "GET /repos/{owner}/{repo}/actions/secrets", "GET /repos/{owner}/{repo}/actions/workflows", "GET /repos/{owner}/{repo}/actions/workflows/{workflow_id}/runs", "GET /repos/{owner}/{repo}/assignees", "GET /repos/{owner}/{repo}/branches", "GET /repos/{owner}/{repo}/check-runs/{check_run_id}/annotations", "GET /repos/{owner}/{repo}/check-suites/{check_suite_id}/check-runs", "GET /repos/{owner}/{repo}/code-scanning/alerts", "GET /repos/{owner}/{repo}/code-scanning/alerts/{alert_number}/instances", "GET /repos/{owner}/{repo}/code-scanning/analyses", "GET /repos/{owner}/{repo}/collaborators", "GET /repos/{owner}/{repo}/comments", "GET /repos/{owner}/{repo}/comments/{comment_id}/reactions", "GET /repos/{owner}/{repo}/commits", "GET /repos/{owner}/{repo}/commits/{commit_sha}/branches-where-head", "GET /repos/{owner}/{repo}/commits/{commit_sha}/comments", "GET /repos/{owner}/{repo}/commits/{commit_sha}/pulls", "GET /repos/{owner}/{repo}/commits/{ref}/check-runs", "GET /repos/{owner}/{repo}/commits/{ref}/check-suites", "GET /repos/{owner}/{repo}/commits/{ref}/statuses", "GET /repos/{owner}/{repo}/contributors", "GET /repos/{owner}/{repo}/deployments", "GET /repos/{owner}/{repo}/deployments/{deployment_id}/statuses", "GET /repos/{owner}/{repo}/events", "GET /repos/{owner}/{repo}/forks", "GET /repos/{owner}/{repo}/git/matching-refs/{ref}", "GET /repos/{owner}/{repo}/hooks", "GET /repos/{owner}/{repo}/invitations", "GET /repos/{owner}/{repo}/issues", "GET /repos/{owner}/{repo}/issues/comments", "GET /repos/{owner}/{repo}/issues/comments/{comment_id}/reactions", "GET /repos/{owner}/{repo}/issues/events", "GET /repos/{owner}/{repo}/issues/{issue_number}/comments", "GET /repos/{owner}/{repo}/issues/{issue_number}/events", "GET /repos/{owner}/{repo}/issues/{issue_number}/labels", "GET /repos/{owner}/{repo}/issues/{issue_number}/reactions", "GET /repos/{owner}/{repo}/issues/{issue_number}/timeline", "GET /repos/{owner}/{repo}/keys", "GET /repos/{owner}/{repo}/labels", "GET /repos/{owner}/{repo}/milestones", "GET /repos/{owner}/{repo}/milestones/{milestone_number}/labels", "GET /repos/{owner}/{repo}/notifications", "GET /repos/{owner}/{repo}/pages/builds", "GET /repos/{owner}/{repo}/projects", "GET /repos/{owner}/{repo}/pulls", "GET /repos/{owner}/{repo}/pulls/comments", "GET /repos/{owner}/{repo}/pulls/comments/{comment_id}/reactions", "GET /repos/{owner}/{repo}/pulls/{pull_number}/comments", "GET /repos/{owner}/{repo}/pulls/{pull_number}/commits", "GET /repos/{owner}/{repo}/pulls/{pull_number}/files", "GET /repos/{owner}/{repo}/pulls/{pull_number}/requested_reviewers", "GET /repos/{owner}/{repo}/pulls/{pull_number}/reviews", "GET /repos/{owner}/{repo}/pulls/{pull_number}/reviews/{review_id}/comments", "GET /repos/{owner}/{repo}/releases", "GET /repos/{owner}/{repo}/releases/{release_id}/assets", "GET /repos/{owner}/{repo}/secret-scanning/alerts", "GET /repos/{owner}/{repo}/stargazers", "GET /repos/{owner}/{repo}/subscribers", "GET /repos/{owner}/{repo}/tags", "GET /repos/{owner}/{repo}/teams", "GET /repositories", "GET /repositories/{repository_id}/environments/{environment_name}/secrets", "GET /scim/v2/enterprises/{enterprise}/Groups", "GET /scim/v2/enterprises/{enterprise}/Users", "GET /scim/v2/organizations/{org}/Users", "GET /search/code", "GET /search/commits", "GET /search/issues", "GET /search/labels", "GET /search/repositories", "GET /search/topics", "GET /search/users", "GET /teams/{team_id}/discussions", "GET /teams/{team_id}/discussions/{discussion_number}/comments", "GET /teams/{team_id}/discussions/{discussion_number}/comments/{comment_number}/reactions", "GET /teams/{team_id}/discussions/{discussion_number}/reactions", "GET /teams/{team_id}/invitations", "GET /teams/{team_id}/members", "GET /teams/{team_id}/projects", "GET /teams/{team_id}/repos", "GET /teams/{team_id}/team-sync/group-mappings", "GET /teams/{team_id}/teams", "GET /user/blocks", "GET /user/emails", "GET /user/followers", "GET /user/following", "GET /user/gpg_keys", "GET /user/installations", "GET /user/installations/{installation_id}/repositories", "GET /user/issues", "GET /user/keys", "GET /user/marketplace_purchases", "GET /user/marketplace_purchases/stubbed", "GET /user/memberships/orgs", "GET /user/migrations", "GET /user/migrations/{migration_id}/repositories", "GET /user/orgs", "GET /user/public_emails", "GET /user/repos", "GET /user/repository_invitations", "GET /user/starred", "GET /user/subscriptions", "GET /user/teams", "GET /users", "GET /users/{username}/events", "GET /users/{username}/events/orgs/{org}", "GET /users/{username}/events/public", "GET /users/{username}/followers", "GET /users/{username}/following", "GET /users/{username}/gists", "GET /users/{username}/gpg_keys", "GET /users/{username}/keys", "GET /users/{username}/orgs", "GET /users/{username}/projects", "GET /users/{username}/received_events", "GET /users/{username}/received_events/public", "GET /users/{username}/repos", "GET /users/{username}/starred", "GET /users/{username}/subscriptions"];
|
|
|
|
function isPaginatingEndpoint(arg) {
|
|
if (typeof arg === "string") {
|
|
return paginatingEndpoints.includes(arg);
|
|
} else {
|
|
return false;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* @param octokit Octokit instance
|
|
* @param options Options passed to Octokit constructor
|
|
*/
|
|
|
|
function paginateRest(octokit) {
|
|
return {
|
|
paginate: Object.assign(paginate.bind(null, octokit), {
|
|
iterator: iterator.bind(null, octokit)
|
|
})
|
|
};
|
|
}
|
|
paginateRest.VERSION = VERSION;
|
|
|
|
exports.composePaginateRest = composePaginateRest;
|
|
exports.isPaginatingEndpoint = isPaginatingEndpoint;
|
|
exports.paginateRest = paginateRest;
|
|
exports.paginatingEndpoints = paginatingEndpoints;
|
|
//# sourceMappingURL=index.js.map
|
|
|
|
|
|
/***/ }),
|
|
|
|
/***/ 3044:
|
|
/***/ ((__unused_webpack_module, exports) => {
|
|
|
|
"use strict";
|
|
|
|
|
|
Object.defineProperty(exports, "__esModule", ({ value: true }));
|
|
|
|
function ownKeys(object, enumerableOnly) {
|
|
var keys = Object.keys(object);
|
|
|
|
if (Object.getOwnPropertySymbols) {
|
|
var symbols = Object.getOwnPropertySymbols(object);
|
|
|
|
if (enumerableOnly) {
|
|
symbols = symbols.filter(function (sym) {
|
|
return Object.getOwnPropertyDescriptor(object, sym).enumerable;
|
|
});
|
|
}
|
|
|
|
keys.push.apply(keys, symbols);
|
|
}
|
|
|
|
return keys;
|
|
}
|
|
|
|
function _objectSpread2(target) {
|
|
for (var i = 1; i < arguments.length; i++) {
|
|
var source = arguments[i] != null ? arguments[i] : {};
|
|
|
|
if (i % 2) {
|
|
ownKeys(Object(source), true).forEach(function (key) {
|
|
_defineProperty(target, key, source[key]);
|
|
});
|
|
} else if (Object.getOwnPropertyDescriptors) {
|
|
Object.defineProperties(target, Object.getOwnPropertyDescriptors(source));
|
|
} else {
|
|
ownKeys(Object(source)).forEach(function (key) {
|
|
Object.defineProperty(target, key, Object.getOwnPropertyDescriptor(source, key));
|
|
});
|
|
}
|
|
}
|
|
|
|
return target;
|
|
}
|
|
|
|
function _defineProperty(obj, key, value) {
|
|
if (key in obj) {
|
|
Object.defineProperty(obj, key, {
|
|
value: value,
|
|
enumerable: true,
|
|
configurable: true,
|
|
writable: true
|
|
});
|
|
} else {
|
|
obj[key] = value;
|
|
}
|
|
|
|
return obj;
|
|
}
|
|
|
|
const Endpoints = {
|
|
actions: {
|
|
addSelectedRepoToOrgSecret: ["PUT /orgs/{org}/actions/secrets/{secret_name}/repositories/{repository_id}"],
|
|
approveWorkflowRun: ["POST /repos/{owner}/{repo}/actions/runs/{run_id}/approve"],
|
|
cancelWorkflowRun: ["POST /repos/{owner}/{repo}/actions/runs/{run_id}/cancel"],
|
|
createOrUpdateEnvironmentSecret: ["PUT /repositories/{repository_id}/environments/{environment_name}/secrets/{secret_name}"],
|
|
createOrUpdateOrgSecret: ["PUT /orgs/{org}/actions/secrets/{secret_name}"],
|
|
createOrUpdateRepoSecret: ["PUT /repos/{owner}/{repo}/actions/secrets/{secret_name}"],
|
|
createRegistrationTokenForOrg: ["POST /orgs/{org}/actions/runners/registration-token"],
|
|
createRegistrationTokenForRepo: ["POST /repos/{owner}/{repo}/actions/runners/registration-token"],
|
|
createRemoveTokenForOrg: ["POST /orgs/{org}/actions/runners/remove-token"],
|
|
createRemoveTokenForRepo: ["POST /repos/{owner}/{repo}/actions/runners/remove-token"],
|
|
createWorkflowDispatch: ["POST /repos/{owner}/{repo}/actions/workflows/{workflow_id}/dispatches"],
|
|
deleteArtifact: ["DELETE /repos/{owner}/{repo}/actions/artifacts/{artifact_id}"],
|
|
deleteEnvironmentSecret: ["DELETE /repositories/{repository_id}/environments/{environment_name}/secrets/{secret_name}"],
|
|
deleteOrgSecret: ["DELETE /orgs/{org}/actions/secrets/{secret_name}"],
|
|
deleteRepoSecret: ["DELETE /repos/{owner}/{repo}/actions/secrets/{secret_name}"],
|
|
deleteSelfHostedRunnerFromOrg: ["DELETE /orgs/{org}/actions/runners/{runner_id}"],
|
|
deleteSelfHostedRunnerFromRepo: ["DELETE /repos/{owner}/{repo}/actions/runners/{runner_id}"],
|
|
deleteWorkflowRun: ["DELETE /repos/{owner}/{repo}/actions/runs/{run_id}"],
|
|
deleteWorkflowRunLogs: ["DELETE /repos/{owner}/{repo}/actions/runs/{run_id}/logs"],
|
|
disableSelectedRepositoryGithubActionsOrganization: ["DELETE /orgs/{org}/actions/permissions/repositories/{repository_id}"],
|
|
disableWorkflow: ["PUT /repos/{owner}/{repo}/actions/workflows/{workflow_id}/disable"],
|
|
downloadArtifact: ["GET /repos/{owner}/{repo}/actions/artifacts/{artifact_id}/{archive_format}"],
|
|
downloadJobLogsForWorkflowRun: ["GET /repos/{owner}/{repo}/actions/jobs/{job_id}/logs"],
|
|
downloadWorkflowRunLogs: ["GET /repos/{owner}/{repo}/actions/runs/{run_id}/logs"],
|
|
enableSelectedRepositoryGithubActionsOrganization: ["PUT /orgs/{org}/actions/permissions/repositories/{repository_id}"],
|
|
enableWorkflow: ["PUT /repos/{owner}/{repo}/actions/workflows/{workflow_id}/enable"],
|
|
getAllowedActionsOrganization: ["GET /orgs/{org}/actions/permissions/selected-actions"],
|
|
getAllowedActionsRepository: ["GET /repos/{owner}/{repo}/actions/permissions/selected-actions"],
|
|
getArtifact: ["GET /repos/{owner}/{repo}/actions/artifacts/{artifact_id}"],
|
|
getEnvironmentPublicKey: ["GET /repositories/{repository_id}/environments/{environment_name}/secrets/public-key"],
|
|
getEnvironmentSecret: ["GET /repositories/{repository_id}/environments/{environment_name}/secrets/{secret_name}"],
|
|
getGithubActionsPermissionsOrganization: ["GET /orgs/{org}/actions/permissions"],
|
|
getGithubActionsPermissionsRepository: ["GET /repos/{owner}/{repo}/actions/permissions"],
|
|
getJobForWorkflowRun: ["GET /repos/{owner}/{repo}/actions/jobs/{job_id}"],
|
|
getOrgPublicKey: ["GET /orgs/{org}/actions/secrets/public-key"],
|
|
getOrgSecret: ["GET /orgs/{org}/actions/secrets/{secret_name}"],
|
|
getPendingDeploymentsForRun: ["GET /repos/{owner}/{repo}/actions/runs/{run_id}/pending_deployments"],
|
|
getRepoPermissions: ["GET /repos/{owner}/{repo}/actions/permissions", {}, {
|
|
renamed: ["actions", "getGithubActionsPermissionsRepository"]
|
|
}],
|
|
getRepoPublicKey: ["GET /repos/{owner}/{repo}/actions/secrets/public-key"],
|
|
getRepoSecret: ["GET /repos/{owner}/{repo}/actions/secrets/{secret_name}"],
|
|
getReviewsForRun: ["GET /repos/{owner}/{repo}/actions/runs/{run_id}/approvals"],
|
|
getSelfHostedRunnerForOrg: ["GET /orgs/{org}/actions/runners/{runner_id}"],
|
|
getSelfHostedRunnerForRepo: ["GET /repos/{owner}/{repo}/actions/runners/{runner_id}"],
|
|
getWorkflow: ["GET /repos/{owner}/{repo}/actions/workflows/{workflow_id}"],
|
|
getWorkflowRun: ["GET /repos/{owner}/{repo}/actions/runs/{run_id}"],
|
|
getWorkflowRunUsage: ["GET /repos/{owner}/{repo}/actions/runs/{run_id}/timing"],
|
|
getWorkflowUsage: ["GET /repos/{owner}/{repo}/actions/workflows/{workflow_id}/timing"],
|
|
listArtifactsForRepo: ["GET /repos/{owner}/{repo}/actions/artifacts"],
|
|
listEnvironmentSecrets: ["GET /repositories/{repository_id}/environments/{environment_name}/secrets"],
|
|
listJobsForWorkflowRun: ["GET /repos/{owner}/{repo}/actions/runs/{run_id}/jobs"],
|
|
listOrgSecrets: ["GET /orgs/{org}/actions/secrets"],
|
|
listRepoSecrets: ["GET /repos/{owner}/{repo}/actions/secrets"],
|
|
listRepoWorkflows: ["GET /repos/{owner}/{repo}/actions/workflows"],
|
|
listRunnerApplicationsForOrg: ["GET /orgs/{org}/actions/runners/downloads"],
|
|
listRunnerApplicationsForRepo: ["GET /repos/{owner}/{repo}/actions/runners/downloads"],
|
|
listSelectedReposForOrgSecret: ["GET /orgs/{org}/actions/secrets/{secret_name}/repositories"],
|
|
listSelectedRepositoriesEnabledGithubActionsOrganization: ["GET /orgs/{org}/actions/permissions/repositories"],
|
|
listSelfHostedRunnersForOrg: ["GET /orgs/{org}/actions/runners"],
|
|
listSelfHostedRunnersForRepo: ["GET /repos/{owner}/{repo}/actions/runners"],
|
|
listWorkflowRunArtifacts: ["GET /repos/{owner}/{repo}/actions/runs/{run_id}/artifacts"],
|
|
listWorkflowRuns: ["GET /repos/{owner}/{repo}/actions/workflows/{workflow_id}/runs"],
|
|
listWorkflowRunsForRepo: ["GET /repos/{owner}/{repo}/actions/runs"],
|
|
reRunWorkflow: ["POST /repos/{owner}/{repo}/actions/runs/{run_id}/rerun"],
|
|
removeSelectedRepoFromOrgSecret: ["DELETE /orgs/{org}/actions/secrets/{secret_name}/repositories/{repository_id}"],
|
|
reviewPendingDeploymentsForRun: ["POST /repos/{owner}/{repo}/actions/runs/{run_id}/pending_deployments"],
|
|
setAllowedActionsOrganization: ["PUT /orgs/{org}/actions/permissions/selected-actions"],
|
|
setAllowedActionsRepository: ["PUT /repos/{owner}/{repo}/actions/permissions/selected-actions"],
|
|
setGithubActionsPermissionsOrganization: ["PUT /orgs/{org}/actions/permissions"],
|
|
setGithubActionsPermissionsRepository: ["PUT /repos/{owner}/{repo}/actions/permissions"],
|
|
setSelectedReposForOrgSecret: ["PUT /orgs/{org}/actions/secrets/{secret_name}/repositories"],
|
|
setSelectedRepositoriesEnabledGithubActionsOrganization: ["PUT /orgs/{org}/actions/permissions/repositories"]
|
|
},
|
|
activity: {
|
|
checkRepoIsStarredByAuthenticatedUser: ["GET /user/starred/{owner}/{repo}"],
|
|
deleteRepoSubscription: ["DELETE /repos/{owner}/{repo}/subscription"],
|
|
deleteThreadSubscription: ["DELETE /notifications/threads/{thread_id}/subscription"],
|
|
getFeeds: ["GET /feeds"],
|
|
getRepoSubscription: ["GET /repos/{owner}/{repo}/subscription"],
|
|
getThread: ["GET /notifications/threads/{thread_id}"],
|
|
getThreadSubscriptionForAuthenticatedUser: ["GET /notifications/threads/{thread_id}/subscription"],
|
|
listEventsForAuthenticatedUser: ["GET /users/{username}/events"],
|
|
listNotificationsForAuthenticatedUser: ["GET /notifications"],
|
|
listOrgEventsForAuthenticatedUser: ["GET /users/{username}/events/orgs/{org}"],
|
|
listPublicEvents: ["GET /events"],
|
|
listPublicEventsForRepoNetwork: ["GET /networks/{owner}/{repo}/events"],
|
|
listPublicEventsForUser: ["GET /users/{username}/events/public"],
|
|
listPublicOrgEvents: ["GET /orgs/{org}/events"],
|
|
listReceivedEventsForUser: ["GET /users/{username}/received_events"],
|
|
listReceivedPublicEventsForUser: ["GET /users/{username}/received_events/public"],
|
|
listRepoEvents: ["GET /repos/{owner}/{repo}/events"],
|
|
listRepoNotificationsForAuthenticatedUser: ["GET /repos/{owner}/{repo}/notifications"],
|
|
listReposStarredByAuthenticatedUser: ["GET /user/starred"],
|
|
listReposStarredByUser: ["GET /users/{username}/starred"],
|
|
listReposWatchedByUser: ["GET /users/{username}/subscriptions"],
|
|
listStargazersForRepo: ["GET /repos/{owner}/{repo}/stargazers"],
|
|
listWatchedReposForAuthenticatedUser: ["GET /user/subscriptions"],
|
|
listWatchersForRepo: ["GET /repos/{owner}/{repo}/subscribers"],
|
|
markNotificationsAsRead: ["PUT /notifications"],
|
|
markRepoNotificationsAsRead: ["PUT /repos/{owner}/{repo}/notifications"],
|
|
markThreadAsRead: ["PATCH /notifications/threads/{thread_id}"],
|
|
setRepoSubscription: ["PUT /repos/{owner}/{repo}/subscription"],
|
|
setThreadSubscription: ["PUT /notifications/threads/{thread_id}/subscription"],
|
|
starRepoForAuthenticatedUser: ["PUT /user/starred/{owner}/{repo}"],
|
|
unstarRepoForAuthenticatedUser: ["DELETE /user/starred/{owner}/{repo}"]
|
|
},
|
|
apps: {
|
|
addRepoToInstallation: ["PUT /user/installations/{installation_id}/repositories/{repository_id}"],
|
|
checkToken: ["POST /applications/{client_id}/token"],
|
|
createContentAttachment: ["POST /content_references/{content_reference_id}/attachments", {
|
|
mediaType: {
|
|
previews: ["corsair"]
|
|
}
|
|
}],
|
|
createContentAttachmentForRepo: ["POST /repos/{owner}/{repo}/content_references/{content_reference_id}/attachments", {
|
|
mediaType: {
|
|
previews: ["corsair"]
|
|
}
|
|
}],
|
|
createFromManifest: ["POST /app-manifests/{code}/conversions"],
|
|
createInstallationAccessToken: ["POST /app/installations/{installation_id}/access_tokens"],
|
|
deleteAuthorization: ["DELETE /applications/{client_id}/grant"],
|
|
deleteInstallation: ["DELETE /app/installations/{installation_id}"],
|
|
deleteToken: ["DELETE /applications/{client_id}/token"],
|
|
getAuthenticated: ["GET /app"],
|
|
getBySlug: ["GET /apps/{app_slug}"],
|
|
getInstallation: ["GET /app/installations/{installation_id}"],
|
|
getOrgInstallation: ["GET /orgs/{org}/installation"],
|
|
getRepoInstallation: ["GET /repos/{owner}/{repo}/installation"],
|
|
getSubscriptionPlanForAccount: ["GET /marketplace_listing/accounts/{account_id}"],
|
|
getSubscriptionPlanForAccountStubbed: ["GET /marketplace_listing/stubbed/accounts/{account_id}"],
|
|
getUserInstallation: ["GET /users/{username}/installation"],
|
|
getWebhookConfigForApp: ["GET /app/hook/config"],
|
|
listAccountsForPlan: ["GET /marketplace_listing/plans/{plan_id}/accounts"],
|
|
listAccountsForPlanStubbed: ["GET /marketplace_listing/stubbed/plans/{plan_id}/accounts"],
|
|
listInstallationReposForAuthenticatedUser: ["GET /user/installations/{installation_id}/repositories"],
|
|
listInstallations: ["GET /app/installations"],
|
|
listInstallationsForAuthenticatedUser: ["GET /user/installations"],
|
|
listPlans: ["GET /marketplace_listing/plans"],
|
|
listPlansStubbed: ["GET /marketplace_listing/stubbed/plans"],
|
|
listReposAccessibleToInstallation: ["GET /installation/repositories"],
|
|
listSubscriptionsForAuthenticatedUser: ["GET /user/marketplace_purchases"],
|
|
listSubscriptionsForAuthenticatedUserStubbed: ["GET /user/marketplace_purchases/stubbed"],
|
|
removeRepoFromInstallation: ["DELETE /user/installations/{installation_id}/repositories/{repository_id}"],
|
|
resetToken: ["PATCH /applications/{client_id}/token"],
|
|
revokeInstallationAccessToken: ["DELETE /installation/token"],
|
|
scopeToken: ["POST /applications/{client_id}/token/scoped"],
|
|
suspendInstallation: ["PUT /app/installations/{installation_id}/suspended"],
|
|
unsuspendInstallation: ["DELETE /app/installations/{installation_id}/suspended"],
|
|
updateWebhookConfigForApp: ["PATCH /app/hook/config"]
|
|
},
|
|
billing: {
|
|
getGithubActionsBillingOrg: ["GET /orgs/{org}/settings/billing/actions"],
|
|
getGithubActionsBillingUser: ["GET /users/{username}/settings/billing/actions"],
|
|
getGithubPackagesBillingOrg: ["GET /orgs/{org}/settings/billing/packages"],
|
|
getGithubPackagesBillingUser: ["GET /users/{username}/settings/billing/packages"],
|
|
getSharedStorageBillingOrg: ["GET /orgs/{org}/settings/billing/shared-storage"],
|
|
getSharedStorageBillingUser: ["GET /users/{username}/settings/billing/shared-storage"]
|
|
},
|
|
checks: {
|
|
create: ["POST /repos/{owner}/{repo}/check-runs"],
|
|
createSuite: ["POST /repos/{owner}/{repo}/check-suites"],
|
|
get: ["GET /repos/{owner}/{repo}/check-runs/{check_run_id}"],
|
|
getSuite: ["GET /repos/{owner}/{repo}/check-suites/{check_suite_id}"],
|
|
listAnnotations: ["GET /repos/{owner}/{repo}/check-runs/{check_run_id}/annotations"],
|
|
listForRef: ["GET /repos/{owner}/{repo}/commits/{ref}/check-runs"],
|
|
listForSuite: ["GET /repos/{owner}/{repo}/check-suites/{check_suite_id}/check-runs"],
|
|
listSuitesForRef: ["GET /repos/{owner}/{repo}/commits/{ref}/check-suites"],
|
|
rerequestSuite: ["POST /repos/{owner}/{repo}/check-suites/{check_suite_id}/rerequest"],
|
|
setSuitesPreferences: ["PATCH /repos/{owner}/{repo}/check-suites/preferences"],
|
|
update: ["PATCH /repos/{owner}/{repo}/check-runs/{check_run_id}"]
|
|
},
|
|
codeScanning: {
|
|
deleteAnalysis: ["DELETE /repos/{owner}/{repo}/code-scanning/analyses/{analysis_id}{?confirm_delete}"],
|
|
getAlert: ["GET /repos/{owner}/{repo}/code-scanning/alerts/{alert_number}", {}, {
|
|
renamedParameters: {
|
|
alert_id: "alert_number"
|
|
}
|
|
}],
|
|
getAnalysis: ["GET /repos/{owner}/{repo}/code-scanning/analyses/{analysis_id}"],
|
|
getSarif: ["GET /repos/{owner}/{repo}/code-scanning/sarifs/{sarif_id}"],
|
|
listAlertInstances: ["GET /repos/{owner}/{repo}/code-scanning/alerts/{alert_number}/instances"],
|
|
listAlertsForRepo: ["GET /repos/{owner}/{repo}/code-scanning/alerts"],
|
|
listAlertsInstances: ["GET /repos/{owner}/{repo}/code-scanning/alerts/{alert_number}/instances", {}, {
|
|
renamed: ["codeScanning", "listAlertInstances"]
|
|
}],
|
|
listRecentAnalyses: ["GET /repos/{owner}/{repo}/code-scanning/analyses"],
|
|
updateAlert: ["PATCH /repos/{owner}/{repo}/code-scanning/alerts/{alert_number}"],
|
|
uploadSarif: ["POST /repos/{owner}/{repo}/code-scanning/sarifs"]
|
|
},
|
|
codesOfConduct: {
|
|
getAllCodesOfConduct: ["GET /codes_of_conduct", {
|
|
mediaType: {
|
|
previews: ["scarlet-witch"]
|
|
}
|
|
}],
|
|
getConductCode: ["GET /codes_of_conduct/{key}", {
|
|
mediaType: {
|
|
previews: ["scarlet-witch"]
|
|
}
|
|
}],
|
|
getForRepo: ["GET /repos/{owner}/{repo}/community/code_of_conduct", {
|
|
mediaType: {
|
|
previews: ["scarlet-witch"]
|
|
}
|
|
}]
|
|
},
|
|
emojis: {
|
|
get: ["GET /emojis"]
|
|
},
|
|
enterpriseAdmin: {
|
|
disableSelectedOrganizationGithubActionsEnterprise: ["DELETE /enterprises/{enterprise}/actions/permissions/organizations/{org_id}"],
|
|
enableSelectedOrganizationGithubActionsEnterprise: ["PUT /enterprises/{enterprise}/actions/permissions/organizations/{org_id}"],
|
|
getAllowedActionsEnterprise: ["GET /enterprises/{enterprise}/actions/permissions/selected-actions"],
|
|
getGithubActionsPermissionsEnterprise: ["GET /enterprises/{enterprise}/actions/permissions"],
|
|
listSelectedOrganizationsEnabledGithubActionsEnterprise: ["GET /enterprises/{enterprise}/actions/permissions/organizations"],
|
|
setAllowedActionsEnterprise: ["PUT /enterprises/{enterprise}/actions/permissions/selected-actions"],
|
|
setGithubActionsPermissionsEnterprise: ["PUT /enterprises/{enterprise}/actions/permissions"],
|
|
setSelectedOrganizationsEnabledGithubActionsEnterprise: ["PUT /enterprises/{enterprise}/actions/permissions/organizations"]
|
|
},
|
|
gists: {
|
|
checkIsStarred: ["GET /gists/{gist_id}/star"],
|
|
create: ["POST /gists"],
|
|
createComment: ["POST /gists/{gist_id}/comments"],
|
|
delete: ["DELETE /gists/{gist_id}"],
|
|
deleteComment: ["DELETE /gists/{gist_id}/comments/{comment_id}"],
|
|
fork: ["POST /gists/{gist_id}/forks"],
|
|
get: ["GET /gists/{gist_id}"],
|
|
getComment: ["GET /gists/{gist_id}/comments/{comment_id}"],
|
|
getRevision: ["GET /gists/{gist_id}/{sha}"],
|
|
list: ["GET /gists"],
|
|
listComments: ["GET /gists/{gist_id}/comments"],
|
|
listCommits: ["GET /gists/{gist_id}/commits"],
|
|
listForUser: ["GET /users/{username}/gists"],
|
|
listForks: ["GET /gists/{gist_id}/forks"],
|
|
listPublic: ["GET /gists/public"],
|
|
listStarred: ["GET /gists/starred"],
|
|
star: ["PUT /gists/{gist_id}/star"],
|
|
unstar: ["DELETE /gists/{gist_id}/star"],
|
|
update: ["PATCH /gists/{gist_id}"],
|
|
updateComment: ["PATCH /gists/{gist_id}/comments/{comment_id}"]
|
|
},
|
|
git: {
|
|
createBlob: ["POST /repos/{owner}/{repo}/git/blobs"],
|
|
createCommit: ["POST /repos/{owner}/{repo}/git/commits"],
|
|
createRef: ["POST /repos/{owner}/{repo}/git/refs"],
|
|
createTag: ["POST /repos/{owner}/{repo}/git/tags"],
|
|
createTree: ["POST /repos/{owner}/{repo}/git/trees"],
|
|
deleteRef: ["DELETE /repos/{owner}/{repo}/git/refs/{ref}"],
|
|
getBlob: ["GET /repos/{owner}/{repo}/git/blobs/{file_sha}"],
|
|
getCommit: ["GET /repos/{owner}/{repo}/git/commits/{commit_sha}"],
|
|
getRef: ["GET /repos/{owner}/{repo}/git/ref/{ref}"],
|
|
getTag: ["GET /repos/{owner}/{repo}/git/tags/{tag_sha}"],
|
|
getTree: ["GET /repos/{owner}/{repo}/git/trees/{tree_sha}"],
|
|
listMatchingRefs: ["GET /repos/{owner}/{repo}/git/matching-refs/{ref}"],
|
|
updateRef: ["PATCH /repos/{owner}/{repo}/git/refs/{ref}"]
|
|
},
|
|
gitignore: {
|
|
getAllTemplates: ["GET /gitignore/templates"],
|
|
getTemplate: ["GET /gitignore/templates/{name}"]
|
|
},
|
|
interactions: {
|
|
getRestrictionsForAuthenticatedUser: ["GET /user/interaction-limits"],
|
|
getRestrictionsForOrg: ["GET /orgs/{org}/interaction-limits"],
|
|
getRestrictionsForRepo: ["GET /repos/{owner}/{repo}/interaction-limits"],
|
|
getRestrictionsForYourPublicRepos: ["GET /user/interaction-limits", {}, {
|
|
renamed: ["interactions", "getRestrictionsForAuthenticatedUser"]
|
|
}],
|
|
removeRestrictionsForAuthenticatedUser: ["DELETE /user/interaction-limits"],
|
|
removeRestrictionsForOrg: ["DELETE /orgs/{org}/interaction-limits"],
|
|
removeRestrictionsForRepo: ["DELETE /repos/{owner}/{repo}/interaction-limits"],
|
|
removeRestrictionsForYourPublicRepos: ["DELETE /user/interaction-limits", {}, {
|
|
renamed: ["interactions", "removeRestrictionsForAuthenticatedUser"]
|
|
}],
|
|
setRestrictionsForAuthenticatedUser: ["PUT /user/interaction-limits"],
|
|
setRestrictionsForOrg: ["PUT /orgs/{org}/interaction-limits"],
|
|
setRestrictionsForRepo: ["PUT /repos/{owner}/{repo}/interaction-limits"],
|
|
setRestrictionsForYourPublicRepos: ["PUT /user/interaction-limits", {}, {
|
|
renamed: ["interactions", "setRestrictionsForAuthenticatedUser"]
|
|
}]
|
|
},
|
|
issues: {
|
|
addAssignees: ["POST /repos/{owner}/{repo}/issues/{issue_number}/assignees"],
|
|
addLabels: ["POST /repos/{owner}/{repo}/issues/{issue_number}/labels"],
|
|
checkUserCanBeAssigned: ["GET /repos/{owner}/{repo}/assignees/{assignee}"],
|
|
create: ["POST /repos/{owner}/{repo}/issues"],
|
|
createComment: ["POST /repos/{owner}/{repo}/issues/{issue_number}/comments"],
|
|
createLabel: ["POST /repos/{owner}/{repo}/labels"],
|
|
createMilestone: ["POST /repos/{owner}/{repo}/milestones"],
|
|
deleteComment: ["DELETE /repos/{owner}/{repo}/issues/comments/{comment_id}"],
|
|
deleteLabel: ["DELETE /repos/{owner}/{repo}/labels/{name}"],
|
|
deleteMilestone: ["DELETE /repos/{owner}/{repo}/milestones/{milestone_number}"],
|
|
get: ["GET /repos/{owner}/{repo}/issues/{issue_number}"],
|
|
getComment: ["GET /repos/{owner}/{repo}/issues/comments/{comment_id}"],
|
|
getEvent: ["GET /repos/{owner}/{repo}/issues/events/{event_id}"],
|
|
getLabel: ["GET /repos/{owner}/{repo}/labels/{name}"],
|
|
getMilestone: ["GET /repos/{owner}/{repo}/milestones/{milestone_number}"],
|
|
list: ["GET /issues"],
|
|
listAssignees: ["GET /repos/{owner}/{repo}/assignees"],
|
|
listComments: ["GET /repos/{owner}/{repo}/issues/{issue_number}/comments"],
|
|
listCommentsForRepo: ["GET /repos/{owner}/{repo}/issues/comments"],
|
|
listEvents: ["GET /repos/{owner}/{repo}/issues/{issue_number}/events"],
|
|
listEventsForRepo: ["GET /repos/{owner}/{repo}/issues/events"],
|
|
listEventsForTimeline: ["GET /repos/{owner}/{repo}/issues/{issue_number}/timeline", {
|
|
mediaType: {
|
|
previews: ["mockingbird"]
|
|
}
|
|
}],
|
|
listForAuthenticatedUser: ["GET /user/issues"],
|
|
listForOrg: ["GET /orgs/{org}/issues"],
|
|
listForRepo: ["GET /repos/{owner}/{repo}/issues"],
|
|
listLabelsForMilestone: ["GET /repos/{owner}/{repo}/milestones/{milestone_number}/labels"],
|
|
listLabelsForRepo: ["GET /repos/{owner}/{repo}/labels"],
|
|
listLabelsOnIssue: ["GET /repos/{owner}/{repo}/issues/{issue_number}/labels"],
|
|
listMilestones: ["GET /repos/{owner}/{repo}/milestones"],
|
|
lock: ["PUT /repos/{owner}/{repo}/issues/{issue_number}/lock"],
|
|
removeAllLabels: ["DELETE /repos/{owner}/{repo}/issues/{issue_number}/labels"],
|
|
removeAssignees: ["DELETE /repos/{owner}/{repo}/issues/{issue_number}/assignees"],
|
|
removeLabel: ["DELETE /repos/{owner}/{repo}/issues/{issue_number}/labels/{name}"],
|
|
setLabels: ["PUT /repos/{owner}/{repo}/issues/{issue_number}/labels"],
|
|
unlock: ["DELETE /repos/{owner}/{repo}/issues/{issue_number}/lock"],
|
|
update: ["PATCH /repos/{owner}/{repo}/issues/{issue_number}"],
|
|
updateComment: ["PATCH /repos/{owner}/{repo}/issues/comments/{comment_id}"],
|
|
updateLabel: ["PATCH /repos/{owner}/{repo}/labels/{name}"],
|
|
updateMilestone: ["PATCH /repos/{owner}/{repo}/milestones/{milestone_number}"]
|
|
},
|
|
licenses: {
|
|
get: ["GET /licenses/{license}"],
|
|
getAllCommonlyUsed: ["GET /licenses"],
|
|
getForRepo: ["GET /repos/{owner}/{repo}/license"]
|
|
},
|
|
markdown: {
|
|
render: ["POST /markdown"],
|
|
renderRaw: ["POST /markdown/raw", {
|
|
headers: {
|
|
"content-type": "text/plain; charset=utf-8"
|
|
}
|
|
}]
|
|
},
|
|
meta: {
|
|
get: ["GET /meta"],
|
|
getOctocat: ["GET /octocat"],
|
|
getZen: ["GET /zen"],
|
|
root: ["GET /"]
|
|
},
|
|
migrations: {
|
|
cancelImport: ["DELETE /repos/{owner}/{repo}/import"],
|
|
deleteArchiveForAuthenticatedUser: ["DELETE /user/migrations/{migration_id}/archive", {
|
|
mediaType: {
|
|
previews: ["wyandotte"]
|
|
}
|
|
}],
|
|
deleteArchiveForOrg: ["DELETE /orgs/{org}/migrations/{migration_id}/archive", {
|
|
mediaType: {
|
|
previews: ["wyandotte"]
|
|
}
|
|
}],
|
|
downloadArchiveForOrg: ["GET /orgs/{org}/migrations/{migration_id}/archive", {
|
|
mediaType: {
|
|
previews: ["wyandotte"]
|
|
}
|
|
}],
|
|
getArchiveForAuthenticatedUser: ["GET /user/migrations/{migration_id}/archive", {
|
|
mediaType: {
|
|
previews: ["wyandotte"]
|
|
}
|
|
}],
|
|
getCommitAuthors: ["GET /repos/{owner}/{repo}/import/authors"],
|
|
getImportStatus: ["GET /repos/{owner}/{repo}/import"],
|
|
getLargeFiles: ["GET /repos/{owner}/{repo}/import/large_files"],
|
|
getStatusForAuthenticatedUser: ["GET /user/migrations/{migration_id}", {
|
|
mediaType: {
|
|
previews: ["wyandotte"]
|
|
}
|
|
}],
|
|
getStatusForOrg: ["GET /orgs/{org}/migrations/{migration_id}", {
|
|
mediaType: {
|
|
previews: ["wyandotte"]
|
|
}
|
|
}],
|
|
listForAuthenticatedUser: ["GET /user/migrations", {
|
|
mediaType: {
|
|
previews: ["wyandotte"]
|
|
}
|
|
}],
|
|
listForOrg: ["GET /orgs/{org}/migrations", {
|
|
mediaType: {
|
|
previews: ["wyandotte"]
|
|
}
|
|
}],
|
|
listReposForOrg: ["GET /orgs/{org}/migrations/{migration_id}/repositories", {
|
|
mediaType: {
|
|
previews: ["wyandotte"]
|
|
}
|
|
}],
|
|
listReposForUser: ["GET /user/migrations/{migration_id}/repositories", {
|
|
mediaType: {
|
|
previews: ["wyandotte"]
|
|
}
|
|
}],
|
|
mapCommitAuthor: ["PATCH /repos/{owner}/{repo}/import/authors/{author_id}"],
|
|
setLfsPreference: ["PATCH /repos/{owner}/{repo}/import/lfs"],
|
|
startForAuthenticatedUser: ["POST /user/migrations"],
|
|
startForOrg: ["POST /orgs/{org}/migrations"],
|
|
startImport: ["PUT /repos/{owner}/{repo}/import"],
|
|
unlockRepoForAuthenticatedUser: ["DELETE /user/migrations/{migration_id}/repos/{repo_name}/lock", {
|
|
mediaType: {
|
|
previews: ["wyandotte"]
|
|
}
|
|
}],
|
|
unlockRepoForOrg: ["DELETE /orgs/{org}/migrations/{migration_id}/repos/{repo_name}/lock", {
|
|
mediaType: {
|
|
previews: ["wyandotte"]
|
|
}
|
|
}],
|
|
updateImport: ["PATCH /repos/{owner}/{repo}/import"]
|
|
},
|
|
orgs: {
|
|
blockUser: ["PUT /orgs/{org}/blocks/{username}"],
|
|
cancelInvitation: ["DELETE /orgs/{org}/invitations/{invitation_id}"],
|
|
checkBlockedUser: ["GET /orgs/{org}/blocks/{username}"],
|
|
checkMembershipForUser: ["GET /orgs/{org}/members/{username}"],
|
|
checkPublicMembershipForUser: ["GET /orgs/{org}/public_members/{username}"],
|
|
convertMemberToOutsideCollaborator: ["PUT /orgs/{org}/outside_collaborators/{username}"],
|
|
createInvitation: ["POST /orgs/{org}/invitations"],
|
|
createWebhook: ["POST /orgs/{org}/hooks"],
|
|
deleteWebhook: ["DELETE /orgs/{org}/hooks/{hook_id}"],
|
|
get: ["GET /orgs/{org}"],
|
|
getMembershipForAuthenticatedUser: ["GET /user/memberships/orgs/{org}"],
|
|
getMembershipForUser: ["GET /orgs/{org}/memberships/{username}"],
|
|
getWebhook: ["GET /orgs/{org}/hooks/{hook_id}"],
|
|
getWebhookConfigForOrg: ["GET /orgs/{org}/hooks/{hook_id}/config"],
|
|
list: ["GET /organizations"],
|
|
listAppInstallations: ["GET /orgs/{org}/installations"],
|
|
listBlockedUsers: ["GET /orgs/{org}/blocks"],
|
|
listFailedInvitations: ["GET /orgs/{org}/failed_invitations"],
|
|
listForAuthenticatedUser: ["GET /user/orgs"],
|
|
listForUser: ["GET /users/{username}/orgs"],
|
|
listInvitationTeams: ["GET /orgs/{org}/invitations/{invitation_id}/teams"],
|
|
listMembers: ["GET /orgs/{org}/members"],
|
|
listMembershipsForAuthenticatedUser: ["GET /user/memberships/orgs"],
|
|
listOutsideCollaborators: ["GET /orgs/{org}/outside_collaborators"],
|
|
listPendingInvitations: ["GET /orgs/{org}/invitations"],
|
|
listPublicMembers: ["GET /orgs/{org}/public_members"],
|
|
listWebhooks: ["GET /orgs/{org}/hooks"],
|
|
pingWebhook: ["POST /orgs/{org}/hooks/{hook_id}/pings"],
|
|
removeMember: ["DELETE /orgs/{org}/members/{username}"],
|
|
removeMembershipForUser: ["DELETE /orgs/{org}/memberships/{username}"],
|
|
removeOutsideCollaborator: ["DELETE /orgs/{org}/outside_collaborators/{username}"],
|
|
removePublicMembershipForAuthenticatedUser: ["DELETE /orgs/{org}/public_members/{username}"],
|
|
setMembershipForUser: ["PUT /orgs/{org}/memberships/{username}"],
|
|
setPublicMembershipForAuthenticatedUser: ["PUT /orgs/{org}/public_members/{username}"],
|
|
unblockUser: ["DELETE /orgs/{org}/blocks/{username}"],
|
|
update: ["PATCH /orgs/{org}"],
|
|
updateMembershipForAuthenticatedUser: ["PATCH /user/memberships/orgs/{org}"],
|
|
updateWebhook: ["PATCH /orgs/{org}/hooks/{hook_id}"],
|
|
updateWebhookConfigForOrg: ["PATCH /orgs/{org}/hooks/{hook_id}/config"]
|
|
},
|
|
packages: {
|
|
deletePackageForAuthenticatedUser: ["DELETE /user/packages/{package_type}/{package_name}"],
|
|
deletePackageForOrg: ["DELETE /orgs/{org}/packages/{package_type}/{package_name}"],
|
|
deletePackageVersionForAuthenticatedUser: ["DELETE /user/packages/{package_type}/{package_name}/versions/{package_version_id}"],
|
|
deletePackageVersionForOrg: ["DELETE /orgs/{org}/packages/{package_type}/{package_name}/versions/{package_version_id}"],
|
|
getAllPackageVersionsForAPackageOwnedByAnOrg: ["GET /orgs/{org}/packages/{package_type}/{package_name}/versions", {}, {
|
|
renamed: ["packages", "getAllPackageVersionsForPackageOwnedByOrg"]
|
|
}],
|
|
getAllPackageVersionsForAPackageOwnedByTheAuthenticatedUser: ["GET /user/packages/{package_type}/{package_name}/versions", {}, {
|
|
renamed: ["packages", "getAllPackageVersionsForPackageOwnedByAuthenticatedUser"]
|
|
}],
|
|
getAllPackageVersionsForPackageOwnedByAuthenticatedUser: ["GET /user/packages/{package_type}/{package_name}/versions"],
|
|
getAllPackageVersionsForPackageOwnedByOrg: ["GET /orgs/{org}/packages/{package_type}/{package_name}/versions"],
|
|
getAllPackageVersionsForPackageOwnedByUser: ["GET /users/{username}/packages/{package_type}/{package_name}/versions"],
|
|
getPackageForAuthenticatedUser: ["GET /user/packages/{package_type}/{package_name}"],
|
|
getPackageForOrganization: ["GET /orgs/{org}/packages/{package_type}/{package_name}"],
|
|
getPackageForUser: ["GET /users/{username}/packages/{package_type}/{package_name}"],
|
|
getPackageVersionForAuthenticatedUser: ["GET /user/packages/{package_type}/{package_name}/versions/{package_version_id}"],
|
|
getPackageVersionForOrganization: ["GET /orgs/{org}/packages/{package_type}/{package_name}/versions/{package_version_id}"],
|
|
getPackageVersionForUser: ["GET /users/{username}/packages/{package_type}/{package_name}/versions/{package_version_id}"],
|
|
restorePackageForAuthenticatedUser: ["POST /user/packages/{package_type}/{package_name}/restore{?token}"],
|
|
restorePackageForOrg: ["POST /orgs/{org}/packages/{package_type}/{package_name}/restore{?token}"],
|
|
restorePackageVersionForAuthenticatedUser: ["POST /user/packages/{package_type}/{package_name}/versions/{package_version_id}/restore"],
|
|
restorePackageVersionForOrg: ["POST /orgs/{org}/packages/{package_type}/{package_name}/versions/{package_version_id}/restore"]
|
|
},
|
|
projects: {
|
|
addCollaborator: ["PUT /projects/{project_id}/collaborators/{username}", {
|
|
mediaType: {
|
|
previews: ["inertia"]
|
|
}
|
|
}],
|
|
createCard: ["POST /projects/columns/{column_id}/cards", {
|
|
mediaType: {
|
|
previews: ["inertia"]
|
|
}
|
|
}],
|
|
createColumn: ["POST /projects/{project_id}/columns", {
|
|
mediaType: {
|
|
previews: ["inertia"]
|
|
}
|
|
}],
|
|
createForAuthenticatedUser: ["POST /user/projects", {
|
|
mediaType: {
|
|
previews: ["inertia"]
|
|
}
|
|
}],
|
|
createForOrg: ["POST /orgs/{org}/projects", {
|
|
mediaType: {
|
|
previews: ["inertia"]
|
|
}
|
|
}],
|
|
createForRepo: ["POST /repos/{owner}/{repo}/projects", {
|
|
mediaType: {
|
|
previews: ["inertia"]
|
|
}
|
|
}],
|
|
delete: ["DELETE /projects/{project_id}", {
|
|
mediaType: {
|
|
previews: ["inertia"]
|
|
}
|
|
}],
|
|
deleteCard: ["DELETE /projects/columns/cards/{card_id}", {
|
|
mediaType: {
|
|
previews: ["inertia"]
|
|
}
|
|
}],
|
|
deleteColumn: ["DELETE /projects/columns/{column_id}", {
|
|
mediaType: {
|
|
previews: ["inertia"]
|
|
}
|
|
}],
|
|
get: ["GET /projects/{project_id}", {
|
|
mediaType: {
|
|
previews: ["inertia"]
|
|
}
|
|
}],
|
|
getCard: ["GET /projects/columns/cards/{card_id}", {
|
|
mediaType: {
|
|
previews: ["inertia"]
|
|
}
|
|
}],
|
|
getColumn: ["GET /projects/columns/{column_id}", {
|
|
mediaType: {
|
|
previews: ["inertia"]
|
|
}
|
|
}],
|
|
getPermissionForUser: ["GET /projects/{project_id}/collaborators/{username}/permission", {
|
|
mediaType: {
|
|
previews: ["inertia"]
|
|
}
|
|
}],
|
|
listCards: ["GET /projects/columns/{column_id}/cards", {
|
|
mediaType: {
|
|
previews: ["inertia"]
|
|
}
|
|
}],
|
|
listCollaborators: ["GET /projects/{project_id}/collaborators", {
|
|
mediaType: {
|
|
previews: ["inertia"]
|
|
}
|
|
}],
|
|
listColumns: ["GET /projects/{project_id}/columns", {
|
|
mediaType: {
|
|
previews: ["inertia"]
|
|
}
|
|
}],
|
|
listForOrg: ["GET /orgs/{org}/projects", {
|
|
mediaType: {
|
|
previews: ["inertia"]
|
|
}
|
|
}],
|
|
listForRepo: ["GET /repos/{owner}/{repo}/projects", {
|
|
mediaType: {
|
|
previews: ["inertia"]
|
|
}
|
|
}],
|
|
listForUser: ["GET /users/{username}/projects", {
|
|
mediaType: {
|
|
previews: ["inertia"]
|
|
}
|
|
}],
|
|
moveCard: ["POST /projects/columns/cards/{card_id}/moves", {
|
|
mediaType: {
|
|
previews: ["inertia"]
|
|
}
|
|
}],
|
|
moveColumn: ["POST /projects/columns/{column_id}/moves", {
|
|
mediaType: {
|
|
previews: ["inertia"]
|
|
}
|
|
}],
|
|
removeCollaborator: ["DELETE /projects/{project_id}/collaborators/{username}", {
|
|
mediaType: {
|
|
previews: ["inertia"]
|
|
}
|
|
}],
|
|
update: ["PATCH /projects/{project_id}", {
|
|
mediaType: {
|
|
previews: ["inertia"]
|
|
}
|
|
}],
|
|
updateCard: ["PATCH /projects/columns/cards/{card_id}", {
|
|
mediaType: {
|
|
previews: ["inertia"]
|
|
}
|
|
}],
|
|
updateColumn: ["PATCH /projects/columns/{column_id}", {
|
|
mediaType: {
|
|
previews: ["inertia"]
|
|
}
|
|
}]
|
|
},
|
|
pulls: {
|
|
checkIfMerged: ["GET /repos/{owner}/{repo}/pulls/{pull_number}/merge"],
|
|
create: ["POST /repos/{owner}/{repo}/pulls"],
|
|
createReplyForReviewComment: ["POST /repos/{owner}/{repo}/pulls/{pull_number}/comments/{comment_id}/replies"],
|
|
createReview: ["POST /repos/{owner}/{repo}/pulls/{pull_number}/reviews"],
|
|
createReviewComment: ["POST /repos/{owner}/{repo}/pulls/{pull_number}/comments"],
|
|
deletePendingReview: ["DELETE /repos/{owner}/{repo}/pulls/{pull_number}/reviews/{review_id}"],
|
|
deleteReviewComment: ["DELETE /repos/{owner}/{repo}/pulls/comments/{comment_id}"],
|
|
dismissReview: ["PUT /repos/{owner}/{repo}/pulls/{pull_number}/reviews/{review_id}/dismissals"],
|
|
get: ["GET /repos/{owner}/{repo}/pulls/{pull_number}"],
|
|
getReview: ["GET /repos/{owner}/{repo}/pulls/{pull_number}/reviews/{review_id}"],
|
|
getReviewComment: ["GET /repos/{owner}/{repo}/pulls/comments/{comment_id}"],
|
|
list: ["GET /repos/{owner}/{repo}/pulls"],
|
|
listCommentsForReview: ["GET /repos/{owner}/{repo}/pulls/{pull_number}/reviews/{review_id}/comments"],
|
|
listCommits: ["GET /repos/{owner}/{repo}/pulls/{pull_number}/commits"],
|
|
listFiles: ["GET /repos/{owner}/{repo}/pulls/{pull_number}/files"],
|
|
listRequestedReviewers: ["GET /repos/{owner}/{repo}/pulls/{pull_number}/requested_reviewers"],
|
|
listReviewComments: ["GET /repos/{owner}/{repo}/pulls/{pull_number}/comments"],
|
|
listReviewCommentsForRepo: ["GET /repos/{owner}/{repo}/pulls/comments"],
|
|
listReviews: ["GET /repos/{owner}/{repo}/pulls/{pull_number}/reviews"],
|
|
merge: ["PUT /repos/{owner}/{repo}/pulls/{pull_number}/merge"],
|
|
removeRequestedReviewers: ["DELETE /repos/{owner}/{repo}/pulls/{pull_number}/requested_reviewers"],
|
|
requestReviewers: ["POST /repos/{owner}/{repo}/pulls/{pull_number}/requested_reviewers"],
|
|
submitReview: ["POST /repos/{owner}/{repo}/pulls/{pull_number}/reviews/{review_id}/events"],
|
|
update: ["PATCH /repos/{owner}/{repo}/pulls/{pull_number}"],
|
|
updateBranch: ["PUT /repos/{owner}/{repo}/pulls/{pull_number}/update-branch", {
|
|
mediaType: {
|
|
previews: ["lydian"]
|
|
}
|
|
}],
|
|
updateReview: ["PUT /repos/{owner}/{repo}/pulls/{pull_number}/reviews/{review_id}"],
|
|
updateReviewComment: ["PATCH /repos/{owner}/{repo}/pulls/comments/{comment_id}"]
|
|
},
|
|
rateLimit: {
|
|
get: ["GET /rate_limit"]
|
|
},
|
|
reactions: {
|
|
createForCommitComment: ["POST /repos/{owner}/{repo}/comments/{comment_id}/reactions", {
|
|
mediaType: {
|
|
previews: ["squirrel-girl"]
|
|
}
|
|
}],
|
|
createForIssue: ["POST /repos/{owner}/{repo}/issues/{issue_number}/reactions", {
|
|
mediaType: {
|
|
previews: ["squirrel-girl"]
|
|
}
|
|
}],
|
|
createForIssueComment: ["POST /repos/{owner}/{repo}/issues/comments/{comment_id}/reactions", {
|
|
mediaType: {
|
|
previews: ["squirrel-girl"]
|
|
}
|
|
}],
|
|
createForPullRequestReviewComment: ["POST /repos/{owner}/{repo}/pulls/comments/{comment_id}/reactions", {
|
|
mediaType: {
|
|
previews: ["squirrel-girl"]
|
|
}
|
|
}],
|
|
createForRelease: ["POST /repos/{owner}/{repo}/releases/{release_id}/reactions", {
|
|
mediaType: {
|
|
previews: ["squirrel-girl"]
|
|
}
|
|
}],
|
|
createForTeamDiscussionCommentInOrg: ["POST /orgs/{org}/teams/{team_slug}/discussions/{discussion_number}/comments/{comment_number}/reactions", {
|
|
mediaType: {
|
|
previews: ["squirrel-girl"]
|
|
}
|
|
}],
|
|
createForTeamDiscussionInOrg: ["POST /orgs/{org}/teams/{team_slug}/discussions/{discussion_number}/reactions", {
|
|
mediaType: {
|
|
previews: ["squirrel-girl"]
|
|
}
|
|
}],
|
|
deleteForCommitComment: ["DELETE /repos/{owner}/{repo}/comments/{comment_id}/reactions/{reaction_id}", {
|
|
mediaType: {
|
|
previews: ["squirrel-girl"]
|
|
}
|
|
}],
|
|
deleteForIssue: ["DELETE /repos/{owner}/{repo}/issues/{issue_number}/reactions/{reaction_id}", {
|
|
mediaType: {
|
|
previews: ["squirrel-girl"]
|
|
}
|
|
}],
|
|
deleteForIssueComment: ["DELETE /repos/{owner}/{repo}/issues/comments/{comment_id}/reactions/{reaction_id}", {
|
|
mediaType: {
|
|
previews: ["squirrel-girl"]
|
|
}
|
|
}],
|
|
deleteForPullRequestComment: ["DELETE /repos/{owner}/{repo}/pulls/comments/{comment_id}/reactions/{reaction_id}", {
|
|
mediaType: {
|
|
previews: ["squirrel-girl"]
|
|
}
|
|
}],
|
|
deleteForTeamDiscussion: ["DELETE /orgs/{org}/teams/{team_slug}/discussions/{discussion_number}/reactions/{reaction_id}", {
|
|
mediaType: {
|
|
previews: ["squirrel-girl"]
|
|
}
|
|
}],
|
|
deleteForTeamDiscussionComment: ["DELETE /orgs/{org}/teams/{team_slug}/discussions/{discussion_number}/comments/{comment_number}/reactions/{reaction_id}", {
|
|
mediaType: {
|
|
previews: ["squirrel-girl"]
|
|
}
|
|
}],
|
|
deleteLegacy: ["DELETE /reactions/{reaction_id}", {
|
|
mediaType: {
|
|
previews: ["squirrel-girl"]
|
|
}
|
|
}, {
|
|
deprecated: "octokit.rest.reactions.deleteLegacy() is deprecated, see https://docs.github.com/rest/reference/reactions/#delete-a-reaction-legacy"
|
|
}],
|
|
listForCommitComment: ["GET /repos/{owner}/{repo}/comments/{comment_id}/reactions", {
|
|
mediaType: {
|
|
previews: ["squirrel-girl"]
|
|
}
|
|
}],
|
|
listForIssue: ["GET /repos/{owner}/{repo}/issues/{issue_number}/reactions", {
|
|
mediaType: {
|
|
previews: ["squirrel-girl"]
|
|
}
|
|
}],
|
|
listForIssueComment: ["GET /repos/{owner}/{repo}/issues/comments/{comment_id}/reactions", {
|
|
mediaType: {
|
|
previews: ["squirrel-girl"]
|
|
}
|
|
}],
|
|
listForPullRequestReviewComment: ["GET /repos/{owner}/{repo}/pulls/comments/{comment_id}/reactions", {
|
|
mediaType: {
|
|
previews: ["squirrel-girl"]
|
|
}
|
|
}],
|
|
listForTeamDiscussionCommentInOrg: ["GET /orgs/{org}/teams/{team_slug}/discussions/{discussion_number}/comments/{comment_number}/reactions", {
|
|
mediaType: {
|
|
previews: ["squirrel-girl"]
|
|
}
|
|
}],
|
|
listForTeamDiscussionInOrg: ["GET /orgs/{org}/teams/{team_slug}/discussions/{discussion_number}/reactions", {
|
|
mediaType: {
|
|
previews: ["squirrel-girl"]
|
|
}
|
|
}]
|
|
},
|
|
repos: {
|
|
acceptInvitation: ["PATCH /user/repository_invitations/{invitation_id}"],
|
|
addAppAccessRestrictions: ["POST /repos/{owner}/{repo}/branches/{branch}/protection/restrictions/apps", {}, {
|
|
mapToData: "apps"
|
|
}],
|
|
addCollaborator: ["PUT /repos/{owner}/{repo}/collaborators/{username}"],
|
|
addStatusCheckContexts: ["POST /repos/{owner}/{repo}/branches/{branch}/protection/required_status_checks/contexts", {}, {
|
|
mapToData: "contexts"
|
|
}],
|
|
addTeamAccessRestrictions: ["POST /repos/{owner}/{repo}/branches/{branch}/protection/restrictions/teams", {}, {
|
|
mapToData: "teams"
|
|
}],
|
|
addUserAccessRestrictions: ["POST /repos/{owner}/{repo}/branches/{branch}/protection/restrictions/users", {}, {
|
|
mapToData: "users"
|
|
}],
|
|
checkCollaborator: ["GET /repos/{owner}/{repo}/collaborators/{username}"],
|
|
checkVulnerabilityAlerts: ["GET /repos/{owner}/{repo}/vulnerability-alerts", {
|
|
mediaType: {
|
|
previews: ["dorian"]
|
|
}
|
|
}],
|
|
compareCommits: ["GET /repos/{owner}/{repo}/compare/{base}...{head}"],
|
|
compareCommitsWithBasehead: ["GET /repos/{owner}/{repo}/compare/{basehead}"],
|
|
createCommitComment: ["POST /repos/{owner}/{repo}/commits/{commit_sha}/comments"],
|
|
createCommitSignatureProtection: ["POST /repos/{owner}/{repo}/branches/{branch}/protection/required_signatures", {
|
|
mediaType: {
|
|
previews: ["zzzax"]
|
|
}
|
|
}],
|
|
createCommitStatus: ["POST /repos/{owner}/{repo}/statuses/{sha}"],
|
|
createDeployKey: ["POST /repos/{owner}/{repo}/keys"],
|
|
createDeployment: ["POST /repos/{owner}/{repo}/deployments"],
|
|
createDeploymentStatus: ["POST /repos/{owner}/{repo}/deployments/{deployment_id}/statuses"],
|
|
createDispatchEvent: ["POST /repos/{owner}/{repo}/dispatches"],
|
|
createForAuthenticatedUser: ["POST /user/repos"],
|
|
createFork: ["POST /repos/{owner}/{repo}/forks"],
|
|
createInOrg: ["POST /orgs/{org}/repos"],
|
|
createOrUpdateEnvironment: ["PUT /repos/{owner}/{repo}/environments/{environment_name}"],
|
|
createOrUpdateFileContents: ["PUT /repos/{owner}/{repo}/contents/{path}"],
|
|
createPagesSite: ["POST /repos/{owner}/{repo}/pages", {
|
|
mediaType: {
|
|
previews: ["switcheroo"]
|
|
}
|
|
}],
|
|
createRelease: ["POST /repos/{owner}/{repo}/releases"],
|
|
createUsingTemplate: ["POST /repos/{template_owner}/{template_repo}/generate", {
|
|
mediaType: {
|
|
previews: ["baptiste"]
|
|
}
|
|
}],
|
|
createWebhook: ["POST /repos/{owner}/{repo}/hooks"],
|
|
declineInvitation: ["DELETE /user/repository_invitations/{invitation_id}"],
|
|
delete: ["DELETE /repos/{owner}/{repo}"],
|
|
deleteAccessRestrictions: ["DELETE /repos/{owner}/{repo}/branches/{branch}/protection/restrictions"],
|
|
deleteAdminBranchProtection: ["DELETE /repos/{owner}/{repo}/branches/{branch}/protection/enforce_admins"],
|
|
deleteAnEnvironment: ["DELETE /repos/{owner}/{repo}/environments/{environment_name}"],
|
|
deleteBranchProtection: ["DELETE /repos/{owner}/{repo}/branches/{branch}/protection"],
|
|
deleteCommitComment: ["DELETE /repos/{owner}/{repo}/comments/{comment_id}"],
|
|
deleteCommitSignatureProtection: ["DELETE /repos/{owner}/{repo}/branches/{branch}/protection/required_signatures", {
|
|
mediaType: {
|
|
previews: ["zzzax"]
|
|
}
|
|
}],
|
|
deleteDeployKey: ["DELETE /repos/{owner}/{repo}/keys/{key_id}"],
|
|
deleteDeployment: ["DELETE /repos/{owner}/{repo}/deployments/{deployment_id}"],
|
|
deleteFile: ["DELETE /repos/{owner}/{repo}/contents/{path}"],
|
|
deleteInvitation: ["DELETE /repos/{owner}/{repo}/invitations/{invitation_id}"],
|
|
deletePagesSite: ["DELETE /repos/{owner}/{repo}/pages", {
|
|
mediaType: {
|
|
previews: ["switcheroo"]
|
|
}
|
|
}],
|
|
deletePullRequestReviewProtection: ["DELETE /repos/{owner}/{repo}/branches/{branch}/protection/required_pull_request_reviews"],
|
|
deleteRelease: ["DELETE /repos/{owner}/{repo}/releases/{release_id}"],
|
|
deleteReleaseAsset: ["DELETE /repos/{owner}/{repo}/releases/assets/{asset_id}"],
|
|
deleteWebhook: ["DELETE /repos/{owner}/{repo}/hooks/{hook_id}"],
|
|
disableAutomatedSecurityFixes: ["DELETE /repos/{owner}/{repo}/automated-security-fixes", {
|
|
mediaType: {
|
|
previews: ["london"]
|
|
}
|
|
}],
|
|
disableVulnerabilityAlerts: ["DELETE /repos/{owner}/{repo}/vulnerability-alerts", {
|
|
mediaType: {
|
|
previews: ["dorian"]
|
|
}
|
|
}],
|
|
downloadArchive: ["GET /repos/{owner}/{repo}/zipball/{ref}", {}, {
|
|
renamed: ["repos", "downloadZipballArchive"]
|
|
}],
|
|
downloadTarballArchive: ["GET /repos/{owner}/{repo}/tarball/{ref}"],
|
|
downloadZipballArchive: ["GET /repos/{owner}/{repo}/zipball/{ref}"],
|
|
enableAutomatedSecurityFixes: ["PUT /repos/{owner}/{repo}/automated-security-fixes", {
|
|
mediaType: {
|
|
previews: ["london"]
|
|
}
|
|
}],
|
|
enableVulnerabilityAlerts: ["PUT /repos/{owner}/{repo}/vulnerability-alerts", {
|
|
mediaType: {
|
|
previews: ["dorian"]
|
|
}
|
|
}],
|
|
get: ["GET /repos/{owner}/{repo}"],
|
|
getAccessRestrictions: ["GET /repos/{owner}/{repo}/branches/{branch}/protection/restrictions"],
|
|
getAdminBranchProtection: ["GET /repos/{owner}/{repo}/branches/{branch}/protection/enforce_admins"],
|
|
getAllEnvironments: ["GET /repos/{owner}/{repo}/environments"],
|
|
getAllStatusCheckContexts: ["GET /repos/{owner}/{repo}/branches/{branch}/protection/required_status_checks/contexts"],
|
|
getAllTopics: ["GET /repos/{owner}/{repo}/topics", {
|
|
mediaType: {
|
|
previews: ["mercy"]
|
|
}
|
|
}],
|
|
getAppsWithAccessToProtectedBranch: ["GET /repos/{owner}/{repo}/branches/{branch}/protection/restrictions/apps"],
|
|
getBranch: ["GET /repos/{owner}/{repo}/branches/{branch}"],
|
|
getBranchProtection: ["GET /repos/{owner}/{repo}/branches/{branch}/protection"],
|
|
getClones: ["GET /repos/{owner}/{repo}/traffic/clones"],
|
|
getCodeFrequencyStats: ["GET /repos/{owner}/{repo}/stats/code_frequency"],
|
|
getCollaboratorPermissionLevel: ["GET /repos/{owner}/{repo}/collaborators/{username}/permission"],
|
|
getCombinedStatusForRef: ["GET /repos/{owner}/{repo}/commits/{ref}/status"],
|
|
getCommit: ["GET /repos/{owner}/{repo}/commits/{ref}"],
|
|
getCommitActivityStats: ["GET /repos/{owner}/{repo}/stats/commit_activity"],
|
|
getCommitComment: ["GET /repos/{owner}/{repo}/comments/{comment_id}"],
|
|
getCommitSignatureProtection: ["GET /repos/{owner}/{repo}/branches/{branch}/protection/required_signatures", {
|
|
mediaType: {
|
|
previews: ["zzzax"]
|
|
}
|
|
}],
|
|
getCommunityProfileMetrics: ["GET /repos/{owner}/{repo}/community/profile"],
|
|
getContent: ["GET /repos/{owner}/{repo}/contents/{path}"],
|
|
getContributorsStats: ["GET /repos/{owner}/{repo}/stats/contributors"],
|
|
getDeployKey: ["GET /repos/{owner}/{repo}/keys/{key_id}"],
|
|
getDeployment: ["GET /repos/{owner}/{repo}/deployments/{deployment_id}"],
|
|
getDeploymentStatus: ["GET /repos/{owner}/{repo}/deployments/{deployment_id}/statuses/{status_id}"],
|
|
getEnvironment: ["GET /repos/{owner}/{repo}/environments/{environment_name}"],
|
|
getLatestPagesBuild: ["GET /repos/{owner}/{repo}/pages/builds/latest"],
|
|
getLatestRelease: ["GET /repos/{owner}/{repo}/releases/latest"],
|
|
getPages: ["GET /repos/{owner}/{repo}/pages"],
|
|
getPagesBuild: ["GET /repos/{owner}/{repo}/pages/builds/{build_id}"],
|
|
getPagesHealthCheck: ["GET /repos/{owner}/{repo}/pages/health"],
|
|
getParticipationStats: ["GET /repos/{owner}/{repo}/stats/participation"],
|
|
getPullRequestReviewProtection: ["GET /repos/{owner}/{repo}/branches/{branch}/protection/required_pull_request_reviews"],
|
|
getPunchCardStats: ["GET /repos/{owner}/{repo}/stats/punch_card"],
|
|
getReadme: ["GET /repos/{owner}/{repo}/readme"],
|
|
getReadmeInDirectory: ["GET /repos/{owner}/{repo}/readme/{dir}"],
|
|
getRelease: ["GET /repos/{owner}/{repo}/releases/{release_id}"],
|
|
getReleaseAsset: ["GET /repos/{owner}/{repo}/releases/assets/{asset_id}"],
|
|
getReleaseByTag: ["GET /repos/{owner}/{repo}/releases/tags/{tag}"],
|
|
getStatusChecksProtection: ["GET /repos/{owner}/{repo}/branches/{branch}/protection/required_status_checks"],
|
|
getTeamsWithAccessToProtectedBranch: ["GET /repos/{owner}/{repo}/branches/{branch}/protection/restrictions/teams"],
|
|
getTopPaths: ["GET /repos/{owner}/{repo}/traffic/popular/paths"],
|
|
getTopReferrers: ["GET /repos/{owner}/{repo}/traffic/popular/referrers"],
|
|
getUsersWithAccessToProtectedBranch: ["GET /repos/{owner}/{repo}/branches/{branch}/protection/restrictions/users"],
|
|
getViews: ["GET /repos/{owner}/{repo}/traffic/views"],
|
|
getWebhook: ["GET /repos/{owner}/{repo}/hooks/{hook_id}"],
|
|
getWebhookConfigForRepo: ["GET /repos/{owner}/{repo}/hooks/{hook_id}/config"],
|
|
listBranches: ["GET /repos/{owner}/{repo}/branches"],
|
|
listBranchesForHeadCommit: ["GET /repos/{owner}/{repo}/commits/{commit_sha}/branches-where-head", {
|
|
mediaType: {
|
|
previews: ["groot"]
|
|
}
|
|
}],
|
|
listCollaborators: ["GET /repos/{owner}/{repo}/collaborators"],
|
|
listCommentsForCommit: ["GET /repos/{owner}/{repo}/commits/{commit_sha}/comments"],
|
|
listCommitCommentsForRepo: ["GET /repos/{owner}/{repo}/comments"],
|
|
listCommitStatusesForRef: ["GET /repos/{owner}/{repo}/commits/{ref}/statuses"],
|
|
listCommits: ["GET /repos/{owner}/{repo}/commits"],
|
|
listContributors: ["GET /repos/{owner}/{repo}/contributors"],
|
|
listDeployKeys: ["GET /repos/{owner}/{repo}/keys"],
|
|
listDeploymentStatuses: ["GET /repos/{owner}/{repo}/deployments/{deployment_id}/statuses"],
|
|
listDeployments: ["GET /repos/{owner}/{repo}/deployments"],
|
|
listForAuthenticatedUser: ["GET /user/repos"],
|
|
listForOrg: ["GET /orgs/{org}/repos"],
|
|
listForUser: ["GET /users/{username}/repos"],
|
|
listForks: ["GET /repos/{owner}/{repo}/forks"],
|
|
listInvitations: ["GET /repos/{owner}/{repo}/invitations"],
|
|
listInvitationsForAuthenticatedUser: ["GET /user/repository_invitations"],
|
|
listLanguages: ["GET /repos/{owner}/{repo}/languages"],
|
|
listPagesBuilds: ["GET /repos/{owner}/{repo}/pages/builds"],
|
|
listPublic: ["GET /repositories"],
|
|
listPullRequestsAssociatedWithCommit: ["GET /repos/{owner}/{repo}/commits/{commit_sha}/pulls", {
|
|
mediaType: {
|
|
previews: ["groot"]
|
|
}
|
|
}],
|
|
listReleaseAssets: ["GET /repos/{owner}/{repo}/releases/{release_id}/assets"],
|
|
listReleases: ["GET /repos/{owner}/{repo}/releases"],
|
|
listTags: ["GET /repos/{owner}/{repo}/tags"],
|
|
listTeams: ["GET /repos/{owner}/{repo}/teams"],
|
|
listWebhooks: ["GET /repos/{owner}/{repo}/hooks"],
|
|
merge: ["POST /repos/{owner}/{repo}/merges"],
|
|
pingWebhook: ["POST /repos/{owner}/{repo}/hooks/{hook_id}/pings"],
|
|
removeAppAccessRestrictions: ["DELETE /repos/{owner}/{repo}/branches/{branch}/protection/restrictions/apps", {}, {
|
|
mapToData: "apps"
|
|
}],
|
|
removeCollaborator: ["DELETE /repos/{owner}/{repo}/collaborators/{username}"],
|
|
removeStatusCheckContexts: ["DELETE /repos/{owner}/{repo}/branches/{branch}/protection/required_status_checks/contexts", {}, {
|
|
mapToData: "contexts"
|
|
}],
|
|
removeStatusCheckProtection: ["DELETE /repos/{owner}/{repo}/branches/{branch}/protection/required_status_checks"],
|
|
removeTeamAccessRestrictions: ["DELETE /repos/{owner}/{repo}/branches/{branch}/protection/restrictions/teams", {}, {
|
|
mapToData: "teams"
|
|
}],
|
|
removeUserAccessRestrictions: ["DELETE /repos/{owner}/{repo}/branches/{branch}/protection/restrictions/users", {}, {
|
|
mapToData: "users"
|
|
}],
|
|
renameBranch: ["POST /repos/{owner}/{repo}/branches/{branch}/rename"],
|
|
replaceAllTopics: ["PUT /repos/{owner}/{repo}/topics", {
|
|
mediaType: {
|
|
previews: ["mercy"]
|
|
}
|
|
}],
|
|
requestPagesBuild: ["POST /repos/{owner}/{repo}/pages/builds"],
|
|
setAdminBranchProtection: ["POST /repos/{owner}/{repo}/branches/{branch}/protection/enforce_admins"],
|
|
setAppAccessRestrictions: ["PUT /repos/{owner}/{repo}/branches/{branch}/protection/restrictions/apps", {}, {
|
|
mapToData: "apps"
|
|
}],
|
|
setStatusCheckContexts: ["PUT /repos/{owner}/{repo}/branches/{branch}/protection/required_status_checks/contexts", {}, {
|
|
mapToData: "contexts"
|
|
}],
|
|
setTeamAccessRestrictions: ["PUT /repos/{owner}/{repo}/branches/{branch}/protection/restrictions/teams", {}, {
|
|
mapToData: "teams"
|
|
}],
|
|
setUserAccessRestrictions: ["PUT /repos/{owner}/{repo}/branches/{branch}/protection/restrictions/users", {}, {
|
|
mapToData: "users"
|
|
}],
|
|
testPushWebhook: ["POST /repos/{owner}/{repo}/hooks/{hook_id}/tests"],
|
|
transfer: ["POST /repos/{owner}/{repo}/transfer"],
|
|
update: ["PATCH /repos/{owner}/{repo}"],
|
|
updateBranchProtection: ["PUT /repos/{owner}/{repo}/branches/{branch}/protection"],
|
|
updateCommitComment: ["PATCH /repos/{owner}/{repo}/comments/{comment_id}"],
|
|
updateInformationAboutPagesSite: ["PUT /repos/{owner}/{repo}/pages"],
|
|
updateInvitation: ["PATCH /repos/{owner}/{repo}/invitations/{invitation_id}"],
|
|
updatePullRequestReviewProtection: ["PATCH /repos/{owner}/{repo}/branches/{branch}/protection/required_pull_request_reviews"],
|
|
updateRelease: ["PATCH /repos/{owner}/{repo}/releases/{release_id}"],
|
|
updateReleaseAsset: ["PATCH /repos/{owner}/{repo}/releases/assets/{asset_id}"],
|
|
updateStatusCheckPotection: ["PATCH /repos/{owner}/{repo}/branches/{branch}/protection/required_status_checks", {}, {
|
|
renamed: ["repos", "updateStatusCheckProtection"]
|
|
}],
|
|
updateStatusCheckProtection: ["PATCH /repos/{owner}/{repo}/branches/{branch}/protection/required_status_checks"],
|
|
updateWebhook: ["PATCH /repos/{owner}/{repo}/hooks/{hook_id}"],
|
|
updateWebhookConfigForRepo: ["PATCH /repos/{owner}/{repo}/hooks/{hook_id}/config"],
|
|
uploadReleaseAsset: ["POST /repos/{owner}/{repo}/releases/{release_id}/assets{?name,label}", {
|
|
baseUrl: "https://uploads.github.com"
|
|
}]
|
|
},
|
|
search: {
|
|
code: ["GET /search/code"],
|
|
commits: ["GET /search/commits", {
|
|
mediaType: {
|
|
previews: ["cloak"]
|
|
}
|
|
}],
|
|
issuesAndPullRequests: ["GET /search/issues"],
|
|
labels: ["GET /search/labels"],
|
|
repos: ["GET /search/repositories"],
|
|
topics: ["GET /search/topics", {
|
|
mediaType: {
|
|
previews: ["mercy"]
|
|
}
|
|
}],
|
|
users: ["GET /search/users"]
|
|
},
|
|
secretScanning: {
|
|
getAlert: ["GET /repos/{owner}/{repo}/secret-scanning/alerts/{alert_number}"],
|
|
listAlertsForRepo: ["GET /repos/{owner}/{repo}/secret-scanning/alerts"],
|
|
updateAlert: ["PATCH /repos/{owner}/{repo}/secret-scanning/alerts/{alert_number}"]
|
|
},
|
|
teams: {
|
|
addOrUpdateMembershipForUserInOrg: ["PUT /orgs/{org}/teams/{team_slug}/memberships/{username}"],
|
|
addOrUpdateProjectPermissionsInOrg: ["PUT /orgs/{org}/teams/{team_slug}/projects/{project_id}", {
|
|
mediaType: {
|
|
previews: ["inertia"]
|
|
}
|
|
}],
|
|
addOrUpdateRepoPermissionsInOrg: ["PUT /orgs/{org}/teams/{team_slug}/repos/{owner}/{repo}"],
|
|
checkPermissionsForProjectInOrg: ["GET /orgs/{org}/teams/{team_slug}/projects/{project_id}", {
|
|
mediaType: {
|
|
previews: ["inertia"]
|
|
}
|
|
}],
|
|
checkPermissionsForRepoInOrg: ["GET /orgs/{org}/teams/{team_slug}/repos/{owner}/{repo}"],
|
|
create: ["POST /orgs/{org}/teams"],
|
|
createDiscussionCommentInOrg: ["POST /orgs/{org}/teams/{team_slug}/discussions/{discussion_number}/comments"],
|
|
createDiscussionInOrg: ["POST /orgs/{org}/teams/{team_slug}/discussions"],
|
|
deleteDiscussionCommentInOrg: ["DELETE /orgs/{org}/teams/{team_slug}/discussions/{discussion_number}/comments/{comment_number}"],
|
|
deleteDiscussionInOrg: ["DELETE /orgs/{org}/teams/{team_slug}/discussions/{discussion_number}"],
|
|
deleteInOrg: ["DELETE /orgs/{org}/teams/{team_slug}"],
|
|
getByName: ["GET /orgs/{org}/teams/{team_slug}"],
|
|
getDiscussionCommentInOrg: ["GET /orgs/{org}/teams/{team_slug}/discussions/{discussion_number}/comments/{comment_number}"],
|
|
getDiscussionInOrg: ["GET /orgs/{org}/teams/{team_slug}/discussions/{discussion_number}"],
|
|
getMembershipForUserInOrg: ["GET /orgs/{org}/teams/{team_slug}/memberships/{username}"],
|
|
list: ["GET /orgs/{org}/teams"],
|
|
listChildInOrg: ["GET /orgs/{org}/teams/{team_slug}/teams"],
|
|
listDiscussionCommentsInOrg: ["GET /orgs/{org}/teams/{team_slug}/discussions/{discussion_number}/comments"],
|
|
listDiscussionsInOrg: ["GET /orgs/{org}/teams/{team_slug}/discussions"],
|
|
listForAuthenticatedUser: ["GET /user/teams"],
|
|
listMembersInOrg: ["GET /orgs/{org}/teams/{team_slug}/members"],
|
|
listPendingInvitationsInOrg: ["GET /orgs/{org}/teams/{team_slug}/invitations"],
|
|
listProjectsInOrg: ["GET /orgs/{org}/teams/{team_slug}/projects", {
|
|
mediaType: {
|
|
previews: ["inertia"]
|
|
}
|
|
}],
|
|
listReposInOrg: ["GET /orgs/{org}/teams/{team_slug}/repos"],
|
|
removeMembershipForUserInOrg: ["DELETE /orgs/{org}/teams/{team_slug}/memberships/{username}"],
|
|
removeProjectInOrg: ["DELETE /orgs/{org}/teams/{team_slug}/projects/{project_id}"],
|
|
removeRepoInOrg: ["DELETE /orgs/{org}/teams/{team_slug}/repos/{owner}/{repo}"],
|
|
updateDiscussionCommentInOrg: ["PATCH /orgs/{org}/teams/{team_slug}/discussions/{discussion_number}/comments/{comment_number}"],
|
|
updateDiscussionInOrg: ["PATCH /orgs/{org}/teams/{team_slug}/discussions/{discussion_number}"],
|
|
updateInOrg: ["PATCH /orgs/{org}/teams/{team_slug}"]
|
|
},
|
|
users: {
|
|
addEmailForAuthenticated: ["POST /user/emails"],
|
|
block: ["PUT /user/blocks/{username}"],
|
|
checkBlocked: ["GET /user/blocks/{username}"],
|
|
checkFollowingForUser: ["GET /users/{username}/following/{target_user}"],
|
|
checkPersonIsFollowedByAuthenticated: ["GET /user/following/{username}"],
|
|
createGpgKeyForAuthenticated: ["POST /user/gpg_keys"],
|
|
createPublicSshKeyForAuthenticated: ["POST /user/keys"],
|
|
deleteEmailForAuthenticated: ["DELETE /user/emails"],
|
|
deleteGpgKeyForAuthenticated: ["DELETE /user/gpg_keys/{gpg_key_id}"],
|
|
deletePublicSshKeyForAuthenticated: ["DELETE /user/keys/{key_id}"],
|
|
follow: ["PUT /user/following/{username}"],
|
|
getAuthenticated: ["GET /user"],
|
|
getByUsername: ["GET /users/{username}"],
|
|
getContextForUser: ["GET /users/{username}/hovercard"],
|
|
getGpgKeyForAuthenticated: ["GET /user/gpg_keys/{gpg_key_id}"],
|
|
getPublicSshKeyForAuthenticated: ["GET /user/keys/{key_id}"],
|
|
list: ["GET /users"],
|
|
listBlockedByAuthenticated: ["GET /user/blocks"],
|
|
listEmailsForAuthenticated: ["GET /user/emails"],
|
|
listFollowedByAuthenticated: ["GET /user/following"],
|
|
listFollowersForAuthenticatedUser: ["GET /user/followers"],
|
|
listFollowersForUser: ["GET /users/{username}/followers"],
|
|
listFollowingForUser: ["GET /users/{username}/following"],
|
|
listGpgKeysForAuthenticated: ["GET /user/gpg_keys"],
|
|
listGpgKeysForUser: ["GET /users/{username}/gpg_keys"],
|
|
listPublicEmailsForAuthenticated: ["GET /user/public_emails"],
|
|
listPublicKeysForUser: ["GET /users/{username}/keys"],
|
|
listPublicSshKeysForAuthenticated: ["GET /user/keys"],
|
|
setPrimaryEmailVisibilityForAuthenticated: ["PATCH /user/email/visibility"],
|
|
unblock: ["DELETE /user/blocks/{username}"],
|
|
unfollow: ["DELETE /user/following/{username}"],
|
|
updateAuthenticated: ["PATCH /user"]
|
|
}
|
|
};
|
|
|
|
const VERSION = "5.3.0";
|
|
|
|
function endpointsToMethods(octokit, endpointsMap) {
|
|
const newMethods = {};
|
|
|
|
for (const [scope, endpoints] of Object.entries(endpointsMap)) {
|
|
for (const [methodName, endpoint] of Object.entries(endpoints)) {
|
|
const [route, defaults, decorations] = endpoint;
|
|
const [method, url] = route.split(/ /);
|
|
const endpointDefaults = Object.assign({
|
|
method,
|
|
url
|
|
}, defaults);
|
|
|
|
if (!newMethods[scope]) {
|
|
newMethods[scope] = {};
|
|
}
|
|
|
|
const scopeMethods = newMethods[scope];
|
|
|
|
if (decorations) {
|
|
scopeMethods[methodName] = decorate(octokit, scope, methodName, endpointDefaults, decorations);
|
|
continue;
|
|
}
|
|
|
|
scopeMethods[methodName] = octokit.request.defaults(endpointDefaults);
|
|
}
|
|
}
|
|
|
|
return newMethods;
|
|
}
|
|
|
|
function decorate(octokit, scope, methodName, defaults, decorations) {
|
|
const requestWithDefaults = octokit.request.defaults(defaults);
|
|
/* istanbul ignore next */
|
|
|
|
function withDecorations(...args) {
|
|
// @ts-ignore https://github.com/microsoft/TypeScript/issues/25488
|
|
let options = requestWithDefaults.endpoint.merge(...args); // There are currently no other decorations than `.mapToData`
|
|
|
|
if (decorations.mapToData) {
|
|
options = Object.assign({}, options, {
|
|
data: options[decorations.mapToData],
|
|
[decorations.mapToData]: undefined
|
|
});
|
|
return requestWithDefaults(options);
|
|
}
|
|
|
|
if (decorations.renamed) {
|
|
const [newScope, newMethodName] = decorations.renamed;
|
|
octokit.log.warn(`octokit.${scope}.${methodName}() has been renamed to octokit.${newScope}.${newMethodName}()`);
|
|
}
|
|
|
|
if (decorations.deprecated) {
|
|
octokit.log.warn(decorations.deprecated);
|
|
}
|
|
|
|
if (decorations.renamedParameters) {
|
|
// @ts-ignore https://github.com/microsoft/TypeScript/issues/25488
|
|
const options = requestWithDefaults.endpoint.merge(...args);
|
|
|
|
for (const [name, alias] of Object.entries(decorations.renamedParameters)) {
|
|
if (name in options) {
|
|
octokit.log.warn(`"${name}" parameter is deprecated for "octokit.${scope}.${methodName}()". Use "${alias}" instead`);
|
|
|
|
if (!(alias in options)) {
|
|
options[alias] = options[name];
|
|
}
|
|
|
|
delete options[name];
|
|
}
|
|
}
|
|
|
|
return requestWithDefaults(options);
|
|
} // @ts-ignore https://github.com/microsoft/TypeScript/issues/25488
|
|
|
|
|
|
return requestWithDefaults(...args);
|
|
}
|
|
|
|
return Object.assign(withDecorations, requestWithDefaults);
|
|
}
|
|
|
|
function restEndpointMethods(octokit) {
|
|
const api = endpointsToMethods(octokit, Endpoints);
|
|
return {
|
|
rest: api
|
|
};
|
|
}
|
|
restEndpointMethods.VERSION = VERSION;
|
|
function legacyRestEndpointMethods(octokit) {
|
|
const api = endpointsToMethods(octokit, Endpoints);
|
|
return _objectSpread2(_objectSpread2({}, api), {}, {
|
|
rest: api
|
|
});
|
|
}
|
|
legacyRestEndpointMethods.VERSION = VERSION;
|
|
|
|
exports.legacyRestEndpointMethods = legacyRestEndpointMethods;
|
|
exports.restEndpointMethods = restEndpointMethods;
|
|
//# sourceMappingURL=index.js.map
|
|
|
|
|
|
/***/ }),
|
|
|
|
/***/ 537:
|
|
/***/ ((__unused_webpack_module, exports, __nccwpck_require__) => {
|
|
|
|
"use strict";
|
|
|
|
|
|
Object.defineProperty(exports, "__esModule", ({ value: true }));
|
|
|
|
function _interopDefault (ex) { return (ex && (typeof ex === 'object') && 'default' in ex) ? ex['default'] : ex; }
|
|
|
|
var deprecation = __nccwpck_require__(8932);
|
|
var once = _interopDefault(__nccwpck_require__(1223));
|
|
|
|
const logOnce = once(deprecation => console.warn(deprecation));
|
|
/**
|
|
* Error with extra properties to help with debugging
|
|
*/
|
|
|
|
class RequestError extends Error {
|
|
constructor(message, statusCode, options) {
|
|
super(message); // Maintains proper stack trace (only available on V8)
|
|
|
|
/* istanbul ignore next */
|
|
|
|
if (Error.captureStackTrace) {
|
|
Error.captureStackTrace(this, this.constructor);
|
|
}
|
|
|
|
this.name = "HttpError";
|
|
this.status = statusCode;
|
|
Object.defineProperty(this, "code", {
|
|
get() {
|
|
logOnce(new deprecation.Deprecation("[@octokit/request-error] `error.code` is deprecated, use `error.status`."));
|
|
return statusCode;
|
|
}
|
|
|
|
});
|
|
this.headers = options.headers || {}; // redact request credentials without mutating original request options
|
|
|
|
const requestCopy = Object.assign({}, options.request);
|
|
|
|
if (options.request.headers.authorization) {
|
|
requestCopy.headers = Object.assign({}, options.request.headers, {
|
|
authorization: options.request.headers.authorization.replace(/ .*$/, " [REDACTED]")
|
|
});
|
|
}
|
|
|
|
requestCopy.url = requestCopy.url // client_id & client_secret can be passed as URL query parameters to increase rate limit
|
|
// see https://developer.github.com/v3/#increasing-the-unauthenticated-rate-limit-for-oauth-applications
|
|
.replace(/\bclient_secret=\w+/g, "client_secret=[REDACTED]") // OAuth tokens can be passed as URL query parameters, although it is not recommended
|
|
// see https://developer.github.com/v3/#oauth2-token-sent-in-a-header
|
|
.replace(/\baccess_token=\w+/g, "access_token=[REDACTED]");
|
|
this.request = requestCopy;
|
|
}
|
|
|
|
}
|
|
|
|
exports.RequestError = RequestError;
|
|
//# sourceMappingURL=index.js.map
|
|
|
|
|
|
/***/ }),
|
|
|
|
/***/ 6234:
|
|
/***/ ((__unused_webpack_module, exports, __nccwpck_require__) => {
|
|
|
|
"use strict";
|
|
|
|
|
|
Object.defineProperty(exports, "__esModule", ({ value: true }));
|
|
|
|
function _interopDefault (ex) { return (ex && (typeof ex === 'object') && 'default' in ex) ? ex['default'] : ex; }
|
|
|
|
var endpoint = __nccwpck_require__(9440);
|
|
var universalUserAgent = __nccwpck_require__(5030);
|
|
var isPlainObject = __nccwpck_require__(3287);
|
|
var nodeFetch = _interopDefault(__nccwpck_require__(467));
|
|
var requestError = __nccwpck_require__(537);
|
|
|
|
const VERSION = "5.4.15";
|
|
|
|
function getBufferResponse(response) {
|
|
return response.arrayBuffer();
|
|
}
|
|
|
|
function fetchWrapper(requestOptions) {
|
|
if (isPlainObject.isPlainObject(requestOptions.body) || Array.isArray(requestOptions.body)) {
|
|
requestOptions.body = JSON.stringify(requestOptions.body);
|
|
}
|
|
|
|
let headers = {};
|
|
let status;
|
|
let url;
|
|
const fetch = requestOptions.request && requestOptions.request.fetch || nodeFetch;
|
|
return fetch(requestOptions.url, Object.assign({
|
|
method: requestOptions.method,
|
|
body: requestOptions.body,
|
|
headers: requestOptions.headers,
|
|
redirect: requestOptions.redirect
|
|
}, // `requestOptions.request.agent` type is incompatible
|
|
// see https://github.com/octokit/types.ts/pull/264
|
|
requestOptions.request)).then(response => {
|
|
url = response.url;
|
|
status = response.status;
|
|
|
|
for (const keyAndValue of response.headers) {
|
|
headers[keyAndValue[0]] = keyAndValue[1];
|
|
}
|
|
|
|
if (status === 204 || status === 205) {
|
|
return;
|
|
} // GitHub API returns 200 for HEAD requests
|
|
|
|
|
|
if (requestOptions.method === "HEAD") {
|
|
if (status < 400) {
|
|
return;
|
|
}
|
|
|
|
throw new requestError.RequestError(response.statusText, status, {
|
|
headers,
|
|
request: requestOptions
|
|
});
|
|
}
|
|
|
|
if (status === 304) {
|
|
throw new requestError.RequestError("Not modified", status, {
|
|
headers,
|
|
request: requestOptions
|
|
});
|
|
}
|
|
|
|
if (status >= 400) {
|
|
return response.text().then(message => {
|
|
const error = new requestError.RequestError(message, status, {
|
|
headers,
|
|
request: requestOptions
|
|
});
|
|
|
|
try {
|
|
let responseBody = JSON.parse(error.message);
|
|
Object.assign(error, responseBody);
|
|
let errors = responseBody.errors; // Assumption `errors` would always be in Array format
|
|
|
|
error.message = error.message + ": " + errors.map(JSON.stringify).join(", ");
|
|
} catch (e) {// ignore, see octokit/rest.js#684
|
|
}
|
|
|
|
throw error;
|
|
});
|
|
}
|
|
|
|
const contentType = response.headers.get("content-type");
|
|
|
|
if (/application\/json/.test(contentType)) {
|
|
return response.json();
|
|
}
|
|
|
|
if (!contentType || /^text\/|charset=utf-8$/.test(contentType)) {
|
|
return response.text();
|
|
}
|
|
|
|
return getBufferResponse(response);
|
|
}).then(data => {
|
|
return {
|
|
status,
|
|
url,
|
|
headers,
|
|
data
|
|
};
|
|
}).catch(error => {
|
|
if (error instanceof requestError.RequestError) {
|
|
throw error;
|
|
}
|
|
|
|
throw new requestError.RequestError(error.message, 500, {
|
|
headers,
|
|
request: requestOptions
|
|
});
|
|
});
|
|
}
|
|
|
|
function withDefaults(oldEndpoint, newDefaults) {
|
|
const endpoint = oldEndpoint.defaults(newDefaults);
|
|
|
|
const newApi = function (route, parameters) {
|
|
const endpointOptions = endpoint.merge(route, parameters);
|
|
|
|
if (!endpointOptions.request || !endpointOptions.request.hook) {
|
|
return fetchWrapper(endpoint.parse(endpointOptions));
|
|
}
|
|
|
|
const request = (route, parameters) => {
|
|
return fetchWrapper(endpoint.parse(endpoint.merge(route, parameters)));
|
|
};
|
|
|
|
Object.assign(request, {
|
|
endpoint,
|
|
defaults: withDefaults.bind(null, endpoint)
|
|
});
|
|
return endpointOptions.request.hook(request, endpointOptions);
|
|
};
|
|
|
|
return Object.assign(newApi, {
|
|
endpoint,
|
|
defaults: withDefaults.bind(null, endpoint)
|
|
});
|
|
}
|
|
|
|
const request = withDefaults(endpoint.endpoint, {
|
|
headers: {
|
|
"user-agent": `octokit-request.js/${VERSION} ${universalUserAgent.getUserAgent()}`
|
|
}
|
|
});
|
|
|
|
exports.request = request;
|
|
//# sourceMappingURL=index.js.map
|
|
|
|
|
|
/***/ }),
|
|
|
|
/***/ 3682:
|
|
/***/ ((module, __unused_webpack_exports, __nccwpck_require__) => {
|
|
|
|
var register = __nccwpck_require__(4670)
|
|
var addHook = __nccwpck_require__(5549)
|
|
var removeHook = __nccwpck_require__(6819)
|
|
|
|
// bind with array of arguments: https://stackoverflow.com/a/21792913
|
|
var bind = Function.bind
|
|
var bindable = bind.bind(bind)
|
|
|
|
function bindApi (hook, state, name) {
|
|
var removeHookRef = bindable(removeHook, null).apply(null, name ? [state, name] : [state])
|
|
hook.api = { remove: removeHookRef }
|
|
hook.remove = removeHookRef
|
|
|
|
;['before', 'error', 'after', 'wrap'].forEach(function (kind) {
|
|
var args = name ? [state, kind, name] : [state, kind]
|
|
hook[kind] = hook.api[kind] = bindable(addHook, null).apply(null, args)
|
|
})
|
|
}
|
|
|
|
function HookSingular () {
|
|
var singularHookName = 'h'
|
|
var singularHookState = {
|
|
registry: {}
|
|
}
|
|
var singularHook = register.bind(null, singularHookState, singularHookName)
|
|
bindApi(singularHook, singularHookState, singularHookName)
|
|
return singularHook
|
|
}
|
|
|
|
function HookCollection () {
|
|
var state = {
|
|
registry: {}
|
|
}
|
|
|
|
var hook = register.bind(null, state)
|
|
bindApi(hook, state)
|
|
|
|
return hook
|
|
}
|
|
|
|
var collectionHookDeprecationMessageDisplayed = false
|
|
function Hook () {
|
|
if (!collectionHookDeprecationMessageDisplayed) {
|
|
console.warn('[before-after-hook]: "Hook()" repurposing warning, use "Hook.Collection()". Read more: https://git.io/upgrade-before-after-hook-to-1.4')
|
|
collectionHookDeprecationMessageDisplayed = true
|
|
}
|
|
return HookCollection()
|
|
}
|
|
|
|
Hook.Singular = HookSingular.bind()
|
|
Hook.Collection = HookCollection.bind()
|
|
|
|
module.exports = Hook
|
|
// expose constructors as a named property for TypeScript
|
|
module.exports.Hook = Hook
|
|
module.exports.Singular = Hook.Singular
|
|
module.exports.Collection = Hook.Collection
|
|
|
|
|
|
/***/ }),
|
|
|
|
/***/ 5549:
|
|
/***/ ((module) => {
|
|
|
|
module.exports = addHook;
|
|
|
|
function addHook(state, kind, name, hook) {
|
|
var orig = hook;
|
|
if (!state.registry[name]) {
|
|
state.registry[name] = [];
|
|
}
|
|
|
|
if (kind === "before") {
|
|
hook = function (method, options) {
|
|
return Promise.resolve()
|
|
.then(orig.bind(null, options))
|
|
.then(method.bind(null, options));
|
|
};
|
|
}
|
|
|
|
if (kind === "after") {
|
|
hook = function (method, options) {
|
|
var result;
|
|
return Promise.resolve()
|
|
.then(method.bind(null, options))
|
|
.then(function (result_) {
|
|
result = result_;
|
|
return orig(result, options);
|
|
})
|
|
.then(function () {
|
|
return result;
|
|
});
|
|
};
|
|
}
|
|
|
|
if (kind === "error") {
|
|
hook = function (method, options) {
|
|
return Promise.resolve()
|
|
.then(method.bind(null, options))
|
|
.catch(function (error) {
|
|
return orig(error, options);
|
|
});
|
|
};
|
|
}
|
|
|
|
state.registry[name].push({
|
|
hook: hook,
|
|
orig: orig,
|
|
});
|
|
}
|
|
|
|
|
|
/***/ }),
|
|
|
|
/***/ 4670:
|
|
/***/ ((module) => {
|
|
|
|
module.exports = register;
|
|
|
|
function register(state, name, method, options) {
|
|
if (typeof method !== "function") {
|
|
throw new Error("method for before hook must be a function");
|
|
}
|
|
|
|
if (!options) {
|
|
options = {};
|
|
}
|
|
|
|
if (Array.isArray(name)) {
|
|
return name.reverse().reduce(function (callback, name) {
|
|
return register.bind(null, state, name, callback, options);
|
|
}, method)();
|
|
}
|
|
|
|
return Promise.resolve().then(function () {
|
|
if (!state.registry[name]) {
|
|
return method(options);
|
|
}
|
|
|
|
return state.registry[name].reduce(function (method, registered) {
|
|
return registered.hook.bind(null, method, options);
|
|
}, method)();
|
|
});
|
|
}
|
|
|
|
|
|
/***/ }),
|
|
|
|
/***/ 6819:
|
|
/***/ ((module) => {
|
|
|
|
module.exports = removeHook;
|
|
|
|
function removeHook(state, name, method) {
|
|
if (!state.registry[name]) {
|
|
return;
|
|
}
|
|
|
|
var index = state.registry[name]
|
|
.map(function (registered) {
|
|
return registered.orig;
|
|
})
|
|
.indexOf(method);
|
|
|
|
if (index === -1) {
|
|
return;
|
|
}
|
|
|
|
state.registry[name].splice(index, 1);
|
|
}
|
|
|
|
|
|
/***/ }),
|
|
|
|
/***/ 8932:
|
|
/***/ ((__unused_webpack_module, exports) => {
|
|
|
|
"use strict";
|
|
|
|
|
|
Object.defineProperty(exports, "__esModule", ({ value: true }));
|
|
|
|
class Deprecation extends Error {
|
|
constructor(message) {
|
|
super(message); // Maintains proper stack trace (only available on V8)
|
|
|
|
/* istanbul ignore next */
|
|
|
|
if (Error.captureStackTrace) {
|
|
Error.captureStackTrace(this, this.constructor);
|
|
}
|
|
|
|
this.name = 'Deprecation';
|
|
}
|
|
|
|
}
|
|
|
|
exports.Deprecation = Deprecation;
|
|
|
|
|
|
/***/ }),
|
|
|
|
/***/ 3287:
|
|
/***/ ((__unused_webpack_module, exports) => {
|
|
|
|
"use strict";
|
|
|
|
|
|
Object.defineProperty(exports, "__esModule", ({ value: true }));
|
|
|
|
/*!
|
|
* is-plain-object <https://github.com/jonschlinkert/is-plain-object>
|
|
*
|
|
* Copyright (c) 2014-2017, Jon Schlinkert.
|
|
* Released under the MIT License.
|
|
*/
|
|
|
|
function isObject(o) {
|
|
return Object.prototype.toString.call(o) === '[object Object]';
|
|
}
|
|
|
|
function isPlainObject(o) {
|
|
var ctor,prot;
|
|
|
|
if (isObject(o) === false) return false;
|
|
|
|
// If has modified constructor
|
|
ctor = o.constructor;
|
|
if (ctor === undefined) return true;
|
|
|
|
// If has modified prototype
|
|
prot = ctor.prototype;
|
|
if (isObject(prot) === false) return false;
|
|
|
|
// If constructor does not have an Object-specific method
|
|
if (prot.hasOwnProperty('isPrototypeOf') === false) {
|
|
return false;
|
|
}
|
|
|
|
// Most likely a plain Object
|
|
return true;
|
|
}
|
|
|
|
exports.isPlainObject = isPlainObject;
|
|
|
|
|
|
/***/ }),
|
|
|
|
/***/ 467:
|
|
/***/ ((module, exports, __nccwpck_require__) => {
|
|
|
|
"use strict";
|
|
|
|
|
|
Object.defineProperty(exports, "__esModule", ({ value: true }));
|
|
|
|
function _interopDefault (ex) { return (ex && (typeof ex === 'object') && 'default' in ex) ? ex['default'] : ex; }
|
|
|
|
var Stream = _interopDefault(__nccwpck_require__(2413));
|
|
var http = _interopDefault(__nccwpck_require__(8605));
|
|
var Url = _interopDefault(__nccwpck_require__(8835));
|
|
var https = _interopDefault(__nccwpck_require__(7211));
|
|
var zlib = _interopDefault(__nccwpck_require__(8761));
|
|
|
|
// Based on https://github.com/tmpvar/jsdom/blob/aa85b2abf07766ff7bf5c1f6daafb3726f2f2db5/lib/jsdom/living/blob.js
|
|
|
|
// fix for "Readable" isn't a named export issue
|
|
const Readable = Stream.Readable;
|
|
|
|
const BUFFER = Symbol('buffer');
|
|
const TYPE = Symbol('type');
|
|
|
|
class Blob {
|
|
constructor() {
|
|
this[TYPE] = '';
|
|
|
|
const blobParts = arguments[0];
|
|
const options = arguments[1];
|
|
|
|
const buffers = [];
|
|
let size = 0;
|
|
|
|
if (blobParts) {
|
|
const a = blobParts;
|
|
const length = Number(a.length);
|
|
for (let i = 0; i < length; i++) {
|
|
const element = a[i];
|
|
let buffer;
|
|
if (element instanceof Buffer) {
|
|
buffer = element;
|
|
} else if (ArrayBuffer.isView(element)) {
|
|
buffer = Buffer.from(element.buffer, element.byteOffset, element.byteLength);
|
|
} else if (element instanceof ArrayBuffer) {
|
|
buffer = Buffer.from(element);
|
|
} else if (element instanceof Blob) {
|
|
buffer = element[BUFFER];
|
|
} else {
|
|
buffer = Buffer.from(typeof element === 'string' ? element : String(element));
|
|
}
|
|
size += buffer.length;
|
|
buffers.push(buffer);
|
|
}
|
|
}
|
|
|
|
this[BUFFER] = Buffer.concat(buffers);
|
|
|
|
let type = options && options.type !== undefined && String(options.type).toLowerCase();
|
|
if (type && !/[^\u0020-\u007E]/.test(type)) {
|
|
this[TYPE] = type;
|
|
}
|
|
}
|
|
get size() {
|
|
return this[BUFFER].length;
|
|
}
|
|
get type() {
|
|
return this[TYPE];
|
|
}
|
|
text() {
|
|
return Promise.resolve(this[BUFFER].toString());
|
|
}
|
|
arrayBuffer() {
|
|
const buf = this[BUFFER];
|
|
const ab = buf.buffer.slice(buf.byteOffset, buf.byteOffset + buf.byteLength);
|
|
return Promise.resolve(ab);
|
|
}
|
|
stream() {
|
|
const readable = new Readable();
|
|
readable._read = function () {};
|
|
readable.push(this[BUFFER]);
|
|
readable.push(null);
|
|
return readable;
|
|
}
|
|
toString() {
|
|
return '[object Blob]';
|
|
}
|
|
slice() {
|
|
const size = this.size;
|
|
|
|
const start = arguments[0];
|
|
const end = arguments[1];
|
|
let relativeStart, relativeEnd;
|
|
if (start === undefined) {
|
|
relativeStart = 0;
|
|
} else if (start < 0) {
|
|
relativeStart = Math.max(size + start, 0);
|
|
} else {
|
|
relativeStart = Math.min(start, size);
|
|
}
|
|
if (end === undefined) {
|
|
relativeEnd = size;
|
|
} else if (end < 0) {
|
|
relativeEnd = Math.max(size + end, 0);
|
|
} else {
|
|
relativeEnd = Math.min(end, size);
|
|
}
|
|
const span = Math.max(relativeEnd - relativeStart, 0);
|
|
|
|
const buffer = this[BUFFER];
|
|
const slicedBuffer = buffer.slice(relativeStart, relativeStart + span);
|
|
const blob = new Blob([], { type: arguments[2] });
|
|
blob[BUFFER] = slicedBuffer;
|
|
return blob;
|
|
}
|
|
}
|
|
|
|
Object.defineProperties(Blob.prototype, {
|
|
size: { enumerable: true },
|
|
type: { enumerable: true },
|
|
slice: { enumerable: true }
|
|
});
|
|
|
|
Object.defineProperty(Blob.prototype, Symbol.toStringTag, {
|
|
value: 'Blob',
|
|
writable: false,
|
|
enumerable: false,
|
|
configurable: true
|
|
});
|
|
|
|
/**
|
|
* fetch-error.js
|
|
*
|
|
* FetchError interface for operational errors
|
|
*/
|
|
|
|
/**
|
|
* Create FetchError instance
|
|
*
|
|
* @param String message Error message for human
|
|
* @param String type Error type for machine
|
|
* @param String systemError For Node.js system error
|
|
* @return FetchError
|
|
*/
|
|
function FetchError(message, type, systemError) {
|
|
Error.call(this, message);
|
|
|
|
this.message = message;
|
|
this.type = type;
|
|
|
|
// when err.type is `system`, err.code contains system error code
|
|
if (systemError) {
|
|
this.code = this.errno = systemError.code;
|
|
}
|
|
|
|
// hide custom error implementation details from end-users
|
|
Error.captureStackTrace(this, this.constructor);
|
|
}
|
|
|
|
FetchError.prototype = Object.create(Error.prototype);
|
|
FetchError.prototype.constructor = FetchError;
|
|
FetchError.prototype.name = 'FetchError';
|
|
|
|
let convert;
|
|
try {
|
|
convert = __nccwpck_require__(2877).convert;
|
|
} catch (e) {}
|
|
|
|
const INTERNALS = Symbol('Body internals');
|
|
|
|
// fix an issue where "PassThrough" isn't a named export for node <10
|
|
const PassThrough = Stream.PassThrough;
|
|
|
|
/**
|
|
* Body mixin
|
|
*
|
|
* Ref: https://fetch.spec.whatwg.org/#body
|
|
*
|
|
* @param Stream body Readable stream
|
|
* @param Object opts Response options
|
|
* @return Void
|
|
*/
|
|
function Body(body) {
|
|
var _this = this;
|
|
|
|
var _ref = arguments.length > 1 && arguments[1] !== undefined ? arguments[1] : {},
|
|
_ref$size = _ref.size;
|
|
|
|
let size = _ref$size === undefined ? 0 : _ref$size;
|
|
var _ref$timeout = _ref.timeout;
|
|
let timeout = _ref$timeout === undefined ? 0 : _ref$timeout;
|
|
|
|
if (body == null) {
|
|
// body is undefined or null
|
|
body = null;
|
|
} else if (isURLSearchParams(body)) {
|
|
// body is a URLSearchParams
|
|
body = Buffer.from(body.toString());
|
|
} else if (isBlob(body)) ; else if (Buffer.isBuffer(body)) ; else if (Object.prototype.toString.call(body) === '[object ArrayBuffer]') {
|
|
// body is ArrayBuffer
|
|
body = Buffer.from(body);
|
|
} else if (ArrayBuffer.isView(body)) {
|
|
// body is ArrayBufferView
|
|
body = Buffer.from(body.buffer, body.byteOffset, body.byteLength);
|
|
} else if (body instanceof Stream) ; else {
|
|
// none of the above
|
|
// coerce to string then buffer
|
|
body = Buffer.from(String(body));
|
|
}
|
|
this[INTERNALS] = {
|
|
body,
|
|
disturbed: false,
|
|
error: null
|
|
};
|
|
this.size = size;
|
|
this.timeout = timeout;
|
|
|
|
if (body instanceof Stream) {
|
|
body.on('error', function (err) {
|
|
const error = err.name === 'AbortError' ? err : new FetchError(`Invalid response body while trying to fetch ${_this.url}: ${err.message}`, 'system', err);
|
|
_this[INTERNALS].error = error;
|
|
});
|
|
}
|
|
}
|
|
|
|
Body.prototype = {
|
|
get body() {
|
|
return this[INTERNALS].body;
|
|
},
|
|
|
|
get bodyUsed() {
|
|
return this[INTERNALS].disturbed;
|
|
},
|
|
|
|
/**
|
|
* Decode response as ArrayBuffer
|
|
*
|
|
* @return Promise
|
|
*/
|
|
arrayBuffer() {
|
|
return consumeBody.call(this).then(function (buf) {
|
|
return buf.buffer.slice(buf.byteOffset, buf.byteOffset + buf.byteLength);
|
|
});
|
|
},
|
|
|
|
/**
|
|
* Return raw response as Blob
|
|
*
|
|
* @return Promise
|
|
*/
|
|
blob() {
|
|
let ct = this.headers && this.headers.get('content-type') || '';
|
|
return consumeBody.call(this).then(function (buf) {
|
|
return Object.assign(
|
|
// Prevent copying
|
|
new Blob([], {
|
|
type: ct.toLowerCase()
|
|
}), {
|
|
[BUFFER]: buf
|
|
});
|
|
});
|
|
},
|
|
|
|
/**
|
|
* Decode response as json
|
|
*
|
|
* @return Promise
|
|
*/
|
|
json() {
|
|
var _this2 = this;
|
|
|
|
return consumeBody.call(this).then(function (buffer) {
|
|
try {
|
|
return JSON.parse(buffer.toString());
|
|
} catch (err) {
|
|
return Body.Promise.reject(new FetchError(`invalid json response body at ${_this2.url} reason: ${err.message}`, 'invalid-json'));
|
|
}
|
|
});
|
|
},
|
|
|
|
/**
|
|
* Decode response as text
|
|
*
|
|
* @return Promise
|
|
*/
|
|
text() {
|
|
return consumeBody.call(this).then(function (buffer) {
|
|
return buffer.toString();
|
|
});
|
|
},
|
|
|
|
/**
|
|
* Decode response as buffer (non-spec api)
|
|
*
|
|
* @return Promise
|
|
*/
|
|
buffer() {
|
|
return consumeBody.call(this);
|
|
},
|
|
|
|
/**
|
|
* Decode response as text, while automatically detecting the encoding and
|
|
* trying to decode to UTF-8 (non-spec api)
|
|
*
|
|
* @return Promise
|
|
*/
|
|
textConverted() {
|
|
var _this3 = this;
|
|
|
|
return consumeBody.call(this).then(function (buffer) {
|
|
return convertBody(buffer, _this3.headers);
|
|
});
|
|
}
|
|
};
|
|
|
|
// In browsers, all properties are enumerable.
|
|
Object.defineProperties(Body.prototype, {
|
|
body: { enumerable: true },
|
|
bodyUsed: { enumerable: true },
|
|
arrayBuffer: { enumerable: true },
|
|
blob: { enumerable: true },
|
|
json: { enumerable: true },
|
|
text: { enumerable: true }
|
|
});
|
|
|
|
Body.mixIn = function (proto) {
|
|
for (const name of Object.getOwnPropertyNames(Body.prototype)) {
|
|
// istanbul ignore else: future proof
|
|
if (!(name in proto)) {
|
|
const desc = Object.getOwnPropertyDescriptor(Body.prototype, name);
|
|
Object.defineProperty(proto, name, desc);
|
|
}
|
|
}
|
|
};
|
|
|
|
/**
|
|
* Consume and convert an entire Body to a Buffer.
|
|
*
|
|
* Ref: https://fetch.spec.whatwg.org/#concept-body-consume-body
|
|
*
|
|
* @return Promise
|
|
*/
|
|
function consumeBody() {
|
|
var _this4 = this;
|
|
|
|
if (this[INTERNALS].disturbed) {
|
|
return Body.Promise.reject(new TypeError(`body used already for: ${this.url}`));
|
|
}
|
|
|
|
this[INTERNALS].disturbed = true;
|
|
|
|
if (this[INTERNALS].error) {
|
|
return Body.Promise.reject(this[INTERNALS].error);
|
|
}
|
|
|
|
let body = this.body;
|
|
|
|
// body is null
|
|
if (body === null) {
|
|
return Body.Promise.resolve(Buffer.alloc(0));
|
|
}
|
|
|
|
// body is blob
|
|
if (isBlob(body)) {
|
|
body = body.stream();
|
|
}
|
|
|
|
// body is buffer
|
|
if (Buffer.isBuffer(body)) {
|
|
return Body.Promise.resolve(body);
|
|
}
|
|
|
|
// istanbul ignore if: should never happen
|
|
if (!(body instanceof Stream)) {
|
|
return Body.Promise.resolve(Buffer.alloc(0));
|
|
}
|
|
|
|
// body is stream
|
|
// get ready to actually consume the body
|
|
let accum = [];
|
|
let accumBytes = 0;
|
|
let abort = false;
|
|
|
|
return new Body.Promise(function (resolve, reject) {
|
|
let resTimeout;
|
|
|
|
// allow timeout on slow response body
|
|
if (_this4.timeout) {
|
|
resTimeout = setTimeout(function () {
|
|
abort = true;
|
|
reject(new FetchError(`Response timeout while trying to fetch ${_this4.url} (over ${_this4.timeout}ms)`, 'body-timeout'));
|
|
}, _this4.timeout);
|
|
}
|
|
|
|
// handle stream errors
|
|
body.on('error', function (err) {
|
|
if (err.name === 'AbortError') {
|
|
// if the request was aborted, reject with this Error
|
|
abort = true;
|
|
reject(err);
|
|
} else {
|
|
// other errors, such as incorrect content-encoding
|
|
reject(new FetchError(`Invalid response body while trying to fetch ${_this4.url}: ${err.message}`, 'system', err));
|
|
}
|
|
});
|
|
|
|
body.on('data', function (chunk) {
|
|
if (abort || chunk === null) {
|
|
return;
|
|
}
|
|
|
|
if (_this4.size && accumBytes + chunk.length > _this4.size) {
|
|
abort = true;
|
|
reject(new FetchError(`content size at ${_this4.url} over limit: ${_this4.size}`, 'max-size'));
|
|
return;
|
|
}
|
|
|
|
accumBytes += chunk.length;
|
|
accum.push(chunk);
|
|
});
|
|
|
|
body.on('end', function () {
|
|
if (abort) {
|
|
return;
|
|
}
|
|
|
|
clearTimeout(resTimeout);
|
|
|
|
try {
|
|
resolve(Buffer.concat(accum, accumBytes));
|
|
} catch (err) {
|
|
// handle streams that have accumulated too much data (issue #414)
|
|
reject(new FetchError(`Could not create Buffer from response body for ${_this4.url}: ${err.message}`, 'system', err));
|
|
}
|
|
});
|
|
});
|
|
}
|
|
|
|
/**
|
|
* Detect buffer encoding and convert to target encoding
|
|
* ref: http://www.w3.org/TR/2011/WD-html5-20110113/parsing.html#determining-the-character-encoding
|
|
*
|
|
* @param Buffer buffer Incoming buffer
|
|
* @param String encoding Target encoding
|
|
* @return String
|
|
*/
|
|
function convertBody(buffer, headers) {
|
|
if (typeof convert !== 'function') {
|
|
throw new Error('The package `encoding` must be installed to use the textConverted() function');
|
|
}
|
|
|
|
const ct = headers.get('content-type');
|
|
let charset = 'utf-8';
|
|
let res, str;
|
|
|
|
// header
|
|
if (ct) {
|
|
res = /charset=([^;]*)/i.exec(ct);
|
|
}
|
|
|
|
// no charset in content type, peek at response body for at most 1024 bytes
|
|
str = buffer.slice(0, 1024).toString();
|
|
|
|
// html5
|
|
if (!res && str) {
|
|
res = /<meta.+?charset=(['"])(.+?)\1/i.exec(str);
|
|
}
|
|
|
|
// html4
|
|
if (!res && str) {
|
|
res = /<meta[\s]+?http-equiv=(['"])content-type\1[\s]+?content=(['"])(.+?)\2/i.exec(str);
|
|
if (!res) {
|
|
res = /<meta[\s]+?content=(['"])(.+?)\1[\s]+?http-equiv=(['"])content-type\3/i.exec(str);
|
|
if (res) {
|
|
res.pop(); // drop last quote
|
|
}
|
|
}
|
|
|
|
if (res) {
|
|
res = /charset=(.*)/i.exec(res.pop());
|
|
}
|
|
}
|
|
|
|
// xml
|
|
if (!res && str) {
|
|
res = /<\?xml.+?encoding=(['"])(.+?)\1/i.exec(str);
|
|
}
|
|
|
|
// found charset
|
|
if (res) {
|
|
charset = res.pop();
|
|
|
|
// prevent decode issues when sites use incorrect encoding
|
|
// ref: https://hsivonen.fi/encoding-menu/
|
|
if (charset === 'gb2312' || charset === 'gbk') {
|
|
charset = 'gb18030';
|
|
}
|
|
}
|
|
|
|
// turn raw buffers into a single utf-8 buffer
|
|
return convert(buffer, 'UTF-8', charset).toString();
|
|
}
|
|
|
|
/**
|
|
* Detect a URLSearchParams object
|
|
* ref: https://github.com/bitinn/node-fetch/issues/296#issuecomment-307598143
|
|
*
|
|
* @param Object obj Object to detect by type or brand
|
|
* @return String
|
|
*/
|
|
function isURLSearchParams(obj) {
|
|
// Duck-typing as a necessary condition.
|
|
if (typeof obj !== 'object' || typeof obj.append !== 'function' || typeof obj.delete !== 'function' || typeof obj.get !== 'function' || typeof obj.getAll !== 'function' || typeof obj.has !== 'function' || typeof obj.set !== 'function') {
|
|
return false;
|
|
}
|
|
|
|
// Brand-checking and more duck-typing as optional condition.
|
|
return obj.constructor.name === 'URLSearchParams' || Object.prototype.toString.call(obj) === '[object URLSearchParams]' || typeof obj.sort === 'function';
|
|
}
|
|
|
|
/**
|
|
* Check if `obj` is a W3C `Blob` object (which `File` inherits from)
|
|
* @param {*} obj
|
|
* @return {boolean}
|
|
*/
|
|
function isBlob(obj) {
|
|
return typeof obj === 'object' && typeof obj.arrayBuffer === 'function' && typeof obj.type === 'string' && typeof obj.stream === 'function' && typeof obj.constructor === 'function' && typeof obj.constructor.name === 'string' && /^(Blob|File)$/.test(obj.constructor.name) && /^(Blob|File)$/.test(obj[Symbol.toStringTag]);
|
|
}
|
|
|
|
/**
|
|
* Clone body given Res/Req instance
|
|
*
|
|
* @param Mixed instance Response or Request instance
|
|
* @return Mixed
|
|
*/
|
|
function clone(instance) {
|
|
let p1, p2;
|
|
let body = instance.body;
|
|
|
|
// don't allow cloning a used body
|
|
if (instance.bodyUsed) {
|
|
throw new Error('cannot clone body after it is used');
|
|
}
|
|
|
|
// check that body is a stream and not form-data object
|
|
// note: we can't clone the form-data object without having it as a dependency
|
|
if (body instanceof Stream && typeof body.getBoundary !== 'function') {
|
|
// tee instance body
|
|
p1 = new PassThrough();
|
|
p2 = new PassThrough();
|
|
body.pipe(p1);
|
|
body.pipe(p2);
|
|
// set instance body to teed body and return the other teed body
|
|
instance[INTERNALS].body = p1;
|
|
body = p2;
|
|
}
|
|
|
|
return body;
|
|
}
|
|
|
|
/**
|
|
* Performs the operation "extract a `Content-Type` value from |object|" as
|
|
* specified in the specification:
|
|
* https://fetch.spec.whatwg.org/#concept-bodyinit-extract
|
|
*
|
|
* This function assumes that instance.body is present.
|
|
*
|
|
* @param Mixed instance Any options.body input
|
|
*/
|
|
function extractContentType(body) {
|
|
if (body === null) {
|
|
// body is null
|
|
return null;
|
|
} else if (typeof body === 'string') {
|
|
// body is string
|
|
return 'text/plain;charset=UTF-8';
|
|
} else if (isURLSearchParams(body)) {
|
|
// body is a URLSearchParams
|
|
return 'application/x-www-form-urlencoded;charset=UTF-8';
|
|
} else if (isBlob(body)) {
|
|
// body is blob
|
|
return body.type || null;
|
|
} else if (Buffer.isBuffer(body)) {
|
|
// body is buffer
|
|
return null;
|
|
} else if (Object.prototype.toString.call(body) === '[object ArrayBuffer]') {
|
|
// body is ArrayBuffer
|
|
return null;
|
|
} else if (ArrayBuffer.isView(body)) {
|
|
// body is ArrayBufferView
|
|
return null;
|
|
} else if (typeof body.getBoundary === 'function') {
|
|
// detect form data input from form-data module
|
|
return `multipart/form-data;boundary=${body.getBoundary()}`;
|
|
} else if (body instanceof Stream) {
|
|
// body is stream
|
|
// can't really do much about this
|
|
return null;
|
|
} else {
|
|
// Body constructor defaults other things to string
|
|
return 'text/plain;charset=UTF-8';
|
|
}
|
|
}
|
|
|
|
/**
|
|
* The Fetch Standard treats this as if "total bytes" is a property on the body.
|
|
* For us, we have to explicitly get it with a function.
|
|
*
|
|
* ref: https://fetch.spec.whatwg.org/#concept-body-total-bytes
|
|
*
|
|
* @param Body instance Instance of Body
|
|
* @return Number? Number of bytes, or null if not possible
|
|
*/
|
|
function getTotalBytes(instance) {
|
|
const body = instance.body;
|
|
|
|
|
|
if (body === null) {
|
|
// body is null
|
|
return 0;
|
|
} else if (isBlob(body)) {
|
|
return body.size;
|
|
} else if (Buffer.isBuffer(body)) {
|
|
// body is buffer
|
|
return body.length;
|
|
} else if (body && typeof body.getLengthSync === 'function') {
|
|
// detect form data input from form-data module
|
|
if (body._lengthRetrievers && body._lengthRetrievers.length == 0 || // 1.x
|
|
body.hasKnownLength && body.hasKnownLength()) {
|
|
// 2.x
|
|
return body.getLengthSync();
|
|
}
|
|
return null;
|
|
} else {
|
|
// body is stream
|
|
return null;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Write a Body to a Node.js WritableStream (e.g. http.Request) object.
|
|
*
|
|
* @param Body instance Instance of Body
|
|
* @return Void
|
|
*/
|
|
function writeToStream(dest, instance) {
|
|
const body = instance.body;
|
|
|
|
|
|
if (body === null) {
|
|
// body is null
|
|
dest.end();
|
|
} else if (isBlob(body)) {
|
|
body.stream().pipe(dest);
|
|
} else if (Buffer.isBuffer(body)) {
|
|
// body is buffer
|
|
dest.write(body);
|
|
dest.end();
|
|
} else {
|
|
// body is stream
|
|
body.pipe(dest);
|
|
}
|
|
}
|
|
|
|
// expose Promise
|
|
Body.Promise = global.Promise;
|
|
|
|
/**
|
|
* headers.js
|
|
*
|
|
* Headers class offers convenient helpers
|
|
*/
|
|
|
|
const invalidTokenRegex = /[^\^_`a-zA-Z\-0-9!#$%&'*+.|~]/;
|
|
const invalidHeaderCharRegex = /[^\t\x20-\x7e\x80-\xff]/;
|
|
|
|
function validateName(name) {
|
|
name = `${name}`;
|
|
if (invalidTokenRegex.test(name) || name === '') {
|
|
throw new TypeError(`${name} is not a legal HTTP header name`);
|
|
}
|
|
}
|
|
|
|
function validateValue(value) {
|
|
value = `${value}`;
|
|
if (invalidHeaderCharRegex.test(value)) {
|
|
throw new TypeError(`${value} is not a legal HTTP header value`);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Find the key in the map object given a header name.
|
|
*
|
|
* Returns undefined if not found.
|
|
*
|
|
* @param String name Header name
|
|
* @return String|Undefined
|
|
*/
|
|
function find(map, name) {
|
|
name = name.toLowerCase();
|
|
for (const key in map) {
|
|
if (key.toLowerCase() === name) {
|
|
return key;
|
|
}
|
|
}
|
|
return undefined;
|
|
}
|
|
|
|
const MAP = Symbol('map');
|
|
class Headers {
|
|
/**
|
|
* Headers class
|
|
*
|
|
* @param Object headers Response headers
|
|
* @return Void
|
|
*/
|
|
constructor() {
|
|
let init = arguments.length > 0 && arguments[0] !== undefined ? arguments[0] : undefined;
|
|
|
|
this[MAP] = Object.create(null);
|
|
|
|
if (init instanceof Headers) {
|
|
const rawHeaders = init.raw();
|
|
const headerNames = Object.keys(rawHeaders);
|
|
|
|
for (const headerName of headerNames) {
|
|
for (const value of rawHeaders[headerName]) {
|
|
this.append(headerName, value);
|
|
}
|
|
}
|
|
|
|
return;
|
|
}
|
|
|
|
// We don't worry about converting prop to ByteString here as append()
|
|
// will handle it.
|
|
if (init == null) ; else if (typeof init === 'object') {
|
|
const method = init[Symbol.iterator];
|
|
if (method != null) {
|
|
if (typeof method !== 'function') {
|
|
throw new TypeError('Header pairs must be iterable');
|
|
}
|
|
|
|
// sequence<sequence<ByteString>>
|
|
// Note: per spec we have to first exhaust the lists then process them
|
|
const pairs = [];
|
|
for (const pair of init) {
|
|
if (typeof pair !== 'object' || typeof pair[Symbol.iterator] !== 'function') {
|
|
throw new TypeError('Each header pair must be iterable');
|
|
}
|
|
pairs.push(Array.from(pair));
|
|
}
|
|
|
|
for (const pair of pairs) {
|
|
if (pair.length !== 2) {
|
|
throw new TypeError('Each header pair must be a name/value tuple');
|
|
}
|
|
this.append(pair[0], pair[1]);
|
|
}
|
|
} else {
|
|
// record<ByteString, ByteString>
|
|
for (const key of Object.keys(init)) {
|
|
const value = init[key];
|
|
this.append(key, value);
|
|
}
|
|
}
|
|
} else {
|
|
throw new TypeError('Provided initializer must be an object');
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Return combined header value given name
|
|
*
|
|
* @param String name Header name
|
|
* @return Mixed
|
|
*/
|
|
get(name) {
|
|
name = `${name}`;
|
|
validateName(name);
|
|
const key = find(this[MAP], name);
|
|
if (key === undefined) {
|
|
return null;
|
|
}
|
|
|
|
return this[MAP][key].join(', ');
|
|
}
|
|
|
|
/**
|
|
* Iterate over all headers
|
|
*
|
|
* @param Function callback Executed for each item with parameters (value, name, thisArg)
|
|
* @param Boolean thisArg `this` context for callback function
|
|
* @return Void
|
|
*/
|
|
forEach(callback) {
|
|
let thisArg = arguments.length > 1 && arguments[1] !== undefined ? arguments[1] : undefined;
|
|
|
|
let pairs = getHeaders(this);
|
|
let i = 0;
|
|
while (i < pairs.length) {
|
|
var _pairs$i = pairs[i];
|
|
const name = _pairs$i[0],
|
|
value = _pairs$i[1];
|
|
|
|
callback.call(thisArg, value, name, this);
|
|
pairs = getHeaders(this);
|
|
i++;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Overwrite header values given name
|
|
*
|
|
* @param String name Header name
|
|
* @param String value Header value
|
|
* @return Void
|
|
*/
|
|
set(name, value) {
|
|
name = `${name}`;
|
|
value = `${value}`;
|
|
validateName(name);
|
|
validateValue(value);
|
|
const key = find(this[MAP], name);
|
|
this[MAP][key !== undefined ? key : name] = [value];
|
|
}
|
|
|
|
/**
|
|
* Append a value onto existing header
|
|
*
|
|
* @param String name Header name
|
|
* @param String value Header value
|
|
* @return Void
|
|
*/
|
|
append(name, value) {
|
|
name = `${name}`;
|
|
value = `${value}`;
|
|
validateName(name);
|
|
validateValue(value);
|
|
const key = find(this[MAP], name);
|
|
if (key !== undefined) {
|
|
this[MAP][key].push(value);
|
|
} else {
|
|
this[MAP][name] = [value];
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Check for header name existence
|
|
*
|
|
* @param String name Header name
|
|
* @return Boolean
|
|
*/
|
|
has(name) {
|
|
name = `${name}`;
|
|
validateName(name);
|
|
return find(this[MAP], name) !== undefined;
|
|
}
|
|
|
|
/**
|
|
* Delete all header values given name
|
|
*
|
|
* @param String name Header name
|
|
* @return Void
|
|
*/
|
|
delete(name) {
|
|
name = `${name}`;
|
|
validateName(name);
|
|
const key = find(this[MAP], name);
|
|
if (key !== undefined) {
|
|
delete this[MAP][key];
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Return raw headers (non-spec api)
|
|
*
|
|
* @return Object
|
|
*/
|
|
raw() {
|
|
return this[MAP];
|
|
}
|
|
|
|
/**
|
|
* Get an iterator on keys.
|
|
*
|
|
* @return Iterator
|
|
*/
|
|
keys() {
|
|
return createHeadersIterator(this, 'key');
|
|
}
|
|
|
|
/**
|
|
* Get an iterator on values.
|
|
*
|
|
* @return Iterator
|
|
*/
|
|
values() {
|
|
return createHeadersIterator(this, 'value');
|
|
}
|
|
|
|
/**
|
|
* Get an iterator on entries.
|
|
*
|
|
* This is the default iterator of the Headers object.
|
|
*
|
|
* @return Iterator
|
|
*/
|
|
[Symbol.iterator]() {
|
|
return createHeadersIterator(this, 'key+value');
|
|
}
|
|
}
|
|
Headers.prototype.entries = Headers.prototype[Symbol.iterator];
|
|
|
|
Object.defineProperty(Headers.prototype, Symbol.toStringTag, {
|
|
value: 'Headers',
|
|
writable: false,
|
|
enumerable: false,
|
|
configurable: true
|
|
});
|
|
|
|
Object.defineProperties(Headers.prototype, {
|
|
get: { enumerable: true },
|
|
forEach: { enumerable: true },
|
|
set: { enumerable: true },
|
|
append: { enumerable: true },
|
|
has: { enumerable: true },
|
|
delete: { enumerable: true },
|
|
keys: { enumerable: true },
|
|
values: { enumerable: true },
|
|
entries: { enumerable: true }
|
|
});
|
|
|
|
function getHeaders(headers) {
|
|
let kind = arguments.length > 1 && arguments[1] !== undefined ? arguments[1] : 'key+value';
|
|
|
|
const keys = Object.keys(headers[MAP]).sort();
|
|
return keys.map(kind === 'key' ? function (k) {
|
|
return k.toLowerCase();
|
|
} : kind === 'value' ? function (k) {
|
|
return headers[MAP][k].join(', ');
|
|
} : function (k) {
|
|
return [k.toLowerCase(), headers[MAP][k].join(', ')];
|
|
});
|
|
}
|
|
|
|
const INTERNAL = Symbol('internal');
|
|
|
|
function createHeadersIterator(target, kind) {
|
|
const iterator = Object.create(HeadersIteratorPrototype);
|
|
iterator[INTERNAL] = {
|
|
target,
|
|
kind,
|
|
index: 0
|
|
};
|
|
return iterator;
|
|
}
|
|
|
|
const HeadersIteratorPrototype = Object.setPrototypeOf({
|
|
next() {
|
|
// istanbul ignore if
|
|
if (!this || Object.getPrototypeOf(this) !== HeadersIteratorPrototype) {
|
|
throw new TypeError('Value of `this` is not a HeadersIterator');
|
|
}
|
|
|
|
var _INTERNAL = this[INTERNAL];
|
|
const target = _INTERNAL.target,
|
|
kind = _INTERNAL.kind,
|
|
index = _INTERNAL.index;
|
|
|
|
const values = getHeaders(target, kind);
|
|
const len = values.length;
|
|
if (index >= len) {
|
|
return {
|
|
value: undefined,
|
|
done: true
|
|
};
|
|
}
|
|
|
|
this[INTERNAL].index = index + 1;
|
|
|
|
return {
|
|
value: values[index],
|
|
done: false
|
|
};
|
|
}
|
|
}, Object.getPrototypeOf(Object.getPrototypeOf([][Symbol.iterator]())));
|
|
|
|
Object.defineProperty(HeadersIteratorPrototype, Symbol.toStringTag, {
|
|
value: 'HeadersIterator',
|
|
writable: false,
|
|
enumerable: false,
|
|
configurable: true
|
|
});
|
|
|
|
/**
|
|
* Export the Headers object in a form that Node.js can consume.
|
|
*
|
|
* @param Headers headers
|
|
* @return Object
|
|
*/
|
|
function exportNodeCompatibleHeaders(headers) {
|
|
const obj = Object.assign({ __proto__: null }, headers[MAP]);
|
|
|
|
// http.request() only supports string as Host header. This hack makes
|
|
// specifying custom Host header possible.
|
|
const hostHeaderKey = find(headers[MAP], 'Host');
|
|
if (hostHeaderKey !== undefined) {
|
|
obj[hostHeaderKey] = obj[hostHeaderKey][0];
|
|
}
|
|
|
|
return obj;
|
|
}
|
|
|
|
/**
|
|
* Create a Headers object from an object of headers, ignoring those that do
|
|
* not conform to HTTP grammar productions.
|
|
*
|
|
* @param Object obj Object of headers
|
|
* @return Headers
|
|
*/
|
|
function createHeadersLenient(obj) {
|
|
const headers = new Headers();
|
|
for (const name of Object.keys(obj)) {
|
|
if (invalidTokenRegex.test(name)) {
|
|
continue;
|
|
}
|
|
if (Array.isArray(obj[name])) {
|
|
for (const val of obj[name]) {
|
|
if (invalidHeaderCharRegex.test(val)) {
|
|
continue;
|
|
}
|
|
if (headers[MAP][name] === undefined) {
|
|
headers[MAP][name] = [val];
|
|
} else {
|
|
headers[MAP][name].push(val);
|
|
}
|
|
}
|
|
} else if (!invalidHeaderCharRegex.test(obj[name])) {
|
|
headers[MAP][name] = [obj[name]];
|
|
}
|
|
}
|
|
return headers;
|
|
}
|
|
|
|
const INTERNALS$1 = Symbol('Response internals');
|
|
|
|
// fix an issue where "STATUS_CODES" aren't a named export for node <10
|
|
const STATUS_CODES = http.STATUS_CODES;
|
|
|
|
/**
|
|
* Response class
|
|
*
|
|
* @param Stream body Readable stream
|
|
* @param Object opts Response options
|
|
* @return Void
|
|
*/
|
|
class Response {
|
|
constructor() {
|
|
let body = arguments.length > 0 && arguments[0] !== undefined ? arguments[0] : null;
|
|
let opts = arguments.length > 1 && arguments[1] !== undefined ? arguments[1] : {};
|
|
|
|
Body.call(this, body, opts);
|
|
|
|
const status = opts.status || 200;
|
|
const headers = new Headers(opts.headers);
|
|
|
|
if (body != null && !headers.has('Content-Type')) {
|
|
const contentType = extractContentType(body);
|
|
if (contentType) {
|
|
headers.append('Content-Type', contentType);
|
|
}
|
|
}
|
|
|
|
this[INTERNALS$1] = {
|
|
url: opts.url,
|
|
status,
|
|
statusText: opts.statusText || STATUS_CODES[status],
|
|
headers,
|
|
counter: opts.counter
|
|
};
|
|
}
|
|
|
|
get url() {
|
|
return this[INTERNALS$1].url || '';
|
|
}
|
|
|
|
get status() {
|
|
return this[INTERNALS$1].status;
|
|
}
|
|
|
|
/**
|
|
* Convenience property representing if the request ended normally
|
|
*/
|
|
get ok() {
|
|
return this[INTERNALS$1].status >= 200 && this[INTERNALS$1].status < 300;
|
|
}
|
|
|
|
get redirected() {
|
|
return this[INTERNALS$1].counter > 0;
|
|
}
|
|
|
|
get statusText() {
|
|
return this[INTERNALS$1].statusText;
|
|
}
|
|
|
|
get headers() {
|
|
return this[INTERNALS$1].headers;
|
|
}
|
|
|
|
/**
|
|
* Clone this response
|
|
*
|
|
* @return Response
|
|
*/
|
|
clone() {
|
|
return new Response(clone(this), {
|
|
url: this.url,
|
|
status: this.status,
|
|
statusText: this.statusText,
|
|
headers: this.headers,
|
|
ok: this.ok,
|
|
redirected: this.redirected
|
|
});
|
|
}
|
|
}
|
|
|
|
Body.mixIn(Response.prototype);
|
|
|
|
Object.defineProperties(Response.prototype, {
|
|
url: { enumerable: true },
|
|
status: { enumerable: true },
|
|
ok: { enumerable: true },
|
|
redirected: { enumerable: true },
|
|
statusText: { enumerable: true },
|
|
headers: { enumerable: true },
|
|
clone: { enumerable: true }
|
|
});
|
|
|
|
Object.defineProperty(Response.prototype, Symbol.toStringTag, {
|
|
value: 'Response',
|
|
writable: false,
|
|
enumerable: false,
|
|
configurable: true
|
|
});
|
|
|
|
const INTERNALS$2 = Symbol('Request internals');
|
|
|
|
// fix an issue where "format", "parse" aren't a named export for node <10
|
|
const parse_url = Url.parse;
|
|
const format_url = Url.format;
|
|
|
|
const streamDestructionSupported = 'destroy' in Stream.Readable.prototype;
|
|
|
|
/**
|
|
* Check if a value is an instance of Request.
|
|
*
|
|
* @param Mixed input
|
|
* @return Boolean
|
|
*/
|
|
function isRequest(input) {
|
|
return typeof input === 'object' && typeof input[INTERNALS$2] === 'object';
|
|
}
|
|
|
|
function isAbortSignal(signal) {
|
|
const proto = signal && typeof signal === 'object' && Object.getPrototypeOf(signal);
|
|
return !!(proto && proto.constructor.name === 'AbortSignal');
|
|
}
|
|
|
|
/**
|
|
* Request class
|
|
*
|
|
* @param Mixed input Url or Request instance
|
|
* @param Object init Custom options
|
|
* @return Void
|
|
*/
|
|
class Request {
|
|
constructor(input) {
|
|
let init = arguments.length > 1 && arguments[1] !== undefined ? arguments[1] : {};
|
|
|
|
let parsedURL;
|
|
|
|
// normalize input
|
|
if (!isRequest(input)) {
|
|
if (input && input.href) {
|
|
// in order to support Node.js' Url objects; though WHATWG's URL objects
|
|
// will fall into this branch also (since their `toString()` will return
|
|
// `href` property anyway)
|
|
parsedURL = parse_url(input.href);
|
|
} else {
|
|
// coerce input to a string before attempting to parse
|
|
parsedURL = parse_url(`${input}`);
|
|
}
|
|
input = {};
|
|
} else {
|
|
parsedURL = parse_url(input.url);
|
|
}
|
|
|
|
let method = init.method || input.method || 'GET';
|
|
method = method.toUpperCase();
|
|
|
|
if ((init.body != null || isRequest(input) && input.body !== null) && (method === 'GET' || method === 'HEAD')) {
|
|
throw new TypeError('Request with GET/HEAD method cannot have body');
|
|
}
|
|
|
|
let inputBody = init.body != null ? init.body : isRequest(input) && input.body !== null ? clone(input) : null;
|
|
|
|
Body.call(this, inputBody, {
|
|
timeout: init.timeout || input.timeout || 0,
|
|
size: init.size || input.size || 0
|
|
});
|
|
|
|
const headers = new Headers(init.headers || input.headers || {});
|
|
|
|
if (inputBody != null && !headers.has('Content-Type')) {
|
|
const contentType = extractContentType(inputBody);
|
|
if (contentType) {
|
|
headers.append('Content-Type', contentType);
|
|
}
|
|
}
|
|
|
|
let signal = isRequest(input) ? input.signal : null;
|
|
if ('signal' in init) signal = init.signal;
|
|
|
|
if (signal != null && !isAbortSignal(signal)) {
|
|
throw new TypeError('Expected signal to be an instanceof AbortSignal');
|
|
}
|
|
|
|
this[INTERNALS$2] = {
|
|
method,
|
|
redirect: init.redirect || input.redirect || 'follow',
|
|
headers,
|
|
parsedURL,
|
|
signal
|
|
};
|
|
|
|
// node-fetch-only options
|
|
this.follow = init.follow !== undefined ? init.follow : input.follow !== undefined ? input.follow : 20;
|
|
this.compress = init.compress !== undefined ? init.compress : input.compress !== undefined ? input.compress : true;
|
|
this.counter = init.counter || input.counter || 0;
|
|
this.agent = init.agent || input.agent;
|
|
}
|
|
|
|
get method() {
|
|
return this[INTERNALS$2].method;
|
|
}
|
|
|
|
get url() {
|
|
return format_url(this[INTERNALS$2].parsedURL);
|
|
}
|
|
|
|
get headers() {
|
|
return this[INTERNALS$2].headers;
|
|
}
|
|
|
|
get redirect() {
|
|
return this[INTERNALS$2].redirect;
|
|
}
|
|
|
|
get signal() {
|
|
return this[INTERNALS$2].signal;
|
|
}
|
|
|
|
/**
|
|
* Clone this request
|
|
*
|
|
* @return Request
|
|
*/
|
|
clone() {
|
|
return new Request(this);
|
|
}
|
|
}
|
|
|
|
Body.mixIn(Request.prototype);
|
|
|
|
Object.defineProperty(Request.prototype, Symbol.toStringTag, {
|
|
value: 'Request',
|
|
writable: false,
|
|
enumerable: false,
|
|
configurable: true
|
|
});
|
|
|
|
Object.defineProperties(Request.prototype, {
|
|
method: { enumerable: true },
|
|
url: { enumerable: true },
|
|
headers: { enumerable: true },
|
|
redirect: { enumerable: true },
|
|
clone: { enumerable: true },
|
|
signal: { enumerable: true }
|
|
});
|
|
|
|
/**
|
|
* Convert a Request to Node.js http request options.
|
|
*
|
|
* @param Request A Request instance
|
|
* @return Object The options object to be passed to http.request
|
|
*/
|
|
function getNodeRequestOptions(request) {
|
|
const parsedURL = request[INTERNALS$2].parsedURL;
|
|
const headers = new Headers(request[INTERNALS$2].headers);
|
|
|
|
// fetch step 1.3
|
|
if (!headers.has('Accept')) {
|
|
headers.set('Accept', '*/*');
|
|
}
|
|
|
|
// Basic fetch
|
|
if (!parsedURL.protocol || !parsedURL.hostname) {
|
|
throw new TypeError('Only absolute URLs are supported');
|
|
}
|
|
|
|
if (!/^https?:$/.test(parsedURL.protocol)) {
|
|
throw new TypeError('Only HTTP(S) protocols are supported');
|
|
}
|
|
|
|
if (request.signal && request.body instanceof Stream.Readable && !streamDestructionSupported) {
|
|
throw new Error('Cancellation of streamed requests with AbortSignal is not supported in node < 8');
|
|
}
|
|
|
|
// HTTP-network-or-cache fetch steps 2.4-2.7
|
|
let contentLengthValue = null;
|
|
if (request.body == null && /^(POST|PUT)$/i.test(request.method)) {
|
|
contentLengthValue = '0';
|
|
}
|
|
if (request.body != null) {
|
|
const totalBytes = getTotalBytes(request);
|
|
if (typeof totalBytes === 'number') {
|
|
contentLengthValue = String(totalBytes);
|
|
}
|
|
}
|
|
if (contentLengthValue) {
|
|
headers.set('Content-Length', contentLengthValue);
|
|
}
|
|
|
|
// HTTP-network-or-cache fetch step 2.11
|
|
if (!headers.has('User-Agent')) {
|
|
headers.set('User-Agent', 'node-fetch/1.0 (+https://github.com/bitinn/node-fetch)');
|
|
}
|
|
|
|
// HTTP-network-or-cache fetch step 2.15
|
|
if (request.compress && !headers.has('Accept-Encoding')) {
|
|
headers.set('Accept-Encoding', 'gzip,deflate');
|
|
}
|
|
|
|
let agent = request.agent;
|
|
if (typeof agent === 'function') {
|
|
agent = agent(parsedURL);
|
|
}
|
|
|
|
if (!headers.has('Connection') && !agent) {
|
|
headers.set('Connection', 'close');
|
|
}
|
|
|
|
// HTTP-network fetch step 4.2
|
|
// chunked encoding is handled by Node.js
|
|
|
|
return Object.assign({}, parsedURL, {
|
|
method: request.method,
|
|
headers: exportNodeCompatibleHeaders(headers),
|
|
agent
|
|
});
|
|
}
|
|
|
|
/**
|
|
* abort-error.js
|
|
*
|
|
* AbortError interface for cancelled requests
|
|
*/
|
|
|
|
/**
|
|
* Create AbortError instance
|
|
*
|
|
* @param String message Error message for human
|
|
* @return AbortError
|
|
*/
|
|
function AbortError(message) {
|
|
Error.call(this, message);
|
|
|
|
this.type = 'aborted';
|
|
this.message = message;
|
|
|
|
// hide custom error implementation details from end-users
|
|
Error.captureStackTrace(this, this.constructor);
|
|
}
|
|
|
|
AbortError.prototype = Object.create(Error.prototype);
|
|
AbortError.prototype.constructor = AbortError;
|
|
AbortError.prototype.name = 'AbortError';
|
|
|
|
// fix an issue where "PassThrough", "resolve" aren't a named export for node <10
|
|
const PassThrough$1 = Stream.PassThrough;
|
|
const resolve_url = Url.resolve;
|
|
|
|
/**
|
|
* Fetch function
|
|
*
|
|
* @param Mixed url Absolute url or Request instance
|
|
* @param Object opts Fetch options
|
|
* @return Promise
|
|
*/
|
|
function fetch(url, opts) {
|
|
|
|
// allow custom promise
|
|
if (!fetch.Promise) {
|
|
throw new Error('native promise missing, set fetch.Promise to your favorite alternative');
|
|
}
|
|
|
|
Body.Promise = fetch.Promise;
|
|
|
|
// wrap http.request into fetch
|
|
return new fetch.Promise(function (resolve, reject) {
|
|
// build request object
|
|
const request = new Request(url, opts);
|
|
const options = getNodeRequestOptions(request);
|
|
|
|
const send = (options.protocol === 'https:' ? https : http).request;
|
|
const signal = request.signal;
|
|
|
|
let response = null;
|
|
|
|
const abort = function abort() {
|
|
let error = new AbortError('The user aborted a request.');
|
|
reject(error);
|
|
if (request.body && request.body instanceof Stream.Readable) {
|
|
request.body.destroy(error);
|
|
}
|
|
if (!response || !response.body) return;
|
|
response.body.emit('error', error);
|
|
};
|
|
|
|
if (signal && signal.aborted) {
|
|
abort();
|
|
return;
|
|
}
|
|
|
|
const abortAndFinalize = function abortAndFinalize() {
|
|
abort();
|
|
finalize();
|
|
};
|
|
|
|
// send request
|
|
const req = send(options);
|
|
let reqTimeout;
|
|
|
|
if (signal) {
|
|
signal.addEventListener('abort', abortAndFinalize);
|
|
}
|
|
|
|
function finalize() {
|
|
req.abort();
|
|
if (signal) signal.removeEventListener('abort', abortAndFinalize);
|
|
clearTimeout(reqTimeout);
|
|
}
|
|
|
|
if (request.timeout) {
|
|
req.once('socket', function (socket) {
|
|
reqTimeout = setTimeout(function () {
|
|
reject(new FetchError(`network timeout at: ${request.url}`, 'request-timeout'));
|
|
finalize();
|
|
}, request.timeout);
|
|
});
|
|
}
|
|
|
|
req.on('error', function (err) {
|
|
reject(new FetchError(`request to ${request.url} failed, reason: ${err.message}`, 'system', err));
|
|
finalize();
|
|
});
|
|
|
|
req.on('response', function (res) {
|
|
clearTimeout(reqTimeout);
|
|
|
|
const headers = createHeadersLenient(res.headers);
|
|
|
|
// HTTP fetch step 5
|
|
if (fetch.isRedirect(res.statusCode)) {
|
|
// HTTP fetch step 5.2
|
|
const location = headers.get('Location');
|
|
|
|
// HTTP fetch step 5.3
|
|
const locationURL = location === null ? null : resolve_url(request.url, location);
|
|
|
|
// HTTP fetch step 5.5
|
|
switch (request.redirect) {
|
|
case 'error':
|
|
reject(new FetchError(`uri requested responds with a redirect, redirect mode is set to error: ${request.url}`, 'no-redirect'));
|
|
finalize();
|
|
return;
|
|
case 'manual':
|
|
// node-fetch-specific step: make manual redirect a bit easier to use by setting the Location header value to the resolved URL.
|
|
if (locationURL !== null) {
|
|
// handle corrupted header
|
|
try {
|
|
headers.set('Location', locationURL);
|
|
} catch (err) {
|
|
// istanbul ignore next: nodejs server prevent invalid response headers, we can't test this through normal request
|
|
reject(err);
|
|
}
|
|
}
|
|
break;
|
|
case 'follow':
|
|
// HTTP-redirect fetch step 2
|
|
if (locationURL === null) {
|
|
break;
|
|
}
|
|
|
|
// HTTP-redirect fetch step 5
|
|
if (request.counter >= request.follow) {
|
|
reject(new FetchError(`maximum redirect reached at: ${request.url}`, 'max-redirect'));
|
|
finalize();
|
|
return;
|
|
}
|
|
|
|
// HTTP-redirect fetch step 6 (counter increment)
|
|
// Create a new Request object.
|
|
const requestOpts = {
|
|
headers: new Headers(request.headers),
|
|
follow: request.follow,
|
|
counter: request.counter + 1,
|
|
agent: request.agent,
|
|
compress: request.compress,
|
|
method: request.method,
|
|
body: request.body,
|
|
signal: request.signal,
|
|
timeout: request.timeout,
|
|
size: request.size
|
|
};
|
|
|
|
// HTTP-redirect fetch step 9
|
|
if (res.statusCode !== 303 && request.body && getTotalBytes(request) === null) {
|
|
reject(new FetchError('Cannot follow redirect with body being a readable stream', 'unsupported-redirect'));
|
|
finalize();
|
|
return;
|
|
}
|
|
|
|
// HTTP-redirect fetch step 11
|
|
if (res.statusCode === 303 || (res.statusCode === 301 || res.statusCode === 302) && request.method === 'POST') {
|
|
requestOpts.method = 'GET';
|
|
requestOpts.body = undefined;
|
|
requestOpts.headers.delete('content-length');
|
|
}
|
|
|
|
// HTTP-redirect fetch step 15
|
|
resolve(fetch(new Request(locationURL, requestOpts)));
|
|
finalize();
|
|
return;
|
|
}
|
|
}
|
|
|
|
// prepare response
|
|
res.once('end', function () {
|
|
if (signal) signal.removeEventListener('abort', abortAndFinalize);
|
|
});
|
|
let body = res.pipe(new PassThrough$1());
|
|
|
|
const response_options = {
|
|
url: request.url,
|
|
status: res.statusCode,
|
|
statusText: res.statusMessage,
|
|
headers: headers,
|
|
size: request.size,
|
|
timeout: request.timeout,
|
|
counter: request.counter
|
|
};
|
|
|
|
// HTTP-network fetch step 12.1.1.3
|
|
const codings = headers.get('Content-Encoding');
|
|
|
|
// HTTP-network fetch step 12.1.1.4: handle content codings
|
|
|
|
// in following scenarios we ignore compression support
|
|
// 1. compression support is disabled
|
|
// 2. HEAD request
|
|
// 3. no Content-Encoding header
|
|
// 4. no content response (204)
|
|
// 5. content not modified response (304)
|
|
if (!request.compress || request.method === 'HEAD' || codings === null || res.statusCode === 204 || res.statusCode === 304) {
|
|
response = new Response(body, response_options);
|
|
resolve(response);
|
|
return;
|
|
}
|
|
|
|
// For Node v6+
|
|
// Be less strict when decoding compressed responses, since sometimes
|
|
// servers send slightly invalid responses that are still accepted
|
|
// by common browsers.
|
|
// Always using Z_SYNC_FLUSH is what cURL does.
|
|
const zlibOptions = {
|
|
flush: zlib.Z_SYNC_FLUSH,
|
|
finishFlush: zlib.Z_SYNC_FLUSH
|
|
};
|
|
|
|
// for gzip
|
|
if (codings == 'gzip' || codings == 'x-gzip') {
|
|
body = body.pipe(zlib.createGunzip(zlibOptions));
|
|
response = new Response(body, response_options);
|
|
resolve(response);
|
|
return;
|
|
}
|
|
|
|
// for deflate
|
|
if (codings == 'deflate' || codings == 'x-deflate') {
|
|
// handle the infamous raw deflate response from old servers
|
|
// a hack for old IIS and Apache servers
|
|
const raw = res.pipe(new PassThrough$1());
|
|
raw.once('data', function (chunk) {
|
|
// see http://stackoverflow.com/questions/37519828
|
|
if ((chunk[0] & 0x0F) === 0x08) {
|
|
body = body.pipe(zlib.createInflate());
|
|
} else {
|
|
body = body.pipe(zlib.createInflateRaw());
|
|
}
|
|
response = new Response(body, response_options);
|
|
resolve(response);
|
|
});
|
|
return;
|
|
}
|
|
|
|
// for br
|
|
if (codings == 'br' && typeof zlib.createBrotliDecompress === 'function') {
|
|
body = body.pipe(zlib.createBrotliDecompress());
|
|
response = new Response(body, response_options);
|
|
resolve(response);
|
|
return;
|
|
}
|
|
|
|
// otherwise, use response as-is
|
|
response = new Response(body, response_options);
|
|
resolve(response);
|
|
});
|
|
|
|
writeToStream(req, request);
|
|
});
|
|
}
|
|
/**
|
|
* Redirect code matching
|
|
*
|
|
* @param Number code Status code
|
|
* @return Boolean
|
|
*/
|
|
fetch.isRedirect = function (code) {
|
|
return code === 301 || code === 302 || code === 303 || code === 307 || code === 308;
|
|
};
|
|
|
|
// expose Promise
|
|
fetch.Promise = global.Promise;
|
|
|
|
module.exports = exports = fetch;
|
|
Object.defineProperty(exports, "__esModule", ({ value: true }));
|
|
exports.default = exports;
|
|
exports.Headers = Headers;
|
|
exports.Request = Request;
|
|
exports.Response = Response;
|
|
exports.FetchError = FetchError;
|
|
|
|
|
|
/***/ }),
|
|
|
|
/***/ 1223:
|
|
/***/ ((module, __unused_webpack_exports, __nccwpck_require__) => {
|
|
|
|
var wrappy = __nccwpck_require__(2940)
|
|
module.exports = wrappy(once)
|
|
module.exports.strict = wrappy(onceStrict)
|
|
|
|
once.proto = once(function () {
|
|
Object.defineProperty(Function.prototype, 'once', {
|
|
value: function () {
|
|
return once(this)
|
|
},
|
|
configurable: true
|
|
})
|
|
|
|
Object.defineProperty(Function.prototype, 'onceStrict', {
|
|
value: function () {
|
|
return onceStrict(this)
|
|
},
|
|
configurable: true
|
|
})
|
|
})
|
|
|
|
function once (fn) {
|
|
var f = function () {
|
|
if (f.called) return f.value
|
|
f.called = true
|
|
return f.value = fn.apply(this, arguments)
|
|
}
|
|
f.called = false
|
|
return f
|
|
}
|
|
|
|
function onceStrict (fn) {
|
|
var f = function () {
|
|
if (f.called)
|
|
throw new Error(f.onceError)
|
|
f.called = true
|
|
return f.value = fn.apply(this, arguments)
|
|
}
|
|
var name = fn.name || 'Function wrapped with `once`'
|
|
f.onceError = name + " shouldn't be called more than once"
|
|
f.called = false
|
|
return f
|
|
}
|
|
|
|
|
|
/***/ }),
|
|
|
|
/***/ 2522:
|
|
/***/ (function(module) {
|
|
|
|
/* global define */
|
|
|
|
(function (root, pluralize) {
|
|
/* istanbul ignore else */
|
|
if (true) {
|
|
// Node.
|
|
module.exports = pluralize();
|
|
} else {}
|
|
})(this, function () {
|
|
// Rule storage - pluralize and singularize need to be run sequentially,
|
|
// while other rules can be optimized using an object for instant lookups.
|
|
var pluralRules = [];
|
|
var singularRules = [];
|
|
var uncountables = {};
|
|
var irregularPlurals = {};
|
|
var irregularSingles = {};
|
|
|
|
/**
|
|
* Sanitize a pluralization rule to a usable regular expression.
|
|
*
|
|
* @param {(RegExp|string)} rule
|
|
* @return {RegExp}
|
|
*/
|
|
function sanitizeRule (rule) {
|
|
if (typeof rule === 'string') {
|
|
return new RegExp('^' + rule + '$', 'i');
|
|
}
|
|
|
|
return rule;
|
|
}
|
|
|
|
/**
|
|
* Pass in a word token to produce a function that can replicate the case on
|
|
* another word.
|
|
*
|
|
* @param {string} word
|
|
* @param {string} token
|
|
* @return {Function}
|
|
*/
|
|
function restoreCase (word, token) {
|
|
// Tokens are an exact match.
|
|
if (word === token) return token;
|
|
|
|
// Lower cased words. E.g. "hello".
|
|
if (word === word.toLowerCase()) return token.toLowerCase();
|
|
|
|
// Upper cased words. E.g. "WHISKY".
|
|
if (word === word.toUpperCase()) return token.toUpperCase();
|
|
|
|
// Title cased words. E.g. "Title".
|
|
if (word[0] === word[0].toUpperCase()) {
|
|
return token.charAt(0).toUpperCase() + token.substr(1).toLowerCase();
|
|
}
|
|
|
|
// Lower cased words. E.g. "test".
|
|
return token.toLowerCase();
|
|
}
|
|
|
|
/**
|
|
* Interpolate a regexp string.
|
|
*
|
|
* @param {string} str
|
|
* @param {Array} args
|
|
* @return {string}
|
|
*/
|
|
function interpolate (str, args) {
|
|
return str.replace(/\$(\d{1,2})/g, function (match, index) {
|
|
return args[index] || '';
|
|
});
|
|
}
|
|
|
|
/**
|
|
* Replace a word using a rule.
|
|
*
|
|
* @param {string} word
|
|
* @param {Array} rule
|
|
* @return {string}
|
|
*/
|
|
function replace (word, rule) {
|
|
return word.replace(rule[0], function (match, index) {
|
|
var result = interpolate(rule[1], arguments);
|
|
|
|
if (match === '') {
|
|
return restoreCase(word[index - 1], result);
|
|
}
|
|
|
|
return restoreCase(match, result);
|
|
});
|
|
}
|
|
|
|
/**
|
|
* Sanitize a word by passing in the word and sanitization rules.
|
|
*
|
|
* @param {string} token
|
|
* @param {string} word
|
|
* @param {Array} rules
|
|
* @return {string}
|
|
*/
|
|
function sanitizeWord (token, word, rules) {
|
|
// Empty string or doesn't need fixing.
|
|
if (!token.length || uncountables.hasOwnProperty(token)) {
|
|
return word;
|
|
}
|
|
|
|
var len = rules.length;
|
|
|
|
// Iterate over the sanitization rules and use the first one to match.
|
|
while (len--) {
|
|
var rule = rules[len];
|
|
|
|
if (rule[0].test(word)) return replace(word, rule);
|
|
}
|
|
|
|
return word;
|
|
}
|
|
|
|
/**
|
|
* Replace a word with the updated word.
|
|
*
|
|
* @param {Object} replaceMap
|
|
* @param {Object} keepMap
|
|
* @param {Array} rules
|
|
* @return {Function}
|
|
*/
|
|
function replaceWord (replaceMap, keepMap, rules) {
|
|
return function (word) {
|
|
// Get the correct token and case restoration functions.
|
|
var token = word.toLowerCase();
|
|
|
|
// Check against the keep object map.
|
|
if (keepMap.hasOwnProperty(token)) {
|
|
return restoreCase(word, token);
|
|
}
|
|
|
|
// Check against the replacement map for a direct word replacement.
|
|
if (replaceMap.hasOwnProperty(token)) {
|
|
return restoreCase(word, replaceMap[token]);
|
|
}
|
|
|
|
// Run all the rules against the word.
|
|
return sanitizeWord(token, word, rules);
|
|
};
|
|
}
|
|
|
|
/**
|
|
* Check if a word is part of the map.
|
|
*/
|
|
function checkWord (replaceMap, keepMap, rules, bool) {
|
|
return function (word) {
|
|
var token = word.toLowerCase();
|
|
|
|
if (keepMap.hasOwnProperty(token)) return true;
|
|
if (replaceMap.hasOwnProperty(token)) return false;
|
|
|
|
return sanitizeWord(token, token, rules) === token;
|
|
};
|
|
}
|
|
|
|
/**
|
|
* Pluralize or singularize a word based on the passed in count.
|
|
*
|
|
* @param {string} word The word to pluralize
|
|
* @param {number} count How many of the word exist
|
|
* @param {boolean} inclusive Whether to prefix with the number (e.g. 3 ducks)
|
|
* @return {string}
|
|
*/
|
|
function pluralize (word, count, inclusive) {
|
|
var pluralized = count === 1
|
|
? pluralize.singular(word) : pluralize.plural(word);
|
|
|
|
return (inclusive ? count + ' ' : '') + pluralized;
|
|
}
|
|
|
|
/**
|
|
* Pluralize a word.
|
|
*
|
|
* @type {Function}
|
|
*/
|
|
pluralize.plural = replaceWord(
|
|
irregularSingles, irregularPlurals, pluralRules
|
|
);
|
|
|
|
/**
|
|
* Check if a word is plural.
|
|
*
|
|
* @type {Function}
|
|
*/
|
|
pluralize.isPlural = checkWord(
|
|
irregularSingles, irregularPlurals, pluralRules
|
|
);
|
|
|
|
/**
|
|
* Singularize a word.
|
|
*
|
|
* @type {Function}
|
|
*/
|
|
pluralize.singular = replaceWord(
|
|
irregularPlurals, irregularSingles, singularRules
|
|
);
|
|
|
|
/**
|
|
* Check if a word is singular.
|
|
*
|
|
* @type {Function}
|
|
*/
|
|
pluralize.isSingular = checkWord(
|
|
irregularPlurals, irregularSingles, singularRules
|
|
);
|
|
|
|
/**
|
|
* Add a pluralization rule to the collection.
|
|
*
|
|
* @param {(string|RegExp)} rule
|
|
* @param {string} replacement
|
|
*/
|
|
pluralize.addPluralRule = function (rule, replacement) {
|
|
pluralRules.push([sanitizeRule(rule), replacement]);
|
|
};
|
|
|
|
/**
|
|
* Add a singularization rule to the collection.
|
|
*
|
|
* @param {(string|RegExp)} rule
|
|
* @param {string} replacement
|
|
*/
|
|
pluralize.addSingularRule = function (rule, replacement) {
|
|
singularRules.push([sanitizeRule(rule), replacement]);
|
|
};
|
|
|
|
/**
|
|
* Add an uncountable word rule.
|
|
*
|
|
* @param {(string|RegExp)} word
|
|
*/
|
|
pluralize.addUncountableRule = function (word) {
|
|
if (typeof word === 'string') {
|
|
uncountables[word.toLowerCase()] = true;
|
|
return;
|
|
}
|
|
|
|
// Set singular and plural references for the word.
|
|
pluralize.addPluralRule(word, '$0');
|
|
pluralize.addSingularRule(word, '$0');
|
|
};
|
|
|
|
/**
|
|
* Add an irregular word definition.
|
|
*
|
|
* @param {string} single
|
|
* @param {string} plural
|
|
*/
|
|
pluralize.addIrregularRule = function (single, plural) {
|
|
plural = plural.toLowerCase();
|
|
single = single.toLowerCase();
|
|
|
|
irregularSingles[single] = plural;
|
|
irregularPlurals[plural] = single;
|
|
};
|
|
|
|
/**
|
|
* Irregular rules.
|
|
*/
|
|
[
|
|
// Pronouns.
|
|
['I', 'we'],
|
|
['me', 'us'],
|
|
['he', 'they'],
|
|
['she', 'they'],
|
|
['them', 'them'],
|
|
['myself', 'ourselves'],
|
|
['yourself', 'yourselves'],
|
|
['itself', 'themselves'],
|
|
['herself', 'themselves'],
|
|
['himself', 'themselves'],
|
|
['themself', 'themselves'],
|
|
['is', 'are'],
|
|
['was', 'were'],
|
|
['has', 'have'],
|
|
['this', 'these'],
|
|
['that', 'those'],
|
|
// Words ending in with a consonant and `o`.
|
|
['echo', 'echoes'],
|
|
['dingo', 'dingoes'],
|
|
['volcano', 'volcanoes'],
|
|
['tornado', 'tornadoes'],
|
|
['torpedo', 'torpedoes'],
|
|
// Ends with `us`.
|
|
['genus', 'genera'],
|
|
['viscus', 'viscera'],
|
|
// Ends with `ma`.
|
|
['stigma', 'stigmata'],
|
|
['stoma', 'stomata'],
|
|
['dogma', 'dogmata'],
|
|
['lemma', 'lemmata'],
|
|
['schema', 'schemata'],
|
|
['anathema', 'anathemata'],
|
|
// Other irregular rules.
|
|
['ox', 'oxen'],
|
|
['axe', 'axes'],
|
|
['die', 'dice'],
|
|
['yes', 'yeses'],
|
|
['foot', 'feet'],
|
|
['eave', 'eaves'],
|
|
['goose', 'geese'],
|
|
['tooth', 'teeth'],
|
|
['quiz', 'quizzes'],
|
|
['human', 'humans'],
|
|
['proof', 'proofs'],
|
|
['carve', 'carves'],
|
|
['valve', 'valves'],
|
|
['looey', 'looies'],
|
|
['thief', 'thieves'],
|
|
['groove', 'grooves'],
|
|
['pickaxe', 'pickaxes'],
|
|
['passerby', 'passersby']
|
|
].forEach(function (rule) {
|
|
return pluralize.addIrregularRule(rule[0], rule[1]);
|
|
});
|
|
|
|
/**
|
|
* Pluralization rules.
|
|
*/
|
|
[
|
|
[/s?$/i, 's'],
|
|
[/[^\u0000-\u007F]$/i, '$0'],
|
|
[/([^aeiou]ese)$/i, '$1'],
|
|
[/(ax|test)is$/i, '$1es'],
|
|
[/(alias|[^aou]us|t[lm]as|gas|ris)$/i, '$1es'],
|
|
[/(e[mn]u)s?$/i, '$1s'],
|
|
[/([^l]ias|[aeiou]las|[ejzr]as|[iu]am)$/i, '$1'],
|
|
[/(alumn|syllab|vir|radi|nucle|fung|cact|stimul|termin|bacill|foc|uter|loc|strat)(?:us|i)$/i, '$1i'],
|
|
[/(alumn|alg|vertebr)(?:a|ae)$/i, '$1ae'],
|
|
[/(seraph|cherub)(?:im)?$/i, '$1im'],
|
|
[/(her|at|gr)o$/i, '$1oes'],
|
|
[/(agend|addend|millenni|dat|extrem|bacteri|desiderat|strat|candelabr|errat|ov|symposi|curricul|automat|quor)(?:a|um)$/i, '$1a'],
|
|
[/(apheli|hyperbat|periheli|asyndet|noumen|phenomen|criteri|organ|prolegomen|hedr|automat)(?:a|on)$/i, '$1a'],
|
|
[/sis$/i, 'ses'],
|
|
[/(?:(kni|wi|li)fe|(ar|l|ea|eo|oa|hoo)f)$/i, '$1$2ves'],
|
|
[/([^aeiouy]|qu)y$/i, '$1ies'],
|
|
[/([^ch][ieo][ln])ey$/i, '$1ies'],
|
|
[/(x|ch|ss|sh|zz)$/i, '$1es'],
|
|
[/(matr|cod|mur|sil|vert|ind|append)(?:ix|ex)$/i, '$1ices'],
|
|
[/\b((?:tit)?m|l)(?:ice|ouse)$/i, '$1ice'],
|
|
[/(pe)(?:rson|ople)$/i, '$1ople'],
|
|
[/(child)(?:ren)?$/i, '$1ren'],
|
|
[/eaux$/i, '$0'],
|
|
[/m[ae]n$/i, 'men'],
|
|
['thou', 'you']
|
|
].forEach(function (rule) {
|
|
return pluralize.addPluralRule(rule[0], rule[1]);
|
|
});
|
|
|
|
/**
|
|
* Singularization rules.
|
|
*/
|
|
[
|
|
[/s$/i, ''],
|
|
[/(ss)$/i, '$1'],
|
|
[/(wi|kni|(?:after|half|high|low|mid|non|night|[^\w]|^)li)ves$/i, '$1fe'],
|
|
[/(ar|(?:wo|[ae])l|[eo][ao])ves$/i, '$1f'],
|
|
[/ies$/i, 'y'],
|
|
[/\b([pl]|zomb|(?:neck|cross)?t|coll|faer|food|gen|goon|group|lass|talk|goal|cut)ies$/i, '$1ie'],
|
|
[/\b(mon|smil)ies$/i, '$1ey'],
|
|
[/\b((?:tit)?m|l)ice$/i, '$1ouse'],
|
|
[/(seraph|cherub)im$/i, '$1'],
|
|
[/(x|ch|ss|sh|zz|tto|go|cho|alias|[^aou]us|t[lm]as|gas|(?:her|at|gr)o|[aeiou]ris)(?:es)?$/i, '$1'],
|
|
[/(analy|diagno|parenthe|progno|synop|the|empha|cri|ne)(?:sis|ses)$/i, '$1sis'],
|
|
[/(movie|twelve|abuse|e[mn]u)s$/i, '$1'],
|
|
[/(test)(?:is|es)$/i, '$1is'],
|
|
[/(alumn|syllab|vir|radi|nucle|fung|cact|stimul|termin|bacill|foc|uter|loc|strat)(?:us|i)$/i, '$1us'],
|
|
[/(agend|addend|millenni|dat|extrem|bacteri|desiderat|strat|candelabr|errat|ov|symposi|curricul|quor)a$/i, '$1um'],
|
|
[/(apheli|hyperbat|periheli|asyndet|noumen|phenomen|criteri|organ|prolegomen|hedr|automat)a$/i, '$1on'],
|
|
[/(alumn|alg|vertebr)ae$/i, '$1a'],
|
|
[/(cod|mur|sil|vert|ind)ices$/i, '$1ex'],
|
|
[/(matr|append)ices$/i, '$1ix'],
|
|
[/(pe)(rson|ople)$/i, '$1rson'],
|
|
[/(child)ren$/i, '$1'],
|
|
[/(eau)x?$/i, '$1'],
|
|
[/men$/i, 'man']
|
|
].forEach(function (rule) {
|
|
return pluralize.addSingularRule(rule[0], rule[1]);
|
|
});
|
|
|
|
/**
|
|
* Uncountable rules.
|
|
*/
|
|
[
|
|
// Singular words with no plurals.
|
|
'adulthood',
|
|
'advice',
|
|
'agenda',
|
|
'aid',
|
|
'aircraft',
|
|
'alcohol',
|
|
'ammo',
|
|
'analytics',
|
|
'anime',
|
|
'athletics',
|
|
'audio',
|
|
'bison',
|
|
'blood',
|
|
'bream',
|
|
'buffalo',
|
|
'butter',
|
|
'carp',
|
|
'cash',
|
|
'chassis',
|
|
'chess',
|
|
'clothing',
|
|
'cod',
|
|
'commerce',
|
|
'cooperation',
|
|
'corps',
|
|
'debris',
|
|
'diabetes',
|
|
'digestion',
|
|
'elk',
|
|
'energy',
|
|
'equipment',
|
|
'excretion',
|
|
'expertise',
|
|
'firmware',
|
|
'flounder',
|
|
'fun',
|
|
'gallows',
|
|
'garbage',
|
|
'graffiti',
|
|
'hardware',
|
|
'headquarters',
|
|
'health',
|
|
'herpes',
|
|
'highjinks',
|
|
'homework',
|
|
'housework',
|
|
'information',
|
|
'jeans',
|
|
'justice',
|
|
'kudos',
|
|
'labour',
|
|
'literature',
|
|
'machinery',
|
|
'mackerel',
|
|
'mail',
|
|
'media',
|
|
'mews',
|
|
'moose',
|
|
'music',
|
|
'mud',
|
|
'manga',
|
|
'news',
|
|
'only',
|
|
'personnel',
|
|
'pike',
|
|
'plankton',
|
|
'pliers',
|
|
'police',
|
|
'pollution',
|
|
'premises',
|
|
'rain',
|
|
'research',
|
|
'rice',
|
|
'salmon',
|
|
'scissors',
|
|
'series',
|
|
'sewage',
|
|
'shambles',
|
|
'shrimp',
|
|
'software',
|
|
'species',
|
|
'staff',
|
|
'swine',
|
|
'tennis',
|
|
'traffic',
|
|
'transportation',
|
|
'trout',
|
|
'tuna',
|
|
'wealth',
|
|
'welfare',
|
|
'whiting',
|
|
'wildebeest',
|
|
'wildlife',
|
|
'you',
|
|
/pok[eé]mon$/i,
|
|
// Regexes.
|
|
/[^aeiou]ese$/i, // "chinese", "japanese"
|
|
/deer$/i, // "deer", "reindeer"
|
|
/fish$/i, // "fish", "blowfish", "angelfish"
|
|
/measles$/i,
|
|
/o[iu]s$/i, // "carnivorous"
|
|
/pox$/i, // "chickpox", "smallpox"
|
|
/sheep$/i
|
|
].forEach(pluralize.addUncountableRule);
|
|
|
|
return pluralize;
|
|
});
|
|
|
|
|
|
/***/ }),
|
|
|
|
/***/ 4294:
|
|
/***/ ((module, __unused_webpack_exports, __nccwpck_require__) => {
|
|
|
|
module.exports = __nccwpck_require__(4219);
|
|
|
|
|
|
/***/ }),
|
|
|
|
/***/ 4219:
|
|
/***/ ((__unused_webpack_module, exports, __nccwpck_require__) => {
|
|
|
|
"use strict";
|
|
|
|
|
|
var net = __nccwpck_require__(1631);
|
|
var tls = __nccwpck_require__(4016);
|
|
var http = __nccwpck_require__(8605);
|
|
var https = __nccwpck_require__(7211);
|
|
var events = __nccwpck_require__(8614);
|
|
var assert = __nccwpck_require__(2357);
|
|
var util = __nccwpck_require__(1669);
|
|
|
|
|
|
exports.httpOverHttp = httpOverHttp;
|
|
exports.httpsOverHttp = httpsOverHttp;
|
|
exports.httpOverHttps = httpOverHttps;
|
|
exports.httpsOverHttps = httpsOverHttps;
|
|
|
|
|
|
function httpOverHttp(options) {
|
|
var agent = new TunnelingAgent(options);
|
|
agent.request = http.request;
|
|
return agent;
|
|
}
|
|
|
|
function httpsOverHttp(options) {
|
|
var agent = new TunnelingAgent(options);
|
|
agent.request = http.request;
|
|
agent.createSocket = createSecureSocket;
|
|
agent.defaultPort = 443;
|
|
return agent;
|
|
}
|
|
|
|
function httpOverHttps(options) {
|
|
var agent = new TunnelingAgent(options);
|
|
agent.request = https.request;
|
|
return agent;
|
|
}
|
|
|
|
function httpsOverHttps(options) {
|
|
var agent = new TunnelingAgent(options);
|
|
agent.request = https.request;
|
|
agent.createSocket = createSecureSocket;
|
|
agent.defaultPort = 443;
|
|
return agent;
|
|
}
|
|
|
|
|
|
function TunnelingAgent(options) {
|
|
var self = this;
|
|
self.options = options || {};
|
|
self.proxyOptions = self.options.proxy || {};
|
|
self.maxSockets = self.options.maxSockets || http.Agent.defaultMaxSockets;
|
|
self.requests = [];
|
|
self.sockets = [];
|
|
|
|
self.on('free', function onFree(socket, host, port, localAddress) {
|
|
var options = toOptions(host, port, localAddress);
|
|
for (var i = 0, len = self.requests.length; i < len; ++i) {
|
|
var pending = self.requests[i];
|
|
if (pending.host === options.host && pending.port === options.port) {
|
|
// Detect the request to connect same origin server,
|
|
// reuse the connection.
|
|
self.requests.splice(i, 1);
|
|
pending.request.onSocket(socket);
|
|
return;
|
|
}
|
|
}
|
|
socket.destroy();
|
|
self.removeSocket(socket);
|
|
});
|
|
}
|
|
util.inherits(TunnelingAgent, events.EventEmitter);
|
|
|
|
TunnelingAgent.prototype.addRequest = function addRequest(req, host, port, localAddress) {
|
|
var self = this;
|
|
var options = mergeOptions({request: req}, self.options, toOptions(host, port, localAddress));
|
|
|
|
if (self.sockets.length >= this.maxSockets) {
|
|
// We are over limit so we'll add it to the queue.
|
|
self.requests.push(options);
|
|
return;
|
|
}
|
|
|
|
// If we are under maxSockets create a new one.
|
|
self.createSocket(options, function(socket) {
|
|
socket.on('free', onFree);
|
|
socket.on('close', onCloseOrRemove);
|
|
socket.on('agentRemove', onCloseOrRemove);
|
|
req.onSocket(socket);
|
|
|
|
function onFree() {
|
|
self.emit('free', socket, options);
|
|
}
|
|
|
|
function onCloseOrRemove(err) {
|
|
self.removeSocket(socket);
|
|
socket.removeListener('free', onFree);
|
|
socket.removeListener('close', onCloseOrRemove);
|
|
socket.removeListener('agentRemove', onCloseOrRemove);
|
|
}
|
|
});
|
|
};
|
|
|
|
TunnelingAgent.prototype.createSocket = function createSocket(options, cb) {
|
|
var self = this;
|
|
var placeholder = {};
|
|
self.sockets.push(placeholder);
|
|
|
|
var connectOptions = mergeOptions({}, self.proxyOptions, {
|
|
method: 'CONNECT',
|
|
path: options.host + ':' + options.port,
|
|
agent: false,
|
|
headers: {
|
|
host: options.host + ':' + options.port
|
|
}
|
|
});
|
|
if (options.localAddress) {
|
|
connectOptions.localAddress = options.localAddress;
|
|
}
|
|
if (connectOptions.proxyAuth) {
|
|
connectOptions.headers = connectOptions.headers || {};
|
|
connectOptions.headers['Proxy-Authorization'] = 'Basic ' +
|
|
new Buffer(connectOptions.proxyAuth).toString('base64');
|
|
}
|
|
|
|
debug('making CONNECT request');
|
|
var connectReq = self.request(connectOptions);
|
|
connectReq.useChunkedEncodingByDefault = false; // for v0.6
|
|
connectReq.once('response', onResponse); // for v0.6
|
|
connectReq.once('upgrade', onUpgrade); // for v0.6
|
|
connectReq.once('connect', onConnect); // for v0.7 or later
|
|
connectReq.once('error', onError);
|
|
connectReq.end();
|
|
|
|
function onResponse(res) {
|
|
// Very hacky. This is necessary to avoid http-parser leaks.
|
|
res.upgrade = true;
|
|
}
|
|
|
|
function onUpgrade(res, socket, head) {
|
|
// Hacky.
|
|
process.nextTick(function() {
|
|
onConnect(res, socket, head);
|
|
});
|
|
}
|
|
|
|
function onConnect(res, socket, head) {
|
|
connectReq.removeAllListeners();
|
|
socket.removeAllListeners();
|
|
|
|
if (res.statusCode !== 200) {
|
|
debug('tunneling socket could not be established, statusCode=%d',
|
|
res.statusCode);
|
|
socket.destroy();
|
|
var error = new Error('tunneling socket could not be established, ' +
|
|
'statusCode=' + res.statusCode);
|
|
error.code = 'ECONNRESET';
|
|
options.request.emit('error', error);
|
|
self.removeSocket(placeholder);
|
|
return;
|
|
}
|
|
if (head.length > 0) {
|
|
debug('got illegal response body from proxy');
|
|
socket.destroy();
|
|
var error = new Error('got illegal response body from proxy');
|
|
error.code = 'ECONNRESET';
|
|
options.request.emit('error', error);
|
|
self.removeSocket(placeholder);
|
|
return;
|
|
}
|
|
debug('tunneling connection has established');
|
|
self.sockets[self.sockets.indexOf(placeholder)] = socket;
|
|
return cb(socket);
|
|
}
|
|
|
|
function onError(cause) {
|
|
connectReq.removeAllListeners();
|
|
|
|
debug('tunneling socket could not be established, cause=%s\n',
|
|
cause.message, cause.stack);
|
|
var error = new Error('tunneling socket could not be established, ' +
|
|
'cause=' + cause.message);
|
|
error.code = 'ECONNRESET';
|
|
options.request.emit('error', error);
|
|
self.removeSocket(placeholder);
|
|
}
|
|
};
|
|
|
|
TunnelingAgent.prototype.removeSocket = function removeSocket(socket) {
|
|
var pos = this.sockets.indexOf(socket)
|
|
if (pos === -1) {
|
|
return;
|
|
}
|
|
this.sockets.splice(pos, 1);
|
|
|
|
var pending = this.requests.shift();
|
|
if (pending) {
|
|
// If we have pending requests and a socket gets closed a new one
|
|
// needs to be created to take over in the pool for the one that closed.
|
|
this.createSocket(pending, function(socket) {
|
|
pending.request.onSocket(socket);
|
|
});
|
|
}
|
|
};
|
|
|
|
function createSecureSocket(options, cb) {
|
|
var self = this;
|
|
TunnelingAgent.prototype.createSocket.call(self, options, function(socket) {
|
|
var hostHeader = options.request.getHeader('host');
|
|
var tlsOptions = mergeOptions({}, self.options, {
|
|
socket: socket,
|
|
servername: hostHeader ? hostHeader.replace(/:.*$/, '') : options.host
|
|
});
|
|
|
|
// 0 is dummy port for v0.6
|
|
var secureSocket = tls.connect(0, tlsOptions);
|
|
self.sockets[self.sockets.indexOf(socket)] = secureSocket;
|
|
cb(secureSocket);
|
|
});
|
|
}
|
|
|
|
|
|
function toOptions(host, port, localAddress) {
|
|
if (typeof host === 'string') { // since v0.10
|
|
return {
|
|
host: host,
|
|
port: port,
|
|
localAddress: localAddress
|
|
};
|
|
}
|
|
return host; // for v0.11 or later
|
|
}
|
|
|
|
function mergeOptions(target) {
|
|
for (var i = 1, len = arguments.length; i < len; ++i) {
|
|
var overrides = arguments[i];
|
|
if (typeof overrides === 'object') {
|
|
var keys = Object.keys(overrides);
|
|
for (var j = 0, keyLen = keys.length; j < keyLen; ++j) {
|
|
var k = keys[j];
|
|
if (overrides[k] !== undefined) {
|
|
target[k] = overrides[k];
|
|
}
|
|
}
|
|
}
|
|
}
|
|
return target;
|
|
}
|
|
|
|
|
|
var debug;
|
|
if (process.env.NODE_DEBUG && /\btunnel\b/.test(process.env.NODE_DEBUG)) {
|
|
debug = function() {
|
|
var args = Array.prototype.slice.call(arguments);
|
|
if (typeof args[0] === 'string') {
|
|
args[0] = 'TUNNEL: ' + args[0];
|
|
} else {
|
|
args.unshift('TUNNEL:');
|
|
}
|
|
console.error.apply(console, args);
|
|
}
|
|
} else {
|
|
debug = function() {};
|
|
}
|
|
exports.debug = debug; // for test
|
|
|
|
|
|
/***/ }),
|
|
|
|
/***/ 5030:
|
|
/***/ ((__unused_webpack_module, exports) => {
|
|
|
|
"use strict";
|
|
|
|
|
|
Object.defineProperty(exports, "__esModule", ({ value: true }));
|
|
|
|
function getUserAgent() {
|
|
if (typeof navigator === "object" && "userAgent" in navigator) {
|
|
return navigator.userAgent;
|
|
}
|
|
|
|
if (typeof process === "object" && "version" in process) {
|
|
return `Node.js/${process.version.substr(1)} (${process.platform}; ${process.arch})`;
|
|
}
|
|
|
|
return "<environment undetectable>";
|
|
}
|
|
|
|
exports.getUserAgent = getUserAgent;
|
|
//# sourceMappingURL=index.js.map
|
|
|
|
|
|
/***/ }),
|
|
|
|
/***/ 2940:
|
|
/***/ ((module) => {
|
|
|
|
// Returns a wrapper function that returns a wrapped callback
|
|
// The wrapper function should do some stuff, and return a
|
|
// presumably different callback function.
|
|
// This makes sure that own properties are retained, so that
|
|
// decorations and such are not lost along the way.
|
|
module.exports = wrappy
|
|
function wrappy (fn, cb) {
|
|
if (fn && cb) return wrappy(fn)(cb)
|
|
|
|
if (typeof fn !== 'function')
|
|
throw new TypeError('need wrapper function')
|
|
|
|
Object.keys(fn).forEach(function (k) {
|
|
wrapper[k] = fn[k]
|
|
})
|
|
|
|
return wrapper
|
|
|
|
function wrapper() {
|
|
var args = new Array(arguments.length)
|
|
for (var i = 0; i < args.length; i++) {
|
|
args[i] = arguments[i]
|
|
}
|
|
var ret = fn.apply(this, args)
|
|
var cb = args[args.length-1]
|
|
if (typeof ret === 'function' && ret !== cb) {
|
|
Object.keys(cb).forEach(function (k) {
|
|
ret[k] = cb[k]
|
|
})
|
|
}
|
|
return ret
|
|
}
|
|
}
|
|
|
|
|
|
/***/ }),
|
|
|
|
/***/ 5506:
|
|
/***/ ((__unused_webpack_module, exports, __nccwpck_require__) => {
|
|
|
|
"use strict";
|
|
|
|
|
|
var PlainValue = __nccwpck_require__(5215);
|
|
var resolveSeq = __nccwpck_require__(4227);
|
|
var Schema = __nccwpck_require__(8021);
|
|
|
|
const defaultOptions = {
|
|
anchorPrefix: 'a',
|
|
customTags: null,
|
|
indent: 2,
|
|
indentSeq: true,
|
|
keepCstNodes: false,
|
|
keepNodeTypes: true,
|
|
keepBlobsInJSON: true,
|
|
mapAsMap: false,
|
|
maxAliasCount: 100,
|
|
prettyErrors: false,
|
|
// TODO Set true in v2
|
|
simpleKeys: false,
|
|
version: '1.2'
|
|
};
|
|
const scalarOptions = {
|
|
get binary() {
|
|
return resolveSeq.binaryOptions;
|
|
},
|
|
|
|
set binary(opt) {
|
|
Object.assign(resolveSeq.binaryOptions, opt);
|
|
},
|
|
|
|
get bool() {
|
|
return resolveSeq.boolOptions;
|
|
},
|
|
|
|
set bool(opt) {
|
|
Object.assign(resolveSeq.boolOptions, opt);
|
|
},
|
|
|
|
get int() {
|
|
return resolveSeq.intOptions;
|
|
},
|
|
|
|
set int(opt) {
|
|
Object.assign(resolveSeq.intOptions, opt);
|
|
},
|
|
|
|
get null() {
|
|
return resolveSeq.nullOptions;
|
|
},
|
|
|
|
set null(opt) {
|
|
Object.assign(resolveSeq.nullOptions, opt);
|
|
},
|
|
|
|
get str() {
|
|
return resolveSeq.strOptions;
|
|
},
|
|
|
|
set str(opt) {
|
|
Object.assign(resolveSeq.strOptions, opt);
|
|
}
|
|
|
|
};
|
|
const documentOptions = {
|
|
'1.0': {
|
|
schema: 'yaml-1.1',
|
|
merge: true,
|
|
tagPrefixes: [{
|
|
handle: '!',
|
|
prefix: PlainValue.defaultTagPrefix
|
|
}, {
|
|
handle: '!!',
|
|
prefix: 'tag:private.yaml.org,2002:'
|
|
}]
|
|
},
|
|
1.1: {
|
|
schema: 'yaml-1.1',
|
|
merge: true,
|
|
tagPrefixes: [{
|
|
handle: '!',
|
|
prefix: '!'
|
|
}, {
|
|
handle: '!!',
|
|
prefix: PlainValue.defaultTagPrefix
|
|
}]
|
|
},
|
|
1.2: {
|
|
schema: 'core',
|
|
merge: false,
|
|
tagPrefixes: [{
|
|
handle: '!',
|
|
prefix: '!'
|
|
}, {
|
|
handle: '!!',
|
|
prefix: PlainValue.defaultTagPrefix
|
|
}]
|
|
}
|
|
};
|
|
|
|
function stringifyTag(doc, tag) {
|
|
if ((doc.version || doc.options.version) === '1.0') {
|
|
const priv = tag.match(/^tag:private\.yaml\.org,2002:([^:/]+)$/);
|
|
if (priv) return '!' + priv[1];
|
|
const vocab = tag.match(/^tag:([a-zA-Z0-9-]+)\.yaml\.org,2002:(.*)/);
|
|
return vocab ? `!${vocab[1]}/${vocab[2]}` : `!${tag.replace(/^tag:/, '')}`;
|
|
}
|
|
|
|
let p = doc.tagPrefixes.find(p => tag.indexOf(p.prefix) === 0);
|
|
|
|
if (!p) {
|
|
const dtp = doc.getDefaults().tagPrefixes;
|
|
p = dtp && dtp.find(p => tag.indexOf(p.prefix) === 0);
|
|
}
|
|
|
|
if (!p) return tag[0] === '!' ? tag : `!<${tag}>`;
|
|
const suffix = tag.substr(p.prefix.length).replace(/[!,[\]{}]/g, ch => ({
|
|
'!': '%21',
|
|
',': '%2C',
|
|
'[': '%5B',
|
|
']': '%5D',
|
|
'{': '%7B',
|
|
'}': '%7D'
|
|
})[ch]);
|
|
return p.handle + suffix;
|
|
}
|
|
|
|
function getTagObject(tags, item) {
|
|
if (item instanceof resolveSeq.Alias) return resolveSeq.Alias;
|
|
|
|
if (item.tag) {
|
|
const match = tags.filter(t => t.tag === item.tag);
|
|
if (match.length > 0) return match.find(t => t.format === item.format) || match[0];
|
|
}
|
|
|
|
let tagObj, obj;
|
|
|
|
if (item instanceof resolveSeq.Scalar) {
|
|
obj = item.value; // TODO: deprecate/remove class check
|
|
|
|
const match = tags.filter(t => t.identify && t.identify(obj) || t.class && obj instanceof t.class);
|
|
tagObj = match.find(t => t.format === item.format) || match.find(t => !t.format);
|
|
} else {
|
|
obj = item;
|
|
tagObj = tags.find(t => t.nodeClass && obj instanceof t.nodeClass);
|
|
}
|
|
|
|
if (!tagObj) {
|
|
const name = obj && obj.constructor ? obj.constructor.name : typeof obj;
|
|
throw new Error(`Tag not resolved for ${name} value`);
|
|
}
|
|
|
|
return tagObj;
|
|
} // needs to be called before value stringifier to allow for circular anchor refs
|
|
|
|
|
|
function stringifyProps(node, tagObj, {
|
|
anchors,
|
|
doc
|
|
}) {
|
|
const props = [];
|
|
const anchor = doc.anchors.getName(node);
|
|
|
|
if (anchor) {
|
|
anchors[anchor] = node;
|
|
props.push(`&${anchor}`);
|
|
}
|
|
|
|
if (node.tag) {
|
|
props.push(stringifyTag(doc, node.tag));
|
|
} else if (!tagObj.default) {
|
|
props.push(stringifyTag(doc, tagObj.tag));
|
|
}
|
|
|
|
return props.join(' ');
|
|
}
|
|
|
|
function stringify(item, ctx, onComment, onChompKeep) {
|
|
const {
|
|
anchors,
|
|
schema
|
|
} = ctx.doc;
|
|
let tagObj;
|
|
|
|
if (!(item instanceof resolveSeq.Node)) {
|
|
const createCtx = {
|
|
aliasNodes: [],
|
|
onTagObj: o => tagObj = o,
|
|
prevObjects: new Map()
|
|
};
|
|
item = schema.createNode(item, true, null, createCtx);
|
|
|
|
for (const alias of createCtx.aliasNodes) {
|
|
alias.source = alias.source.node;
|
|
let name = anchors.getName(alias.source);
|
|
|
|
if (!name) {
|
|
name = anchors.newName();
|
|
anchors.map[name] = alias.source;
|
|
}
|
|
}
|
|
}
|
|
|
|
if (item instanceof resolveSeq.Pair) return item.toString(ctx, onComment, onChompKeep);
|
|
if (!tagObj) tagObj = getTagObject(schema.tags, item);
|
|
const props = stringifyProps(item, tagObj, ctx);
|
|
if (props.length > 0) ctx.indentAtStart = (ctx.indentAtStart || 0) + props.length + 1;
|
|
const str = typeof tagObj.stringify === 'function' ? tagObj.stringify(item, ctx, onComment, onChompKeep) : item instanceof resolveSeq.Scalar ? resolveSeq.stringifyString(item, ctx, onComment, onChompKeep) : item.toString(ctx, onComment, onChompKeep);
|
|
if (!props) return str;
|
|
return item instanceof resolveSeq.Scalar || str[0] === '{' || str[0] === '[' ? `${props} ${str}` : `${props}\n${ctx.indent}${str}`;
|
|
}
|
|
|
|
class Anchors {
|
|
static validAnchorNode(node) {
|
|
return node instanceof resolveSeq.Scalar || node instanceof resolveSeq.YAMLSeq || node instanceof resolveSeq.YAMLMap;
|
|
}
|
|
|
|
constructor(prefix) {
|
|
PlainValue._defineProperty(this, "map", Object.create(null));
|
|
|
|
this.prefix = prefix;
|
|
}
|
|
|
|
createAlias(node, name) {
|
|
this.setAnchor(node, name);
|
|
return new resolveSeq.Alias(node);
|
|
}
|
|
|
|
createMergePair(...sources) {
|
|
const merge = new resolveSeq.Merge();
|
|
merge.value.items = sources.map(s => {
|
|
if (s instanceof resolveSeq.Alias) {
|
|
if (s.source instanceof resolveSeq.YAMLMap) return s;
|
|
} else if (s instanceof resolveSeq.YAMLMap) {
|
|
return this.createAlias(s);
|
|
}
|
|
|
|
throw new Error('Merge sources must be Map nodes or their Aliases');
|
|
});
|
|
return merge;
|
|
}
|
|
|
|
getName(node) {
|
|
const {
|
|
map
|
|
} = this;
|
|
return Object.keys(map).find(a => map[a] === node);
|
|
}
|
|
|
|
getNames() {
|
|
return Object.keys(this.map);
|
|
}
|
|
|
|
getNode(name) {
|
|
return this.map[name];
|
|
}
|
|
|
|
newName(prefix) {
|
|
if (!prefix) prefix = this.prefix;
|
|
const names = Object.keys(this.map);
|
|
|
|
for (let i = 1; true; ++i) {
|
|
const name = `${prefix}${i}`;
|
|
if (!names.includes(name)) return name;
|
|
}
|
|
} // During parsing, map & aliases contain CST nodes
|
|
|
|
|
|
resolveNodes() {
|
|
const {
|
|
map,
|
|
_cstAliases
|
|
} = this;
|
|
Object.keys(map).forEach(a => {
|
|
map[a] = map[a].resolved;
|
|
});
|
|
|
|
_cstAliases.forEach(a => {
|
|
a.source = a.source.resolved;
|
|
});
|
|
|
|
delete this._cstAliases;
|
|
}
|
|
|
|
setAnchor(node, name) {
|
|
if (node != null && !Anchors.validAnchorNode(node)) {
|
|
throw new Error('Anchors may only be set for Scalar, Seq and Map nodes');
|
|
}
|
|
|
|
if (name && /[\x00-\x19\s,[\]{}]/.test(name)) {
|
|
throw new Error('Anchor names must not contain whitespace or control characters');
|
|
}
|
|
|
|
const {
|
|
map
|
|
} = this;
|
|
const prev = node && Object.keys(map).find(a => map[a] === node);
|
|
|
|
if (prev) {
|
|
if (!name) {
|
|
return prev;
|
|
} else if (prev !== name) {
|
|
delete map[prev];
|
|
map[name] = node;
|
|
}
|
|
} else {
|
|
if (!name) {
|
|
if (!node) return null;
|
|
name = this.newName();
|
|
}
|
|
|
|
map[name] = node;
|
|
}
|
|
|
|
return name;
|
|
}
|
|
|
|
}
|
|
|
|
const visit = (node, tags) => {
|
|
if (node && typeof node === 'object') {
|
|
const {
|
|
tag
|
|
} = node;
|
|
|
|
if (node instanceof resolveSeq.Collection) {
|
|
if (tag) tags[tag] = true;
|
|
node.items.forEach(n => visit(n, tags));
|
|
} else if (node instanceof resolveSeq.Pair) {
|
|
visit(node.key, tags);
|
|
visit(node.value, tags);
|
|
} else if (node instanceof resolveSeq.Scalar) {
|
|
if (tag) tags[tag] = true;
|
|
}
|
|
}
|
|
|
|
return tags;
|
|
};
|
|
|
|
const listTagNames = node => Object.keys(visit(node, {}));
|
|
|
|
function parseContents(doc, contents) {
|
|
const comments = {
|
|
before: [],
|
|
after: []
|
|
};
|
|
let body = undefined;
|
|
let spaceBefore = false;
|
|
|
|
for (const node of contents) {
|
|
if (node.valueRange) {
|
|
if (body !== undefined) {
|
|
const msg = 'Document contains trailing content not separated by a ... or --- line';
|
|
doc.errors.push(new PlainValue.YAMLSyntaxError(node, msg));
|
|
break;
|
|
}
|
|
|
|
const res = resolveSeq.resolveNode(doc, node);
|
|
|
|
if (spaceBefore) {
|
|
res.spaceBefore = true;
|
|
spaceBefore = false;
|
|
}
|
|
|
|
body = res;
|
|
} else if (node.comment !== null) {
|
|
const cc = body === undefined ? comments.before : comments.after;
|
|
cc.push(node.comment);
|
|
} else if (node.type === PlainValue.Type.BLANK_LINE) {
|
|
spaceBefore = true;
|
|
|
|
if (body === undefined && comments.before.length > 0 && !doc.commentBefore) {
|
|
// space-separated comments at start are parsed as document comments
|
|
doc.commentBefore = comments.before.join('\n');
|
|
comments.before = [];
|
|
}
|
|
}
|
|
}
|
|
|
|
doc.contents = body || null;
|
|
|
|
if (!body) {
|
|
doc.comment = comments.before.concat(comments.after).join('\n') || null;
|
|
} else {
|
|
const cb = comments.before.join('\n');
|
|
|
|
if (cb) {
|
|
const cbNode = body instanceof resolveSeq.Collection && body.items[0] ? body.items[0] : body;
|
|
cbNode.commentBefore = cbNode.commentBefore ? `${cb}\n${cbNode.commentBefore}` : cb;
|
|
}
|
|
|
|
doc.comment = comments.after.join('\n') || null;
|
|
}
|
|
}
|
|
|
|
function resolveTagDirective({
|
|
tagPrefixes
|
|
}, directive) {
|
|
const [handle, prefix] = directive.parameters;
|
|
|
|
if (!handle || !prefix) {
|
|
const msg = 'Insufficient parameters given for %TAG directive';
|
|
throw new PlainValue.YAMLSemanticError(directive, msg);
|
|
}
|
|
|
|
if (tagPrefixes.some(p => p.handle === handle)) {
|
|
const msg = 'The %TAG directive must only be given at most once per handle in the same document.';
|
|
throw new PlainValue.YAMLSemanticError(directive, msg);
|
|
}
|
|
|
|
return {
|
|
handle,
|
|
prefix
|
|
};
|
|
}
|
|
|
|
function resolveYamlDirective(doc, directive) {
|
|
let [version] = directive.parameters;
|
|
if (directive.name === 'YAML:1.0') version = '1.0';
|
|
|
|
if (!version) {
|
|
const msg = 'Insufficient parameters given for %YAML directive';
|
|
throw new PlainValue.YAMLSemanticError(directive, msg);
|
|
}
|
|
|
|
if (!documentOptions[version]) {
|
|
const v0 = doc.version || doc.options.version;
|
|
const msg = `Document will be parsed as YAML ${v0} rather than YAML ${version}`;
|
|
doc.warnings.push(new PlainValue.YAMLWarning(directive, msg));
|
|
}
|
|
|
|
return version;
|
|
}
|
|
|
|
function parseDirectives(doc, directives, prevDoc) {
|
|
const directiveComments = [];
|
|
let hasDirectives = false;
|
|
|
|
for (const directive of directives) {
|
|
const {
|
|
comment,
|
|
name
|
|
} = directive;
|
|
|
|
switch (name) {
|
|
case 'TAG':
|
|
try {
|
|
doc.tagPrefixes.push(resolveTagDirective(doc, directive));
|
|
} catch (error) {
|
|
doc.errors.push(error);
|
|
}
|
|
|
|
hasDirectives = true;
|
|
break;
|
|
|
|
case 'YAML':
|
|
case 'YAML:1.0':
|
|
if (doc.version) {
|
|
const msg = 'The %YAML directive must only be given at most once per document.';
|
|
doc.errors.push(new PlainValue.YAMLSemanticError(directive, msg));
|
|
}
|
|
|
|
try {
|
|
doc.version = resolveYamlDirective(doc, directive);
|
|
} catch (error) {
|
|
doc.errors.push(error);
|
|
}
|
|
|
|
hasDirectives = true;
|
|
break;
|
|
|
|
default:
|
|
if (name) {
|
|
const msg = `YAML only supports %TAG and %YAML directives, and not %${name}`;
|
|
doc.warnings.push(new PlainValue.YAMLWarning(directive, msg));
|
|
}
|
|
|
|
}
|
|
|
|
if (comment) directiveComments.push(comment);
|
|
}
|
|
|
|
if (prevDoc && !hasDirectives && '1.1' === (doc.version || prevDoc.version || doc.options.version)) {
|
|
const copyTagPrefix = ({
|
|
handle,
|
|
prefix
|
|
}) => ({
|
|
handle,
|
|
prefix
|
|
});
|
|
|
|
doc.tagPrefixes = prevDoc.tagPrefixes.map(copyTagPrefix);
|
|
doc.version = prevDoc.version;
|
|
}
|
|
|
|
doc.commentBefore = directiveComments.join('\n') || null;
|
|
}
|
|
|
|
function assertCollection(contents) {
|
|
if (contents instanceof resolveSeq.Collection) return true;
|
|
throw new Error('Expected a YAML collection as document contents');
|
|
}
|
|
|
|
class Document {
|
|
constructor(options) {
|
|
this.anchors = new Anchors(options.anchorPrefix);
|
|
this.commentBefore = null;
|
|
this.comment = null;
|
|
this.contents = null;
|
|
this.directivesEndMarker = null;
|
|
this.errors = [];
|
|
this.options = options;
|
|
this.schema = null;
|
|
this.tagPrefixes = [];
|
|
this.version = null;
|
|
this.warnings = [];
|
|
}
|
|
|
|
add(value) {
|
|
assertCollection(this.contents);
|
|
return this.contents.add(value);
|
|
}
|
|
|
|
addIn(path, value) {
|
|
assertCollection(this.contents);
|
|
this.contents.addIn(path, value);
|
|
}
|
|
|
|
delete(key) {
|
|
assertCollection(this.contents);
|
|
return this.contents.delete(key);
|
|
}
|
|
|
|
deleteIn(path) {
|
|
if (resolveSeq.isEmptyPath(path)) {
|
|
if (this.contents == null) return false;
|
|
this.contents = null;
|
|
return true;
|
|
}
|
|
|
|
assertCollection(this.contents);
|
|
return this.contents.deleteIn(path);
|
|
}
|
|
|
|
getDefaults() {
|
|
return Document.defaults[this.version] || Document.defaults[this.options.version] || {};
|
|
}
|
|
|
|
get(key, keepScalar) {
|
|
return this.contents instanceof resolveSeq.Collection ? this.contents.get(key, keepScalar) : undefined;
|
|
}
|
|
|
|
getIn(path, keepScalar) {
|
|
if (resolveSeq.isEmptyPath(path)) return !keepScalar && this.contents instanceof resolveSeq.Scalar ? this.contents.value : this.contents;
|
|
return this.contents instanceof resolveSeq.Collection ? this.contents.getIn(path, keepScalar) : undefined;
|
|
}
|
|
|
|
has(key) {
|
|
return this.contents instanceof resolveSeq.Collection ? this.contents.has(key) : false;
|
|
}
|
|
|
|
hasIn(path) {
|
|
if (resolveSeq.isEmptyPath(path)) return this.contents !== undefined;
|
|
return this.contents instanceof resolveSeq.Collection ? this.contents.hasIn(path) : false;
|
|
}
|
|
|
|
set(key, value) {
|
|
assertCollection(this.contents);
|
|
this.contents.set(key, value);
|
|
}
|
|
|
|
setIn(path, value) {
|
|
if (resolveSeq.isEmptyPath(path)) this.contents = value;else {
|
|
assertCollection(this.contents);
|
|
this.contents.setIn(path, value);
|
|
}
|
|
}
|
|
|
|
setSchema(id, customTags) {
|
|
if (!id && !customTags && this.schema) return;
|
|
if (typeof id === 'number') id = id.toFixed(1);
|
|
|
|
if (id === '1.0' || id === '1.1' || id === '1.2') {
|
|
if (this.version) this.version = id;else this.options.version = id;
|
|
delete this.options.schema;
|
|
} else if (id && typeof id === 'string') {
|
|
this.options.schema = id;
|
|
}
|
|
|
|
if (Array.isArray(customTags)) this.options.customTags = customTags;
|
|
const opt = Object.assign({}, this.getDefaults(), this.options);
|
|
this.schema = new Schema.Schema(opt);
|
|
}
|
|
|
|
parse(node, prevDoc) {
|
|
if (this.options.keepCstNodes) this.cstNode = node;
|
|
if (this.options.keepNodeTypes) this.type = 'DOCUMENT';
|
|
const {
|
|
directives = [],
|
|
contents = [],
|
|
directivesEndMarker,
|
|
error,
|
|
valueRange
|
|
} = node;
|
|
|
|
if (error) {
|
|
if (!error.source) error.source = this;
|
|
this.errors.push(error);
|
|
}
|
|
|
|
parseDirectives(this, directives, prevDoc);
|
|
if (directivesEndMarker) this.directivesEndMarker = true;
|
|
this.range = valueRange ? [valueRange.start, valueRange.end] : null;
|
|
this.setSchema();
|
|
this.anchors._cstAliases = [];
|
|
parseContents(this, contents);
|
|
this.anchors.resolveNodes();
|
|
|
|
if (this.options.prettyErrors) {
|
|
for (const error of this.errors) if (error instanceof PlainValue.YAMLError) error.makePretty();
|
|
|
|
for (const warn of this.warnings) if (warn instanceof PlainValue.YAMLError) warn.makePretty();
|
|
}
|
|
|
|
return this;
|
|
}
|
|
|
|
listNonDefaultTags() {
|
|
return listTagNames(this.contents).filter(t => t.indexOf(Schema.Schema.defaultPrefix) !== 0);
|
|
}
|
|
|
|
setTagPrefix(handle, prefix) {
|
|
if (handle[0] !== '!' || handle[handle.length - 1] !== '!') throw new Error('Handle must start and end with !');
|
|
|
|
if (prefix) {
|
|
const prev = this.tagPrefixes.find(p => p.handle === handle);
|
|
if (prev) prev.prefix = prefix;else this.tagPrefixes.push({
|
|
handle,
|
|
prefix
|
|
});
|
|
} else {
|
|
this.tagPrefixes = this.tagPrefixes.filter(p => p.handle !== handle);
|
|
}
|
|
}
|
|
|
|
toJSON(arg, onAnchor) {
|
|
const {
|
|
keepBlobsInJSON,
|
|
mapAsMap,
|
|
maxAliasCount
|
|
} = this.options;
|
|
const keep = keepBlobsInJSON && (typeof arg !== 'string' || !(this.contents instanceof resolveSeq.Scalar));
|
|
const ctx = {
|
|
doc: this,
|
|
indentStep: ' ',
|
|
keep,
|
|
mapAsMap: keep && !!mapAsMap,
|
|
maxAliasCount,
|
|
stringify // Requiring directly in Pair would create circular dependencies
|
|
|
|
};
|
|
const anchorNames = Object.keys(this.anchors.map);
|
|
if (anchorNames.length > 0) ctx.anchors = new Map(anchorNames.map(name => [this.anchors.map[name], {
|
|
alias: [],
|
|
aliasCount: 0,
|
|
count: 1
|
|
}]));
|
|
const res = resolveSeq.toJSON(this.contents, arg, ctx);
|
|
if (typeof onAnchor === 'function' && ctx.anchors) for (const {
|
|
count,
|
|
res
|
|
} of ctx.anchors.values()) onAnchor(res, count);
|
|
return res;
|
|
}
|
|
|
|
toString() {
|
|
if (this.errors.length > 0) throw new Error('Document with errors cannot be stringified');
|
|
const indentSize = this.options.indent;
|
|
|
|
if (!Number.isInteger(indentSize) || indentSize <= 0) {
|
|
const s = JSON.stringify(indentSize);
|
|
throw new Error(`"indent" option must be a positive integer, not ${s}`);
|
|
}
|
|
|
|
this.setSchema();
|
|
const lines = [];
|
|
let hasDirectives = false;
|
|
|
|
if (this.version) {
|
|
let vd = '%YAML 1.2';
|
|
|
|
if (this.schema.name === 'yaml-1.1') {
|
|
if (this.version === '1.0') vd = '%YAML:1.0';else if (this.version === '1.1') vd = '%YAML 1.1';
|
|
}
|
|
|
|
lines.push(vd);
|
|
hasDirectives = true;
|
|
}
|
|
|
|
const tagNames = this.listNonDefaultTags();
|
|
this.tagPrefixes.forEach(({
|
|
handle,
|
|
prefix
|
|
}) => {
|
|
if (tagNames.some(t => t.indexOf(prefix) === 0)) {
|
|
lines.push(`%TAG ${handle} ${prefix}`);
|
|
hasDirectives = true;
|
|
}
|
|
});
|
|
if (hasDirectives || this.directivesEndMarker) lines.push('---');
|
|
|
|
if (this.commentBefore) {
|
|
if (hasDirectives || !this.directivesEndMarker) lines.unshift('');
|
|
lines.unshift(this.commentBefore.replace(/^/gm, '#'));
|
|
}
|
|
|
|
const ctx = {
|
|
anchors: Object.create(null),
|
|
doc: this,
|
|
indent: '',
|
|
indentStep: ' '.repeat(indentSize),
|
|
stringify // Requiring directly in nodes would create circular dependencies
|
|
|
|
};
|
|
let chompKeep = false;
|
|
let contentComment = null;
|
|
|
|
if (this.contents) {
|
|
if (this.contents instanceof resolveSeq.Node) {
|
|
if (this.contents.spaceBefore && (hasDirectives || this.directivesEndMarker)) lines.push('');
|
|
if (this.contents.commentBefore) lines.push(this.contents.commentBefore.replace(/^/gm, '#')); // top-level block scalars need to be indented if followed by a comment
|
|
|
|
ctx.forceBlockIndent = !!this.comment;
|
|
contentComment = this.contents.comment;
|
|
}
|
|
|
|
const onChompKeep = contentComment ? null : () => chompKeep = true;
|
|
const body = stringify(this.contents, ctx, () => contentComment = null, onChompKeep);
|
|
lines.push(resolveSeq.addComment(body, '', contentComment));
|
|
} else if (this.contents !== undefined) {
|
|
lines.push(stringify(this.contents, ctx));
|
|
}
|
|
|
|
if (this.comment) {
|
|
if ((!chompKeep || contentComment) && lines[lines.length - 1] !== '') lines.push('');
|
|
lines.push(this.comment.replace(/^/gm, '#'));
|
|
}
|
|
|
|
return lines.join('\n') + '\n';
|
|
}
|
|
|
|
}
|
|
|
|
PlainValue._defineProperty(Document, "defaults", documentOptions);
|
|
|
|
exports.Document = Document;
|
|
exports.defaultOptions = defaultOptions;
|
|
exports.scalarOptions = scalarOptions;
|
|
|
|
|
|
/***/ }),
|
|
|
|
/***/ 5215:
|
|
/***/ ((__unused_webpack_module, exports) => {
|
|
|
|
"use strict";
|
|
|
|
|
|
const Char = {
|
|
ANCHOR: '&',
|
|
COMMENT: '#',
|
|
TAG: '!',
|
|
DIRECTIVES_END: '-',
|
|
DOCUMENT_END: '.'
|
|
};
|
|
const Type = {
|
|
ALIAS: 'ALIAS',
|
|
BLANK_LINE: 'BLANK_LINE',
|
|
BLOCK_FOLDED: 'BLOCK_FOLDED',
|
|
BLOCK_LITERAL: 'BLOCK_LITERAL',
|
|
COMMENT: 'COMMENT',
|
|
DIRECTIVE: 'DIRECTIVE',
|
|
DOCUMENT: 'DOCUMENT',
|
|
FLOW_MAP: 'FLOW_MAP',
|
|
FLOW_SEQ: 'FLOW_SEQ',
|
|
MAP: 'MAP',
|
|
MAP_KEY: 'MAP_KEY',
|
|
MAP_VALUE: 'MAP_VALUE',
|
|
PLAIN: 'PLAIN',
|
|
QUOTE_DOUBLE: 'QUOTE_DOUBLE',
|
|
QUOTE_SINGLE: 'QUOTE_SINGLE',
|
|
SEQ: 'SEQ',
|
|
SEQ_ITEM: 'SEQ_ITEM'
|
|
};
|
|
const defaultTagPrefix = 'tag:yaml.org,2002:';
|
|
const defaultTags = {
|
|
MAP: 'tag:yaml.org,2002:map',
|
|
SEQ: 'tag:yaml.org,2002:seq',
|
|
STR: 'tag:yaml.org,2002:str'
|
|
};
|
|
|
|
function findLineStarts(src) {
|
|
const ls = [0];
|
|
let offset = src.indexOf('\n');
|
|
|
|
while (offset !== -1) {
|
|
offset += 1;
|
|
ls.push(offset);
|
|
offset = src.indexOf('\n', offset);
|
|
}
|
|
|
|
return ls;
|
|
}
|
|
|
|
function getSrcInfo(cst) {
|
|
let lineStarts, src;
|
|
|
|
if (typeof cst === 'string') {
|
|
lineStarts = findLineStarts(cst);
|
|
src = cst;
|
|
} else {
|
|
if (Array.isArray(cst)) cst = cst[0];
|
|
|
|
if (cst && cst.context) {
|
|
if (!cst.lineStarts) cst.lineStarts = findLineStarts(cst.context.src);
|
|
lineStarts = cst.lineStarts;
|
|
src = cst.context.src;
|
|
}
|
|
}
|
|
|
|
return {
|
|
lineStarts,
|
|
src
|
|
};
|
|
}
|
|
/**
|
|
* @typedef {Object} LinePos - One-indexed position in the source
|
|
* @property {number} line
|
|
* @property {number} col
|
|
*/
|
|
|
|
/**
|
|
* Determine the line/col position matching a character offset.
|
|
*
|
|
* Accepts a source string or a CST document as the second parameter. With
|
|
* the latter, starting indices for lines are cached in the document as
|
|
* `lineStarts: number[]`.
|
|
*
|
|
* Returns a one-indexed `{ line, col }` location if found, or
|
|
* `undefined` otherwise.
|
|
*
|
|
* @param {number} offset
|
|
* @param {string|Document|Document[]} cst
|
|
* @returns {?LinePos}
|
|
*/
|
|
|
|
|
|
function getLinePos(offset, cst) {
|
|
if (typeof offset !== 'number' || offset < 0) return null;
|
|
const {
|
|
lineStarts,
|
|
src
|
|
} = getSrcInfo(cst);
|
|
if (!lineStarts || !src || offset > src.length) return null;
|
|
|
|
for (let i = 0; i < lineStarts.length; ++i) {
|
|
const start = lineStarts[i];
|
|
|
|
if (offset < start) {
|
|
return {
|
|
line: i,
|
|
col: offset - lineStarts[i - 1] + 1
|
|
};
|
|
}
|
|
|
|
if (offset === start) return {
|
|
line: i + 1,
|
|
col: 1
|
|
};
|
|
}
|
|
|
|
const line = lineStarts.length;
|
|
return {
|
|
line,
|
|
col: offset - lineStarts[line - 1] + 1
|
|
};
|
|
}
|
|
/**
|
|
* Get a specified line from the source.
|
|
*
|
|
* Accepts a source string or a CST document as the second parameter. With
|
|
* the latter, starting indices for lines are cached in the document as
|
|
* `lineStarts: number[]`.
|
|
*
|
|
* Returns the line as a string if found, or `null` otherwise.
|
|
*
|
|
* @param {number} line One-indexed line number
|
|
* @param {string|Document|Document[]} cst
|
|
* @returns {?string}
|
|
*/
|
|
|
|
function getLine(line, cst) {
|
|
const {
|
|
lineStarts,
|
|
src
|
|
} = getSrcInfo(cst);
|
|
if (!lineStarts || !(line >= 1) || line > lineStarts.length) return null;
|
|
const start = lineStarts[line - 1];
|
|
let end = lineStarts[line]; // undefined for last line; that's ok for slice()
|
|
|
|
while (end && end > start && src[end - 1] === '\n') --end;
|
|
|
|
return src.slice(start, end);
|
|
}
|
|
/**
|
|
* Pretty-print the starting line from the source indicated by the range `pos`
|
|
*
|
|
* Trims output to `maxWidth` chars while keeping the starting column visible,
|
|
* using `…` at either end to indicate dropped characters.
|
|
*
|
|
* Returns a two-line string (or `null`) with `\n` as separator; the second line
|
|
* will hold appropriately indented `^` marks indicating the column range.
|
|
*
|
|
* @param {Object} pos
|
|
* @param {LinePos} pos.start
|
|
* @param {LinePos} [pos.end]
|
|
* @param {string|Document|Document[]*} cst
|
|
* @param {number} [maxWidth=80]
|
|
* @returns {?string}
|
|
*/
|
|
|
|
function getPrettyContext({
|
|
start,
|
|
end
|
|
}, cst, maxWidth = 80) {
|
|
let src = getLine(start.line, cst);
|
|
if (!src) return null;
|
|
let {
|
|
col
|
|
} = start;
|
|
|
|
if (src.length > maxWidth) {
|
|
if (col <= maxWidth - 10) {
|
|
src = src.substr(0, maxWidth - 1) + '…';
|
|
} else {
|
|
const halfWidth = Math.round(maxWidth / 2);
|
|
if (src.length > col + halfWidth) src = src.substr(0, col + halfWidth - 1) + '…';
|
|
col -= src.length - maxWidth;
|
|
src = '…' + src.substr(1 - maxWidth);
|
|
}
|
|
}
|
|
|
|
let errLen = 1;
|
|
let errEnd = '';
|
|
|
|
if (end) {
|
|
if (end.line === start.line && col + (end.col - start.col) <= maxWidth + 1) {
|
|
errLen = end.col - start.col;
|
|
} else {
|
|
errLen = Math.min(src.length + 1, maxWidth) - col;
|
|
errEnd = '…';
|
|
}
|
|
}
|
|
|
|
const offset = col > 1 ? ' '.repeat(col - 1) : '';
|
|
const err = '^'.repeat(errLen);
|
|
return `${src}\n${offset}${err}${errEnd}`;
|
|
}
|
|
|
|
class Range {
|
|
static copy(orig) {
|
|
return new Range(orig.start, orig.end);
|
|
}
|
|
|
|
constructor(start, end) {
|
|
this.start = start;
|
|
this.end = end || start;
|
|
}
|
|
|
|
isEmpty() {
|
|
return typeof this.start !== 'number' || !this.end || this.end <= this.start;
|
|
}
|
|
/**
|
|
* Set `origStart` and `origEnd` to point to the original source range for
|
|
* this node, which may differ due to dropped CR characters.
|
|
*
|
|
* @param {number[]} cr - Positions of dropped CR characters
|
|
* @param {number} offset - Starting index of `cr` from the last call
|
|
* @returns {number} - The next offset, matching the one found for `origStart`
|
|
*/
|
|
|
|
|
|
setOrigRange(cr, offset) {
|
|
const {
|
|
start,
|
|
end
|
|
} = this;
|
|
|
|
if (cr.length === 0 || end <= cr[0]) {
|
|
this.origStart = start;
|
|
this.origEnd = end;
|
|
return offset;
|
|
}
|
|
|
|
let i = offset;
|
|
|
|
while (i < cr.length) {
|
|
if (cr[i] > start) break;else ++i;
|
|
}
|
|
|
|
this.origStart = start + i;
|
|
const nextOffset = i;
|
|
|
|
while (i < cr.length) {
|
|
// if end was at \n, it should now be at \r
|
|
if (cr[i] >= end) break;else ++i;
|
|
}
|
|
|
|
this.origEnd = end + i;
|
|
return nextOffset;
|
|
}
|
|
|
|
}
|
|
|
|
/** Root class of all nodes */
|
|
|
|
class Node {
|
|
static addStringTerminator(src, offset, str) {
|
|
if (str[str.length - 1] === '\n') return str;
|
|
const next = Node.endOfWhiteSpace(src, offset);
|
|
return next >= src.length || src[next] === '\n' ? str + '\n' : str;
|
|
} // ^(---|...)
|
|
|
|
|
|
static atDocumentBoundary(src, offset, sep) {
|
|
const ch0 = src[offset];
|
|
if (!ch0) return true;
|
|
const prev = src[offset - 1];
|
|
if (prev && prev !== '\n') return false;
|
|
|
|
if (sep) {
|
|
if (ch0 !== sep) return false;
|
|
} else {
|
|
if (ch0 !== Char.DIRECTIVES_END && ch0 !== Char.DOCUMENT_END) return false;
|
|
}
|
|
|
|
const ch1 = src[offset + 1];
|
|
const ch2 = src[offset + 2];
|
|
if (ch1 !== ch0 || ch2 !== ch0) return false;
|
|
const ch3 = src[offset + 3];
|
|
return !ch3 || ch3 === '\n' || ch3 === '\t' || ch3 === ' ';
|
|
}
|
|
|
|
static endOfIdentifier(src, offset) {
|
|
let ch = src[offset];
|
|
const isVerbatim = ch === '<';
|
|
const notOk = isVerbatim ? ['\n', '\t', ' ', '>'] : ['\n', '\t', ' ', '[', ']', '{', '}', ','];
|
|
|
|
while (ch && notOk.indexOf(ch) === -1) ch = src[offset += 1];
|
|
|
|
if (isVerbatim && ch === '>') offset += 1;
|
|
return offset;
|
|
}
|
|
|
|
static endOfIndent(src, offset) {
|
|
let ch = src[offset];
|
|
|
|
while (ch === ' ') ch = src[offset += 1];
|
|
|
|
return offset;
|
|
}
|
|
|
|
static endOfLine(src, offset) {
|
|
let ch = src[offset];
|
|
|
|
while (ch && ch !== '\n') ch = src[offset += 1];
|
|
|
|
return offset;
|
|
}
|
|
|
|
static endOfWhiteSpace(src, offset) {
|
|
let ch = src[offset];
|
|
|
|
while (ch === '\t' || ch === ' ') ch = src[offset += 1];
|
|
|
|
return offset;
|
|
}
|
|
|
|
static startOfLine(src, offset) {
|
|
let ch = src[offset - 1];
|
|
if (ch === '\n') return offset;
|
|
|
|
while (ch && ch !== '\n') ch = src[offset -= 1];
|
|
|
|
return offset + 1;
|
|
}
|
|
/**
|
|
* End of indentation, or null if the line's indent level is not more
|
|
* than `indent`
|
|
*
|
|
* @param {string} src
|
|
* @param {number} indent
|
|
* @param {number} lineStart
|
|
* @returns {?number}
|
|
*/
|
|
|
|
|
|
static endOfBlockIndent(src, indent, lineStart) {
|
|
const inEnd = Node.endOfIndent(src, lineStart);
|
|
|
|
if (inEnd > lineStart + indent) {
|
|
return inEnd;
|
|
} else {
|
|
const wsEnd = Node.endOfWhiteSpace(src, inEnd);
|
|
const ch = src[wsEnd];
|
|
if (!ch || ch === '\n') return wsEnd;
|
|
}
|
|
|
|
return null;
|
|
}
|
|
|
|
static atBlank(src, offset, endAsBlank) {
|
|
const ch = src[offset];
|
|
return ch === '\n' || ch === '\t' || ch === ' ' || endAsBlank && !ch;
|
|
}
|
|
|
|
static nextNodeIsIndented(ch, indentDiff, indicatorAsIndent) {
|
|
if (!ch || indentDiff < 0) return false;
|
|
if (indentDiff > 0) return true;
|
|
return indicatorAsIndent && ch === '-';
|
|
} // should be at line or string end, or at next non-whitespace char
|
|
|
|
|
|
static normalizeOffset(src, offset) {
|
|
const ch = src[offset];
|
|
return !ch ? offset : ch !== '\n' && src[offset - 1] === '\n' ? offset - 1 : Node.endOfWhiteSpace(src, offset);
|
|
} // fold single newline into space, multiple newlines to N - 1 newlines
|
|
// presumes src[offset] === '\n'
|
|
|
|
|
|
static foldNewline(src, offset, indent) {
|
|
let inCount = 0;
|
|
let error = false;
|
|
let fold = '';
|
|
let ch = src[offset + 1];
|
|
|
|
while (ch === ' ' || ch === '\t' || ch === '\n') {
|
|
switch (ch) {
|
|
case '\n':
|
|
inCount = 0;
|
|
offset += 1;
|
|
fold += '\n';
|
|
break;
|
|
|
|
case '\t':
|
|
if (inCount <= indent) error = true;
|
|
offset = Node.endOfWhiteSpace(src, offset + 2) - 1;
|
|
break;
|
|
|
|
case ' ':
|
|
inCount += 1;
|
|
offset += 1;
|
|
break;
|
|
}
|
|
|
|
ch = src[offset + 1];
|
|
}
|
|
|
|
if (!fold) fold = ' ';
|
|
if (ch && inCount <= indent) error = true;
|
|
return {
|
|
fold,
|
|
offset,
|
|
error
|
|
};
|
|
}
|
|
|
|
constructor(type, props, context) {
|
|
Object.defineProperty(this, 'context', {
|
|
value: context || null,
|
|
writable: true
|
|
});
|
|
this.error = null;
|
|
this.range = null;
|
|
this.valueRange = null;
|
|
this.props = props || [];
|
|
this.type = type;
|
|
this.value = null;
|
|
}
|
|
|
|
getPropValue(idx, key, skipKey) {
|
|
if (!this.context) return null;
|
|
const {
|
|
src
|
|
} = this.context;
|
|
const prop = this.props[idx];
|
|
return prop && src[prop.start] === key ? src.slice(prop.start + (skipKey ? 1 : 0), prop.end) : null;
|
|
}
|
|
|
|
get anchor() {
|
|
for (let i = 0; i < this.props.length; ++i) {
|
|
const anchor = this.getPropValue(i, Char.ANCHOR, true);
|
|
if (anchor != null) return anchor;
|
|
}
|
|
|
|
return null;
|
|
}
|
|
|
|
get comment() {
|
|
const comments = [];
|
|
|
|
for (let i = 0; i < this.props.length; ++i) {
|
|
const comment = this.getPropValue(i, Char.COMMENT, true);
|
|
if (comment != null) comments.push(comment);
|
|
}
|
|
|
|
return comments.length > 0 ? comments.join('\n') : null;
|
|
}
|
|
|
|
commentHasRequiredWhitespace(start) {
|
|
const {
|
|
src
|
|
} = this.context;
|
|
if (this.header && start === this.header.end) return false;
|
|
if (!this.valueRange) return false;
|
|
const {
|
|
end
|
|
} = this.valueRange;
|
|
return start !== end || Node.atBlank(src, end - 1);
|
|
}
|
|
|
|
get hasComment() {
|
|
if (this.context) {
|
|
const {
|
|
src
|
|
} = this.context;
|
|
|
|
for (let i = 0; i < this.props.length; ++i) {
|
|
if (src[this.props[i].start] === Char.COMMENT) return true;
|
|
}
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
get hasProps() {
|
|
if (this.context) {
|
|
const {
|
|
src
|
|
} = this.context;
|
|
|
|
for (let i = 0; i < this.props.length; ++i) {
|
|
if (src[this.props[i].start] !== Char.COMMENT) return true;
|
|
}
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
get includesTrailingLines() {
|
|
return false;
|
|
}
|
|
|
|
get jsonLike() {
|
|
const jsonLikeTypes = [Type.FLOW_MAP, Type.FLOW_SEQ, Type.QUOTE_DOUBLE, Type.QUOTE_SINGLE];
|
|
return jsonLikeTypes.indexOf(this.type) !== -1;
|
|
}
|
|
|
|
get rangeAsLinePos() {
|
|
if (!this.range || !this.context) return undefined;
|
|
const start = getLinePos(this.range.start, this.context.root);
|
|
if (!start) return undefined;
|
|
const end = getLinePos(this.range.end, this.context.root);
|
|
return {
|
|
start,
|
|
end
|
|
};
|
|
}
|
|
|
|
get rawValue() {
|
|
if (!this.valueRange || !this.context) return null;
|
|
const {
|
|
start,
|
|
end
|
|
} = this.valueRange;
|
|
return this.context.src.slice(start, end);
|
|
}
|
|
|
|
get tag() {
|
|
for (let i = 0; i < this.props.length; ++i) {
|
|
const tag = this.getPropValue(i, Char.TAG, false);
|
|
|
|
if (tag != null) {
|
|
if (tag[1] === '<') {
|
|
return {
|
|
verbatim: tag.slice(2, -1)
|
|
};
|
|
} else {
|
|
// eslint-disable-next-line no-unused-vars
|
|
const [_, handle, suffix] = tag.match(/^(.*!)([^!]*)$/);
|
|
return {
|
|
handle,
|
|
suffix
|
|
};
|
|
}
|
|
}
|
|
}
|
|
|
|
return null;
|
|
}
|
|
|
|
get valueRangeContainsNewline() {
|
|
if (!this.valueRange || !this.context) return false;
|
|
const {
|
|
start,
|
|
end
|
|
} = this.valueRange;
|
|
const {
|
|
src
|
|
} = this.context;
|
|
|
|
for (let i = start; i < end; ++i) {
|
|
if (src[i] === '\n') return true;
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
parseComment(start) {
|
|
const {
|
|
src
|
|
} = this.context;
|
|
|
|
if (src[start] === Char.COMMENT) {
|
|
const end = Node.endOfLine(src, start + 1);
|
|
const commentRange = new Range(start, end);
|
|
this.props.push(commentRange);
|
|
return end;
|
|
}
|
|
|
|
return start;
|
|
}
|
|
/**
|
|
* Populates the `origStart` and `origEnd` values of all ranges for this
|
|
* node. Extended by child classes to handle descendant nodes.
|
|
*
|
|
* @param {number[]} cr - Positions of dropped CR characters
|
|
* @param {number} offset - Starting index of `cr` from the last call
|
|
* @returns {number} - The next offset, matching the one found for `origStart`
|
|
*/
|
|
|
|
|
|
setOrigRanges(cr, offset) {
|
|
if (this.range) offset = this.range.setOrigRange(cr, offset);
|
|
if (this.valueRange) this.valueRange.setOrigRange(cr, offset);
|
|
this.props.forEach(prop => prop.setOrigRange(cr, offset));
|
|
return offset;
|
|
}
|
|
|
|
toString() {
|
|
const {
|
|
context: {
|
|
src
|
|
},
|
|
range,
|
|
value
|
|
} = this;
|
|
if (value != null) return value;
|
|
const str = src.slice(range.start, range.end);
|
|
return Node.addStringTerminator(src, range.end, str);
|
|
}
|
|
|
|
}
|
|
|
|
class YAMLError extends Error {
|
|
constructor(name, source, message) {
|
|
if (!message || !(source instanceof Node)) throw new Error(`Invalid arguments for new ${name}`);
|
|
super();
|
|
this.name = name;
|
|
this.message = message;
|
|
this.source = source;
|
|
}
|
|
|
|
makePretty() {
|
|
if (!this.source) return;
|
|
this.nodeType = this.source.type;
|
|
const cst = this.source.context && this.source.context.root;
|
|
|
|
if (typeof this.offset === 'number') {
|
|
this.range = new Range(this.offset, this.offset + 1);
|
|
const start = cst && getLinePos(this.offset, cst);
|
|
|
|
if (start) {
|
|
const end = {
|
|
line: start.line,
|
|
col: start.col + 1
|
|
};
|
|
this.linePos = {
|
|
start,
|
|
end
|
|
};
|
|
}
|
|
|
|
delete this.offset;
|
|
} else {
|
|
this.range = this.source.range;
|
|
this.linePos = this.source.rangeAsLinePos;
|
|
}
|
|
|
|
if (this.linePos) {
|
|
const {
|
|
line,
|
|
col
|
|
} = this.linePos.start;
|
|
this.message += ` at line ${line}, column ${col}`;
|
|
const ctx = cst && getPrettyContext(this.linePos, cst);
|
|
if (ctx) this.message += `:\n\n${ctx}\n`;
|
|
}
|
|
|
|
delete this.source;
|
|
}
|
|
|
|
}
|
|
class YAMLReferenceError extends YAMLError {
|
|
constructor(source, message) {
|
|
super('YAMLReferenceError', source, message);
|
|
}
|
|
|
|
}
|
|
class YAMLSemanticError extends YAMLError {
|
|
constructor(source, message) {
|
|
super('YAMLSemanticError', source, message);
|
|
}
|
|
|
|
}
|
|
class YAMLSyntaxError extends YAMLError {
|
|
constructor(source, message) {
|
|
super('YAMLSyntaxError', source, message);
|
|
}
|
|
|
|
}
|
|
class YAMLWarning extends YAMLError {
|
|
constructor(source, message) {
|
|
super('YAMLWarning', source, message);
|
|
}
|
|
|
|
}
|
|
|
|
function _defineProperty(obj, key, value) {
|
|
if (key in obj) {
|
|
Object.defineProperty(obj, key, {
|
|
value: value,
|
|
enumerable: true,
|
|
configurable: true,
|
|
writable: true
|
|
});
|
|
} else {
|
|
obj[key] = value;
|
|
}
|
|
|
|
return obj;
|
|
}
|
|
|
|
class PlainValue extends Node {
|
|
static endOfLine(src, start, inFlow) {
|
|
let ch = src[start];
|
|
let offset = start;
|
|
|
|
while (ch && ch !== '\n') {
|
|
if (inFlow && (ch === '[' || ch === ']' || ch === '{' || ch === '}' || ch === ',')) break;
|
|
const next = src[offset + 1];
|
|
if (ch === ':' && (!next || next === '\n' || next === '\t' || next === ' ' || inFlow && next === ',')) break;
|
|
if ((ch === ' ' || ch === '\t') && next === '#') break;
|
|
offset += 1;
|
|
ch = next;
|
|
}
|
|
|
|
return offset;
|
|
}
|
|
|
|
get strValue() {
|
|
if (!this.valueRange || !this.context) return null;
|
|
let {
|
|
start,
|
|
end
|
|
} = this.valueRange;
|
|
const {
|
|
src
|
|
} = this.context;
|
|
let ch = src[end - 1];
|
|
|
|
while (start < end && (ch === '\n' || ch === '\t' || ch === ' ')) ch = src[--end - 1];
|
|
|
|
let str = '';
|
|
|
|
for (let i = start; i < end; ++i) {
|
|
const ch = src[i];
|
|
|
|
if (ch === '\n') {
|
|
const {
|
|
fold,
|
|
offset
|
|
} = Node.foldNewline(src, i, -1);
|
|
str += fold;
|
|
i = offset;
|
|
} else if (ch === ' ' || ch === '\t') {
|
|
// trim trailing whitespace
|
|
const wsStart = i;
|
|
let next = src[i + 1];
|
|
|
|
while (i < end && (next === ' ' || next === '\t')) {
|
|
i += 1;
|
|
next = src[i + 1];
|
|
}
|
|
|
|
if (next !== '\n') str += i > wsStart ? src.slice(wsStart, i + 1) : ch;
|
|
} else {
|
|
str += ch;
|
|
}
|
|
}
|
|
|
|
const ch0 = src[start];
|
|
|
|
switch (ch0) {
|
|
case '\t':
|
|
{
|
|
const msg = 'Plain value cannot start with a tab character';
|
|
const errors = [new YAMLSemanticError(this, msg)];
|
|
return {
|
|
errors,
|
|
str
|
|
};
|
|
}
|
|
|
|
case '@':
|
|
case '`':
|
|
{
|
|
const msg = `Plain value cannot start with reserved character ${ch0}`;
|
|
const errors = [new YAMLSemanticError(this, msg)];
|
|
return {
|
|
errors,
|
|
str
|
|
};
|
|
}
|
|
|
|
default:
|
|
return str;
|
|
}
|
|
}
|
|
|
|
parseBlockValue(start) {
|
|
const {
|
|
indent,
|
|
inFlow,
|
|
src
|
|
} = this.context;
|
|
let offset = start;
|
|
let valueEnd = start;
|
|
|
|
for (let ch = src[offset]; ch === '\n'; ch = src[offset]) {
|
|
if (Node.atDocumentBoundary(src, offset + 1)) break;
|
|
const end = Node.endOfBlockIndent(src, indent, offset + 1);
|
|
if (end === null || src[end] === '#') break;
|
|
|
|
if (src[end] === '\n') {
|
|
offset = end;
|
|
} else {
|
|
valueEnd = PlainValue.endOfLine(src, end, inFlow);
|
|
offset = valueEnd;
|
|
}
|
|
}
|
|
|
|
if (this.valueRange.isEmpty()) this.valueRange.start = start;
|
|
this.valueRange.end = valueEnd;
|
|
return valueEnd;
|
|
}
|
|
/**
|
|
* Parses a plain value from the source
|
|
*
|
|
* Accepted forms are:
|
|
* ```
|
|
* #comment
|
|
*
|
|
* first line
|
|
*
|
|
* first line #comment
|
|
*
|
|
* first line
|
|
* block
|
|
* lines
|
|
*
|
|
* #comment
|
|
* block
|
|
* lines
|
|
* ```
|
|
* where block lines are empty or have an indent level greater than `indent`.
|
|
*
|
|
* @param {ParseContext} context
|
|
* @param {number} start - Index of first character
|
|
* @returns {number} - Index of the character after this scalar, may be `\n`
|
|
*/
|
|
|
|
|
|
parse(context, start) {
|
|
this.context = context;
|
|
const {
|
|
inFlow,
|
|
src
|
|
} = context;
|
|
let offset = start;
|
|
const ch = src[offset];
|
|
|
|
if (ch && ch !== '#' && ch !== '\n') {
|
|
offset = PlainValue.endOfLine(src, start, inFlow);
|
|
}
|
|
|
|
this.valueRange = new Range(start, offset);
|
|
offset = Node.endOfWhiteSpace(src, offset);
|
|
offset = this.parseComment(offset);
|
|
|
|
if (!this.hasComment || this.valueRange.isEmpty()) {
|
|
offset = this.parseBlockValue(offset);
|
|
}
|
|
|
|
return offset;
|
|
}
|
|
|
|
}
|
|
|
|
exports.Char = Char;
|
|
exports.Node = Node;
|
|
exports.PlainValue = PlainValue;
|
|
exports.Range = Range;
|
|
exports.Type = Type;
|
|
exports.YAMLError = YAMLError;
|
|
exports.YAMLReferenceError = YAMLReferenceError;
|
|
exports.YAMLSemanticError = YAMLSemanticError;
|
|
exports.YAMLSyntaxError = YAMLSyntaxError;
|
|
exports.YAMLWarning = YAMLWarning;
|
|
exports._defineProperty = _defineProperty;
|
|
exports.defaultTagPrefix = defaultTagPrefix;
|
|
exports.defaultTags = defaultTags;
|
|
|
|
|
|
/***/ }),
|
|
|
|
/***/ 8021:
|
|
/***/ ((__unused_webpack_module, exports, __nccwpck_require__) => {
|
|
|
|
"use strict";
|
|
|
|
|
|
var PlainValue = __nccwpck_require__(5215);
|
|
var resolveSeq = __nccwpck_require__(4227);
|
|
var warnings = __nccwpck_require__(6003);
|
|
|
|
function createMap(schema, obj, ctx) {
|
|
const map = new resolveSeq.YAMLMap(schema);
|
|
|
|
if (obj instanceof Map) {
|
|
for (const [key, value] of obj) map.items.push(schema.createPair(key, value, ctx));
|
|
} else if (obj && typeof obj === 'object') {
|
|
for (const key of Object.keys(obj)) map.items.push(schema.createPair(key, obj[key], ctx));
|
|
}
|
|
|
|
if (typeof schema.sortMapEntries === 'function') {
|
|
map.items.sort(schema.sortMapEntries);
|
|
}
|
|
|
|
return map;
|
|
}
|
|
|
|
const map = {
|
|
createNode: createMap,
|
|
default: true,
|
|
nodeClass: resolveSeq.YAMLMap,
|
|
tag: 'tag:yaml.org,2002:map',
|
|
resolve: resolveSeq.resolveMap
|
|
};
|
|
|
|
function createSeq(schema, obj, ctx) {
|
|
const seq = new resolveSeq.YAMLSeq(schema);
|
|
|
|
if (obj && obj[Symbol.iterator]) {
|
|
for (const it of obj) {
|
|
const v = schema.createNode(it, ctx.wrapScalars, null, ctx);
|
|
seq.items.push(v);
|
|
}
|
|
}
|
|
|
|
return seq;
|
|
}
|
|
|
|
const seq = {
|
|
createNode: createSeq,
|
|
default: true,
|
|
nodeClass: resolveSeq.YAMLSeq,
|
|
tag: 'tag:yaml.org,2002:seq',
|
|
resolve: resolveSeq.resolveSeq
|
|
};
|
|
|
|
const string = {
|
|
identify: value => typeof value === 'string',
|
|
default: true,
|
|
tag: 'tag:yaml.org,2002:str',
|
|
resolve: resolveSeq.resolveString,
|
|
|
|
stringify(item, ctx, onComment, onChompKeep) {
|
|
ctx = Object.assign({
|
|
actualString: true
|
|
}, ctx);
|
|
return resolveSeq.stringifyString(item, ctx, onComment, onChompKeep);
|
|
},
|
|
|
|
options: resolveSeq.strOptions
|
|
};
|
|
|
|
const failsafe = [map, seq, string];
|
|
|
|
/* global BigInt */
|
|
|
|
const intIdentify$2 = value => typeof value === 'bigint' || Number.isInteger(value);
|
|
|
|
const intResolve$1 = (src, part, radix) => resolveSeq.intOptions.asBigInt ? BigInt(src) : parseInt(part, radix);
|
|
|
|
function intStringify$1(node, radix, prefix) {
|
|
const {
|
|
value
|
|
} = node;
|
|
if (intIdentify$2(value) && value >= 0) return prefix + value.toString(radix);
|
|
return resolveSeq.stringifyNumber(node);
|
|
}
|
|
|
|
const nullObj = {
|
|
identify: value => value == null,
|
|
createNode: (schema, value, ctx) => ctx.wrapScalars ? new resolveSeq.Scalar(null) : null,
|
|
default: true,
|
|
tag: 'tag:yaml.org,2002:null',
|
|
test: /^(?:~|[Nn]ull|NULL)?$/,
|
|
resolve: () => null,
|
|
options: resolveSeq.nullOptions,
|
|
stringify: () => resolveSeq.nullOptions.nullStr
|
|
};
|
|
const boolObj = {
|
|
identify: value => typeof value === 'boolean',
|
|
default: true,
|
|
tag: 'tag:yaml.org,2002:bool',
|
|
test: /^(?:[Tt]rue|TRUE|[Ff]alse|FALSE)$/,
|
|
resolve: str => str[0] === 't' || str[0] === 'T',
|
|
options: resolveSeq.boolOptions,
|
|
stringify: ({
|
|
value
|
|
}) => value ? resolveSeq.boolOptions.trueStr : resolveSeq.boolOptions.falseStr
|
|
};
|
|
const octObj = {
|
|
identify: value => intIdentify$2(value) && value >= 0,
|
|
default: true,
|
|
tag: 'tag:yaml.org,2002:int',
|
|
format: 'OCT',
|
|
test: /^0o([0-7]+)$/,
|
|
resolve: (str, oct) => intResolve$1(str, oct, 8),
|
|
options: resolveSeq.intOptions,
|
|
stringify: node => intStringify$1(node, 8, '0o')
|
|
};
|
|
const intObj = {
|
|
identify: intIdentify$2,
|
|
default: true,
|
|
tag: 'tag:yaml.org,2002:int',
|
|
test: /^[-+]?[0-9]+$/,
|
|
resolve: str => intResolve$1(str, str, 10),
|
|
options: resolveSeq.intOptions,
|
|
stringify: resolveSeq.stringifyNumber
|
|
};
|
|
const hexObj = {
|
|
identify: value => intIdentify$2(value) && value >= 0,
|
|
default: true,
|
|
tag: 'tag:yaml.org,2002:int',
|
|
format: 'HEX',
|
|
test: /^0x([0-9a-fA-F]+)$/,
|
|
resolve: (str, hex) => intResolve$1(str, hex, 16),
|
|
options: resolveSeq.intOptions,
|
|
stringify: node => intStringify$1(node, 16, '0x')
|
|
};
|
|
const nanObj = {
|
|
identify: value => typeof value === 'number',
|
|
default: true,
|
|
tag: 'tag:yaml.org,2002:float',
|
|
test: /^(?:[-+]?\.inf|(\.nan))$/i,
|
|
resolve: (str, nan) => nan ? NaN : str[0] === '-' ? Number.NEGATIVE_INFINITY : Number.POSITIVE_INFINITY,
|
|
stringify: resolveSeq.stringifyNumber
|
|
};
|
|
const expObj = {
|
|
identify: value => typeof value === 'number',
|
|
default: true,
|
|
tag: 'tag:yaml.org,2002:float',
|
|
format: 'EXP',
|
|
test: /^[-+]?(?:\.[0-9]+|[0-9]+(?:\.[0-9]*)?)[eE][-+]?[0-9]+$/,
|
|
resolve: str => parseFloat(str),
|
|
stringify: ({
|
|
value
|
|
}) => Number(value).toExponential()
|
|
};
|
|
const floatObj = {
|
|
identify: value => typeof value === 'number',
|
|
default: true,
|
|
tag: 'tag:yaml.org,2002:float',
|
|
test: /^[-+]?(?:\.([0-9]+)|[0-9]+\.([0-9]*))$/,
|
|
|
|
resolve(str, frac1, frac2) {
|
|
const frac = frac1 || frac2;
|
|
const node = new resolveSeq.Scalar(parseFloat(str));
|
|
if (frac && frac[frac.length - 1] === '0') node.minFractionDigits = frac.length;
|
|
return node;
|
|
},
|
|
|
|
stringify: resolveSeq.stringifyNumber
|
|
};
|
|
const core = failsafe.concat([nullObj, boolObj, octObj, intObj, hexObj, nanObj, expObj, floatObj]);
|
|
|
|
/* global BigInt */
|
|
|
|
const intIdentify$1 = value => typeof value === 'bigint' || Number.isInteger(value);
|
|
|
|
const stringifyJSON = ({
|
|
value
|
|
}) => JSON.stringify(value);
|
|
|
|
const json = [map, seq, {
|
|
identify: value => typeof value === 'string',
|
|
default: true,
|
|
tag: 'tag:yaml.org,2002:str',
|
|
resolve: resolveSeq.resolveString,
|
|
stringify: stringifyJSON
|
|
}, {
|
|
identify: value => value == null,
|
|
createNode: (schema, value, ctx) => ctx.wrapScalars ? new resolveSeq.Scalar(null) : null,
|
|
default: true,
|
|
tag: 'tag:yaml.org,2002:null',
|
|
test: /^null$/,
|
|
resolve: () => null,
|
|
stringify: stringifyJSON
|
|
}, {
|
|
identify: value => typeof value === 'boolean',
|
|
default: true,
|
|
tag: 'tag:yaml.org,2002:bool',
|
|
test: /^true|false$/,
|
|
resolve: str => str === 'true',
|
|
stringify: stringifyJSON
|
|
}, {
|
|
identify: intIdentify$1,
|
|
default: true,
|
|
tag: 'tag:yaml.org,2002:int',
|
|
test: /^-?(?:0|[1-9][0-9]*)$/,
|
|
resolve: str => resolveSeq.intOptions.asBigInt ? BigInt(str) : parseInt(str, 10),
|
|
stringify: ({
|
|
value
|
|
}) => intIdentify$1(value) ? value.toString() : JSON.stringify(value)
|
|
}, {
|
|
identify: value => typeof value === 'number',
|
|
default: true,
|
|
tag: 'tag:yaml.org,2002:float',
|
|
test: /^-?(?:0|[1-9][0-9]*)(?:\.[0-9]*)?(?:[eE][-+]?[0-9]+)?$/,
|
|
resolve: str => parseFloat(str),
|
|
stringify: stringifyJSON
|
|
}];
|
|
|
|
json.scalarFallback = str => {
|
|
throw new SyntaxError(`Unresolved plain scalar ${JSON.stringify(str)}`);
|
|
};
|
|
|
|
/* global BigInt */
|
|
|
|
const boolStringify = ({
|
|
value
|
|
}) => value ? resolveSeq.boolOptions.trueStr : resolveSeq.boolOptions.falseStr;
|
|
|
|
const intIdentify = value => typeof value === 'bigint' || Number.isInteger(value);
|
|
|
|
function intResolve(sign, src, radix) {
|
|
let str = src.replace(/_/g, '');
|
|
|
|
if (resolveSeq.intOptions.asBigInt) {
|
|
switch (radix) {
|
|
case 2:
|
|
str = `0b${str}`;
|
|
break;
|
|
|
|
case 8:
|
|
str = `0o${str}`;
|
|
break;
|
|
|
|
case 16:
|
|
str = `0x${str}`;
|
|
break;
|
|
}
|
|
|
|
const n = BigInt(str);
|
|
return sign === '-' ? BigInt(-1) * n : n;
|
|
}
|
|
|
|
const n = parseInt(str, radix);
|
|
return sign === '-' ? -1 * n : n;
|
|
}
|
|
|
|
function intStringify(node, radix, prefix) {
|
|
const {
|
|
value
|
|
} = node;
|
|
|
|
if (intIdentify(value)) {
|
|
const str = value.toString(radix);
|
|
return value < 0 ? '-' + prefix + str.substr(1) : prefix + str;
|
|
}
|
|
|
|
return resolveSeq.stringifyNumber(node);
|
|
}
|
|
|
|
const yaml11 = failsafe.concat([{
|
|
identify: value => value == null,
|
|
createNode: (schema, value, ctx) => ctx.wrapScalars ? new resolveSeq.Scalar(null) : null,
|
|
default: true,
|
|
tag: 'tag:yaml.org,2002:null',
|
|
test: /^(?:~|[Nn]ull|NULL)?$/,
|
|
resolve: () => null,
|
|
options: resolveSeq.nullOptions,
|
|
stringify: () => resolveSeq.nullOptions.nullStr
|
|
}, {
|
|
identify: value => typeof value === 'boolean',
|
|
default: true,
|
|
tag: 'tag:yaml.org,2002:bool',
|
|
test: /^(?:Y|y|[Yy]es|YES|[Tt]rue|TRUE|[Oo]n|ON)$/,
|
|
resolve: () => true,
|
|
options: resolveSeq.boolOptions,
|
|
stringify: boolStringify
|
|
}, {
|
|
identify: value => typeof value === 'boolean',
|
|
default: true,
|
|
tag: 'tag:yaml.org,2002:bool',
|
|
test: /^(?:N|n|[Nn]o|NO|[Ff]alse|FALSE|[Oo]ff|OFF)$/i,
|
|
resolve: () => false,
|
|
options: resolveSeq.boolOptions,
|
|
stringify: boolStringify
|
|
}, {
|
|
identify: intIdentify,
|
|
default: true,
|
|
tag: 'tag:yaml.org,2002:int',
|
|
format: 'BIN',
|
|
test: /^([-+]?)0b([0-1_]+)$/,
|
|
resolve: (str, sign, bin) => intResolve(sign, bin, 2),
|
|
stringify: node => intStringify(node, 2, '0b')
|
|
}, {
|
|
identify: intIdentify,
|
|
default: true,
|
|
tag: 'tag:yaml.org,2002:int',
|
|
format: 'OCT',
|
|
test: /^([-+]?)0([0-7_]+)$/,
|
|
resolve: (str, sign, oct) => intResolve(sign, oct, 8),
|
|
stringify: node => intStringify(node, 8, '0')
|
|
}, {
|
|
identify: intIdentify,
|
|
default: true,
|
|
tag: 'tag:yaml.org,2002:int',
|
|
test: /^([-+]?)([0-9][0-9_]*)$/,
|
|
resolve: (str, sign, abs) => intResolve(sign, abs, 10),
|
|
stringify: resolveSeq.stringifyNumber
|
|
}, {
|
|
identify: intIdentify,
|
|
default: true,
|
|
tag: 'tag:yaml.org,2002:int',
|
|
format: 'HEX',
|
|
test: /^([-+]?)0x([0-9a-fA-F_]+)$/,
|
|
resolve: (str, sign, hex) => intResolve(sign, hex, 16),
|
|
stringify: node => intStringify(node, 16, '0x')
|
|
}, {
|
|
identify: value => typeof value === 'number',
|
|
default: true,
|
|
tag: 'tag:yaml.org,2002:float',
|
|
test: /^(?:[-+]?\.inf|(\.nan))$/i,
|
|
resolve: (str, nan) => nan ? NaN : str[0] === '-' ? Number.NEGATIVE_INFINITY : Number.POSITIVE_INFINITY,
|
|
stringify: resolveSeq.stringifyNumber
|
|
}, {
|
|
identify: value => typeof value === 'number',
|
|
default: true,
|
|
tag: 'tag:yaml.org,2002:float',
|
|
format: 'EXP',
|
|
test: /^[-+]?([0-9][0-9_]*)?(\.[0-9_]*)?[eE][-+]?[0-9]+$/,
|
|
resolve: str => parseFloat(str.replace(/_/g, '')),
|
|
stringify: ({
|
|
value
|
|
}) => Number(value).toExponential()
|
|
}, {
|
|
identify: value => typeof value === 'number',
|
|
default: true,
|
|
tag: 'tag:yaml.org,2002:float',
|
|
test: /^[-+]?(?:[0-9][0-9_]*)?\.([0-9_]*)$/,
|
|
|
|
resolve(str, frac) {
|
|
const node = new resolveSeq.Scalar(parseFloat(str.replace(/_/g, '')));
|
|
|
|
if (frac) {
|
|
const f = frac.replace(/_/g, '');
|
|
if (f[f.length - 1] === '0') node.minFractionDigits = f.length;
|
|
}
|
|
|
|
return node;
|
|
},
|
|
|
|
stringify: resolveSeq.stringifyNumber
|
|
}], warnings.binary, warnings.omap, warnings.pairs, warnings.set, warnings.intTime, warnings.floatTime, warnings.timestamp);
|
|
|
|
const schemas = {
|
|
core,
|
|
failsafe,
|
|
json,
|
|
yaml11
|
|
};
|
|
const tags = {
|
|
binary: warnings.binary,
|
|
bool: boolObj,
|
|
float: floatObj,
|
|
floatExp: expObj,
|
|
floatNaN: nanObj,
|
|
floatTime: warnings.floatTime,
|
|
int: intObj,
|
|
intHex: hexObj,
|
|
intOct: octObj,
|
|
intTime: warnings.intTime,
|
|
map,
|
|
null: nullObj,
|
|
omap: warnings.omap,
|
|
pairs: warnings.pairs,
|
|
seq,
|
|
set: warnings.set,
|
|
timestamp: warnings.timestamp
|
|
};
|
|
|
|
function findTagObject(value, tagName, tags) {
|
|
if (tagName) {
|
|
const match = tags.filter(t => t.tag === tagName);
|
|
const tagObj = match.find(t => !t.format) || match[0];
|
|
if (!tagObj) throw new Error(`Tag ${tagName} not found`);
|
|
return tagObj;
|
|
} // TODO: deprecate/remove class check
|
|
|
|
|
|
return tags.find(t => (t.identify && t.identify(value) || t.class && value instanceof t.class) && !t.format);
|
|
}
|
|
|
|
function createNode(value, tagName, ctx) {
|
|
if (value instanceof resolveSeq.Node) return value;
|
|
const {
|
|
defaultPrefix,
|
|
onTagObj,
|
|
prevObjects,
|
|
schema,
|
|
wrapScalars
|
|
} = ctx;
|
|
if (tagName && tagName.startsWith('!!')) tagName = defaultPrefix + tagName.slice(2);
|
|
let tagObj = findTagObject(value, tagName, schema.tags);
|
|
|
|
if (!tagObj) {
|
|
if (typeof value.toJSON === 'function') value = value.toJSON();
|
|
if (!value || typeof value !== 'object') return wrapScalars ? new resolveSeq.Scalar(value) : value;
|
|
tagObj = value instanceof Map ? map : value[Symbol.iterator] ? seq : map;
|
|
}
|
|
|
|
if (onTagObj) {
|
|
onTagObj(tagObj);
|
|
delete ctx.onTagObj;
|
|
} // Detect duplicate references to the same object & use Alias nodes for all
|
|
// after first. The `obj` wrapper allows for circular references to resolve.
|
|
|
|
|
|
const obj = {
|
|
value: undefined,
|
|
node: undefined
|
|
};
|
|
|
|
if (value && typeof value === 'object' && prevObjects) {
|
|
const prev = prevObjects.get(value);
|
|
|
|
if (prev) {
|
|
const alias = new resolveSeq.Alias(prev); // leaves source dirty; must be cleaned by caller
|
|
|
|
ctx.aliasNodes.push(alias); // defined along with prevObjects
|
|
|
|
return alias;
|
|
}
|
|
|
|
obj.value = value;
|
|
prevObjects.set(value, obj);
|
|
}
|
|
|
|
obj.node = tagObj.createNode ? tagObj.createNode(ctx.schema, value, ctx) : wrapScalars ? new resolveSeq.Scalar(value) : value;
|
|
if (tagName && obj.node instanceof resolveSeq.Node) obj.node.tag = tagName;
|
|
return obj.node;
|
|
}
|
|
|
|
function getSchemaTags(schemas, knownTags, customTags, schemaId) {
|
|
let tags = schemas[schemaId.replace(/\W/g, '')]; // 'yaml-1.1' -> 'yaml11'
|
|
|
|
if (!tags) {
|
|
const keys = Object.keys(schemas).map(key => JSON.stringify(key)).join(', ');
|
|
throw new Error(`Unknown schema "${schemaId}"; use one of ${keys}`);
|
|
}
|
|
|
|
if (Array.isArray(customTags)) {
|
|
for (const tag of customTags) tags = tags.concat(tag);
|
|
} else if (typeof customTags === 'function') {
|
|
tags = customTags(tags.slice());
|
|
}
|
|
|
|
for (let i = 0; i < tags.length; ++i) {
|
|
const tag = tags[i];
|
|
|
|
if (typeof tag === 'string') {
|
|
const tagObj = knownTags[tag];
|
|
|
|
if (!tagObj) {
|
|
const keys = Object.keys(knownTags).map(key => JSON.stringify(key)).join(', ');
|
|
throw new Error(`Unknown custom tag "${tag}"; use one of ${keys}`);
|
|
}
|
|
|
|
tags[i] = tagObj;
|
|
}
|
|
}
|
|
|
|
return tags;
|
|
}
|
|
|
|
const sortMapEntriesByKey = (a, b) => a.key < b.key ? -1 : a.key > b.key ? 1 : 0;
|
|
|
|
class Schema {
|
|
// TODO: remove in v2
|
|
// TODO: remove in v2
|
|
constructor({
|
|
customTags,
|
|
merge,
|
|
schema,
|
|
sortMapEntries,
|
|
tags: deprecatedCustomTags
|
|
}) {
|
|
this.merge = !!merge;
|
|
this.name = schema;
|
|
this.sortMapEntries = sortMapEntries === true ? sortMapEntriesByKey : sortMapEntries || null;
|
|
if (!customTags && deprecatedCustomTags) warnings.warnOptionDeprecation('tags', 'customTags');
|
|
this.tags = getSchemaTags(schemas, tags, customTags || deprecatedCustomTags, schema);
|
|
}
|
|
|
|
createNode(value, wrapScalars, tagName, ctx) {
|
|
const baseCtx = {
|
|
defaultPrefix: Schema.defaultPrefix,
|
|
schema: this,
|
|
wrapScalars
|
|
};
|
|
const createCtx = ctx ? Object.assign(ctx, baseCtx) : baseCtx;
|
|
return createNode(value, tagName, createCtx);
|
|
}
|
|
|
|
createPair(key, value, ctx) {
|
|
if (!ctx) ctx = {
|
|
wrapScalars: true
|
|
};
|
|
const k = this.createNode(key, ctx.wrapScalars, null, ctx);
|
|
const v = this.createNode(value, ctx.wrapScalars, null, ctx);
|
|
return new resolveSeq.Pair(k, v);
|
|
}
|
|
|
|
}
|
|
|
|
PlainValue._defineProperty(Schema, "defaultPrefix", PlainValue.defaultTagPrefix);
|
|
|
|
PlainValue._defineProperty(Schema, "defaultTags", PlainValue.defaultTags);
|
|
|
|
exports.Schema = Schema;
|
|
|
|
|
|
/***/ }),
|
|
|
|
/***/ 5065:
|
|
/***/ ((__unused_webpack_module, exports, __nccwpck_require__) => {
|
|
|
|
"use strict";
|
|
|
|
|
|
var parseCst = __nccwpck_require__(445);
|
|
var Document$1 = __nccwpck_require__(5506);
|
|
var Schema = __nccwpck_require__(8021);
|
|
var PlainValue = __nccwpck_require__(5215);
|
|
var warnings = __nccwpck_require__(6003);
|
|
__nccwpck_require__(4227);
|
|
|
|
function createNode(value, wrapScalars = true, tag) {
|
|
if (tag === undefined && typeof wrapScalars === 'string') {
|
|
tag = wrapScalars;
|
|
wrapScalars = true;
|
|
}
|
|
|
|
const options = Object.assign({}, Document$1.Document.defaults[Document$1.defaultOptions.version], Document$1.defaultOptions);
|
|
const schema = new Schema.Schema(options);
|
|
return schema.createNode(value, wrapScalars, tag);
|
|
}
|
|
|
|
class Document extends Document$1.Document {
|
|
constructor(options) {
|
|
super(Object.assign({}, Document$1.defaultOptions, options));
|
|
}
|
|
|
|
}
|
|
|
|
function parseAllDocuments(src, options) {
|
|
const stream = [];
|
|
let prev;
|
|
|
|
for (const cstDoc of parseCst.parse(src)) {
|
|
const doc = new Document(options);
|
|
doc.parse(cstDoc, prev);
|
|
stream.push(doc);
|
|
prev = doc;
|
|
}
|
|
|
|
return stream;
|
|
}
|
|
|
|
function parseDocument(src, options) {
|
|
const cst = parseCst.parse(src);
|
|
const doc = new Document(options).parse(cst[0]);
|
|
|
|
if (cst.length > 1) {
|
|
const errMsg = 'Source contains multiple documents; please use YAML.parseAllDocuments()';
|
|
doc.errors.unshift(new PlainValue.YAMLSemanticError(cst[1], errMsg));
|
|
}
|
|
|
|
return doc;
|
|
}
|
|
|
|
function parse(src, options) {
|
|
const doc = parseDocument(src, options);
|
|
doc.warnings.forEach(warning => warnings.warn(warning));
|
|
if (doc.errors.length > 0) throw doc.errors[0];
|
|
return doc.toJSON();
|
|
}
|
|
|
|
function stringify(value, options) {
|
|
const doc = new Document(options);
|
|
doc.contents = value;
|
|
return String(doc);
|
|
}
|
|
|
|
const YAML = {
|
|
createNode,
|
|
defaultOptions: Document$1.defaultOptions,
|
|
Document,
|
|
parse,
|
|
parseAllDocuments,
|
|
parseCST: parseCst.parse,
|
|
parseDocument,
|
|
scalarOptions: Document$1.scalarOptions,
|
|
stringify
|
|
};
|
|
|
|
exports.YAML = YAML;
|
|
|
|
|
|
/***/ }),
|
|
|
|
/***/ 445:
|
|
/***/ ((__unused_webpack_module, exports, __nccwpck_require__) => {
|
|
|
|
"use strict";
|
|
|
|
|
|
var PlainValue = __nccwpck_require__(5215);
|
|
|
|
class BlankLine extends PlainValue.Node {
|
|
constructor() {
|
|
super(PlainValue.Type.BLANK_LINE);
|
|
}
|
|
/* istanbul ignore next */
|
|
|
|
|
|
get includesTrailingLines() {
|
|
// This is never called from anywhere, but if it were,
|
|
// this is the value it should return.
|
|
return true;
|
|
}
|
|
/**
|
|
* Parses a blank line from the source
|
|
*
|
|
* @param {ParseContext} context
|
|
* @param {number} start - Index of first \n character
|
|
* @returns {number} - Index of the character after this
|
|
*/
|
|
|
|
|
|
parse(context, start) {
|
|
this.context = context;
|
|
this.range = new PlainValue.Range(start, start + 1);
|
|
return start + 1;
|
|
}
|
|
|
|
}
|
|
|
|
class CollectionItem extends PlainValue.Node {
|
|
constructor(type, props) {
|
|
super(type, props);
|
|
this.node = null;
|
|
}
|
|
|
|
get includesTrailingLines() {
|
|
return !!this.node && this.node.includesTrailingLines;
|
|
}
|
|
/**
|
|
* @param {ParseContext} context
|
|
* @param {number} start - Index of first character
|
|
* @returns {number} - Index of the character after this
|
|
*/
|
|
|
|
|
|
parse(context, start) {
|
|
this.context = context;
|
|
const {
|
|
parseNode,
|
|
src
|
|
} = context;
|
|
let {
|
|
atLineStart,
|
|
lineStart
|
|
} = context;
|
|
if (!atLineStart && this.type === PlainValue.Type.SEQ_ITEM) this.error = new PlainValue.YAMLSemanticError(this, 'Sequence items must not have preceding content on the same line');
|
|
const indent = atLineStart ? start - lineStart : context.indent;
|
|
let offset = PlainValue.Node.endOfWhiteSpace(src, start + 1);
|
|
let ch = src[offset];
|
|
const inlineComment = ch === '#';
|
|
const comments = [];
|
|
let blankLine = null;
|
|
|
|
while (ch === '\n' || ch === '#') {
|
|
if (ch === '#') {
|
|
const end = PlainValue.Node.endOfLine(src, offset + 1);
|
|
comments.push(new PlainValue.Range(offset, end));
|
|
offset = end;
|
|
} else {
|
|
atLineStart = true;
|
|
lineStart = offset + 1;
|
|
const wsEnd = PlainValue.Node.endOfWhiteSpace(src, lineStart);
|
|
|
|
if (src[wsEnd] === '\n' && comments.length === 0) {
|
|
blankLine = new BlankLine();
|
|
lineStart = blankLine.parse({
|
|
src
|
|
}, lineStart);
|
|
}
|
|
|
|
offset = PlainValue.Node.endOfIndent(src, lineStart);
|
|
}
|
|
|
|
ch = src[offset];
|
|
}
|
|
|
|
if (PlainValue.Node.nextNodeIsIndented(ch, offset - (lineStart + indent), this.type !== PlainValue.Type.SEQ_ITEM)) {
|
|
this.node = parseNode({
|
|
atLineStart,
|
|
inCollection: false,
|
|
indent,
|
|
lineStart,
|
|
parent: this
|
|
}, offset);
|
|
} else if (ch && lineStart > start + 1) {
|
|
offset = lineStart - 1;
|
|
}
|
|
|
|
if (this.node) {
|
|
if (blankLine) {
|
|
// Only blank lines preceding non-empty nodes are captured. Note that
|
|
// this means that collection item range start indices do not always
|
|
// increase monotonically. -- eemeli/yaml#126
|
|
const items = context.parent.items || context.parent.contents;
|
|
if (items) items.push(blankLine);
|
|
}
|
|
|
|
if (comments.length) Array.prototype.push.apply(this.props, comments);
|
|
offset = this.node.range.end;
|
|
} else {
|
|
if (inlineComment) {
|
|
const c = comments[0];
|
|
this.props.push(c);
|
|
offset = c.end;
|
|
} else {
|
|
offset = PlainValue.Node.endOfLine(src, start + 1);
|
|
}
|
|
}
|
|
|
|
const end = this.node ? this.node.valueRange.end : offset;
|
|
this.valueRange = new PlainValue.Range(start, end);
|
|
return offset;
|
|
}
|
|
|
|
setOrigRanges(cr, offset) {
|
|
offset = super.setOrigRanges(cr, offset);
|
|
return this.node ? this.node.setOrigRanges(cr, offset) : offset;
|
|
}
|
|
|
|
toString() {
|
|
const {
|
|
context: {
|
|
src
|
|
},
|
|
node,
|
|
range,
|
|
value
|
|
} = this;
|
|
if (value != null) return value;
|
|
const str = node ? src.slice(range.start, node.range.start) + String(node) : src.slice(range.start, range.end);
|
|
return PlainValue.Node.addStringTerminator(src, range.end, str);
|
|
}
|
|
|
|
}
|
|
|
|
class Comment extends PlainValue.Node {
|
|
constructor() {
|
|
super(PlainValue.Type.COMMENT);
|
|
}
|
|
/**
|
|
* Parses a comment line from the source
|
|
*
|
|
* @param {ParseContext} context
|
|
* @param {number} start - Index of first character
|
|
* @returns {number} - Index of the character after this scalar
|
|
*/
|
|
|
|
|
|
parse(context, start) {
|
|
this.context = context;
|
|
const offset = this.parseComment(start);
|
|
this.range = new PlainValue.Range(start, offset);
|
|
return offset;
|
|
}
|
|
|
|
}
|
|
|
|
function grabCollectionEndComments(node) {
|
|
let cnode = node;
|
|
|
|
while (cnode instanceof CollectionItem) cnode = cnode.node;
|
|
|
|
if (!(cnode instanceof Collection)) return null;
|
|
const len = cnode.items.length;
|
|
let ci = -1;
|
|
|
|
for (let i = len - 1; i >= 0; --i) {
|
|
const n = cnode.items[i];
|
|
|
|
if (n.type === PlainValue.Type.COMMENT) {
|
|
// Keep sufficiently indented comments with preceding node
|
|
const {
|
|
indent,
|
|
lineStart
|
|
} = n.context;
|
|
if (indent > 0 && n.range.start >= lineStart + indent) break;
|
|
ci = i;
|
|
} else if (n.type === PlainValue.Type.BLANK_LINE) ci = i;else break;
|
|
}
|
|
|
|
if (ci === -1) return null;
|
|
const ca = cnode.items.splice(ci, len - ci);
|
|
const prevEnd = ca[0].range.start;
|
|
|
|
while (true) {
|
|
cnode.range.end = prevEnd;
|
|
if (cnode.valueRange && cnode.valueRange.end > prevEnd) cnode.valueRange.end = prevEnd;
|
|
if (cnode === node) break;
|
|
cnode = cnode.context.parent;
|
|
}
|
|
|
|
return ca;
|
|
}
|
|
class Collection extends PlainValue.Node {
|
|
static nextContentHasIndent(src, offset, indent) {
|
|
const lineStart = PlainValue.Node.endOfLine(src, offset) + 1;
|
|
offset = PlainValue.Node.endOfWhiteSpace(src, lineStart);
|
|
const ch = src[offset];
|
|
if (!ch) return false;
|
|
if (offset >= lineStart + indent) return true;
|
|
if (ch !== '#' && ch !== '\n') return false;
|
|
return Collection.nextContentHasIndent(src, offset, indent);
|
|
}
|
|
|
|
constructor(firstItem) {
|
|
super(firstItem.type === PlainValue.Type.SEQ_ITEM ? PlainValue.Type.SEQ : PlainValue.Type.MAP);
|
|
|
|
for (let i = firstItem.props.length - 1; i >= 0; --i) {
|
|
if (firstItem.props[i].start < firstItem.context.lineStart) {
|
|
// props on previous line are assumed by the collection
|
|
this.props = firstItem.props.slice(0, i + 1);
|
|
firstItem.props = firstItem.props.slice(i + 1);
|
|
const itemRange = firstItem.props[0] || firstItem.valueRange;
|
|
firstItem.range.start = itemRange.start;
|
|
break;
|
|
}
|
|
}
|
|
|
|
this.items = [firstItem];
|
|
const ec = grabCollectionEndComments(firstItem);
|
|
if (ec) Array.prototype.push.apply(this.items, ec);
|
|
}
|
|
|
|
get includesTrailingLines() {
|
|
return this.items.length > 0;
|
|
}
|
|
/**
|
|
* @param {ParseContext} context
|
|
* @param {number} start - Index of first character
|
|
* @returns {number} - Index of the character after this
|
|
*/
|
|
|
|
|
|
parse(context, start) {
|
|
this.context = context;
|
|
const {
|
|
parseNode,
|
|
src
|
|
} = context; // It's easier to recalculate lineStart here rather than tracking down the
|
|
// last context from which to read it -- eemeli/yaml#2
|
|
|
|
let lineStart = PlainValue.Node.startOfLine(src, start);
|
|
const firstItem = this.items[0]; // First-item context needs to be correct for later comment handling
|
|
// -- eemeli/yaml#17
|
|
|
|
firstItem.context.parent = this;
|
|
this.valueRange = PlainValue.Range.copy(firstItem.valueRange);
|
|
const indent = firstItem.range.start - firstItem.context.lineStart;
|
|
let offset = start;
|
|
offset = PlainValue.Node.normalizeOffset(src, offset);
|
|
let ch = src[offset];
|
|
let atLineStart = PlainValue.Node.endOfWhiteSpace(src, lineStart) === offset;
|
|
let prevIncludesTrailingLines = false;
|
|
|
|
while (ch) {
|
|
while (ch === '\n' || ch === '#') {
|
|
if (atLineStart && ch === '\n' && !prevIncludesTrailingLines) {
|
|
const blankLine = new BlankLine();
|
|
offset = blankLine.parse({
|
|
src
|
|
}, offset);
|
|
this.valueRange.end = offset;
|
|
|
|
if (offset >= src.length) {
|
|
ch = null;
|
|
break;
|
|
}
|
|
|
|
this.items.push(blankLine);
|
|
offset -= 1; // blankLine.parse() consumes terminal newline
|
|
} else if (ch === '#') {
|
|
if (offset < lineStart + indent && !Collection.nextContentHasIndent(src, offset, indent)) {
|
|
return offset;
|
|
}
|
|
|
|
const comment = new Comment();
|
|
offset = comment.parse({
|
|
indent,
|
|
lineStart,
|
|
src
|
|
}, offset);
|
|
this.items.push(comment);
|
|
this.valueRange.end = offset;
|
|
|
|
if (offset >= src.length) {
|
|
ch = null;
|
|
break;
|
|
}
|
|
}
|
|
|
|
lineStart = offset + 1;
|
|
offset = PlainValue.Node.endOfIndent(src, lineStart);
|
|
|
|
if (PlainValue.Node.atBlank(src, offset)) {
|
|
const wsEnd = PlainValue.Node.endOfWhiteSpace(src, offset);
|
|
const next = src[wsEnd];
|
|
|
|
if (!next || next === '\n' || next === '#') {
|
|
offset = wsEnd;
|
|
}
|
|
}
|
|
|
|
ch = src[offset];
|
|
atLineStart = true;
|
|
}
|
|
|
|
if (!ch) {
|
|
break;
|
|
}
|
|
|
|
if (offset !== lineStart + indent && (atLineStart || ch !== ':')) {
|
|
if (offset < lineStart + indent) {
|
|
if (lineStart > start) offset = lineStart;
|
|
break;
|
|
} else if (!this.error) {
|
|
const msg = 'All collection items must start at the same column';
|
|
this.error = new PlainValue.YAMLSyntaxError(this, msg);
|
|
}
|
|
}
|
|
|
|
if (firstItem.type === PlainValue.Type.SEQ_ITEM) {
|
|
if (ch !== '-') {
|
|
if (lineStart > start) offset = lineStart;
|
|
break;
|
|
}
|
|
} else if (ch === '-' && !this.error) {
|
|
// map key may start with -, as long as it's followed by a non-whitespace char
|
|
const next = src[offset + 1];
|
|
|
|
if (!next || next === '\n' || next === '\t' || next === ' ') {
|
|
const msg = 'A collection cannot be both a mapping and a sequence';
|
|
this.error = new PlainValue.YAMLSyntaxError(this, msg);
|
|
}
|
|
}
|
|
|
|
const node = parseNode({
|
|
atLineStart,
|
|
inCollection: true,
|
|
indent,
|
|
lineStart,
|
|
parent: this
|
|
}, offset);
|
|
if (!node) return offset; // at next document start
|
|
|
|
this.items.push(node);
|
|
this.valueRange.end = node.valueRange.end;
|
|
offset = PlainValue.Node.normalizeOffset(src, node.range.end);
|
|
ch = src[offset];
|
|
atLineStart = false;
|
|
prevIncludesTrailingLines = node.includesTrailingLines; // Need to reset lineStart and atLineStart here if preceding node's range
|
|
// has advanced to check the current line's indentation level
|
|
// -- eemeli/yaml#10 & eemeli/yaml#38
|
|
|
|
if (ch) {
|
|
let ls = offset - 1;
|
|
let prev = src[ls];
|
|
|
|
while (prev === ' ' || prev === '\t') prev = src[--ls];
|
|
|
|
if (prev === '\n') {
|
|
lineStart = ls + 1;
|
|
atLineStart = true;
|
|
}
|
|
}
|
|
|
|
const ec = grabCollectionEndComments(node);
|
|
if (ec) Array.prototype.push.apply(this.items, ec);
|
|
}
|
|
|
|
return offset;
|
|
}
|
|
|
|
setOrigRanges(cr, offset) {
|
|
offset = super.setOrigRanges(cr, offset);
|
|
this.items.forEach(node => {
|
|
offset = node.setOrigRanges(cr, offset);
|
|
});
|
|
return offset;
|
|
}
|
|
|
|
toString() {
|
|
const {
|
|
context: {
|
|
src
|
|
},
|
|
items,
|
|
range,
|
|
value
|
|
} = this;
|
|
if (value != null) return value;
|
|
let str = src.slice(range.start, items[0].range.start) + String(items[0]);
|
|
|
|
for (let i = 1; i < items.length; ++i) {
|
|
const item = items[i];
|
|
const {
|
|
atLineStart,
|
|
indent
|
|
} = item.context;
|
|
if (atLineStart) for (let i = 0; i < indent; ++i) str += ' ';
|
|
str += String(item);
|
|
}
|
|
|
|
return PlainValue.Node.addStringTerminator(src, range.end, str);
|
|
}
|
|
|
|
}
|
|
|
|
class Directive extends PlainValue.Node {
|
|
constructor() {
|
|
super(PlainValue.Type.DIRECTIVE);
|
|
this.name = null;
|
|
}
|
|
|
|
get parameters() {
|
|
const raw = this.rawValue;
|
|
return raw ? raw.trim().split(/[ \t]+/) : [];
|
|
}
|
|
|
|
parseName(start) {
|
|
const {
|
|
src
|
|
} = this.context;
|
|
let offset = start;
|
|
let ch = src[offset];
|
|
|
|
while (ch && ch !== '\n' && ch !== '\t' && ch !== ' ') ch = src[offset += 1];
|
|
|
|
this.name = src.slice(start, offset);
|
|
return offset;
|
|
}
|
|
|
|
parseParameters(start) {
|
|
const {
|
|
src
|
|
} = this.context;
|
|
let offset = start;
|
|
let ch = src[offset];
|
|
|
|
while (ch && ch !== '\n' && ch !== '#') ch = src[offset += 1];
|
|
|
|
this.valueRange = new PlainValue.Range(start, offset);
|
|
return offset;
|
|
}
|
|
|
|
parse(context, start) {
|
|
this.context = context;
|
|
let offset = this.parseName(start + 1);
|
|
offset = this.parseParameters(offset);
|
|
offset = this.parseComment(offset);
|
|
this.range = new PlainValue.Range(start, offset);
|
|
return offset;
|
|
}
|
|
|
|
}
|
|
|
|
class Document extends PlainValue.Node {
|
|
static startCommentOrEndBlankLine(src, start) {
|
|
const offset = PlainValue.Node.endOfWhiteSpace(src, start);
|
|
const ch = src[offset];
|
|
return ch === '#' || ch === '\n' ? offset : start;
|
|
}
|
|
|
|
constructor() {
|
|
super(PlainValue.Type.DOCUMENT);
|
|
this.directives = null;
|
|
this.contents = null;
|
|
this.directivesEndMarker = null;
|
|
this.documentEndMarker = null;
|
|
}
|
|
|
|
parseDirectives(start) {
|
|
const {
|
|
src
|
|
} = this.context;
|
|
this.directives = [];
|
|
let atLineStart = true;
|
|
let hasDirectives = false;
|
|
let offset = start;
|
|
|
|
while (!PlainValue.Node.atDocumentBoundary(src, offset, PlainValue.Char.DIRECTIVES_END)) {
|
|
offset = Document.startCommentOrEndBlankLine(src, offset);
|
|
|
|
switch (src[offset]) {
|
|
case '\n':
|
|
if (atLineStart) {
|
|
const blankLine = new BlankLine();
|
|
offset = blankLine.parse({
|
|
src
|
|
}, offset);
|
|
|
|
if (offset < src.length) {
|
|
this.directives.push(blankLine);
|
|
}
|
|
} else {
|
|
offset += 1;
|
|
atLineStart = true;
|
|
}
|
|
|
|
break;
|
|
|
|
case '#':
|
|
{
|
|
const comment = new Comment();
|
|
offset = comment.parse({
|
|
src
|
|
}, offset);
|
|
this.directives.push(comment);
|
|
atLineStart = false;
|
|
}
|
|
break;
|
|
|
|
case '%':
|
|
{
|
|
const directive = new Directive();
|
|
offset = directive.parse({
|
|
parent: this,
|
|
src
|
|
}, offset);
|
|
this.directives.push(directive);
|
|
hasDirectives = true;
|
|
atLineStart = false;
|
|
}
|
|
break;
|
|
|
|
default:
|
|
if (hasDirectives) {
|
|
this.error = new PlainValue.YAMLSemanticError(this, 'Missing directives-end indicator line');
|
|
} else if (this.directives.length > 0) {
|
|
this.contents = this.directives;
|
|
this.directives = [];
|
|
}
|
|
|
|
return offset;
|
|
}
|
|
}
|
|
|
|
if (src[offset]) {
|
|
this.directivesEndMarker = new PlainValue.Range(offset, offset + 3);
|
|
return offset + 3;
|
|
}
|
|
|
|
if (hasDirectives) {
|
|
this.error = new PlainValue.YAMLSemanticError(this, 'Missing directives-end indicator line');
|
|
} else if (this.directives.length > 0) {
|
|
this.contents = this.directives;
|
|
this.directives = [];
|
|
}
|
|
|
|
return offset;
|
|
}
|
|
|
|
parseContents(start) {
|
|
const {
|
|
parseNode,
|
|
src
|
|
} = this.context;
|
|
if (!this.contents) this.contents = [];
|
|
let lineStart = start;
|
|
|
|
while (src[lineStart - 1] === '-') lineStart -= 1;
|
|
|
|
let offset = PlainValue.Node.endOfWhiteSpace(src, start);
|
|
let atLineStart = lineStart === start;
|
|
this.valueRange = new PlainValue.Range(offset);
|
|
|
|
while (!PlainValue.Node.atDocumentBoundary(src, offset, PlainValue.Char.DOCUMENT_END)) {
|
|
switch (src[offset]) {
|
|
case '\n':
|
|
if (atLineStart) {
|
|
const blankLine = new BlankLine();
|
|
offset = blankLine.parse({
|
|
src
|
|
}, offset);
|
|
|
|
if (offset < src.length) {
|
|
this.contents.push(blankLine);
|
|
}
|
|
} else {
|
|
offset += 1;
|
|
atLineStart = true;
|
|
}
|
|
|
|
lineStart = offset;
|
|
break;
|
|
|
|
case '#':
|
|
{
|
|
const comment = new Comment();
|
|
offset = comment.parse({
|
|
src
|
|
}, offset);
|
|
this.contents.push(comment);
|
|
atLineStart = false;
|
|
}
|
|
break;
|
|
|
|
default:
|
|
{
|
|
const iEnd = PlainValue.Node.endOfIndent(src, offset);
|
|
const context = {
|
|
atLineStart,
|
|
indent: -1,
|
|
inFlow: false,
|
|
inCollection: false,
|
|
lineStart,
|
|
parent: this
|
|
};
|
|
const node = parseNode(context, iEnd);
|
|
if (!node) return this.valueRange.end = iEnd; // at next document start
|
|
|
|
this.contents.push(node);
|
|
offset = node.range.end;
|
|
atLineStart = false;
|
|
const ec = grabCollectionEndComments(node);
|
|
if (ec) Array.prototype.push.apply(this.contents, ec);
|
|
}
|
|
}
|
|
|
|
offset = Document.startCommentOrEndBlankLine(src, offset);
|
|
}
|
|
|
|
this.valueRange.end = offset;
|
|
|
|
if (src[offset]) {
|
|
this.documentEndMarker = new PlainValue.Range(offset, offset + 3);
|
|
offset += 3;
|
|
|
|
if (src[offset]) {
|
|
offset = PlainValue.Node.endOfWhiteSpace(src, offset);
|
|
|
|
if (src[offset] === '#') {
|
|
const comment = new Comment();
|
|
offset = comment.parse({
|
|
src
|
|
}, offset);
|
|
this.contents.push(comment);
|
|
}
|
|
|
|
switch (src[offset]) {
|
|
case '\n':
|
|
offset += 1;
|
|
break;
|
|
|
|
case undefined:
|
|
break;
|
|
|
|
default:
|
|
this.error = new PlainValue.YAMLSyntaxError(this, 'Document end marker line cannot have a non-comment suffix');
|
|
}
|
|
}
|
|
}
|
|
|
|
return offset;
|
|
}
|
|
/**
|
|
* @param {ParseContext} context
|
|
* @param {number} start - Index of first character
|
|
* @returns {number} - Index of the character after this
|
|
*/
|
|
|
|
|
|
parse(context, start) {
|
|
context.root = this;
|
|
this.context = context;
|
|
const {
|
|
src
|
|
} = context;
|
|
let offset = src.charCodeAt(start) === 0xfeff ? start + 1 : start; // skip BOM
|
|
|
|
offset = this.parseDirectives(offset);
|
|
offset = this.parseContents(offset);
|
|
return offset;
|
|
}
|
|
|
|
setOrigRanges(cr, offset) {
|
|
offset = super.setOrigRanges(cr, offset);
|
|
this.directives.forEach(node => {
|
|
offset = node.setOrigRanges(cr, offset);
|
|
});
|
|
if (this.directivesEndMarker) offset = this.directivesEndMarker.setOrigRange(cr, offset);
|
|
this.contents.forEach(node => {
|
|
offset = node.setOrigRanges(cr, offset);
|
|
});
|
|
if (this.documentEndMarker) offset = this.documentEndMarker.setOrigRange(cr, offset);
|
|
return offset;
|
|
}
|
|
|
|
toString() {
|
|
const {
|
|
contents,
|
|
directives,
|
|
value
|
|
} = this;
|
|
if (value != null) return value;
|
|
let str = directives.join('');
|
|
|
|
if (contents.length > 0) {
|
|
if (directives.length > 0 || contents[0].type === PlainValue.Type.COMMENT) str += '---\n';
|
|
str += contents.join('');
|
|
}
|
|
|
|
if (str[str.length - 1] !== '\n') str += '\n';
|
|
return str;
|
|
}
|
|
|
|
}
|
|
|
|
class Alias extends PlainValue.Node {
|
|
/**
|
|
* Parses an *alias from the source
|
|
*
|
|
* @param {ParseContext} context
|
|
* @param {number} start - Index of first character
|
|
* @returns {number} - Index of the character after this scalar
|
|
*/
|
|
parse(context, start) {
|
|
this.context = context;
|
|
const {
|
|
src
|
|
} = context;
|
|
let offset = PlainValue.Node.endOfIdentifier(src, start + 1);
|
|
this.valueRange = new PlainValue.Range(start + 1, offset);
|
|
offset = PlainValue.Node.endOfWhiteSpace(src, offset);
|
|
offset = this.parseComment(offset);
|
|
return offset;
|
|
}
|
|
|
|
}
|
|
|
|
const Chomp = {
|
|
CLIP: 'CLIP',
|
|
KEEP: 'KEEP',
|
|
STRIP: 'STRIP'
|
|
};
|
|
class BlockValue extends PlainValue.Node {
|
|
constructor(type, props) {
|
|
super(type, props);
|
|
this.blockIndent = null;
|
|
this.chomping = Chomp.CLIP;
|
|
this.header = null;
|
|
}
|
|
|
|
get includesTrailingLines() {
|
|
return this.chomping === Chomp.KEEP;
|
|
}
|
|
|
|
get strValue() {
|
|
if (!this.valueRange || !this.context) return null;
|
|
let {
|
|
start,
|
|
end
|
|
} = this.valueRange;
|
|
const {
|
|
indent,
|
|
src
|
|
} = this.context;
|
|
if (this.valueRange.isEmpty()) return '';
|
|
let lastNewLine = null;
|
|
let ch = src[end - 1];
|
|
|
|
while (ch === '\n' || ch === '\t' || ch === ' ') {
|
|
end -= 1;
|
|
|
|
if (end <= start) {
|
|
if (this.chomping === Chomp.KEEP) break;else return ''; // probably never happens
|
|
}
|
|
|
|
if (ch === '\n') lastNewLine = end;
|
|
ch = src[end - 1];
|
|
}
|
|
|
|
let keepStart = end + 1;
|
|
|
|
if (lastNewLine) {
|
|
if (this.chomping === Chomp.KEEP) {
|
|
keepStart = lastNewLine;
|
|
end = this.valueRange.end;
|
|
} else {
|
|
end = lastNewLine;
|
|
}
|
|
}
|
|
|
|
const bi = indent + this.blockIndent;
|
|
const folded = this.type === PlainValue.Type.BLOCK_FOLDED;
|
|
let atStart = true;
|
|
let str = '';
|
|
let sep = '';
|
|
let prevMoreIndented = false;
|
|
|
|
for (let i = start; i < end; ++i) {
|
|
for (let j = 0; j < bi; ++j) {
|
|
if (src[i] !== ' ') break;
|
|
i += 1;
|
|
}
|
|
|
|
const ch = src[i];
|
|
|
|
if (ch === '\n') {
|
|
if (sep === '\n') str += '\n';else sep = '\n';
|
|
} else {
|
|
const lineEnd = PlainValue.Node.endOfLine(src, i);
|
|
const line = src.slice(i, lineEnd);
|
|
i = lineEnd;
|
|
|
|
if (folded && (ch === ' ' || ch === '\t') && i < keepStart) {
|
|
if (sep === ' ') sep = '\n';else if (!prevMoreIndented && !atStart && sep === '\n') sep = '\n\n';
|
|
str += sep + line; //+ ((lineEnd < end && src[lineEnd]) || '')
|
|
|
|
sep = lineEnd < end && src[lineEnd] || '';
|
|
prevMoreIndented = true;
|
|
} else {
|
|
str += sep + line;
|
|
sep = folded && i < keepStart ? ' ' : '\n';
|
|
prevMoreIndented = false;
|
|
}
|
|
|
|
if (atStart && line !== '') atStart = false;
|
|
}
|
|
}
|
|
|
|
return this.chomping === Chomp.STRIP ? str : str + '\n';
|
|
}
|
|
|
|
parseBlockHeader(start) {
|
|
const {
|
|
src
|
|
} = this.context;
|
|
let offset = start + 1;
|
|
let bi = '';
|
|
|
|
while (true) {
|
|
const ch = src[offset];
|
|
|
|
switch (ch) {
|
|
case '-':
|
|
this.chomping = Chomp.STRIP;
|
|
break;
|
|
|
|
case '+':
|
|
this.chomping = Chomp.KEEP;
|
|
break;
|
|
|
|
case '0':
|
|
case '1':
|
|
case '2':
|
|
case '3':
|
|
case '4':
|
|
case '5':
|
|
case '6':
|
|
case '7':
|
|
case '8':
|
|
case '9':
|
|
bi += ch;
|
|
break;
|
|
|
|
default:
|
|
this.blockIndent = Number(bi) || null;
|
|
this.header = new PlainValue.Range(start, offset);
|
|
return offset;
|
|
}
|
|
|
|
offset += 1;
|
|
}
|
|
}
|
|
|
|
parseBlockValue(start) {
|
|
const {
|
|
indent,
|
|
src
|
|
} = this.context;
|
|
const explicit = !!this.blockIndent;
|
|
let offset = start;
|
|
let valueEnd = start;
|
|
let minBlockIndent = 1;
|
|
|
|
for (let ch = src[offset]; ch === '\n'; ch = src[offset]) {
|
|
offset += 1;
|
|
if (PlainValue.Node.atDocumentBoundary(src, offset)) break;
|
|
const end = PlainValue.Node.endOfBlockIndent(src, indent, offset); // should not include tab?
|
|
|
|
if (end === null) break;
|
|
const ch = src[end];
|
|
const lineIndent = end - (offset + indent);
|
|
|
|
if (!this.blockIndent) {
|
|
// no explicit block indent, none yet detected
|
|
if (src[end] !== '\n') {
|
|
// first line with non-whitespace content
|
|
if (lineIndent < minBlockIndent) {
|
|
const msg = 'Block scalars with more-indented leading empty lines must use an explicit indentation indicator';
|
|
this.error = new PlainValue.YAMLSemanticError(this, msg);
|
|
}
|
|
|
|
this.blockIndent = lineIndent;
|
|
} else if (lineIndent > minBlockIndent) {
|
|
// empty line with more whitespace
|
|
minBlockIndent = lineIndent;
|
|
}
|
|
} else if (ch && ch !== '\n' && lineIndent < this.blockIndent) {
|
|
if (src[end] === '#') break;
|
|
|
|
if (!this.error) {
|
|
const src = explicit ? 'explicit indentation indicator' : 'first line';
|
|
const msg = `Block scalars must not be less indented than their ${src}`;
|
|
this.error = new PlainValue.YAMLSemanticError(this, msg);
|
|
}
|
|
}
|
|
|
|
if (src[end] === '\n') {
|
|
offset = end;
|
|
} else {
|
|
offset = valueEnd = PlainValue.Node.endOfLine(src, end);
|
|
}
|
|
}
|
|
|
|
if (this.chomping !== Chomp.KEEP) {
|
|
offset = src[valueEnd] ? valueEnd + 1 : valueEnd;
|
|
}
|
|
|
|
this.valueRange = new PlainValue.Range(start + 1, offset);
|
|
return offset;
|
|
}
|
|
/**
|
|
* Parses a block value from the source
|
|
*
|
|
* Accepted forms are:
|
|
* ```
|
|
* BS
|
|
* block
|
|
* lines
|
|
*
|
|
* BS #comment
|
|
* block
|
|
* lines
|
|
* ```
|
|
* where the block style BS matches the regexp `[|>][-+1-9]*` and block lines
|
|
* are empty or have an indent level greater than `indent`.
|
|
*
|
|
* @param {ParseContext} context
|
|
* @param {number} start - Index of first character
|
|
* @returns {number} - Index of the character after this block
|
|
*/
|
|
|
|
|
|
parse(context, start) {
|
|
this.context = context;
|
|
const {
|
|
src
|
|
} = context;
|
|
let offset = this.parseBlockHeader(start);
|
|
offset = PlainValue.Node.endOfWhiteSpace(src, offset);
|
|
offset = this.parseComment(offset);
|
|
offset = this.parseBlockValue(offset);
|
|
return offset;
|
|
}
|
|
|
|
setOrigRanges(cr, offset) {
|
|
offset = super.setOrigRanges(cr, offset);
|
|
return this.header ? this.header.setOrigRange(cr, offset) : offset;
|
|
}
|
|
|
|
}
|
|
|
|
class FlowCollection extends PlainValue.Node {
|
|
constructor(type, props) {
|
|
super(type, props);
|
|
this.items = null;
|
|
}
|
|
|
|
prevNodeIsJsonLike(idx = this.items.length) {
|
|
const node = this.items[idx - 1];
|
|
return !!node && (node.jsonLike || node.type === PlainValue.Type.COMMENT && this.prevNodeIsJsonLike(idx - 1));
|
|
}
|
|
/**
|
|
* @param {ParseContext} context
|
|
* @param {number} start - Index of first character
|
|
* @returns {number} - Index of the character after this
|
|
*/
|
|
|
|
|
|
parse(context, start) {
|
|
this.context = context;
|
|
const {
|
|
parseNode,
|
|
src
|
|
} = context;
|
|
let {
|
|
indent,
|
|
lineStart
|
|
} = context;
|
|
let char = src[start]; // { or [
|
|
|
|
this.items = [{
|
|
char,
|
|
offset: start
|
|
}];
|
|
let offset = PlainValue.Node.endOfWhiteSpace(src, start + 1);
|
|
char = src[offset];
|
|
|
|
while (char && char !== ']' && char !== '}') {
|
|
switch (char) {
|
|
case '\n':
|
|
{
|
|
lineStart = offset + 1;
|
|
const wsEnd = PlainValue.Node.endOfWhiteSpace(src, lineStart);
|
|
|
|
if (src[wsEnd] === '\n') {
|
|
const blankLine = new BlankLine();
|
|
lineStart = blankLine.parse({
|
|
src
|
|
}, lineStart);
|
|
this.items.push(blankLine);
|
|
}
|
|
|
|
offset = PlainValue.Node.endOfIndent(src, lineStart);
|
|
|
|
if (offset <= lineStart + indent) {
|
|
char = src[offset];
|
|
|
|
if (offset < lineStart + indent || char !== ']' && char !== '}') {
|
|
const msg = 'Insufficient indentation in flow collection';
|
|
this.error = new PlainValue.YAMLSemanticError(this, msg);
|
|
}
|
|
}
|
|
}
|
|
break;
|
|
|
|
case ',':
|
|
{
|
|
this.items.push({
|
|
char,
|
|
offset
|
|
});
|
|
offset += 1;
|
|
}
|
|
break;
|
|
|
|
case '#':
|
|
{
|
|
const comment = new Comment();
|
|
offset = comment.parse({
|
|
src
|
|
}, offset);
|
|
this.items.push(comment);
|
|
}
|
|
break;
|
|
|
|
case '?':
|
|
case ':':
|
|
{
|
|
const next = src[offset + 1];
|
|
|
|
if (next === '\n' || next === '\t' || next === ' ' || next === ',' || // in-flow : after JSON-like key does not need to be followed by whitespace
|
|
char === ':' && this.prevNodeIsJsonLike()) {
|
|
this.items.push({
|
|
char,
|
|
offset
|
|
});
|
|
offset += 1;
|
|
break;
|
|
}
|
|
}
|
|
// fallthrough
|
|
|
|
default:
|
|
{
|
|
const node = parseNode({
|
|
atLineStart: false,
|
|
inCollection: false,
|
|
inFlow: true,
|
|
indent: -1,
|
|
lineStart,
|
|
parent: this
|
|
}, offset);
|
|
|
|
if (!node) {
|
|
// at next document start
|
|
this.valueRange = new PlainValue.Range(start, offset);
|
|
return offset;
|
|
}
|
|
|
|
this.items.push(node);
|
|
offset = PlainValue.Node.normalizeOffset(src, node.range.end);
|
|
}
|
|
}
|
|
|
|
offset = PlainValue.Node.endOfWhiteSpace(src, offset);
|
|
char = src[offset];
|
|
}
|
|
|
|
this.valueRange = new PlainValue.Range(start, offset + 1);
|
|
|
|
if (char) {
|
|
this.items.push({
|
|
char,
|
|
offset
|
|
});
|
|
offset = PlainValue.Node.endOfWhiteSpace(src, offset + 1);
|
|
offset = this.parseComment(offset);
|
|
}
|
|
|
|
return offset;
|
|
}
|
|
|
|
setOrigRanges(cr, offset) {
|
|
offset = super.setOrigRanges(cr, offset);
|
|
this.items.forEach(node => {
|
|
if (node instanceof PlainValue.Node) {
|
|
offset = node.setOrigRanges(cr, offset);
|
|
} else if (cr.length === 0) {
|
|
node.origOffset = node.offset;
|
|
} else {
|
|
let i = offset;
|
|
|
|
while (i < cr.length) {
|
|
if (cr[i] > node.offset) break;else ++i;
|
|
}
|
|
|
|
node.origOffset = node.offset + i;
|
|
offset = i;
|
|
}
|
|
});
|
|
return offset;
|
|
}
|
|
|
|
toString() {
|
|
const {
|
|
context: {
|
|
src
|
|
},
|
|
items,
|
|
range,
|
|
value
|
|
} = this;
|
|
if (value != null) return value;
|
|
const nodes = items.filter(item => item instanceof PlainValue.Node);
|
|
let str = '';
|
|
let prevEnd = range.start;
|
|
nodes.forEach(node => {
|
|
const prefix = src.slice(prevEnd, node.range.start);
|
|
prevEnd = node.range.end;
|
|
str += prefix + String(node);
|
|
|
|
if (str[str.length - 1] === '\n' && src[prevEnd - 1] !== '\n' && src[prevEnd] === '\n') {
|
|
// Comment range does not include the terminal newline, but its
|
|
// stringified value does. Without this fix, newlines at comment ends
|
|
// get duplicated.
|
|
prevEnd += 1;
|
|
}
|
|
});
|
|
str += src.slice(prevEnd, range.end);
|
|
return PlainValue.Node.addStringTerminator(src, range.end, str);
|
|
}
|
|
|
|
}
|
|
|
|
class QuoteDouble extends PlainValue.Node {
|
|
static endOfQuote(src, offset) {
|
|
let ch = src[offset];
|
|
|
|
while (ch && ch !== '"') {
|
|
offset += ch === '\\' ? 2 : 1;
|
|
ch = src[offset];
|
|
}
|
|
|
|
return offset + 1;
|
|
}
|
|
/**
|
|
* @returns {string | { str: string, errors: YAMLSyntaxError[] }}
|
|
*/
|
|
|
|
|
|
get strValue() {
|
|
if (!this.valueRange || !this.context) return null;
|
|
const errors = [];
|
|
const {
|
|
start,
|
|
end
|
|
} = this.valueRange;
|
|
const {
|
|
indent,
|
|
src
|
|
} = this.context;
|
|
if (src[end - 1] !== '"') errors.push(new PlainValue.YAMLSyntaxError(this, 'Missing closing "quote')); // Using String#replace is too painful with escaped newlines preceded by
|
|
// escaped backslashes; also, this should be faster.
|
|
|
|
let str = '';
|
|
|
|
for (let i = start + 1; i < end - 1; ++i) {
|
|
const ch = src[i];
|
|
|
|
if (ch === '\n') {
|
|
if (PlainValue.Node.atDocumentBoundary(src, i + 1)) errors.push(new PlainValue.YAMLSemanticError(this, 'Document boundary indicators are not allowed within string values'));
|
|
const {
|
|
fold,
|
|
offset,
|
|
error
|
|
} = PlainValue.Node.foldNewline(src, i, indent);
|
|
str += fold;
|
|
i = offset;
|
|
if (error) errors.push(new PlainValue.YAMLSemanticError(this, 'Multi-line double-quoted string needs to be sufficiently indented'));
|
|
} else if (ch === '\\') {
|
|
i += 1;
|
|
|
|
switch (src[i]) {
|
|
case '0':
|
|
str += '\0';
|
|
break;
|
|
// null character
|
|
|
|
case 'a':
|
|
str += '\x07';
|
|
break;
|
|
// bell character
|
|
|
|
case 'b':
|
|
str += '\b';
|
|
break;
|
|
// backspace
|
|
|
|
case 'e':
|
|
str += '\x1b';
|
|
break;
|
|
// escape character
|
|
|
|
case 'f':
|
|
str += '\f';
|
|
break;
|
|
// form feed
|
|
|
|
case 'n':
|
|
str += '\n';
|
|
break;
|
|
// line feed
|
|
|
|
case 'r':
|
|
str += '\r';
|
|
break;
|
|
// carriage return
|
|
|
|
case 't':
|
|
str += '\t';
|
|
break;
|
|
// horizontal tab
|
|
|
|
case 'v':
|
|
str += '\v';
|
|
break;
|
|
// vertical tab
|
|
|
|
case 'N':
|
|
str += '\u0085';
|
|
break;
|
|
// Unicode next line
|
|
|
|
case '_':
|
|
str += '\u00a0';
|
|
break;
|
|
// Unicode non-breaking space
|
|
|
|
case 'L':
|
|
str += '\u2028';
|
|
break;
|
|
// Unicode line separator
|
|
|
|
case 'P':
|
|
str += '\u2029';
|
|
break;
|
|
// Unicode paragraph separator
|
|
|
|
case ' ':
|
|
str += ' ';
|
|
break;
|
|
|
|
case '"':
|
|
str += '"';
|
|
break;
|
|
|
|
case '/':
|
|
str += '/';
|
|
break;
|
|
|
|
case '\\':
|
|
str += '\\';
|
|
break;
|
|
|
|
case '\t':
|
|
str += '\t';
|
|
break;
|
|
|
|
case 'x':
|
|
str += this.parseCharCode(i + 1, 2, errors);
|
|
i += 2;
|
|
break;
|
|
|
|
case 'u':
|
|
str += this.parseCharCode(i + 1, 4, errors);
|
|
i += 4;
|
|
break;
|
|
|
|
case 'U':
|
|
str += this.parseCharCode(i + 1, 8, errors);
|
|
i += 8;
|
|
break;
|
|
|
|
case '\n':
|
|
// skip escaped newlines, but still trim the following line
|
|
while (src[i + 1] === ' ' || src[i + 1] === '\t') i += 1;
|
|
|
|
break;
|
|
|
|
default:
|
|
errors.push(new PlainValue.YAMLSyntaxError(this, `Invalid escape sequence ${src.substr(i - 1, 2)}`));
|
|
str += '\\' + src[i];
|
|
}
|
|
} else if (ch === ' ' || ch === '\t') {
|
|
// trim trailing whitespace
|
|
const wsStart = i;
|
|
let next = src[i + 1];
|
|
|
|
while (next === ' ' || next === '\t') {
|
|
i += 1;
|
|
next = src[i + 1];
|
|
}
|
|
|
|
if (next !== '\n') str += i > wsStart ? src.slice(wsStart, i + 1) : ch;
|
|
} else {
|
|
str += ch;
|
|
}
|
|
}
|
|
|
|
return errors.length > 0 ? {
|
|
errors,
|
|
str
|
|
} : str;
|
|
}
|
|
|
|
parseCharCode(offset, length, errors) {
|
|
const {
|
|
src
|
|
} = this.context;
|
|
const cc = src.substr(offset, length);
|
|
const ok = cc.length === length && /^[0-9a-fA-F]+$/.test(cc);
|
|
const code = ok ? parseInt(cc, 16) : NaN;
|
|
|
|
if (isNaN(code)) {
|
|
errors.push(new PlainValue.YAMLSyntaxError(this, `Invalid escape sequence ${src.substr(offset - 2, length + 2)}`));
|
|
return src.substr(offset - 2, length + 2);
|
|
}
|
|
|
|
return String.fromCodePoint(code);
|
|
}
|
|
/**
|
|
* Parses a "double quoted" value from the source
|
|
*
|
|
* @param {ParseContext} context
|
|
* @param {number} start - Index of first character
|
|
* @returns {number} - Index of the character after this scalar
|
|
*/
|
|
|
|
|
|
parse(context, start) {
|
|
this.context = context;
|
|
const {
|
|
src
|
|
} = context;
|
|
let offset = QuoteDouble.endOfQuote(src, start + 1);
|
|
this.valueRange = new PlainValue.Range(start, offset);
|
|
offset = PlainValue.Node.endOfWhiteSpace(src, offset);
|
|
offset = this.parseComment(offset);
|
|
return offset;
|
|
}
|
|
|
|
}
|
|
|
|
class QuoteSingle extends PlainValue.Node {
|
|
static endOfQuote(src, offset) {
|
|
let ch = src[offset];
|
|
|
|
while (ch) {
|
|
if (ch === "'") {
|
|
if (src[offset + 1] !== "'") break;
|
|
ch = src[offset += 2];
|
|
} else {
|
|
ch = src[offset += 1];
|
|
}
|
|
}
|
|
|
|
return offset + 1;
|
|
}
|
|
/**
|
|
* @returns {string | { str: string, errors: YAMLSyntaxError[] }}
|
|
*/
|
|
|
|
|
|
get strValue() {
|
|
if (!this.valueRange || !this.context) return null;
|
|
const errors = [];
|
|
const {
|
|
start,
|
|
end
|
|
} = this.valueRange;
|
|
const {
|
|
indent,
|
|
src
|
|
} = this.context;
|
|
if (src[end - 1] !== "'") errors.push(new PlainValue.YAMLSyntaxError(this, "Missing closing 'quote"));
|
|
let str = '';
|
|
|
|
for (let i = start + 1; i < end - 1; ++i) {
|
|
const ch = src[i];
|
|
|
|
if (ch === '\n') {
|
|
if (PlainValue.Node.atDocumentBoundary(src, i + 1)) errors.push(new PlainValue.YAMLSemanticError(this, 'Document boundary indicators are not allowed within string values'));
|
|
const {
|
|
fold,
|
|
offset,
|
|
error
|
|
} = PlainValue.Node.foldNewline(src, i, indent);
|
|
str += fold;
|
|
i = offset;
|
|
if (error) errors.push(new PlainValue.YAMLSemanticError(this, 'Multi-line single-quoted string needs to be sufficiently indented'));
|
|
} else if (ch === "'") {
|
|
str += ch;
|
|
i += 1;
|
|
if (src[i] !== "'") errors.push(new PlainValue.YAMLSyntaxError(this, 'Unescaped single quote? This should not happen.'));
|
|
} else if (ch === ' ' || ch === '\t') {
|
|
// trim trailing whitespace
|
|
const wsStart = i;
|
|
let next = src[i + 1];
|
|
|
|
while (next === ' ' || next === '\t') {
|
|
i += 1;
|
|
next = src[i + 1];
|
|
}
|
|
|
|
if (next !== '\n') str += i > wsStart ? src.slice(wsStart, i + 1) : ch;
|
|
} else {
|
|
str += ch;
|
|
}
|
|
}
|
|
|
|
return errors.length > 0 ? {
|
|
errors,
|
|
str
|
|
} : str;
|
|
}
|
|
/**
|
|
* Parses a 'single quoted' value from the source
|
|
*
|
|
* @param {ParseContext} context
|
|
* @param {number} start - Index of first character
|
|
* @returns {number} - Index of the character after this scalar
|
|
*/
|
|
|
|
|
|
parse(context, start) {
|
|
this.context = context;
|
|
const {
|
|
src
|
|
} = context;
|
|
let offset = QuoteSingle.endOfQuote(src, start + 1);
|
|
this.valueRange = new PlainValue.Range(start, offset);
|
|
offset = PlainValue.Node.endOfWhiteSpace(src, offset);
|
|
offset = this.parseComment(offset);
|
|
return offset;
|
|
}
|
|
|
|
}
|
|
|
|
function createNewNode(type, props) {
|
|
switch (type) {
|
|
case PlainValue.Type.ALIAS:
|
|
return new Alias(type, props);
|
|
|
|
case PlainValue.Type.BLOCK_FOLDED:
|
|
case PlainValue.Type.BLOCK_LITERAL:
|
|
return new BlockValue(type, props);
|
|
|
|
case PlainValue.Type.FLOW_MAP:
|
|
case PlainValue.Type.FLOW_SEQ:
|
|
return new FlowCollection(type, props);
|
|
|
|
case PlainValue.Type.MAP_KEY:
|
|
case PlainValue.Type.MAP_VALUE:
|
|
case PlainValue.Type.SEQ_ITEM:
|
|
return new CollectionItem(type, props);
|
|
|
|
case PlainValue.Type.COMMENT:
|
|
case PlainValue.Type.PLAIN:
|
|
return new PlainValue.PlainValue(type, props);
|
|
|
|
case PlainValue.Type.QUOTE_DOUBLE:
|
|
return new QuoteDouble(type, props);
|
|
|
|
case PlainValue.Type.QUOTE_SINGLE:
|
|
return new QuoteSingle(type, props);
|
|
|
|
/* istanbul ignore next */
|
|
|
|
default:
|
|
return null;
|
|
// should never happen
|
|
}
|
|
}
|
|
/**
|
|
* @param {boolean} atLineStart - Node starts at beginning of line
|
|
* @param {boolean} inFlow - true if currently in a flow context
|
|
* @param {boolean} inCollection - true if currently in a collection context
|
|
* @param {number} indent - Current level of indentation
|
|
* @param {number} lineStart - Start of the current line
|
|
* @param {Node} parent - The parent of the node
|
|
* @param {string} src - Source of the YAML document
|
|
*/
|
|
|
|
|
|
class ParseContext {
|
|
static parseType(src, offset, inFlow) {
|
|
switch (src[offset]) {
|
|
case '*':
|
|
return PlainValue.Type.ALIAS;
|
|
|
|
case '>':
|
|
return PlainValue.Type.BLOCK_FOLDED;
|
|
|
|
case '|':
|
|
return PlainValue.Type.BLOCK_LITERAL;
|
|
|
|
case '{':
|
|
return PlainValue.Type.FLOW_MAP;
|
|
|
|
case '[':
|
|
return PlainValue.Type.FLOW_SEQ;
|
|
|
|
case '?':
|
|
return !inFlow && PlainValue.Node.atBlank(src, offset + 1, true) ? PlainValue.Type.MAP_KEY : PlainValue.Type.PLAIN;
|
|
|
|
case ':':
|
|
return !inFlow && PlainValue.Node.atBlank(src, offset + 1, true) ? PlainValue.Type.MAP_VALUE : PlainValue.Type.PLAIN;
|
|
|
|
case '-':
|
|
return !inFlow && PlainValue.Node.atBlank(src, offset + 1, true) ? PlainValue.Type.SEQ_ITEM : PlainValue.Type.PLAIN;
|
|
|
|
case '"':
|
|
return PlainValue.Type.QUOTE_DOUBLE;
|
|
|
|
case "'":
|
|
return PlainValue.Type.QUOTE_SINGLE;
|
|
|
|
default:
|
|
return PlainValue.Type.PLAIN;
|
|
}
|
|
}
|
|
|
|
constructor(orig = {}, {
|
|
atLineStart,
|
|
inCollection,
|
|
inFlow,
|
|
indent,
|
|
lineStart,
|
|
parent
|
|
} = {}) {
|
|
PlainValue._defineProperty(this, "parseNode", (overlay, start) => {
|
|
if (PlainValue.Node.atDocumentBoundary(this.src, start)) return null;
|
|
const context = new ParseContext(this, overlay);
|
|
const {
|
|
props,
|
|
type,
|
|
valueStart
|
|
} = context.parseProps(start);
|
|
const node = createNewNode(type, props);
|
|
let offset = node.parse(context, valueStart);
|
|
node.range = new PlainValue.Range(start, offset);
|
|
/* istanbul ignore if */
|
|
|
|
if (offset <= start) {
|
|
// This should never happen, but if it does, let's make sure to at least
|
|
// step one character forward to avoid a busy loop.
|
|
node.error = new Error(`Node#parse consumed no characters`);
|
|
node.error.parseEnd = offset;
|
|
node.error.source = node;
|
|
node.range.end = start + 1;
|
|
}
|
|
|
|
if (context.nodeStartsCollection(node)) {
|
|
if (!node.error && !context.atLineStart && context.parent.type === PlainValue.Type.DOCUMENT) {
|
|
node.error = new PlainValue.YAMLSyntaxError(node, 'Block collection must not have preceding content here (e.g. directives-end indicator)');
|
|
}
|
|
|
|
const collection = new Collection(node);
|
|
offset = collection.parse(new ParseContext(context), offset);
|
|
collection.range = new PlainValue.Range(start, offset);
|
|
return collection;
|
|
}
|
|
|
|
return node;
|
|
});
|
|
|
|
this.atLineStart = atLineStart != null ? atLineStart : orig.atLineStart || false;
|
|
this.inCollection = inCollection != null ? inCollection : orig.inCollection || false;
|
|
this.inFlow = inFlow != null ? inFlow : orig.inFlow || false;
|
|
this.indent = indent != null ? indent : orig.indent;
|
|
this.lineStart = lineStart != null ? lineStart : orig.lineStart;
|
|
this.parent = parent != null ? parent : orig.parent || {};
|
|
this.root = orig.root;
|
|
this.src = orig.src;
|
|
}
|
|
|
|
nodeStartsCollection(node) {
|
|
const {
|
|
inCollection,
|
|
inFlow,
|
|
src
|
|
} = this;
|
|
if (inCollection || inFlow) return false;
|
|
if (node instanceof CollectionItem) return true; // check for implicit key
|
|
|
|
let offset = node.range.end;
|
|
if (src[offset] === '\n' || src[offset - 1] === '\n') return false;
|
|
offset = PlainValue.Node.endOfWhiteSpace(src, offset);
|
|
return src[offset] === ':';
|
|
} // Anchor and tag are before type, which determines the node implementation
|
|
// class; hence this intermediate step.
|
|
|
|
|
|
parseProps(offset) {
|
|
const {
|
|
inFlow,
|
|
parent,
|
|
src
|
|
} = this;
|
|
const props = [];
|
|
let lineHasProps = false;
|
|
offset = this.atLineStart ? PlainValue.Node.endOfIndent(src, offset) : PlainValue.Node.endOfWhiteSpace(src, offset);
|
|
let ch = src[offset];
|
|
|
|
while (ch === PlainValue.Char.ANCHOR || ch === PlainValue.Char.COMMENT || ch === PlainValue.Char.TAG || ch === '\n') {
|
|
if (ch === '\n') {
|
|
let inEnd = offset;
|
|
let lineStart;
|
|
|
|
do {
|
|
lineStart = inEnd + 1;
|
|
inEnd = PlainValue.Node.endOfIndent(src, lineStart);
|
|
} while (src[inEnd] === '\n');
|
|
|
|
const indentDiff = inEnd - (lineStart + this.indent);
|
|
const noIndicatorAsIndent = parent.type === PlainValue.Type.SEQ_ITEM && parent.context.atLineStart;
|
|
if (src[inEnd] !== '#' && !PlainValue.Node.nextNodeIsIndented(src[inEnd], indentDiff, !noIndicatorAsIndent)) break;
|
|
this.atLineStart = true;
|
|
this.lineStart = lineStart;
|
|
lineHasProps = false;
|
|
offset = inEnd;
|
|
} else if (ch === PlainValue.Char.COMMENT) {
|
|
const end = PlainValue.Node.endOfLine(src, offset + 1);
|
|
props.push(new PlainValue.Range(offset, end));
|
|
offset = end;
|
|
} else {
|
|
let end = PlainValue.Node.endOfIdentifier(src, offset + 1);
|
|
|
|
if (ch === PlainValue.Char.TAG && src[end] === ',' && /^[a-zA-Z0-9-]+\.[a-zA-Z0-9-]+,\d\d\d\d(-\d\d){0,2}\/\S/.test(src.slice(offset + 1, end + 13))) {
|
|
// Let's presume we're dealing with a YAML 1.0 domain tag here, rather
|
|
// than an empty but 'foo.bar' private-tagged node in a flow collection
|
|
// followed without whitespace by a plain string starting with a year
|
|
// or date divided by something.
|
|
end = PlainValue.Node.endOfIdentifier(src, end + 5);
|
|
}
|
|
|
|
props.push(new PlainValue.Range(offset, end));
|
|
lineHasProps = true;
|
|
offset = PlainValue.Node.endOfWhiteSpace(src, end);
|
|
}
|
|
|
|
ch = src[offset];
|
|
} // '- &a : b' has an anchor on an empty node
|
|
|
|
|
|
if (lineHasProps && ch === ':' && PlainValue.Node.atBlank(src, offset + 1, true)) offset -= 1;
|
|
const type = ParseContext.parseType(src, offset, inFlow);
|
|
return {
|
|
props,
|
|
type,
|
|
valueStart: offset
|
|
};
|
|
}
|
|
/**
|
|
* Parses a node from the source
|
|
* @param {ParseContext} overlay
|
|
* @param {number} start - Index of first non-whitespace character for the node
|
|
* @returns {?Node} - null if at a document boundary
|
|
*/
|
|
|
|
|
|
}
|
|
|
|
// Published as 'yaml/parse-cst'
|
|
function parse(src) {
|
|
const cr = [];
|
|
|
|
if (src.indexOf('\r') !== -1) {
|
|
src = src.replace(/\r\n?/g, (match, offset) => {
|
|
if (match.length > 1) cr.push(offset);
|
|
return '\n';
|
|
});
|
|
}
|
|
|
|
const documents = [];
|
|
let offset = 0;
|
|
|
|
do {
|
|
const doc = new Document();
|
|
const context = new ParseContext({
|
|
src
|
|
});
|
|
offset = doc.parse(context, offset);
|
|
documents.push(doc);
|
|
} while (offset < src.length);
|
|
|
|
documents.setOrigRanges = () => {
|
|
if (cr.length === 0) return false;
|
|
|
|
for (let i = 1; i < cr.length; ++i) cr[i] -= i;
|
|
|
|
let crOffset = 0;
|
|
|
|
for (let i = 0; i < documents.length; ++i) {
|
|
crOffset = documents[i].setOrigRanges(cr, crOffset);
|
|
}
|
|
|
|
cr.splice(0, cr.length);
|
|
return true;
|
|
};
|
|
|
|
documents.toString = () => documents.join('...\n');
|
|
|
|
return documents;
|
|
}
|
|
|
|
exports.parse = parse;
|
|
|
|
|
|
/***/ }),
|
|
|
|
/***/ 4227:
|
|
/***/ ((__unused_webpack_module, exports, __nccwpck_require__) => {
|
|
|
|
"use strict";
|
|
|
|
|
|
var PlainValue = __nccwpck_require__(5215);
|
|
|
|
function addCommentBefore(str, indent, comment) {
|
|
if (!comment) return str;
|
|
const cc = comment.replace(/[\s\S]^/gm, `$&${indent}#`);
|
|
return `#${cc}\n${indent}${str}`;
|
|
}
|
|
function addComment(str, indent, comment) {
|
|
return !comment ? str : comment.indexOf('\n') === -1 ? `${str} #${comment}` : `${str}\n` + comment.replace(/^/gm, `${indent || ''}#`);
|
|
}
|
|
|
|
class Node {}
|
|
|
|
function toJSON(value, arg, ctx) {
|
|
if (Array.isArray(value)) return value.map((v, i) => toJSON(v, String(i), ctx));
|
|
|
|
if (value && typeof value.toJSON === 'function') {
|
|
const anchor = ctx && ctx.anchors && ctx.anchors.get(value);
|
|
if (anchor) ctx.onCreate = res => {
|
|
anchor.res = res;
|
|
delete ctx.onCreate;
|
|
};
|
|
const res = value.toJSON(arg, ctx);
|
|
if (anchor && ctx.onCreate) ctx.onCreate(res);
|
|
return res;
|
|
}
|
|
|
|
if ((!ctx || !ctx.keep) && typeof value === 'bigint') return Number(value);
|
|
return value;
|
|
}
|
|
|
|
class Scalar extends Node {
|
|
constructor(value) {
|
|
super();
|
|
this.value = value;
|
|
}
|
|
|
|
toJSON(arg, ctx) {
|
|
return ctx && ctx.keep ? this.value : toJSON(this.value, arg, ctx);
|
|
}
|
|
|
|
toString() {
|
|
return String(this.value);
|
|
}
|
|
|
|
}
|
|
|
|
function collectionFromPath(schema, path, value) {
|
|
let v = value;
|
|
|
|
for (let i = path.length - 1; i >= 0; --i) {
|
|
const k = path[i];
|
|
|
|
if (Number.isInteger(k) && k >= 0) {
|
|
const a = [];
|
|
a[k] = v;
|
|
v = a;
|
|
} else {
|
|
const o = {};
|
|
Object.defineProperty(o, k, {
|
|
value: v,
|
|
writable: true,
|
|
enumerable: true,
|
|
configurable: true
|
|
});
|
|
v = o;
|
|
}
|
|
}
|
|
|
|
return schema.createNode(v, false);
|
|
} // null, undefined, or an empty non-string iterable (e.g. [])
|
|
|
|
|
|
const isEmptyPath = path => path == null || typeof path === 'object' && path[Symbol.iterator]().next().done;
|
|
class Collection extends Node {
|
|
constructor(schema) {
|
|
super();
|
|
|
|
PlainValue._defineProperty(this, "items", []);
|
|
|
|
this.schema = schema;
|
|
}
|
|
|
|
addIn(path, value) {
|
|
if (isEmptyPath(path)) this.add(value);else {
|
|
const [key, ...rest] = path;
|
|
const node = this.get(key, true);
|
|
if (node instanceof Collection) node.addIn(rest, value);else if (node === undefined && this.schema) this.set(key, collectionFromPath(this.schema, rest, value));else throw new Error(`Expected YAML collection at ${key}. Remaining path: ${rest}`);
|
|
}
|
|
}
|
|
|
|
deleteIn([key, ...rest]) {
|
|
if (rest.length === 0) return this.delete(key);
|
|
const node = this.get(key, true);
|
|
if (node instanceof Collection) return node.deleteIn(rest);else throw new Error(`Expected YAML collection at ${key}. Remaining path: ${rest}`);
|
|
}
|
|
|
|
getIn([key, ...rest], keepScalar) {
|
|
const node = this.get(key, true);
|
|
if (rest.length === 0) return !keepScalar && node instanceof Scalar ? node.value : node;else return node instanceof Collection ? node.getIn(rest, keepScalar) : undefined;
|
|
}
|
|
|
|
hasAllNullValues() {
|
|
return this.items.every(node => {
|
|
if (!node || node.type !== 'PAIR') return false;
|
|
const n = node.value;
|
|
return n == null || n instanceof Scalar && n.value == null && !n.commentBefore && !n.comment && !n.tag;
|
|
});
|
|
}
|
|
|
|
hasIn([key, ...rest]) {
|
|
if (rest.length === 0) return this.has(key);
|
|
const node = this.get(key, true);
|
|
return node instanceof Collection ? node.hasIn(rest) : false;
|
|
}
|
|
|
|
setIn([key, ...rest], value) {
|
|
if (rest.length === 0) {
|
|
this.set(key, value);
|
|
} else {
|
|
const node = this.get(key, true);
|
|
if (node instanceof Collection) node.setIn(rest, value);else if (node === undefined && this.schema) this.set(key, collectionFromPath(this.schema, rest, value));else throw new Error(`Expected YAML collection at ${key}. Remaining path: ${rest}`);
|
|
}
|
|
} // overridden in implementations
|
|
|
|
/* istanbul ignore next */
|
|
|
|
|
|
toJSON() {
|
|
return null;
|
|
}
|
|
|
|
toString(ctx, {
|
|
blockItem,
|
|
flowChars,
|
|
isMap,
|
|
itemIndent
|
|
}, onComment, onChompKeep) {
|
|
const {
|
|
indent,
|
|
indentStep,
|
|
stringify
|
|
} = ctx;
|
|
const inFlow = this.type === PlainValue.Type.FLOW_MAP || this.type === PlainValue.Type.FLOW_SEQ || ctx.inFlow;
|
|
if (inFlow) itemIndent += indentStep;
|
|
const allNullValues = isMap && this.hasAllNullValues();
|
|
ctx = Object.assign({}, ctx, {
|
|
allNullValues,
|
|
indent: itemIndent,
|
|
inFlow,
|
|
type: null
|
|
});
|
|
let chompKeep = false;
|
|
let hasItemWithNewLine = false;
|
|
const nodes = this.items.reduce((nodes, item, i) => {
|
|
let comment;
|
|
|
|
if (item) {
|
|
if (!chompKeep && item.spaceBefore) nodes.push({
|
|
type: 'comment',
|
|
str: ''
|
|
});
|
|
if (item.commentBefore) item.commentBefore.match(/^.*$/gm).forEach(line => {
|
|
nodes.push({
|
|
type: 'comment',
|
|
str: `#${line}`
|
|
});
|
|
});
|
|
if (item.comment) comment = item.comment;
|
|
if (inFlow && (!chompKeep && item.spaceBefore || item.commentBefore || item.comment || item.key && (item.key.commentBefore || item.key.comment) || item.value && (item.value.commentBefore || item.value.comment))) hasItemWithNewLine = true;
|
|
}
|
|
|
|
chompKeep = false;
|
|
let str = stringify(item, ctx, () => comment = null, () => chompKeep = true);
|
|
if (inFlow && !hasItemWithNewLine && str.includes('\n')) hasItemWithNewLine = true;
|
|
if (inFlow && i < this.items.length - 1) str += ',';
|
|
str = addComment(str, itemIndent, comment);
|
|
if (chompKeep && (comment || inFlow)) chompKeep = false;
|
|
nodes.push({
|
|
type: 'item',
|
|
str
|
|
});
|
|
return nodes;
|
|
}, []);
|
|
let str;
|
|
|
|
if (nodes.length === 0) {
|
|
str = flowChars.start + flowChars.end;
|
|
} else if (inFlow) {
|
|
const {
|
|
start,
|
|
end
|
|
} = flowChars;
|
|
const strings = nodes.map(n => n.str);
|
|
|
|
if (hasItemWithNewLine || strings.reduce((sum, str) => sum + str.length + 2, 2) > Collection.maxFlowStringSingleLineLength) {
|
|
str = start;
|
|
|
|
for (const s of strings) {
|
|
str += s ? `\n${indentStep}${indent}${s}` : '\n';
|
|
}
|
|
|
|
str += `\n${indent}${end}`;
|
|
} else {
|
|
str = `${start} ${strings.join(' ')} ${end}`;
|
|
}
|
|
} else {
|
|
const strings = nodes.map(blockItem);
|
|
str = strings.shift();
|
|
|
|
for (const s of strings) str += s ? `\n${indent}${s}` : '\n';
|
|
}
|
|
|
|
if (this.comment) {
|
|
str += '\n' + this.comment.replace(/^/gm, `${indent}#`);
|
|
if (onComment) onComment();
|
|
} else if (chompKeep && onChompKeep) onChompKeep();
|
|
|
|
return str;
|
|
}
|
|
|
|
}
|
|
|
|
PlainValue._defineProperty(Collection, "maxFlowStringSingleLineLength", 60);
|
|
|
|
function asItemIndex(key) {
|
|
let idx = key instanceof Scalar ? key.value : key;
|
|
if (idx && typeof idx === 'string') idx = Number(idx);
|
|
return Number.isInteger(idx) && idx >= 0 ? idx : null;
|
|
}
|
|
|
|
class YAMLSeq extends Collection {
|
|
add(value) {
|
|
this.items.push(value);
|
|
}
|
|
|
|
delete(key) {
|
|
const idx = asItemIndex(key);
|
|
if (typeof idx !== 'number') return false;
|
|
const del = this.items.splice(idx, 1);
|
|
return del.length > 0;
|
|
}
|
|
|
|
get(key, keepScalar) {
|
|
const idx = asItemIndex(key);
|
|
if (typeof idx !== 'number') return undefined;
|
|
const it = this.items[idx];
|
|
return !keepScalar && it instanceof Scalar ? it.value : it;
|
|
}
|
|
|
|
has(key) {
|
|
const idx = asItemIndex(key);
|
|
return typeof idx === 'number' && idx < this.items.length;
|
|
}
|
|
|
|
set(key, value) {
|
|
const idx = asItemIndex(key);
|
|
if (typeof idx !== 'number') throw new Error(`Expected a valid index, not ${key}.`);
|
|
this.items[idx] = value;
|
|
}
|
|
|
|
toJSON(_, ctx) {
|
|
const seq = [];
|
|
if (ctx && ctx.onCreate) ctx.onCreate(seq);
|
|
let i = 0;
|
|
|
|
for (const item of this.items) seq.push(toJSON(item, String(i++), ctx));
|
|
|
|
return seq;
|
|
}
|
|
|
|
toString(ctx, onComment, onChompKeep) {
|
|
if (!ctx) return JSON.stringify(this);
|
|
return super.toString(ctx, {
|
|
blockItem: n => n.type === 'comment' ? n.str : `- ${n.str}`,
|
|
flowChars: {
|
|
start: '[',
|
|
end: ']'
|
|
},
|
|
isMap: false,
|
|
itemIndent: (ctx.indent || '') + ' '
|
|
}, onComment, onChompKeep);
|
|
}
|
|
|
|
}
|
|
|
|
const stringifyKey = (key, jsKey, ctx) => {
|
|
if (jsKey === null) return '';
|
|
if (typeof jsKey !== 'object') return String(jsKey);
|
|
if (key instanceof Node && ctx && ctx.doc) return key.toString({
|
|
anchors: Object.create(null),
|
|
doc: ctx.doc,
|
|
indent: '',
|
|
indentStep: ctx.indentStep,
|
|
inFlow: true,
|
|
inStringifyKey: true,
|
|
stringify: ctx.stringify
|
|
});
|
|
return JSON.stringify(jsKey);
|
|
};
|
|
|
|
class Pair extends Node {
|
|
constructor(key, value = null) {
|
|
super();
|
|
this.key = key;
|
|
this.value = value;
|
|
this.type = Pair.Type.PAIR;
|
|
}
|
|
|
|
get commentBefore() {
|
|
return this.key instanceof Node ? this.key.commentBefore : undefined;
|
|
}
|
|
|
|
set commentBefore(cb) {
|
|
if (this.key == null) this.key = new Scalar(null);
|
|
if (this.key instanceof Node) this.key.commentBefore = cb;else {
|
|
const msg = 'Pair.commentBefore is an alias for Pair.key.commentBefore. To set it, the key must be a Node.';
|
|
throw new Error(msg);
|
|
}
|
|
}
|
|
|
|
addToJSMap(ctx, map) {
|
|
const key = toJSON(this.key, '', ctx);
|
|
|
|
if (map instanceof Map) {
|
|
const value = toJSON(this.value, key, ctx);
|
|
map.set(key, value);
|
|
} else if (map instanceof Set) {
|
|
map.add(key);
|
|
} else {
|
|
const stringKey = stringifyKey(this.key, key, ctx);
|
|
const value = toJSON(this.value, stringKey, ctx);
|
|
if (stringKey in map) Object.defineProperty(map, stringKey, {
|
|
value,
|
|
writable: true,
|
|
enumerable: true,
|
|
configurable: true
|
|
});else map[stringKey] = value;
|
|
}
|
|
|
|
return map;
|
|
}
|
|
|
|
toJSON(_, ctx) {
|
|
const pair = ctx && ctx.mapAsMap ? new Map() : {};
|
|
return this.addToJSMap(ctx, pair);
|
|
}
|
|
|
|
toString(ctx, onComment, onChompKeep) {
|
|
if (!ctx || !ctx.doc) return JSON.stringify(this);
|
|
const {
|
|
indent: indentSize,
|
|
indentSeq,
|
|
simpleKeys
|
|
} = ctx.doc.options;
|
|
let {
|
|
key,
|
|
value
|
|
} = this;
|
|
let keyComment = key instanceof Node && key.comment;
|
|
|
|
if (simpleKeys) {
|
|
if (keyComment) {
|
|
throw new Error('With simple keys, key nodes cannot have comments');
|
|
}
|
|
|
|
if (key instanceof Collection) {
|
|
const msg = 'With simple keys, collection cannot be used as a key value';
|
|
throw new Error(msg);
|
|
}
|
|
}
|
|
|
|
let explicitKey = !simpleKeys && (!key || keyComment || (key instanceof Node ? key instanceof Collection || key.type === PlainValue.Type.BLOCK_FOLDED || key.type === PlainValue.Type.BLOCK_LITERAL : typeof key === 'object'));
|
|
const {
|
|
doc,
|
|
indent,
|
|
indentStep,
|
|
stringify
|
|
} = ctx;
|
|
ctx = Object.assign({}, ctx, {
|
|
implicitKey: !explicitKey,
|
|
indent: indent + indentStep
|
|
});
|
|
let chompKeep = false;
|
|
let str = stringify(key, ctx, () => keyComment = null, () => chompKeep = true);
|
|
str = addComment(str, ctx.indent, keyComment);
|
|
|
|
if (!explicitKey && str.length > 1024) {
|
|
if (simpleKeys) throw new Error('With simple keys, single line scalar must not span more than 1024 characters');
|
|
explicitKey = true;
|
|
}
|
|
|
|
if (ctx.allNullValues && !simpleKeys) {
|
|
if (this.comment) {
|
|
str = addComment(str, ctx.indent, this.comment);
|
|
if (onComment) onComment();
|
|
} else if (chompKeep && !keyComment && onChompKeep) onChompKeep();
|
|
|
|
return ctx.inFlow && !explicitKey ? str : `? ${str}`;
|
|
}
|
|
|
|
str = explicitKey ? `? ${str}\n${indent}:` : `${str}:`;
|
|
|
|
if (this.comment) {
|
|
// expected (but not strictly required) to be a single-line comment
|
|
str = addComment(str, ctx.indent, this.comment);
|
|
if (onComment) onComment();
|
|
}
|
|
|
|
let vcb = '';
|
|
let valueComment = null;
|
|
|
|
if (value instanceof Node) {
|
|
if (value.spaceBefore) vcb = '\n';
|
|
|
|
if (value.commentBefore) {
|
|
const cs = value.commentBefore.replace(/^/gm, `${ctx.indent}#`);
|
|
vcb += `\n${cs}`;
|
|
}
|
|
|
|
valueComment = value.comment;
|
|
} else if (value && typeof value === 'object') {
|
|
value = doc.schema.createNode(value, true);
|
|
}
|
|
|
|
ctx.implicitKey = false;
|
|
if (!explicitKey && !this.comment && value instanceof Scalar) ctx.indentAtStart = str.length + 1;
|
|
chompKeep = false;
|
|
|
|
if (!indentSeq && indentSize >= 2 && !ctx.inFlow && !explicitKey && value instanceof YAMLSeq && value.type !== PlainValue.Type.FLOW_SEQ && !value.tag && !doc.anchors.getName(value)) {
|
|
// If indentSeq === false, consider '- ' as part of indentation where possible
|
|
ctx.indent = ctx.indent.substr(2);
|
|
}
|
|
|
|
const valueStr = stringify(value, ctx, () => valueComment = null, () => chompKeep = true);
|
|
let ws = ' ';
|
|
|
|
if (vcb || this.comment) {
|
|
ws = `${vcb}\n${ctx.indent}`;
|
|
} else if (!explicitKey && value instanceof Collection) {
|
|
const flow = valueStr[0] === '[' || valueStr[0] === '{';
|
|
if (!flow || valueStr.includes('\n')) ws = `\n${ctx.indent}`;
|
|
} else if (valueStr[0] === '\n') ws = '';
|
|
|
|
if (chompKeep && !valueComment && onChompKeep) onChompKeep();
|
|
return addComment(str + ws + valueStr, ctx.indent, valueComment);
|
|
}
|
|
|
|
}
|
|
|
|
PlainValue._defineProperty(Pair, "Type", {
|
|
PAIR: 'PAIR',
|
|
MERGE_PAIR: 'MERGE_PAIR'
|
|
});
|
|
|
|
const getAliasCount = (node, anchors) => {
|
|
if (node instanceof Alias) {
|
|
const anchor = anchors.get(node.source);
|
|
return anchor.count * anchor.aliasCount;
|
|
} else if (node instanceof Collection) {
|
|
let count = 0;
|
|
|
|
for (const item of node.items) {
|
|
const c = getAliasCount(item, anchors);
|
|
if (c > count) count = c;
|
|
}
|
|
|
|
return count;
|
|
} else if (node instanceof Pair) {
|
|
const kc = getAliasCount(node.key, anchors);
|
|
const vc = getAliasCount(node.value, anchors);
|
|
return Math.max(kc, vc);
|
|
}
|
|
|
|
return 1;
|
|
};
|
|
|
|
class Alias extends Node {
|
|
static stringify({
|
|
range,
|
|
source
|
|
}, {
|
|
anchors,
|
|
doc,
|
|
implicitKey,
|
|
inStringifyKey
|
|
}) {
|
|
let anchor = Object.keys(anchors).find(a => anchors[a] === source);
|
|
if (!anchor && inStringifyKey) anchor = doc.anchors.getName(source) || doc.anchors.newName();
|
|
if (anchor) return `*${anchor}${implicitKey ? ' ' : ''}`;
|
|
const msg = doc.anchors.getName(source) ? 'Alias node must be after source node' : 'Source node not found for alias node';
|
|
throw new Error(`${msg} [${range}]`);
|
|
}
|
|
|
|
constructor(source) {
|
|
super();
|
|
this.source = source;
|
|
this.type = PlainValue.Type.ALIAS;
|
|
}
|
|
|
|
set tag(t) {
|
|
throw new Error('Alias nodes cannot have tags');
|
|
}
|
|
|
|
toJSON(arg, ctx) {
|
|
if (!ctx) return toJSON(this.source, arg, ctx);
|
|
const {
|
|
anchors,
|
|
maxAliasCount
|
|
} = ctx;
|
|
const anchor = anchors.get(this.source);
|
|
/* istanbul ignore if */
|
|
|
|
if (!anchor || anchor.res === undefined) {
|
|
const msg = 'This should not happen: Alias anchor was not resolved?';
|
|
if (this.cstNode) throw new PlainValue.YAMLReferenceError(this.cstNode, msg);else throw new ReferenceError(msg);
|
|
}
|
|
|
|
if (maxAliasCount >= 0) {
|
|
anchor.count += 1;
|
|
if (anchor.aliasCount === 0) anchor.aliasCount = getAliasCount(this.source, anchors);
|
|
|
|
if (anchor.count * anchor.aliasCount > maxAliasCount) {
|
|
const msg = 'Excessive alias count indicates a resource exhaustion attack';
|
|
if (this.cstNode) throw new PlainValue.YAMLReferenceError(this.cstNode, msg);else throw new ReferenceError(msg);
|
|
}
|
|
}
|
|
|
|
return anchor.res;
|
|
} // Only called when stringifying an alias mapping key while constructing
|
|
// Object output.
|
|
|
|
|
|
toString(ctx) {
|
|
return Alias.stringify(this, ctx);
|
|
}
|
|
|
|
}
|
|
|
|
PlainValue._defineProperty(Alias, "default", true);
|
|
|
|
function findPair(items, key) {
|
|
const k = key instanceof Scalar ? key.value : key;
|
|
|
|
for (const it of items) {
|
|
if (it instanceof Pair) {
|
|
if (it.key === key || it.key === k) return it;
|
|
if (it.key && it.key.value === k) return it;
|
|
}
|
|
}
|
|
|
|
return undefined;
|
|
}
|
|
class YAMLMap extends Collection {
|
|
add(pair, overwrite) {
|
|
if (!pair) pair = new Pair(pair);else if (!(pair instanceof Pair)) pair = new Pair(pair.key || pair, pair.value);
|
|
const prev = findPair(this.items, pair.key);
|
|
const sortEntries = this.schema && this.schema.sortMapEntries;
|
|
|
|
if (prev) {
|
|
if (overwrite) prev.value = pair.value;else throw new Error(`Key ${pair.key} already set`);
|
|
} else if (sortEntries) {
|
|
const i = this.items.findIndex(item => sortEntries(pair, item) < 0);
|
|
if (i === -1) this.items.push(pair);else this.items.splice(i, 0, pair);
|
|
} else {
|
|
this.items.push(pair);
|
|
}
|
|
}
|
|
|
|
delete(key) {
|
|
const it = findPair(this.items, key);
|
|
if (!it) return false;
|
|
const del = this.items.splice(this.items.indexOf(it), 1);
|
|
return del.length > 0;
|
|
}
|
|
|
|
get(key, keepScalar) {
|
|
const it = findPair(this.items, key);
|
|
const node = it && it.value;
|
|
return !keepScalar && node instanceof Scalar ? node.value : node;
|
|
}
|
|
|
|
has(key) {
|
|
return !!findPair(this.items, key);
|
|
}
|
|
|
|
set(key, value) {
|
|
this.add(new Pair(key, value), true);
|
|
}
|
|
/**
|
|
* @param {*} arg ignored
|
|
* @param {*} ctx Conversion context, originally set in Document#toJSON()
|
|
* @param {Class} Type If set, forces the returned collection type
|
|
* @returns {*} Instance of Type, Map, or Object
|
|
*/
|
|
|
|
|
|
toJSON(_, ctx, Type) {
|
|
const map = Type ? new Type() : ctx && ctx.mapAsMap ? new Map() : {};
|
|
if (ctx && ctx.onCreate) ctx.onCreate(map);
|
|
|
|
for (const item of this.items) item.addToJSMap(ctx, map);
|
|
|
|
return map;
|
|
}
|
|
|
|
toString(ctx, onComment, onChompKeep) {
|
|
if (!ctx) return JSON.stringify(this);
|
|
|
|
for (const item of this.items) {
|
|
if (!(item instanceof Pair)) throw new Error(`Map items must all be pairs; found ${JSON.stringify(item)} instead`);
|
|
}
|
|
|
|
return super.toString(ctx, {
|
|
blockItem: n => n.str,
|
|
flowChars: {
|
|
start: '{',
|
|
end: '}'
|
|
},
|
|
isMap: true,
|
|
itemIndent: ctx.indent || ''
|
|
}, onComment, onChompKeep);
|
|
}
|
|
|
|
}
|
|
|
|
const MERGE_KEY = '<<';
|
|
class Merge extends Pair {
|
|
constructor(pair) {
|
|
if (pair instanceof Pair) {
|
|
let seq = pair.value;
|
|
|
|
if (!(seq instanceof YAMLSeq)) {
|
|
seq = new YAMLSeq();
|
|
seq.items.push(pair.value);
|
|
seq.range = pair.value.range;
|
|
}
|
|
|
|
super(pair.key, seq);
|
|
this.range = pair.range;
|
|
} else {
|
|
super(new Scalar(MERGE_KEY), new YAMLSeq());
|
|
}
|
|
|
|
this.type = Pair.Type.MERGE_PAIR;
|
|
} // If the value associated with a merge key is a single mapping node, each of
|
|
// its key/value pairs is inserted into the current mapping, unless the key
|
|
// already exists in it. If the value associated with the merge key is a
|
|
// sequence, then this sequence is expected to contain mapping nodes and each
|
|
// of these nodes is merged in turn according to its order in the sequence.
|
|
// Keys in mapping nodes earlier in the sequence override keys specified in
|
|
// later mapping nodes. -- http://yaml.org/type/merge.html
|
|
|
|
|
|
addToJSMap(ctx, map) {
|
|
for (const {
|
|
source
|
|
} of this.value.items) {
|
|
if (!(source instanceof YAMLMap)) throw new Error('Merge sources must be maps');
|
|
const srcMap = source.toJSON(null, ctx, Map);
|
|
|
|
for (const [key, value] of srcMap) {
|
|
if (map instanceof Map) {
|
|
if (!map.has(key)) map.set(key, value);
|
|
} else if (map instanceof Set) {
|
|
map.add(key);
|
|
} else if (!Object.prototype.hasOwnProperty.call(map, key)) {
|
|
Object.defineProperty(map, key, {
|
|
value,
|
|
writable: true,
|
|
enumerable: true,
|
|
configurable: true
|
|
});
|
|
}
|
|
}
|
|
}
|
|
|
|
return map;
|
|
}
|
|
|
|
toString(ctx, onComment) {
|
|
const seq = this.value;
|
|
if (seq.items.length > 1) return super.toString(ctx, onComment);
|
|
this.value = seq.items[0];
|
|
const str = super.toString(ctx, onComment);
|
|
this.value = seq;
|
|
return str;
|
|
}
|
|
|
|
}
|
|
|
|
const binaryOptions = {
|
|
defaultType: PlainValue.Type.BLOCK_LITERAL,
|
|
lineWidth: 76
|
|
};
|
|
const boolOptions = {
|
|
trueStr: 'true',
|
|
falseStr: 'false'
|
|
};
|
|
const intOptions = {
|
|
asBigInt: false
|
|
};
|
|
const nullOptions = {
|
|
nullStr: 'null'
|
|
};
|
|
const strOptions = {
|
|
defaultType: PlainValue.Type.PLAIN,
|
|
doubleQuoted: {
|
|
jsonEncoding: false,
|
|
minMultiLineLength: 40
|
|
},
|
|
fold: {
|
|
lineWidth: 80,
|
|
minContentWidth: 20
|
|
}
|
|
};
|
|
|
|
function resolveScalar(str, tags, scalarFallback) {
|
|
for (const {
|
|
format,
|
|
test,
|
|
resolve
|
|
} of tags) {
|
|
if (test) {
|
|
const match = str.match(test);
|
|
|
|
if (match) {
|
|
let res = resolve.apply(null, match);
|
|
if (!(res instanceof Scalar)) res = new Scalar(res);
|
|
if (format) res.format = format;
|
|
return res;
|
|
}
|
|
}
|
|
}
|
|
|
|
if (scalarFallback) str = scalarFallback(str);
|
|
return new Scalar(str);
|
|
}
|
|
|
|
const FOLD_FLOW = 'flow';
|
|
const FOLD_BLOCK = 'block';
|
|
const FOLD_QUOTED = 'quoted'; // presumes i+1 is at the start of a line
|
|
// returns index of last newline in more-indented block
|
|
|
|
const consumeMoreIndentedLines = (text, i) => {
|
|
let ch = text[i + 1];
|
|
|
|
while (ch === ' ' || ch === '\t') {
|
|
do {
|
|
ch = text[i += 1];
|
|
} while (ch && ch !== '\n');
|
|
|
|
ch = text[i + 1];
|
|
}
|
|
|
|
return i;
|
|
};
|
|
/**
|
|
* Tries to keep input at up to `lineWidth` characters, splitting only on spaces
|
|
* not followed by newlines or spaces unless `mode` is `'quoted'`. Lines are
|
|
* terminated with `\n` and started with `indent`.
|
|
*
|
|
* @param {string} text
|
|
* @param {string} indent
|
|
* @param {string} [mode='flow'] `'block'` prevents more-indented lines
|
|
* from being folded; `'quoted'` allows for `\` escapes, including escaped
|
|
* newlines
|
|
* @param {Object} options
|
|
* @param {number} [options.indentAtStart] Accounts for leading contents on
|
|
* the first line, defaulting to `indent.length`
|
|
* @param {number} [options.lineWidth=80]
|
|
* @param {number} [options.minContentWidth=20] Allow highly indented lines to
|
|
* stretch the line width or indent content from the start
|
|
* @param {function} options.onFold Called once if the text is folded
|
|
* @param {function} options.onFold Called once if any line of text exceeds
|
|
* lineWidth characters
|
|
*/
|
|
|
|
|
|
function foldFlowLines(text, indent, mode, {
|
|
indentAtStart,
|
|
lineWidth = 80,
|
|
minContentWidth = 20,
|
|
onFold,
|
|
onOverflow
|
|
}) {
|
|
if (!lineWidth || lineWidth < 0) return text;
|
|
const endStep = Math.max(1 + minContentWidth, 1 + lineWidth - indent.length);
|
|
if (text.length <= endStep) return text;
|
|
const folds = [];
|
|
const escapedFolds = {};
|
|
let end = lineWidth - indent.length;
|
|
|
|
if (typeof indentAtStart === 'number') {
|
|
if (indentAtStart > lineWidth - Math.max(2, minContentWidth)) folds.push(0);else end = lineWidth - indentAtStart;
|
|
}
|
|
|
|
let split = undefined;
|
|
let prev = undefined;
|
|
let overflow = false;
|
|
let i = -1;
|
|
let escStart = -1;
|
|
let escEnd = -1;
|
|
|
|
if (mode === FOLD_BLOCK) {
|
|
i = consumeMoreIndentedLines(text, i);
|
|
if (i !== -1) end = i + endStep;
|
|
}
|
|
|
|
for (let ch; ch = text[i += 1];) {
|
|
if (mode === FOLD_QUOTED && ch === '\\') {
|
|
escStart = i;
|
|
|
|
switch (text[i + 1]) {
|
|
case 'x':
|
|
i += 3;
|
|
break;
|
|
|
|
case 'u':
|
|
i += 5;
|
|
break;
|
|
|
|
case 'U':
|
|
i += 9;
|
|
break;
|
|
|
|
default:
|
|
i += 1;
|
|
}
|
|
|
|
escEnd = i;
|
|
}
|
|
|
|
if (ch === '\n') {
|
|
if (mode === FOLD_BLOCK) i = consumeMoreIndentedLines(text, i);
|
|
end = i + endStep;
|
|
split = undefined;
|
|
} else {
|
|
if (ch === ' ' && prev && prev !== ' ' && prev !== '\n' && prev !== '\t') {
|
|
// space surrounded by non-space can be replaced with newline + indent
|
|
const next = text[i + 1];
|
|
if (next && next !== ' ' && next !== '\n' && next !== '\t') split = i;
|
|
}
|
|
|
|
if (i >= end) {
|
|
if (split) {
|
|
folds.push(split);
|
|
end = split + endStep;
|
|
split = undefined;
|
|
} else if (mode === FOLD_QUOTED) {
|
|
// white-space collected at end may stretch past lineWidth
|
|
while (prev === ' ' || prev === '\t') {
|
|
prev = ch;
|
|
ch = text[i += 1];
|
|
overflow = true;
|
|
} // Account for newline escape, but don't break preceding escape
|
|
|
|
|
|
const j = i > escEnd + 1 ? i - 2 : escStart - 1; // Bail out if lineWidth & minContentWidth are shorter than an escape string
|
|
|
|
if (escapedFolds[j]) return text;
|
|
folds.push(j);
|
|
escapedFolds[j] = true;
|
|
end = j + endStep;
|
|
split = undefined;
|
|
} else {
|
|
overflow = true;
|
|
}
|
|
}
|
|
}
|
|
|
|
prev = ch;
|
|
}
|
|
|
|
if (overflow && onOverflow) onOverflow();
|
|
if (folds.length === 0) return text;
|
|
if (onFold) onFold();
|
|
let res = text.slice(0, folds[0]);
|
|
|
|
for (let i = 0; i < folds.length; ++i) {
|
|
const fold = folds[i];
|
|
const end = folds[i + 1] || text.length;
|
|
if (fold === 0) res = `\n${indent}${text.slice(0, end)}`;else {
|
|
if (mode === FOLD_QUOTED && escapedFolds[fold]) res += `${text[fold]}\\`;
|
|
res += `\n${indent}${text.slice(fold + 1, end)}`;
|
|
}
|
|
}
|
|
|
|
return res;
|
|
}
|
|
|
|
const getFoldOptions = ({
|
|
indentAtStart
|
|
}) => indentAtStart ? Object.assign({
|
|
indentAtStart
|
|
}, strOptions.fold) : strOptions.fold; // Also checks for lines starting with %, as parsing the output as YAML 1.1 will
|
|
// presume that's starting a new document.
|
|
|
|
|
|
const containsDocumentMarker = str => /^(%|---|\.\.\.)/m.test(str);
|
|
|
|
function lineLengthOverLimit(str, lineWidth, indentLength) {
|
|
if (!lineWidth || lineWidth < 0) return false;
|
|
const limit = lineWidth - indentLength;
|
|
const strLen = str.length;
|
|
if (strLen <= limit) return false;
|
|
|
|
for (let i = 0, start = 0; i < strLen; ++i) {
|
|
if (str[i] === '\n') {
|
|
if (i - start > limit) return true;
|
|
start = i + 1;
|
|
if (strLen - start <= limit) return false;
|
|
}
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
function doubleQuotedString(value, ctx) {
|
|
const {
|
|
implicitKey
|
|
} = ctx;
|
|
const {
|
|
jsonEncoding,
|
|
minMultiLineLength
|
|
} = strOptions.doubleQuoted;
|
|
const json = JSON.stringify(value);
|
|
if (jsonEncoding) return json;
|
|
const indent = ctx.indent || (containsDocumentMarker(value) ? ' ' : '');
|
|
let str = '';
|
|
let start = 0;
|
|
|
|
for (let i = 0, ch = json[i]; ch; ch = json[++i]) {
|
|
if (ch === ' ' && json[i + 1] === '\\' && json[i + 2] === 'n') {
|
|
// space before newline needs to be escaped to not be folded
|
|
str += json.slice(start, i) + '\\ ';
|
|
i += 1;
|
|
start = i;
|
|
ch = '\\';
|
|
}
|
|
|
|
if (ch === '\\') switch (json[i + 1]) {
|
|
case 'u':
|
|
{
|
|
str += json.slice(start, i);
|
|
const code = json.substr(i + 2, 4);
|
|
|
|
switch (code) {
|
|
case '0000':
|
|
str += '\\0';
|
|
break;
|
|
|
|
case '0007':
|
|
str += '\\a';
|
|
break;
|
|
|
|
case '000b':
|
|
str += '\\v';
|
|
break;
|
|
|
|
case '001b':
|
|
str += '\\e';
|
|
break;
|
|
|
|
case '0085':
|
|
str += '\\N';
|
|
break;
|
|
|
|
case '00a0':
|
|
str += '\\_';
|
|
break;
|
|
|
|
case '2028':
|
|
str += '\\L';
|
|
break;
|
|
|
|
case '2029':
|
|
str += '\\P';
|
|
break;
|
|
|
|
default:
|
|
if (code.substr(0, 2) === '00') str += '\\x' + code.substr(2);else str += json.substr(i, 6);
|
|
}
|
|
|
|
i += 5;
|
|
start = i + 1;
|
|
}
|
|
break;
|
|
|
|
case 'n':
|
|
if (implicitKey || json[i + 2] === '"' || json.length < minMultiLineLength) {
|
|
i += 1;
|
|
} else {
|
|
// folding will eat first newline
|
|
str += json.slice(start, i) + '\n\n';
|
|
|
|
while (json[i + 2] === '\\' && json[i + 3] === 'n' && json[i + 4] !== '"') {
|
|
str += '\n';
|
|
i += 2;
|
|
}
|
|
|
|
str += indent; // space after newline needs to be escaped to not be folded
|
|
|
|
if (json[i + 2] === ' ') str += '\\';
|
|
i += 1;
|
|
start = i + 1;
|
|
}
|
|
|
|
break;
|
|
|
|
default:
|
|
i += 1;
|
|
}
|
|
}
|
|
|
|
str = start ? str + json.slice(start) : json;
|
|
return implicitKey ? str : foldFlowLines(str, indent, FOLD_QUOTED, getFoldOptions(ctx));
|
|
}
|
|
|
|
function singleQuotedString(value, ctx) {
|
|
if (ctx.implicitKey) {
|
|
if (/\n/.test(value)) return doubleQuotedString(value, ctx);
|
|
} else {
|
|
// single quoted string can't have leading or trailing whitespace around newline
|
|
if (/[ \t]\n|\n[ \t]/.test(value)) return doubleQuotedString(value, ctx);
|
|
}
|
|
|
|
const indent = ctx.indent || (containsDocumentMarker(value) ? ' ' : '');
|
|
const res = "'" + value.replace(/'/g, "''").replace(/\n+/g, `$&\n${indent}`) + "'";
|
|
return ctx.implicitKey ? res : foldFlowLines(res, indent, FOLD_FLOW, getFoldOptions(ctx));
|
|
}
|
|
|
|
function blockString({
|
|
comment,
|
|
type,
|
|
value
|
|
}, ctx, onComment, onChompKeep) {
|
|
// 1. Block can't end in whitespace unless the last line is non-empty.
|
|
// 2. Strings consisting of only whitespace are best rendered explicitly.
|
|
if (/\n[\t ]+$/.test(value) || /^\s*$/.test(value)) {
|
|
return doubleQuotedString(value, ctx);
|
|
}
|
|
|
|
const indent = ctx.indent || (ctx.forceBlockIndent || containsDocumentMarker(value) ? ' ' : '');
|
|
const indentSize = indent ? '2' : '1'; // root is at -1
|
|
|
|
const literal = type === PlainValue.Type.BLOCK_FOLDED ? false : type === PlainValue.Type.BLOCK_LITERAL ? true : !lineLengthOverLimit(value, strOptions.fold.lineWidth, indent.length);
|
|
let header = literal ? '|' : '>';
|
|
if (!value) return header + '\n';
|
|
let wsStart = '';
|
|
let wsEnd = '';
|
|
value = value.replace(/[\n\t ]*$/, ws => {
|
|
const n = ws.indexOf('\n');
|
|
|
|
if (n === -1) {
|
|
header += '-'; // strip
|
|
} else if (value === ws || n !== ws.length - 1) {
|
|
header += '+'; // keep
|
|
|
|
if (onChompKeep) onChompKeep();
|
|
}
|
|
|
|
wsEnd = ws.replace(/\n$/, '');
|
|
return '';
|
|
}).replace(/^[\n ]*/, ws => {
|
|
if (ws.indexOf(' ') !== -1) header += indentSize;
|
|
const m = ws.match(/ +$/);
|
|
|
|
if (m) {
|
|
wsStart = ws.slice(0, -m[0].length);
|
|
return m[0];
|
|
} else {
|
|
wsStart = ws;
|
|
return '';
|
|
}
|
|
});
|
|
if (wsEnd) wsEnd = wsEnd.replace(/\n+(?!\n|$)/g, `$&${indent}`);
|
|
if (wsStart) wsStart = wsStart.replace(/\n+/g, `$&${indent}`);
|
|
|
|
if (comment) {
|
|
header += ' #' + comment.replace(/ ?[\r\n]+/g, ' ');
|
|
if (onComment) onComment();
|
|
}
|
|
|
|
if (!value) return `${header}${indentSize}\n${indent}${wsEnd}`;
|
|
|
|
if (literal) {
|
|
value = value.replace(/\n+/g, `$&${indent}`);
|
|
return `${header}\n${indent}${wsStart}${value}${wsEnd}`;
|
|
}
|
|
|
|
value = value.replace(/\n+/g, '\n$&').replace(/(?:^|\n)([\t ].*)(?:([\n\t ]*)\n(?![\n\t ]))?/g, '$1$2') // more-indented lines aren't folded
|
|
// ^ ind.line ^ empty ^ capture next empty lines only at end of indent
|
|
.replace(/\n+/g, `$&${indent}`);
|
|
const body = foldFlowLines(`${wsStart}${value}${wsEnd}`, indent, FOLD_BLOCK, strOptions.fold);
|
|
return `${header}\n${indent}${body}`;
|
|
}
|
|
|
|
function plainString(item, ctx, onComment, onChompKeep) {
|
|
const {
|
|
comment,
|
|
type,
|
|
value
|
|
} = item;
|
|
const {
|
|
actualString,
|
|
implicitKey,
|
|
indent,
|
|
inFlow
|
|
} = ctx;
|
|
|
|
if (implicitKey && /[\n[\]{},]/.test(value) || inFlow && /[[\]{},]/.test(value)) {
|
|
return doubleQuotedString(value, ctx);
|
|
}
|
|
|
|
if (!value || /^[\n\t ,[\]{}#&*!|>'"%@`]|^[?-]$|^[?-][ \t]|[\n:][ \t]|[ \t]\n|[\n\t ]#|[\n\t :]$/.test(value)) {
|
|
// not allowed:
|
|
// - empty string, '-' or '?'
|
|
// - start with an indicator character (except [?:-]) or /[?-] /
|
|
// - '\n ', ': ' or ' \n' anywhere
|
|
// - '#' not preceded by a non-space char
|
|
// - end with ' ' or ':'
|
|
return implicitKey || inFlow || value.indexOf('\n') === -1 ? value.indexOf('"') !== -1 && value.indexOf("'") === -1 ? singleQuotedString(value, ctx) : doubleQuotedString(value, ctx) : blockString(item, ctx, onComment, onChompKeep);
|
|
}
|
|
|
|
if (!implicitKey && !inFlow && type !== PlainValue.Type.PLAIN && value.indexOf('\n') !== -1) {
|
|
// Where allowed & type not set explicitly, prefer block style for multiline strings
|
|
return blockString(item, ctx, onComment, onChompKeep);
|
|
}
|
|
|
|
if (indent === '' && containsDocumentMarker(value)) {
|
|
ctx.forceBlockIndent = true;
|
|
return blockString(item, ctx, onComment, onChompKeep);
|
|
}
|
|
|
|
const str = value.replace(/\n+/g, `$&\n${indent}`); // Verify that output will be parsed as a string, as e.g. plain numbers and
|
|
// booleans get parsed with those types in v1.2 (e.g. '42', 'true' & '0.9e-3'),
|
|
// and others in v1.1.
|
|
|
|
if (actualString) {
|
|
const {
|
|
tags
|
|
} = ctx.doc.schema;
|
|
const resolved = resolveScalar(str, tags, tags.scalarFallback).value;
|
|
if (typeof resolved !== 'string') return doubleQuotedString(value, ctx);
|
|
}
|
|
|
|
const body = implicitKey ? str : foldFlowLines(str, indent, FOLD_FLOW, getFoldOptions(ctx));
|
|
|
|
if (comment && !inFlow && (body.indexOf('\n') !== -1 || comment.indexOf('\n') !== -1)) {
|
|
if (onComment) onComment();
|
|
return addCommentBefore(body, indent, comment);
|
|
}
|
|
|
|
return body;
|
|
}
|
|
|
|
function stringifyString(item, ctx, onComment, onChompKeep) {
|
|
const {
|
|
defaultType
|
|
} = strOptions;
|
|
const {
|
|
implicitKey,
|
|
inFlow
|
|
} = ctx;
|
|
let {
|
|
type,
|
|
value
|
|
} = item;
|
|
|
|
if (typeof value !== 'string') {
|
|
value = String(value);
|
|
item = Object.assign({}, item, {
|
|
value
|
|
});
|
|
}
|
|
|
|
const _stringify = _type => {
|
|
switch (_type) {
|
|
case PlainValue.Type.BLOCK_FOLDED:
|
|
case PlainValue.Type.BLOCK_LITERAL:
|
|
return blockString(item, ctx, onComment, onChompKeep);
|
|
|
|
case PlainValue.Type.QUOTE_DOUBLE:
|
|
return doubleQuotedString(value, ctx);
|
|
|
|
case PlainValue.Type.QUOTE_SINGLE:
|
|
return singleQuotedString(value, ctx);
|
|
|
|
case PlainValue.Type.PLAIN:
|
|
return plainString(item, ctx, onComment, onChompKeep);
|
|
|
|
default:
|
|
return null;
|
|
}
|
|
};
|
|
|
|
if (type !== PlainValue.Type.QUOTE_DOUBLE && /[\x00-\x08\x0b-\x1f\x7f-\x9f]/.test(value)) {
|
|
// force double quotes on control characters
|
|
type = PlainValue.Type.QUOTE_DOUBLE;
|
|
} else if ((implicitKey || inFlow) && (type === PlainValue.Type.BLOCK_FOLDED || type === PlainValue.Type.BLOCK_LITERAL)) {
|
|
// should not happen; blocks are not valid inside flow containers
|
|
type = PlainValue.Type.QUOTE_DOUBLE;
|
|
}
|
|
|
|
let res = _stringify(type);
|
|
|
|
if (res === null) {
|
|
res = _stringify(defaultType);
|
|
if (res === null) throw new Error(`Unsupported default string type ${defaultType}`);
|
|
}
|
|
|
|
return res;
|
|
}
|
|
|
|
function stringifyNumber({
|
|
format,
|
|
minFractionDigits,
|
|
tag,
|
|
value
|
|
}) {
|
|
if (typeof value === 'bigint') return String(value);
|
|
if (!isFinite(value)) return isNaN(value) ? '.nan' : value < 0 ? '-.inf' : '.inf';
|
|
let n = JSON.stringify(value);
|
|
|
|
if (!format && minFractionDigits && (!tag || tag === 'tag:yaml.org,2002:float') && /^\d/.test(n)) {
|
|
let i = n.indexOf('.');
|
|
|
|
if (i < 0) {
|
|
i = n.length;
|
|
n += '.';
|
|
}
|
|
|
|
let d = minFractionDigits - (n.length - i - 1);
|
|
|
|
while (d-- > 0) n += '0';
|
|
}
|
|
|
|
return n;
|
|
}
|
|
|
|
function checkFlowCollectionEnd(errors, cst) {
|
|
let char, name;
|
|
|
|
switch (cst.type) {
|
|
case PlainValue.Type.FLOW_MAP:
|
|
char = '}';
|
|
name = 'flow map';
|
|
break;
|
|
|
|
case PlainValue.Type.FLOW_SEQ:
|
|
char = ']';
|
|
name = 'flow sequence';
|
|
break;
|
|
|
|
default:
|
|
errors.push(new PlainValue.YAMLSemanticError(cst, 'Not a flow collection!?'));
|
|
return;
|
|
}
|
|
|
|
let lastItem;
|
|
|
|
for (let i = cst.items.length - 1; i >= 0; --i) {
|
|
const item = cst.items[i];
|
|
|
|
if (!item || item.type !== PlainValue.Type.COMMENT) {
|
|
lastItem = item;
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (lastItem && lastItem.char !== char) {
|
|
const msg = `Expected ${name} to end with ${char}`;
|
|
let err;
|
|
|
|
if (typeof lastItem.offset === 'number') {
|
|
err = new PlainValue.YAMLSemanticError(cst, msg);
|
|
err.offset = lastItem.offset + 1;
|
|
} else {
|
|
err = new PlainValue.YAMLSemanticError(lastItem, msg);
|
|
if (lastItem.range && lastItem.range.end) err.offset = lastItem.range.end - lastItem.range.start;
|
|
}
|
|
|
|
errors.push(err);
|
|
}
|
|
}
|
|
function checkFlowCommentSpace(errors, comment) {
|
|
const prev = comment.context.src[comment.range.start - 1];
|
|
|
|
if (prev !== '\n' && prev !== '\t' && prev !== ' ') {
|
|
const msg = 'Comments must be separated from other tokens by white space characters';
|
|
errors.push(new PlainValue.YAMLSemanticError(comment, msg));
|
|
}
|
|
}
|
|
function getLongKeyError(source, key) {
|
|
const sk = String(key);
|
|
const k = sk.substr(0, 8) + '...' + sk.substr(-8);
|
|
return new PlainValue.YAMLSemanticError(source, `The "${k}" key is too long`);
|
|
}
|
|
function resolveComments(collection, comments) {
|
|
for (const {
|
|
afterKey,
|
|
before,
|
|
comment
|
|
} of comments) {
|
|
let item = collection.items[before];
|
|
|
|
if (!item) {
|
|
if (comment !== undefined) {
|
|
if (collection.comment) collection.comment += '\n' + comment;else collection.comment = comment;
|
|
}
|
|
} else {
|
|
if (afterKey && item.value) item = item.value;
|
|
|
|
if (comment === undefined) {
|
|
if (afterKey || !item.commentBefore) item.spaceBefore = true;
|
|
} else {
|
|
if (item.commentBefore) item.commentBefore += '\n' + comment;else item.commentBefore = comment;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// on error, will return { str: string, errors: Error[] }
|
|
function resolveString(doc, node) {
|
|
const res = node.strValue;
|
|
if (!res) return '';
|
|
if (typeof res === 'string') return res;
|
|
res.errors.forEach(error => {
|
|
if (!error.source) error.source = node;
|
|
doc.errors.push(error);
|
|
});
|
|
return res.str;
|
|
}
|
|
|
|
function resolveTagHandle(doc, node) {
|
|
const {
|
|
handle,
|
|
suffix
|
|
} = node.tag;
|
|
let prefix = doc.tagPrefixes.find(p => p.handle === handle);
|
|
|
|
if (!prefix) {
|
|
const dtp = doc.getDefaults().tagPrefixes;
|
|
if (dtp) prefix = dtp.find(p => p.handle === handle);
|
|
if (!prefix) throw new PlainValue.YAMLSemanticError(node, `The ${handle} tag handle is non-default and was not declared.`);
|
|
}
|
|
|
|
if (!suffix) throw new PlainValue.YAMLSemanticError(node, `The ${handle} tag has no suffix.`);
|
|
|
|
if (handle === '!' && (doc.version || doc.options.version) === '1.0') {
|
|
if (suffix[0] === '^') {
|
|
doc.warnings.push(new PlainValue.YAMLWarning(node, 'YAML 1.0 ^ tag expansion is not supported'));
|
|
return suffix;
|
|
}
|
|
|
|
if (/[:/]/.test(suffix)) {
|
|
// word/foo -> tag:word.yaml.org,2002:foo
|
|
const vocab = suffix.match(/^([a-z0-9-]+)\/(.*)/i);
|
|
return vocab ? `tag:${vocab[1]}.yaml.org,2002:${vocab[2]}` : `tag:${suffix}`;
|
|
}
|
|
}
|
|
|
|
return prefix.prefix + decodeURIComponent(suffix);
|
|
}
|
|
|
|
function resolveTagName(doc, node) {
|
|
const {
|
|
tag,
|
|
type
|
|
} = node;
|
|
let nonSpecific = false;
|
|
|
|
if (tag) {
|
|
const {
|
|
handle,
|
|
suffix,
|
|
verbatim
|
|
} = tag;
|
|
|
|
if (verbatim) {
|
|
if (verbatim !== '!' && verbatim !== '!!') return verbatim;
|
|
const msg = `Verbatim tags aren't resolved, so ${verbatim} is invalid.`;
|
|
doc.errors.push(new PlainValue.YAMLSemanticError(node, msg));
|
|
} else if (handle === '!' && !suffix) {
|
|
nonSpecific = true;
|
|
} else {
|
|
try {
|
|
return resolveTagHandle(doc, node);
|
|
} catch (error) {
|
|
doc.errors.push(error);
|
|
}
|
|
}
|
|
}
|
|
|
|
switch (type) {
|
|
case PlainValue.Type.BLOCK_FOLDED:
|
|
case PlainValue.Type.BLOCK_LITERAL:
|
|
case PlainValue.Type.QUOTE_DOUBLE:
|
|
case PlainValue.Type.QUOTE_SINGLE:
|
|
return PlainValue.defaultTags.STR;
|
|
|
|
case PlainValue.Type.FLOW_MAP:
|
|
case PlainValue.Type.MAP:
|
|
return PlainValue.defaultTags.MAP;
|
|
|
|
case PlainValue.Type.FLOW_SEQ:
|
|
case PlainValue.Type.SEQ:
|
|
return PlainValue.defaultTags.SEQ;
|
|
|
|
case PlainValue.Type.PLAIN:
|
|
return nonSpecific ? PlainValue.defaultTags.STR : null;
|
|
|
|
default:
|
|
return null;
|
|
}
|
|
}
|
|
|
|
function resolveByTagName(doc, node, tagName) {
|
|
const {
|
|
tags
|
|
} = doc.schema;
|
|
const matchWithTest = [];
|
|
|
|
for (const tag of tags) {
|
|
if (tag.tag === tagName) {
|
|
if (tag.test) matchWithTest.push(tag);else {
|
|
const res = tag.resolve(doc, node);
|
|
return res instanceof Collection ? res : new Scalar(res);
|
|
}
|
|
}
|
|
}
|
|
|
|
const str = resolveString(doc, node);
|
|
if (typeof str === 'string' && matchWithTest.length > 0) return resolveScalar(str, matchWithTest, tags.scalarFallback);
|
|
return null;
|
|
}
|
|
|
|
function getFallbackTagName({
|
|
type
|
|
}) {
|
|
switch (type) {
|
|
case PlainValue.Type.FLOW_MAP:
|
|
case PlainValue.Type.MAP:
|
|
return PlainValue.defaultTags.MAP;
|
|
|
|
case PlainValue.Type.FLOW_SEQ:
|
|
case PlainValue.Type.SEQ:
|
|
return PlainValue.defaultTags.SEQ;
|
|
|
|
default:
|
|
return PlainValue.defaultTags.STR;
|
|
}
|
|
}
|
|
|
|
function resolveTag(doc, node, tagName) {
|
|
try {
|
|
const res = resolveByTagName(doc, node, tagName);
|
|
|
|
if (res) {
|
|
if (tagName && node.tag) res.tag = tagName;
|
|
return res;
|
|
}
|
|
} catch (error) {
|
|
/* istanbul ignore if */
|
|
if (!error.source) error.source = node;
|
|
doc.errors.push(error);
|
|
return null;
|
|
}
|
|
|
|
try {
|
|
const fallback = getFallbackTagName(node);
|
|
if (!fallback) throw new Error(`The tag ${tagName} is unavailable`);
|
|
const msg = `The tag ${tagName} is unavailable, falling back to ${fallback}`;
|
|
doc.warnings.push(new PlainValue.YAMLWarning(node, msg));
|
|
const res = resolveByTagName(doc, node, fallback);
|
|
res.tag = tagName;
|
|
return res;
|
|
} catch (error) {
|
|
const refError = new PlainValue.YAMLReferenceError(node, error.message);
|
|
refError.stack = error.stack;
|
|
doc.errors.push(refError);
|
|
return null;
|
|
}
|
|
}
|
|
|
|
const isCollectionItem = node => {
|
|
if (!node) return false;
|
|
const {
|
|
type
|
|
} = node;
|
|
return type === PlainValue.Type.MAP_KEY || type === PlainValue.Type.MAP_VALUE || type === PlainValue.Type.SEQ_ITEM;
|
|
};
|
|
|
|
function resolveNodeProps(errors, node) {
|
|
const comments = {
|
|
before: [],
|
|
after: []
|
|
};
|
|
let hasAnchor = false;
|
|
let hasTag = false;
|
|
const props = isCollectionItem(node.context.parent) ? node.context.parent.props.concat(node.props) : node.props;
|
|
|
|
for (const {
|
|
start,
|
|
end
|
|
} of props) {
|
|
switch (node.context.src[start]) {
|
|
case PlainValue.Char.COMMENT:
|
|
{
|
|
if (!node.commentHasRequiredWhitespace(start)) {
|
|
const msg = 'Comments must be separated from other tokens by white space characters';
|
|
errors.push(new PlainValue.YAMLSemanticError(node, msg));
|
|
}
|
|
|
|
const {
|
|
header,
|
|
valueRange
|
|
} = node;
|
|
const cc = valueRange && (start > valueRange.start || header && start > header.start) ? comments.after : comments.before;
|
|
cc.push(node.context.src.slice(start + 1, end));
|
|
break;
|
|
}
|
|
// Actual anchor & tag resolution is handled by schema, here we just complain
|
|
|
|
case PlainValue.Char.ANCHOR:
|
|
if (hasAnchor) {
|
|
const msg = 'A node can have at most one anchor';
|
|
errors.push(new PlainValue.YAMLSemanticError(node, msg));
|
|
}
|
|
|
|
hasAnchor = true;
|
|
break;
|
|
|
|
case PlainValue.Char.TAG:
|
|
if (hasTag) {
|
|
const msg = 'A node can have at most one tag';
|
|
errors.push(new PlainValue.YAMLSemanticError(node, msg));
|
|
}
|
|
|
|
hasTag = true;
|
|
break;
|
|
}
|
|
}
|
|
|
|
return {
|
|
comments,
|
|
hasAnchor,
|
|
hasTag
|
|
};
|
|
}
|
|
|
|
function resolveNodeValue(doc, node) {
|
|
const {
|
|
anchors,
|
|
errors,
|
|
schema
|
|
} = doc;
|
|
|
|
if (node.type === PlainValue.Type.ALIAS) {
|
|
const name = node.rawValue;
|
|
const src = anchors.getNode(name);
|
|
|
|
if (!src) {
|
|
const msg = `Aliased anchor not found: ${name}`;
|
|
errors.push(new PlainValue.YAMLReferenceError(node, msg));
|
|
return null;
|
|
} // Lazy resolution for circular references
|
|
|
|
|
|
const res = new Alias(src);
|
|
|
|
anchors._cstAliases.push(res);
|
|
|
|
return res;
|
|
}
|
|
|
|
const tagName = resolveTagName(doc, node);
|
|
if (tagName) return resolveTag(doc, node, tagName);
|
|
|
|
if (node.type !== PlainValue.Type.PLAIN) {
|
|
const msg = `Failed to resolve ${node.type} node here`;
|
|
errors.push(new PlainValue.YAMLSyntaxError(node, msg));
|
|
return null;
|
|
}
|
|
|
|
try {
|
|
const str = resolveString(doc, node);
|
|
return resolveScalar(str, schema.tags, schema.tags.scalarFallback);
|
|
} catch (error) {
|
|
if (!error.source) error.source = node;
|
|
errors.push(error);
|
|
return null;
|
|
}
|
|
} // sets node.resolved on success
|
|
|
|
|
|
function resolveNode(doc, node) {
|
|
if (!node) return null;
|
|
if (node.error) doc.errors.push(node.error);
|
|
const {
|
|
comments,
|
|
hasAnchor,
|
|
hasTag
|
|
} = resolveNodeProps(doc.errors, node);
|
|
|
|
if (hasAnchor) {
|
|
const {
|
|
anchors
|
|
} = doc;
|
|
const name = node.anchor;
|
|
const prev = anchors.getNode(name); // At this point, aliases for any preceding node with the same anchor
|
|
// name have already been resolved, so it may safely be renamed.
|
|
|
|
if (prev) anchors.map[anchors.newName(name)] = prev; // During parsing, we need to store the CST node in anchors.map as
|
|
// anchors need to be available during resolution to allow for
|
|
// circular references.
|
|
|
|
anchors.map[name] = node;
|
|
}
|
|
|
|
if (node.type === PlainValue.Type.ALIAS && (hasAnchor || hasTag)) {
|
|
const msg = 'An alias node must not specify any properties';
|
|
doc.errors.push(new PlainValue.YAMLSemanticError(node, msg));
|
|
}
|
|
|
|
const res = resolveNodeValue(doc, node);
|
|
|
|
if (res) {
|
|
res.range = [node.range.start, node.range.end];
|
|
if (doc.options.keepCstNodes) res.cstNode = node;
|
|
if (doc.options.keepNodeTypes) res.type = node.type;
|
|
const cb = comments.before.join('\n');
|
|
|
|
if (cb) {
|
|
res.commentBefore = res.commentBefore ? `${res.commentBefore}\n${cb}` : cb;
|
|
}
|
|
|
|
const ca = comments.after.join('\n');
|
|
if (ca) res.comment = res.comment ? `${res.comment}\n${ca}` : ca;
|
|
}
|
|
|
|
return node.resolved = res;
|
|
}
|
|
|
|
function resolveMap(doc, cst) {
|
|
if (cst.type !== PlainValue.Type.MAP && cst.type !== PlainValue.Type.FLOW_MAP) {
|
|
const msg = `A ${cst.type} node cannot be resolved as a mapping`;
|
|
doc.errors.push(new PlainValue.YAMLSyntaxError(cst, msg));
|
|
return null;
|
|
}
|
|
|
|
const {
|
|
comments,
|
|
items
|
|
} = cst.type === PlainValue.Type.FLOW_MAP ? resolveFlowMapItems(doc, cst) : resolveBlockMapItems(doc, cst);
|
|
const map = new YAMLMap();
|
|
map.items = items;
|
|
resolveComments(map, comments);
|
|
let hasCollectionKey = false;
|
|
|
|
for (let i = 0; i < items.length; ++i) {
|
|
const {
|
|
key: iKey
|
|
} = items[i];
|
|
if (iKey instanceof Collection) hasCollectionKey = true;
|
|
|
|
if (doc.schema.merge && iKey && iKey.value === MERGE_KEY) {
|
|
items[i] = new Merge(items[i]);
|
|
const sources = items[i].value.items;
|
|
let error = null;
|
|
sources.some(node => {
|
|
if (node instanceof Alias) {
|
|
// During parsing, alias sources are CST nodes; to account for
|
|
// circular references their resolved values can't be used here.
|
|
const {
|
|
type
|
|
} = node.source;
|
|
if (type === PlainValue.Type.MAP || type === PlainValue.Type.FLOW_MAP) return false;
|
|
return error = 'Merge nodes aliases can only point to maps';
|
|
}
|
|
|
|
return error = 'Merge nodes can only have Alias nodes as values';
|
|
});
|
|
if (error) doc.errors.push(new PlainValue.YAMLSemanticError(cst, error));
|
|
} else {
|
|
for (let j = i + 1; j < items.length; ++j) {
|
|
const {
|
|
key: jKey
|
|
} = items[j];
|
|
|
|
if (iKey === jKey || iKey && jKey && Object.prototype.hasOwnProperty.call(iKey, 'value') && iKey.value === jKey.value) {
|
|
const msg = `Map keys must be unique; "${iKey}" is repeated`;
|
|
doc.errors.push(new PlainValue.YAMLSemanticError(cst, msg));
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
if (hasCollectionKey && !doc.options.mapAsMap) {
|
|
const warn = 'Keys with collection values will be stringified as YAML due to JS Object restrictions. Use mapAsMap: true to avoid this.';
|
|
doc.warnings.push(new PlainValue.YAMLWarning(cst, warn));
|
|
}
|
|
|
|
cst.resolved = map;
|
|
return map;
|
|
}
|
|
|
|
const valueHasPairComment = ({
|
|
context: {
|
|
lineStart,
|
|
node,
|
|
src
|
|
},
|
|
props
|
|
}) => {
|
|
if (props.length === 0) return false;
|
|
const {
|
|
start
|
|
} = props[0];
|
|
if (node && start > node.valueRange.start) return false;
|
|
if (src[start] !== PlainValue.Char.COMMENT) return false;
|
|
|
|
for (let i = lineStart; i < start; ++i) if (src[i] === '\n') return false;
|
|
|
|
return true;
|
|
};
|
|
|
|
function resolvePairComment(item, pair) {
|
|
if (!valueHasPairComment(item)) return;
|
|
const comment = item.getPropValue(0, PlainValue.Char.COMMENT, true);
|
|
let found = false;
|
|
const cb = pair.value.commentBefore;
|
|
|
|
if (cb && cb.startsWith(comment)) {
|
|
pair.value.commentBefore = cb.substr(comment.length + 1);
|
|
found = true;
|
|
} else {
|
|
const cc = pair.value.comment;
|
|
|
|
if (!item.node && cc && cc.startsWith(comment)) {
|
|
pair.value.comment = cc.substr(comment.length + 1);
|
|
found = true;
|
|
}
|
|
}
|
|
|
|
if (found) pair.comment = comment;
|
|
}
|
|
|
|
function resolveBlockMapItems(doc, cst) {
|
|
const comments = [];
|
|
const items = [];
|
|
let key = undefined;
|
|
let keyStart = null;
|
|
|
|
for (let i = 0; i < cst.items.length; ++i) {
|
|
const item = cst.items[i];
|
|
|
|
switch (item.type) {
|
|
case PlainValue.Type.BLANK_LINE:
|
|
comments.push({
|
|
afterKey: !!key,
|
|
before: items.length
|
|
});
|
|
break;
|
|
|
|
case PlainValue.Type.COMMENT:
|
|
comments.push({
|
|
afterKey: !!key,
|
|
before: items.length,
|
|
comment: item.comment
|
|
});
|
|
break;
|
|
|
|
case PlainValue.Type.MAP_KEY:
|
|
if (key !== undefined) items.push(new Pair(key));
|
|
if (item.error) doc.errors.push(item.error);
|
|
key = resolveNode(doc, item.node);
|
|
keyStart = null;
|
|
break;
|
|
|
|
case PlainValue.Type.MAP_VALUE:
|
|
{
|
|
if (key === undefined) key = null;
|
|
if (item.error) doc.errors.push(item.error);
|
|
|
|
if (!item.context.atLineStart && item.node && item.node.type === PlainValue.Type.MAP && !item.node.context.atLineStart) {
|
|
const msg = 'Nested mappings are not allowed in compact mappings';
|
|
doc.errors.push(new PlainValue.YAMLSemanticError(item.node, msg));
|
|
}
|
|
|
|
let valueNode = item.node;
|
|
|
|
if (!valueNode && item.props.length > 0) {
|
|
// Comments on an empty mapping value need to be preserved, so we
|
|
// need to construct a minimal empty node here to use instead of the
|
|
// missing `item.node`. -- eemeli/yaml#19
|
|
valueNode = new PlainValue.PlainValue(PlainValue.Type.PLAIN, []);
|
|
valueNode.context = {
|
|
parent: item,
|
|
src: item.context.src
|
|
};
|
|
const pos = item.range.start + 1;
|
|
valueNode.range = {
|
|
start: pos,
|
|
end: pos
|
|
};
|
|
valueNode.valueRange = {
|
|
start: pos,
|
|
end: pos
|
|
};
|
|
|
|
if (typeof item.range.origStart === 'number') {
|
|
const origPos = item.range.origStart + 1;
|
|
valueNode.range.origStart = valueNode.range.origEnd = origPos;
|
|
valueNode.valueRange.origStart = valueNode.valueRange.origEnd = origPos;
|
|
}
|
|
}
|
|
|
|
const pair = new Pair(key, resolveNode(doc, valueNode));
|
|
resolvePairComment(item, pair);
|
|
items.push(pair);
|
|
|
|
if (key && typeof keyStart === 'number') {
|
|
if (item.range.start > keyStart + 1024) doc.errors.push(getLongKeyError(cst, key));
|
|
}
|
|
|
|
key = undefined;
|
|
keyStart = null;
|
|
}
|
|
break;
|
|
|
|
default:
|
|
if (key !== undefined) items.push(new Pair(key));
|
|
key = resolveNode(doc, item);
|
|
keyStart = item.range.start;
|
|
if (item.error) doc.errors.push(item.error);
|
|
|
|
next: for (let j = i + 1;; ++j) {
|
|
const nextItem = cst.items[j];
|
|
|
|
switch (nextItem && nextItem.type) {
|
|
case PlainValue.Type.BLANK_LINE:
|
|
case PlainValue.Type.COMMENT:
|
|
continue next;
|
|
|
|
case PlainValue.Type.MAP_VALUE:
|
|
break next;
|
|
|
|
default:
|
|
{
|
|
const msg = 'Implicit map keys need to be followed by map values';
|
|
doc.errors.push(new PlainValue.YAMLSemanticError(item, msg));
|
|
break next;
|
|
}
|
|
}
|
|
}
|
|
|
|
if (item.valueRangeContainsNewline) {
|
|
const msg = 'Implicit map keys need to be on a single line';
|
|
doc.errors.push(new PlainValue.YAMLSemanticError(item, msg));
|
|
}
|
|
|
|
}
|
|
}
|
|
|
|
if (key !== undefined) items.push(new Pair(key));
|
|
return {
|
|
comments,
|
|
items
|
|
};
|
|
}
|
|
|
|
function resolveFlowMapItems(doc, cst) {
|
|
const comments = [];
|
|
const items = [];
|
|
let key = undefined;
|
|
let explicitKey = false;
|
|
let next = '{';
|
|
|
|
for (let i = 0; i < cst.items.length; ++i) {
|
|
const item = cst.items[i];
|
|
|
|
if (typeof item.char === 'string') {
|
|
const {
|
|
char,
|
|
offset
|
|
} = item;
|
|
|
|
if (char === '?' && key === undefined && !explicitKey) {
|
|
explicitKey = true;
|
|
next = ':';
|
|
continue;
|
|
}
|
|
|
|
if (char === ':') {
|
|
if (key === undefined) key = null;
|
|
|
|
if (next === ':') {
|
|
next = ',';
|
|
continue;
|
|
}
|
|
} else {
|
|
if (explicitKey) {
|
|
if (key === undefined && char !== ',') key = null;
|
|
explicitKey = false;
|
|
}
|
|
|
|
if (key !== undefined) {
|
|
items.push(new Pair(key));
|
|
key = undefined;
|
|
|
|
if (char === ',') {
|
|
next = ':';
|
|
continue;
|
|
}
|
|
}
|
|
}
|
|
|
|
if (char === '}') {
|
|
if (i === cst.items.length - 1) continue;
|
|
} else if (char === next) {
|
|
next = ':';
|
|
continue;
|
|
}
|
|
|
|
const msg = `Flow map contains an unexpected ${char}`;
|
|
const err = new PlainValue.YAMLSyntaxError(cst, msg);
|
|
err.offset = offset;
|
|
doc.errors.push(err);
|
|
} else if (item.type === PlainValue.Type.BLANK_LINE) {
|
|
comments.push({
|
|
afterKey: !!key,
|
|
before: items.length
|
|
});
|
|
} else if (item.type === PlainValue.Type.COMMENT) {
|
|
checkFlowCommentSpace(doc.errors, item);
|
|
comments.push({
|
|
afterKey: !!key,
|
|
before: items.length,
|
|
comment: item.comment
|
|
});
|
|
} else if (key === undefined) {
|
|
if (next === ',') doc.errors.push(new PlainValue.YAMLSemanticError(item, 'Separator , missing in flow map'));
|
|
key = resolveNode(doc, item);
|
|
} else {
|
|
if (next !== ',') doc.errors.push(new PlainValue.YAMLSemanticError(item, 'Indicator : missing in flow map entry'));
|
|
items.push(new Pair(key, resolveNode(doc, item)));
|
|
key = undefined;
|
|
explicitKey = false;
|
|
}
|
|
}
|
|
|
|
checkFlowCollectionEnd(doc.errors, cst);
|
|
if (key !== undefined) items.push(new Pair(key));
|
|
return {
|
|
comments,
|
|
items
|
|
};
|
|
}
|
|
|
|
function resolveSeq(doc, cst) {
|
|
if (cst.type !== PlainValue.Type.SEQ && cst.type !== PlainValue.Type.FLOW_SEQ) {
|
|
const msg = `A ${cst.type} node cannot be resolved as a sequence`;
|
|
doc.errors.push(new PlainValue.YAMLSyntaxError(cst, msg));
|
|
return null;
|
|
}
|
|
|
|
const {
|
|
comments,
|
|
items
|
|
} = cst.type === PlainValue.Type.FLOW_SEQ ? resolveFlowSeqItems(doc, cst) : resolveBlockSeqItems(doc, cst);
|
|
const seq = new YAMLSeq();
|
|
seq.items = items;
|
|
resolveComments(seq, comments);
|
|
|
|
if (!doc.options.mapAsMap && items.some(it => it instanceof Pair && it.key instanceof Collection)) {
|
|
const warn = 'Keys with collection values will be stringified as YAML due to JS Object restrictions. Use mapAsMap: true to avoid this.';
|
|
doc.warnings.push(new PlainValue.YAMLWarning(cst, warn));
|
|
}
|
|
|
|
cst.resolved = seq;
|
|
return seq;
|
|
}
|
|
|
|
function resolveBlockSeqItems(doc, cst) {
|
|
const comments = [];
|
|
const items = [];
|
|
|
|
for (let i = 0; i < cst.items.length; ++i) {
|
|
const item = cst.items[i];
|
|
|
|
switch (item.type) {
|
|
case PlainValue.Type.BLANK_LINE:
|
|
comments.push({
|
|
before: items.length
|
|
});
|
|
break;
|
|
|
|
case PlainValue.Type.COMMENT:
|
|
comments.push({
|
|
comment: item.comment,
|
|
before: items.length
|
|
});
|
|
break;
|
|
|
|
case PlainValue.Type.SEQ_ITEM:
|
|
if (item.error) doc.errors.push(item.error);
|
|
items.push(resolveNode(doc, item.node));
|
|
|
|
if (item.hasProps) {
|
|
const msg = 'Sequence items cannot have tags or anchors before the - indicator';
|
|
doc.errors.push(new PlainValue.YAMLSemanticError(item, msg));
|
|
}
|
|
|
|
break;
|
|
|
|
default:
|
|
if (item.error) doc.errors.push(item.error);
|
|
doc.errors.push(new PlainValue.YAMLSyntaxError(item, `Unexpected ${item.type} node in sequence`));
|
|
}
|
|
}
|
|
|
|
return {
|
|
comments,
|
|
items
|
|
};
|
|
}
|
|
|
|
function resolveFlowSeqItems(doc, cst) {
|
|
const comments = [];
|
|
const items = [];
|
|
let explicitKey = false;
|
|
let key = undefined;
|
|
let keyStart = null;
|
|
let next = '[';
|
|
let prevItem = null;
|
|
|
|
for (let i = 0; i < cst.items.length; ++i) {
|
|
const item = cst.items[i];
|
|
|
|
if (typeof item.char === 'string') {
|
|
const {
|
|
char,
|
|
offset
|
|
} = item;
|
|
|
|
if (char !== ':' && (explicitKey || key !== undefined)) {
|
|
if (explicitKey && key === undefined) key = next ? items.pop() : null;
|
|
items.push(new Pair(key));
|
|
explicitKey = false;
|
|
key = undefined;
|
|
keyStart = null;
|
|
}
|
|
|
|
if (char === next) {
|
|
next = null;
|
|
} else if (!next && char === '?') {
|
|
explicitKey = true;
|
|
} else if (next !== '[' && char === ':' && key === undefined) {
|
|
if (next === ',') {
|
|
key = items.pop();
|
|
|
|
if (key instanceof Pair) {
|
|
const msg = 'Chaining flow sequence pairs is invalid';
|
|
const err = new PlainValue.YAMLSemanticError(cst, msg);
|
|
err.offset = offset;
|
|
doc.errors.push(err);
|
|
}
|
|
|
|
if (!explicitKey && typeof keyStart === 'number') {
|
|
const keyEnd = item.range ? item.range.start : item.offset;
|
|
if (keyEnd > keyStart + 1024) doc.errors.push(getLongKeyError(cst, key));
|
|
const {
|
|
src
|
|
} = prevItem.context;
|
|
|
|
for (let i = keyStart; i < keyEnd; ++i) if (src[i] === '\n') {
|
|
const msg = 'Implicit keys of flow sequence pairs need to be on a single line';
|
|
doc.errors.push(new PlainValue.YAMLSemanticError(prevItem, msg));
|
|
break;
|
|
}
|
|
}
|
|
} else {
|
|
key = null;
|
|
}
|
|
|
|
keyStart = null;
|
|
explicitKey = false;
|
|
next = null;
|
|
} else if (next === '[' || char !== ']' || i < cst.items.length - 1) {
|
|
const msg = `Flow sequence contains an unexpected ${char}`;
|
|
const err = new PlainValue.YAMLSyntaxError(cst, msg);
|
|
err.offset = offset;
|
|
doc.errors.push(err);
|
|
}
|
|
} else if (item.type === PlainValue.Type.BLANK_LINE) {
|
|
comments.push({
|
|
before: items.length
|
|
});
|
|
} else if (item.type === PlainValue.Type.COMMENT) {
|
|
checkFlowCommentSpace(doc.errors, item);
|
|
comments.push({
|
|
comment: item.comment,
|
|
before: items.length
|
|
});
|
|
} else {
|
|
if (next) {
|
|
const msg = `Expected a ${next} in flow sequence`;
|
|
doc.errors.push(new PlainValue.YAMLSemanticError(item, msg));
|
|
}
|
|
|
|
const value = resolveNode(doc, item);
|
|
|
|
if (key === undefined) {
|
|
items.push(value);
|
|
prevItem = item;
|
|
} else {
|
|
items.push(new Pair(key, value));
|
|
key = undefined;
|
|
}
|
|
|
|
keyStart = item.range.start;
|
|
next = ',';
|
|
}
|
|
}
|
|
|
|
checkFlowCollectionEnd(doc.errors, cst);
|
|
if (key !== undefined) items.push(new Pair(key));
|
|
return {
|
|
comments,
|
|
items
|
|
};
|
|
}
|
|
|
|
exports.Alias = Alias;
|
|
exports.Collection = Collection;
|
|
exports.Merge = Merge;
|
|
exports.Node = Node;
|
|
exports.Pair = Pair;
|
|
exports.Scalar = Scalar;
|
|
exports.YAMLMap = YAMLMap;
|
|
exports.YAMLSeq = YAMLSeq;
|
|
exports.addComment = addComment;
|
|
exports.binaryOptions = binaryOptions;
|
|
exports.boolOptions = boolOptions;
|
|
exports.findPair = findPair;
|
|
exports.intOptions = intOptions;
|
|
exports.isEmptyPath = isEmptyPath;
|
|
exports.nullOptions = nullOptions;
|
|
exports.resolveMap = resolveMap;
|
|
exports.resolveNode = resolveNode;
|
|
exports.resolveSeq = resolveSeq;
|
|
exports.resolveString = resolveString;
|
|
exports.strOptions = strOptions;
|
|
exports.stringifyNumber = stringifyNumber;
|
|
exports.stringifyString = stringifyString;
|
|
exports.toJSON = toJSON;
|
|
|
|
|
|
/***/ }),
|
|
|
|
/***/ 6003:
|
|
/***/ ((__unused_webpack_module, exports, __nccwpck_require__) => {
|
|
|
|
"use strict";
|
|
|
|
|
|
var PlainValue = __nccwpck_require__(5215);
|
|
var resolveSeq = __nccwpck_require__(4227);
|
|
|
|
/* global atob, btoa, Buffer */
|
|
const binary = {
|
|
identify: value => value instanceof Uint8Array,
|
|
// Buffer inherits from Uint8Array
|
|
default: false,
|
|
tag: 'tag:yaml.org,2002:binary',
|
|
|
|
/**
|
|
* Returns a Buffer in node and an Uint8Array in browsers
|
|
*
|
|
* To use the resulting buffer as an image, you'll want to do something like:
|
|
*
|
|
* const blob = new Blob([buffer], { type: 'image/jpeg' })
|
|
* document.querySelector('#photo').src = URL.createObjectURL(blob)
|
|
*/
|
|
resolve: (doc, node) => {
|
|
const src = resolveSeq.resolveString(doc, node);
|
|
|
|
if (typeof Buffer === 'function') {
|
|
return Buffer.from(src, 'base64');
|
|
} else if (typeof atob === 'function') {
|
|
// On IE 11, atob() can't handle newlines
|
|
const str = atob(src.replace(/[\n\r]/g, ''));
|
|
const buffer = new Uint8Array(str.length);
|
|
|
|
for (let i = 0; i < str.length; ++i) buffer[i] = str.charCodeAt(i);
|
|
|
|
return buffer;
|
|
} else {
|
|
const msg = 'This environment does not support reading binary tags; either Buffer or atob is required';
|
|
doc.errors.push(new PlainValue.YAMLReferenceError(node, msg));
|
|
return null;
|
|
}
|
|
},
|
|
options: resolveSeq.binaryOptions,
|
|
stringify: ({
|
|
comment,
|
|
type,
|
|
value
|
|
}, ctx, onComment, onChompKeep) => {
|
|
let src;
|
|
|
|
if (typeof Buffer === 'function') {
|
|
src = value instanceof Buffer ? value.toString('base64') : Buffer.from(value.buffer).toString('base64');
|
|
} else if (typeof btoa === 'function') {
|
|
let s = '';
|
|
|
|
for (let i = 0; i < value.length; ++i) s += String.fromCharCode(value[i]);
|
|
|
|
src = btoa(s);
|
|
} else {
|
|
throw new Error('This environment does not support writing binary tags; either Buffer or btoa is required');
|
|
}
|
|
|
|
if (!type) type = resolveSeq.binaryOptions.defaultType;
|
|
|
|
if (type === PlainValue.Type.QUOTE_DOUBLE) {
|
|
value = src;
|
|
} else {
|
|
const {
|
|
lineWidth
|
|
} = resolveSeq.binaryOptions;
|
|
const n = Math.ceil(src.length / lineWidth);
|
|
const lines = new Array(n);
|
|
|
|
for (let i = 0, o = 0; i < n; ++i, o += lineWidth) {
|
|
lines[i] = src.substr(o, lineWidth);
|
|
}
|
|
|
|
value = lines.join(type === PlainValue.Type.BLOCK_LITERAL ? '\n' : ' ');
|
|
}
|
|
|
|
return resolveSeq.stringifyString({
|
|
comment,
|
|
type,
|
|
value
|
|
}, ctx, onComment, onChompKeep);
|
|
}
|
|
};
|
|
|
|
function parsePairs(doc, cst) {
|
|
const seq = resolveSeq.resolveSeq(doc, cst);
|
|
|
|
for (let i = 0; i < seq.items.length; ++i) {
|
|
let item = seq.items[i];
|
|
if (item instanceof resolveSeq.Pair) continue;else if (item instanceof resolveSeq.YAMLMap) {
|
|
if (item.items.length > 1) {
|
|
const msg = 'Each pair must have its own sequence indicator';
|
|
throw new PlainValue.YAMLSemanticError(cst, msg);
|
|
}
|
|
|
|
const pair = item.items[0] || new resolveSeq.Pair();
|
|
if (item.commentBefore) pair.commentBefore = pair.commentBefore ? `${item.commentBefore}\n${pair.commentBefore}` : item.commentBefore;
|
|
if (item.comment) pair.comment = pair.comment ? `${item.comment}\n${pair.comment}` : item.comment;
|
|
item = pair;
|
|
}
|
|
seq.items[i] = item instanceof resolveSeq.Pair ? item : new resolveSeq.Pair(item);
|
|
}
|
|
|
|
return seq;
|
|
}
|
|
function createPairs(schema, iterable, ctx) {
|
|
const pairs = new resolveSeq.YAMLSeq(schema);
|
|
pairs.tag = 'tag:yaml.org,2002:pairs';
|
|
|
|
for (const it of iterable) {
|
|
let key, value;
|
|
|
|
if (Array.isArray(it)) {
|
|
if (it.length === 2) {
|
|
key = it[0];
|
|
value = it[1];
|
|
} else throw new TypeError(`Expected [key, value] tuple: ${it}`);
|
|
} else if (it && it instanceof Object) {
|
|
const keys = Object.keys(it);
|
|
|
|
if (keys.length === 1) {
|
|
key = keys[0];
|
|
value = it[key];
|
|
} else throw new TypeError(`Expected { key: value } tuple: ${it}`);
|
|
} else {
|
|
key = it;
|
|
}
|
|
|
|
const pair = schema.createPair(key, value, ctx);
|
|
pairs.items.push(pair);
|
|
}
|
|
|
|
return pairs;
|
|
}
|
|
const pairs = {
|
|
default: false,
|
|
tag: 'tag:yaml.org,2002:pairs',
|
|
resolve: parsePairs,
|
|
createNode: createPairs
|
|
};
|
|
|
|
class YAMLOMap extends resolveSeq.YAMLSeq {
|
|
constructor() {
|
|
super();
|
|
|
|
PlainValue._defineProperty(this, "add", resolveSeq.YAMLMap.prototype.add.bind(this));
|
|
|
|
PlainValue._defineProperty(this, "delete", resolveSeq.YAMLMap.prototype.delete.bind(this));
|
|
|
|
PlainValue._defineProperty(this, "get", resolveSeq.YAMLMap.prototype.get.bind(this));
|
|
|
|
PlainValue._defineProperty(this, "has", resolveSeq.YAMLMap.prototype.has.bind(this));
|
|
|
|
PlainValue._defineProperty(this, "set", resolveSeq.YAMLMap.prototype.set.bind(this));
|
|
|
|
this.tag = YAMLOMap.tag;
|
|
}
|
|
|
|
toJSON(_, ctx) {
|
|
const map = new Map();
|
|
if (ctx && ctx.onCreate) ctx.onCreate(map);
|
|
|
|
for (const pair of this.items) {
|
|
let key, value;
|
|
|
|
if (pair instanceof resolveSeq.Pair) {
|
|
key = resolveSeq.toJSON(pair.key, '', ctx);
|
|
value = resolveSeq.toJSON(pair.value, key, ctx);
|
|
} else {
|
|
key = resolveSeq.toJSON(pair, '', ctx);
|
|
}
|
|
|
|
if (map.has(key)) throw new Error('Ordered maps must not include duplicate keys');
|
|
map.set(key, value);
|
|
}
|
|
|
|
return map;
|
|
}
|
|
|
|
}
|
|
|
|
PlainValue._defineProperty(YAMLOMap, "tag", 'tag:yaml.org,2002:omap');
|
|
|
|
function parseOMap(doc, cst) {
|
|
const pairs = parsePairs(doc, cst);
|
|
const seenKeys = [];
|
|
|
|
for (const {
|
|
key
|
|
} of pairs.items) {
|
|
if (key instanceof resolveSeq.Scalar) {
|
|
if (seenKeys.includes(key.value)) {
|
|
const msg = 'Ordered maps must not include duplicate keys';
|
|
throw new PlainValue.YAMLSemanticError(cst, msg);
|
|
} else {
|
|
seenKeys.push(key.value);
|
|
}
|
|
}
|
|
}
|
|
|
|
return Object.assign(new YAMLOMap(), pairs);
|
|
}
|
|
|
|
function createOMap(schema, iterable, ctx) {
|
|
const pairs = createPairs(schema, iterable, ctx);
|
|
const omap = new YAMLOMap();
|
|
omap.items = pairs.items;
|
|
return omap;
|
|
}
|
|
|
|
const omap = {
|
|
identify: value => value instanceof Map,
|
|
nodeClass: YAMLOMap,
|
|
default: false,
|
|
tag: 'tag:yaml.org,2002:omap',
|
|
resolve: parseOMap,
|
|
createNode: createOMap
|
|
};
|
|
|
|
class YAMLSet extends resolveSeq.YAMLMap {
|
|
constructor() {
|
|
super();
|
|
this.tag = YAMLSet.tag;
|
|
}
|
|
|
|
add(key) {
|
|
const pair = key instanceof resolveSeq.Pair ? key : new resolveSeq.Pair(key);
|
|
const prev = resolveSeq.findPair(this.items, pair.key);
|
|
if (!prev) this.items.push(pair);
|
|
}
|
|
|
|
get(key, keepPair) {
|
|
const pair = resolveSeq.findPair(this.items, key);
|
|
return !keepPair && pair instanceof resolveSeq.Pair ? pair.key instanceof resolveSeq.Scalar ? pair.key.value : pair.key : pair;
|
|
}
|
|
|
|
set(key, value) {
|
|
if (typeof value !== 'boolean') throw new Error(`Expected boolean value for set(key, value) in a YAML set, not ${typeof value}`);
|
|
const prev = resolveSeq.findPair(this.items, key);
|
|
|
|
if (prev && !value) {
|
|
this.items.splice(this.items.indexOf(prev), 1);
|
|
} else if (!prev && value) {
|
|
this.items.push(new resolveSeq.Pair(key));
|
|
}
|
|
}
|
|
|
|
toJSON(_, ctx) {
|
|
return super.toJSON(_, ctx, Set);
|
|
}
|
|
|
|
toString(ctx, onComment, onChompKeep) {
|
|
if (!ctx) return JSON.stringify(this);
|
|
if (this.hasAllNullValues()) return super.toString(ctx, onComment, onChompKeep);else throw new Error('Set items must all have null values');
|
|
}
|
|
|
|
}
|
|
|
|
PlainValue._defineProperty(YAMLSet, "tag", 'tag:yaml.org,2002:set');
|
|
|
|
function parseSet(doc, cst) {
|
|
const map = resolveSeq.resolveMap(doc, cst);
|
|
if (!map.hasAllNullValues()) throw new PlainValue.YAMLSemanticError(cst, 'Set items must all have null values');
|
|
return Object.assign(new YAMLSet(), map);
|
|
}
|
|
|
|
function createSet(schema, iterable, ctx) {
|
|
const set = new YAMLSet();
|
|
|
|
for (const value of iterable) set.items.push(schema.createPair(value, null, ctx));
|
|
|
|
return set;
|
|
}
|
|
|
|
const set = {
|
|
identify: value => value instanceof Set,
|
|
nodeClass: YAMLSet,
|
|
default: false,
|
|
tag: 'tag:yaml.org,2002:set',
|
|
resolve: parseSet,
|
|
createNode: createSet
|
|
};
|
|
|
|
const parseSexagesimal = (sign, parts) => {
|
|
const n = parts.split(':').reduce((n, p) => n * 60 + Number(p), 0);
|
|
return sign === '-' ? -n : n;
|
|
}; // hhhh:mm:ss.sss
|
|
|
|
|
|
const stringifySexagesimal = ({
|
|
value
|
|
}) => {
|
|
if (isNaN(value) || !isFinite(value)) return resolveSeq.stringifyNumber(value);
|
|
let sign = '';
|
|
|
|
if (value < 0) {
|
|
sign = '-';
|
|
value = Math.abs(value);
|
|
}
|
|
|
|
const parts = [value % 60]; // seconds, including ms
|
|
|
|
if (value < 60) {
|
|
parts.unshift(0); // at least one : is required
|
|
} else {
|
|
value = Math.round((value - parts[0]) / 60);
|
|
parts.unshift(value % 60); // minutes
|
|
|
|
if (value >= 60) {
|
|
value = Math.round((value - parts[0]) / 60);
|
|
parts.unshift(value); // hours
|
|
}
|
|
}
|
|
|
|
return sign + parts.map(n => n < 10 ? '0' + String(n) : String(n)).join(':').replace(/000000\d*$/, '') // % 60 may introduce error
|
|
;
|
|
};
|
|
|
|
const intTime = {
|
|
identify: value => typeof value === 'number',
|
|
default: true,
|
|
tag: 'tag:yaml.org,2002:int',
|
|
format: 'TIME',
|
|
test: /^([-+]?)([0-9][0-9_]*(?::[0-5]?[0-9])+)$/,
|
|
resolve: (str, sign, parts) => parseSexagesimal(sign, parts.replace(/_/g, '')),
|
|
stringify: stringifySexagesimal
|
|
};
|
|
const floatTime = {
|
|
identify: value => typeof value === 'number',
|
|
default: true,
|
|
tag: 'tag:yaml.org,2002:float',
|
|
format: 'TIME',
|
|
test: /^([-+]?)([0-9][0-9_]*(?::[0-5]?[0-9])+\.[0-9_]*)$/,
|
|
resolve: (str, sign, parts) => parseSexagesimal(sign, parts.replace(/_/g, '')),
|
|
stringify: stringifySexagesimal
|
|
};
|
|
const timestamp = {
|
|
identify: value => value instanceof Date,
|
|
default: true,
|
|
tag: 'tag:yaml.org,2002:timestamp',
|
|
// If the time zone is omitted, the timestamp is assumed to be specified in UTC. The time part
|
|
// may be omitted altogether, resulting in a date format. In such a case, the time part is
|
|
// assumed to be 00:00:00Z (start of day, UTC).
|
|
test: RegExp('^(?:' + '([0-9]{4})-([0-9]{1,2})-([0-9]{1,2})' + // YYYY-Mm-Dd
|
|
'(?:(?:t|T|[ \\t]+)' + // t | T | whitespace
|
|
'([0-9]{1,2}):([0-9]{1,2}):([0-9]{1,2}(\\.[0-9]+)?)' + // Hh:Mm:Ss(.ss)?
|
|
'(?:[ \\t]*(Z|[-+][012]?[0-9](?::[0-9]{2})?))?' + // Z | +5 | -03:30
|
|
')?' + ')$'),
|
|
resolve: (str, year, month, day, hour, minute, second, millisec, tz) => {
|
|
if (millisec) millisec = (millisec + '00').substr(1, 3);
|
|
let date = Date.UTC(year, month - 1, day, hour || 0, minute || 0, second || 0, millisec || 0);
|
|
|
|
if (tz && tz !== 'Z') {
|
|
let d = parseSexagesimal(tz[0], tz.slice(1));
|
|
if (Math.abs(d) < 30) d *= 60;
|
|
date -= 60000 * d;
|
|
}
|
|
|
|
return new Date(date);
|
|
},
|
|
stringify: ({
|
|
value
|
|
}) => value.toISOString().replace(/((T00:00)?:00)?\.000Z$/, '')
|
|
};
|
|
|
|
/* global console, process, YAML_SILENCE_DEPRECATION_WARNINGS, YAML_SILENCE_WARNINGS */
|
|
function shouldWarn(deprecation) {
|
|
const env = typeof process !== 'undefined' && process.env || {};
|
|
|
|
if (deprecation) {
|
|
if (typeof YAML_SILENCE_DEPRECATION_WARNINGS !== 'undefined') return !YAML_SILENCE_DEPRECATION_WARNINGS;
|
|
return !env.YAML_SILENCE_DEPRECATION_WARNINGS;
|
|
}
|
|
|
|
if (typeof YAML_SILENCE_WARNINGS !== 'undefined') return !YAML_SILENCE_WARNINGS;
|
|
return !env.YAML_SILENCE_WARNINGS;
|
|
}
|
|
|
|
function warn(warning, type) {
|
|
if (shouldWarn(false)) {
|
|
const emit = typeof process !== 'undefined' && process.emitWarning; // This will throw in Jest if `warning` is an Error instance due to
|
|
// https://github.com/facebook/jest/issues/2549
|
|
|
|
if (emit) emit(warning, type);else {
|
|
// eslint-disable-next-line no-console
|
|
console.warn(type ? `${type}: ${warning}` : warning);
|
|
}
|
|
}
|
|
}
|
|
function warnFileDeprecation(filename) {
|
|
if (shouldWarn(true)) {
|
|
const path = filename.replace(/.*yaml[/\\]/i, '').replace(/\.js$/, '').replace(/\\/g, '/');
|
|
warn(`The endpoint 'yaml/${path}' will be removed in a future release.`, 'DeprecationWarning');
|
|
}
|
|
}
|
|
const warned = {};
|
|
function warnOptionDeprecation(name, alternative) {
|
|
if (!warned[name] && shouldWarn(true)) {
|
|
warned[name] = true;
|
|
let msg = `The option '${name}' will be removed in a future release`;
|
|
msg += alternative ? `, use '${alternative}' instead.` : '.';
|
|
warn(msg, 'DeprecationWarning');
|
|
}
|
|
}
|
|
|
|
exports.binary = binary;
|
|
exports.floatTime = floatTime;
|
|
exports.intTime = intTime;
|
|
exports.omap = omap;
|
|
exports.pairs = pairs;
|
|
exports.set = set;
|
|
exports.timestamp = timestamp;
|
|
exports.warn = warn;
|
|
exports.warnFileDeprecation = warnFileDeprecation;
|
|
exports.warnOptionDeprecation = warnOptionDeprecation;
|
|
|
|
|
|
/***/ }),
|
|
|
|
/***/ 3552:
|
|
/***/ ((module, __unused_webpack_exports, __nccwpck_require__) => {
|
|
|
|
module.exports = __nccwpck_require__(5065).YAML
|
|
|
|
|
|
/***/ }),
|
|
|
|
/***/ 2877:
|
|
/***/ ((module) => {
|
|
|
|
module.exports = eval("require")("encoding");
|
|
|
|
|
|
/***/ }),
|
|
|
|
/***/ 2357:
|
|
/***/ ((module) => {
|
|
|
|
"use strict";
|
|
module.exports = require("assert");
|
|
|
|
/***/ }),
|
|
|
|
/***/ 8614:
|
|
/***/ ((module) => {
|
|
|
|
"use strict";
|
|
module.exports = require("events");
|
|
|
|
/***/ }),
|
|
|
|
/***/ 5747:
|
|
/***/ ((module) => {
|
|
|
|
"use strict";
|
|
module.exports = require("fs");
|
|
|
|
/***/ }),
|
|
|
|
/***/ 8605:
|
|
/***/ ((module) => {
|
|
|
|
"use strict";
|
|
module.exports = require("http");
|
|
|
|
/***/ }),
|
|
|
|
/***/ 7211:
|
|
/***/ ((module) => {
|
|
|
|
"use strict";
|
|
module.exports = require("https");
|
|
|
|
/***/ }),
|
|
|
|
/***/ 1631:
|
|
/***/ ((module) => {
|
|
|
|
"use strict";
|
|
module.exports = require("net");
|
|
|
|
/***/ }),
|
|
|
|
/***/ 2087:
|
|
/***/ ((module) => {
|
|
|
|
"use strict";
|
|
module.exports = require("os");
|
|
|
|
/***/ }),
|
|
|
|
/***/ 5622:
|
|
/***/ ((module) => {
|
|
|
|
"use strict";
|
|
module.exports = require("path");
|
|
|
|
/***/ }),
|
|
|
|
/***/ 2413:
|
|
/***/ ((module) => {
|
|
|
|
"use strict";
|
|
module.exports = require("stream");
|
|
|
|
/***/ }),
|
|
|
|
/***/ 4016:
|
|
/***/ ((module) => {
|
|
|
|
"use strict";
|
|
module.exports = require("tls");
|
|
|
|
/***/ }),
|
|
|
|
/***/ 8835:
|
|
/***/ ((module) => {
|
|
|
|
"use strict";
|
|
module.exports = require("url");
|
|
|
|
/***/ }),
|
|
|
|
/***/ 1669:
|
|
/***/ ((module) => {
|
|
|
|
"use strict";
|
|
module.exports = require("util");
|
|
|
|
/***/ }),
|
|
|
|
/***/ 8761:
|
|
/***/ ((module) => {
|
|
|
|
"use strict";
|
|
module.exports = require("zlib");
|
|
|
|
/***/ })
|
|
|
|
/******/ });
|
|
/************************************************************************/
|
|
/******/ // The module cache
|
|
/******/ var __webpack_module_cache__ = {};
|
|
/******/
|
|
/******/ // The require function
|
|
/******/ function __nccwpck_require__(moduleId) {
|
|
/******/ // Check if module is in cache
|
|
/******/ var cachedModule = __webpack_module_cache__[moduleId];
|
|
/******/ if (cachedModule !== undefined) {
|
|
/******/ return cachedModule.exports;
|
|
/******/ }
|
|
/******/ // Create a new module (and put it into the cache)
|
|
/******/ var module = __webpack_module_cache__[moduleId] = {
|
|
/******/ // no module.id needed
|
|
/******/ // no module.loaded needed
|
|
/******/ exports: {}
|
|
/******/ };
|
|
/******/
|
|
/******/ // Execute the module function
|
|
/******/ var threw = true;
|
|
/******/ try {
|
|
/******/ __webpack_modules__[moduleId].call(module.exports, module, module.exports, __nccwpck_require__);
|
|
/******/ threw = false;
|
|
/******/ } finally {
|
|
/******/ if(threw) delete __webpack_module_cache__[moduleId];
|
|
/******/ }
|
|
/******/
|
|
/******/ // Return the exports of the module
|
|
/******/ return module.exports;
|
|
/******/ }
|
|
/******/
|
|
/************************************************************************/
|
|
/******/ /* webpack/runtime/compat get default export */
|
|
/******/ (() => {
|
|
/******/ // getDefaultExport function for compatibility with non-harmony modules
|
|
/******/ __nccwpck_require__.n = (module) => {
|
|
/******/ var getter = module && module.__esModule ?
|
|
/******/ () => (module['default']) :
|
|
/******/ () => (module);
|
|
/******/ __nccwpck_require__.d(getter, { a: getter });
|
|
/******/ return getter;
|
|
/******/ };
|
|
/******/ })();
|
|
/******/
|
|
/******/ /* webpack/runtime/define property getters */
|
|
/******/ (() => {
|
|
/******/ // define getter functions for harmony exports
|
|
/******/ __nccwpck_require__.d = (exports, definition) => {
|
|
/******/ for(var key in definition) {
|
|
/******/ if(__nccwpck_require__.o(definition, key) && !__nccwpck_require__.o(exports, key)) {
|
|
/******/ Object.defineProperty(exports, key, { enumerable: true, get: definition[key] });
|
|
/******/ }
|
|
/******/ }
|
|
/******/ };
|
|
/******/ })();
|
|
/******/
|
|
/******/ /* webpack/runtime/hasOwnProperty shorthand */
|
|
/******/ (() => {
|
|
/******/ __nccwpck_require__.o = (obj, prop) => (Object.prototype.hasOwnProperty.call(obj, prop))
|
|
/******/ })();
|
|
/******/
|
|
/******/ /* webpack/runtime/make namespace object */
|
|
/******/ (() => {
|
|
/******/ // define __esModule on exports
|
|
/******/ __nccwpck_require__.r = (exports) => {
|
|
/******/ if(typeof Symbol !== 'undefined' && Symbol.toStringTag) {
|
|
/******/ Object.defineProperty(exports, Symbol.toStringTag, { value: 'Module' });
|
|
/******/ }
|
|
/******/ Object.defineProperty(exports, '__esModule', { value: true });
|
|
/******/ };
|
|
/******/ })();
|
|
/******/
|
|
/******/ /* webpack/runtime/compat */
|
|
/******/
|
|
/******/ if (typeof __nccwpck_require__ !== 'undefined') __nccwpck_require__.ab = __dirname + "/";
|
|
/******/
|
|
/************************************************************************/
|
|
var __webpack_exports__ = {};
|
|
// This entry need to be wrapped in an IIFE because it need to be in strict mode.
|
|
(() => {
|
|
"use strict";
|
|
// ESM COMPAT FLAG
|
|
__nccwpck_require__.r(__webpack_exports__);
|
|
|
|
// EXPORTS
|
|
__nccwpck_require__.d(__webpack_exports__, {
|
|
"run": () => (/* binding */ run)
|
|
});
|
|
|
|
// EXTERNAL MODULE: ./node_modules/@actions/core/lib/core.js
|
|
var core = __nccwpck_require__(2186);
|
|
// EXTERNAL MODULE: ./node_modules/@actions/github/lib/github.js
|
|
var github = __nccwpck_require__(5438);
|
|
// EXTERNAL MODULE: ./node_modules/@octokit/request-error/dist-node/index.js
|
|
var dist_node = __nccwpck_require__(537);
|
|
;// CONCATENATED MODULE: ./src/dependabot/verified_commits.ts
|
|
var __awaiter = (undefined && undefined.__awaiter) || function (thisArg, _arguments, P, generator) {
|
|
function adopt(value) { return value instanceof P ? value : new P(function (resolve) { resolve(value); }); }
|
|
return new (P || (P = Promise))(function (resolve, reject) {
|
|
function fulfilled(value) { try { step(generator.next(value)); } catch (e) { reject(e); } }
|
|
function rejected(value) { try { step(generator["throw"](value)); } catch (e) { reject(e); } }
|
|
function step(result) { result.done ? resolve(result.value) : adopt(result.value).then(fulfilled, rejected); }
|
|
step((generator = generator.apply(thisArg, _arguments || [])).next());
|
|
});
|
|
};
|
|
|
|
const DEPENDABOT_LOGIN = 'dependabot[bot]';
|
|
function getMessage(client, context) {
|
|
var _a;
|
|
return __awaiter(this, void 0, void 0, function* () {
|
|
core.debug('Verifying the job is for an authentic Dependabot Pull Request');
|
|
const { pull_request: pr } = context.payload;
|
|
if (!pr) {
|
|
core.warning("Event payload missing `pull_request` key. Make sure you're " +
|
|
'triggering this action on the `pull_request` or `pull_request_target` events.');
|
|
return false;
|
|
}
|
|
// Don't bother hitting the API if the PR author isn't Dependabot
|
|
if (pr.user.login !== DEPENDABOT_LOGIN) {
|
|
core.debug(`PR author '${pr.user.login}' is not Dependabot.`);
|
|
return false;
|
|
}
|
|
core.debug('Verifying the Pull Request contents are from Dependabot');
|
|
const { data: commits } = yield client.rest.pulls.listCommits({
|
|
owner: context.repo.owner,
|
|
repo: context.repo.repo,
|
|
pull_number: pr.number
|
|
});
|
|
if (commits.length > 1) {
|
|
warnOtherCommits();
|
|
return false;
|
|
}
|
|
const { commit, author } = commits[0];
|
|
if ((author === null || author === void 0 ? void 0 : author.login) !== DEPENDABOT_LOGIN) {
|
|
warnOtherCommits();
|
|
return false;
|
|
}
|
|
if (!((_a = commit.verification) === null || _a === void 0 ? void 0 : _a.verified)) {
|
|
// TODO: Promote to setFailed
|
|
core.warning("Dependabot's commit signature is not verified, refusing to proceed.");
|
|
return false;
|
|
}
|
|
return commit.message;
|
|
});
|
|
}
|
|
function warnOtherCommits() {
|
|
core.warning("It looks like this PR has contains commits that aren't part of a Dependabot update. " +
|
|
"Try using '@dependabot rebase' to remove merge commits or '@dependabot recreate' to remove " +
|
|
'any non-Dependabot changes.');
|
|
}
|
|
function getAlert(name, version, directory, client, context) {
|
|
var _a, _b, _c, _d, _e;
|
|
return __awaiter(this, void 0, void 0, function* () {
|
|
const alerts = yield client.graphql(`
|
|
{
|
|
repository(owner: "${context.repo.owner}", name: "${context.repo.repo}") {
|
|
vulnerabilityAlerts(first: 100) {
|
|
nodes {
|
|
vulnerableManifestFilename
|
|
vulnerableManifestPath
|
|
vulnerableRequirements
|
|
state
|
|
securityVulnerability {
|
|
package { name }
|
|
}
|
|
securityAdvisory {
|
|
cvss { score }
|
|
ghsaId
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}`);
|
|
const nodes = (_b = (_a = alerts === null || alerts === void 0 ? void 0 : alerts.repository) === null || _a === void 0 ? void 0 : _a.vulnerabilityAlerts) === null || _b === void 0 ? void 0 : _b.nodes;
|
|
const found = nodes.find(a => (version === '' || a.vulnerableRequirements === `= ${version}`) &&
|
|
trimSlashes(a.vulnerableManifestPath) === `${trimSlashes(directory)}/${a.vulnerableManifestFilename}` &&
|
|
a.securityVulnerability.package.name === name);
|
|
return {
|
|
alertState: (_c = found === null || found === void 0 ? void 0 : found.state) !== null && _c !== void 0 ? _c : '',
|
|
ghsaId: (_d = found === null || found === void 0 ? void 0 : found.securityAdvisory.ghsaId) !== null && _d !== void 0 ? _d : '',
|
|
cvss: (_e = found === null || found === void 0 ? void 0 : found.securityAdvisory.cvss.score) !== null && _e !== void 0 ? _e : 0.0
|
|
};
|
|
});
|
|
}
|
|
function trimSlashes(value) {
|
|
return value.replace(/^\/+/, '').replace(/\/+$/, '');
|
|
}
|
|
|
|
// EXTERNAL MODULE: ./node_modules/yaml/index.js
|
|
var yaml = __nccwpck_require__(3552);
|
|
;// CONCATENATED MODULE: ./src/dependabot/update_metadata.ts
|
|
var update_metadata_awaiter = (undefined && undefined.__awaiter) || function (thisArg, _arguments, P, generator) {
|
|
function adopt(value) { return value instanceof P ? value : new P(function (resolve) { resolve(value); }); }
|
|
return new (P || (P = Promise))(function (resolve, reject) {
|
|
function fulfilled(value) { try { step(generator.next(value)); } catch (e) { reject(e); } }
|
|
function rejected(value) { try { step(generator["throw"](value)); } catch (e) { reject(e); } }
|
|
function step(result) { result.done ? resolve(result.value) : adopt(result.value).then(fulfilled, rejected); }
|
|
step((generator = generator.apply(thisArg, _arguments || [])).next());
|
|
});
|
|
};
|
|
|
|
function parse(commitMessage, branchName, mainBranch, lookup) {
|
|
var _a, _b, _c, _d;
|
|
return update_metadata_awaiter(this, void 0, void 0, function* () {
|
|
const bumpFragment = commitMessage.match(/^Bumps .* from (?<from>\d[^ ]*) to (?<to>\d[^ ]*)\.$/m);
|
|
const yamlFragment = commitMessage.match(/^-{3}\n(?<dependencies>[\S|\s]*?)\n^\.{3}\n/m);
|
|
if ((yamlFragment === null || yamlFragment === void 0 ? void 0 : yamlFragment.groups) && branchName.startsWith('dependabot')) {
|
|
const data = yaml.parse(yamlFragment.groups.dependencies);
|
|
// Since we are on the `dependabot` branch (9 letters), the 10th letter in the branch name is the delimiter
|
|
const delim = branchName[10];
|
|
const chunks = branchName.split(delim);
|
|
const prev = (_b = (_a = bumpFragment === null || bumpFragment === void 0 ? void 0 : bumpFragment.groups) === null || _a === void 0 ? void 0 : _a.from) !== null && _b !== void 0 ? _b : '';
|
|
const next = (_d = (_c = bumpFragment === null || bumpFragment === void 0 ? void 0 : bumpFragment.groups) === null || _c === void 0 ? void 0 : _c.to) !== null && _d !== void 0 ? _d : '';
|
|
if (data['updated-dependencies']) {
|
|
return yield Promise.all(data['updated-dependencies'].map((dependency, index) => update_metadata_awaiter(this, void 0, void 0, function* () {
|
|
const dirname = `/${chunks.slice(2, -1 * (1 + (dependency['dependency-name'].match(/\//g) || []).length)).join(delim) || ''}`;
|
|
return Object.assign({ dependencyName: dependency['dependency-name'], dependencyType: dependency['dependency-type'], updateType: dependency['update-type'], directory: dirname, packageEcosystem: chunks[1], targetBranch: mainBranch, prevVersion: index === 0 ? prev : '', newVersion: index === 0 ? next : '' }, yield lookup(dependency['dependency-name'], index === 0 ? prev : '', dirname));
|
|
})));
|
|
}
|
|
}
|
|
return Promise.resolve([]);
|
|
});
|
|
}
|
|
|
|
// EXTERNAL MODULE: ./node_modules/pluralize/pluralize.js
|
|
var pluralize = __nccwpck_require__(2522);
|
|
var pluralize_default = /*#__PURE__*/__nccwpck_require__.n(pluralize);
|
|
;// CONCATENATED MODULE: ./src/dependabot/output.ts
|
|
|
|
|
|
const DEPENDENCY_TYPES_PRIORITY = [
|
|
'direct:production',
|
|
'direct:development',
|
|
'indirect'
|
|
];
|
|
const UPDATE_TYPES_PRIORITY = [
|
|
'version-update:semver-major',
|
|
'version-update:semver-minor',
|
|
'version-update:semver-patch'
|
|
];
|
|
function set(updatedDependencies) {
|
|
const dependencyNames = updatedDependencies.map(dependency => {
|
|
return dependency.dependencyName;
|
|
}).join(', ');
|
|
const dependencyType = maxDependencyTypes(updatedDependencies);
|
|
const updateType = maxSemver(updatedDependencies);
|
|
const firstDependency = updatedDependencies[0];
|
|
const directory = firstDependency === null || firstDependency === void 0 ? void 0 : firstDependency.directory;
|
|
const ecosystem = firstDependency === null || firstDependency === void 0 ? void 0 : firstDependency.packageEcosystem;
|
|
const target = firstDependency === null || firstDependency === void 0 ? void 0 : firstDependency.targetBranch;
|
|
const prevVersion = firstDependency === null || firstDependency === void 0 ? void 0 : firstDependency.prevVersion;
|
|
const newVersion = firstDependency === null || firstDependency === void 0 ? void 0 : firstDependency.newVersion;
|
|
const alertState = firstDependency === null || firstDependency === void 0 ? void 0 : firstDependency.alertState;
|
|
const ghsaId = firstDependency === null || firstDependency === void 0 ? void 0 : firstDependency.ghsaId;
|
|
const cvss = firstDependency === null || firstDependency === void 0 ? void 0 : firstDependency.cvss;
|
|
core.startGroup(`Outputting metadata for ${pluralize_default()('updated dependency', updatedDependencies.length, true)}`);
|
|
core.info(`outputs.dependency-names: ${dependencyNames}`);
|
|
core.info(`outputs.dependency-type: ${dependencyType}`);
|
|
core.info(`outputs.update-type: ${updateType}`);
|
|
core.info(`outputs.directory: ${directory}`);
|
|
core.info(`outputs.package-ecosystem: ${ecosystem}`);
|
|
core.info(`outputs.target-branch: ${target}`);
|
|
core.info(`outputs.previous-version: ${prevVersion}`);
|
|
core.info(`outputs.new-version: ${newVersion}`);
|
|
core.info(`outputs.alert-state: ${alertState}`);
|
|
core.info(`outputs.ghsa-id: ${ghsaId}`);
|
|
core.info(`outputs.cvss: ${cvss}`);
|
|
core.endGroup();
|
|
core.setOutput('updated-dependencies-json', updatedDependencies);
|
|
core.setOutput('dependency-names', dependencyNames);
|
|
core.setOutput('dependency-type', dependencyType);
|
|
core.setOutput('update-type', updateType);
|
|
core.setOutput('directory', directory);
|
|
core.setOutput('package-ecosystem', ecosystem);
|
|
core.setOutput('target-branch', target);
|
|
core.setOutput('previous-version', prevVersion);
|
|
core.setOutput('new-version', newVersion);
|
|
core.setOutput('alert-state', alertState);
|
|
core.setOutput('ghsa-id', ghsaId);
|
|
core.setOutput('cvss', cvss);
|
|
}
|
|
function maxDependencyTypes(updatedDependencies) {
|
|
const dependencyTypes = updatedDependencies.reduce(function (dependencyTypes, dependency) {
|
|
dependencyTypes.add(dependency.dependencyType);
|
|
return dependencyTypes;
|
|
}, new Set());
|
|
return DEPENDENCY_TYPES_PRIORITY.find(dependencyType => dependencyTypes.has(dependencyType)) || 'unknown';
|
|
}
|
|
function maxSemver(updatedDependencies) {
|
|
const semverLevels = updatedDependencies.reduce(function (semverLevels, dependency) {
|
|
semverLevels.add(dependency.updateType);
|
|
return semverLevels;
|
|
}, new Set());
|
|
return UPDATE_TYPES_PRIORITY.find(semverLevel => semverLevels.has(semverLevel)) || null;
|
|
}
|
|
|
|
;// CONCATENATED MODULE: ./src/dependabot/util.ts
|
|
function parseNwo(nwo) {
|
|
const [owner, name] = nwo.split('/');
|
|
if (!owner || !name) {
|
|
throw new Error(`'${nwo}' does not appear to be a valid repository NWO`);
|
|
}
|
|
return { owner: owner, repo: name };
|
|
}
|
|
function getBranchNames(context) {
|
|
const { pull_request: pr } = context.payload;
|
|
return { headName: (pr === null || pr === void 0 ? void 0 : pr.head.ref) || '', baseName: pr === null || pr === void 0 ? void 0 : pr.base.ref };
|
|
}
|
|
|
|
;// CONCATENATED MODULE: ./src/main.ts
|
|
var main_awaiter = (undefined && undefined.__awaiter) || function (thisArg, _arguments, P, generator) {
|
|
function adopt(value) { return value instanceof P ? value : new P(function (resolve) { resolve(value); }); }
|
|
return new (P || (P = Promise))(function (resolve, reject) {
|
|
function fulfilled(value) { try { step(generator.next(value)); } catch (e) { reject(e); } }
|
|
function rejected(value) { try { step(generator["throw"](value)); } catch (e) { reject(e); } }
|
|
function step(result) { result.done ? resolve(result.value) : adopt(result.value).then(fulfilled, rejected); }
|
|
step((generator = generator.apply(thisArg, _arguments || [])).next());
|
|
});
|
|
};
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
function run() {
|
|
return main_awaiter(this, void 0, void 0, function* () {
|
|
const token = core.getInput('github-token');
|
|
if (!token) {
|
|
/* eslint-disable no-template-curly-in-string */
|
|
core.setFailed('github-token is not set! Please add \'github-token: "${{ secrets.GITHUB_TOKEN }}"\' to your workflow file.');
|
|
/* eslint-enable no-template-curly-in-string */
|
|
return;
|
|
}
|
|
try {
|
|
const githubClient = github.getOctokit(token);
|
|
// Validate the job
|
|
const commitMessage = yield getMessage(githubClient, github.context);
|
|
const branchNames = getBranchNames(github.context);
|
|
const alertLookup = (name, version, directory) => getAlert(name, version, directory, githubClient, github.context);
|
|
if (commitMessage) {
|
|
// Parse metadata
|
|
core.info('Parsing Dependabot metadata');
|
|
const updatedDependencies = yield parse(commitMessage, branchNames.headName, branchNames.baseName, alertLookup);
|
|
if (updatedDependencies.length > 0) {
|
|
set(updatedDependencies);
|
|
}
|
|
else {
|
|
core.setFailed('PR does not contain metadata, nothing to do.');
|
|
}
|
|
}
|
|
else {
|
|
core.setFailed('PR is not from Dependabot, nothing to do.');
|
|
}
|
|
}
|
|
catch (error) {
|
|
if (error instanceof dist_node.RequestError) {
|
|
core.setFailed(`Api Error: (${error.status}) ${error.message}`);
|
|
return;
|
|
}
|
|
if (error instanceof Error) {
|
|
core.setFailed(error.message);
|
|
}
|
|
else {
|
|
core.setFailed('There was an unexpected error.');
|
|
}
|
|
}
|
|
});
|
|
}
|
|
run();
|
|
|
|
})();
|
|
|
|
module.exports = __webpack_exports__;
|
|
/******/ })()
|
|
; |