forked from M3-Academy/challenge-algorithms-v2.0
980 lines
29 KiB
JavaScript
980 lines
29 KiB
JavaScript
'use strict';
|
|
|
|
Object.defineProperty(exports, '__esModule', {
|
|
value: true
|
|
});
|
|
exports.createScriptTransformer = createScriptTransformer;
|
|
exports.createTranspilingRequire = createTranspilingRequire;
|
|
function _crypto() {
|
|
const data = require('crypto');
|
|
_crypto = function () {
|
|
return data;
|
|
};
|
|
return data;
|
|
}
|
|
function path() {
|
|
const data = _interopRequireWildcard(require('path'));
|
|
path = function () {
|
|
return data;
|
|
};
|
|
return data;
|
|
}
|
|
function _core() {
|
|
const data = require('@babel/core');
|
|
_core = function () {
|
|
return data;
|
|
};
|
|
return data;
|
|
}
|
|
function _babelPluginIstanbul() {
|
|
const data = _interopRequireDefault(require('babel-plugin-istanbul'));
|
|
_babelPluginIstanbul = function () {
|
|
return data;
|
|
};
|
|
return data;
|
|
}
|
|
function _convertSourceMap() {
|
|
const data = require('convert-source-map');
|
|
_convertSourceMap = function () {
|
|
return data;
|
|
};
|
|
return data;
|
|
}
|
|
function _fastJsonStableStringify() {
|
|
const data = _interopRequireDefault(require('fast-json-stable-stringify'));
|
|
_fastJsonStableStringify = function () {
|
|
return data;
|
|
};
|
|
return data;
|
|
}
|
|
function fs() {
|
|
const data = _interopRequireWildcard(require('graceful-fs'));
|
|
fs = function () {
|
|
return data;
|
|
};
|
|
return data;
|
|
}
|
|
function _pirates() {
|
|
const data = require('pirates');
|
|
_pirates = function () {
|
|
return data;
|
|
};
|
|
return data;
|
|
}
|
|
function _slash() {
|
|
const data = _interopRequireDefault(require('slash'));
|
|
_slash = function () {
|
|
return data;
|
|
};
|
|
return data;
|
|
}
|
|
function _writeFileAtomic() {
|
|
const data = require('write-file-atomic');
|
|
_writeFileAtomic = function () {
|
|
return data;
|
|
};
|
|
return data;
|
|
}
|
|
function _jestHasteMap() {
|
|
const data = _interopRequireDefault(require('jest-haste-map'));
|
|
_jestHasteMap = function () {
|
|
return data;
|
|
};
|
|
return data;
|
|
}
|
|
function _jestUtil() {
|
|
const data = require('jest-util');
|
|
_jestUtil = function () {
|
|
return data;
|
|
};
|
|
return data;
|
|
}
|
|
var _enhanceUnexpectedTokenMessage = _interopRequireDefault(
|
|
require('./enhanceUnexpectedTokenMessage')
|
|
);
|
|
var _runtimeErrorsAndWarnings = require('./runtimeErrorsAndWarnings');
|
|
var _shouldInstrument = _interopRequireDefault(require('./shouldInstrument'));
|
|
function _interopRequireDefault(obj) {
|
|
return obj && obj.__esModule ? obj : {default: obj};
|
|
}
|
|
function _getRequireWildcardCache(nodeInterop) {
|
|
if (typeof WeakMap !== 'function') return null;
|
|
var cacheBabelInterop = new WeakMap();
|
|
var cacheNodeInterop = new WeakMap();
|
|
return (_getRequireWildcardCache = function (nodeInterop) {
|
|
return nodeInterop ? cacheNodeInterop : cacheBabelInterop;
|
|
})(nodeInterop);
|
|
}
|
|
function _interopRequireWildcard(obj, nodeInterop) {
|
|
if (!nodeInterop && obj && obj.__esModule) {
|
|
return obj;
|
|
}
|
|
if (obj === null || (typeof obj !== 'object' && typeof obj !== 'function')) {
|
|
return {default: obj};
|
|
}
|
|
var cache = _getRequireWildcardCache(nodeInterop);
|
|
if (cache && cache.has(obj)) {
|
|
return cache.get(obj);
|
|
}
|
|
var newObj = {};
|
|
var hasPropertyDescriptor =
|
|
Object.defineProperty && Object.getOwnPropertyDescriptor;
|
|
for (var key in obj) {
|
|
if (key !== 'default' && Object.prototype.hasOwnProperty.call(obj, key)) {
|
|
var desc = hasPropertyDescriptor
|
|
? Object.getOwnPropertyDescriptor(obj, key)
|
|
: null;
|
|
if (desc && (desc.get || desc.set)) {
|
|
Object.defineProperty(newObj, key, desc);
|
|
} else {
|
|
newObj[key] = obj[key];
|
|
}
|
|
}
|
|
}
|
|
newObj.default = obj;
|
|
if (cache) {
|
|
cache.set(obj, newObj);
|
|
}
|
|
return newObj;
|
|
}
|
|
/**
|
|
* Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved.
|
|
*
|
|
* This source code is licensed under the MIT license found in the
|
|
* LICENSE file in the root directory of this source tree.
|
|
*/
|
|
|
|
// @ts-expect-error: should just be `require.resolve`, but the tests mess that up
|
|
|
|
// Use `require` to avoid TS rootDir
|
|
const {version: VERSION} = require('../package.json');
|
|
// This data structure is used to avoid recalculating some data every time that
|
|
// we need to transform a file. Since ScriptTransformer is instantiated for each
|
|
// file we need to keep this object in the local scope of this module.
|
|
const projectCaches = new Map();
|
|
|
|
// To reset the cache for specific changesets (rather than package version).
|
|
const CACHE_VERSION = '1';
|
|
async function waitForPromiseWithCleanup(promise, cleanup) {
|
|
try {
|
|
await promise;
|
|
} finally {
|
|
cleanup();
|
|
}
|
|
}
|
|
|
|
// type predicate
|
|
function isTransformerFactory(t) {
|
|
return typeof t.createTransformer === 'function';
|
|
}
|
|
class ScriptTransformer {
|
|
_cache;
|
|
_transformCache = new Map();
|
|
_transformsAreLoaded = false;
|
|
constructor(_config, _cacheFS) {
|
|
this._config = _config;
|
|
this._cacheFS = _cacheFS;
|
|
const configString = (0, _fastJsonStableStringify().default)(this._config);
|
|
let projectCache = projectCaches.get(configString);
|
|
if (!projectCache) {
|
|
projectCache = {
|
|
configString,
|
|
ignorePatternsRegExp: calcIgnorePatternRegExp(this._config),
|
|
transformRegExp: calcTransformRegExp(this._config),
|
|
transformedFiles: new Map()
|
|
};
|
|
projectCaches.set(configString, projectCache);
|
|
}
|
|
this._cache = projectCache;
|
|
}
|
|
_buildCacheKeyFromFileInfo(
|
|
fileData,
|
|
filename,
|
|
transformOptions,
|
|
transformerCacheKey
|
|
) {
|
|
if (transformerCacheKey != null) {
|
|
return (0, _crypto().createHash)('sha1')
|
|
.update(transformerCacheKey)
|
|
.update(CACHE_VERSION)
|
|
.digest('hex')
|
|
.substring(0, 32);
|
|
}
|
|
return (0, _crypto().createHash)('sha1')
|
|
.update(fileData)
|
|
.update(transformOptions.configString)
|
|
.update(transformOptions.instrument ? 'instrument' : '')
|
|
.update(filename)
|
|
.update(CACHE_VERSION)
|
|
.digest('hex')
|
|
.substring(0, 32);
|
|
}
|
|
_getCacheKey(fileData, filename, options) {
|
|
const configString = this._cache.configString;
|
|
const {transformer, transformerConfig = {}} =
|
|
this._getTransformer(filename) ?? {};
|
|
let transformerCacheKey = undefined;
|
|
const transformOptions = {
|
|
...options,
|
|
cacheFS: this._cacheFS,
|
|
config: this._config,
|
|
configString,
|
|
transformerConfig
|
|
};
|
|
if (typeof transformer?.getCacheKey === 'function') {
|
|
transformerCacheKey = transformer.getCacheKey(
|
|
fileData,
|
|
filename,
|
|
transformOptions
|
|
);
|
|
}
|
|
return this._buildCacheKeyFromFileInfo(
|
|
fileData,
|
|
filename,
|
|
transformOptions,
|
|
transformerCacheKey
|
|
);
|
|
}
|
|
async _getCacheKeyAsync(fileData, filename, options) {
|
|
const configString = this._cache.configString;
|
|
const {transformer, transformerConfig = {}} =
|
|
this._getTransformer(filename) ?? {};
|
|
let transformerCacheKey = undefined;
|
|
const transformOptions = {
|
|
...options,
|
|
cacheFS: this._cacheFS,
|
|
config: this._config,
|
|
configString,
|
|
transformerConfig
|
|
};
|
|
if (transformer) {
|
|
const getCacheKey =
|
|
transformer.getCacheKeyAsync ?? transformer.getCacheKey;
|
|
if (typeof getCacheKey === 'function') {
|
|
transformerCacheKey = await getCacheKey(
|
|
fileData,
|
|
filename,
|
|
transformOptions
|
|
);
|
|
}
|
|
}
|
|
return this._buildCacheKeyFromFileInfo(
|
|
fileData,
|
|
filename,
|
|
transformOptions,
|
|
transformerCacheKey
|
|
);
|
|
}
|
|
_createFolderFromCacheKey(filename, cacheKey) {
|
|
const HasteMapClass = _jestHasteMap().default.getStatic(this._config);
|
|
const baseCacheDir = HasteMapClass.getCacheFilePath(
|
|
this._config.cacheDirectory,
|
|
`jest-transform-cache-${this._config.id}`,
|
|
VERSION
|
|
);
|
|
// Create sub folders based on the cacheKey to avoid creating one
|
|
// directory with many files.
|
|
const cacheDir = path().join(baseCacheDir, cacheKey[0] + cacheKey[1]);
|
|
const cacheFilenamePrefix = path()
|
|
.basename(filename, path().extname(filename))
|
|
.replace(/\W/g, '');
|
|
const cachePath = (0, _slash().default)(
|
|
path().join(cacheDir, `${cacheFilenamePrefix}_${cacheKey}`)
|
|
);
|
|
(0, _jestUtil().createDirectory)(cacheDir);
|
|
return cachePath;
|
|
}
|
|
_getFileCachePath(filename, content, options) {
|
|
const cacheKey = this._getCacheKey(content, filename, options);
|
|
return this._createFolderFromCacheKey(filename, cacheKey);
|
|
}
|
|
async _getFileCachePathAsync(filename, content, options) {
|
|
const cacheKey = await this._getCacheKeyAsync(content, filename, options);
|
|
return this._createFolderFromCacheKey(filename, cacheKey);
|
|
}
|
|
_getTransformPath(filename) {
|
|
const transformRegExp = this._cache.transformRegExp;
|
|
if (!transformRegExp) {
|
|
return undefined;
|
|
}
|
|
for (let i = 0; i < transformRegExp.length; i++) {
|
|
if (transformRegExp[i][0].test(filename)) {
|
|
return transformRegExp[i][1];
|
|
}
|
|
}
|
|
return undefined;
|
|
}
|
|
async loadTransformers() {
|
|
await Promise.all(
|
|
this._config.transform.map(
|
|
async ([, transformPath, transformerConfig]) => {
|
|
let transformer = await (0, _jestUtil().requireOrImportModule)(
|
|
transformPath
|
|
);
|
|
if (transformer == null) {
|
|
throw new Error(
|
|
(0, _runtimeErrorsAndWarnings.makeInvalidTransformerError)(
|
|
transformPath
|
|
)
|
|
);
|
|
}
|
|
if (isTransformerFactory(transformer)) {
|
|
transformer = transformer.createTransformer(transformerConfig);
|
|
}
|
|
if (
|
|
typeof transformer.process !== 'function' &&
|
|
typeof transformer.processAsync !== 'function'
|
|
) {
|
|
throw new Error(
|
|
(0, _runtimeErrorsAndWarnings.makeInvalidTransformerError)(
|
|
transformPath
|
|
)
|
|
);
|
|
}
|
|
const res = {
|
|
transformer,
|
|
transformerConfig
|
|
};
|
|
this._transformCache.set(transformPath, res);
|
|
}
|
|
)
|
|
);
|
|
this._transformsAreLoaded = true;
|
|
}
|
|
_getTransformer(filename) {
|
|
if (!this._transformsAreLoaded) {
|
|
throw new Error(
|
|
'Jest: Transformers have not been loaded yet - make sure to run `loadTransformers` and wait for it to complete before starting to transform files'
|
|
);
|
|
}
|
|
if (this._config.transform.length === 0) {
|
|
return null;
|
|
}
|
|
const transformPath = this._getTransformPath(filename);
|
|
if (transformPath == null) {
|
|
return null;
|
|
}
|
|
const cached = this._transformCache.get(transformPath);
|
|
if (cached != null) {
|
|
return cached;
|
|
}
|
|
throw new Error(
|
|
`Jest was unable to load the transformer defined for ${filename}. This is a bug in Jest, please open up an issue`
|
|
);
|
|
}
|
|
_instrumentFile(filename, input, canMapToInput, options) {
|
|
const inputCode = typeof input === 'string' ? input : input.code;
|
|
const inputMap = typeof input === 'string' ? null : input.map;
|
|
const result = (0, _core().transformSync)(inputCode, {
|
|
auxiliaryCommentBefore: ' istanbul ignore next ',
|
|
babelrc: false,
|
|
caller: {
|
|
name: '@jest/transform',
|
|
supportsDynamicImport: options.supportsDynamicImport,
|
|
supportsExportNamespaceFrom: options.supportsExportNamespaceFrom,
|
|
supportsStaticESM: options.supportsStaticESM,
|
|
supportsTopLevelAwait: options.supportsTopLevelAwait
|
|
},
|
|
configFile: false,
|
|
filename,
|
|
plugins: [
|
|
[
|
|
_babelPluginIstanbul().default,
|
|
{
|
|
compact: false,
|
|
// files outside `cwd` will not be instrumented
|
|
cwd: this._config.rootDir,
|
|
exclude: [],
|
|
extension: false,
|
|
inputSourceMap: inputMap,
|
|
useInlineSourceMaps: false
|
|
}
|
|
]
|
|
],
|
|
sourceMaps: canMapToInput ? 'both' : false
|
|
});
|
|
if (result?.code != null) {
|
|
return result;
|
|
}
|
|
return input;
|
|
}
|
|
_buildTransformResult(
|
|
filename,
|
|
cacheFilePath,
|
|
content,
|
|
transformer,
|
|
shouldCallTransform,
|
|
options,
|
|
processed,
|
|
sourceMapPath
|
|
) {
|
|
let transformed = {
|
|
code: content,
|
|
map: null
|
|
};
|
|
if (transformer && shouldCallTransform) {
|
|
if (processed != null && typeof processed.code === 'string') {
|
|
transformed = processed;
|
|
} else {
|
|
const transformPath = this._getTransformPath(filename);
|
|
invariant(transformPath);
|
|
throw new Error(
|
|
(0, _runtimeErrorsAndWarnings.makeInvalidReturnValueError)(
|
|
transformPath
|
|
)
|
|
);
|
|
}
|
|
}
|
|
if (transformed.map == null || transformed.map === '') {
|
|
try {
|
|
//Could be a potential freeze here.
|
|
//See: https://github.com/facebook/jest/pull/5177#discussion_r158883570
|
|
const inlineSourceMap = (0, _convertSourceMap().fromSource)(
|
|
transformed.code
|
|
);
|
|
if (inlineSourceMap) {
|
|
transformed.map = inlineSourceMap.toObject();
|
|
}
|
|
} catch {
|
|
const transformPath = this._getTransformPath(filename);
|
|
invariant(transformPath);
|
|
console.warn(
|
|
(0, _runtimeErrorsAndWarnings.makeInvalidSourceMapWarning)(
|
|
filename,
|
|
transformPath
|
|
)
|
|
);
|
|
}
|
|
}
|
|
|
|
// That means that the transform has a custom instrumentation
|
|
// logic and will handle it based on `config.collectCoverage` option
|
|
const transformWillInstrument =
|
|
shouldCallTransform && transformer && transformer.canInstrument;
|
|
|
|
// Apply instrumentation to the code if necessary, keeping the instrumented code and new map
|
|
let map = transformed.map;
|
|
let code;
|
|
if (transformWillInstrument !== true && options.instrument) {
|
|
/**
|
|
* We can map the original source code to the instrumented code ONLY if
|
|
* - the process of transforming the code produced a source map e.g. ts-jest
|
|
* - we did not transform the source code
|
|
*
|
|
* Otherwise we cannot make any statements about how the instrumented code corresponds to the original code,
|
|
* and we should NOT emit any source maps
|
|
*
|
|
*/
|
|
const shouldEmitSourceMaps =
|
|
(transformer != null && map != null) || transformer == null;
|
|
const instrumented = this._instrumentFile(
|
|
filename,
|
|
transformed,
|
|
shouldEmitSourceMaps,
|
|
options
|
|
);
|
|
code =
|
|
typeof instrumented === 'string' ? instrumented : instrumented.code;
|
|
map = typeof instrumented === 'string' ? null : instrumented.map;
|
|
} else {
|
|
code = transformed.code;
|
|
}
|
|
if (map != null) {
|
|
const sourceMapContent =
|
|
typeof map === 'string' ? map : JSON.stringify(map);
|
|
invariant(sourceMapPath, 'We should always have default sourceMapPath');
|
|
writeCacheFile(sourceMapPath, sourceMapContent);
|
|
} else {
|
|
sourceMapPath = null;
|
|
}
|
|
writeCodeCacheFile(cacheFilePath, code);
|
|
return {
|
|
code,
|
|
originalCode: content,
|
|
sourceMapPath
|
|
};
|
|
}
|
|
transformSource(filepath, content, options) {
|
|
const filename = (0, _jestUtil().tryRealpath)(filepath);
|
|
const {transformer, transformerConfig = {}} =
|
|
this._getTransformer(filename) ?? {};
|
|
const cacheFilePath = this._getFileCachePath(filename, content, options);
|
|
const sourceMapPath = `${cacheFilePath}.map`;
|
|
// Ignore cache if `config.cache` is set (--no-cache)
|
|
const code = this._config.cache ? readCodeCacheFile(cacheFilePath) : null;
|
|
if (code != null) {
|
|
// This is broken: we return the code, and a path for the source map
|
|
// directly from the cache. But, nothing ensures the source map actually
|
|
// matches that source code. They could have gotten out-of-sync in case
|
|
// two separate processes write concurrently to the same cache files.
|
|
return {
|
|
code,
|
|
originalCode: content,
|
|
sourceMapPath
|
|
};
|
|
}
|
|
let processed = null;
|
|
let shouldCallTransform = false;
|
|
if (transformer && this.shouldTransform(filename)) {
|
|
shouldCallTransform = true;
|
|
assertSyncTransformer(transformer, this._getTransformPath(filename));
|
|
processed = transformer.process(content, filename, {
|
|
...options,
|
|
cacheFS: this._cacheFS,
|
|
config: this._config,
|
|
configString: this._cache.configString,
|
|
transformerConfig
|
|
});
|
|
}
|
|
return this._buildTransformResult(
|
|
filename,
|
|
cacheFilePath,
|
|
content,
|
|
transformer,
|
|
shouldCallTransform,
|
|
options,
|
|
processed,
|
|
sourceMapPath
|
|
);
|
|
}
|
|
async transformSourceAsync(filepath, content, options) {
|
|
const filename = (0, _jestUtil().tryRealpath)(filepath);
|
|
const {transformer, transformerConfig = {}} =
|
|
this._getTransformer(filename) ?? {};
|
|
const cacheFilePath = await this._getFileCachePathAsync(
|
|
filename,
|
|
content,
|
|
options
|
|
);
|
|
const sourceMapPath = `${cacheFilePath}.map`;
|
|
// Ignore cache if `config.cache` is set (--no-cache)
|
|
const code = this._config.cache ? readCodeCacheFile(cacheFilePath) : null;
|
|
if (code != null) {
|
|
// This is broken: we return the code, and a path for the source map
|
|
// directly from the cache. But, nothing ensures the source map actually
|
|
// matches that source code. They could have gotten out-of-sync in case
|
|
// two separate processes write concurrently to the same cache files.
|
|
return {
|
|
code,
|
|
originalCode: content,
|
|
sourceMapPath
|
|
};
|
|
}
|
|
let processed = null;
|
|
let shouldCallTransform = false;
|
|
if (transformer && this.shouldTransform(filename)) {
|
|
shouldCallTransform = true;
|
|
const process = transformer.processAsync ?? transformer.process;
|
|
|
|
// This is probably dead code since `_getTransformerAsync` already asserts this
|
|
invariant(
|
|
typeof process === 'function',
|
|
'A transformer must always export either a `process` or `processAsync`'
|
|
);
|
|
processed = await process(content, filename, {
|
|
...options,
|
|
cacheFS: this._cacheFS,
|
|
config: this._config,
|
|
configString: this._cache.configString,
|
|
transformerConfig
|
|
});
|
|
}
|
|
return this._buildTransformResult(
|
|
filename,
|
|
cacheFilePath,
|
|
content,
|
|
transformer,
|
|
shouldCallTransform,
|
|
options,
|
|
processed,
|
|
sourceMapPath
|
|
);
|
|
}
|
|
async _transformAndBuildScriptAsync(
|
|
filename,
|
|
options,
|
|
transformOptions,
|
|
fileSource
|
|
) {
|
|
const {isInternalModule} = options;
|
|
let fileContent = fileSource ?? this._cacheFS.get(filename);
|
|
if (fileContent == null) {
|
|
fileContent = fs().readFileSync(filename, 'utf8');
|
|
this._cacheFS.set(filename, fileContent);
|
|
}
|
|
const content = stripShebang(fileContent);
|
|
let code = content;
|
|
let sourceMapPath = null;
|
|
const willTransform =
|
|
isInternalModule !== true &&
|
|
(transformOptions.instrument || this.shouldTransform(filename));
|
|
try {
|
|
if (willTransform) {
|
|
const transformedSource = await this.transformSourceAsync(
|
|
filename,
|
|
content,
|
|
transformOptions
|
|
);
|
|
code = transformedSource.code;
|
|
sourceMapPath = transformedSource.sourceMapPath;
|
|
}
|
|
return {
|
|
code,
|
|
originalCode: content,
|
|
sourceMapPath
|
|
};
|
|
} catch (e) {
|
|
if (!(e instanceof Error)) {
|
|
throw e;
|
|
}
|
|
throw (0, _enhanceUnexpectedTokenMessage.default)(e);
|
|
}
|
|
}
|
|
_transformAndBuildScript(filename, options, transformOptions, fileSource) {
|
|
const {isInternalModule} = options;
|
|
let fileContent = fileSource ?? this._cacheFS.get(filename);
|
|
if (fileContent == null) {
|
|
fileContent = fs().readFileSync(filename, 'utf8');
|
|
this._cacheFS.set(filename, fileContent);
|
|
}
|
|
const content = stripShebang(fileContent);
|
|
let code = content;
|
|
let sourceMapPath = null;
|
|
const willTransform =
|
|
isInternalModule !== true &&
|
|
(transformOptions.instrument || this.shouldTransform(filename));
|
|
try {
|
|
if (willTransform) {
|
|
const transformedSource = this.transformSource(
|
|
filename,
|
|
content,
|
|
transformOptions
|
|
);
|
|
code = transformedSource.code;
|
|
sourceMapPath = transformedSource.sourceMapPath;
|
|
}
|
|
return {
|
|
code,
|
|
originalCode: content,
|
|
sourceMapPath
|
|
};
|
|
} catch (e) {
|
|
if (!(e instanceof Error)) {
|
|
throw e;
|
|
}
|
|
throw (0, _enhanceUnexpectedTokenMessage.default)(e);
|
|
}
|
|
}
|
|
async transformAsync(filename, options, fileSource) {
|
|
const instrument =
|
|
options.coverageProvider === 'babel' &&
|
|
(0, _shouldInstrument.default)(filename, options, this._config);
|
|
const scriptCacheKey = getScriptCacheKey(filename, instrument);
|
|
let result = this._cache.transformedFiles.get(scriptCacheKey);
|
|
if (result) {
|
|
return result;
|
|
}
|
|
result = await this._transformAndBuildScriptAsync(
|
|
filename,
|
|
options,
|
|
{
|
|
...options,
|
|
instrument
|
|
},
|
|
fileSource
|
|
);
|
|
if (scriptCacheKey) {
|
|
this._cache.transformedFiles.set(scriptCacheKey, result);
|
|
}
|
|
return result;
|
|
}
|
|
transform(filename, options, fileSource) {
|
|
const instrument =
|
|
options.coverageProvider === 'babel' &&
|
|
(0, _shouldInstrument.default)(filename, options, this._config);
|
|
const scriptCacheKey = getScriptCacheKey(filename, instrument);
|
|
let result = this._cache.transformedFiles.get(scriptCacheKey);
|
|
if (result) {
|
|
return result;
|
|
}
|
|
result = this._transformAndBuildScript(
|
|
filename,
|
|
options,
|
|
{
|
|
...options,
|
|
instrument
|
|
},
|
|
fileSource
|
|
);
|
|
if (scriptCacheKey) {
|
|
this._cache.transformedFiles.set(scriptCacheKey, result);
|
|
}
|
|
return result;
|
|
}
|
|
transformJson(filename, options, fileSource) {
|
|
const {isInternalModule} = options;
|
|
const willTransform =
|
|
isInternalModule !== true && this.shouldTransform(filename);
|
|
if (willTransform) {
|
|
const {code: transformedJsonSource} = this.transformSource(
|
|
filename,
|
|
fileSource,
|
|
{
|
|
...options,
|
|
instrument: false
|
|
}
|
|
);
|
|
return transformedJsonSource;
|
|
}
|
|
return fileSource;
|
|
}
|
|
async requireAndTranspileModule(
|
|
moduleName,
|
|
callback,
|
|
options = {
|
|
applyInteropRequireDefault: true,
|
|
instrument: false,
|
|
supportsDynamicImport: false,
|
|
supportsExportNamespaceFrom: false,
|
|
supportsStaticESM: false,
|
|
supportsTopLevelAwait: false
|
|
}
|
|
) {
|
|
let transforming = false;
|
|
const {applyInteropRequireDefault, ...transformOptions} = options;
|
|
const revertHook = (0, _pirates().addHook)(
|
|
(code, filename) => {
|
|
try {
|
|
transforming = true;
|
|
return (
|
|
this.transformSource(filename, code, transformOptions).code || code
|
|
);
|
|
} finally {
|
|
transforming = false;
|
|
}
|
|
},
|
|
{
|
|
// Exclude `mjs` extension when addHook because pirates don't support hijack es module
|
|
exts: this._config.moduleFileExtensions
|
|
.filter(ext => ext !== 'mjs')
|
|
.map(ext => `.${ext}`),
|
|
ignoreNodeModules: false,
|
|
matcher: filename => {
|
|
if (transforming) {
|
|
// Don't transform any dependency required by the transformer itself
|
|
return false;
|
|
}
|
|
return this.shouldTransform(filename);
|
|
}
|
|
}
|
|
);
|
|
try {
|
|
const module = await (0, _jestUtil().requireOrImportModule)(
|
|
moduleName,
|
|
applyInteropRequireDefault
|
|
);
|
|
if (!callback) {
|
|
revertHook();
|
|
return module;
|
|
}
|
|
const cbResult = callback(module);
|
|
if ((0, _jestUtil().isPromise)(cbResult)) {
|
|
return await waitForPromiseWithCleanup(cbResult, revertHook).then(
|
|
() => module
|
|
);
|
|
}
|
|
return module;
|
|
} finally {
|
|
revertHook();
|
|
}
|
|
}
|
|
shouldTransform(filename) {
|
|
const ignoreRegexp = this._cache.ignorePatternsRegExp;
|
|
const isIgnored = ignoreRegexp ? ignoreRegexp.test(filename) : false;
|
|
return this._config.transform.length !== 0 && !isIgnored;
|
|
}
|
|
}
|
|
|
|
// TODO: do we need to define the generics twice?
|
|
async function createTranspilingRequire(config) {
|
|
const transformer = await createScriptTransformer(config);
|
|
return async function requireAndTranspileModule(
|
|
resolverPath,
|
|
applyInteropRequireDefault = false
|
|
) {
|
|
const transpiledModule = await transformer.requireAndTranspileModule(
|
|
resolverPath,
|
|
// eslint-disable-next-line @typescript-eslint/no-empty-function
|
|
() => {},
|
|
{
|
|
applyInteropRequireDefault,
|
|
instrument: false,
|
|
supportsDynamicImport: false,
|
|
// this might be true, depending on node version.
|
|
supportsExportNamespaceFrom: false,
|
|
supportsStaticESM: false,
|
|
supportsTopLevelAwait: false
|
|
}
|
|
);
|
|
return transpiledModule;
|
|
};
|
|
}
|
|
const removeFile = path => {
|
|
try {
|
|
fs().unlinkSync(path);
|
|
} catch {}
|
|
};
|
|
const stripShebang = content => {
|
|
// If the file data starts with a shebang remove it. Leaves the empty line
|
|
// to keep stack trace line numbers correct.
|
|
if (content.startsWith('#!')) {
|
|
return content.replace(/^#!.*/, '');
|
|
} else {
|
|
return content;
|
|
}
|
|
};
|
|
|
|
/**
|
|
* This is like `writeCacheFile` but with an additional sanity checksum. We
|
|
* cannot use the same technique for source maps because we expose source map
|
|
* cache file paths directly to callsites, with the expectation they can read
|
|
* it right away. This is not a great system, because source map cache file
|
|
* could get corrupted, out-of-sync, etc.
|
|
*/
|
|
function writeCodeCacheFile(cachePath, code) {
|
|
const checksum = (0, _crypto().createHash)('sha1')
|
|
.update(code)
|
|
.digest('hex')
|
|
.substring(0, 32);
|
|
writeCacheFile(cachePath, `${checksum}\n${code}`);
|
|
}
|
|
|
|
/**
|
|
* Read counterpart of `writeCodeCacheFile`. We verify that the content of the
|
|
* file matches the checksum, in case some kind of corruption happened. This
|
|
* could happen if an older version of `jest-runtime` writes non-atomically to
|
|
* the same cache, for example.
|
|
*/
|
|
function readCodeCacheFile(cachePath) {
|
|
const content = readCacheFile(cachePath);
|
|
if (content == null) {
|
|
return null;
|
|
}
|
|
const code = content.substring(33);
|
|
const checksum = (0, _crypto().createHash)('sha1')
|
|
.update(code)
|
|
.digest('hex')
|
|
.substring(0, 32);
|
|
if (checksum === content.substring(0, 32)) {
|
|
return code;
|
|
}
|
|
return null;
|
|
}
|
|
|
|
/**
|
|
* Writing to the cache atomically relies on 'rename' being atomic on most
|
|
* file systems. Doing atomic write reduces the risk of corruption by avoiding
|
|
* two processes to write to the same file at the same time. It also reduces
|
|
* the risk of reading a file that's being overwritten at the same time.
|
|
*/
|
|
const writeCacheFile = (cachePath, fileData) => {
|
|
try {
|
|
(0, _writeFileAtomic().sync)(cachePath, fileData, {
|
|
encoding: 'utf8',
|
|
fsync: false
|
|
});
|
|
} catch (e) {
|
|
if (!(e instanceof Error)) {
|
|
throw e;
|
|
}
|
|
if (cacheWriteErrorSafeToIgnore(e, cachePath)) {
|
|
return;
|
|
}
|
|
e.message = `jest: failed to cache transform results in: ${cachePath}\nFailure message: ${e.message}`;
|
|
removeFile(cachePath);
|
|
throw e;
|
|
}
|
|
};
|
|
|
|
/**
|
|
* On Windows, renames are not atomic, leading to EPERM exceptions when two
|
|
* processes attempt to rename to the same target file at the same time.
|
|
* If the target file exists we can be reasonably sure another process has
|
|
* legitimately won a cache write race and ignore the error.
|
|
*/
|
|
const cacheWriteErrorSafeToIgnore = (e, cachePath) =>
|
|
process.platform === 'win32' &&
|
|
e.code === 'EPERM' &&
|
|
fs().existsSync(cachePath);
|
|
const readCacheFile = cachePath => {
|
|
if (!fs().existsSync(cachePath)) {
|
|
return null;
|
|
}
|
|
let fileData;
|
|
try {
|
|
fileData = fs().readFileSync(cachePath, 'utf8');
|
|
} catch (e) {
|
|
if (!(e instanceof Error)) {
|
|
throw e;
|
|
}
|
|
// on windows write-file-atomic is not atomic which can
|
|
// result in this error
|
|
if (e.code === 'ENOENT' && process.platform === 'win32') {
|
|
return null;
|
|
}
|
|
e.message = `jest: failed to read cache file: ${cachePath}\nFailure message: ${e.message}`;
|
|
removeFile(cachePath);
|
|
throw e;
|
|
}
|
|
if (fileData == null) {
|
|
// We must have somehow created the file but failed to write to it,
|
|
// let's delete it and retry.
|
|
removeFile(cachePath);
|
|
}
|
|
return fileData;
|
|
};
|
|
const getScriptCacheKey = (filename, instrument) => {
|
|
const mtime = fs().statSync(filename).mtime;
|
|
return `${filename}_${mtime.getTime()}${instrument ? '_instrumented' : ''}`;
|
|
};
|
|
const calcIgnorePatternRegExp = config => {
|
|
if (
|
|
config.transformIgnorePatterns == null ||
|
|
config.transformIgnorePatterns.length === 0
|
|
) {
|
|
return undefined;
|
|
}
|
|
return new RegExp(config.transformIgnorePatterns.join('|'));
|
|
};
|
|
const calcTransformRegExp = config => {
|
|
if (!config.transform.length) {
|
|
return undefined;
|
|
}
|
|
const transformRegexp = [];
|
|
for (let i = 0; i < config.transform.length; i++) {
|
|
transformRegexp.push([
|
|
new RegExp(config.transform[i][0]),
|
|
config.transform[i][1],
|
|
config.transform[i][2]
|
|
]);
|
|
}
|
|
return transformRegexp;
|
|
};
|
|
function invariant(condition, message) {
|
|
if (condition == null || condition === false || condition === '') {
|
|
throw new Error(message);
|
|
}
|
|
}
|
|
function assertSyncTransformer(transformer, name) {
|
|
invariant(name);
|
|
invariant(
|
|
typeof transformer.process === 'function',
|
|
(0, _runtimeErrorsAndWarnings.makeInvalidSyncTransformerError)(name)
|
|
);
|
|
}
|
|
async function createScriptTransformer(config, cacheFS = new Map()) {
|
|
const transformer = new ScriptTransformer(config, cacheFS);
|
|
await transformer.loadTransformers();
|
|
return transformer;
|
|
}
|