mirror of https://github.com/webpack/webpack.git
				
				
				
			
		
			
				
	
	
		
			1333 lines
		
	
	
		
			41 KiB
		
	
	
	
		
			JavaScript
		
	
	
	
			
		
		
	
	
			1333 lines
		
	
	
		
			41 KiB
		
	
	
	
		
			JavaScript
		
	
	
	
| /*
 | |
| 	MIT License http://www.opensource.org/licenses/mit-license.php
 | |
| 	Author Tobias Koppers @sokra
 | |
| */
 | |
| 
 | |
| "use strict";
 | |
| 
 | |
| const AsyncDependencyToInitialChunkError = require("./AsyncDependencyToInitialChunkError");
 | |
| const { connectChunkGroupParentAndChild } = require("./GraphHelpers");
 | |
| const ModuleGraphConnection = require("./ModuleGraphConnection");
 | |
| const { getEntryRuntime, mergeRuntime } = require("./util/runtime");
 | |
| 
 | |
| /** @typedef {import("./AsyncDependenciesBlock")} AsyncDependenciesBlock */
 | |
| /** @typedef {import("./Chunk")} Chunk */
 | |
| /** @typedef {import("./ChunkGroup")} ChunkGroup */
 | |
| /** @typedef {import("./Compilation")} Compilation */
 | |
| /** @typedef {import("./DependenciesBlock")} DependenciesBlock */
 | |
| /** @typedef {import("./Dependency")} Dependency */
 | |
| /** @typedef {import("./Dependency").DependencyLocation} DependencyLocation */
 | |
| /** @typedef {import("./Entrypoint")} Entrypoint */
 | |
| /** @typedef {import("./Module")} Module */
 | |
| /** @typedef {import("./ModuleGraph")} ModuleGraph */
 | |
| /** @typedef {import("./ModuleGraphConnection").ConnectionState} ConnectionState */
 | |
| /** @typedef {import("./logging/Logger").Logger} Logger */
 | |
| /** @typedef {import("./util/runtime").RuntimeSpec} RuntimeSpec */
 | |
| 
 | |
| /**
 | |
|  * @typedef {object} QueueItem
 | |
|  * @property {number} action
 | |
|  * @property {DependenciesBlock} block
 | |
|  * @property {Module} module
 | |
|  * @property {Chunk} chunk
 | |
|  * @property {ChunkGroup} chunkGroup
 | |
|  * @property {ChunkGroupInfo} chunkGroupInfo
 | |
|  */
 | |
| 
 | |
| /**
 | |
|  * @typedef {object} ChunkGroupInfo
 | |
|  * @property {ChunkGroup} chunkGroup the chunk group
 | |
|  * @property {RuntimeSpec} runtime the runtimes
 | |
|  * @property {bigint | undefined} minAvailableModules current minimal set of modules available at this point
 | |
|  * @property {bigint[]} availableModulesToBeMerged enqueued updates to the minimal set of available modules
 | |
|  * @property {Set<Module>=} skippedItems modules that were skipped because module is already available in parent chunks (need to reconsider when minAvailableModules is shrinking)
 | |
|  * @property {Set<[Module, ModuleGraphConnection[]]>=} skippedModuleConnections referenced modules that where skipped because they were not active in this runtime
 | |
|  * @property {bigint | undefined} resultingAvailableModules set of modules available including modules from this chunk group
 | |
|  * @property {Set<ChunkGroupInfo> | undefined} children set of children chunk groups, that will be revisited when availableModules shrink
 | |
|  * @property {Set<ChunkGroupInfo> | undefined} availableSources set of chunk groups that are the source for minAvailableModules
 | |
|  * @property {Set<ChunkGroupInfo> | undefined} availableChildren set of chunk groups which depend on the this chunk group as availableSource
 | |
|  * @property {number} preOrderIndex next pre order index
 | |
|  * @property {number} postOrderIndex next post order index
 | |
|  * @property {boolean} chunkLoading has a chunk loading mechanism
 | |
|  * @property {boolean} asyncChunks create async chunks
 | |
|  */
 | |
| 
 | |
| /**
 | |
|  * @typedef {object} BlockChunkGroupConnection
 | |
|  * @property {ChunkGroupInfo} originChunkGroupInfo origin chunk group
 | |
|  * @property {ChunkGroup} chunkGroup referenced chunk group
 | |
|  */
 | |
| 
 | |
| /** @typedef {(Module | ConnectionState | ModuleGraphConnection)[]} BlockModulesInTuples */
 | |
| /** @typedef {(Module | ConnectionState | ModuleGraphConnection[])[]} BlockModulesInFlattenTuples */
 | |
| /** @typedef {Map<DependenciesBlock, BlockModulesInFlattenTuples>} BlockModulesMap */
 | |
| /** @typedef {Map<Chunk, bigint>} MaskByChunk */
 | |
| /** @typedef {Set<DependenciesBlock>} BlocksWithNestedBlocks */
 | |
| /** @typedef {Map<AsyncDependenciesBlock, BlockChunkGroupConnection[]>} BlockConnections */
 | |
| /** @typedef {Map<ChunkGroup, ChunkGroupInfo>} ChunkGroupInfoMap */
 | |
| /** @typedef {Set<ChunkGroup>} AllCreatedChunkGroups */
 | |
| /** @typedef {Map<Entrypoint, Module[]>} InputEntrypointsAndModules */
 | |
| 
 | |
| const ZERO_BIGINT = BigInt(0);
 | |
| const ONE_BIGINT = BigInt(1);
 | |
| 
 | |
| /**
 | |
|  * @param {bigint} mask The mask to test
 | |
|  * @param {number} ordinal The ordinal of the bit to test
 | |
|  * @returns {boolean} If the ordinal-th bit is set in the mask
 | |
|  */
 | |
| const isOrdinalSetInMask = (mask, ordinal) =>
 | |
| 	BigInt.asUintN(1, mask >> BigInt(ordinal)) !== ZERO_BIGINT;
 | |
| 
 | |
| /**
 | |
|  * @param {ModuleGraphConnection[]} connections list of connections
 | |
|  * @param {RuntimeSpec} runtime for which runtime
 | |
|  * @returns {ConnectionState} connection state
 | |
|  */
 | |
| const getActiveStateOfConnections = (connections, runtime) => {
 | |
| 	let merged = connections[0].getActiveState(runtime);
 | |
| 	if (merged === true) return true;
 | |
| 	for (let i = 1; i < connections.length; i++) {
 | |
| 		const c = connections[i];
 | |
| 		merged = ModuleGraphConnection.addConnectionStates(
 | |
| 			merged,
 | |
| 			c.getActiveState(runtime)
 | |
| 		);
 | |
| 		if (merged === true) return true;
 | |
| 	}
 | |
| 	return merged;
 | |
| };
 | |
| 
 | |
| /**
 | |
|  * @param {Module} module module
 | |
|  * @param {ModuleGraph} moduleGraph module graph
 | |
|  * @param {RuntimeSpec} runtime runtime
 | |
|  * @param {BlockModulesMap} blockModulesMap block modules map
 | |
|  */
 | |
| const extractBlockModules = (module, moduleGraph, runtime, blockModulesMap) => {
 | |
| 	/** @type {DependenciesBlock | undefined} */
 | |
| 	let blockCache;
 | |
| 	/** @type {BlockModulesInTuples | undefined} */
 | |
| 	let modules;
 | |
| 
 | |
| 	/** @type {BlockModulesInTuples[]} */
 | |
| 	const arrays = [];
 | |
| 
 | |
| 	/** @type {DependenciesBlock[]} */
 | |
| 	const queue = [module];
 | |
| 	while (queue.length > 0) {
 | |
| 		const block = /** @type {DependenciesBlock} */ (queue.pop());
 | |
| 		/** @type {Module[]} */
 | |
| 		const arr = [];
 | |
| 		arrays.push(arr);
 | |
| 		blockModulesMap.set(block, arr);
 | |
| 		for (const b of block.blocks) {
 | |
| 			queue.push(b);
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	for (const connection of moduleGraph.getOutgoingConnections(module)) {
 | |
| 		const d = connection.dependency;
 | |
| 		// We skip connections without dependency
 | |
| 		if (!d) continue;
 | |
| 		const m = connection.module;
 | |
| 		// We skip connections without Module pointer
 | |
| 		if (!m) continue;
 | |
| 		// We skip weak connections
 | |
| 		if (connection.weak) continue;
 | |
| 
 | |
| 		const block = moduleGraph.getParentBlock(d);
 | |
| 		let index = moduleGraph.getParentBlockIndex(d);
 | |
| 
 | |
| 		// deprecated fallback
 | |
| 		if (index < 0) {
 | |
| 			index = /** @type {DependenciesBlock} */ (block).dependencies.indexOf(d);
 | |
| 		}
 | |
| 
 | |
| 		if (blockCache !== block) {
 | |
| 			modules =
 | |
| 				/** @type {BlockModulesInTuples} */
 | |
| 				(
 | |
| 					blockModulesMap.get(
 | |
| 						(blockCache = /** @type {DependenciesBlock} */ (block))
 | |
| 					)
 | |
| 				);
 | |
| 		}
 | |
| 
 | |
| 		const i = index * 3;
 | |
| 		/** @type {BlockModulesInTuples} */
 | |
| 		(modules)[i] = m;
 | |
| 		/** @type {BlockModulesInTuples} */
 | |
| 		(modules)[i + 1] = connection.getActiveState(runtime);
 | |
| 		/** @type {BlockModulesInTuples} */
 | |
| 		(modules)[i + 2] = connection;
 | |
| 	}
 | |
| 
 | |
| 	for (const modules of arrays) {
 | |
| 		if (modules.length === 0) continue;
 | |
| 		let indexMap;
 | |
| 		let length = 0;
 | |
| 		outer: for (let j = 0; j < modules.length; j += 3) {
 | |
| 			const m = modules[j];
 | |
| 			if (m === undefined) continue;
 | |
| 			const state = /** @type {ConnectionState} */ (modules[j + 1]);
 | |
| 			const connection = /** @type {ModuleGraphConnection} */ (modules[j + 2]);
 | |
| 			if (indexMap === undefined) {
 | |
| 				let i = 0;
 | |
| 				for (; i < length; i += 3) {
 | |
| 					if (modules[i] === m) {
 | |
| 						const merged = /** @type {ConnectionState} */ (modules[i + 1]);
 | |
| 						/** @type {ModuleGraphConnection[]} */
 | |
| 						(/** @type {unknown} */ (modules[i + 2])).push(connection);
 | |
| 						if (merged === true) continue outer;
 | |
| 						modules[i + 1] = ModuleGraphConnection.addConnectionStates(
 | |
| 							merged,
 | |
| 							state
 | |
| 						);
 | |
| 						continue outer;
 | |
| 					}
 | |
| 				}
 | |
| 				modules[length] = m;
 | |
| 				length++;
 | |
| 				modules[length] = state;
 | |
| 				length++;
 | |
| 				/** @type {ModuleGraphConnection[]} */
 | |
| 				(/** @type {unknown} */ (modules[length])) = [connection];
 | |
| 				length++;
 | |
| 				if (length > 30) {
 | |
| 					// To avoid worse case performance, we will use an index map for
 | |
| 					// linear cost access, which allows to maintain O(n) complexity
 | |
| 					// while keeping allocations down to a minimum
 | |
| 					indexMap = new Map();
 | |
| 					for (let i = 0; i < length; i += 3) {
 | |
| 						indexMap.set(modules[i], i + 1);
 | |
| 					}
 | |
| 				}
 | |
| 			} else {
 | |
| 				const idx = indexMap.get(m);
 | |
| 				if (idx !== undefined) {
 | |
| 					const merged = /** @type {ConnectionState} */ (modules[idx]);
 | |
| 					/** @type {ModuleGraphConnection[]} */
 | |
| 					(/** @type {unknown} */ (modules[idx + 1])).push(connection);
 | |
| 					if (merged === true) continue outer;
 | |
| 					modules[idx] = ModuleGraphConnection.addConnectionStates(
 | |
| 						merged,
 | |
| 						state
 | |
| 					);
 | |
| 				} else {
 | |
| 					modules[length] = m;
 | |
| 					length++;
 | |
| 					modules[length] = state;
 | |
| 					indexMap.set(m, length);
 | |
| 					length++;
 | |
| 					/** @type {ModuleGraphConnection[]} */
 | |
| 					(
 | |
| 						/** @type {unknown} */
 | |
| 						(modules[length])
 | |
| 					) = [connection];
 | |
| 					length++;
 | |
| 				}
 | |
| 			}
 | |
| 		}
 | |
| 		modules.length = length;
 | |
| 	}
 | |
| };
 | |
| 
 | |
| /**
 | |
|  *
 | |
|  * @param {Logger} logger a logger
 | |
|  * @param {Compilation} compilation the compilation
 | |
|  * @param {InputEntrypointsAndModules} inputEntrypointsAndModules chunk groups which are processed with the modules
 | |
|  * @param {ChunkGroupInfoMap} chunkGroupInfoMap mapping from chunk group to available modules
 | |
|  * @param {BlockConnections} blockConnections connection for blocks
 | |
|  * @param {BlocksWithNestedBlocks} blocksWithNestedBlocks flag for blocks that have nested blocks
 | |
|  * @param {AllCreatedChunkGroups} allCreatedChunkGroups filled with all chunk groups that are created here
 | |
|  * @param {MaskByChunk} maskByChunk module content mask by chunk
 | |
|  */
 | |
| const visitModules = (
 | |
| 	logger,
 | |
| 	compilation,
 | |
| 	inputEntrypointsAndModules,
 | |
| 	chunkGroupInfoMap,
 | |
| 	blockConnections,
 | |
| 	blocksWithNestedBlocks,
 | |
| 	allCreatedChunkGroups,
 | |
| 	maskByChunk
 | |
| ) => {
 | |
| 	const { moduleGraph, chunkGraph, moduleMemCaches } = compilation;
 | |
| 
 | |
| 	const blockModulesRuntimeMap = new Map();
 | |
| 
 | |
| 	/** @type {BlockModulesMap | undefined} */
 | |
| 	let blockModulesMap;
 | |
| 
 | |
| 	/** @type {Map<Module, number>} */
 | |
| 	const ordinalByModule = new Map();
 | |
| 
 | |
| 	/**
 | |
| 	 * @param {Module} module The module to look up
 | |
| 	 * @returns {number} The ordinal of the module in masks
 | |
| 	 */
 | |
| 	const getModuleOrdinal = module => {
 | |
| 		let ordinal = ordinalByModule.get(module);
 | |
| 		if (ordinal === undefined) {
 | |
| 			ordinal = ordinalByModule.size;
 | |
| 			ordinalByModule.set(module, ordinal);
 | |
| 		}
 | |
| 		return ordinal;
 | |
| 	};
 | |
| 
 | |
| 	for (const chunk of compilation.chunks) {
 | |
| 		let mask = ZERO_BIGINT;
 | |
| 		for (const m of chunkGraph.getChunkModulesIterable(chunk)) {
 | |
| 			mask |= ONE_BIGINT << BigInt(getModuleOrdinal(m));
 | |
| 		}
 | |
| 		maskByChunk.set(chunk, mask);
 | |
| 	}
 | |
| 
 | |
| 	/**
 | |
| 	 *
 | |
| 	 * @param {DependenciesBlock} block block
 | |
| 	 * @param {RuntimeSpec} runtime runtime
 | |
| 	 * @returns {BlockModulesInFlattenTuples} block modules in flatten tuples
 | |
| 	 */
 | |
| 	const getBlockModules = (block, runtime) => {
 | |
| 		blockModulesMap = blockModulesRuntimeMap.get(runtime);
 | |
| 		if (blockModulesMap === undefined) {
 | |
| 			blockModulesMap = new Map();
 | |
| 			blockModulesRuntimeMap.set(runtime, blockModulesMap);
 | |
| 		}
 | |
| 		let blockModules = blockModulesMap.get(block);
 | |
| 		if (blockModules !== undefined) return blockModules;
 | |
| 		const module = /** @type {Module} */ (block.getRootBlock());
 | |
| 		const memCache = moduleMemCaches && moduleMemCaches.get(module);
 | |
| 		if (memCache !== undefined) {
 | |
| 			const map = memCache.provide(
 | |
| 				"bundleChunkGraph.blockModules",
 | |
| 				runtime,
 | |
| 				() => {
 | |
| 					logger.time("visitModules: prepare");
 | |
| 					const map = new Map();
 | |
| 					extractBlockModules(module, moduleGraph, runtime, map);
 | |
| 					logger.timeAggregate("visitModules: prepare");
 | |
| 					return map;
 | |
| 				}
 | |
| 			);
 | |
| 			for (const [block, blockModules] of map)
 | |
| 				blockModulesMap.set(block, blockModules);
 | |
| 			return map.get(block);
 | |
| 		} else {
 | |
| 			logger.time("visitModules: prepare");
 | |
| 			extractBlockModules(module, moduleGraph, runtime, blockModulesMap);
 | |
| 			blockModules =
 | |
| 				/** @type {BlockModulesInFlattenTuples} */
 | |
| 				(blockModulesMap.get(block));
 | |
| 			logger.timeAggregate("visitModules: prepare");
 | |
| 			return blockModules;
 | |
| 		}
 | |
| 	};
 | |
| 
 | |
| 	let statProcessedQueueItems = 0;
 | |
| 	let statProcessedBlocks = 0;
 | |
| 	let statConnectedChunkGroups = 0;
 | |
| 	let statProcessedChunkGroupsForMerging = 0;
 | |
| 	let statMergedAvailableModuleSets = 0;
 | |
| 	let statForkedAvailableModules = 0;
 | |
| 	let statForkedAvailableModulesCount = 0;
 | |
| 	let statForkedAvailableModulesCountPlus = 0;
 | |
| 	let statForkedMergedModulesCount = 0;
 | |
| 	let statForkedMergedModulesCountPlus = 0;
 | |
| 	let statForkedResultModulesCount = 0;
 | |
| 	let statChunkGroupInfoUpdated = 0;
 | |
| 	let statChildChunkGroupsReconnected = 0;
 | |
| 
 | |
| 	let nextChunkGroupIndex = 0;
 | |
| 	let nextFreeModulePreOrderIndex = 0;
 | |
| 	let nextFreeModulePostOrderIndex = 0;
 | |
| 
 | |
| 	/** @type {Map<DependenciesBlock, ChunkGroupInfo>} */
 | |
| 	const blockChunkGroups = new Map();
 | |
| 
 | |
| 	/** @type {Map<ChunkGroupInfo, DependenciesBlock>} */
 | |
| 	const blockByChunkGroups = new Map();
 | |
| 
 | |
| 	/** @type {Map<string, ChunkGroupInfo>} */
 | |
| 	const namedChunkGroups = new Map();
 | |
| 
 | |
| 	/** @type {Map<string, ChunkGroupInfo>} */
 | |
| 	const namedAsyncEntrypoints = new Map();
 | |
| 
 | |
| 	/** @type {Set<ChunkGroupInfo>} */
 | |
| 	const outdatedOrderIndexChunkGroups = new Set();
 | |
| 
 | |
| 	const ADD_AND_ENTER_ENTRY_MODULE = 0;
 | |
| 	const ADD_AND_ENTER_MODULE = 1;
 | |
| 	const ENTER_MODULE = 2;
 | |
| 	const PROCESS_BLOCK = 3;
 | |
| 	const PROCESS_ENTRY_BLOCK = 4;
 | |
| 	const LEAVE_MODULE = 5;
 | |
| 
 | |
| 	/** @type {QueueItem[]} */
 | |
| 	let queue = [];
 | |
| 
 | |
| 	/** @type {Map<ChunkGroupInfo, Set<ChunkGroupInfo>>} */
 | |
| 	const queueConnect = new Map();
 | |
| 	/** @type {Set<ChunkGroupInfo>} */
 | |
| 	const chunkGroupsForCombining = new Set();
 | |
| 
 | |
| 	// Fill queue with entrypoint modules
 | |
| 	// Create ChunkGroupInfo for entrypoints
 | |
| 	for (const [chunkGroup, modules] of inputEntrypointsAndModules) {
 | |
| 		const runtime = getEntryRuntime(
 | |
| 			compilation,
 | |
| 			/** @type {string} */ (chunkGroup.name),
 | |
| 			chunkGroup.options
 | |
| 		);
 | |
| 		/** @type {ChunkGroupInfo} */
 | |
| 		const chunkGroupInfo = {
 | |
| 			chunkGroup,
 | |
| 			runtime,
 | |
| 			minAvailableModules: undefined,
 | |
| 			availableModulesToBeMerged: [],
 | |
| 			skippedItems: undefined,
 | |
| 			resultingAvailableModules: undefined,
 | |
| 			children: undefined,
 | |
| 			availableSources: undefined,
 | |
| 			availableChildren: undefined,
 | |
| 			preOrderIndex: 0,
 | |
| 			postOrderIndex: 0,
 | |
| 			chunkLoading:
 | |
| 				chunkGroup.options.chunkLoading !== undefined
 | |
| 					? chunkGroup.options.chunkLoading !== false
 | |
| 					: compilation.outputOptions.chunkLoading !== false,
 | |
| 			asyncChunks:
 | |
| 				chunkGroup.options.asyncChunks !== undefined
 | |
| 					? chunkGroup.options.asyncChunks
 | |
| 					: compilation.outputOptions.asyncChunks !== false
 | |
| 		};
 | |
| 		chunkGroup.index = nextChunkGroupIndex++;
 | |
| 		if (chunkGroup.getNumberOfParents() > 0) {
 | |
| 			// minAvailableModules for child entrypoints are unknown yet, set to undefined.
 | |
| 			// This means no module is added until other sets are merged into
 | |
| 			// this minAvailableModules (by the parent entrypoints)
 | |
| 			const skippedItems = new Set(modules);
 | |
| 			chunkGroupInfo.skippedItems = skippedItems;
 | |
| 			chunkGroupsForCombining.add(chunkGroupInfo);
 | |
| 		} else {
 | |
| 			// The application may start here: We start with an empty list of available modules
 | |
| 			chunkGroupInfo.minAvailableModules = ZERO_BIGINT;
 | |
| 			const chunk = chunkGroup.getEntrypointChunk();
 | |
| 			for (const module of modules) {
 | |
| 				queue.push({
 | |
| 					action: ADD_AND_ENTER_MODULE,
 | |
| 					block: module,
 | |
| 					module,
 | |
| 					chunk,
 | |
| 					chunkGroup,
 | |
| 					chunkGroupInfo
 | |
| 				});
 | |
| 			}
 | |
| 		}
 | |
| 		chunkGroupInfoMap.set(chunkGroup, chunkGroupInfo);
 | |
| 		if (chunkGroup.name) {
 | |
| 			namedChunkGroups.set(chunkGroup.name, chunkGroupInfo);
 | |
| 		}
 | |
| 	}
 | |
| 	// Fill availableSources with parent-child dependencies between entrypoints
 | |
| 	for (const chunkGroupInfo of chunkGroupsForCombining) {
 | |
| 		const { chunkGroup } = chunkGroupInfo;
 | |
| 		chunkGroupInfo.availableSources = new Set();
 | |
| 		for (const parent of chunkGroup.parentsIterable) {
 | |
| 			const parentChunkGroupInfo =
 | |
| 				/** @type {ChunkGroupInfo} */
 | |
| 				(chunkGroupInfoMap.get(parent));
 | |
| 			chunkGroupInfo.availableSources.add(parentChunkGroupInfo);
 | |
| 			if (parentChunkGroupInfo.availableChildren === undefined) {
 | |
| 				parentChunkGroupInfo.availableChildren = new Set();
 | |
| 			}
 | |
| 			parentChunkGroupInfo.availableChildren.add(chunkGroupInfo);
 | |
| 		}
 | |
| 	}
 | |
| 	// pop() is used to read from the queue
 | |
| 	// so it need to be reversed to be iterated in
 | |
| 	// correct order
 | |
| 	queue.reverse();
 | |
| 
 | |
| 	/** @type {Set<ChunkGroupInfo>} */
 | |
| 	const outdatedChunkGroupInfo = new Set();
 | |
| 	/** @type {Set<ChunkGroupInfo>} */
 | |
| 	const chunkGroupsForMerging = new Set();
 | |
| 	/** @type {QueueItem[]} */
 | |
| 	let queueDelayed = [];
 | |
| 
 | |
| 	/** @type {[Module, ModuleGraphConnection[]][]} */
 | |
| 	const skipConnectionBuffer = [];
 | |
| 	/** @type {Module[]} */
 | |
| 	const skipBuffer = [];
 | |
| 	/** @type {QueueItem[]} */
 | |
| 	const queueBuffer = [];
 | |
| 
 | |
| 	/** @type {Module} */
 | |
| 	let module;
 | |
| 	/** @type {Chunk} */
 | |
| 	let chunk;
 | |
| 	/** @type {ChunkGroup} */
 | |
| 	let chunkGroup;
 | |
| 	/** @type {DependenciesBlock} */
 | |
| 	let block;
 | |
| 	/** @type {ChunkGroupInfo} */
 | |
| 	let chunkGroupInfo;
 | |
| 
 | |
| 	// For each async Block in graph
 | |
| 	/**
 | |
| 	 * @param {AsyncDependenciesBlock} b iterating over each Async DepBlock
 | |
| 	 * @returns {void}
 | |
| 	 */
 | |
| 	const iteratorBlock = b => {
 | |
| 		// 1. We create a chunk group with single chunk in it for this Block
 | |
| 		// but only once (blockChunkGroups map)
 | |
| 		/** @type {ChunkGroupInfo | undefined} */
 | |
| 		let cgi = blockChunkGroups.get(b);
 | |
| 		/** @type {ChunkGroup | undefined} */
 | |
| 		let c;
 | |
| 		/** @type {Entrypoint | undefined} */
 | |
| 		let entrypoint;
 | |
| 		const entryOptions = b.groupOptions && b.groupOptions.entryOptions;
 | |
| 		if (cgi === undefined) {
 | |
| 			const chunkName = (b.groupOptions && b.groupOptions.name) || b.chunkName;
 | |
| 			if (entryOptions) {
 | |
| 				cgi = namedAsyncEntrypoints.get(/** @type {string} */ (chunkName));
 | |
| 				if (!cgi) {
 | |
| 					entrypoint = compilation.addAsyncEntrypoint(
 | |
| 						entryOptions,
 | |
| 						module,
 | |
| 						/** @type {DependencyLocation} */ (b.loc),
 | |
| 						/** @type {string} */ (b.request)
 | |
| 					);
 | |
| 					maskByChunk.set(entrypoint.chunks[0], ZERO_BIGINT);
 | |
| 					entrypoint.index = nextChunkGroupIndex++;
 | |
| 					cgi = {
 | |
| 						chunkGroup: entrypoint,
 | |
| 						runtime: entrypoint.options.runtime || entrypoint.name,
 | |
| 						minAvailableModules: ZERO_BIGINT,
 | |
| 						availableModulesToBeMerged: [],
 | |
| 						skippedItems: undefined,
 | |
| 						resultingAvailableModules: undefined,
 | |
| 						children: undefined,
 | |
| 						availableSources: undefined,
 | |
| 						availableChildren: undefined,
 | |
| 						preOrderIndex: 0,
 | |
| 						postOrderIndex: 0,
 | |
| 						chunkLoading:
 | |
| 							entryOptions.chunkLoading !== undefined
 | |
| 								? entryOptions.chunkLoading !== false
 | |
| 								: chunkGroupInfo.chunkLoading,
 | |
| 						asyncChunks:
 | |
| 							entryOptions.asyncChunks !== undefined
 | |
| 								? entryOptions.asyncChunks
 | |
| 								: chunkGroupInfo.asyncChunks
 | |
| 					};
 | |
| 					chunkGroupInfoMap.set(entrypoint, cgi);
 | |
| 
 | |
| 					chunkGraph.connectBlockAndChunkGroup(b, entrypoint);
 | |
| 					if (chunkName) {
 | |
| 						namedAsyncEntrypoints.set(chunkName, cgi);
 | |
| 					}
 | |
| 				} else {
 | |
| 					entrypoint = /** @type {Entrypoint} */ (cgi.chunkGroup);
 | |
| 					// TODO merge entryOptions
 | |
| 					entrypoint.addOrigin(
 | |
| 						module,
 | |
| 						/** @type {DependencyLocation} */ (b.loc),
 | |
| 						/** @type {string} */ (b.request)
 | |
| 					);
 | |
| 					chunkGraph.connectBlockAndChunkGroup(b, entrypoint);
 | |
| 				}
 | |
| 
 | |
| 				// 2. We enqueue the DependenciesBlock for traversal
 | |
| 				queueDelayed.push({
 | |
| 					action: PROCESS_ENTRY_BLOCK,
 | |
| 					block: b,
 | |
| 					module,
 | |
| 					chunk: entrypoint.chunks[0],
 | |
| 					chunkGroup: entrypoint,
 | |
| 					chunkGroupInfo: cgi
 | |
| 				});
 | |
| 			} else if (!chunkGroupInfo.asyncChunks || !chunkGroupInfo.chunkLoading) {
 | |
| 				// Just queue the block into the current chunk group
 | |
| 				queue.push({
 | |
| 					action: PROCESS_BLOCK,
 | |
| 					block: b,
 | |
| 					module,
 | |
| 					chunk,
 | |
| 					chunkGroup,
 | |
| 					chunkGroupInfo
 | |
| 				});
 | |
| 			} else {
 | |
| 				cgi = chunkName ? namedChunkGroups.get(chunkName) : undefined;
 | |
| 				if (!cgi) {
 | |
| 					c = compilation.addChunkInGroup(
 | |
| 						b.groupOptions || b.chunkName,
 | |
| 						module,
 | |
| 						/** @type {DependencyLocation} */ (b.loc),
 | |
| 						/** @type {string} */ (b.request)
 | |
| 					);
 | |
| 					maskByChunk.set(c.chunks[0], ZERO_BIGINT);
 | |
| 					c.index = nextChunkGroupIndex++;
 | |
| 					cgi = {
 | |
| 						chunkGroup: c,
 | |
| 						runtime: chunkGroupInfo.runtime,
 | |
| 						minAvailableModules: undefined,
 | |
| 						availableModulesToBeMerged: [],
 | |
| 						skippedItems: undefined,
 | |
| 						resultingAvailableModules: undefined,
 | |
| 						children: undefined,
 | |
| 						availableSources: undefined,
 | |
| 						availableChildren: undefined,
 | |
| 						preOrderIndex: 0,
 | |
| 						postOrderIndex: 0,
 | |
| 						chunkLoading: chunkGroupInfo.chunkLoading,
 | |
| 						asyncChunks: chunkGroupInfo.asyncChunks
 | |
| 					};
 | |
| 					allCreatedChunkGroups.add(c);
 | |
| 					chunkGroupInfoMap.set(c, cgi);
 | |
| 					if (chunkName) {
 | |
| 						namedChunkGroups.set(chunkName, cgi);
 | |
| 					}
 | |
| 				} else {
 | |
| 					c = cgi.chunkGroup;
 | |
| 					if (c.isInitial()) {
 | |
| 						compilation.errors.push(
 | |
| 							new AsyncDependencyToInitialChunkError(
 | |
| 								/** @type {string} */ (chunkName),
 | |
| 								module,
 | |
| 								/** @type {DependencyLocation} */ (b.loc)
 | |
| 							)
 | |
| 						);
 | |
| 						c = chunkGroup;
 | |
| 					} else {
 | |
| 						c.addOptions(b.groupOptions);
 | |
| 					}
 | |
| 					c.addOrigin(
 | |
| 						module,
 | |
| 						/** @type {DependencyLocation} */ (b.loc),
 | |
| 						/** @type {string} */ (b.request)
 | |
| 					);
 | |
| 				}
 | |
| 				blockConnections.set(b, []);
 | |
| 			}
 | |
| 			blockChunkGroups.set(b, /** @type {ChunkGroupInfo} */ (cgi));
 | |
| 			blockByChunkGroups.set(/** @type {ChunkGroupInfo} */ (cgi), b);
 | |
| 		} else if (entryOptions) {
 | |
| 			entrypoint = /** @type {Entrypoint} */ (cgi.chunkGroup);
 | |
| 		} else {
 | |
| 			c = cgi.chunkGroup;
 | |
| 		}
 | |
| 
 | |
| 		if (c !== undefined) {
 | |
| 			// 2. We store the connection for the block
 | |
| 			// to connect it later if needed
 | |
| 			/** @type {BlockChunkGroupConnection[]} */
 | |
| 			(blockConnections.get(b)).push({
 | |
| 				originChunkGroupInfo: chunkGroupInfo,
 | |
| 				chunkGroup: c
 | |
| 			});
 | |
| 
 | |
| 			// 3. We enqueue the chunk group info creation/updating
 | |
| 			let connectList = queueConnect.get(chunkGroupInfo);
 | |
| 			if (connectList === undefined) {
 | |
| 				connectList = new Set();
 | |
| 				queueConnect.set(chunkGroupInfo, connectList);
 | |
| 			}
 | |
| 			connectList.add(/** @type {ChunkGroupInfo} */ (cgi));
 | |
| 
 | |
| 			// TODO check if this really need to be done for each traversal
 | |
| 			// or if it is enough when it's queued when created
 | |
| 			// 4. We enqueue the DependenciesBlock for traversal
 | |
| 			queueDelayed.push({
 | |
| 				action: PROCESS_BLOCK,
 | |
| 				block: b,
 | |
| 				module,
 | |
| 				chunk: c.chunks[0],
 | |
| 				chunkGroup: c,
 | |
| 				chunkGroupInfo: /** @type {ChunkGroupInfo} */ (cgi)
 | |
| 			});
 | |
| 		} else if (entrypoint !== undefined) {
 | |
| 			chunkGroupInfo.chunkGroup.addAsyncEntrypoint(entrypoint);
 | |
| 		}
 | |
| 	};
 | |
| 
 | |
| 	/**
 | |
| 	 * @param {DependenciesBlock} block the block
 | |
| 	 * @returns {void}
 | |
| 	 */
 | |
| 	const processBlock = block => {
 | |
| 		statProcessedBlocks++;
 | |
| 		// get prepared block info
 | |
| 		const blockModules = getBlockModules(block, chunkGroupInfo.runtime);
 | |
| 
 | |
| 		if (blockModules !== undefined) {
 | |
| 			const minAvailableModules =
 | |
| 				/** @type {bigint} */
 | |
| 				(chunkGroupInfo.minAvailableModules);
 | |
| 			// Buffer items because order need to be reversed to get indices correct
 | |
| 			// Traverse all referenced modules
 | |
| 			for (let i = 0, len = blockModules.length; i < len; i += 3) {
 | |
| 				const refModule = /** @type {Module} */ (blockModules[i]);
 | |
| 				// For single comparisons this might be cheaper
 | |
| 				const isModuleInChunk = chunkGraph.isModuleInChunk(refModule, chunk);
 | |
| 
 | |
| 				if (isModuleInChunk) {
 | |
| 					// skip early if already connected
 | |
| 					continue;
 | |
| 				}
 | |
| 
 | |
| 				const refOrdinal = /** @type {number} */ getModuleOrdinal(refModule);
 | |
| 				const activeState = /** @type {ConnectionState} */ (
 | |
| 					blockModules[i + 1]
 | |
| 				);
 | |
| 				if (activeState !== true) {
 | |
| 					const connections = /** @type {ModuleGraphConnection[]} */ (
 | |
| 						blockModules[i + 2]
 | |
| 					);
 | |
| 					skipConnectionBuffer.push([refModule, connections]);
 | |
| 					// We skip inactive connections
 | |
| 					if (activeState === false) continue;
 | |
| 				} else if (isOrdinalSetInMask(minAvailableModules, refOrdinal)) {
 | |
| 					// already in parent chunks, skip it for now
 | |
| 					skipBuffer.push(refModule);
 | |
| 					continue;
 | |
| 				}
 | |
| 				// enqueue, then add and enter to be in the correct order
 | |
| 				// this is relevant with circular dependencies
 | |
| 				queueBuffer.push({
 | |
| 					action: activeState === true ? ADD_AND_ENTER_MODULE : PROCESS_BLOCK,
 | |
| 					block: refModule,
 | |
| 					module: refModule,
 | |
| 					chunk,
 | |
| 					chunkGroup,
 | |
| 					chunkGroupInfo
 | |
| 				});
 | |
| 			}
 | |
| 			// Add buffered items in reverse order
 | |
| 			if (skipConnectionBuffer.length > 0) {
 | |
| 				let { skippedModuleConnections } = chunkGroupInfo;
 | |
| 				if (skippedModuleConnections === undefined) {
 | |
| 					chunkGroupInfo.skippedModuleConnections = skippedModuleConnections =
 | |
| 						new Set();
 | |
| 				}
 | |
| 				for (let i = skipConnectionBuffer.length - 1; i >= 0; i--) {
 | |
| 					skippedModuleConnections.add(skipConnectionBuffer[i]);
 | |
| 				}
 | |
| 				skipConnectionBuffer.length = 0;
 | |
| 			}
 | |
| 			if (skipBuffer.length > 0) {
 | |
| 				let { skippedItems } = chunkGroupInfo;
 | |
| 				if (skippedItems === undefined) {
 | |
| 					chunkGroupInfo.skippedItems = skippedItems = new Set();
 | |
| 				}
 | |
| 				for (let i = skipBuffer.length - 1; i >= 0; i--) {
 | |
| 					skippedItems.add(skipBuffer[i]);
 | |
| 				}
 | |
| 				skipBuffer.length = 0;
 | |
| 			}
 | |
| 			if (queueBuffer.length > 0) {
 | |
| 				for (let i = queueBuffer.length - 1; i >= 0; i--) {
 | |
| 					queue.push(queueBuffer[i]);
 | |
| 				}
 | |
| 				queueBuffer.length = 0;
 | |
| 			}
 | |
| 		}
 | |
| 
 | |
| 		// Traverse all Blocks
 | |
| 		for (const b of block.blocks) {
 | |
| 			iteratorBlock(b);
 | |
| 		}
 | |
| 
 | |
| 		if (block.blocks.length > 0 && module !== block) {
 | |
| 			blocksWithNestedBlocks.add(block);
 | |
| 		}
 | |
| 	};
 | |
| 
 | |
| 	/**
 | |
| 	 * @param {DependenciesBlock} block the block
 | |
| 	 * @returns {void}
 | |
| 	 */
 | |
| 	const processEntryBlock = block => {
 | |
| 		statProcessedBlocks++;
 | |
| 		// get prepared block info
 | |
| 		const blockModules = getBlockModules(block, chunkGroupInfo.runtime);
 | |
| 
 | |
| 		if (blockModules !== undefined) {
 | |
| 			// Traverse all referenced modules in reverse order
 | |
| 			for (let i = blockModules.length - 3; i >= 0; i -= 3) {
 | |
| 				const refModule = /** @type {Module} */ (blockModules[i]);
 | |
| 				const activeState = /** @type {ConnectionState} */ (
 | |
| 					blockModules[i + 1]
 | |
| 				);
 | |
| 				// enqueue, then add and enter to be in the correct order
 | |
| 				// this is relevant with circular dependencies
 | |
| 				queue.push({
 | |
| 					action:
 | |
| 						activeState === true ? ADD_AND_ENTER_ENTRY_MODULE : PROCESS_BLOCK,
 | |
| 					block: refModule,
 | |
| 					module: refModule,
 | |
| 					chunk,
 | |
| 					chunkGroup,
 | |
| 					chunkGroupInfo
 | |
| 				});
 | |
| 			}
 | |
| 		}
 | |
| 
 | |
| 		// Traverse all Blocks
 | |
| 		for (const b of block.blocks) {
 | |
| 			iteratorBlock(b);
 | |
| 		}
 | |
| 
 | |
| 		if (block.blocks.length > 0 && module !== block) {
 | |
| 			blocksWithNestedBlocks.add(block);
 | |
| 		}
 | |
| 	};
 | |
| 
 | |
| 	const processQueue = () => {
 | |
| 		while (queue.length) {
 | |
| 			statProcessedQueueItems++;
 | |
| 			const queueItem = /** @type {QueueItem} */ (queue.pop());
 | |
| 			module = queueItem.module;
 | |
| 			block = queueItem.block;
 | |
| 			chunk = queueItem.chunk;
 | |
| 			chunkGroup = queueItem.chunkGroup;
 | |
| 			chunkGroupInfo = queueItem.chunkGroupInfo;
 | |
| 
 | |
| 			switch (queueItem.action) {
 | |
| 				case ADD_AND_ENTER_ENTRY_MODULE:
 | |
| 					chunkGraph.connectChunkAndEntryModule(
 | |
| 						chunk,
 | |
| 						module,
 | |
| 						/** @type {Entrypoint} */ (chunkGroup)
 | |
| 					);
 | |
| 				// fallthrough
 | |
| 				case ADD_AND_ENTER_MODULE: {
 | |
| 					const isModuleInChunk = chunkGraph.isModuleInChunk(module, chunk);
 | |
| 
 | |
| 					if (isModuleInChunk) {
 | |
| 						// already connected, skip it
 | |
| 						break;
 | |
| 					}
 | |
| 					// We connect Module and Chunk
 | |
| 					chunkGraph.connectChunkAndModule(chunk, module);
 | |
| 					const moduleOrdinal = getModuleOrdinal(module);
 | |
| 					let chunkMask = /** @type {bigint} */ (maskByChunk.get(chunk));
 | |
| 					chunkMask |= ONE_BIGINT << BigInt(moduleOrdinal);
 | |
| 					maskByChunk.set(chunk, chunkMask);
 | |
| 				}
 | |
| 				// fallthrough
 | |
| 				case ENTER_MODULE: {
 | |
| 					const index = chunkGroup.getModulePreOrderIndex(module);
 | |
| 					if (index === undefined) {
 | |
| 						chunkGroup.setModulePreOrderIndex(
 | |
| 							module,
 | |
| 							chunkGroupInfo.preOrderIndex++
 | |
| 						);
 | |
| 					}
 | |
| 
 | |
| 					if (
 | |
| 						moduleGraph.setPreOrderIndexIfUnset(
 | |
| 							module,
 | |
| 							nextFreeModulePreOrderIndex
 | |
| 						)
 | |
| 					) {
 | |
| 						nextFreeModulePreOrderIndex++;
 | |
| 					}
 | |
| 
 | |
| 					// reuse queueItem
 | |
| 					queueItem.action = LEAVE_MODULE;
 | |
| 					queue.push(queueItem);
 | |
| 				}
 | |
| 				// fallthrough
 | |
| 				case PROCESS_BLOCK: {
 | |
| 					processBlock(block);
 | |
| 					break;
 | |
| 				}
 | |
| 				case PROCESS_ENTRY_BLOCK: {
 | |
| 					processEntryBlock(block);
 | |
| 					break;
 | |
| 				}
 | |
| 				case LEAVE_MODULE: {
 | |
| 					const index = chunkGroup.getModulePostOrderIndex(module);
 | |
| 					if (index === undefined) {
 | |
| 						chunkGroup.setModulePostOrderIndex(
 | |
| 							module,
 | |
| 							chunkGroupInfo.postOrderIndex++
 | |
| 						);
 | |
| 					}
 | |
| 
 | |
| 					if (
 | |
| 						moduleGraph.setPostOrderIndexIfUnset(
 | |
| 							module,
 | |
| 							nextFreeModulePostOrderIndex
 | |
| 						)
 | |
| 					) {
 | |
| 						nextFreeModulePostOrderIndex++;
 | |
| 					}
 | |
| 					break;
 | |
| 				}
 | |
| 			}
 | |
| 		}
 | |
| 	};
 | |
| 
 | |
| 	/**
 | |
| 	 * @param {ChunkGroupInfo} chunkGroupInfo The info object for the chunk group
 | |
| 	 * @returns {bigint} The mask of available modules after the chunk group
 | |
| 	 */
 | |
| 	const calculateResultingAvailableModules = chunkGroupInfo => {
 | |
| 		if (chunkGroupInfo.resultingAvailableModules !== undefined)
 | |
| 			return chunkGroupInfo.resultingAvailableModules;
 | |
| 
 | |
| 		let resultingAvailableModules = /** @type {bigint} */ (
 | |
| 			chunkGroupInfo.minAvailableModules
 | |
| 		);
 | |
| 
 | |
| 		// add the modules from the chunk group to the set
 | |
| 		for (const chunk of chunkGroupInfo.chunkGroup.chunks) {
 | |
| 			const mask = /** @type {bigint} */ (maskByChunk.get(chunk));
 | |
| 			resultingAvailableModules |= mask;
 | |
| 		}
 | |
| 
 | |
| 		return (chunkGroupInfo.resultingAvailableModules =
 | |
| 			resultingAvailableModules);
 | |
| 	};
 | |
| 
 | |
| 	const processConnectQueue = () => {
 | |
| 		// Figure out new parents for chunk groups
 | |
| 		// to get new available modules for these children
 | |
| 		for (const [chunkGroupInfo, targets] of queueConnect) {
 | |
| 			// 1. Add new targets to the list of children
 | |
| 			if (chunkGroupInfo.children === undefined) {
 | |
| 				chunkGroupInfo.children = targets;
 | |
| 			} else {
 | |
| 				for (const target of targets) {
 | |
| 					chunkGroupInfo.children.add(target);
 | |
| 				}
 | |
| 			}
 | |
| 
 | |
| 			// 2. Calculate resulting available modules
 | |
| 			const resultingAvailableModules =
 | |
| 				calculateResultingAvailableModules(chunkGroupInfo);
 | |
| 
 | |
| 			const runtime = chunkGroupInfo.runtime;
 | |
| 
 | |
| 			// 3. Update chunk group info
 | |
| 			for (const target of targets) {
 | |
| 				target.availableModulesToBeMerged.push(resultingAvailableModules);
 | |
| 				chunkGroupsForMerging.add(target);
 | |
| 				const oldRuntime = target.runtime;
 | |
| 				const newRuntime = mergeRuntime(oldRuntime, runtime);
 | |
| 				if (oldRuntime !== newRuntime) {
 | |
| 					target.runtime = newRuntime;
 | |
| 					outdatedChunkGroupInfo.add(target);
 | |
| 				}
 | |
| 			}
 | |
| 
 | |
| 			statConnectedChunkGroups += targets.size;
 | |
| 		}
 | |
| 		queueConnect.clear();
 | |
| 	};
 | |
| 
 | |
| 	const processChunkGroupsForMerging = () => {
 | |
| 		statProcessedChunkGroupsForMerging += chunkGroupsForMerging.size;
 | |
| 
 | |
| 		// Execute the merge
 | |
| 		for (const info of chunkGroupsForMerging) {
 | |
| 			const availableModulesToBeMerged = info.availableModulesToBeMerged;
 | |
| 			const cachedMinAvailableModules = info.minAvailableModules;
 | |
| 			let minAvailableModules = cachedMinAvailableModules;
 | |
| 
 | |
| 			statMergedAvailableModuleSets += availableModulesToBeMerged.length;
 | |
| 
 | |
| 			for (const availableModules of availableModulesToBeMerged) {
 | |
| 				if (minAvailableModules === undefined) {
 | |
| 					minAvailableModules = availableModules;
 | |
| 				} else {
 | |
| 					minAvailableModules &= availableModules;
 | |
| 				}
 | |
| 			}
 | |
| 
 | |
| 			const changed = minAvailableModules !== cachedMinAvailableModules;
 | |
| 
 | |
| 			availableModulesToBeMerged.length = 0;
 | |
| 			if (changed) {
 | |
| 				info.minAvailableModules = minAvailableModules;
 | |
| 				info.resultingAvailableModules = undefined;
 | |
| 				outdatedChunkGroupInfo.add(info);
 | |
| 			}
 | |
| 		}
 | |
| 		chunkGroupsForMerging.clear();
 | |
| 	};
 | |
| 
 | |
| 	const processChunkGroupsForCombining = () => {
 | |
| 		for (const info of chunkGroupsForCombining) {
 | |
| 			for (const source of /** @type {Set<ChunkGroupInfo>} */ (
 | |
| 				info.availableSources
 | |
| 			)) {
 | |
| 				if (source.minAvailableModules === undefined) {
 | |
| 					chunkGroupsForCombining.delete(info);
 | |
| 					break;
 | |
| 				}
 | |
| 			}
 | |
| 		}
 | |
| 
 | |
| 		for (const info of chunkGroupsForCombining) {
 | |
| 			let availableModules = ZERO_BIGINT;
 | |
| 			// combine minAvailableModules from all resultingAvailableModules
 | |
| 			for (const source of /** @type {Set<ChunkGroupInfo>} */ (
 | |
| 				info.availableSources
 | |
| 			)) {
 | |
| 				const resultingAvailableModules =
 | |
| 					calculateResultingAvailableModules(source);
 | |
| 				availableModules |= resultingAvailableModules;
 | |
| 			}
 | |
| 			info.minAvailableModules = availableModules;
 | |
| 			info.resultingAvailableModules = undefined;
 | |
| 			outdatedChunkGroupInfo.add(info);
 | |
| 		}
 | |
| 		chunkGroupsForCombining.clear();
 | |
| 	};
 | |
| 
 | |
| 	const processOutdatedChunkGroupInfo = () => {
 | |
| 		statChunkGroupInfoUpdated += outdatedChunkGroupInfo.size;
 | |
| 		// Revisit skipped elements
 | |
| 		for (const info of outdatedChunkGroupInfo) {
 | |
| 			// 1. Reconsider skipped items
 | |
| 			if (info.skippedItems !== undefined) {
 | |
| 				const minAvailableModules =
 | |
| 					/** @type {bigint} */
 | |
| 					(info.minAvailableModules);
 | |
| 				for (const module of info.skippedItems) {
 | |
| 					const ordinal = getModuleOrdinal(module);
 | |
| 					if (!isOrdinalSetInMask(minAvailableModules, ordinal)) {
 | |
| 						queue.push({
 | |
| 							action: ADD_AND_ENTER_MODULE,
 | |
| 							block: module,
 | |
| 							module,
 | |
| 							chunk: info.chunkGroup.chunks[0],
 | |
| 							chunkGroup: info.chunkGroup,
 | |
| 							chunkGroupInfo: info
 | |
| 						});
 | |
| 						info.skippedItems.delete(module);
 | |
| 					}
 | |
| 				}
 | |
| 			}
 | |
| 
 | |
| 			// 2. Reconsider skipped connections
 | |
| 			if (info.skippedModuleConnections !== undefined) {
 | |
| 				const minAvailableModules =
 | |
| 					/** @type {bigint} */
 | |
| 					(info.minAvailableModules);
 | |
| 				for (const entry of info.skippedModuleConnections) {
 | |
| 					const [module, connections] = entry;
 | |
| 					const activeState = getActiveStateOfConnections(
 | |
| 						connections,
 | |
| 						info.runtime
 | |
| 					);
 | |
| 					if (activeState === false) continue;
 | |
| 					if (activeState === true) {
 | |
| 						const ordinal = getModuleOrdinal(module);
 | |
| 						info.skippedModuleConnections.delete(entry);
 | |
| 						if (isOrdinalSetInMask(minAvailableModules, ordinal)) {
 | |
| 							/** @type {NonNullable<ChunkGroupInfo["skippedItems"]>} */
 | |
| 							(info.skippedItems).add(module);
 | |
| 							continue;
 | |
| 						}
 | |
| 					}
 | |
| 					queue.push({
 | |
| 						action: activeState === true ? ADD_AND_ENTER_MODULE : PROCESS_BLOCK,
 | |
| 						block: module,
 | |
| 						module,
 | |
| 						chunk: info.chunkGroup.chunks[0],
 | |
| 						chunkGroup: info.chunkGroup,
 | |
| 						chunkGroupInfo: info
 | |
| 					});
 | |
| 				}
 | |
| 			}
 | |
| 
 | |
| 			// 2. Reconsider children chunk groups
 | |
| 			if (info.children !== undefined) {
 | |
| 				statChildChunkGroupsReconnected += info.children.size;
 | |
| 				for (const cgi of info.children) {
 | |
| 					let connectList = queueConnect.get(info);
 | |
| 					if (connectList === undefined) {
 | |
| 						connectList = new Set();
 | |
| 						queueConnect.set(info, connectList);
 | |
| 					}
 | |
| 					connectList.add(cgi);
 | |
| 				}
 | |
| 			}
 | |
| 
 | |
| 			// 3. Reconsider chunk groups for combining
 | |
| 			if (info.availableChildren !== undefined) {
 | |
| 				for (const cgi of info.availableChildren) {
 | |
| 					chunkGroupsForCombining.add(cgi);
 | |
| 				}
 | |
| 			}
 | |
| 			outdatedOrderIndexChunkGroups.add(info);
 | |
| 		}
 | |
| 		outdatedChunkGroupInfo.clear();
 | |
| 	};
 | |
| 
 | |
| 	// Iterative traversal of the Module graph
 | |
| 	// Recursive would be simpler to write but could result in Stack Overflows
 | |
| 	while (queue.length || queueConnect.size) {
 | |
| 		logger.time("visitModules: visiting");
 | |
| 		processQueue();
 | |
| 		logger.timeAggregateEnd("visitModules: prepare");
 | |
| 		logger.timeEnd("visitModules: visiting");
 | |
| 
 | |
| 		if (chunkGroupsForCombining.size > 0) {
 | |
| 			logger.time("visitModules: combine available modules");
 | |
| 			processChunkGroupsForCombining();
 | |
| 			logger.timeEnd("visitModules: combine available modules");
 | |
| 		}
 | |
| 
 | |
| 		if (queueConnect.size > 0) {
 | |
| 			logger.time("visitModules: calculating available modules");
 | |
| 			processConnectQueue();
 | |
| 			logger.timeEnd("visitModules: calculating available modules");
 | |
| 
 | |
| 			if (chunkGroupsForMerging.size > 0) {
 | |
| 				logger.time("visitModules: merging available modules");
 | |
| 				processChunkGroupsForMerging();
 | |
| 				logger.timeEnd("visitModules: merging available modules");
 | |
| 			}
 | |
| 		}
 | |
| 
 | |
| 		if (outdatedChunkGroupInfo.size > 0) {
 | |
| 			logger.time("visitModules: check modules for revisit");
 | |
| 			processOutdatedChunkGroupInfo();
 | |
| 			logger.timeEnd("visitModules: check modules for revisit");
 | |
| 		}
 | |
| 
 | |
| 		// Run queueDelayed when all items of the queue are processed
 | |
| 		// This is important to get the global indexing correct
 | |
| 		// Async blocks should be processed after all sync blocks are processed
 | |
| 		if (queue.length === 0) {
 | |
| 			const tempQueue = queue;
 | |
| 			queue = queueDelayed.reverse();
 | |
| 			queueDelayed = tempQueue;
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	for (const info of outdatedOrderIndexChunkGroups) {
 | |
| 		const { chunkGroup, runtime } = info;
 | |
| 
 | |
| 		const block = blockByChunkGroups.get(info);
 | |
| 
 | |
| 		if (!block) {
 | |
| 			continue;
 | |
| 		}
 | |
| 
 | |
| 		let preOrderIndex = 0;
 | |
| 		let postOrderIndex = 0;
 | |
| 
 | |
| 		/**
 | |
| 		 * @param {DependenciesBlock} current current
 | |
| 		 * @param {BlocksWithNestedBlocks} visited visited dependencies blocks
 | |
| 		 */
 | |
| 		const process = (current, visited) => {
 | |
| 			const blockModules = getBlockModules(current, runtime);
 | |
| 			if (blockModules === undefined) {
 | |
| 				return;
 | |
| 			}
 | |
| 
 | |
| 			for (let i = 0, len = blockModules.length; i < len; i += 3) {
 | |
| 				const activeState = /** @type {ConnectionState} */ (
 | |
| 					blockModules[i + 1]
 | |
| 				);
 | |
| 				if (activeState === false) {
 | |
| 					continue;
 | |
| 				}
 | |
| 				const refModule = /** @type {Module} */ (blockModules[i]);
 | |
| 				if (visited.has(refModule)) {
 | |
| 					continue;
 | |
| 				}
 | |
| 
 | |
| 				visited.add(refModule);
 | |
| 
 | |
| 				if (refModule) {
 | |
| 					chunkGroup.setModulePreOrderIndex(refModule, preOrderIndex++);
 | |
| 					process(refModule, visited);
 | |
| 					chunkGroup.setModulePostOrderIndex(refModule, postOrderIndex++);
 | |
| 				}
 | |
| 			}
 | |
| 		};
 | |
| 
 | |
| 		process(block, new Set());
 | |
| 	}
 | |
| 	outdatedOrderIndexChunkGroups.clear();
 | |
| 	ordinalByModule.clear();
 | |
| 
 | |
| 	logger.log(
 | |
| 		`${statProcessedQueueItems} queue items processed (${statProcessedBlocks} blocks)`
 | |
| 	);
 | |
| 	logger.log(`${statConnectedChunkGroups} chunk groups connected`);
 | |
| 	logger.log(
 | |
| 		`${statProcessedChunkGroupsForMerging} chunk groups processed for merging (${statMergedAvailableModuleSets} module sets, ${statForkedAvailableModules} forked, ${statForkedAvailableModulesCount} + ${statForkedAvailableModulesCountPlus} modules forked, ${statForkedMergedModulesCount} + ${statForkedMergedModulesCountPlus} modules merged into fork, ${statForkedResultModulesCount} resulting modules)`
 | |
| 	);
 | |
| 	logger.log(
 | |
| 		`${statChunkGroupInfoUpdated} chunk group info updated (${statChildChunkGroupsReconnected} already connected chunk groups reconnected)`
 | |
| 	);
 | |
| };
 | |
| 
 | |
| /**
 | |
|  *
 | |
|  * @param {Compilation} compilation the compilation
 | |
|  * @param {BlocksWithNestedBlocks} blocksWithNestedBlocks flag for blocks that have nested blocks
 | |
|  * @param {BlockConnections} blockConnections connection for blocks
 | |
|  * @param {MaskByChunk} maskByChunk mapping from chunk to module mask
 | |
|  */
 | |
| const connectChunkGroups = (
 | |
| 	compilation,
 | |
| 	blocksWithNestedBlocks,
 | |
| 	blockConnections,
 | |
| 	maskByChunk
 | |
| ) => {
 | |
| 	const { chunkGraph } = compilation;
 | |
| 
 | |
| 	/**
 | |
| 	 * Helper function to check if all modules of a chunk are available
 | |
| 	 *
 | |
| 	 * @param {ChunkGroup} chunkGroup the chunkGroup to scan
 | |
| 	 * @param {bigint} availableModules the comparator set
 | |
| 	 * @returns {boolean} return true if all modules of a chunk are available
 | |
| 	 */
 | |
| 	const areModulesAvailable = (chunkGroup, availableModules) => {
 | |
| 		for (const chunk of chunkGroup.chunks) {
 | |
| 			const chunkMask = /** @type {bigint} */ (maskByChunk.get(chunk));
 | |
| 			if ((chunkMask & availableModules) !== chunkMask) return false;
 | |
| 		}
 | |
| 		return true;
 | |
| 	};
 | |
| 
 | |
| 	// For each edge in the basic chunk graph
 | |
| 	for (const [block, connections] of blockConnections) {
 | |
| 		// 1. Check if connection is needed
 | |
| 		// When none of the dependencies need to be connected
 | |
| 		// we can skip all of them
 | |
| 		// It's not possible to filter each item so it doesn't create inconsistent
 | |
| 		// connections and modules can only create one version
 | |
| 		// TODO maybe decide this per runtime
 | |
| 		if (
 | |
| 			// TODO is this needed?
 | |
| 			!blocksWithNestedBlocks.has(block) &&
 | |
| 			connections.every(({ chunkGroup, originChunkGroupInfo }) =>
 | |
| 				areModulesAvailable(
 | |
| 					chunkGroup,
 | |
| 					/** @type {bigint} */ (originChunkGroupInfo.resultingAvailableModules)
 | |
| 				)
 | |
| 			)
 | |
| 		) {
 | |
| 			continue;
 | |
| 		}
 | |
| 
 | |
| 		// 2. Foreach edge
 | |
| 		for (let i = 0; i < connections.length; i++) {
 | |
| 			const { chunkGroup, originChunkGroupInfo } = connections[i];
 | |
| 
 | |
| 			// 3. Connect block with chunk
 | |
| 			chunkGraph.connectBlockAndChunkGroup(block, chunkGroup);
 | |
| 
 | |
| 			// 4. Connect chunk with parent
 | |
| 			connectChunkGroupParentAndChild(
 | |
| 				originChunkGroupInfo.chunkGroup,
 | |
| 				chunkGroup
 | |
| 			);
 | |
| 		}
 | |
| 	}
 | |
| };
 | |
| 
 | |
| /**
 | |
|  * Remove all unconnected chunk groups
 | |
|  * @param {Compilation} compilation the compilation
 | |
|  * @param {Iterable<ChunkGroup>} allCreatedChunkGroups all chunk groups that where created before
 | |
|  */
 | |
| const cleanupUnconnectedGroups = (compilation, allCreatedChunkGroups) => {
 | |
| 	const { chunkGraph } = compilation;
 | |
| 
 | |
| 	for (const chunkGroup of allCreatedChunkGroups) {
 | |
| 		if (chunkGroup.getNumberOfParents() === 0) {
 | |
| 			for (const chunk of chunkGroup.chunks) {
 | |
| 				compilation.chunks.delete(chunk);
 | |
| 				chunkGraph.disconnectChunk(chunk);
 | |
| 			}
 | |
| 			chunkGraph.disconnectChunkGroup(chunkGroup);
 | |
| 			chunkGroup.remove();
 | |
| 		}
 | |
| 	}
 | |
| };
 | |
| 
 | |
| /**
 | |
|  * This method creates the Chunk graph from the Module graph
 | |
|  * @param {Compilation} compilation the compilation
 | |
|  * @param {InputEntrypointsAndModules} inputEntrypointsAndModules chunk groups which are processed with the modules
 | |
|  * @returns {void}
 | |
|  */
 | |
| const buildChunkGraph = (compilation, inputEntrypointsAndModules) => {
 | |
| 	const logger = compilation.getLogger("webpack.buildChunkGraph");
 | |
| 
 | |
| 	// SHARED STATE
 | |
| 
 | |
| 	/** @type {BlockConnections} */
 | |
| 	const blockConnections = new Map();
 | |
| 
 | |
| 	/** @type {AllCreatedChunkGroups} */
 | |
| 	const allCreatedChunkGroups = new Set();
 | |
| 
 | |
| 	/** @type {ChunkGroupInfoMap} */
 | |
| 	const chunkGroupInfoMap = new Map();
 | |
| 
 | |
| 	/** @type {BlocksWithNestedBlocks} */
 | |
| 	const blocksWithNestedBlocks = new Set();
 | |
| 
 | |
| 	/** @type {MaskByChunk} */
 | |
| 	const maskByChunk = new Map();
 | |
| 
 | |
| 	// PART ONE
 | |
| 
 | |
| 	logger.time("visitModules");
 | |
| 	visitModules(
 | |
| 		logger,
 | |
| 		compilation,
 | |
| 		inputEntrypointsAndModules,
 | |
| 		chunkGroupInfoMap,
 | |
| 		blockConnections,
 | |
| 		blocksWithNestedBlocks,
 | |
| 		allCreatedChunkGroups,
 | |
| 		maskByChunk
 | |
| 	);
 | |
| 	logger.timeEnd("visitModules");
 | |
| 
 | |
| 	// PART TWO
 | |
| 
 | |
| 	logger.time("connectChunkGroups");
 | |
| 	connectChunkGroups(
 | |
| 		compilation,
 | |
| 		blocksWithNestedBlocks,
 | |
| 		blockConnections,
 | |
| 		maskByChunk
 | |
| 	);
 | |
| 	logger.timeEnd("connectChunkGroups");
 | |
| 
 | |
| 	for (const [chunkGroup, chunkGroupInfo] of chunkGroupInfoMap) {
 | |
| 		for (const chunk of chunkGroup.chunks)
 | |
| 			chunk.runtime = mergeRuntime(chunk.runtime, chunkGroupInfo.runtime);
 | |
| 	}
 | |
| 
 | |
| 	// Cleanup work
 | |
| 
 | |
| 	logger.time("cleanup");
 | |
| 	cleanupUnconnectedGroups(compilation, allCreatedChunkGroups);
 | |
| 	logger.timeEnd("cleanup");
 | |
| };
 | |
| 
 | |
| module.exports = buildChunkGraph;
 |