2019-07-25 14:07:55 +08:00
|
|
|
/*
|
|
|
|
MIT License http://www.opensource.org/licenses/mit-license.php
|
|
|
|
Author Tobias Koppers @sokra
|
|
|
|
*/
|
|
|
|
|
|
|
|
"use strict";
|
|
|
|
|
|
|
|
const AsyncDependencyToInitialChunkError = require("./AsyncDependencyToInitialChunkError");
|
2020-02-07 17:05:51 +08:00
|
|
|
const { connectChunkGroupParentAndChild } = require("./GraphHelpers");
|
2019-07-25 14:07:55 +08:00
|
|
|
|
|
|
|
/** @typedef {import("./AsyncDependenciesBlock")} AsyncDependenciesBlock */
|
|
|
|
/** @typedef {import("./Chunk")} Chunk */
|
|
|
|
/** @typedef {import("./ChunkGroup")} ChunkGroup */
|
|
|
|
/** @typedef {import("./Compilation")} Compilation */
|
|
|
|
/** @typedef {import("./DependenciesBlock")} DependenciesBlock */
|
|
|
|
/** @typedef {import("./Dependency")} Dependency */
|
|
|
|
/** @typedef {import("./Entrypoint")} Entrypoint */
|
|
|
|
/** @typedef {import("./Module")} Module */
|
|
|
|
/** @typedef {import("./ModuleGraph")} ModuleGraph */
|
2019-07-26 15:56:55 +08:00
|
|
|
/** @typedef {import("./logging/Logger").Logger} Logger */
|
2019-07-25 14:07:55 +08:00
|
|
|
|
2019-07-25 17:25:29 +08:00
|
|
|
/**
|
|
|
|
* @typedef {Object} QueueItem
|
|
|
|
* @property {number} action
|
|
|
|
* @property {DependenciesBlock} block
|
|
|
|
* @property {Module} module
|
|
|
|
* @property {Chunk} chunk
|
|
|
|
* @property {ChunkGroup} chunkGroup
|
2020-01-28 22:44:03 +08:00
|
|
|
* @property {ChunkGroupInfo} chunkGroupInfo
|
2019-07-25 17:25:29 +08:00
|
|
|
*/
|
|
|
|
|
2020-01-31 23:38:04 +08:00
|
|
|
/** @typedef {Set<Module> & { plus: Set<Module> }} ModuleSetPlus */
|
|
|
|
|
2019-07-25 17:25:29 +08:00
|
|
|
/**
|
|
|
|
* @typedef {Object} ChunkGroupInfo
|
2019-08-27 18:25:02 +08:00
|
|
|
* @property {ChunkGroup} chunkGroup the chunk group
|
2020-02-01 23:09:43 +08:00
|
|
|
* @property {ModuleSetPlus} minAvailableModules current minimal set of modules available at this point
|
2019-07-25 21:39:54 +08:00
|
|
|
* @property {boolean} minAvailableModulesOwned true, if minAvailableModules is owned and can be modified
|
2020-01-31 23:38:04 +08:00
|
|
|
* @property {ModuleSetPlus[]} availableModulesToBeMerged enqueued updates to the minimal set of available modules
|
2020-01-29 01:28:53 +08:00
|
|
|
* @property {Set<Module>=} skippedItems modules that were skipped because module is already available in parent chunks (need to reconsider when minAvailableModules is shrinking)
|
2020-01-31 23:38:04 +08:00
|
|
|
* @property {ModuleSetPlus} resultingAvailableModules set of modules available including modules from this chunk group
|
2020-01-28 22:55:51 +08:00
|
|
|
* @property {Set<ChunkGroupInfo>} children set of children chunk groups, that will be revisited when availableModules shrink
|
2020-02-07 17:05:51 +08:00
|
|
|
* @property {Set<ChunkGroupInfo>} availableSources set of chunk groups that are the source for minAvailableModules
|
|
|
|
* @property {Set<ChunkGroupInfo>} availableChildren set of chunk groups which depend on the this chunk group as availableSource
|
2020-01-28 22:13:10 +08:00
|
|
|
* @property {number} preOrderIndex next pre order index
|
|
|
|
* @property {number} postOrderIndex next post order index
|
2019-07-25 17:25:29 +08:00
|
|
|
*/
|
|
|
|
|
|
|
|
/**
|
|
|
|
* @typedef {Object} ChunkGroupDep
|
|
|
|
* @property {AsyncDependenciesBlock} block referencing block
|
|
|
|
* @property {ChunkGroup} chunkGroup referenced chunk group
|
|
|
|
*/
|
|
|
|
|
2020-02-01 23:09:43 +08:00
|
|
|
const EMPTY_SET = /** @type {ModuleSetPlus} */ (new Set());
|
|
|
|
EMPTY_SET.plus = EMPTY_SET;
|
2020-01-31 23:38:04 +08:00
|
|
|
|
2019-07-25 14:07:55 +08:00
|
|
|
/**
|
2020-01-31 23:38:04 +08:00
|
|
|
* @param {ModuleSetPlus} a first set
|
|
|
|
* @param {ModuleSetPlus} b second set
|
2019-07-25 14:07:55 +08:00
|
|
|
* @returns {number} cmp
|
|
|
|
*/
|
|
|
|
const bySetSize = (a, b) => {
|
2020-01-31 23:38:04 +08:00
|
|
|
return b.size + b.plus.size - a.size - a.plus.size;
|
2019-07-25 14:07:55 +08:00
|
|
|
};
|
|
|
|
|
|
|
|
/**
|
2019-10-30 07:09:51 +08:00
|
|
|
* Extracts block to modules mapping from all modules
|
2019-07-25 14:07:55 +08:00
|
|
|
* @param {Compilation} compilation the compilation
|
2019-10-30 07:09:51 +08:00
|
|
|
* @returns {Map<DependenciesBlock, Iterable<Module>>} the mapping block to modules
|
2019-07-25 14:07:55 +08:00
|
|
|
*/
|
2020-01-31 23:54:42 +08:00
|
|
|
const extractBlockModulesMap = compilation => {
|
2019-10-30 07:09:51 +08:00
|
|
|
const { moduleGraph } = compilation;
|
2019-07-25 14:07:55 +08:00
|
|
|
|
2019-10-30 07:09:51 +08:00
|
|
|
/** @type {Map<DependenciesBlock, Iterable<Module>>} */
|
|
|
|
const blockModulesMap = new Map();
|
2019-07-25 14:07:55 +08:00
|
|
|
|
2019-10-30 07:09:51 +08:00
|
|
|
const blockQueue = new Set();
|
2019-07-25 14:07:55 +08:00
|
|
|
|
|
|
|
for (const module of compilation.modules) {
|
2019-10-30 07:09:51 +08:00
|
|
|
/** @type {WeakMap<Dependency, Module>} */
|
|
|
|
let moduleMap;
|
|
|
|
|
|
|
|
for (const connection of moduleGraph.getOutgoingConnections(module)) {
|
|
|
|
const d = connection.dependency;
|
|
|
|
// We skip connections without dependency
|
|
|
|
if (!d) continue;
|
|
|
|
const m = connection.module;
|
|
|
|
// We skip connections without Module pointer
|
|
|
|
if (!m) continue;
|
|
|
|
// We skip weak connections
|
|
|
|
if (connection.weak) continue;
|
|
|
|
// We skip inactive connections
|
|
|
|
if (!connection.active) continue;
|
|
|
|
// Store Dependency to Module mapping in local map
|
|
|
|
// to allow to access it faster compared to
|
|
|
|
// moduleGraph.getConnection()
|
|
|
|
if (moduleMap === undefined) {
|
|
|
|
moduleMap = new WeakMap();
|
2019-07-25 14:07:55 +08:00
|
|
|
}
|
2019-10-30 07:09:51 +08:00
|
|
|
moduleMap.set(connection.dependency, m);
|
|
|
|
}
|
2019-07-25 14:07:55 +08:00
|
|
|
|
2019-10-30 07:09:51 +08:00
|
|
|
blockQueue.clear();
|
|
|
|
blockQueue.add(module);
|
|
|
|
for (const block of blockQueue) {
|
|
|
|
let modules;
|
|
|
|
|
|
|
|
if (moduleMap !== undefined && block.dependencies) {
|
|
|
|
for (const dep of block.dependencies) {
|
|
|
|
const module = moduleMap.get(dep);
|
|
|
|
if (module !== undefined) {
|
|
|
|
if (modules === undefined) {
|
|
|
|
modules = new Set();
|
|
|
|
blockModulesMap.set(block, modules);
|
|
|
|
}
|
|
|
|
modules.add(module);
|
|
|
|
}
|
|
|
|
}
|
2019-07-25 14:07:55 +08:00
|
|
|
}
|
|
|
|
|
2019-10-30 07:09:51 +08:00
|
|
|
if (block.blocks) {
|
|
|
|
for (const b of block.blocks) {
|
|
|
|
blockQueue.add(b);
|
|
|
|
}
|
|
|
|
}
|
2019-07-25 14:07:55 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-10-30 07:09:51 +08:00
|
|
|
return blockModulesMap;
|
2019-07-25 14:07:55 +08:00
|
|
|
};
|
|
|
|
|
|
|
|
/**
|
2019-07-25 17:25:29 +08:00
|
|
|
*
|
2019-07-26 15:56:55 +08:00
|
|
|
* @param {Logger} logger a logger
|
2019-07-25 14:07:55 +08:00
|
|
|
* @param {Compilation} compilation the compilation
|
2020-02-07 17:05:51 +08:00
|
|
|
* @param {Entrypoint[]} inputEntrypoints input groups
|
2019-07-25 17:25:29 +08:00
|
|
|
* @param {Map<ChunkGroup, ChunkGroupInfo>} chunkGroupInfoMap mapping from chunk group to available modules
|
2020-03-18 15:54:43 +08:00
|
|
|
* @param {Map<ChunkGroup, ChunkGroupDep[]>} chunkGroupDependencies dependencies for chunk groups
|
2019-07-25 17:25:29 +08:00
|
|
|
* @param {Set<DependenciesBlock>} blocksWithNestedBlocks flag for blocks that have nested blocks
|
|
|
|
* @param {Set<ChunkGroup>} allCreatedChunkGroups filled with all chunk groups that are created here
|
2019-07-25 14:07:55 +08:00
|
|
|
*/
|
2019-07-25 17:25:29 +08:00
|
|
|
const visitModules = (
|
2019-07-26 15:56:55 +08:00
|
|
|
logger,
|
2019-07-25 17:25:29 +08:00
|
|
|
compilation,
|
2020-02-07 17:05:51 +08:00
|
|
|
inputEntrypoints,
|
2019-07-25 17:25:29 +08:00
|
|
|
chunkGroupInfoMap,
|
2020-03-18 15:54:43 +08:00
|
|
|
chunkGroupDependencies,
|
2019-07-25 17:25:29 +08:00
|
|
|
blocksWithNestedBlocks,
|
|
|
|
allCreatedChunkGroups
|
|
|
|
) => {
|
2020-01-28 22:44:03 +08:00
|
|
|
const { moduleGraph, chunkGraph } = compilation;
|
2019-07-25 17:25:29 +08:00
|
|
|
|
2019-07-26 15:56:55 +08:00
|
|
|
logger.time("visitModules: prepare");
|
2020-01-31 23:54:42 +08:00
|
|
|
const blockModulesMap = extractBlockModulesMap(compilation);
|
2019-07-25 14:07:55 +08:00
|
|
|
|
2019-09-26 21:51:40 +08:00
|
|
|
let nextChunkGroupIndex = 0;
|
2019-07-25 14:07:55 +08:00
|
|
|
let nextFreeModulePreOrderIndex = 0;
|
|
|
|
let nextFreeModulePostOrderIndex = 0;
|
|
|
|
|
2020-01-28 22:44:03 +08:00
|
|
|
/** @type {Map<DependenciesBlock, ChunkGroupInfo>} */
|
2019-07-25 14:07:55 +08:00
|
|
|
const blockChunkGroups = new Map();
|
|
|
|
|
2020-01-28 22:44:03 +08:00
|
|
|
/** @type {Map<string, ChunkGroupInfo>} */
|
|
|
|
const namedChunkGroups = new Map();
|
|
|
|
|
2019-07-25 14:07:55 +08:00
|
|
|
const ADD_AND_ENTER_MODULE = 0;
|
|
|
|
const ENTER_MODULE = 1;
|
|
|
|
const PROCESS_BLOCK = 2;
|
|
|
|
const LEAVE_MODULE = 3;
|
|
|
|
|
2020-01-17 04:22:05 +08:00
|
|
|
/** @type {QueueItem[]} */
|
|
|
|
let queue = [];
|
|
|
|
|
2020-02-07 17:05:51 +08:00
|
|
|
/** @type {Map<ChunkGroupInfo, Set<ChunkGroupInfo>>} */
|
|
|
|
const queueConnect = new Map();
|
|
|
|
/** @type {Set<ChunkGroupInfo>} */
|
|
|
|
const chunkGroupsForCombining = new Set();
|
|
|
|
|
|
|
|
// Fill queue with entrypoint modules
|
|
|
|
// Create ChunkGroupInfo for entrypoints
|
|
|
|
for (const chunkGroup of inputEntrypoints) {
|
|
|
|
/** @type {ChunkGroupInfo} */
|
2020-01-28 22:44:03 +08:00
|
|
|
const chunkGroupInfo = {
|
|
|
|
chunkGroup,
|
2020-02-07 17:05:51 +08:00
|
|
|
minAvailableModules: undefined,
|
2020-02-01 23:09:43 +08:00
|
|
|
minAvailableModulesOwned: false,
|
2020-01-28 22:44:03 +08:00
|
|
|
availableModulesToBeMerged: [],
|
2020-01-29 01:28:53 +08:00
|
|
|
skippedItems: undefined,
|
2020-01-28 22:44:03 +08:00
|
|
|
resultingAvailableModules: undefined,
|
|
|
|
children: undefined,
|
2020-02-07 17:05:51 +08:00
|
|
|
availableSources: undefined,
|
|
|
|
availableChildren: undefined,
|
2020-01-28 22:44:03 +08:00
|
|
|
preOrderIndex: 0,
|
|
|
|
postOrderIndex: 0
|
|
|
|
};
|
2019-09-26 21:51:40 +08:00
|
|
|
chunkGroup.index = nextChunkGroupIndex++;
|
2020-02-07 17:05:51 +08:00
|
|
|
if (chunkGroup.getNumberOfParents() > 0) {
|
|
|
|
// minAvailableModules for child entrypoints are unknown yet, set to undefined.
|
|
|
|
// This means no module is added until other sets are merged into
|
|
|
|
// this minAvailableModules (by the parent entrypoints)
|
|
|
|
const skippedItems = new Set();
|
|
|
|
for (const chunk of chunkGroup.chunks) {
|
|
|
|
for (const module of chunkGraph.getChunkEntryModulesIterable(chunk)) {
|
|
|
|
skippedItems.add(module);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
chunkGroupInfo.skippedItems = skippedItems;
|
|
|
|
chunkGroupsForCombining.add(chunkGroupInfo);
|
|
|
|
} else {
|
|
|
|
// The application may start here: We start with an empty list of available modules
|
|
|
|
chunkGroupInfo.minAvailableModules = EMPTY_SET;
|
|
|
|
for (const chunk of chunkGroup.chunks) {
|
|
|
|
for (const module of chunkGraph.getChunkEntryModulesIterable(chunk)) {
|
|
|
|
queue.push({
|
|
|
|
action: ADD_AND_ENTER_MODULE,
|
|
|
|
block: module,
|
|
|
|
module,
|
|
|
|
chunk,
|
|
|
|
chunkGroup,
|
|
|
|
chunkGroupInfo
|
|
|
|
});
|
|
|
|
}
|
2019-07-25 14:07:55 +08:00
|
|
|
}
|
|
|
|
}
|
2020-01-28 22:44:03 +08:00
|
|
|
chunkGroupInfoMap.set(chunkGroup, chunkGroupInfo);
|
|
|
|
if (chunkGroup.name) {
|
|
|
|
namedChunkGroups.set(chunkGroup.name, chunkGroupInfo);
|
|
|
|
}
|
2020-01-17 04:22:05 +08:00
|
|
|
}
|
2020-02-07 17:05:51 +08:00
|
|
|
// Fill availableSources with parent-child dependencies between entrypoints
|
|
|
|
for (const chunkGroupInfo of chunkGroupsForCombining) {
|
|
|
|
const { chunkGroup } = chunkGroupInfo;
|
|
|
|
chunkGroupInfo.availableSources = new Set();
|
|
|
|
for (const parent of chunkGroup.parentsIterable) {
|
|
|
|
const parentChunkGroupInfo = chunkGroupInfoMap.get(parent);
|
|
|
|
chunkGroupInfo.availableSources.add(parentChunkGroupInfo);
|
|
|
|
if (parentChunkGroupInfo.availableChildren === undefined) {
|
|
|
|
parentChunkGroupInfo.availableChildren = new Set();
|
|
|
|
}
|
|
|
|
parentChunkGroupInfo.availableChildren.add(chunkGroupInfo);
|
|
|
|
}
|
|
|
|
}
|
2020-01-17 04:22:05 +08:00
|
|
|
// pop() is used to read from the queue
|
|
|
|
// so it need to be reversed to be iterated in
|
|
|
|
// correct order
|
|
|
|
queue.reverse();
|
2019-07-25 14:07:55 +08:00
|
|
|
|
2019-07-26 14:05:16 +08:00
|
|
|
/** @type {Set<ChunkGroupInfo>} */
|
|
|
|
const outdatedChunkGroupInfo = new Set();
|
2020-02-07 17:05:51 +08:00
|
|
|
/** @type {Set<ChunkGroupInfo>} */
|
|
|
|
const chunkGroupsForMerging = new Set();
|
2019-07-25 14:07:55 +08:00
|
|
|
/** @type {QueueItem[]} */
|
|
|
|
let queueDelayed = [];
|
2019-07-26 14:05:16 +08:00
|
|
|
|
2019-07-26 15:56:55 +08:00
|
|
|
logger.timeEnd("visitModules: prepare");
|
2019-07-25 14:07:55 +08:00
|
|
|
|
2020-01-29 01:28:53 +08:00
|
|
|
/** @type {Module[]} */
|
|
|
|
const skipBuffer = [];
|
|
|
|
/** @type {QueueItem[]} */
|
|
|
|
const queueBuffer = [];
|
|
|
|
|
2019-07-25 14:07:55 +08:00
|
|
|
/** @type {Module} */
|
|
|
|
let module;
|
|
|
|
/** @type {Chunk} */
|
|
|
|
let chunk;
|
|
|
|
/** @type {ChunkGroup} */
|
|
|
|
let chunkGroup;
|
|
|
|
/** @type {DependenciesBlock} */
|
|
|
|
let block;
|
2020-01-28 22:13:10 +08:00
|
|
|
/** @type {ChunkGroupInfo} */
|
|
|
|
let chunkGroupInfo;
|
2019-07-25 14:07:55 +08:00
|
|
|
|
|
|
|
// For each async Block in graph
|
|
|
|
/**
|
|
|
|
* @param {AsyncDependenciesBlock} b iterating over each Async DepBlock
|
|
|
|
* @returns {void}
|
|
|
|
*/
|
|
|
|
const iteratorBlock = b => {
|
2020-03-18 15:54:43 +08:00
|
|
|
// 1. We create a chunk group for this Block
|
2019-07-25 14:07:55 +08:00
|
|
|
// but only once (blockChunkGroups map)
|
2020-01-28 22:44:03 +08:00
|
|
|
let cgi = blockChunkGroups.get(b);
|
|
|
|
/** @type {ChunkGroup} */
|
|
|
|
let c;
|
|
|
|
if (cgi === undefined) {
|
2020-01-28 22:13:10 +08:00
|
|
|
const chunkName = (b.groupOptions && b.groupOptions.name) || b.chunkName;
|
2020-01-28 22:44:03 +08:00
|
|
|
cgi = namedChunkGroups.get(chunkName);
|
|
|
|
if (!cgi) {
|
2019-07-25 14:07:55 +08:00
|
|
|
c = compilation.addChunkInGroup(
|
|
|
|
b.groupOptions || b.chunkName,
|
|
|
|
module,
|
|
|
|
b.loc,
|
|
|
|
b.request
|
|
|
|
);
|
2020-01-28 22:13:10 +08:00
|
|
|
c.index = nextChunkGroupIndex++;
|
2020-01-28 22:44:03 +08:00
|
|
|
cgi = {
|
2020-01-28 22:13:10 +08:00
|
|
|
chunkGroup: c,
|
|
|
|
minAvailableModules: undefined,
|
|
|
|
minAvailableModulesOwned: undefined,
|
|
|
|
availableModulesToBeMerged: [],
|
2020-01-29 01:28:53 +08:00
|
|
|
skippedItems: undefined,
|
2020-01-28 22:13:10 +08:00
|
|
|
resultingAvailableModules: undefined,
|
|
|
|
children: undefined,
|
2020-02-07 17:05:51 +08:00
|
|
|
availableSources: undefined,
|
|
|
|
availableChildren: undefined,
|
2020-01-28 22:13:10 +08:00
|
|
|
preOrderIndex: 0,
|
|
|
|
postOrderIndex: 0
|
2020-01-28 22:44:03 +08:00
|
|
|
};
|
2019-07-25 14:07:55 +08:00
|
|
|
allCreatedChunkGroups.add(c);
|
2020-01-28 22:44:03 +08:00
|
|
|
chunkGroupInfoMap.set(c, cgi);
|
|
|
|
if (chunkName) {
|
|
|
|
namedChunkGroups.set(chunkName, cgi);
|
|
|
|
}
|
2020-01-28 22:13:10 +08:00
|
|
|
} else {
|
2020-01-28 22:44:03 +08:00
|
|
|
c = cgi.chunkGroup;
|
2020-01-28 22:13:10 +08:00
|
|
|
if (c.isInitial()) {
|
|
|
|
compilation.errors.push(
|
|
|
|
new AsyncDependencyToInitialChunkError(chunkName, module, b.loc)
|
|
|
|
);
|
|
|
|
c = chunkGroup;
|
|
|
|
}
|
|
|
|
c.addOptions(b.groupOptions);
|
|
|
|
c.addOrigin(module, b.loc, b.request);
|
2019-07-25 14:07:55 +08:00
|
|
|
}
|
2020-01-28 22:44:03 +08:00
|
|
|
blockChunkGroups.set(b, cgi);
|
|
|
|
} else {
|
|
|
|
c = cgi.chunkGroup;
|
2019-07-25 14:07:55 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
// 2. We store the Block+Chunk mapping as dependency for the chunk
|
2020-03-18 15:54:43 +08:00
|
|
|
let deps = chunkGroupDependencies.get(chunkGroup);
|
|
|
|
if (!deps) chunkGroupDependencies.set(chunkGroup, (deps = []));
|
2019-07-25 14:07:55 +08:00
|
|
|
deps.push({
|
|
|
|
block: b,
|
2019-07-25 17:25:29 +08:00
|
|
|
chunkGroup: c
|
2019-07-25 14:07:55 +08:00
|
|
|
});
|
|
|
|
|
2019-07-25 17:25:29 +08:00
|
|
|
// 3. We create/update the chunk group info
|
2020-01-28 22:46:47 +08:00
|
|
|
let connectList = queueConnect.get(chunkGroupInfo);
|
2019-07-25 17:25:29 +08:00
|
|
|
if (connectList === undefined) {
|
|
|
|
connectList = new Set();
|
2020-01-28 22:46:47 +08:00
|
|
|
queueConnect.set(chunkGroupInfo, connectList);
|
2019-07-25 17:25:29 +08:00
|
|
|
}
|
2020-01-28 22:55:51 +08:00
|
|
|
connectList.add(cgi);
|
2019-07-25 17:25:29 +08:00
|
|
|
|
|
|
|
// 4. We enqueue the DependenciesBlock for traversal
|
2019-07-25 14:07:55 +08:00
|
|
|
queueDelayed.push({
|
|
|
|
action: PROCESS_BLOCK,
|
|
|
|
block: b,
|
|
|
|
module: module,
|
|
|
|
chunk: c.chunks[0],
|
2020-01-28 22:44:03 +08:00
|
|
|
chunkGroup: c,
|
|
|
|
chunkGroupInfo: cgi
|
2019-07-25 14:07:55 +08:00
|
|
|
});
|
|
|
|
};
|
|
|
|
|
2020-02-07 17:05:51 +08:00
|
|
|
const processQueue = () => {
|
|
|
|
while (queue.length) {
|
|
|
|
const queueItem = queue.pop();
|
|
|
|
module = queueItem.module;
|
|
|
|
block = queueItem.block;
|
|
|
|
chunk = queueItem.chunk;
|
|
|
|
chunkGroup = queueItem.chunkGroup;
|
|
|
|
chunkGroupInfo = queueItem.chunkGroupInfo;
|
|
|
|
|
|
|
|
switch (queueItem.action) {
|
|
|
|
case ADD_AND_ENTER_MODULE: {
|
|
|
|
if (chunkGraph.isModuleInChunk(module, chunk)) {
|
|
|
|
// already connected, skip it
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
// We connect Module and Chunk
|
|
|
|
chunkGraph.connectChunkAndModule(chunk, module);
|
|
|
|
}
|
|
|
|
// fallthrough
|
|
|
|
case ENTER_MODULE: {
|
|
|
|
const index = chunkGroup.getModulePreOrderIndex(module);
|
|
|
|
if (index === undefined) {
|
|
|
|
chunkGroup.setModulePreOrderIndex(
|
|
|
|
module,
|
|
|
|
chunkGroupInfo.preOrderIndex++
|
|
|
|
);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (
|
|
|
|
moduleGraph.setPreOrderIndexIfUnset(
|
|
|
|
module,
|
|
|
|
nextFreeModulePreOrderIndex
|
|
|
|
)
|
|
|
|
) {
|
|
|
|
nextFreeModulePreOrderIndex++;
|
|
|
|
}
|
|
|
|
|
|
|
|
// reuse queueItem
|
|
|
|
queueItem.action = LEAVE_MODULE;
|
|
|
|
queue.push(queueItem);
|
|
|
|
}
|
|
|
|
// fallthrough
|
|
|
|
case PROCESS_BLOCK: {
|
|
|
|
// get prepared block info
|
|
|
|
const blockModules = blockModulesMap.get(block);
|
|
|
|
|
|
|
|
if (blockModules !== undefined) {
|
|
|
|
const { minAvailableModules } = chunkGroupInfo;
|
|
|
|
// Buffer items because order need to be reversed to get indices correct
|
|
|
|
// Traverse all referenced modules
|
|
|
|
for (const refModule of blockModules) {
|
|
|
|
if (chunkGraph.isModuleInChunk(refModule, chunk)) {
|
|
|
|
// skip early if already connected
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
if (
|
|
|
|
minAvailableModules.has(refModule) ||
|
|
|
|
minAvailableModules.plus.has(refModule)
|
|
|
|
) {
|
|
|
|
// already in parent chunks, skip it for now
|
|
|
|
skipBuffer.push(refModule);
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
// enqueue, then add and enter to be in the correct order
|
|
|
|
// this is relevant with circular dependencies
|
|
|
|
queueBuffer.push({
|
|
|
|
action: ADD_AND_ENTER_MODULE,
|
|
|
|
block: refModule,
|
|
|
|
module: refModule,
|
|
|
|
chunk,
|
|
|
|
chunkGroup,
|
|
|
|
chunkGroupInfo
|
|
|
|
});
|
|
|
|
}
|
|
|
|
// Add buffered items in reverse order
|
|
|
|
if (skipBuffer.length > 0) {
|
|
|
|
let { skippedItems } = chunkGroupInfo;
|
|
|
|
if (skippedItems === undefined) {
|
|
|
|
chunkGroupInfo.skippedItems = skippedItems = new Set();
|
|
|
|
}
|
|
|
|
for (let i = skipBuffer.length - 1; i >= 0; i--) {
|
|
|
|
skippedItems.add(skipBuffer[i]);
|
|
|
|
}
|
|
|
|
skipBuffer.length = 0;
|
|
|
|
}
|
|
|
|
if (queueBuffer.length > 0) {
|
|
|
|
for (let i = queueBuffer.length - 1; i >= 0; i--) {
|
|
|
|
queue.push(queueBuffer[i]);
|
|
|
|
}
|
|
|
|
queueBuffer.length = 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Traverse all Blocks
|
|
|
|
for (const b of block.blocks) iteratorBlock(b);
|
|
|
|
|
|
|
|
if (block.blocks.length > 0 && module !== block) {
|
|
|
|
blocksWithNestedBlocks.add(block);
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
case LEAVE_MODULE: {
|
|
|
|
const index = chunkGroup.getModulePostOrderIndex(module);
|
|
|
|
if (index === undefined) {
|
|
|
|
chunkGroup.setModulePostOrderIndex(
|
|
|
|
module,
|
|
|
|
chunkGroupInfo.postOrderIndex++
|
|
|
|
);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (
|
|
|
|
moduleGraph.setPostOrderIndexIfUnset(
|
|
|
|
module,
|
|
|
|
nextFreeModulePostOrderIndex
|
|
|
|
)
|
|
|
|
) {
|
|
|
|
nextFreeModulePostOrderIndex++;
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
const calculateResultingAvailableModules = chunkGroupInfo => {
|
|
|
|
if (chunkGroupInfo.resultingAvailableModules)
|
|
|
|
return chunkGroupInfo.resultingAvailableModules;
|
|
|
|
|
|
|
|
const minAvailableModules = chunkGroupInfo.minAvailableModules;
|
|
|
|
|
|
|
|
// Create a new Set of available modules at this point
|
|
|
|
// We want to be as lazy as possible. There are multiple ways doing this:
|
|
|
|
// Note that resultingAvailableModules is stored as "(a) + (b)" as it's a ModuleSetPlus
|
|
|
|
// - resultingAvailableModules = (modules of chunk) + (minAvailableModules + minAvailableModules.plus)
|
|
|
|
// - resultingAvailableModules = (minAvailableModules + modules of chunk) + (minAvailableModules.plus)
|
|
|
|
// We choose one depending on the size of minAvailableModules vs minAvailableModules.plus
|
|
|
|
|
|
|
|
let resultingAvailableModules;
|
|
|
|
if (minAvailableModules.size > minAvailableModules.plus.size) {
|
|
|
|
// resultingAvailableModules = (modules of chunk) + (minAvailableModules + minAvailableModules.plus)
|
|
|
|
resultingAvailableModules = /** @type {Set<Module> & {plus: Set<Module>}} */ (new Set());
|
|
|
|
for (const module of minAvailableModules.plus)
|
|
|
|
minAvailableModules.add(module);
|
|
|
|
minAvailableModules.plus = EMPTY_SET;
|
|
|
|
resultingAvailableModules.plus = minAvailableModules;
|
|
|
|
chunkGroupInfo.minAvailableModulesOwned = false;
|
|
|
|
} else {
|
|
|
|
// resultingAvailableModules = (minAvailableModules + modules of chunk) + (minAvailableModules.plus)
|
|
|
|
resultingAvailableModules = /** @type {Set<Module> & {plus: Set<Module>}} */ (new Set(
|
|
|
|
minAvailableModules
|
|
|
|
));
|
|
|
|
resultingAvailableModules.plus = minAvailableModules.plus;
|
|
|
|
}
|
|
|
|
|
|
|
|
// add the modules from the chunk group to the set
|
|
|
|
for (const chunk of chunkGroupInfo.chunkGroup.chunks) {
|
|
|
|
for (const m of chunkGraph.getChunkModulesIterable(chunk)) {
|
|
|
|
resultingAvailableModules.add(m);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return (chunkGroupInfo.resultingAvailableModules = resultingAvailableModules);
|
|
|
|
};
|
|
|
|
|
2020-01-31 23:38:04 +08:00
|
|
|
const processConnectQueue = () => {
|
|
|
|
// Figure out new parents for chunk groups
|
|
|
|
// to get new available modules for these children
|
|
|
|
for (const [chunkGroupInfo, targets] of queueConnect) {
|
2020-02-07 17:05:51 +08:00
|
|
|
// 1. Add new targets to the list of children
|
2020-01-31 23:38:04 +08:00
|
|
|
if (chunkGroupInfo.children === undefined) {
|
|
|
|
chunkGroupInfo.children = targets;
|
|
|
|
} else {
|
|
|
|
for (const target of targets) {
|
|
|
|
chunkGroupInfo.children.add(target);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-02-07 17:05:51 +08:00
|
|
|
// 2. Calculate resulting available modules
|
|
|
|
const resultingAvailableModules = calculateResultingAvailableModules(
|
|
|
|
chunkGroupInfo
|
|
|
|
);
|
|
|
|
|
|
|
|
// 3. Update chunk group info
|
2020-01-31 23:38:04 +08:00
|
|
|
for (const target of targets) {
|
|
|
|
target.availableModulesToBeMerged.push(resultingAvailableModules);
|
2020-02-07 17:05:51 +08:00
|
|
|
chunkGroupsForMerging.add(target);
|
2020-01-31 23:38:04 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
queueConnect.clear();
|
|
|
|
};
|
|
|
|
|
2020-02-07 17:05:51 +08:00
|
|
|
const processChunkGroupsForMerging = () => {
|
2020-01-31 23:38:04 +08:00
|
|
|
// Execute the merge
|
2020-02-07 17:05:51 +08:00
|
|
|
for (const info of chunkGroupsForMerging) {
|
2020-01-31 23:38:04 +08:00
|
|
|
const availableModulesToBeMerged = info.availableModulesToBeMerged;
|
|
|
|
let cachedMinAvailableModules = info.minAvailableModules;
|
|
|
|
|
|
|
|
// 1. Get minimal available modules
|
|
|
|
// It doesn't make sense to traverse a chunk again with more available modules.
|
|
|
|
// This step calculates the minimal available modules and skips traversal when
|
|
|
|
// the list didn't shrink.
|
|
|
|
if (availableModulesToBeMerged.length > 1) {
|
|
|
|
availableModulesToBeMerged.sort(bySetSize);
|
|
|
|
}
|
|
|
|
let changed = false;
|
2020-02-01 23:09:43 +08:00
|
|
|
merge: for (const availableModules of availableModulesToBeMerged) {
|
2020-01-31 23:38:04 +08:00
|
|
|
if (cachedMinAvailableModules === undefined) {
|
|
|
|
cachedMinAvailableModules = availableModules;
|
|
|
|
info.minAvailableModules = cachedMinAvailableModules;
|
|
|
|
info.minAvailableModulesOwned = false;
|
|
|
|
changed = true;
|
|
|
|
} else {
|
|
|
|
if (info.minAvailableModulesOwned) {
|
|
|
|
// We own it and can modify it
|
|
|
|
for (const m of cachedMinAvailableModules) {
|
|
|
|
if (!availableModules.has(m) && !availableModules.plus.has(m)) {
|
|
|
|
cachedMinAvailableModules.delete(m);
|
|
|
|
changed = true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
for (const m of cachedMinAvailableModules) {
|
|
|
|
if (!availableModules.has(m) && !availableModules.plus.has(m)) {
|
|
|
|
// cachedMinAvailableModules need to be modified
|
|
|
|
// but we don't own it
|
|
|
|
// construct a new Set as intersection of cachedMinAvailableModules and availableModules
|
2020-02-01 23:09:43 +08:00
|
|
|
const newSet = /** @type {ModuleSetPlus} */ (new Set());
|
|
|
|
newSet.plus = EMPTY_SET;
|
2020-01-31 23:38:04 +08:00
|
|
|
const iterator = cachedMinAvailableModules[Symbol.iterator]();
|
2020-02-01 23:09:43 +08:00
|
|
|
// fast forward add all modules until m
|
2020-01-31 23:38:04 +08:00
|
|
|
/** @type {IteratorResult<Module>} */
|
|
|
|
let it;
|
|
|
|
while (!(it = iterator.next()).done) {
|
|
|
|
const module = it.value;
|
|
|
|
if (module === m) break;
|
|
|
|
newSet.add(module);
|
|
|
|
}
|
2020-02-01 23:09:43 +08:00
|
|
|
// check the remaining modules before adding
|
|
|
|
while (!(it = iterator.next()).done) {
|
|
|
|
const module = it.value;
|
|
|
|
if (
|
|
|
|
availableModules.has(module) ||
|
|
|
|
availableModules.plus.has(module)
|
|
|
|
) {
|
|
|
|
newSet.add(module);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// also check all modules in cachedMinAvailableModules.plus
|
|
|
|
for (const module of cachedMinAvailableModules.plus) {
|
|
|
|
if (
|
|
|
|
availableModules.has(module) ||
|
|
|
|
availableModules.plus.has(module)
|
|
|
|
) {
|
|
|
|
newSet.add(module);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
cachedMinAvailableModules = newSet;
|
|
|
|
info.minAvailableModulesOwned = true;
|
|
|
|
info.minAvailableModules = newSet;
|
|
|
|
changed = true;
|
|
|
|
continue merge;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
for (const m of cachedMinAvailableModules.plus) {
|
|
|
|
if (!availableModules.has(m) && !availableModules.plus.has(m)) {
|
|
|
|
// cachedMinAvailableModules need to be modified
|
|
|
|
// but we don't own it
|
|
|
|
// construct a new Set as intersection of cachedMinAvailableModules and availableModules
|
|
|
|
// we already know that all modules directly from cachedMinAvailableModules are in availableModules too
|
|
|
|
const newSet = /** @type {ModuleSetPlus} */ (new Set(
|
|
|
|
cachedMinAvailableModules
|
|
|
|
));
|
|
|
|
newSet.plus = EMPTY_SET;
|
|
|
|
const iterator = cachedMinAvailableModules.plus[
|
|
|
|
Symbol.iterator
|
|
|
|
]();
|
|
|
|
// fast forward add all modules until m
|
|
|
|
/** @type {IteratorResult<Module>} */
|
|
|
|
let it;
|
|
|
|
while (!(it = iterator.next()).done) {
|
|
|
|
const module = it.value;
|
|
|
|
if (module === m) break;
|
|
|
|
newSet.add(module);
|
|
|
|
}
|
|
|
|
// check the remaining modules before adding
|
2020-01-31 23:38:04 +08:00
|
|
|
while (!(it = iterator.next()).done) {
|
|
|
|
const module = it.value;
|
|
|
|
if (
|
|
|
|
availableModules.has(module) ||
|
|
|
|
availableModules.plus.has(module)
|
|
|
|
) {
|
|
|
|
newSet.add(module);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
cachedMinAvailableModules = newSet;
|
|
|
|
info.minAvailableModulesOwned = true;
|
|
|
|
info.minAvailableModules = newSet;
|
|
|
|
changed = true;
|
2020-02-01 23:09:43 +08:00
|
|
|
continue merge;
|
2020-01-31 23:38:04 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
availableModulesToBeMerged.length = 0;
|
2020-02-07 17:05:51 +08:00
|
|
|
if (changed) {
|
|
|
|
info.resultingAvailableModules = undefined;
|
|
|
|
outdatedChunkGroupInfo.add(info);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
const processChunkGroupsForCombining = () => {
|
|
|
|
loop: for (const info of chunkGroupsForCombining) {
|
|
|
|
for (const source of info.availableSources) {
|
|
|
|
if (!source.minAvailableModules) continue loop;
|
|
|
|
}
|
|
|
|
const availableModules = /** @type {ModuleSetPlus} */ (new Set());
|
|
|
|
availableModules.plus = EMPTY_SET;
|
|
|
|
const mergeSet = set => {
|
|
|
|
if (set.size > availableModules.plus.size) {
|
|
|
|
for (const item of availableModules.plus) availableModules.add(item);
|
|
|
|
availableModules.plus = set;
|
|
|
|
} else {
|
|
|
|
for (const item of set) availableModules.add(item);
|
|
|
|
}
|
|
|
|
};
|
|
|
|
// combine minAvailableModules from all resultingAvailableModules
|
|
|
|
for (const source of info.availableSources) {
|
|
|
|
const resultingAvailableModules = calculateResultingAvailableModules(
|
|
|
|
source
|
|
|
|
);
|
|
|
|
mergeSet(resultingAvailableModules);
|
|
|
|
mergeSet(resultingAvailableModules.plus);
|
|
|
|
}
|
|
|
|
info.minAvailableModules = availableModules;
|
|
|
|
info.minAvailableModulesOwned = false;
|
|
|
|
info.resultingAvailableModules = undefined;
|
|
|
|
outdatedChunkGroupInfo.add(info);
|
|
|
|
}
|
|
|
|
chunkGroupsForCombining.clear();
|
|
|
|
};
|
2020-01-31 23:38:04 +08:00
|
|
|
|
2020-02-07 17:05:51 +08:00
|
|
|
const processOutdatedChunkGroupInfo = () => {
|
|
|
|
// Revisit skipped elements
|
|
|
|
for (const info of outdatedChunkGroupInfo) {
|
|
|
|
// 1. Reconsider skipped items
|
2020-01-31 23:38:04 +08:00
|
|
|
if (info.skippedItems !== undefined) {
|
2020-02-07 17:05:51 +08:00
|
|
|
const { minAvailableModules } = info;
|
2020-01-31 23:38:04 +08:00
|
|
|
for (const module of info.skippedItems) {
|
2020-02-06 18:15:57 +08:00
|
|
|
if (
|
2020-02-07 17:05:51 +08:00
|
|
|
!minAvailableModules.has(module) &&
|
|
|
|
!minAvailableModules.plus.has(module)
|
2020-02-06 18:15:57 +08:00
|
|
|
) {
|
2020-01-31 23:38:04 +08:00
|
|
|
queue.push({
|
|
|
|
action: ADD_AND_ENTER_MODULE,
|
|
|
|
block: module,
|
|
|
|
module,
|
|
|
|
chunk: info.chunkGroup.chunks[0],
|
|
|
|
chunkGroup: info.chunkGroup,
|
|
|
|
chunkGroupInfo: info
|
|
|
|
});
|
|
|
|
info.skippedItems.delete(module);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-02-07 17:05:51 +08:00
|
|
|
// 2. Reconsider children chunk groups
|
2020-01-31 23:38:04 +08:00
|
|
|
if (info.children !== undefined) {
|
|
|
|
for (const cgi of info.children) {
|
|
|
|
let connectList = queueConnect.get(info);
|
|
|
|
if (connectList === undefined) {
|
|
|
|
connectList = new Set();
|
|
|
|
queueConnect.set(info, connectList);
|
|
|
|
}
|
|
|
|
connectList.add(cgi);
|
|
|
|
}
|
|
|
|
}
|
2020-02-07 17:05:51 +08:00
|
|
|
|
|
|
|
// 3. Reconsider chunk groups for combining
|
|
|
|
if (info.availableChildren !== undefined) {
|
|
|
|
for (const cgi of info.availableChildren) {
|
|
|
|
chunkGroupsForCombining.add(cgi);
|
|
|
|
}
|
|
|
|
}
|
2020-01-31 23:38:04 +08:00
|
|
|
}
|
|
|
|
outdatedChunkGroupInfo.clear();
|
|
|
|
};
|
|
|
|
|
2019-07-25 14:07:55 +08:00
|
|
|
// Iterative traversal of the Module graph
|
|
|
|
// Recursive would be simpler to write but could result in Stack Overflows
|
2020-02-07 17:05:51 +08:00
|
|
|
while (queue.length || queueConnect.size) {
|
2019-07-26 15:56:55 +08:00
|
|
|
logger.time("visitModules: visiting");
|
2020-02-07 17:05:51 +08:00
|
|
|
processQueue();
|
|
|
|
logger.timeEnd("visitModules: visiting");
|
2019-07-25 14:07:55 +08:00
|
|
|
|
2020-02-07 17:05:51 +08:00
|
|
|
if (chunkGroupsForCombining.size > 0) {
|
|
|
|
logger.time("visitModules: combine available modules");
|
|
|
|
processChunkGroupsForCombining();
|
|
|
|
logger.timeEnd("visitModules: combine available modules");
|
2019-07-25 14:07:55 +08:00
|
|
|
}
|
2019-07-26 14:05:16 +08:00
|
|
|
|
2020-02-07 17:05:51 +08:00
|
|
|
if (queueConnect.size > 0) {
|
2019-07-26 15:56:55 +08:00
|
|
|
logger.time("visitModules: calculating available modules");
|
2020-01-31 23:38:04 +08:00
|
|
|
processConnectQueue();
|
2019-07-26 15:56:55 +08:00
|
|
|
logger.timeEnd("visitModules: calculating available modules");
|
2019-07-26 14:05:16 +08:00
|
|
|
|
2020-02-07 17:05:51 +08:00
|
|
|
if (chunkGroupsForMerging.size > 0) {
|
2019-07-26 15:56:55 +08:00
|
|
|
logger.time("visitModules: merging available modules");
|
2020-02-07 17:05:51 +08:00
|
|
|
processChunkGroupsForMerging();
|
2019-07-26 15:56:55 +08:00
|
|
|
logger.timeEnd("visitModules: merging available modules");
|
2019-07-25 17:25:29 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-02-07 17:05:51 +08:00
|
|
|
if (outdatedChunkGroupInfo.size > 0) {
|
|
|
|
logger.time("visitModules: check modules for revist");
|
|
|
|
processOutdatedChunkGroupInfo();
|
|
|
|
logger.timeEnd("visitModules: check modules for revist");
|
|
|
|
}
|
|
|
|
|
2019-07-25 17:25:29 +08:00
|
|
|
// Run queueDelayed when all items of the queue are processed
|
2020-01-03 14:21:54 +08:00
|
|
|
// This is important to get the global indexing correct
|
2019-07-25 17:25:29 +08:00
|
|
|
// Async blocks should be processed after all sync blocks are processed
|
|
|
|
if (queue.length === 0) {
|
|
|
|
const tempQueue = queue;
|
|
|
|
queue = queueDelayed.reverse();
|
|
|
|
queueDelayed = tempQueue;
|
|
|
|
}
|
2019-07-25 14:07:55 +08:00
|
|
|
}
|
2019-07-25 17:25:29 +08:00
|
|
|
};
|
|
|
|
|
|
|
|
/**
|
|
|
|
*
|
|
|
|
* @param {Compilation} compilation the compilation
|
|
|
|
* @param {Set<DependenciesBlock>} blocksWithNestedBlocks flag for blocks that have nested blocks
|
2020-03-18 15:54:43 +08:00
|
|
|
* @param {Map<ChunkGroup, ChunkGroupDep[]>} chunkGroupDependencies dependencies for chunk groups
|
2019-07-25 17:25:29 +08:00
|
|
|
* @param {Map<ChunkGroup, ChunkGroupInfo>} chunkGroupInfoMap mapping from chunk group to available modules
|
|
|
|
*/
|
|
|
|
const connectChunkGroups = (
|
|
|
|
compilation,
|
|
|
|
blocksWithNestedBlocks,
|
2020-03-18 15:54:43 +08:00
|
|
|
chunkGroupDependencies,
|
2019-07-25 17:25:29 +08:00
|
|
|
chunkGroupInfoMap
|
|
|
|
) => {
|
|
|
|
const { chunkGraph } = compilation;
|
|
|
|
|
2020-01-31 23:38:04 +08:00
|
|
|
/** @type {ModuleSetPlus} */
|
2019-07-25 17:25:29 +08:00
|
|
|
let resultingAvailableModules;
|
2019-07-25 14:07:55 +08:00
|
|
|
|
|
|
|
/**
|
|
|
|
* Helper function to check if all modules of a chunk are available
|
|
|
|
*
|
|
|
|
* @param {ChunkGroup} chunkGroup the chunkGroup to scan
|
2020-01-31 23:38:04 +08:00
|
|
|
* @param {ModuleSetPlus} availableModules the comparator set
|
2019-07-25 14:07:55 +08:00
|
|
|
* @returns {boolean} return true if all modules of a chunk are available
|
|
|
|
*/
|
|
|
|
const areModulesAvailable = (chunkGroup, availableModules) => {
|
|
|
|
for (const chunk of chunkGroup.chunks) {
|
|
|
|
for (const module of chunkGraph.getChunkModulesIterable(chunk)) {
|
2020-01-31 23:38:04 +08:00
|
|
|
if (!availableModules.has(module) && !availableModules.plus.has(module))
|
|
|
|
return false;
|
2019-07-25 14:07:55 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
return true;
|
|
|
|
};
|
|
|
|
|
|
|
|
// For each edge in the basic chunk graph
|
|
|
|
/**
|
|
|
|
* @param {ChunkGroupDep} dep the dependency used for filtering
|
|
|
|
* @returns {boolean} used to filter "edges" (aka Dependencies) that were pointing
|
|
|
|
* to modules that are already available. Also filters circular dependencies in the chunks graph
|
|
|
|
*/
|
|
|
|
const filterFn = dep => {
|
|
|
|
const depChunkGroup = dep.chunkGroup;
|
2019-07-25 17:25:29 +08:00
|
|
|
// TODO is this needed?
|
2019-07-25 14:07:55 +08:00
|
|
|
if (blocksWithNestedBlocks.has(dep.block)) return true;
|
2019-07-25 17:25:29 +08:00
|
|
|
if (areModulesAvailable(depChunkGroup, resultingAvailableModules)) {
|
2019-07-25 14:07:55 +08:00
|
|
|
return false; // break all modules are already available
|
|
|
|
}
|
|
|
|
return true;
|
|
|
|
};
|
|
|
|
|
2019-07-25 17:25:29 +08:00
|
|
|
// For all deps, check if chunk groups need to be connected
|
2020-03-18 15:54:43 +08:00
|
|
|
for (const [chunkGroup, deps] of chunkGroupDependencies) {
|
2019-07-25 14:07:55 +08:00
|
|
|
if (deps.length === 0) continue;
|
|
|
|
|
2019-07-25 17:25:29 +08:00
|
|
|
// 1. Get info from chunk group info map
|
|
|
|
const info = chunkGroupInfoMap.get(chunkGroup);
|
|
|
|
resultingAvailableModules = info.resultingAvailableModules;
|
2019-07-25 14:07:55 +08:00
|
|
|
|
2019-07-25 17:25:29 +08:00
|
|
|
// 2. Foreach edge
|
2019-07-25 14:07:55 +08:00
|
|
|
for (let i = 0; i < deps.length; i++) {
|
|
|
|
const dep = deps[i];
|
|
|
|
|
|
|
|
// Filter inline, rather than creating a new array from `.filter()`
|
2019-07-25 17:25:29 +08:00
|
|
|
// TODO check if inlining filterFn makes sense here
|
2019-07-25 14:07:55 +08:00
|
|
|
if (!filterFn(dep)) {
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
const depChunkGroup = dep.chunkGroup;
|
|
|
|
const depBlock = dep.block;
|
|
|
|
|
|
|
|
// 5. Connect block with chunk
|
|
|
|
chunkGraph.connectBlockAndChunkGroup(depBlock, depChunkGroup);
|
|
|
|
|
|
|
|
// 6. Connect chunk with parent
|
|
|
|
connectChunkGroupParentAndChild(chunkGroup, depChunkGroup);
|
|
|
|
}
|
|
|
|
}
|
2019-07-25 17:25:29 +08:00
|
|
|
};
|
|
|
|
|
|
|
|
/**
|
|
|
|
* Remove all unconnected chunk groups
|
|
|
|
* @param {Compilation} compilation the compilation
|
|
|
|
* @param {Iterable<ChunkGroup>} allCreatedChunkGroups all chunk groups that where created before
|
|
|
|
*/
|
|
|
|
const cleanupUnconnectedGroups = (compilation, allCreatedChunkGroups) => {
|
|
|
|
const { chunkGraph } = compilation;
|
2019-07-25 14:07:55 +08:00
|
|
|
|
|
|
|
for (const chunkGroup of allCreatedChunkGroups) {
|
|
|
|
if (chunkGroup.getNumberOfParents() === 0) {
|
|
|
|
for (const chunk of chunkGroup.chunks) {
|
|
|
|
compilation.chunks.delete(chunk);
|
|
|
|
chunkGraph.disconnectChunk(chunk);
|
|
|
|
}
|
|
|
|
chunkGraph.disconnectChunkGroup(chunkGroup);
|
|
|
|
chunkGroup.remove();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
2019-07-25 17:25:29 +08:00
|
|
|
/**
|
|
|
|
* This method creates the Chunk graph from the Module graph
|
|
|
|
* @param {Compilation} compilation the compilation
|
2020-02-07 17:05:51 +08:00
|
|
|
* @param {Entrypoint[]} inputEntrypoints chunk groups which are processed
|
2019-07-25 17:25:29 +08:00
|
|
|
* @returns {void}
|
|
|
|
*/
|
2020-02-07 17:05:51 +08:00
|
|
|
const buildChunkGraph = (compilation, inputEntrypoints) => {
|
2019-07-26 15:56:55 +08:00
|
|
|
const logger = compilation.getLogger("webpack.buildChunkGraph");
|
|
|
|
|
2019-07-25 17:25:29 +08:00
|
|
|
// SHARED STATE
|
|
|
|
|
|
|
|
/** @type {Map<ChunkGroup, ChunkGroupDep[]>} */
|
2020-03-18 15:54:43 +08:00
|
|
|
const chunkGroupDependencies = new Map();
|
2019-07-25 17:25:29 +08:00
|
|
|
|
|
|
|
/** @type {Set<ChunkGroup>} */
|
|
|
|
const allCreatedChunkGroups = new Set();
|
|
|
|
|
|
|
|
/** @type {Map<ChunkGroup, ChunkGroupInfo>} */
|
|
|
|
const chunkGroupInfoMap = new Map();
|
|
|
|
|
|
|
|
/** @type {Set<DependenciesBlock>} */
|
|
|
|
const blocksWithNestedBlocks = new Set();
|
|
|
|
|
|
|
|
// PART ONE
|
|
|
|
|
2020-02-01 00:10:53 +08:00
|
|
|
logger.time("visitModules");
|
2019-07-25 17:25:29 +08:00
|
|
|
visitModules(
|
2019-07-26 15:56:55 +08:00
|
|
|
logger,
|
2019-07-25 17:25:29 +08:00
|
|
|
compilation,
|
2020-02-07 17:05:51 +08:00
|
|
|
inputEntrypoints,
|
2019-07-25 17:25:29 +08:00
|
|
|
chunkGroupInfoMap,
|
2020-03-18 15:54:43 +08:00
|
|
|
chunkGroupDependencies,
|
2019-07-25 17:25:29 +08:00
|
|
|
blocksWithNestedBlocks,
|
|
|
|
allCreatedChunkGroups
|
|
|
|
);
|
2020-02-01 00:10:53 +08:00
|
|
|
logger.timeEnd("visitModules");
|
2019-07-25 17:25:29 +08:00
|
|
|
|
|
|
|
// PART TWO
|
|
|
|
|
2019-07-26 15:56:55 +08:00
|
|
|
logger.time("connectChunkGroups");
|
2019-07-25 17:25:29 +08:00
|
|
|
connectChunkGroups(
|
|
|
|
compilation,
|
|
|
|
blocksWithNestedBlocks,
|
2020-03-18 15:54:43 +08:00
|
|
|
chunkGroupDependencies,
|
2019-07-25 17:25:29 +08:00
|
|
|
chunkGroupInfoMap
|
|
|
|
);
|
2019-07-26 15:56:55 +08:00
|
|
|
logger.timeEnd("connectChunkGroups");
|
2019-07-25 17:25:29 +08:00
|
|
|
|
|
|
|
// Cleaup work
|
|
|
|
|
2019-07-26 15:56:55 +08:00
|
|
|
logger.time("cleanup");
|
2019-07-25 17:25:29 +08:00
|
|
|
cleanupUnconnectedGroups(compilation, allCreatedChunkGroups);
|
2019-07-26 15:56:55 +08:00
|
|
|
logger.timeEnd("cleanup");
|
2019-07-25 17:25:29 +08:00
|
|
|
};
|
|
|
|
|
2019-07-25 14:07:55 +08:00
|
|
|
module.exports = buildChunkGraph;
|