(globalThis.TURBOPACK = globalThis.TURBOPACK || []).push([ typeof document === "object" ? document.currentScript : undefined, {}, {"otherChunks":["static/chunks/[turbopack]_browser_dev_hmr-client_hmr-client_ts_fd44f5a4._.js","static/chunks/node_modules_next_dist_compiled_2ce9398a._.js","static/chunks/node_modules_next_dist_client_8f19e6fb._.js","static/chunks/node_modules_next_dist_2ecbf5fa._.js","static/chunks/node_modules_@swc_helpers_cjs_00636ac3._.js"],"runtimeModuleIds":["[project]/node_modules/next/dist/compiled/@next/react-refresh-utils/dist/runtime.js [app-client] (ecmascript)","[project]/node_modules/next/dist/client/app-next-turbopack.js [app-client] (ecmascript)"]} ]); (() => { if (!Array.isArray(globalThis.TURBOPACK)) { return; } const CHUNK_BASE_PATH = "/_next/"; const CHUNK_SUFFIX_PATH = ""; const RELATIVE_ROOT_PATH = "/ROOT"; const RUNTIME_PUBLIC_PATH = "/_next/"; /** * This file contains runtime types and functions that are shared between all * TurboPack ECMAScript runtimes. * * It will be prepended to the runtime code of each runtime. */ /* eslint-disable @typescript-eslint/no-unused-vars */ /// const REEXPORTED_OBJECTS = Symbol("reexported objects"); const hasOwnProperty = Object.prototype.hasOwnProperty; const toStringTag = typeof Symbol !== "undefined" && Symbol.toStringTag; function defineProp(obj, name, options) { if (!hasOwnProperty.call(obj, name)) Object.defineProperty(obj, name, options); } /** * Adds the getters to the exports object. */ function esm(exports, getters) { defineProp(exports, "__esModule", { value: true }); if (toStringTag) defineProp(exports, toStringTag, { value: "Module" }); for(const key in getters){ const item = getters[key]; if (Array.isArray(item)) { defineProp(exports, key, { get: item[0], set: item[1], enumerable: true }); } else { defineProp(exports, key, { get: item, enumerable: true }); } } Object.seal(exports); } /** * Makes the module an ESM with exports */ function esmExport(module, exports, getters) { module.namespaceObject = module.exports; esm(exports, getters); } function ensureDynamicExports(module, exports) { let reexportedObjects = module[REEXPORTED_OBJECTS]; if (!reexportedObjects) { reexportedObjects = module[REEXPORTED_OBJECTS] = []; module.exports = module.namespaceObject = new Proxy(exports, { get (target, prop) { if (hasOwnProperty.call(target, prop) || prop === "default" || prop === "__esModule") { return Reflect.get(target, prop); } for (const obj of reexportedObjects){ const value = Reflect.get(obj, prop); if (value !== undefined) return value; } return undefined; }, ownKeys (target) { const keys = Reflect.ownKeys(target); for (const obj of reexportedObjects){ for (const key of Reflect.ownKeys(obj)){ if (key !== "default" && !keys.includes(key)) keys.push(key); } } return keys; } }); } } /** * Dynamically exports properties from an object */ function dynamicExport(module, exports, object) { ensureDynamicExports(module, exports); if (typeof object === "object" && object !== null) { module[REEXPORTED_OBJECTS].push(object); } } function exportValue(module, value) { module.exports = value; } function exportNamespace(module, namespace) { module.exports = module.namespaceObject = namespace; } function createGetter(obj, key) { return ()=>obj[key]; } /** * @returns prototype of the object */ const getProto = Object.getPrototypeOf ? (obj)=>Object.getPrototypeOf(obj) : (obj)=>obj.__proto__; /** Prototypes that are not expanded for exports */ const LEAF_PROTOTYPES = [ null, getProto({}), getProto([]), getProto(getProto) ]; /** * @param raw * @param ns * @param allowExportDefault * * `false`: will have the raw module as default export * * `true`: will have the default property as default export */ function interopEsm(raw, ns, allowExportDefault) { const getters = Object.create(null); for(let current = raw; (typeof current === "object" || typeof current === "function") && !LEAF_PROTOTYPES.includes(current); current = getProto(current)){ for (const key of Object.getOwnPropertyNames(current)){ getters[key] = createGetter(raw, key); } } // this is not really correct // we should set the `default` getter if the imported module is a `.cjs file` if (!(allowExportDefault && "default" in getters)) { getters["default"] = ()=>raw; } esm(ns, getters); return ns; } function createNS(raw) { if (typeof raw === "function") { return function(...args) { return raw.apply(this, args); }; } else { return Object.create(null); } } function esmImport(sourceModule, id) { const module = getOrInstantiateModuleFromParent(id, sourceModule); if (module.error) throw module.error; // any ES module has to have `module.namespaceObject` defined. if (module.namespaceObject) return module.namespaceObject; // only ESM can be an async module, so we don't need to worry about exports being a promise here. const raw = module.exports; return module.namespaceObject = interopEsm(raw, createNS(raw), raw && raw.__esModule); } // Add a simple runtime require so that environments without one can still pass // `typeof require` CommonJS checks so that exports are correctly registered. const runtimeRequire = // @ts-ignore typeof require === "function" ? require : function require1() { throw new Error("Unexpected use of runtime require"); }; function commonJsRequire(sourceModule, id) { const module = getOrInstantiateModuleFromParent(id, sourceModule); if (module.error) throw module.error; return module.exports; } /** * `require.context` and require/import expression runtime. */ function moduleContext(map) { function moduleContext(id) { if (hasOwnProperty.call(map, id)) { return map[id].module(); } const e = new Error(`Cannot find module '${id}'`); e.code = "MODULE_NOT_FOUND"; throw e; } moduleContext.keys = ()=>{ return Object.keys(map); }; moduleContext.resolve = (id)=>{ if (hasOwnProperty.call(map, id)) { return map[id].id(); } const e = new Error(`Cannot find module '${id}'`); e.code = "MODULE_NOT_FOUND"; throw e; }; moduleContext.import = async (id)=>{ return await moduleContext(id); }; return moduleContext; } /** * Returns the path of a chunk defined by its data. */ function getChunkPath(chunkData) { return typeof chunkData === "string" ? chunkData : chunkData.path; } function isPromise(maybePromise) { return maybePromise != null && typeof maybePromise === "object" && "then" in maybePromise && typeof maybePromise.then === "function"; } function isAsyncModuleExt(obj) { return turbopackQueues in obj; } function createPromise() { let resolve; let reject; const promise = new Promise((res, rej)=>{ reject = rej; resolve = res; }); return { promise, resolve: resolve, reject: reject }; } // everything below is adapted from webpack // https://github.com/webpack/webpack/blob/6be4065ade1e252c1d8dcba4af0f43e32af1bdc1/lib/runtime/AsyncModuleRuntimeModule.js#L13 const turbopackQueues = Symbol("turbopack queues"); const turbopackExports = Symbol("turbopack exports"); const turbopackError = Symbol("turbopack error"); function resolveQueue(queue) { if (queue && queue.status !== 1) { queue.status = 1; queue.forEach((fn)=>fn.queueCount--); queue.forEach((fn)=>fn.queueCount-- ? fn.queueCount++ : fn()); } } function wrapDeps(deps) { return deps.map((dep)=>{ if (dep !== null && typeof dep === "object") { if (isAsyncModuleExt(dep)) return dep; if (isPromise(dep)) { const queue = Object.assign([], { status: 0 }); const obj = { [turbopackExports]: {}, [turbopackQueues]: (fn)=>fn(queue) }; dep.then((res)=>{ obj[turbopackExports] = res; resolveQueue(queue); }, (err)=>{ obj[turbopackError] = err; resolveQueue(queue); }); return obj; } } return { [turbopackExports]: dep, [turbopackQueues]: ()=>{} }; }); } function asyncModule(module, body, hasAwait) { const queue = hasAwait ? Object.assign([], { status: -1 }) : undefined; const depQueues = new Set(); const { resolve, reject, promise: rawPromise } = createPromise(); const promise = Object.assign(rawPromise, { [turbopackExports]: module.exports, [turbopackQueues]: (fn)=>{ queue && fn(queue); depQueues.forEach(fn); promise["catch"](()=>{}); } }); const attributes = { get () { return promise; }, set (v) { // Calling `esmExport` leads to this. if (v !== promise) { promise[turbopackExports] = v; } } }; Object.defineProperty(module, "exports", attributes); Object.defineProperty(module, "namespaceObject", attributes); function handleAsyncDependencies(deps) { const currentDeps = wrapDeps(deps); const getResult = ()=>currentDeps.map((d)=>{ if (d[turbopackError]) throw d[turbopackError]; return d[turbopackExports]; }); const { promise, resolve } = createPromise(); const fn = Object.assign(()=>resolve(getResult), { queueCount: 0 }); function fnQueue(q) { if (q !== queue && !depQueues.has(q)) { depQueues.add(q); if (q && q.status === 0) { fn.queueCount++; q.push(fn); } } } currentDeps.map((dep)=>dep[turbopackQueues](fnQueue)); return fn.queueCount ? promise : getResult(); } function asyncResult(err) { if (err) { reject(promise[turbopackError] = err); } else { resolve(promise[turbopackExports]); } resolveQueue(queue); } body(handleAsyncDependencies, asyncResult); if (queue && queue.status === -1) { queue.status = 0; } } /** * A pseudo "fake" URL object to resolve to its relative path. * * When UrlRewriteBehavior is set to relative, calls to the `new URL()` will construct url without base using this * runtime function to generate context-agnostic urls between different rendering context, i.e ssr / client to avoid * hydration mismatch. * * This is based on webpack's existing implementation: * https://github.com/webpack/webpack/blob/87660921808566ef3b8796f8df61bd79fc026108/lib/runtime/RelativeUrlRuntimeModule.js */ const relativeURL = function relativeURL(inputUrl) { const realUrl = new URL(inputUrl, "x:/"); const values = {}; for(const key in realUrl)values[key] = realUrl[key]; values.href = inputUrl; values.pathname = inputUrl.replace(/[?#].*/, ""); values.origin = values.protocol = ""; values.toString = values.toJSON = (..._args)=>inputUrl; for(const key in values)Object.defineProperty(this, key, { enumerable: true, configurable: true, value: values[key] }); }; relativeURL.prototype = URL.prototype; /** * Utility function to ensure all variants of an enum are handled. */ function invariant(never, computeMessage) { throw new Error(`Invariant: ${computeMessage(never)}`); } /** * A stub function to make `require` available but non-functional in ESM. */ function requireStub(_moduleId) { throw new Error("dynamic usage of require is not supported"); } /** * This file contains runtime types and functions that are shared between all * Turbopack *development* ECMAScript runtimes. * * It will be appended to the runtime code of each runtime right after the * shared runtime utils. */ /* eslint-disable @typescript-eslint/no-unused-vars */ /// /// // Used in WebWorkers to tell the runtime about the chunk base path var SourceType = /*#__PURE__*/ function(SourceType) { /** * The module was instantiated because it was included in an evaluated chunk's * runtime. */ SourceType[SourceType["Runtime"] = 0] = "Runtime"; /** * The module was instantiated because a parent module imported it. */ SourceType[SourceType["Parent"] = 1] = "Parent"; /** * The module was instantiated because it was included in a chunk's hot module * update. */ SourceType[SourceType["Update"] = 2] = "Update"; return SourceType; }(SourceType || {}); const moduleFactories = Object.create(null); /** * Module IDs that are instantiated as part of the runtime of a chunk. */ const runtimeModules = new Set(); /** * Map from module ID to the chunks that contain this module. * * In HMR, we need to keep track of which modules are contained in which so * chunks. This is so we don't eagerly dispose of a module when it is removed * from chunk A, but still exists in chunk B. */ const moduleChunksMap = new Map(); /** * Map from a chunk path to all modules it contains. */ const chunkModulesMap = new Map(); /** * Chunk lists that contain a runtime. When these chunk lists receive an update * that can't be reconciled with the current state of the page, we need to * reload the runtime entirely. */ const runtimeChunkLists = new Set(); /** * Map from a chunk list to the chunk paths it contains. */ const chunkListChunksMap = new Map(); /** * Map from a chunk path to the chunk lists it belongs to. */ const chunkChunkListsMap = new Map(); const availableModules = new Map(); const availableModuleChunks = new Map(); async function loadChunk(source, chunkData) { if (typeof chunkData === "string") { return loadChunkPath(source, chunkData); } const includedList = chunkData.included || []; const modulesPromises = includedList.map((included)=>{ if (moduleFactories[included]) return true; return availableModules.get(included); }); if (modulesPromises.length > 0 && modulesPromises.every((p)=>p)) { // When all included items are already loaded or loading, we can skip loading ourselves return Promise.all(modulesPromises); } const includedModuleChunksList = chunkData.moduleChunks || []; const moduleChunksPromises = includedModuleChunksList.map((included)=>{ // TODO(alexkirsz) Do we need this check? // if (moduleFactories[included]) return true; return availableModuleChunks.get(included); }).filter((p)=>p); let promise; if (moduleChunksPromises.length > 0) { // Some module chunks are already loaded or loading. if (moduleChunksPromises.length === includedModuleChunksList.length) { // When all included module chunks are already loaded or loading, we can skip loading ourselves return Promise.all(moduleChunksPromises); } const moduleChunksToLoad = new Set(); for (const moduleChunk of includedModuleChunksList){ if (!availableModuleChunks.has(moduleChunk)) { moduleChunksToLoad.add(moduleChunk); } } for (const moduleChunkToLoad of moduleChunksToLoad){ const promise = loadChunkPath(source, moduleChunkToLoad); availableModuleChunks.set(moduleChunkToLoad, promise); moduleChunksPromises.push(promise); } promise = Promise.all(moduleChunksPromises); } else { promise = loadChunkPath(source, chunkData.path); // Mark all included module chunks as loading if they are not already loaded or loading. for (const includedModuleChunk of includedModuleChunksList){ if (!availableModuleChunks.has(includedModuleChunk)) { availableModuleChunks.set(includedModuleChunk, promise); } } } for (const included of includedList){ if (!availableModules.has(included)) { // It might be better to race old and new promises, but it's rare that the new promise will be faster than a request started earlier. // In production it's even more rare, because the chunk optimization tries to deduplicate modules anyway. availableModules.set(included, promise); } } return promise; } async function loadChunkByUrl(source, chunkUrl) { try { await BACKEND.loadChunk(chunkUrl, source); } catch (error) { let loadReason; switch(source.type){ case 0: loadReason = `as a runtime dependency of chunk ${source.chunkPath}`; break; case 1: loadReason = `from module ${source.parentId}`; break; case 2: loadReason = "from an HMR update"; break; default: invariant(source, (source)=>`Unknown source type: ${source?.type}`); } throw new Error(`Failed to load chunk ${chunkUrl} ${loadReason}${error ? `: ${error}` : ""}`, error ? { cause: error } : undefined); } } async function loadChunkPath(source, chunkPath) { const url = getChunkRelativeUrl(chunkPath); return loadChunkByUrl(source, url); } /** * Returns an absolute url to an asset. */ function createResolvePathFromModule(resolver) { return function resolvePathFromModule(moduleId) { const exported = resolver(moduleId); return exported?.default ?? exported; }; } /** * no-op for browser * @param modulePath */ function resolveAbsolutePath(modulePath) { return `/ROOT/${modulePath ?? ""}`; } function getWorkerBlobURL(chunks) { let bootstrap = `self.TURBOPACK_WORKER_LOCATION = ${JSON.stringify(location.origin)}; self.TURBOPACK_NEXT_CHUNK_URLS = ${JSON.stringify(chunks.reverse().map(getChunkRelativeUrl), null, 2)}; importScripts(...self.TURBOPACK_NEXT_CHUNK_URLS.map(c => self.TURBOPACK_WORKER_LOCATION + c).reverse());`; let blob = new Blob([ bootstrap ], { type: "text/javascript" }); return URL.createObjectURL(blob); } /** * Adds a module to a chunk. */ function addModuleToChunk(moduleId, chunkPath) { let moduleChunks = moduleChunksMap.get(moduleId); if (!moduleChunks) { moduleChunks = new Set([ chunkPath ]); moduleChunksMap.set(moduleId, moduleChunks); } else { moduleChunks.add(chunkPath); } let chunkModules = chunkModulesMap.get(chunkPath); if (!chunkModules) { chunkModules = new Set([ moduleId ]); chunkModulesMap.set(chunkPath, chunkModules); } else { chunkModules.add(moduleId); } } /** * Returns the first chunk that included a module. * This is used by the Node.js backend, hence why it's marked as unused in this * file. */ function getFirstModuleChunk(moduleId) { const moduleChunkPaths = moduleChunksMap.get(moduleId); if (moduleChunkPaths == null) { return null; } return moduleChunkPaths.values().next().value; } /** * Instantiates a runtime module. */ function instantiateRuntimeModule(moduleId, chunkPath) { return instantiateModule(moduleId, { type: 0, chunkPath }); } /** * Returns the URL relative to the origin where a chunk can be fetched from. */ function getChunkRelativeUrl(chunkPath) { return `${CHUNK_BASE_PATH}${chunkPath.split("/").map((p)=>encodeURIComponent(p)).join("/")}${CHUNK_SUFFIX_PATH}`; } function getPathFromScript(chunkScript) { if (typeof chunkScript === "string") { return chunkScript; } const chunkUrl = typeof TURBOPACK_NEXT_CHUNK_URLS !== "undefined" ? TURBOPACK_NEXT_CHUNK_URLS.pop() : chunkScript.getAttribute("src"); const src = decodeURIComponent(chunkUrl.replace(/[?#].*$/, "")); const path = src.startsWith(CHUNK_BASE_PATH) ? src.slice(CHUNK_BASE_PATH.length) : src; return path; } /** * Marks a chunk list as a runtime chunk list. There can be more than one * runtime chunk list. For instance, integration tests can have multiple chunk * groups loaded at runtime, each with its own chunk list. */ function markChunkListAsRuntime(chunkListPath) { runtimeChunkLists.add(chunkListPath); } function registerChunk([chunkScript, chunkModules, runtimeParams]) { const chunkPath = getPathFromScript(chunkScript); for (const [moduleId, moduleFactory] of Object.entries(chunkModules)){ if (!moduleFactories[moduleId]) { moduleFactories[moduleId] = moduleFactory; } addModuleToChunk(moduleId, chunkPath); } return BACKEND.registerChunk(chunkPath, runtimeParams); } const regexJsUrl = /\.js(?:\?[^#]*)?(?:#.*)?$/; /** * Checks if a given path/URL ends with .js, optionally followed by ?query or #fragment. */ function isJs(chunkUrlOrPath) { return regexJsUrl.test(chunkUrlOrPath); } const regexCssUrl = /\.css(?:\?[^#]*)?(?:#.*)?$/; /** * Checks if a given path/URL ends with .css, optionally followed by ?query or #fragment. */ function isCss(chunkUrl) { return regexCssUrl.test(chunkUrl); } /// /// /// /** * This file contains runtime types and functions that are shared between all * Turbopack *development* ECMAScript runtimes. * * It will be appended to the runtime code of each runtime right after the * shared runtime utils. */ /* eslint-disable @typescript-eslint/no-unused-vars */ const devModuleCache = Object.create(null); class UpdateApplyError extends Error { name = "UpdateApplyError"; dependencyChain; constructor(message, dependencyChain){ super(message); this.dependencyChain = dependencyChain; } } /** * Maps module IDs to persisted data between executions of their hot module * implementation (`hot.data`). */ const moduleHotData = new Map(); /** * Maps module instances to their hot module state. */ const moduleHotState = new Map(); /** * Modules that call `module.hot.invalidate()` (while being updated). */ const queuedInvalidatedModules = new Set(); /** * Gets or instantiates a runtime module. */ // @ts-ignore function getOrInstantiateRuntimeModule(moduleId, chunkPath) { const module = devModuleCache[moduleId]; if (module) { if (module.error) { throw module.error; } return module; } // @ts-ignore return instantiateModule(moduleId, { type: SourceType.Runtime, chunkPath }); } /** * Retrieves a module from the cache, or instantiate it if it is not cached. */ // @ts-ignore Defined in `runtime-utils.ts` const getOrInstantiateModuleFromParent = (id, sourceModule)=>{ if (!sourceModule.hot.active) { console.warn(`Unexpected import of module ${id} from module ${sourceModule.id}, which was deleted by an HMR update`); } const module = devModuleCache[id]; if (sourceModule.children.indexOf(id) === -1) { sourceModule.children.push(id); } if (module) { if (module.parents.indexOf(sourceModule.id) === -1) { module.parents.push(sourceModule.id); } return module; } return instantiateModule(id, { type: SourceType.Parent, parentId: sourceModule.id }); }; // @ts-ignore Defined in `runtime-base.ts` function instantiateModule(id, source) { const moduleFactory = moduleFactories[id]; if (typeof moduleFactory !== "function") { // This can happen if modules incorrectly handle HMR disposes/updates, // e.g. when they keep a `setTimeout` around which still executes old code // and contains e.g. a `require("something")` call. let instantiationReason; switch(source.type){ case SourceType.Runtime: instantiationReason = `as a runtime entry of chunk ${source.chunkPath}`; break; case SourceType.Parent: instantiationReason = `because it was required from module ${source.parentId}`; break; case SourceType.Update: instantiationReason = "because of an HMR update"; break; default: invariant(source, (source)=>`Unknown source type: ${source?.type}`); } throw new Error(`Module ${id} was instantiated ${instantiationReason}, but the module factory is not available. It might have been deleted in an HMR update.`); } const hotData = moduleHotData.get(id); const { hot, hotState } = createModuleHot(id, hotData); let parents; switch(source.type){ case SourceType.Runtime: runtimeModules.add(id); parents = []; break; case SourceType.Parent: // No need to add this module as a child of the parent module here, this // has already been taken care of in `getOrInstantiateModuleFromParent`. parents = [ source.parentId ]; break; case SourceType.Update: parents = source.parents || []; break; default: invariant(source, (source)=>`Unknown source type: ${source?.type}`); } const module = { exports: {}, error: undefined, loaded: false, id, parents, children: [], namespaceObject: undefined, hot }; devModuleCache[id] = module; moduleHotState.set(module, hotState); // NOTE(alexkirsz) This can fail when the module encounters a runtime error. try { const sourceInfo = { type: SourceType.Parent, parentId: id }; runModuleExecutionHooks(module, (refresh)=>{ const r = commonJsRequire.bind(null, module); moduleFactory.call(module.exports, augmentContext({ a: asyncModule.bind(null, module), e: module.exports, r: commonJsRequire.bind(null, module), t: runtimeRequire, f: moduleContext, i: esmImport.bind(null, module), s: esmExport.bind(null, module, module.exports), j: dynamicExport.bind(null, module, module.exports), v: exportValue.bind(null, module), n: exportNamespace.bind(null, module), m: module, c: devModuleCache, M: moduleFactories, l: loadChunk.bind(null, sourceInfo), L: loadChunkByUrl.bind(null, sourceInfo), w: loadWebAssembly.bind(null, sourceInfo), u: loadWebAssemblyModule.bind(null, sourceInfo), g: globalThis, P: resolveAbsolutePath, U: relativeURL, k: refresh, R: createResolvePathFromModule(r), b: getWorkerBlobURL, z: requireStub, d: typeof module.id === "string" ? module.id.replace(/(^|\/)\/+$/, "") : module.id })); }); } catch (error) { module.error = error; throw error; } module.loaded = true; if (module.namespaceObject && module.exports !== module.namespaceObject) { // in case of a circular dependency: cjs1 -> esm2 -> cjs1 interopEsm(module.exports, module.namespaceObject); } return module; } /** * NOTE(alexkirsz) Webpack has a "module execution" interception hook that * Next.js' React Refresh runtime hooks into to add module context to the * refresh registry. */ function runModuleExecutionHooks(module, executeModule) { const cleanupReactRefreshIntercept = typeof globalThis.$RefreshInterceptModuleExecution$ === "function" ? globalThis.$RefreshInterceptModuleExecution$(module.id) : ()=>{}; try { executeModule({ register: globalThis.$RefreshReg$, signature: globalThis.$RefreshSig$, registerExports: registerExportsAndSetupBoundaryForReactRefresh }); } catch (e) { throw e; } finally{ // Always cleanup the intercept, even if module execution failed. cleanupReactRefreshIntercept(); } } /** * This is adapted from https://github.com/vercel/next.js/blob/3466862d9dc9c8bb3131712134d38757b918d1c0/packages/react-refresh-utils/internal/ReactRefreshModule.runtime.ts */ function registerExportsAndSetupBoundaryForReactRefresh(module, helpers) { const currentExports = module.exports; const prevExports = module.hot.data.prevExports ?? null; helpers.registerExportsForReactRefresh(currentExports, module.id); // A module can be accepted automatically based on its exports, e.g. when // it is a Refresh Boundary. if (helpers.isReactRefreshBoundary(currentExports)) { // Save the previous exports on update, so we can compare the boundary // signatures. module.hot.dispose((data)=>{ data.prevExports = currentExports; }); // Unconditionally accept an update to this module, we'll check if it's // still a Refresh Boundary later. module.hot.accept(); // This field is set when the previous version of this module was a // Refresh Boundary, letting us know we need to check for invalidation or // enqueue an update. if (prevExports !== null) { // A boundary can become ineligible if its exports are incompatible // with the previous exports. // // For example, if you add/remove/change exports, we'll want to // re-execute the importing modules, and force those components to // re-render. Similarly, if you convert a class component to a // function, we want to invalidate the boundary. if (helpers.shouldInvalidateReactRefreshBoundary(helpers.getRefreshBoundarySignature(prevExports), helpers.getRefreshBoundarySignature(currentExports))) { module.hot.invalidate(); } else { helpers.scheduleUpdate(); } } } else { // Since we just executed the code for the module, it's possible that the // new exports made it ineligible for being a boundary. // We only care about the case when we were _previously_ a boundary, // because we already accepted this update (accidental side effect). const isNoLongerABoundary = prevExports !== null; if (isNoLongerABoundary) { module.hot.invalidate(); } } } function formatDependencyChain(dependencyChain) { return `Dependency chain: ${dependencyChain.join(" -> ")}`; } function computeOutdatedModules(added, modified) { const newModuleFactories = new Map(); for (const [moduleId, entry] of added){ if (entry != null) { newModuleFactories.set(moduleId, _eval(entry)); } } const outdatedModules = computedInvalidatedModules(modified.keys()); for (const [moduleId, entry] of modified){ newModuleFactories.set(moduleId, _eval(entry)); } return { outdatedModules, newModuleFactories }; } function computedInvalidatedModules(invalidated) { const outdatedModules = new Set(); for (const moduleId of invalidated){ const effect = getAffectedModuleEffects(moduleId); switch(effect.type){ case "unaccepted": throw new UpdateApplyError(`cannot apply update: unaccepted module. ${formatDependencyChain(effect.dependencyChain)}.`, effect.dependencyChain); case "self-declined": throw new UpdateApplyError(`cannot apply update: self-declined module. ${formatDependencyChain(effect.dependencyChain)}.`, effect.dependencyChain); case "accepted": for (const outdatedModuleId of effect.outdatedModules){ outdatedModules.add(outdatedModuleId); } break; // TODO(alexkirsz) Dependencies: handle dependencies effects. default: invariant(effect, (effect)=>`Unknown effect type: ${effect?.type}`); } } return outdatedModules; } function computeOutdatedSelfAcceptedModules(outdatedModules) { const outdatedSelfAcceptedModules = []; for (const moduleId of outdatedModules){ const module = devModuleCache[moduleId]; const hotState = moduleHotState.get(module); if (module && hotState.selfAccepted && !hotState.selfInvalidated) { outdatedSelfAcceptedModules.push({ moduleId, errorHandler: hotState.selfAccepted }); } } return outdatedSelfAcceptedModules; } /** * Adds, deletes, and moves modules between chunks. This must happen before the * dispose phase as it needs to know which modules were removed from all chunks, * which we can only compute *after* taking care of added and moved modules. */ function updateChunksPhase(chunksAddedModules, chunksDeletedModules) { for (const [chunkPath, addedModuleIds] of chunksAddedModules){ for (const moduleId of addedModuleIds){ addModuleToChunk(moduleId, chunkPath); } } const disposedModules = new Set(); for (const [chunkPath, addedModuleIds] of chunksDeletedModules){ for (const moduleId of addedModuleIds){ if (removeModuleFromChunk(moduleId, chunkPath)) { disposedModules.add(moduleId); } } } return { disposedModules }; } function disposePhase(outdatedModules, disposedModules) { for (const moduleId of outdatedModules){ disposeModule(moduleId, "replace"); } for (const moduleId of disposedModules){ disposeModule(moduleId, "clear"); } // Removing modules from the module cache is a separate step. // We also want to keep track of previous parents of the outdated modules. const outdatedModuleParents = new Map(); for (const moduleId of outdatedModules){ const oldModule = devModuleCache[moduleId]; outdatedModuleParents.set(moduleId, oldModule?.parents); delete devModuleCache[moduleId]; } // TODO(alexkirsz) Dependencies: remove outdated dependency from module // children. return { outdatedModuleParents }; } /** * Disposes of an instance of a module. * * Returns the persistent hot data that should be kept for the next module * instance. * * NOTE: mode = "replace" will not remove modules from the devModuleCache * This must be done in a separate step afterwards. * This is important because all modules need to be disposed to update the * parent/child relationships before they are actually removed from the devModuleCache. * If this was done in this method, the following disposeModule calls won't find * the module from the module id in the cache. */ function disposeModule(moduleId, mode) { const module = devModuleCache[moduleId]; if (!module) { return; } const hotState = moduleHotState.get(module); const data = {}; // Run the `hot.dispose` handler, if any, passing in the persistent // `hot.data` object. for (const disposeHandler of hotState.disposeHandlers){ disposeHandler(data); } // This used to warn in `getOrInstantiateModuleFromParent` when a disposed // module is still importing other modules. module.hot.active = false; moduleHotState.delete(module); // TODO(alexkirsz) Dependencies: delete the module from outdated deps. // Remove the disposed module from its children's parent list. // It will be added back once the module re-instantiates and imports its // children again. for (const childId of module.children){ const child = devModuleCache[childId]; if (!child) { continue; } const idx = child.parents.indexOf(module.id); if (idx >= 0) { child.parents.splice(idx, 1); } } switch(mode){ case "clear": delete devModuleCache[module.id]; moduleHotData.delete(module.id); break; case "replace": moduleHotData.set(module.id, data); break; default: invariant(mode, (mode)=>`invalid mode: ${mode}`); } } function applyPhase(outdatedSelfAcceptedModules, newModuleFactories, outdatedModuleParents, reportError) { // Update module factories. for (const [moduleId, factory] of newModuleFactories.entries()){ moduleFactories[moduleId] = factory; } // TODO(alexkirsz) Run new runtime entries here. // TODO(alexkirsz) Dependencies: call accept handlers for outdated deps. // Re-instantiate all outdated self-accepted modules. for (const { moduleId, errorHandler } of outdatedSelfAcceptedModules){ try { instantiateModule(moduleId, { type: SourceType.Update, parents: outdatedModuleParents.get(moduleId) }); } catch (err) { if (typeof errorHandler === "function") { try { errorHandler(err, { moduleId, module: devModuleCache[moduleId] }); } catch (err2) { reportError(err2); reportError(err); } } else { reportError(err); } } } } function applyUpdate(update) { switch(update.type){ case "ChunkListUpdate": applyChunkListUpdate(update); break; default: invariant(update, (update)=>`Unknown update type: ${update.type}`); } } function applyChunkListUpdate(update) { if (update.merged != null) { for (const merged of update.merged){ switch(merged.type){ case "EcmascriptMergedUpdate": applyEcmascriptMergedUpdate(merged); break; default: invariant(merged, (merged)=>`Unknown merged type: ${merged.type}`); } } } if (update.chunks != null) { for (const [chunkPath, chunkUpdate] of Object.entries(update.chunks)){ const chunkUrl = getChunkRelativeUrl(chunkPath); switch(chunkUpdate.type){ case "added": BACKEND.loadChunk(chunkUrl, { type: SourceType.Update }); break; case "total": DEV_BACKEND.reloadChunk?.(chunkUrl); break; case "deleted": DEV_BACKEND.unloadChunk?.(chunkUrl); break; case "partial": invariant(chunkUpdate.instruction, (instruction)=>`Unknown partial instruction: ${JSON.stringify(instruction)}.`); break; default: invariant(chunkUpdate, (chunkUpdate)=>`Unknown chunk update type: ${chunkUpdate.type}`); } } } } function applyEcmascriptMergedUpdate(update) { const { entries = {}, chunks = {} } = update; const { added, modified, chunksAdded, chunksDeleted } = computeChangedModules(entries, chunks); const { outdatedModules, newModuleFactories } = computeOutdatedModules(added, modified); const { disposedModules } = updateChunksPhase(chunksAdded, chunksDeleted); applyInternal(outdatedModules, disposedModules, newModuleFactories); } function applyInvalidatedModules(outdatedModules) { if (queuedInvalidatedModules.size > 0) { computedInvalidatedModules(queuedInvalidatedModules).forEach((moduleId)=>{ outdatedModules.add(moduleId); }); queuedInvalidatedModules.clear(); } return outdatedModules; } function applyInternal(outdatedModules, disposedModules, newModuleFactories) { outdatedModules = applyInvalidatedModules(outdatedModules); const outdatedSelfAcceptedModules = computeOutdatedSelfAcceptedModules(outdatedModules); const { outdatedModuleParents } = disposePhase(outdatedModules, disposedModules); // we want to continue on error and only throw the error after we tried applying all updates let error; function reportError(err) { if (!error) error = err; } applyPhase(outdatedSelfAcceptedModules, newModuleFactories, outdatedModuleParents, reportError); if (error) { throw error; } if (queuedInvalidatedModules.size > 0) { applyInternal(new Set(), [], new Map()); } } function computeChangedModules(entries, updates) { const chunksAdded = new Map(); const chunksDeleted = new Map(); const added = new Map(); const modified = new Map(); const deleted = new Set(); for (const [chunkPath, mergedChunkUpdate] of Object.entries(updates)){ switch(mergedChunkUpdate.type){ case "added": { const updateAdded = new Set(mergedChunkUpdate.modules); for (const moduleId of updateAdded){ added.set(moduleId, entries[moduleId]); } chunksAdded.set(chunkPath, updateAdded); break; } case "deleted": { // We could also use `mergedChunkUpdate.modules` here. const updateDeleted = new Set(chunkModulesMap.get(chunkPath)); for (const moduleId of updateDeleted){ deleted.add(moduleId); } chunksDeleted.set(chunkPath, updateDeleted); break; } case "partial": { const updateAdded = new Set(mergedChunkUpdate.added); const updateDeleted = new Set(mergedChunkUpdate.deleted); for (const moduleId of updateAdded){ added.set(moduleId, entries[moduleId]); } for (const moduleId of updateDeleted){ deleted.add(moduleId); } chunksAdded.set(chunkPath, updateAdded); chunksDeleted.set(chunkPath, updateDeleted); break; } default: invariant(mergedChunkUpdate, (mergedChunkUpdate)=>`Unknown merged chunk update type: ${mergedChunkUpdate.type}`); } } // If a module was added from one chunk and deleted from another in the same update, // consider it to be modified, as it means the module was moved from one chunk to another // AND has new code in a single update. for (const moduleId of added.keys()){ if (deleted.has(moduleId)) { added.delete(moduleId); deleted.delete(moduleId); } } for (const [moduleId, entry] of Object.entries(entries)){ // Modules that haven't been added to any chunk but have new code are considered // to be modified. // This needs to be under the previous loop, as we need it to get rid of modules // that were added and deleted in the same update. if (!added.has(moduleId)) { modified.set(moduleId, entry); } } return { added, deleted, modified, chunksAdded, chunksDeleted }; } function getAffectedModuleEffects(moduleId) { const outdatedModules = new Set(); const queue = [ { moduleId, dependencyChain: [] } ]; let nextItem; while(nextItem = queue.shift()){ const { moduleId, dependencyChain } = nextItem; if (moduleId != null) { if (outdatedModules.has(moduleId)) { continue; } outdatedModules.add(moduleId); } // We've arrived at the runtime of the chunk, which means that nothing // else above can accept this update. if (moduleId === undefined) { return { type: "unaccepted", dependencyChain }; } const module = devModuleCache[moduleId]; const hotState = moduleHotState.get(module); if (// The module is not in the cache. Since this is a "modified" update, // it means that the module was never instantiated before. !module || hotState.selfAccepted && !hotState.selfInvalidated) { continue; } if (hotState.selfDeclined) { return { type: "self-declined", dependencyChain, moduleId }; } if (runtimeModules.has(moduleId)) { queue.push({ moduleId: undefined, dependencyChain: [ ...dependencyChain, moduleId ] }); continue; } for (const parentId of module.parents){ const parent = devModuleCache[parentId]; if (!parent) { continue; } // TODO(alexkirsz) Dependencies: check accepted and declined // dependencies here. queue.push({ moduleId: parentId, dependencyChain: [ ...dependencyChain, moduleId ] }); } } return { type: "accepted", moduleId, outdatedModules }; } function handleApply(chunkListPath, update) { switch(update.type){ case "partial": { // This indicates that the update is can be applied to the current state of the application. applyUpdate(update.instruction); break; } case "restart": { // This indicates that there is no way to apply the update to the // current state of the application, and that the application must be // restarted. DEV_BACKEND.restart(); break; } case "notFound": { // This indicates that the chunk list no longer exists: either the dynamic import which created it was removed, // or the page itself was deleted. // If it is a dynamic import, we simply discard all modules that the chunk has exclusive access to. // If it is a runtime chunk list, we restart the application. if (runtimeChunkLists.has(chunkListPath)) { DEV_BACKEND.restart(); } else { disposeChunkList(chunkListPath); } break; } default: throw new Error(`Unknown update type: ${update.type}`); } } function createModuleHot(moduleId, hotData) { const hotState = { selfAccepted: false, selfDeclined: false, selfInvalidated: false, disposeHandlers: [] }; const hot = { // TODO(alexkirsz) This is not defined in the HMR API. It was used to // decide whether to warn whenever an HMR-disposed module required other // modules. We might want to remove it. active: true, data: hotData ?? {}, // TODO(alexkirsz) Support full (dep, callback, errorHandler) form. accept: (modules, _callback, _errorHandler)=>{ if (modules === undefined) { hotState.selfAccepted = true; } else if (typeof modules === "function") { hotState.selfAccepted = modules; } else { throw new Error("unsupported `accept` signature"); } }, decline: (dep)=>{ if (dep === undefined) { hotState.selfDeclined = true; } else { throw new Error("unsupported `decline` signature"); } }, dispose: (callback)=>{ hotState.disposeHandlers.push(callback); }, addDisposeHandler: (callback)=>{ hotState.disposeHandlers.push(callback); }, removeDisposeHandler: (callback)=>{ const idx = hotState.disposeHandlers.indexOf(callback); if (idx >= 0) { hotState.disposeHandlers.splice(idx, 1); } }, invalidate: ()=>{ hotState.selfInvalidated = true; queuedInvalidatedModules.add(moduleId); }, // NOTE(alexkirsz) This is part of the management API, which we don't // implement, but the Next.js React Refresh runtime uses this to decide // whether to schedule an update. status: ()=>"idle", // NOTE(alexkirsz) Since we always return "idle" for now, these are no-ops. addStatusHandler: (_handler)=>{}, removeStatusHandler: (_handler)=>{}, // NOTE(jridgewell) Check returns the list of updated modules, but we don't // want the webpack code paths to ever update (the turbopack paths handle // this already). check: ()=>Promise.resolve(null) }; return { hot, hotState }; } /** * Removes a module from a chunk. * Returns `true` if there are no remaining chunks including this module. */ function removeModuleFromChunk(moduleId, chunkPath) { const moduleChunks = moduleChunksMap.get(moduleId); moduleChunks.delete(chunkPath); const chunkModules = chunkModulesMap.get(chunkPath); chunkModules.delete(moduleId); const noRemainingModules = chunkModules.size === 0; if (noRemainingModules) { chunkModulesMap.delete(chunkPath); } const noRemainingChunks = moduleChunks.size === 0; if (noRemainingChunks) { moduleChunksMap.delete(moduleId); } return noRemainingChunks; } /** * Disposes of a chunk list and its corresponding exclusive chunks. */ function disposeChunkList(chunkListPath) { const chunkPaths = chunkListChunksMap.get(chunkListPath); if (chunkPaths == null) { return false; } chunkListChunksMap.delete(chunkListPath); for (const chunkPath of chunkPaths){ const chunkChunkLists = chunkChunkListsMap.get(chunkPath); chunkChunkLists.delete(chunkListPath); if (chunkChunkLists.size === 0) { chunkChunkListsMap.delete(chunkPath); disposeChunk(chunkPath); } } // We must also dispose of the chunk list's chunk itself to ensure it may // be reloaded properly in the future. const chunkListUrl = getChunkRelativeUrl(chunkListPath); DEV_BACKEND.unloadChunk?.(chunkListUrl); return true; } /** * Disposes of a chunk and its corresponding exclusive modules. * * @returns Whether the chunk was disposed of. */ function disposeChunk(chunkPath) { const chunkUrl = getChunkRelativeUrl(chunkPath); // This should happen whether the chunk has any modules in it or not. // For instance, CSS chunks have no modules in them, but they still need to be unloaded. DEV_BACKEND.unloadChunk?.(chunkUrl); const chunkModules = chunkModulesMap.get(chunkPath); if (chunkModules == null) { return false; } chunkModules.delete(chunkPath); for (const moduleId of chunkModules){ const moduleChunks = moduleChunksMap.get(moduleId); moduleChunks.delete(chunkPath); const noRemainingChunks = moduleChunks.size === 0; if (noRemainingChunks) { moduleChunksMap.delete(moduleId); disposeModule(moduleId, "clear"); availableModules.delete(moduleId); } } return true; } /** * Subscribes to chunk list updates from the update server and applies them. */ function registerChunkList(chunkList) { const chunkListScript = chunkList.script; const chunkListPath = getPathFromScript(chunkListScript); // The "chunk" is also registered to finish the loading in the backend BACKEND.registerChunk(chunkListPath); globalThis.TURBOPACK_CHUNK_UPDATE_LISTENERS.push([ chunkListPath, handleApply.bind(null, chunkListPath) ]); // Adding chunks to chunk lists and vice versa. const chunkPaths = new Set(chunkList.chunks.map(getChunkPath)); chunkListChunksMap.set(chunkListPath, chunkPaths); for (const chunkPath of chunkPaths){ let chunkChunkLists = chunkChunkListsMap.get(chunkPath); if (!chunkChunkLists) { chunkChunkLists = new Set([ chunkListPath ]); chunkChunkListsMap.set(chunkPath, chunkChunkLists); } else { chunkChunkLists.add(chunkListPath); } } if (chunkList.source === "entry") { markChunkListAsRuntime(chunkListPath); } } globalThis.TURBOPACK_CHUNK_UPDATE_LISTENERS ??= []; /** * This file contains the runtime code specific to the Turbopack development * ECMAScript DOM runtime. * * It will be appended to the base development runtime code. */ /* eslint-disable @typescript-eslint/no-unused-vars */ /// /// let BACKEND; function augmentContext(context) { return context; } function fetchWebAssembly(wasmChunkPath) { return fetch(getChunkRelativeUrl(wasmChunkPath)); } async function loadWebAssembly(_source, wasmChunkPath, importsObj) { const req = fetchWebAssembly(wasmChunkPath); const { instance } = await WebAssembly.instantiateStreaming(req, importsObj); return instance.exports; } async function loadWebAssemblyModule(_source, wasmChunkPath) { const req = fetchWebAssembly(wasmChunkPath); return await WebAssembly.compileStreaming(req); } /** * Maps chunk paths to the corresponding resolver. */ const chunkResolvers = new Map(); (()=>{ BACKEND = { async registerChunk (chunkPath, params) { const chunkUrl = getChunkRelativeUrl(chunkPath); const resolver = getOrCreateResolver(chunkUrl); resolver.resolve(); if (params == null) { return; } for (const otherChunkData of params.otherChunks){ const otherChunkPath = getChunkPath(otherChunkData); const otherChunkUrl = getChunkRelativeUrl(otherChunkPath); // Chunk might have started loading, so we want to avoid triggering another load. getOrCreateResolver(otherChunkUrl); } // This waits for chunks to be loaded, but also marks included items as available. await Promise.all(params.otherChunks.map((otherChunkData)=>loadChunk({ type: SourceType.Runtime, chunkPath }, otherChunkData))); if (params.runtimeModuleIds.length > 0) { for (const moduleId of params.runtimeModuleIds){ getOrInstantiateRuntimeModule(moduleId, chunkPath); } } }, /** * Loads the given chunk, and returns a promise that resolves once the chunk * has been loaded. */ loadChunk (chunkUrl, source) { return doLoadChunk(chunkUrl, source); } }; function getOrCreateResolver(chunkUrl) { let resolver = chunkResolvers.get(chunkUrl); if (!resolver) { let resolve; let reject; const promise = new Promise((innerResolve, innerReject)=>{ resolve = innerResolve; reject = innerReject; }); resolver = { resolved: false, promise, resolve: ()=>{ resolver.resolved = true; resolve(); }, reject: reject }; chunkResolvers.set(chunkUrl, resolver); } return resolver; } /** * Loads the given chunk, and returns a promise that resolves once the chunk * has been loaded. */ function doLoadChunk(chunkUrl, source) { const resolver = getOrCreateResolver(chunkUrl); if (resolver.resolved) { return resolver.promise; } if (source.type === SourceType.Runtime) { // We don't need to load chunks references from runtime code, as they're already // present in the DOM. if (isCss(chunkUrl)) { // CSS chunks do not register themselves, and as such must be marked as // loaded instantly. resolver.resolve(); } // We need to wait for JS chunks to register themselves within `registerChunk` // before we can start instantiating runtime modules, hence the absence of // `resolver.resolve()` in this branch. return resolver.promise; } if (typeof importScripts === "function") { // We're in a web worker if (isCss(chunkUrl)) { // ignore } else if (isJs(chunkUrl)) { self.TURBOPACK_NEXT_CHUNK_URLS.push(chunkUrl); importScripts(TURBOPACK_WORKER_LOCATION + chunkUrl); } else { throw new Error(`can't infer type of chunk from URL ${chunkUrl} in worker`); } } else { // TODO(PACK-2140): remove this once all filenames are guaranteed to be escaped. const decodedChunkUrl = decodeURI(chunkUrl); if (isCss(chunkUrl)) { const previousLinks = document.querySelectorAll(`link[rel=stylesheet][href="${chunkUrl}"],link[rel=stylesheet][href^="${chunkUrl}?"],link[rel=stylesheet][href="${decodedChunkUrl}"],link[rel=stylesheet][href^="${decodedChunkUrl}?"]`); if (previousLinks.length > 0) { // CSS chunks do not register themselves, and as such must be marked as // loaded instantly. resolver.resolve(); } else { const link = document.createElement("link"); link.rel = "stylesheet"; link.href = chunkUrl; link.onerror = ()=>{ resolver.reject(); }; link.onload = ()=>{ // CSS chunks do not register themselves, and as such must be marked as // loaded instantly. resolver.resolve(); }; document.body.appendChild(link); } } else if (isJs(chunkUrl)) { const previousScripts = document.querySelectorAll(`script[src="${chunkUrl}"],script[src^="${chunkUrl}?"],script[src="${decodedChunkUrl}"],script[src^="${decodedChunkUrl}?"]`); if (previousScripts.length > 0) { // There is this edge where the script already failed loading, but we // can't detect that. The Promise will never resolve in this case. for (const script of Array.from(previousScripts)){ script.addEventListener("error", ()=>{ resolver.reject(); }); } } else { const script = document.createElement("script"); script.src = chunkUrl; // We'll only mark the chunk as loaded once the script has been executed, // which happens in `registerChunk`. Hence the absence of `resolve()` in // this branch. script.onerror = ()=>{ resolver.reject(); }; document.body.appendChild(script); } } else { throw new Error(`can't infer type of chunk from URL ${chunkUrl}`); } } return resolver.promise; } })(); /** * This file contains the runtime code specific to the Turbopack development * ECMAScript DOM runtime. * * It will be appended to the base development runtime code. */ /* eslint-disable @typescript-eslint/no-unused-vars */ /// /// /// /// let DEV_BACKEND; (()=>{ DEV_BACKEND = { unloadChunk (chunkUrl) { deleteResolver(chunkUrl); // TODO(PACK-2140): remove this once all filenames are guaranteed to be escaped. const decodedChunkUrl = decodeURI(chunkUrl); if (isCss(chunkUrl)) { const links = document.querySelectorAll(`link[href="${chunkUrl}"],link[href^="${chunkUrl}?"],link[href="${decodedChunkUrl}"],link[href^="${decodedChunkUrl}?"]`); for (const link of Array.from(links)){ link.remove(); } } else if (isJs(chunkUrl)) { // Unloading a JS chunk would have no effect, as it lives in the JS // runtime once evaluated. // However, we still want to remove the script tag from the DOM to keep // the HTML somewhat consistent from the user's perspective. const scripts = document.querySelectorAll(`script[src="${chunkUrl}"],script[src^="${chunkUrl}?"],script[src="${decodedChunkUrl}"],script[src^="${decodedChunkUrl}?"]`); for (const script of Array.from(scripts)){ script.remove(); } } else { throw new Error(`can't infer type of chunk from URL ${chunkUrl}`); } }, reloadChunk (chunkUrl) { return new Promise((resolve, reject)=>{ if (!isCss(chunkUrl)) { reject(new Error("The DOM backend can only reload CSS chunks")); return; } const decodedChunkUrl = decodeURI(chunkUrl); const previousLinks = document.querySelectorAll(`link[rel=stylesheet][href="${chunkUrl}"],link[rel=stylesheet][href^="${chunkUrl}?"],link[rel=stylesheet][href="${decodedChunkUrl}"],link[rel=stylesheet][href^="${decodedChunkUrl}?"]`); if (previousLinks.length === 0) { reject(new Error(`No link element found for chunk ${chunkUrl}`)); return; } const link = document.createElement("link"); link.rel = "stylesheet"; if (navigator.userAgent.includes("Firefox")) { // Firefox won't reload CSS files that were previously loaded on the current page, // we need to add a query param to make sure CSS is actually reloaded from the server. // // I believe this is this issue: https://bugzilla.mozilla.org/show_bug.cgi?id=1037506 // // Safari has a similar issue, but only if you have a `` tag // pointing to the same URL as the stylesheet: https://bugs.webkit.org/show_bug.cgi?id=187726 link.href = `${chunkUrl}?ts=${Date.now()}`; } else { link.href = chunkUrl; } link.onerror = ()=>{ reject(); }; link.onload = ()=>{ // First load the new CSS, then remove the old ones. This prevents visible // flickering that would happen in-between removing the previous CSS and // loading the new one. for (const previousLink of Array.from(previousLinks))previousLink.remove(); // CSS chunks do not register themselves, and as such must be marked as // loaded instantly. resolve(); }; // Make sure to insert the new CSS right after the previous one, so that // its precedence is higher. previousLinks[0].parentElement.insertBefore(link, previousLinks[0].nextSibling); }); }, restart: ()=>self.location.reload() }; function deleteResolver(chunkUrl) { chunkResolvers.delete(chunkUrl); } })(); function _eval({ code, url, map }) { code += `\n\n//# sourceURL=${encodeURI(location.origin + CHUNK_BASE_PATH + url + CHUNK_SUFFIX_PATH)}`; if (map) { code += `\n//# sourceMappingURL=data:application/json;charset=utf-8;base64,${btoa(// btoa doesn't handle nonlatin characters, so escape them as \x sequences // See https://stackoverflow.com/a/26603875 unescape(encodeURIComponent(map)))}`; } // eslint-disable-next-line no-eval return eval(code); } const chunksToRegister = globalThis.TURBOPACK; globalThis.TURBOPACK = { push: registerChunk }; chunksToRegister.forEach(registerChunk); const chunkListsToRegister = globalThis.TURBOPACK_CHUNK_LISTS || []; chunkListsToRegister.forEach(registerChunkList); globalThis.TURBOPACK_CHUNK_LISTS = { push: registerChunkList }; })(); //# sourceMappingURL=_93808211._.js.map