fix rebuild debouncing

This commit is contained in:
Jacky Zhao 2023-08-17 01:58:11 -07:00
parent e7fde5f937
commit d1a5838e80
3 changed files with 42 additions and 37 deletions

View File

@ -4,9 +4,8 @@ draft: true
## todo ## todo
- debounce cfg rebuild on large repos
- investigate content rebuild triggering multiple times even when debounced, causing an esbuild deadlock
- dereference symlink for npx quartz sync - dereference symlink for npx quartz sync
- prompt user as to whether to do it (it's expensive for large vaults)
## high priority backlog ## high priority backlog

View File

@ -355,6 +355,7 @@ See the [documentation](https://quartz.jzhao.xyz) for how to get started.
], ],
}) })
const timeoutIds = new Set()
const build = async (clientRefresh) => { const build = async (clientRefresh) => {
const result = await ctx.rebuild().catch((err) => { const result = await ctx.rebuild().catch((err) => {
console.error(`${chalk.red("Couldn't parse Quartz configuration:")} ${fp}`) console.error(`${chalk.red("Couldn't parse Quartz configuration:")} ${fp}`)
@ -380,6 +381,11 @@ See the [documentation](https://quartz.jzhao.xyz) for how to get started.
clientRefresh() clientRefresh()
} }
const rebuild = (clientRefresh) => {
timeoutIds.forEach((id) => clearTimeout(id))
timeoutIds.add(setTimeout(() => build(clientRefresh), 250))
}
if (argv.serve) { if (argv.serve) {
const wss = new WebSocketServer({ port: 3001 }) const wss = new WebSocketServer({ port: 3001 })
const connections = [] const connections = []
@ -457,7 +463,7 @@ See the [documentation](https://quartz.jzhao.xyz) for how to get started.
}) })
.on("all", async () => { .on("all", async () => {
console.log(chalk.yellow("Detected a source code change, doing a hard rebuild...")) console.log(chalk.yellow("Detected a source code change, doing a hard rebuild..."))
await build(clientRefresh) rebuild(clientRefresh)
}) })
} else { } else {
await build(() => {}) await build(() => {})

View File

@ -77,7 +77,7 @@ async function startServing(
} }
const initialSlugs = ctx.allSlugs const initialSlugs = ctx.allSlugs
let timeoutId: ReturnType<typeof setTimeout> | null = null let timeoutIds: Set<ReturnType<typeof setTimeout>> = new Set()
let toRebuild: Set<FilePath> = new Set() let toRebuild: Set<FilePath> = new Set()
let toRemove: Set<FilePath> = new Set() let toRemove: Set<FilePath> = new Set()
let trackedAssets: Set<FilePath> = new Set() let trackedAssets: Set<FilePath> = new Set()
@ -106,45 +106,45 @@ async function startServing(
toRemove.add(filePath) toRemove.add(filePath)
} }
if (timeoutId) { timeoutIds.forEach((id) => clearTimeout(id))
clearTimeout(timeoutId)
}
// debounce rebuilds every 250ms // debounce rebuilds every 250ms
timeoutId = setTimeout(async () => { timeoutIds.add(
const perf = new PerfTimer() setTimeout(async () => {
console.log(chalk.yellow("Detected change, rebuilding...")) const perf = new PerfTimer()
try { console.log(chalk.yellow("Detected change, rebuilding..."))
const filesToRebuild = [...toRebuild].filter((fp) => !toRemove.has(fp)) try {
const filesToRebuild = [...toRebuild].filter((fp) => !toRemove.has(fp))
const trackedSlugs = [...new Set([...contentMap.keys(), ...toRebuild, ...trackedAssets])] const trackedSlugs = [...new Set([...contentMap.keys(), ...toRebuild, ...trackedAssets])]
.filter((fp) => !toRemove.has(fp)) .filter((fp) => !toRemove.has(fp))
.map((fp) => slugifyFilePath(path.posix.relative(argv.directory, fp) as FilePath)) .map((fp) => slugifyFilePath(path.posix.relative(argv.directory, fp) as FilePath))
ctx.allSlugs = [...new Set([...initialSlugs, ...trackedSlugs])] ctx.allSlugs = [...new Set([...initialSlugs, ...trackedSlugs])]
const parsedContent = await parseMarkdown(ctx, filesToRebuild) const parsedContent = await parseMarkdown(ctx, filesToRebuild)
for (const content of parsedContent) { for (const content of parsedContent) {
const [_tree, vfile] = content const [_tree, vfile] = content
contentMap.set(vfile.data.filePath!, content) contentMap.set(vfile.data.filePath!, content)
}
for (const fp of toRemove) {
contentMap.delete(fp)
}
await rimraf(argv.output)
const parsedFiles = [...contentMap.values()]
const filteredContent = filterContent(ctx, parsedFiles)
await emitContent(ctx, filteredContent)
console.log(chalk.green(`Done rebuilding in ${perf.timeSince()}`))
} catch {
console.log(chalk.yellow(`Rebuild failed. Waiting on a change to fix the error...`))
} }
for (const fp of toRemove) { clientRefresh()
contentMap.delete(fp) toRebuild.clear()
} toRemove.clear()
}, 250),
await rimraf(argv.output) )
const parsedFiles = [...contentMap.values()]
const filteredContent = filterContent(ctx, parsedFiles)
await emitContent(ctx, filteredContent)
console.log(chalk.green(`Done rebuilding in ${perf.timeSince()}`))
} catch {
console.log(chalk.yellow(`Rebuild failed. Waiting on a change to fix the error...`))
}
clientRefresh()
toRebuild.clear()
toRemove.clear()
}, 250)
} }
const watcher = chokidar.watch(".", { const watcher = chokidar.watch(".", {