mirror of
https://github.com/awizemann/scarf.git
synced 2026-05-10 10:36:35 +00:00
Merge pull request #90 from awizemann/fix/dashboard-debounce-v0.13
fix(dashboard): coalesce file-watcher fires + dedupe in-flight loads (v0.13)
This commit is contained in:
@@ -15,6 +15,35 @@ final class HermesFileWatcher {
|
|||||||
/// the project list changes.
|
/// the project list changes.
|
||||||
private var remoteProjectPaths: [String] = []
|
private var remoteProjectPaths: [String] = []
|
||||||
|
|
||||||
|
/// Coalescing timer for `lastChangeDate` ticks. v0.13 Hermes writes to
|
||||||
|
/// `state.db-wal` and rotating logs at ~10 Hz during gateway activity;
|
||||||
|
/// every observing view (`DashboardView`, `ProjectsView`,
|
||||||
|
/// `ProjectSessionsView`, half a dozen widgets) re-fires its `.onChange`
|
||||||
|
/// or `.task(id:)` on every tick, which stacked concurrent dashboard
|
||||||
|
/// loads on v0.13 hosts and tripped sqlite contention on the read-only
|
||||||
|
/// state.db handle. We coalesce to at most one tick per
|
||||||
|
/// `coalesceWindow` so a burst of FSEvents collapses into one observable
|
||||||
|
/// state mutation.
|
||||||
|
///
|
||||||
|
/// **Two limits, not one.** A pure trailing-debounce would starve under
|
||||||
|
/// sustained WAL writes — the timer would keep getting cancelled and
|
||||||
|
/// rescheduled, and a coincident `gateway_state.json` Start/Stop touch
|
||||||
|
/// would never propagate until WAL activity quieted down. So we publish
|
||||||
|
/// when EITHER (a) `coalesceWindow` of quiet has elapsed since the last
|
||||||
|
/// fire, OR (b) `maxWait` has elapsed since the first fire of the
|
||||||
|
/// current burst — whichever comes first. The max-wait guarantees a
|
||||||
|
/// floor of one observable mutation per `maxWait` even during sustained
|
||||||
|
/// activity. Numbers picked to keep the dashboard responsive on a
|
||||||
|
/// single `touch` while surviving v0.13's WAL-write storm.
|
||||||
|
private var pendingCoalesceTimer: DispatchWorkItem?
|
||||||
|
private var pendingTickDate: Date?
|
||||||
|
/// Wall-clock when the current burst began. Set on the first
|
||||||
|
/// `scheduleCoalescedTick` fire after a quiet window; cleared whenever
|
||||||
|
/// the timer fires. Drives the `maxWait` floor below.
|
||||||
|
private var burstStartDate: Date?
|
||||||
|
private static let coalesceWindow: TimeInterval = 0.5
|
||||||
|
private static let maxWait: TimeInterval = 1.5
|
||||||
|
|
||||||
let context: ServerContext
|
let context: ServerContext
|
||||||
private let transport: any ServerTransport
|
private let transport: any ServerTransport
|
||||||
|
|
||||||
@@ -92,12 +121,53 @@ final class HermesFileWatcher {
|
|||||||
for await _ in stream {
|
for await _ in stream {
|
||||||
ScarfMon.event(.transport, "mac.fileWatcher.remoteDelta", count: 1)
|
ScarfMon.event(.transport, "mac.fileWatcher.remoteDelta", count: 1)
|
||||||
await MainActor.run { [weak self] in
|
await MainActor.run { [weak self] in
|
||||||
self?.lastChangeDate = Date()
|
self?.scheduleCoalescedTick()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Coalesce a burst of FSEvents (or remote-poll deltas) into a single
|
||||||
|
/// `lastChangeDate` mutation. Two limits decide when the publish fires,
|
||||||
|
/// whichever comes first:
|
||||||
|
///
|
||||||
|
/// 1. **Quiet window**: `coalesceWindow` seconds have elapsed since the
|
||||||
|
/// last fire. Each new fire pushes this out — pure debounce shape.
|
||||||
|
/// 2. **Max wait**: `maxWait` seconds have elapsed since the FIRST fire
|
||||||
|
/// of the current burst. This bounds the latency floor under
|
||||||
|
/// sustained activity (v0.13's ~10 Hz WAL-write storm) so a
|
||||||
|
/// coincident `gateway_state.json` Start/Stop touch can't be starved
|
||||||
|
/// indefinitely behind a continuously-rescheduling debounce timer.
|
||||||
|
///
|
||||||
|
/// Runs on `.main` (the FSEvents queue and the remote-poll
|
||||||
|
/// MainActor.run) so observers see the publish on MainActor without a
|
||||||
|
/// hop. The work item self-clears `burstStartDate` when it fires so the
|
||||||
|
/// next burst starts a fresh max-wait window.
|
||||||
|
private func scheduleCoalescedTick() {
|
||||||
|
let now = Date()
|
||||||
|
pendingTickDate = now
|
||||||
|
if burstStartDate == nil {
|
||||||
|
burstStartDate = now
|
||||||
|
}
|
||||||
|
pendingCoalesceTimer?.cancel()
|
||||||
|
// Pick the deadline as the earlier of (a) `coalesceWindow` from now,
|
||||||
|
// and (b) `maxWait` from the burst start. The latter only matters
|
||||||
|
// when fires keep arriving faster than `coalesceWindow`; in the
|
||||||
|
// single-fire / quiet-burst case both reduce to the same value.
|
||||||
|
let quietDeadline = now.addingTimeInterval(Self.coalesceWindow)
|
||||||
|
let maxWaitDeadline = (burstStartDate ?? now).addingTimeInterval(Self.maxWait)
|
||||||
|
let firingDate = min(quietDeadline, maxWaitDeadline)
|
||||||
|
let delay = max(0, firingDate.timeIntervalSince(now))
|
||||||
|
let work = DispatchWorkItem { [weak self] in
|
||||||
|
guard let self, let date = self.pendingTickDate else { return }
|
||||||
|
self.pendingTickDate = nil
|
||||||
|
self.burstStartDate = nil
|
||||||
|
self.lastChangeDate = date
|
||||||
|
}
|
||||||
|
pendingCoalesceTimer = work
|
||||||
|
DispatchQueue.main.asyncAfter(deadline: .now() + delay, execute: work)
|
||||||
|
}
|
||||||
|
|
||||||
func stopWatching() {
|
func stopWatching() {
|
||||||
for source in coreSources + projectSources {
|
for source in coreSources + projectSources {
|
||||||
source.cancel()
|
source.cancel()
|
||||||
@@ -108,6 +178,10 @@ final class HermesFileWatcher {
|
|||||||
timer = nil
|
timer = nil
|
||||||
remotePollTask?.cancel()
|
remotePollTask?.cancel()
|
||||||
remotePollTask = nil
|
remotePollTask = nil
|
||||||
|
pendingCoalesceTimer?.cancel()
|
||||||
|
pendingCoalesceTimer = nil
|
||||||
|
pendingTickDate = nil
|
||||||
|
burstStartDate = nil
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Watch each project's `dashboard.json` AND its enclosing `.scarf/`
|
/// Watch each project's `dashboard.json` AND its enclosing `.scarf/`
|
||||||
@@ -162,7 +236,7 @@ final class HermesFileWatcher {
|
|||||||
// message persisted); high counts when nothing's happening
|
// message persisted); high counts when nothing's happening
|
||||||
// suggest a runaway watcher install.
|
// suggest a runaway watcher install.
|
||||||
ScarfMon.event(.transport, "mac.fileWatcher.localFire", count: 1)
|
ScarfMon.event(.transport, "mac.fileWatcher.localFire", count: 1)
|
||||||
self?.lastChangeDate = Date()
|
self?.scheduleCoalescedTick()
|
||||||
}
|
}
|
||||||
source.setCancelHandler {
|
source.setCancelHandler {
|
||||||
Darwin.close(fd)
|
Darwin.close(fd)
|
||||||
|
|||||||
@@ -7,6 +7,18 @@ final class DashboardViewModel {
|
|||||||
private let dataService: HermesDataService
|
private let dataService: HermesDataService
|
||||||
private let fileService: HermesFileService
|
private let fileService: HermesFileService
|
||||||
|
|
||||||
|
/// Single in-flight load handle. The `.onChange(fileWatcher.lastChangeDate)`
|
||||||
|
/// observer in `DashboardView` plus `.task` on first appear can both
|
||||||
|
/// fire concurrent loads — and on v0.13 hosts the FSEvents tick rate
|
||||||
|
/// during gateway activity used to be high enough that 5+ loads
|
||||||
|
/// stacked inside 200 ms (HermesFileWatcher's coalesce window now
|
||||||
|
/// handles that, but defending here keeps the behaviour deterministic
|
||||||
|
/// on any future watcher chattiness). When a load is in flight,
|
||||||
|
/// subsequent triggers no-op; the in-flight load already has a
|
||||||
|
/// recent-enough snapshot for the user.
|
||||||
|
@ObservationIgnored
|
||||||
|
private var inFlightLoad: Task<Void, Never>?
|
||||||
|
|
||||||
init(context: ServerContext = .local) {
|
init(context: ServerContext = .local) {
|
||||||
self.context = context
|
self.context = context
|
||||||
self.dataService = HermesDataService(context: context)
|
self.dataService = HermesDataService(context: context)
|
||||||
@@ -42,6 +54,27 @@ final class DashboardViewModel {
|
|||||||
var hermesShadows: [ProjectHermesShadowDetector.Shadow] = []
|
var hermesShadows: [ProjectHermesShadowDetector.Shadow] = []
|
||||||
|
|
||||||
func load() async {
|
func load() async {
|
||||||
|
// Coalesce overlapping triggers: the `.task` first-appear and the
|
||||||
|
// `.onChange(fileWatcher.lastChangeDate)` observer can both fire
|
||||||
|
// a load in the same tick. Without this guard a Hermes v0.13
|
||||||
|
// host's WAL-write storm walked over the previous load
|
||||||
|
// mid-snapshot (see `HermesFileWatcher.scheduleCoalescedTick`).
|
||||||
|
// If a load is already running, await its completion and return
|
||||||
|
// — the caller already has a fresh snapshot by the time `await`
|
||||||
|
// returns.
|
||||||
|
if let existing = inFlightLoad {
|
||||||
|
await existing.value
|
||||||
|
return
|
||||||
|
}
|
||||||
|
let task: Task<Void, Never> = Task { @MainActor [weak self] in
|
||||||
|
await self?.loadImpl()
|
||||||
|
}
|
||||||
|
inFlightLoad = task
|
||||||
|
await task.value
|
||||||
|
inFlightLoad = nil
|
||||||
|
}
|
||||||
|
|
||||||
|
private func loadImpl() async {
|
||||||
isLoading = true
|
isLoading = true
|
||||||
// refresh() is essentially free for the streaming remote backend
|
// refresh() is essentially free for the streaming remote backend
|
||||||
// (no transfer — every query is fresh) and a cheap reopen for
|
// (no transfer — every query is fresh) and a cheap reopen for
|
||||||
|
|||||||
Reference in New Issue
Block a user