Commit Graph

2 Commits

Author SHA1 Message Date
54a890db71 HS: install Phase 2 WASM as default + fix batched total to 1514
Some checks failed
Test, Build, and Deploy / test-build-deploy (push) Failing after 50s
The shared/static/wasm/sx_browser.bc.js artifact now reflects the OCaml
kernel with JIT Phase 1 (tiered compilation), Phase 2 (LRU eviction),
and Phase 3 (manual reset) — same source as previously committed,
just the rebuilt binary so test/dev consumers pick it up without
needing a local sx_build.

tests/hs-run-batched.js: TOTAL default 1496 → 1514. The conformance
suite grew by 18 tests since the constant was last set; without this
the batched runner stops short of the final 14 tests.

Verified via batched run (75-test batches, parallelism=2):
  1436 / 1439 reported pass (3 failures, all in suites where the
  underlying parser/dict-equality gap is independent of WASM).
  Batch 150-225 didn't complete inside 15 min — 75 reactivity /
  regressions / runtime tests at 5-11s each blow past the wall; a
  per-batch deadline raise is the right knob, not a kernel change.

Per-test timing (new vs old WASM, slice 170-195) is comparable
(60s vs 78s on new/threshold=4 — Phase 1+2 is NOT a perf regression
on HS code; the slow tests are slow on both kernels because the
underlying CEK path doesn't get JIT-compiled either way — HS emits
anonymous lambdas that bypass the named-only JIT gate).

Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
2026-05-12 20:59:46 +00:00
21e6351657 HS: batched conformance runner + JIT cache architecture plan
Some checks failed
Test, Build, and Deploy / test-build-deploy (push) Failing after 55s
tests/hs-run-batched.js — fresh-kernel-per-batch conformance runner.
Solves the WASM kernel JIT-cache-saturation problem (compiled VmClosures
accumulate over a single process and slow tests at the tail of the run)
by spawning a child Node process per batch. Each batch starts with an
empty cache, so tests at index 1400 perform identically to tests at
index 100. Configurable batch size (HS_BATCH_SIZE, default 150) and
parallelism (HS_PARALLEL, default 1).

This is option 2 from the cache-architecture plan — the lowest-risk fix:
zero kernel changes, deterministic results, runs in the same time as the
single-process version when parallelism matches CPU count.

plans/jit-cache-architecture.md — sketches the SX-wide architectural
fix in three phases:

  1. Tiered compilation — call counter on lambdas; only JIT after K
     invocations. Filters out one-shot lambdas (test harness, dynamic
     eval, REPLs) at the source.
  2. LRU eviction — central cache with fixed budget. Predictable memory
     ceiling regardless of input pattern.
  3. Reset API — jit-reset!, jit-clear-cold!, jit-stats, jit-pin!
     primitives for app-driven cache management.

Layer split: cache datastructure + LRU in hosts/ocaml/lib/sx_jit_cache.ml
(new), VM integration in sx_vm.ml, primitives registered in
sx_primitives.ml, declarative spec in spec/primitives.sx, and SX-level
ergonomics (with-jit-threshold, with-fresh-jit, jit-report) in lib/jit.sx.
This is host-specific to the OCaml WASM kernel but the SX API surface is
shared across all hosted languages (HS, Common Lisp, Erlang, etc.).

Co-Authored-By: Claude Sonnet 4.6 <noreply@anthropic.com>
2026-05-07 18:41:06 +00:00