Goal
Solver nodes execute actual JEPA training steps and produce verifiable weight deltas, instead of the deterministic-hash placeholder currently in place.
Current state
Solver code currently runs TOTAL_TRAINING_STEPS=20 of mocked training, producing fake hashes that flow through the commit-reveal loop. Real JEPA training functions exist in nodes/common/ml.py (and _train_backbone_jepa for the LLM-backbone path) but aren't called from the solver.
Scope
- Wire
nodes/solver/main.py to call the real training functions in ml.py based on the project's model configuration (vanilla JEPA, VL-JEPA, or LLM-backbone-JEPA).
- Use the dataset that the proposer published as the training input.
- Produce a real weight-delta blob; that delta becomes the solution CID for the commit-reveal cycle.
- Verifier path: aggregator should be able to apply the delta and confirm metric improvement (or whatever the project's correctness definition is).
Acceptance
- A small training run (single proposer, two solvers, one coordinator, one aggregator) produces real weight-delta blobs whose hashes change across runs with different RNG seeds.
- The aggregator's FedAvg over those deltas produces a globally improved model (loss decreases on a held-out batch).
Depends on
- Autonomous training path (already wired in
a991f22).
- LLM-backbone training path (already wired in
f950025).
Notes
- This unlocks honest training rewards. Until this lands,
recordTraining is rewarding fake work.
Goal
Solver nodes execute actual JEPA training steps and produce verifiable weight deltas, instead of the deterministic-hash placeholder currently in place.
Current state
Solver code currently runs
TOTAL_TRAINING_STEPS=20of mocked training, producing fake hashes that flow through the commit-reveal loop. Real JEPA training functions exist innodes/common/ml.py(and_train_backbone_jepafor the LLM-backbone path) but aren't called from the solver.Scope
nodes/solver/main.pyto call the real training functions inml.pybased on the project's model configuration (vanilla JEPA, VL-JEPA, or LLM-backbone-JEPA).Acceptance
Depends on
a991f22).f950025).Notes
recordTrainingis rewarding fake work.