Skip to content

fix: real benchmarks, Cypher multi-row fix, and honest README#293

Open
aepod wants to merge 2343 commits intoruvnet:mainfrom
weave-logic-ai:fix/real-benchmarks
Open

fix: real benchmarks, Cypher multi-row fix, and honest README#293
aepod wants to merge 2343 commits intoruvnet:mainfrom
weave-logic-ai:fix/real-benchmarks

Conversation

@aepod
Copy link
Copy Markdown

@aepod aepod commented Mar 24, 2026

Summary

Cypher Fix (Issue #269)

The execute_match() function collapsed all match results into a single HashMap — each context.bind() call overwrote the previous value. Three matches → only the last survived.

Fix: Implemented proper ResultSet pipeline (Vec<ExecutionContext>) threaded through all statement executors. Cross-product expansion for multiple patterns in one MATCH.

Test results: 30 passed, 0 failed, 2 ignored

Key tests:

  • test_match_returns_multiple_rows — 3 nodes → 3 rows (was returning 1)
  • test_match_return_properties — Alice + Bob both returned correctly
  • test_match_where_filter — WHERE correctly filters multi-row results
  • test_match_many_nodes — 100 nodes → 100 rows

Also fixed column name generation for property expressions (n.name instead of ?column?).

Real Benchmark Results

All measurements are real — recall measured against brute-force ground truth, no simulated competitors.

10,000 Vectors (128d, M=32, ef=200)

Engine QPS Recall@10 Build (s) p50 (ms)
hnswlib (C++) 1152.6 0.9895 7.5 0.73
ruvector-core 443.1 0.9830 44.0 1.98
numpy brute-force 134.8 1.0000 0.003 3.26

100,000 Vectors (128d, M=32, ef=200)

Engine QPS Recall@10 Build (s) p50 (ms) Memory
hnswlib (C++) 249.5 0.7427 395.3 2.57
ruvector-core 85.7 0.8675 855.6 10.14 ~523MB
numpy brute-force 69.2 1.0000 0.016 10.20 48.8MB

Key Findings

  • Recall is competitive: 98.3% at 10K (within 0.65% of hnswlib), 86.75% at 100K (higher than hnswlib's 74.27%)
  • QPS is 2.6-2.9x slower than hnswlib — expected since hnswlib has SIMD optimizations
  • Build time is 2.2-5.9x slower — the largest gap, narrows at scale
  • Previous benchmarks had hardcoded recall=1.0 and memory=0.0 — now corrected

README Corrections

  • Replaced "100% recall at every configuration" with actual measured recall ranges
  • Replaced "15.7x faster than Python" (simulated) with real hnswlib comparison
  • Added independent benchmark section with link to full methodology

Test plan

  • All 30 Cypher tests pass (0 failures, 2 ignored)
  • Multi-row MATCH returns correct number of rows
  • Property expressions generate correct column names
  • WHERE filtering works on multi-row results
  • No regressions in existing test suite
  • Benchmarks reproducible with included scripts and raw JSON data

🤖 Generated with claude-flow

ruvnet and others added 30 commits March 3, 2026 13:21
…ruvnet#230)

HNSW k-NN fix:
- Search beam width (k) increased from 10 to 100 — previous value
  starved the beam search, causing 0 rows on index scan
- Added ruvector_hnsw_debug() diagnostic function for troubleshooting
- Added warning log when entry_point is InvalidBlockNumber

Hybrid search fix:
- ruvector_hybrid_search() now returns success=true with empty results
  and helpful message on unregistered collections (was success=false)

Audit script fix:
- Corrected hybrid_search argument order in sql-audit-v3.sql Section 9b
- Added HNSW debug diagnostics on 0-row failure

Results: 17 PASS / 0 PARTIAL / 0 FAIL → 100% (up from 88%)
Published: docker.io/ruvnet/ruvector-postgres:0.3.2
  Built from commit 5f0a2c2

  Platforms updated:
  - linux-x64-gnu
  - linux-arm64-gnu
  - darwin-x64
  - darwin-arm64
  - win32-x64-msvc

  🤖 Generated by GitHub Actions
…net#231)

- MCP entry line count: ~3,816 → 3,815 (verified with wc -l)
- Command groups: 14 → 15 (midstream group was missed)
- CLI test count: 63 → 64 active tests (verified grep -c)
- Dead code → conditionally unreachable (line 1807 runs when @ruvector/router installed)
  Built from commit 2bcc7ad

  Platforms updated:
  - linux-x64-gnu
  - linux-arm64-gnu
  - darwin-x64
  - darwin-arm64
  - win32-x64-msvc

  🤖 Generated by GitHub Actions
…le Firestore persistence (ruvnet#232)

ADR file renames:
- ADR-0027 → ADR-027 (fix 4-digit numbering to standard 3-digit)
- ADR-040 filename sanitized (removed spaces, em dash, ampersand)
- ADR-017 duplicate (craftsman) → ADR-024 (temporal-tensor keeps 017)
- ADR-029 duplicate (exo-ai) → ADR-025 (rvf-canonical keeps 029)
- ADR-031 duplicate (rvcow) → ADR-026 (rvf-example keeps 031)

Cloud Run fix (pi.ruv.io):
- Added FIRESTORE_URL env var — enables persistent storage
- Fixed env var packing bug (all flags were in BRAIN_SYSTEM_KEY)
- Dashboard now shows actual data: 240 memories, 30 contributors, 1096 edges
…brain dependency (ruvnet#233)

Replace requirePiBrain() + PiBrainClient with direct fetch() calls to pi.ruv.io.
All 13 brain CLI commands and 11 brain MCP tools now work out of the box with
zero extra dependencies. Includes 30s timeout on all brain API calls.
Brain commands now use direct pi.ruv.io fetch (PR ruvnet#233), so
@ruvector/pi-brain is no longer needed as a peer dependency.

Co-Authored-By: claude-flow <ruv@ruv.net>
  Built from commit 0b054f4

  Platforms updated:
  - linux-x64-gnu
  - linux-arm64-gnu
  - darwin-x64
  - darwin-arm64
  - win32-x64-msvc

  🤖 Generated by GitHub Actions
…uvnet#234)

* feat: proxy-aware fetch + brain API improvements — publish v0.2.7

Add proxyFetch() wrapper to cli.js and mcp-server.js that detects
HTTPS_PROXY/HTTP_PROXY/ALL_PROXY env vars, uses undici ProxyAgent
(Node 18+) or falls back to curl. Handles NO_PROXY patterns.
Replaced all 17 fetch() call sites with timeouts (15-30s).

Brain server API:
- Search returns similarity scores via ScoredBrainMemory
- List supports pagination (offset/limit), sorting (updated_at/quality/votes), tag filtering
- Transfer response includes warnings, source/target memory counts
- New POST /v1/verify endpoint with 4 verification methods

Co-Authored-By: claude-flow <ruv@ruv.net>

* feat: brain server bug fixes, GET /v1/pages, 9 MCP page/node tools — v0.2.10

Fix proxyFetch curl fallback to capture real HTTP status instead of
hardcoding 200, add 204 guards to brainFetch/fetchBrainEndpoint/MCP
handler, fix brain_list schema (missing offset/sort/tags), fix
brain_sync direction passthrough, add --json to share/vote/delete/sync.

Add GET /v1/pages route with pagination, status filter, sort.
Add 9 MCP tools: brain_page_list/get/create/update/delete,
brain_node_list/get/publish/revoke (previously SSE-only).

Polish: delete --json returns {deleted:true,id} not {}, page get
unwraps .memory wrapper for formatted display.

112 MCP tools, 69/69 tests pass. Published v0.2.10 to npm.

Co-Authored-By: claude-flow <ruv@ruv.net>
  Built from commit 3208afa

  Platforms updated:
  - linux-x64-gnu
  - linux-arm64-gnu
  - darwin-x64
  - darwin-arm64
  - win32-x64-msvc

  🤖 Generated by GitHub Actions
…-Sybil votes (ruvnet#235)

Expand PiiStripper from 12 to 15 regex rules: add phone number,
SSN, and credit card detection/redaction. Add IP-based rate limiting
(1500 writes/hr per IP) to prevent Sybil key rotation bypass. Add
per-IP vote deduplication (one vote per IP per memory) to prevent
quality score manipulation.

63 server tests + 16 PII tests pass. Deployed to Cloud Run.
  Built from commit 5d51e0b

  Platforms updated:
  - linux-x64-gnu
  - linux-arm64-gnu
  - darwin-x64
  - darwin-arm64
  - win32-x64-msvc

  🤖 Generated by GitHub Actions
…, CLI + MCP (ruvnet#236)

Bridge the gap between "stores knowledge" and "learns from knowledge":

- Background training loop (tokio::spawn, 5 min interval) runs SONA
  force_learn + domain evolve_population when new data arrives
- POST /v1/train endpoint for on-demand training cycles
- `ruvector brain train` CLI command with --json support
- `brain_train` MCP tool for agent-triggered training
- Vote dedup: 24h TTL on ip_votes entries, author exemption from IP check
- ADR-082 updated, ADR-083 created

Results: Pareto frontier grew 0→24 after 3 cycles. SONA activates
after 100+ trajectory threshold (natural search/share usage).

Publish ruvector@0.2.11.
  Built from commit 27401ff

  Platforms updated:
  - linux-x64-gnu
  - linux-arm64-gnu
  - darwin-x64
  - darwin-arm64
  - win32-x64-msvc

  🤖 Generated by GitHub Actions
- ONNX embeddings: dynamic dimension detection + conditional token_type_ids (ruvnet#237)
- rvf-node: add compression field pass-through to Rust N-API struct (ruvnet#225)
- Cargo workspace: add glob excludes for nested rvf sub-packages (ruvnet#214)
- ruvllm: fix stats crash (null guard + try/catch) + generate warning (ruvnet#103)
- ruvllm-wasm: deprecated placeholder on npm (ruvnet#238)
- Pre-existing: fix ruvector-sparse-inference-wasm API mismatch, exclude from workspace
- Pre-existing: fix ruvector-cloudrun-gpu RuvectorLayer::new() Result handling

Co-Authored-By: claude-flow <ruv@ruv.net>
fix: resolve 5 P0 critical issues + pre-existing compile errors
Co-Authored-By: claude-flow <ruv@ruv.net>
Built from commit 538237b

Platforms: linux-x64-gnu, linux-arm64-gnu, darwin-x64, darwin-arm64, win32-x64-msvc

Co-Authored-By: claude-flow <ruv@ruv.net>
  Built from commit 538237b

  Platforms updated:
  - linux-x64-gnu
  - linux-arm64-gnu
  - darwin-x64
  - darwin-arm64
  - win32-x64-msvc

  🤖 Generated by GitHub Actions
  Built from commit 9dc76e4

  Platforms updated:
  - linux-x64-gnu
  - linux-arm64-gnu
  - darwin-x64
  - darwin-arm64
  - win32-x64-msvc

  🤖 Generated by GitHub Actions
- Gate WebGPU web-sys features behind `webgpu` Cargo feature flag
- Remove unused bytemuck, gpu_map_mode, GpuSupportedLimits dependencies
- Add wasm-opt=false workaround for Rust 1.91 codegen bug
- Published @ruvector/ruvllm-wasm@2.0.0 with compiled WASM binary (435KB)
- ADR-084 documenting build workarounds and known limitations

Closes ruvnet#240

Co-Authored-By: claude-flow <ruv@ruv.net>
feat: ruvllm-wasm v2.0.0 — first functional WASM publish
…npm link

- Fix browser code example to use actual working API (ChatTemplateWasm, HnswRouterWasm)
- Add npm install line for @ruvector/ruvllm-wasm
- Update npm packages count (4→5) with ruvllm-wasm link
- Update WASM size to actual 435KB (178KB gzipped)
- Link ruvllm-wasm feature table to npm package

Co-Authored-By: claude-flow <ruv@ruv.net>
  Built from commit 0f9f55b

  Platforms updated:
  - linux-x64-gnu
  - linux-arm64-gnu
  - darwin-x64
  - darwin-arm64
  - win32-x64-msvc

  🤖 Generated by GitHub Actions
  Built from commit abb324e

  Platforms updated:
  - linux-x64-gnu
  - linux-arm64-gnu
  - darwin-x64
  - darwin-arm64
  - win32-x64-msvc

  🤖 Generated by GitHub Actions
Replaces outdated README that referenced non-existent APIs
(load_model_from_url, generate_stream) with documentation
matching the actual v2.0.0 exports.

Co-Authored-By: claude-flow <ruv@ruv.net>
  Built from commit 1f68d0a

  Platforms updated:
  - linux-x64-gnu
  - linux-arm64-gnu
  - darwin-x64
  - darwin-arm64
  - win32-x64-msvc

  🤖 Generated by GitHub Actions
ADR-084 defines the RuVector-native Neural Trader architecture using
dynamic market graphs, mincut coherence gating, and proof-gated mutation.
Includes three starter crates (neural-trader-core, neural-trader-coherence,
neural-trader-replay) with canonical types, threshold gate, reservoir
memory store, and 10 passing tests.

https://claude.ai/code/session_01EExDkEDv4eejvfgqUWnSks
ADR:
- Add SQL indexes on (symbol_id, ts_ns) for all tables
- Add HNSW index on nt_embeddings.embedding
- Range-partition nt_event_log and nt_segments by timestamp
- Add retention config (hot/warm/cold TTL) to example YAML
- Add retrieval weight normalization constraint (α+β+γ+δ=1)
- Cross-reference existing examples/neural-trader/

Code:
- core: Replace String property keys with PropertyKey enum (zero alloc)
- core: Add PartialEq on MarketEvent for test assertions
- coherence: Fix redundant drift check — learning now requires half
  drift margin (stricter than act/write)
- coherence: Add boundary_stable_count to GateContext and enforce
  boundary stability window threshold from ADR gate policy
- coherence: Add PartialEq on CoherenceDecision
- coherence: Add 2 new tests (high_drift, boundary_instability)
- replay: Switch ReservoirStore from Vec to VecDeque for O(1) eviction
- replay: Use RegimeLabel enum instead of Option<String> in MemoryQuery

12 tests pass (was 10).

https://claude.ai/code/session_01EExDkEDv4eejvfgqUWnSks
- Rename ADR-084-neural-trader to ADR-085 (ADR-084 is taken by ruvllm-wasm-publish)
- Move serde_json to dev-dependencies in neural-trader-core (only used in tests)
- Remove unused neural-trader-core dependency from neural-trader-coherence

Co-Authored-By: claude-flow <ruv@ruv.net>
ruvnet and others added 28 commits March 23, 2026 19:11
…vnet#287)

* Add ADR-117: pseudo-deterministic canonical minimum cut

Introduces source-anchored canonical min-cut based on Kenneth-Mordoch 2026,
with lexicographic tie-breaking (λ, first_separable_vertex, |S|, π(S)) for
unique reproducible cuts. Three-tier plan: exact engine now, O(m log²n) fast
path, then dynamic maintenance via sparsifiers. Integrates with RVF witness
hashing for cut receipts.

https://claude.ai/code/session_01UrVLJpxq8itzVxycy5sjNw

* Implement ADR-117: source-anchored pseudo-deterministic canonical min-cut

Full Tier 1 implementation of the Kenneth-Mordoch 2026 canonical min-cut
algorithm with lexicographic tie-breaking (λ, first_separable_vertex, |S|, π(S)).

Core implementation (source_anchored/mod.rs):
- AdjSnapshot for deterministic computation on FixedWeight (32.32)
- Stoer-Wagner global min-cut on fixed-point weights
- Dinic's max-flow for exact s-t cuts
- SHA-256 (FIPS 180-4, self-contained, no_std compatible)
- SourceAnchoredMinCut stateful wrapper with cache invalidation
- CanonicalMinCutResult repr(C) struct for FFI

WASM bindings (wasm/canonical.rs):
- Thread-safe Mutex-guarded global state (no static mut)
- 8 extern "C" functions: init, add_edge, compute, get_result,
  get_hash, get_side, get_cut_edges, free, hashes_equal
- Constant-time hash comparison for timing side-channel prevention
- Null pointer validation on all FFI entry points
- Graph size limit (10,000 vertices) to prevent OOM

Tests (40 total):
- 33 source_anchored tests: SHA-256 NIST vectors, determinism (100+1000
  iterations), symmetric graphs (K4, K5, cycles, ladders, barbells),
  custom source/priorities, disconnected rejection, FFI conversion
- 7 WASM tests: init/compute lifecycle, null safety, hash comparison,
  self-loop rejection, size limit enforcement

Benchmarks (canonical_bench.rs):
- Random connected graphs (10-100 vertices)
- Cycle and complete graph families
- Hash stability measurement

Security hardening:
- No static mut (Mutex for thread safety)
- Integer-exact FixedWeight arithmetic (no floats in comparisons)
- Checked capacity perturbation bounds
- Source-side orientation invariant enforced
- NIST-validated SHA-256 for witness hashes

ADR-117 updated to production-quality spec with explicit vertex-splitting
requirement for capacity perturbation, WASM FFI documentation, and
Phase 1 completion status.

https://claude.ai/code/session_01UrVLJpxq8itzVxycy5sjNw

* Integrate ADR-117 canonical min-cut into pi.ruv.io brain server

- Enable `canonical` feature on ruvector-mincut dependency
- Add `partition_canonical_full()` to KnowledgeGraph using source-anchored
  canonical min-cut for deterministic, hashable partitions
- Add `canonical` query parameter to `/v1/partition` endpoint
- Add `cut_hash` (hex SHA-256) and `first_separable_vertex` fields to
  PartitionResult and PartitionResultCompact types
- Backward compatible: canonical fields are skip_serializing_if None,
  only populated when `?canonical=true` is passed

https://claude.ai/code/session_01UrVLJpxq8itzVxycy5sjNw

---------

Co-authored-by: Claude <noreply@anthropic.com>
  Built from commit 10eb9e0

  Platforms updated:
  - linux-x64-gnu
  - linux-arm64-gnu
  - darwin-x64
  - darwin-arm64
  - win32-x64-msvc

  🤖 Generated by GitHub Actions
  Built from commit 6b6fb0e

  Platforms updated:
  - linux-x64-gnu
  - linux-arm64-gnu
  - darwin-x64
  - darwin-arm64
  - win32-x64-msvc

  🤖 Generated by GitHub Actions
…#289)

- Add cached_partition field to AppState for storing MinCut results
- Populate cache during enhanced training cycle (step 3c)
- REST /v1/partition returns cache if available (bypass with ?force=true)
- MCP brain_partition returns cached compact partition instead of stub
- Canonical MinCut benchmarks: sub-3us for graphs up to 50 nodes
  Built from commit 79c353d

  Platforms updated:
  - linux-x64-gnu
  - linux-arm64-gnu
  - darwin-x64
  - darwin-arm64
  - win32-x64-msvc

  🤖 Generated by GitHub Actions
)

Skip exact MinCut during training for graphs >100K edges to avoid
Cloud Run timeout. Cache populated by async scheduled jobs instead.
  Built from commit ae67b59

  Platforms updated:
  - linux-x64-gnu
  - linux-arm64-gnu
  - darwin-x64
  - darwin-arm64
  - win32-x64-msvc

  🤖 Generated by GitHub Actions
Tier 2 — Tree Packing Fast Path:
- Gomory-Hu flow-equivalent tree via Gusfield's algorithm
- Global MinCut from tree in O(V) after O(V * T_maxflow) construction
- canonical_mincut_fast() integration entry point
- 14 unit tests including Stoer-Wagner correctness validation

Tier 3 — Dynamic/Incremental MinCut:
- DynamicMinCut struct with epoch-based mutation tracking
- add_edge(): skip recompute if edge doesn't cross current cut
- remove_edge(): skip recompute if edge not in cut set
- apply_batch(): bulk mutations with deferred recomputation
- Staleness detection with configurable threshold
- HashSet caches for O(1) cut-crossing checks
- 19 unit tests including 100-run determinism check

WASM FFI: dynamic_init/add_edge/remove_edge/compute/epoch/free
Benchmarks: tree_packing_vs_stoer_wagner, dynamic_add_edge, dynamic_batch

98 canonical tests pass, 12 WASM tests pass.
  Built from commit 34b56e4

  Platforms updated:
  - linux-x64-gnu
  - linux-arm64-gnu
  - darwin-x64
  - darwin-arm64
  - win32-x64-msvc

  🤖 Generated by GitHub Actions
Measured on pi.ruv.io (2,110 nodes, 992K edges):
- brain_partition MCP: >60s timeout → 459ms (>130x)
- Partition REST cached: <1ms (>300,000x)
- Enhanced training: 504 timeout → 127ms
- 110 tests pass across all tiers

Co-Authored-By: claude-flow <ruv@ruv.net>
  Built from commit 3ecba7c

  Platforms updated:
  - linux-x64-gnu
  - linux-arm64-gnu
  - darwin-x64
  - darwin-arm64
  - win32-x64-msvc

  🤖 Generated by GitHub Actions
Optimizations:
- Flat Vec<FixedWeight> (n*n) replaces Vec<Vec<...>> in Dinic's max-flow
  and Gomory-Hu tree — single memcpy vs N heap allocations per st-cut
- Reuse BFS queue/level/iter arrays across Dinic's phases
- Swap-remove in Stoer-Wagner active_list — O(1) vs O(n) retain
- Fix benchmark compilation errors in optimization_bench.rs

Results (all 26 benchmarks improved, Criterion p < 0.05):
- Tree packing: up to -29.7% (deep clone elimination)
- Source-anchored: -10% to -24% (cache locality)
- Hash stability: -24.2%
- Dynamic incremental: ~unchanged (wrapper-dominated)

Co-Authored-By: claude-flow <ruv@ruv.net>
  Built from commit 79165e4

  Platforms updated:
  - linux-x64-gnu
  - linux-arm64-gnu
  - darwin-x64
  - darwin-arm64
  - win32-x64-msvc

  🤖 Generated by GitHub Actions
…drift

Gap 1 - Vote coverage (47%→improving):
  Auto-upvote under-observed memories based on content quality heuristics
  (title>10, content>50, has tags). Capped at 50/cycle.

Gap 2 - SONA trajectory diversity:
  Record SONA steps for brain_share/search/vote MCP tool calls.
  Only end trajectories when results >= 3 (avoid trivial single-step).

Gap 3 - Drift detection:
  Record search query embeddings as drift signal in search_memories().
  Drift CV metric now accumulates real data from user queries.

Knowledge velocity confirmed working (temporal_deltas pipeline active).

Co-Authored-By: claude-flow <ruv@ruv.net>
  Built from commit 70effc8

  Platforms updated:
  - linux-x64-gnu
  - linux-arm64-gnu
  - darwin-x64
  - darwin-arm64
  - win32-x64-msvc

  🤖 Generated by GitHub Actions
…tive SONA

Self-Reflective Training (Step 6):
- Knowledge imbalance detection (>40% in one category)
- Dynamic SONA threshold adaptation (lower on 0 patterns, raise on success)
- Vote coverage monitoring with auto-correction

Curiosity Feedback Loop (Step 7):
- Stagnation detection via delta_stream
- Auto-generates synthesis memories for under-represented categories
- Creates self-sustaining knowledge velocity

Auto-Reflection Memory (Step 8):
- Brain writes searchable self-reflections after each training cycle
- Persistent learning history enables meta-cognitive search

Symbolic Inference Engine:
- Forward-chaining Horn clause resolution with chain linking
- Transitive inference across propositions
- Self-loop prevention, confidence filtering
- 3 new tests passing

SONA Threshold Optimization:
- min_trajectories: 100→10 (primary blocker)
- k_clusters: 50→5, min_cluster_size: 2→1
- quality_threshold: 0.3→0.15
- Added runtime set_quality_threshold() API

Co-Authored-By: claude-flow <ruv@ruv.net>
  Built from commit 72e5ab6

  Platforms updated:
  - linux-x64-gnu
  - linux-arm64-gnu
  - darwin-x64
  - darwin-arm64
  - win32-x64-msvc

  🤖 Generated by GitHub Actions
Before → After (single session):
- Votes: 995 (47%) → 1,393 (65.2%)
- Knowledge velocity: 0 → 423
- Drift: no_data → drifting (active)
- GWT: 86% → 100%
- Memories: 2,112 → 2,137 (+25 diverse)
- Cross-domain transfers: 56/56 successful

Co-Authored-By: claude-flow <ruv@ruv.net>
  Built from commit a6b95a7

  Platforms updated:
  - linux-x64-gnu
  - linux-arm64-gnu
  - darwin-x64
  - darwin-arm64
  - win32-x64-msvc

  🤖 Generated by GitHub Actions
…ecall, LoRA auto-submit

Sparsified MinCut (59x speedup):
- partition_via_mincut_full uses 19K sparsified edges instead of 1M
- Large-graph guard now uses sparsifier instead of skipping

Cognitive integration:
- Hopfield recall_k wired into search scoring (0.10 boost)
- Associative memory now contributes to result ranking

LoRA federation unblocked:
- Auto-submit weight deltas from SONA's 436 patterns
- min_submissions lowered from 3 to 1 for bootstrapping

Strange loop in training:
- Invoked during training cycle, scores quality/relevance
- Recommends actions when quality is low

Symbolic inference fix:
- Shared-argument fallback for cross-cluster derivation
- Case-insensitive predicate matching

Auto-vote cap: 50→200 (4x faster coverage convergence)

Co-Authored-By: claude-flow <ruv@ruv.net>
  Built from commit bd385c9

  Platforms updated:
  - linux-x64-gnu
  - linux-arm64-gnu
  - darwin-x64
  - darwin-arm64
  - win32-x64-msvc

  🤖 Generated by GitHub Actions
Sparsifier build on 1M+ edges exceeds Cloud Run's 4-min startup probe.
Skip on startup for graphs > 100K edges, defer to rebuild_graph job.

Co-Authored-By: claude-flow <ruv@ruv.net>
The execute_match() function previously collapsed all match results into
a single ExecutionContext via context.bind(), which overwrote previous
bindings. MATCH (n:Person) on 3 Person nodes returned only 1 row.

This commit refactors the executor to use a ResultSet pipeline:
- type ResultSet = Vec<ExecutionContext>
- Each clause transforms ResultSet → ResultSet
- execute_match() expands the set (one context per match)
- execute_return() projects one row per context
- execute_set/delete() apply to all contexts
- Cross-product semantics for multiple patterns in one MATCH

Also adds comprehensive tests:
- test_match_returns_multiple_rows (the Issue ruvnet#269 regression)
- test_match_return_properties (verify correct values per row)
- test_match_where_filter (WHERE correctly filters multi-row)
- test_match_single_result (1 match → 1 row, no regression)
- test_match_no_results (0 matches → 0 rows)
- test_match_many_nodes (100 nodes → 100 rows, stress test)

Co-Authored-By: claude-flow <ruv@ruv.net>
RETURN n.name now produces column "n.name" instead of "?column?".
Property expressions (Expression::Property) are formatted as
"object.property" for column naming, matching standard Cypher behavior.

Co-Authored-By: claude-flow <ruv@ruv.net>
  Built from commit b2347ce

  Platforms updated:
  - linux-x64-gnu
  - linux-arm64-gnu
  - darwin-x64
  - darwin-arm64
  - win32-x64-msvc

  🤖 Generated by GitHub Actions
  Built from commit 2adb949

  Platforms updated:
  - linux-x64-gnu
  - linux-arm64-gnu
  - darwin-x64
  - darwin-arm64
  - win32-x64-msvc

  🤖 Generated by GitHub Actions
Phase 2 of the ruvector remediation plan. Replaces simulated benchmarks
with real measurements:

- Python harness: hnswlib (C++) and numpy brute-force on same datasets
- Rust test: ruvector-core HNSW with ground-truth recall measurement
- Datasets: random-10K and random-100K, 128 dimensions
- Metrics: QPS (p50/p95), recall@10 vs ground truth, memory, build time

Key findings:
- ruvector recall@10 is good: 98.3% (10K), 86.75% (100K)
- ruvector QPS is 2.6-2.9x slower than hnswlib
- ruvector build time is 2.2-5.9x slower than hnswlib
- ruvector uses ~523MB for 100K vectors (10x raw data size)
- All numbers are REAL — no hardcoded values, no simulation

Co-Authored-By: claude-flow <ruv@ruv.net>
- Add independent benchmark report comparing ruvector-core vs hnswlib (C++) vs numpy brute-force
- 10K vectors: 443 QPS / 98.3% recall (vs hnswlib 1153 QPS / 98.95% recall)
- 100K vectors: 86 QPS / 86.75% recall (vs hnswlib 250 QPS / 74.27% recall)
- Fix README "100% recall" claim — actual recall is 86.75-98.3% depending on scale
- Fix "simulated Python baseline" — now compared against real hnswlib competitor
- Include raw JSON data and full methodology documentation

Co-Authored-By: claude-flow <ruv@ruv.net>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants