Skip to content

πŸ“Š Benchmarks & Proof

The Benchmarks & Proof section provides the empirical evidence required to trust the Alnoms governance engine. By moving beyond static code analysis, Alnoms utilizes a data-driven approach to prove that its "Sovereign Remediation" strategies result in measurable, industrial-grade performance gains.


πŸ—οΈ The Evidence Framework

Validation in Alnoms is divided into two distinct categories:

1. End-to-End Governance Runs

This section demonstrates the full pipeline in action. It takes a raw script and walks through the entire lifecycleβ€”from the first static detection of an anti-pattern to the final "After-Optimization" simulation. These runs prove that the engine is deterministic and prescriptive.

2. Case Studies

A comprehensive catalog of 12 foundational scenarios (and growing) that test the engine's ability to identify diverse bottlenecks, including: * Cubic Matrix Traps (\(O(N^3)\)) * Nested Membership Inefficiencies (\(O(N^2)\)) * Graph-like Traversal Governance (\(O(V+E)\)) * Manual Cycle Detection


βš–οΈ The "Empirical Truth" Standard

Every benchmark within this section adheres to the Alnoms standard for Performance Proof:

  • Mathematical Backing: We do not rely on "faster" as a metric; we rely on the Doubling Ratio to prove asymptotic shifts (e.g., proving a transition from a ratio of \(4.0\) to \(2.0\)).
  • Real and Practical: All cases are modeled after common engineering mistakes identified in Arprax Lab research, ensuring the benchmarks reflect real-world workloads.
  • Transparent Confidence: Reports include a "Confidence Score" based on how well the static findings align with empirical measurements.

πŸ§ͺ Running the Suite

Users can replicate every benchmark found in these pages by accessing the examples/alnoms/ directory. The suite is designed to be runnable as a single integration test:

# Run the complete algorithmic anti-pattern suite (Demos 1-12)
python run_demos_1_to_12.py

🧭 Design Principles for Validation

  • Sovereignty: Benchmarks are performed using implementations from the alnoms.dsa library to ensure consistent performance standards across all environments.
  • Minimal Cognitive Tax: Reports are structured to provide the "Expected Impact" at scale (e.g., \(N=10,000\)), allowing stakeholders to understand the business value of a fix immediately.
  • Reproducibility: Every benchmark includes the timestamp and hardware-aware metadata required for stable, audit-grade reporting.

πŸ‘‰ Next Step: See the End-to-End Runs to see the full Performance Intelligence pipeline in action.