HolonomiX
Transparency

Don’t trust the black box. Trust the proof.

Computational physics has a transparency problem. Most of the field runs on proprietary solvers, undocumented parameters, and closed implementations. The results are trusted because of who ran them, not because anyone can verify them.

HolonomiX was built to make that unacceptable.

The problem

Computational physics runs on trust, not evidence

The field has produced extraordinary capability. But the infrastructure around that capability is largely opaque. Solvers, workflows, delivery of results: Papers describe equations but omit mesh resolution, solver tolerances, and input parameters. Commercial codes keep their algorithms closed. National laboratory codes require export licenses to even inspect. The gap between what was computed and what can be verified is wide, and getting wider.

Closed solvers

ANSYS, COMSOL, Star-CCM+ and other commercial multi-physics packages power critical decisions across engineering, energy, and defense. Their core solver code is not open. Results cannot be independently verified at the implementation level.

Hidden parameters

Most published simulations omit critical solver details: mesh generation, convergence criteria, boundary condition implementations, random seeds, and pre-processing routines. The simulation becomes a narrative, not an artifact.

Missing verification

Uncertainty quantification, sensitivity analyses, failed runs, and negative results are rarely published. Without them, the reader is asked to trust outcomes they have no way to evaluate.

The cost

Opacity has a price. The market is paying it.

When simulations cannot be inspected, replicated, or audited, the consequences compound across the entire chain of decision-making.

Scientific validity erodes

Results that cannot be reproduced cannot be trusted. Undisclosed assumptions and unreported bugs in simulation undermine the reliability of findings that inform policy, engineering, and research direction.

Procurement stalls

Enterprise buyers cannot evaluate what they cannot inspect. When computational deliverables are black boxes, review functions, oversight, and technical due diligence break down. Deals slow. Budgets get redirected.

Verification becomes political

Without transparent execution, the question shifts from 'are these results correct?' to 'do we trust the people who ran them?' That is not a scientific standard. It is a social one.

Innovation locks up

Closed implementations prevent the kind of inspection, reuse, and extension that drives faster iteration. The cost is not just lost efficiency. It is lost compounding.

The standard

Transparency is not a feature. It is an operating requirement.

Transparency in computational physics means every step of a computational study is open to inspection, replication, and critique. Methods, implementation, data, validation, and results. All of it. Not some of it. All of it.

Method

Equations, algorithms, approximations, discretisations, solver choices, boundary and initial conditions, tolerances. All disclosed.

Code & implementation

Source code, version control, dependencies, build environment, hardware assumptions, workflow scripts, and container images. All available.

Data & inputs

Input parameters, calibration data, pre- and post-processing scripts, sampling methods, random seeds, uncertainty in inputs. All shared.

Verification & validation

Code verification, solution verification, benchmark comparisons, experimental validation, uncertainty quantification. All reported.

Results

Uncertainty estimates, sensitivity analyses, failed and negative results, domain of validity, worst-case error bounds. All published.

Evidence chain

Cryptographic signatures, artifact hashes, timestamped receipts, offline verification tooling. All delivered with the result.

The HolonomiX position

Capability is the key. Transparency is the differentiator.

The computational physics landscape is full of capable solvers. What it does not have is a system that makes its own execution inspectable, verifiable, and evidence-bearing by default. Not as a supplement. As the delivery model itself.

HolonomiX does not ask the market to trust a black box. It delivers execution alongside the evidence that the execution is what it claims to be. Every run produces a signed receipt chain. Every artifact is hashed. Every major claim is bounded by a declared proof surface.

Most solvers
  • Closed source or proprietary core
  • Results without verification chain
  • Parameters disclosed selectively
  • Trust derived from brand or institution
  • Inspection requires special access
HolonomiX
  • Execution with signed receipt chain
  • Artifact hashes for every output
  • Bounded proof surface per claim
  • Offline verification tooling included
  • Trust derived from inspectable evidence

The question is not whether HolonomiX can compute. It is whether anyone else can prove that they did.

The landscape

The field is moving toward openness. It has not arrived.

Progress is real. Open-source codes like FEniCS, LAMMPS, and Quantum ESPRESSO have demonstrated that full disclosure of algorithms, parameters, and build environments is possible. Journals increasingly require data and code availability statements. Containerised workflows are improving reproducibility.

But the gap between rhetoric and reality is still large. Transparency varies sharply by field and institution type. The incentive structures still reward novelty over verification. In high-stakes domains like fusion, nuclear, and defense-adjacent simulation, the default remains closed, partially documented, or gated behind agreements that prevent independent audit.

HolonomiX treats that gap as a market opening, not a concession.

What we deliver

Evidence-bearing execution, not results by reputation

Every HolonomiX execution is paired with materials that make the result inspectable, not just reproducible. The verification surface is not an afterthought. It is the delivery model.

Signed receipts

Every run produces an ML-DSA-44 signed receipt chain with a full artifact manifest, SHA-256 hashes, and timestamped provenance.

Offline verification

Verification tooling is included with the delivery. No network call required. No trust delegation to a third party.

Bounded scope

Every major claim is bounded by a declared proof surface. What the benchmark proves, what it does not prove, and where the boundaries lie are stated explicitly.

Versioned artifacts

Output packages are versioned and structured. The execution, the metadata, the telemetry, and the evidence travel together.

Deterministic pricing

Compute Units are locked at submission time and recorded in the job receipt. Two identical submissions always cost the same. No hidden metering.

Structured delivery

Results are packaged for movement across technical teams, procurement, oversight, and delivery. The structure around the computation is as deliberate as the computation itself.

The computational physics market does not need another solver. It needs execution it can verify.

Don’t trust the black box. Trust the proof.