Philosophy of ARF¶
ARF is not just a set of algorithms or a policy engine. It is a coherent design philosophy for making autonomous AI safe, accountable, and commercially viable. This page captures the underlying beliefs, trade‑offs, and first principles that guide every architectural decision. Wherever possible, we link to the actual code that embodies these ideas (the core engine is proprietary, but its behaviour is exactly as described in the public specification).
1. Core Beliefs¶
1.1 Reliability is a system property, not a model output¶
No single number (accuracy, F1, etc.) can capture operational reliability. Reliability emerges from clear boundaries, structured policy evaluation, strong observability, and careful separation of concerns. ARF therefore separates:
- Risk estimation –
RiskEngine(inrisk_engine.py) combines conjugate Beta priors, HMC logistic regression, and optional hyperpriors. - Policy enforcement –
PolicyEvaluator(used ingovernance_loop.py) returns a list of violations; any violation forcesDENY. - Advisory reasoning – the
HealingIntentcontains a human‑readablejustificationandrisk_factorsbreakdown. - Execution gating – the enterprise layer (Rust
ExecutionLadder) validates gates (license, confidence, risk, rollback, causal).
1.2 Probabilistic inputs, deterministic actions¶
AI models are inherently probabilistic, but production systems require deterministic, auditable actions. ARF therefore:
- Uses Bayesian risk scores only as advisory inputs – the risk score is a posterior mean (
conjugate_riskinrisk_engine.py), but the final action is chosen by expected loss minimisation (seegovernance_loop.py, lines whereL_approve,L_deny,L_escalateare computed). - Never uses fixed probability thresholds. Instead, it computes
expected_lossesand selects the action with the minimum expected loss (or overrides via policy or epistemic gate). - Applies epistemic uncertainty (
psi_meaningovernance_loop.py) to escalate, not to guess.
1.3 Trust through transparency, not black‑boxes¶
Explainability is a first‑class requirement. Every decision must answer: What happened? Why? What would change the outcome? ARF provides:
- Structured justifications (the
explanationstring fromrisk_engine.calculate_risk()). - Confidence intervals (derived from Beta posterior variance).
- Causal explanations (via the enterprise
CausalExplainer).
1.4 Pilot‑first, honest boundaries¶
The core engine is proprietary and access‑controlled, offered under outcome‑based pricing. This is not a restriction – it is a commitment to sustainability and continuous improvement. The public specification, demo UI, and pitch deck are fully open source (Apache 2.0). This clear boundary ensures:
- Enterprises can evaluate ARF without operational risk.
- The core engine remains focused, fast, and commercially supported.
- The open‑source community can contribute to the spec and demos.
2. Design Tenets (with code evidence)¶
| Tenet | Implementation in Code |
|---|---|
| Determinism at the boundary | GovernanceLoop.run() (same inputs → same decision) – no hidden randomness. The only “random” part is Monte Carlo sampling for CVaR, which is advisory. |
| Separation of concerns | GovernanceLoop orchestrates independent modules: RiskEngine, PolicyEvaluator, CostEstimator, RAGGraphMemory, etc. |
| Extensibility without corruption | Optional components (hyperpriors, epistemic gate, predictive engine) are enabled via configuration flags (use_hyperpriors, enable_epistemic). |
| Safety by default | When uncertain, USE_EPISTEMIC_GATE and psi_mean > EPISTEMIC_ESCALATION_THRESHOLD force ESCALATE (see governance_loop.py). Policy violations always force DENY. |
| Economic alignment | Outcome‑based pricing is enforced in the enterprise layer; the OSS constants (COST_FP, COST_FN, etc.) reflect real business trade‑offs (e.g., COST_FP > COST_FN penalises false positives more). |
3. Human‑Centric AI¶
Psychological principles (see psychology.md) are not “nice to have” – they are mechanically enforced:
- Trust calibration – the
confidencefield inHealingIntentis set to1.0 - psi_mean, and the frontend displays credible intervals. - Cognitive load reduction – the dashboard prioritises critical alerts, then system status, then trends.
- Bias mitigation – pessimistic priors are hardcoded in
PRIORS(e.g., database actions haveBeta(1.5, 8.0)– prior mean ≈0.16). Periodic fairness audits can be run on enterprise audit logs. - Human‑in‑the‑loop – automatic escalation when
psi_mean > EPISTEMIC_ESCALATION_THRESHOLDor when policy violations occur.
4. Mathematical Minimalism¶
ARF uses the minimum mathematics necessary for safety and explainability:
- Conjugate Beta‑Binomial priors – fast, interpretable, updateable (
BetaStoreinrisk_engine.py). - HMC (offline) – captures complex interactions (time, role, environment) without over‑engineering (
HMCModel). - Hyperprior shrinkage – shares strength across sparse categories (
HyperpriorBetaStoreuses Pyro SVI). - Expected loss minimisation – chooses action by balancing risk, cost, and uncertainty (formulas in
governance_loop.pyusing constants likeCOST_FP,COST_IMPACT, etc.). - Lyapunov stability – provides local asymptotic guarantees for healing actions (enterprise layer, not in OSS, but described in
mathematics.md).
We do not add complexity for its own sake. Every formula serves a clear operational purpose.
5. Commercial Realism¶
ARF is not a research project – it is a product. The philosophy acknowledges that:
- Open source does not mean free infrastructure – the core engine is proprietary to ensure sustained development, support, and security.
- Pilot‑first, outcome‑based pricing aligns incentives: customers pay when ARF reduces risk, not for idle evaluations.
- Transparency about limitations – the public spec honestly states what is public vs. private, what is implemented vs. planned. The constants file (
constants.py) explicitly defines OSS boundaries (EXECUTION_ALLOWED = False,MCP_MODES_ALLOWED = ("advisory",)).
6. Where Philosophy Meets Code (Concrete File References)¶
The principles above are not abstract – they are enforced in the actual implementation:
governance_loop.py– deterministic expected loss minimisation, policy override, epistemic gate, and creation ofHealingIntent.risk_engine.py– hybrid Bayesian fusion with caps:w_hmc = min(0.6, n/n0),w_hyper = min(0.3, ...), ensuring weights sum to 1.constants.py– cost coefficients (COST_FP,COST_FN,COST_REVIEW, etc.) and OSS boundaries (EXECUTION_ALLOWED = False,MAX_INCIDENT_NODES = 1000, etc.).
These files are proprietary, but their behaviour is exactly as described in the public specification. The philosophy tab serves as a bridge between high‑level principles and the actual code.
7. See Also¶
design.md– architectural decisions that implement these principles.psychology.md– psychological foundations of trust and explainability.governance.md– the governance loop that turns philosophy into action.mathematics.md– mathematical formulations referenced here.