Public Trust Layer

Reality Engine — Prediction vs Execution

FYOS does not publish a success dashboard here. This surface exists to show where the model tracks reality, where it drifts, and how much trust each layer deserves.

This page separates four kinds of evidence:

  • Prediction Alignment — Did predictions match realized direction/magnitude?
  • Economic Viability — Did opportunities remain viable after fees, decay, downside, and execution realities?
  • Bounded Planner — Portfolio-level validation on limited cohorts only
  • Coverage Limits — What remains weakly validated or incomplete?
Validated observations
4497

Product-wide sample

Median prediction error
0.05%

Alignment metric only

Current reliability state
35.0

Trust layer metric

Last audited snapshot
Mar 28, 2026, 04:51 PM
How to use this page

Use this surface to evaluate model trust, not opportunity attractiveness. Alignment and reliability describe how closely FYOS tracks realized outcomes; survivability, half-life, and capacity show where caution still matters.

Prediction Alignment
Product-wide

Prediction Accuracy

How closely predictions matched realized direction and magnitude within the intended prediction contract. Good alignment does NOT imply economic viability.

Prediction vs Realized

Aggregated reality check

Median predicted net return
-0.46%
Median realized net return
-0.43%
Mean bias
0.06%
Directional accuracy
99.89%
Predictions aligned with realized edgeneutral

No systematic overstatement detected; mean bias is 0.06% and directional flips remain 99.89% of snapshots.

Model Reliability

Current trust signal, not a success badge

Empirical reliability score
35.0
12630 validated rows
empirical sample live
Moderate confidencemedium

Model behaves cautiously with mixed evidence.

Reliability reflects alignment with realized outcomes, not profitability.

Error Distribution

Observed prediction error bands

Median0.05%
P750.15%
P900.43%
Wide error bandshigh

Prediction error dispersion is currently high within the validated sample.

Exchange-Level Alignment

Accuracy by exchange

ExchangeMedian ErrorDir. AccuracySamples
BINANCE0.05%99.89%1,839
BYBIT0.08%99.93%1,485
OKX0.05%99.83%1,173
Prediction alignment measures whether the model correctly forecasts direction and magnitude. It does NOT measure whether opportunities were economically viable after costs and execution.
Economic Viability

Economic Viability

Whether opportunities remained economically viable after fees, decay, mirage adjustments, and execution realities. This is separate from prediction accuracy.

Legacy Survivability BranchDeprecated

Archived validation context only

Legacy survivability rate

Deprecated diagnostic only. Not used in primary scoring, ranking, or allocation.

Legacy sign-flip rate
0.11%

Deprecated diagnostic only. Retained for archived comparison context.

Legacy observations remain visible here only to document cleanup progress, not as primary truth semantics.

Mirage ValidationGross/Net Gap

Predicted vs actual reality gap

Median mirage error
547.50%

This measures gross-to-net explanatory gap, not prediction failure

Underpenalty rate
29.27%

Missed decay events

Overpenalty rate
66.47%

Conservative estimates

Based on 12520 validated observations

Decay-Adjusted Return ValidationNet-Facing Metric

Post-decay return accuracy

Median error
547.50%
Directional accuracy
70.81%

Predictions with correct sign vs realized

Based on 12630 validated observations

Economic viability validation shows whether mirage-adjusted returns matched reality. Large mirage errors are expected - they represent the gross-to-net gap, not model failure.
Bounded Planner Validation
Sparse Samplen=3

Planner Validation

Portfolio-level validation on bounded cohorts only. These metrics are NOT product-wide system truth - they represent results from specific portfolio replay scenarios.

Return prediction errorBounded
0.00%
Beat expected returnBounded
100.00%

Portfolios that outperformed expectations

Validated runsBounded
3

Sparse sample - interpret with caution

Only 3 planner runs validated. This is a bounded cohort result, not product-wide truth. Metrics will stabilize as more portfolios are replayed.
Coverage & Validation Limits

Coverage & Validation Limits

What remains weakly validated, incomplete, or limited in scope. These caveats should inform how much weight to give other metrics.

Half-Life AccuracyWeakest Layer

Predicted vs actual decay

Median absolute errorExploratory
20.2h
Observed fractionExploratory
4.19%

Censorship-heavy observation window

Decay modeling gap notedhigh

Deviation of 20.19h hours marks this as the weakest validated layer.

Capacity RealityReference

Expected vs realized capacity

Reference median capacity
$22,121
Interpretation
Capacity modeled, not validated

Used as a soft constraint, not an audited execution limit.

Coverage & Observability

Model visibility signals

Tracked opportunities
1615

Active venue points observed

Validation row counts
Reliability:12630
Legacy survivability:0
Decay (half-life):4037
Half-life remains the weakest validated layer with limited observed decay events. Capacity validation is reference-only. Coverage shows observability, not edge quality.

Final verdict

This page reports model accountability—alignment, drift, and validation gaps—so trust is earned empirically.

Methodology

How predictions are evaluated

FYOS compares prediction snapshots against realized execution outcomes, profiles error distributions, validates realized economic drift empirically, and exposes weak areas instead of hiding them behind performance claims.

Cookie preferences

We use cookies to improve analytics and user experience. You can accept or reject non-essential cookies. Learn more in our Privacy Policy.