Works for any time-dependant prediction task • Finance, ML, Sales, Weather, Energy

Did Your Predictions
Actually Work?

Stop measuring how close you were. Start measuring if you would have won.

📊

Upload Your Predictions

Drop any CSV with predictions & actual outcomes (6 columns maximum)
Works for stocks, sales, weather, sports-anything

📥 Download Example CSV

Your Results

Success Score
-
How useful these predictions actually are

📈 Forecast vs Actuals

Forecast Comparison

Direction Accuracy
-
Did you predict up/down correctly?
Profit Score
-
If you acted on these, would you win?
Efficiency Ratio
-
Balance of usefulness vs accuracy
Risk Score
-
Reward relative to downside risk
Worst Case
-
Biggest mistake in your predictions

What's the Success Score?

Think of it like this: If you bet money on your predictions, would you make profit?

Why Not Just "Accuracy"?

Being close isn't enough. Wrong direction can lose money.

Who Uses This?

Anyone who needs predictions that work, not just look good on paper.

Real Examples That Make Sense

This works for any prediction, not just finance

📈 Stock Trading

Scenario: You predict a stock at $101, it goes to $99.
Traditional: "Close! Only 2% error."
Quantsynth: "Wrong direction = you lost money."

☁️ Weather Forecasting

Scenario: You predict 75°F, actual is 73°F.
Traditional: "Good job!"
Quantsynth: "Did people bring jackets? That's what matters."

🏀 Sports Betting

Scenario: You predict a 10-point win, they win by 3.
Traditional: "Pretty close!"
Quantsynth: "You predicted a win and they won = profit."

💼 Sales Forecasting

Scenario: Predict $100K sales, actual is $80K.
Traditional: "20% error"
Quantsynth: "Did you overstaff? That costs real money."

🤖 ML Model Selection

Scenario: Model A: 95% accurate. Model B: 88% accurate.
Traditional: "Pick A"
Quantsynth: "Which one makes fewer expensive mistakes?"

🏠 Real Estate

Scenario: Predict price increase, it decreases 5%.
Traditional: "Small error"
Quantsynth: "Wrong direction = you bought at the peak."

Why Traditional Metrics Miss The Point

Comparing what matters vs what gets measured

WHAT MATTERS TRADITIONAL APPROACH QUANTSYNTH APPROACH
Getting Direction Right ✕ Completely ignored (only cares about distance) ✓ Primary focus (80% of success)
Actual Success Rate ✕ No connection to outcomes ✓ Directly measures if it works
When You're Wrong ✕ Big mistakes kill your score ✓ Understands some errors matter more
Confidence Levels ✕ Treats all predictions equally ✓ Rewards being confident when right
Real-World Use ✕ Great metrics, bad decisions ✓ Optimizes for actual utility
Industry Adoption ✕ Used because it's familiar ✓ Used by quant funds making real bets

Metrics That Reflect Real Outcomes

Not how close your predictions were — but how useful they actually are

🎯

Forecast Investment Score (FIS)

FIS measures whether a prediction would have led to a good decision. Instead of rewarding forecasts for being close to the target, it rewards them for being right in the moments that matter.

  • Correct direction matters more than small numerical errors
  • Consistent signals are trusted more than lucky ones
  • Weak or non-committal predictions are penalized
How to read it:
FIS ranges from 0 to 1. Higher values mean the prediction behaves like a reliable, decision-ready signal.
⚖️

Confidence–Efficiency Ratio (CER)

CER answers a different question: How much confidence do you earn per unit of error? It balances how strong a prediction is with how efficiently it achieves that strength.

  • Rewards confident forecasts that justify their errors
  • Penalizes noisy or inefficient predictions
  • Comparable across different problems and scales
How to read it:
Higher CER means the model delivers trustworthy confidence without excessive error.

How These Metrics Work Together

FIS tells you whether a prediction is worth acting on. CER tells you whether that confidence is efficiently earned. Together, they reveal why traditional accuracy metrics can be misleading.

Why R² Fails When It Matters Most

Visualizing the difference in what gets rewarded

📉

High R², Low FIS

The "Precise Loser"

Example Forecast:
Predicted: $101.00 → Actual: $99.00
Predicted: $102.50 → Actual: $98.00
Predicted: $104.00 → Actual: $95.00
R² Score: 0.89
FIS Score: 0.08

Close to the target price, but wrong on every direction. R² says "great!" FIS says "you'd lose money on every trade."

📈

Low R², High FIS

The "Imprecise Winner"

Example Forecast:
Predicted: $110.00 → Actual: $105.00
Predicted: $115.00 → Actual: $108.00
Predicted: $120.00 → Actual: $112.00
R² Score: -0.52
FIS Score: 0.87

Wrong magnitude, but correct on every direction. R² says "terrible!" FIS says "you'd profit on every trade."

We Measure Economic Density

While R² rewards models for being near the target, FIS rewards models for being right on the trade. It's the difference between academic accuracy and profitable action.

Stress-Tested for Black Swans

Your model looks great in calm markets. What about chaos?

🌪️
Volatility
Analysis

Copula-Based Tail Dependence

Does Your Model Fail When It Matters?

Most evaluation metrics treat all days equally. But in reality, extreme events are where fortunes are made or lost. A model that's "pretty good" 95% of the time but completely wrong during 5% black swan events is worthless.

Quantsynth's Approach:

We use regime-dependent copula analysis to separately evaluate your performance during calm vs volatile periods. Your score reflects how you perform when the market moves ±5% in a day, not just the easy days.

This is the same mathematics used by quant hedge funds to ensure their strategies don't blow up during market stress. Now you have it for your predictions.

Stop Gaming the System

Traditional metrics are easy to cheat. Ours aren't.

🎯

The Flat Forecast Trick

Predict the same value every time. You'll often get decent MSE and MAE scores because you're "consistently close to the average."

FIS Score: 0.00

We catch it instantly. Flat forecasts = zero useful information.

📊

The Sample Size Game

Get lucky on 10 predictions? Traditional metrics will give you a great score even though it's statistically meaningless.

10 predictions: λ = 1.3
100 predictions: λ = 2.0

Dynamic λ adapts: we're skeptical of small samples, confident in large ones.

🔍

The Magnitude Bluff

Exaggerate every move. You'll be wrong on magnitude but R² won't penalize you much if you're directionally inconsistent.

Traditional CER: R² / MSE
Quantsynth CER: FIS² / MASE

MASE-adjusted scoring makes magnitude errors impossible to hide.

"It's an honest mirror for your models. You can't hide poor performance behind statistical noise."

See The Difference in Action

How traditional metrics get fooled by common prediction patterns

⚠️ The Lagged Predictor

This model just predicts yesterday's price. It's incredibly common (moving averages, momentum strategies) and looks great on traditional metrics because it's "always close." But it's completely untradeable.

Sample Predictions:
Day 1: Predict $100 → Actual $102
Day 2: Predict $102 → Actual $98
Day 3: Predict $98 → Actual $103
Day 4: Predict $103 → Actual $101
Why it fails:

By the time you get the signal, the move already happened. You're always entering after the profitable moment. It's like reading yesterday's news.

METRIC COMPARISON

R² Score 0.92

✓ "Excellent fit!"

MAE 1.8

✓ "Very accurate!"

FIS Score 0.02

✕ "Worthless for trading"

Real Trading Result: You'd lose money on transaction costs alone. Every trade is a day late.

Built for Professional Use

The transparency and privacy you need for sensitive data

🔒

Privacy First

Predictions processed in-memory. We evaluate your utility, we don't steal your alpha. No data storage, no tracking, no leaks.

📐

Open Methodology

Math is documented. Implementation stays private. You know what we do, just not how we optimize it.

📊

Full Transparency

Download reports include all the necessary data to complete your analysis. No black boxes.

Statistically Validated

50,000+ Monte Carlo simulations. Bootstrap confidence intervals. Peer-reviewed methodology. Not guesswork.

Built by Francisco Cardoso · LinkedIn