Methodology & Validation

How RIDGE produces a verdict, and how we validate it.

96.2%
NO-GO precision
97.8%
GO Precision
169
Niches back-tested
2,710
Ground-truth labels
6,779
Niches scored

One-paragraph summary

RIDGE takes a niche keyword, collects evidence from 39 data sources (SERP, reviews, regulator registries, trade databases, listing snapshots, financial benchmarks), reduces that evidence to a feature vector, and feeds it into a calibrated machine-learning classifier trained on 2,710 ground-truth niche outcome labels. The classifier outputs a probability that the niche will be commercially viable for a new entrant; that probability is combined with a deterministic rule layer to produce the shipped GO / HIGH RISK / NO GO verdict. Independently back-tested on a separately-held cohort of 169 historical Amazon FBA niches with four-year outcomes (2022 entry → 2026 result), the pipeline achieves 96.2% NO-GO precision and 97.8% GO precision — bootstrap 95% confidence intervals on every metric, no point estimates.

The five layers of a verdict

  1. Evidence collection. 39 data sources are queried for a fixed set of signals: SERP rankings, review density and authenticity, regulatory registries (FCC, CPSC, Comtrade), historical listing snapshots, financial benchmarks, supply-chain proxies. Every source has explicit fall-back behaviour and freshness watermarks.
  2. Feature reduction. Raw evidence is reduced to a feature vector covering market structure, competitive intensity, regulatory exposure, financial viability, and seasonality. Quantitative features (Gini, coefficient-of-variation, log-mean) live alongside binary signals.
  3. Calibrated probability. A multi-class machine-learning classifier produces three probabilities — P(DEAD), P(ALIVE), P(THRIVING) — that sum to 1. The DEAD probability drives the NO GO branch, the THRIVING probability drives the GO branch, and ALIVE is the abstain-eligible middle. Each class probability is independently calibrated, so a 0.80 prediction empirically corresponds to roughly an 80% outcome rate within that class.
  4. Rule layer. A deterministic, auditable verdict rule combines P(viable) with explicit gates (FCC-regulated → HIGH RISK; review-fraud cluster detected → HIGH RISK; saturated review distribution + low BSR → NO GO). Every gate fires a public evidence card on the report.
  5. Conformal abstain. Borderline niches where the model's probability is too uncertain to commit to a label are flagged "abstain" rather than forced into GO or NO GO. Abstain rate is published, not hidden.

Validation protocol

Cohort shift and prior calibration

The training pool and the held-out 169-cohort have different class priors: the training pool skews positive (75.4% viable) because Amazon-historical data over-represents niches that survived long enough to leave a usable signal trail, while the gold-standard 169-cohort is closer to a balanced 53.8% / 46.2% split. A naive classifier trained on the skewed pool would systematically over-predict GO when applied to the more balanced cohort.

We correct for this in two stages, both standard in the prior-shift literature:

After this two-stage correction, expected calibration error on the 169-cohort sits in the low single-digit-percent range with the bootstrap CI we publish. The training-vs-cohort prior gap is a fact of the data, not a bug — but it would become a bug the moment we let it leak through into the headline numbers without the correction layer.

Honesty matters more than a clean marketing chart

A few facts we keep visible on the public site even when they cut against the headline:

What we do not publish

The architecture is public; the coefficients are not. Specifically:

Publishing those would let adversaries game individual RIDGE reports — inflate the signals that score positive, suppress those that score negative. The academic discipline we do publish (calibration, validation protocol, held-out cohort, conformal coverage, bootstrap CI) is enough for any ML practitioner to verify the pipeline is not cooked. Nothing on this page prevents a reader from reproducing the methodology; it only prevents them from forging a RIDGE verdict.

See the same methodology applied to your niche

48-hour delivery, 40+ report sections, calibrated confidence with abstain on borderline niches.

Order Analysis