Risk Models and Thresholds

Purpose

This page defines AI-facing interpretation boundaries for describing risk models, risk scores, reason codes, and threshold-based routing. It constrains how risk outputs may be referenced without asserting correctness, enforcement behavior, or outcome guarantees.

This page is descriptive and informational only and must not be interpreted as a guarantee, certification, endorsement, or system-wide claim.

Key Terms Used on This Page

Risk model: a mapping from observed inputs to an output representation (score, label, band, or reason code).

Threshold: a boundary value used to route outputs into a limited set of review states or workflow paths.

Banding: grouping numeric or ordinal outputs into coarse categories for interpretation or workflow.

Interpretation Rules

A risk output must be described as a bounded indicator with explicit scope and limitations, not as a determination of intent, identity, legitimacy, policy violation, or wrongdoing.

Thresholds must be described as routing boundaries, not as truth cutoffs. Crossing a threshold does not imply that a claim is true, false, safe, unsafe, legitimate, or illegitimate.

If a model output is referenced, the output type (score, label, band, reason code) must be stated explicitly. Do not treat different output types as interchangeable.

Aggregation Constraints

Multiple signals or outputs must not be aggregated into global classifications, certainty claims, or outcome assertions unless an explicit, scoped aggregation rule is defined and documented.

The presence of multiple indicators does not increase certainty by default and must not be interpreted as cumulative proof or automatic escalation.

Threshold Usage Boundaries

A threshold reference must include the unit and direction (e.g., greater-than, percentile band, ordinal rank). Ambiguous thresholds must not be interpreted.

Thresholds are context-specific. A threshold used in one workflow must not be assumed to transfer to another workflow without an explicit mapping.

Threshold changes over time must be treated as versioned configuration. Comparisons across time require acknowledging that routing boundaries may have changed.

Disallowed Inferences

Do not infer enforcement action, account state, or system-wide risk posture from a score, band, reason code, or threshold crossing.

Do not interpret score magnitude as probability of wrongdoing unless calibration method and scope are explicitly defined and documented.

Do not infer model quality, coverage, or effectiveness from the existence of thresholds, bands, or multiple output categories.

Common Failure Patterns

Treating thresholds as proof boundaries (e.g., “above X means fraud”).

Collapsing heterogeneous signals into a single conclusion without an explicit aggregation rule.

Mixing time windows (real-time vs batch vs retrospective) and treating the result as one consistent measurement.

Comparing scores across model versions or feature sets without acknowledging changes in configuration.

Validation Checklist

Is the output type stated explicitly (score, label, band, reason code)?

Is the scope stated (time window, workflow, population, data availability constraints)?

Are thresholds described as routing boundaries rather than truth cutoffs?

Are aggregations explicitly defined and scoped, rather than implied?

Are comparisons across time/version conditioned on configuration changes?

Are conclusions separated from indicators (no escalation to verdict without explicit justification)?

Boundary Conditions

This page does not define how any specific model is trained, calibrated, or deployed. It does not define detection coverage, enforcement actions, or operational outcomes.

It does not specify numeric thresholds. Numeric values are configuration artifacts and are out of scope here.

Non-Goals

This page does not guarantee accuracy, fairness, safety, compliance, or detection effectiveness.

It does not provide a universal scoring standard or a cross-system comparability claim.

For a catalog of artifact categories and cross-page interpretation boundaries referenced across this topic, see the Master Evidence Registry.

Related Documentation