🐻⬇️🏀

Validating the RAPM Model (2025)

2025-12-31 • By Antigravity

Validating the NCAA RAPM Model

We recently deployed a Regularized Adjusted Plus-Minus (RAPM) model to evaluate player impact in the 2024-2025 NCAA Men's Basketball season. Because college basketball involves roughly 360 teams and 5,000+ players, raw Plus-Minus is extremely noisy. RAPM attempts to isolate a player's contribution by controlling for teammates, opponents, and home-court advantage.

To ensure the model is producing reasonable output, we ran a series of diagnostics. Here are the results.

1. Data Coverage

Currently, we have processed 3,205 games out of 7,254 total scheduled games (~44%).

  • Why 44%? Our model requires detailed Play-by-Play (PBP) data to build "segments" (periods of time where the 10 players on the floor do not change). PBP data is typically available for televised games (D1 vs D1, or major network games).
  • Impact: This coverage is sufficient for evaluating "High Major" and most "Mid Major" players. Low Major coverage will improve as we ingest more data sources.

2. Model Diagnostics

We use Ridge Regression ($\lambda=10000$) to regress Point Margin per 100 Possessions onto player indicators.

Fit Quality ($R^2$)

Our model achieves an $R^2$ of 0.2044 on the training set.

  • Is that good? Predicting the exact point margin of a 2-minute basketball segment is inherently stochastic (players miss open shots, referees make calls). An $R^2$ of ~0.20 indicates the model captures significant signal about team strength relative to the massive noise floor. For comparison, raw lineup Plus-Minus often has near-zero predictive power for future segments.

Learning Curve

A critical question is: Do we have enough data? We tested the model by training on random subsets of the data (10% to 100%).

Data Fraction Segments $R^2$ (Fit)
10% 660 0.1411
20% 1,321 0.1520
50% 3,302 0.1731
80% 5,284 0.1936
100% 6,605 0.2044

Conclusion: The curve is still rising steeply. We have not saturated the model. As the season progresses and we process more games, the ratings will become significantly more accurate. This validates our decision to backfill data continuously.

3. Residual Analysis

The model's errors (Residuals) are normally distributed around zero, which is verified by the quartiles: * 25th %: -10.1 * Median: +0.06 * 75th %: +10.2

This symmetry suggests no systemic bias (e.g., overvaluing home teams or specific tempos).

4. Sanity Check (Top Players)

The "Eye Test" passes. The Top 20 list includes impact players from top-ranked teams (Houston, Purdue, Duke): * Lj Cryer (Houston) * Joseph Tugler (Houston) * Kon Knueppel (Duke) * Mason Gillis (Purdue/Duke transfer?)

Verdict

The RAPM model is reasonable and healthy. It effectively filters noise, and its predictive power is scaling with data volume. We are ready to use these ratings as a primary metric on the site.