## 1 Introduction

In sports, as in many other industries and research fields, data analysis has become an essential ingredient of management. Sports teams, traditionally run by people with experience playing and/or coaching, now rely on statistical models to measure player ability and inform strategy decisions (Lewis 2004; Oliver 2004). Over the years, the quantity, scope, and sophistication of these models has expanded, reflecting new data sources, methodological developments, and increasing interest in the field of sports analytics. Despite their inherent promise, new developments in sports analytics have created a clutter of metrics. For example, there are at least three different calculations of the WAR (“Wins Above Replacement”) metric in baseball (Baumer, Jensen and Matthews 2015), all of which have the same hypothetical estimand. In general, any individual making a management, coaching, or gambling decision has potentially dozens of metrics at their disposal, but finding the right metrics to support a given decision can be daunting. We seek to ameliorate this problem by proposing a set of “meta-metrics” that describe which metrics provide the most unique and reliable information for decision-makers. Our methods are simple to implement and applicable to any sport so they should be of broad interest to the sports analytics community.

The core idea of our work is that quantifying sources of variability – and how these sources are related across metrics, players, and time – is essential for understanding how sports metrics can be used. In this paper, we consider three different sources of variation, which we classify differently depending on the use-case. These are (1) intrinsic player skill, (2) context, e.g. influence of teammates, and (3) chance, i.e. sampling variability. Each of these sources can vary across seasons and between players. We consider each player metric to be composed of a combination of these sources of variation, and in this paper we discuss several diagnostics that can be used to assess how well certain metrics are able to measure, control for, and average across these sources of variation, depending on what is required by the decision-maker.

The primary purpose of constructing our meta-metrics is to categorize the sources of variation in the data as *signal* and *noise*. The signal corresponds to variation that is the key input into a decision process, e.g., a player’s ability to operate in a given system, whereas the *noise* is variation that we choose not to explain either because of complexity or lack of information (e.g., complex team interactions or minuscule variations in a player’s release between shots). When relevant we condition on observed contextual information (e.g. player position) to create more reliable and interpretable signals.

For a metric to be useful for a particular decision, its treatment of variation needs to match up with the decision that is being made. For example, consider two distinct tasks in which metrics are often deployed – attribution, where we wish to credit a portion of a team’s success to a given player for, e.g., year-end awards, and acquisition, where we wish to assess whether a player should be added to or retained on a team. The classification of signal and noise in these decision tasks is very different. For attribution, we do not care whether a player can repeat their performance in another season (or arguably even how much of their performance was due to chance), whereas repeatability is a central question in player acquisition. That is, chance and team context are still relevant signals when making an attribution decision, but are sources of noise for an acquisition decision.

While we can isolate some player-wise, season-wise, and team-wise variation by subsetting the data, all measurements that we take are confounded with chance. Further “skills” are abstract concepts that are often collapsed together. With this in mind, we define three meta-metrics that can be used to answer the following questions of player performance metrics:

**Discrimination**: Does the metric reliably differentiate between players?**Stability**: Does the metric measure a quantity that is stable over time?**Independence**: Does the metric provide new information?

Our discrimination meta-metric quantifies how useful a metric is for distinguishing between players within a given season, whereas our stability meta-metric measures how much a metric varies season to season due to changes in context and player skill after removing chance variation. The independence meta-metric quantifies how much information in one metric is already captured by a set of other metrics. Our meta-metrics are based on ideas that have a long history in statistics (e.g., analysis of variance) and psychometrics (e.g., Cronbach’s alpha) (Fisher 1925; Kuder and Richardson 1937; Cronbach 1951) but have not received widespread treatment in sports. The limited work quantifying the reliability of metrics in sports mostly appears in blogs (Blackport 2014; Sprigings 2014; Arthur 2015) and our hope is to formalize and generalize some of the ideas discussed in these articles. We start in Section 2 by motivating and defining three meta-metrics and discuss how to estimate them in Section 3. Section 4 demonstrates the application of these meta-metrics to player performance in the National Basketball Association (NBA) and National Hockey League (NHL). Lastly, in Section 5 we discuss building new metrics and adjusting existing ones in order to improve their meta-analytic properties.

## 2 Defining meta-metrics

Throughout this paper, we write the 3-dimensional array of players, seasons and metrics as *X*, with *X _{spm}* the value of metric

*m*for player

*p*from season

*s*. Our meta-metrics are all R-squared style statistics and can be understood as functions of the (co)variances along the three dimensions of

*X*. As a useful example, consider a model for a metric

*m*that varies over time

*s*and between players

*p*in a linear mixed effects model:

where

and [*μ*, *σ*^{2}] represents a distribution with mean *μ* and variance *σ*^{2}. The terms *Z*_{*} can be thought of as random effects, while *ϵ*_{spm} represents the variation induced by the sampling effects of a season – for instance, binomial variation in made shot percentage given a finite sample size; for an infinitely long season, we would observe *ϵ*_{spm} = 0. *Z _{spm}*, on the other hand, reflects true value (above average) of the “skill”

*m*of player

*p*in season

*s*. This model encodes four sources of variation, although we only intend to discuss three. The extra parameter,

In this representation, we can recognize

The linear mixed effects model (1) may be a reasonable choice for some metrics and, due to its simplicity, provides a convenient example to illustrate our meta-metrics. However, an exchangeable, additive model is not appropriate for many of the metrics we consider. A major practical challenge in our analysis is that all of the metrics have unique distributions with distinct support – for example, percentages are constrained to the unit interval, while many per game or per season statistics are discrete and strictly positive. Other advanced metrics like “plus-minus” or “value over replacement” (VORP) in basketball are continuous real-valued metrics that can be negative or positive.

To define meta-metrics with full generality, consider the random variable *X*, which is a single entry *X _{spm}* chosen randomly from

*X*. Randomness in

*X*thus occurs both from sampling the indexes

*S*,

*P*, and ℳ of

*X*, as well as intrinsic variability in

*X*due to finite season lengths. We will then use the notational shorthand

_{spm}and analogously for *E _{sm}*[

*X*],

*V*[

_{sm}*X*],

*E*[

_{m}*X*], etc. For example,

*E*[

_{sm}*V*[

_{spm}*X*]] is the average over all players of the intrinsic variability in

*X*for metric

_{spm}*m*during season

*s*, or ∑

_{p}

*Var*[

*X*] /

_{spm}*N*, where

_{sm}*N*is the number of entries of

_{sm}*X*.

_{s ⋅ m}### 2.1 Discrimination

For a metric measuring player ability to be applicable, it must be a useful tool for discriminating between different players. This implies that most of the variability between players reflects true variation in player ability and not chance variation or noise from small sample sizes. As a useful baseline for discrimination, we compare the average intrinsic variability of a metric to the total between player variation in this metric. Similar approaches which partially inspired our version of this metric have been used in analyzing Major League Baseball data (Tango, Lichtman and Dolphin 2007; Arthur 2015).

To characterize the discriminative power of a metric, we need to quantify the fraction of total between player variance that is due to chance and the fraction that is due to signal. By the law of total variance, the between player variance can be decomposed as

Here, *V _{sm}*[

*X*] corresponds to the total variation in metric

*m*between players in season

*s*, whereas

*E*[

_{sm}*V*[

_{spm}*X*]] is the average (across players) sampling variability for metric

*m*in season

*s*. With this decomposition in mind, we define the discriminative power of a metric

*m*in season

*s*as

Intuitively, this describes the fraction (between 0 and 1) of between-player variance in metric *m* (in season *s*) due to true differences in player ability. Discrimination meta-metrics for different seasons can be combined as

It is helpful to understand the discrimination estimand for the linear mixed effects model defined in Equation 1. When this model holds, *s*). Thus, the discrimination meta-metric under the linear mixed effects model is simply

### 2.2 Stability

In addition to discrimination, which is a meta-metric that describes variation within a single season, it is important to understand how much an individual player’s metric varies from season to season. The notion of stability is particularly important in sports management when making decisions about future acquisitions. For a stable metric, we have more confidence that this year’s performance will be predictive of next year’s performance. A metric can be unstable if it is particularly context dependent (e.g. the player’s performance varies significantly depending on who their teammates are) or if a player’s intrinsic skill set tends to change year to year (e.g. through offseason practice or injury).

Consequently, we define stability as a metric, which describes how much we expect a single player metric to vary over time after removing chance variability. This metric specifically targets the sensitivity of a metric to change in context or intrinsic player skill over time. Mathematically, we define stability as:

with 0 ≤ *S*_{m} ≤ 1 (see Appendix for proof). Here, *V _{pm}*[

*X*] is the between-season variability in metric

*m*for player

*p*; thus, the numerator in (4) averages the between-season variability in metric

*m*, minus sampling variance, over all players. The denominator is the total variation for metric

*m*minus sampling variance. Again, this metric can be easily understood under the assumption of an exchangeable linear model (Equation 1).:

This estimand reflects the fraction of total variance (with sampling variability removed) that is due to within-player changes over time. If the within player variance is as large as the total variance, then *S*_{m} = 0 whereas if a metric is constant over time, then *S*_{m} = 1.

### 2.3 Independence

When multiple metrics measure similar aspects of a player’s ability, we should not treat these metrics as independent pieces of information. This is especially important for decision makers in sports management who use these metrics to inform decisions. Accurate assessments of player ability can only be achieved by appropriately synthesizing the available information. As such, we present a method for quantifying the dependencies between metrics that can help decision makers make sense of the growing number of data summaries.

For some advanced metrics we know their exact formula in terms of basic box score statistics, but this is not always the case. For instance, it is much more challenging to assess the relationships between new and complex model based NBA metrics like adjusted plus minus (Sill 2010), EPV-Added (Cervone et al. 2016) and counterpoints (Franks et al. 2015), which are model-based metrics that incorporate both game-log and player tracking data. Most importantly, even basic box score statistics that are not functionally related will be correlated if they measure similar aspects of intrinsic player skill (e.g., blocks and rebounds in basketball are highly correlated due to their association with height).

As such, we present a general approach for expressing dependencies among an arbitrary set of metrics measuring multiple players’ styles and abilities across multiple seasons. Specifically, we propose a Gaussian copula model in which the dependencies between metrics are expressed with a latent multivariate normal distribution. Assuming we have ℳ metrics of interest, let *Z _{sp}* be an ℳ-vector of metrics for player

*p*during season

*s*, and

where *C* is a ℳ × ℳ correlation matrix, and *m*. We define the independence score of a metric *m* given a condition set of other metrics, ℳ, as

For the latent variables *Z*, this corresponds to one minus the R-squared for the regression of *Z _{m}* on the latent variables

*Z*with

_{q}*q*in ℳ. Metrics for which

*m*is independent from all metrics in ℳ.

We also run a principal component analysis (PCA) on *C* to evaluate the amount of independent information in a set of metrics. If *UΛ U ^{T}* is the eigendecomposition of

*C*, with Λ = diag(

*λ*

_{1}, …

*λ*

_{ℳ}) the diagonal matrix of eigenvalues, then we can interpret

*k*principal components (Mardia, Kent and Bibby 1980). When

*F*

_{k}is large for small

*k*then there is significant redundancy in the set of metrics, and thus dimension reduction is possible.

## 3 Inference

In order to calculate discrimination *D*_{m} and stability *S*_{m}, we need estimates of *V _{spm}*[

*X*],

*V*[

_{sm}*X*],

*V*[

_{pm}*X*] and

*V*[

_{m}*X*]. Rather than establish a parametric model for each metric (e.g. the linear mixed effects model (1)), we use nonparametric methods to estimate these variances. Specifically, to estimate the sampling distribution of

*X*within each season (e.g.,

*Var*[

*X*], or equivalently

_{spm}*V*[

_{spm}*X*], for all

*s*,

*p*,

*m*), we use the bootstrap (Efron and Tibshirani, 1986). For each team, we resample (with replacement) every game played in a season and reconstruct end-of-season metrics for each player. We use the sample variance of these resampled metrics, BV[

*X*], to estimate the intrinsic variation in each player-season metric

_{spm}*X*. We estimate

_{spm}*V*[

_{sm}*X*],

*V*[

_{pm}*X*] and

*V*[

_{m}*X*] using sample moments.

Our implementation of the bootstrap deserves some discussion. Although we run the bootstrap by resampling (with replacement) each game, there are other plausible fine grained units of replication (e.g. we could resample by individual possessions or according to one-minute intervals). However, unlike baseball, hockey and basketball involve complex continuous dynamics and long-term within-game dependences that make it hard to justify the use of higher resolution units of replication. By resampling games, we replicate any within-game correlations that might affect player metrics. For example, there is a tendency for star players to be benched during “garbage time” at the end of games when the score differential is large, but this event depends on the performance of those players earlier in the game. Because our goal is to estimate variances of functions of season totals, capturing this within-game correlation in the bootstrapping procedure is critical. Finally, we work with games as the unit of replication for practical reasons – it is much simpler to resample at the game level since it does not require access to play-by-play data. In short, there are other resampling methods that replicate within-game correlations, but our approach appears to be the simplest.

With this approach in mind, assuming *P* players, our estimator for discrimination is simply

where *m* over the players in season *s*. Similarly, the stability estimator for a metric *m* is

where *m* for player *p* over all seasons, *S _{p}* is the number of seasons played by player

*p*.

All independence meta-metrics are defined as a function of the latent correlation matrix *C* from the copula model presented in Equation 6. To estimate *C*, we use the semi-parametric rank-likelihood approach developed by Hoff (2007). This method is appealing because we eschew the need to directly estimate the marginal density of the metrics, *F _{m}*. We fit the model using the R package

*sbgcop*(Hoff 2012). Using this software, we can model the dependencies for both continuous and discrete valued metrics with missing values.

In Section 4, we use *m* we first estimate the independence score

## 4 Results

To demonstrate the utility of our meta-metrics, we analyze metrics from both basketball (NBA) and hockey (NHL), including both traditional and “advanced” (model-derived) metrics. We scraped data relevant for over 70 NBA metrics from all players and seasons from the year 2000 onwards from basketball-reference.com (Sports Reference LLC 2016a). We also scraped data for 40 NHL metrics recorded from the year 2006 onwards (Sports Reference LLC 2016b). For both seasons we use regular season data only. We also use the R package *nhlscrapr* to gather NHL gamelog data. Where appropriate, we normalized metrics by minutes played or possessions played to ameliorate the impact of anomalous events in our data range, such as injuries and work stoppages; this approach sacrifices no generality, since minutes/possessions can also be treated as metrics. In the Appendix, we provide a glossary of all of the metrics evaluated in this paper. A repository for the replication code is available on GitHub: https://github.com/afranks86/meta-analytics.

### 4.1 Analysis of NBA metrics

In Figure 1 we plot the stability and discrimination meta-metrics for many of the NBA metrics available on basketball-reference.com. When computing the discrimination and stability meta-metrics, we exclude data from players with fewer than 250 minutes played in a season. This leads to 5182 player-seasons of data from the year 2000 onwards. Discrimination scores are estimated from gamelog data for the year 2015 only. There were no minutes or sample size restrictions for the independence analyses: 7507 player-seasons of data from the year 2000 onwards were used for these analyses.

For basic box score statistics, discrimination and stability scores match intuition. Metrics like rebounds, blocks and assists, which are strong indicators of player position, are highly discriminative and stable because of the relatively large between player variance. As another example, free throw percentage is a relatively non-discriminative statistic within-season but very stable over time. This makes sense because free throw shooting requires little athleticism (e.g., does not change with age or health) and is isolated from larger team strategy and personnel (e.g., teammates do not have an effect on a player’s free throw ability).

Our results also highlight the distinction between pure rate statistics (e.g., per-game or per-minute metrics) and those that incorporate total playing time. Metrics based on total minutes played are highly discriminative but less stable, whereas per-minute or per-game metrics are less discriminative but more stable. One reason for this is that injuries affect total minutes or games played in a season, but generally have less effect on per-game or per-minute metrics. This is an important observation when comparing the most reliable metrics since it is more meaningful to compare metrics of a similar type (rate-based vs total).

WS/48, ORtg, DRtg and BPM metrics are rate-based metrics whereas WS and VORP based metrics incorporate total minutes played (Sports Reference LLC 2016a). WS and VORP are more discriminative than the rate based statistics primarily because MP increases discrimination, not because there is stronger signal about player ability. Rate based metrics are more relevant for estimating player skill whereas total metrics are more relevant for identifying overall end of season contributions (e.g. for deciding the MVP). Since these classes of metrics serve different purposes, in general they should not be compared directly. Our results show moderately improved stability and discriminative power of the BPM-based metrics over other rate-based metrics like WS/48, ORTg and DRtg. Similarly, we can see that for the omnibus metrics which incorporate total minutes played, VORP is more reliable in both dimensions than total WS.

Perhaps the most striking result is the unreliability of empirical three point percentage. It is both the least stable and least discriminative of the metrics that we evaluate. Amazingly, over 50% of the variation in three point percentage between players who take a non-negligible number of three point shots (at least 10) in a given season is due to chance. This is likely because differences between three point shooters’ true shooting percentage tend to be relatively small, and as such, chance variation tends to be the dominant source of variation. Moreover, contextual variation like a team’s ability to create open shots for a player affect the stability of three point percentage.

Finally, we use independence meta-metrics to explore the dependencies between available NBA metrics. In Figure 2 we plot the independence curves described in Section 3. Of the metrics that we examine, steals (STL) appear to provide some of the most unique information. This is evidenced by the fact that the

Not surprisingly, there is a significant amount of redundancy across available metrics. Principal component analysis (PCA) on the full correlation matrix *C* suggests that we can explain over 75% of the dependencies in the data using only the first 15 out of 65 principal components, i.e., *F*_{15} ≈ 0.75. Meanwhile, PCA of the sub-matrix *F*_{1} = 0.75, that is, the first component explains 75% of the variation in these five metrics. This means that much of the information in these 5 metrics can be compressed into a single metric that reflects the same latent attributes of player skill. In contrast, for the defensive metrics presented in Figure 2, *F*_{k} as a function of the number of components *k* for all metrics ℳ and the subsets ℳ_{o} and ℳ_{d}. Figure 8 illustrates a hierarchical clustering of these metrics based on these dependencies.

### 4.2 Analysis of NHL metrics

NHL analytics is a much younger field than NBA analytics, and as a consequence there are fewer available metrics to analyze. In Figure 3A we plot the estimated discrimination and stability scores for many of the hockey metrics available on hockey-reference.com. When computing the discrimination and stability scores, we exclude data from players with fewer than 500 minutes played in a season. This leads to 4291 player-seasons of data from the year 2000 onwards. Discrimination scores are estimated from gamelog data for the year 2015 only. There were no minutes or sample size restrictions for the independence analyses: 7270 player-seasons of data from the year 2000 onwards were used for these analyses.

Again, we find that metrics like hits (HIT), blocks (BLK) and shots (S) which are strong indicators for player type are the most discriminative and stable because of the large between-player variance.

Our results can be used to inform several debates in the NHL analytics community. For example, our results highlight the low discrimination of plus-minus (“±”) in hockey, which can be explained by the relative paucity of goals scored per game. For this reason, NHL analysts typically focus more on shot attempts (including shots on goal, missed shots and blocked shots). In this context, it is often debated whether it is better to use Corsi- or Fenwick-based statistics (Peterson 2014). Fenwick-based statistics incorporate shots and misses whereas Corsi-based statistics additionally incorporate blocked shots. Our results indicate that with the addition of blocks, Corsi metrics (e.g. “CF% rel” and “CF%”) are both more discriminative and stable than the Fenwick metrics.

In Figure 3B we plot the estimated independence scores as a function of the number of statistics in the conditional set for five different metrics. Like steals in the NBA, we found that takeaways (TK) provide the most unique information relative to the other 39 metrics. Here, *oZS*% ∈ ℳ and significantly larger otherwise. This is clear from the large uptick in the independence score of dZS% after removing oZS% from ℳ.

Once again, the analysis of the dependencies among metrics reveals significant redundancy in information across NHL metrics. We can explain over 90% of the variation in the data using only 15 out of 40 principal components, that is *F*_{15} = 0.90 (Figure 10). Figure 4 illustrates a hierarchical clustering of these metrics based on these dependencies.

## 5 Constructing novel metrics

In addition to providing useful benchmarks on the quality of different metrics, the meta-metrics can motivate the design of new and improved metrics or be used to justify the superiority of new metrics over traditional ones. Here we provide two examples in which novel metrics improve upon existing metrics in at least one of the meta-metrics. In the first example, we use a hierarchical model to shrink empirical estimates of three point ability in basketball. We demonstrate that this model-based estimate is both more stable and discriminative than the simple percentage metric. In the second example, we propose a method for creating a set of new metrics that are all mutually independent.

### 5.1 Shrinkage estimators

Model-based adjustments of common box score statistics can reduce sampling variability and thus lead to improvements in discrimination and stability. In Section 4.1, we showed how three point percentage was one of the least discriminative and stable metrics in basketball and thus an improved estimator of three point making ability is warranted. We define three point ability using the notation introduced in Section 2 as *E _{sp(3P%)}*[

*X*] , i.e. the expected three point percentage for player

*p*in season

*s*, and propose a model-based estimate of this quantity that is both more stable and discriminative than the observed percentage.

For this model, we assume an independent hierarchical Bernoulli model for the three point ability of each player:

where *p* in season *s*, *π*_{sp} = *E _{sp(3P%)}*[

*X*] is the estimand of interest,

*n*is the number of attempts,

_{sp}*p*, and

*π*

_{sp}over time. We use the R package

*gbp*for empirical Bayes inference of

*π*

_{sp}and

*r*, which controls the amount of shrinkage (Tak et al. 2016). In Figure 5 we plot the original and shrunken estimates for LeBron James’ three point ability over his career.

_{p}We can compute discrimination and stability estimates for the estimated three point ability derived from this model using the same approach outlined in Section 3. Although the empirical Bayes’ procedure yields probability intervals for all estimates, we can still compute the frequentist variability using the bootstrap (e.g. see Efron (2015)). In Figure 1 we highlight the comparison between observed three point percentage and the empirical Bayes estimate in red. Observed three point percentage is an unbiased estimate of three point ability but is highly unreliable. The Bayes estimate is biased for all players, but theory suggests that the estimates have lower mean squared error due to a reduction in variance (Efron and Morris 1975). The improved stability and discrimination of the empirical Bayes estimate is consistent with this fact.

There are, of course, other shrinkage estimators that may be more appropriate depending on the context. For instance, rather than shrink each estimate toward a player’s career average, we could derive estimators based on shrinkage to the league average for each season. This might be especially reasonable for rookie players or players with little career playing time. More sophisticated extensions might involve shrinking estimates to a regression surface based on covariates like how long the player has been in the league. The choice of shrinkage model should reflect the decision-maker’s classification of observed variation into signal and noise and assumptions about the relationships between those sources of variation. The shrinkage scheme used in this section classifies both chance variation and inter-season variation as noise, and treats player-specific variation is independent.

### 5.2 Principal component metrics

The dependency model proposed in Section 2.3 provides a natural way to derive new metrics that describe orthogonal aspects of player ability. In particular, the eigendecomposition of the latent correlation matrix, *C*, (Equation 6) can be used to develop a (smaller) set of new metrics, which, by construction, are mutually independent and explain much of the variation in the original set. If the latent normal variables *Z* defined in Equation 6 were known, then we could compute the principal components of this matrix to derive a new set of orthogonal metrics. The principal components are defined as *W* = *ZU* where *U* is the matrix of eigenvectors of *C*. Then, by definition, *W* ∼ MVN(0, *I*) and thus *k*, where *W _{j}*,

*j*≠

*k*. We estimate

*Z*by normalizing

*X*, that is

_{m}is the empirical CDF of

*X*. Our estimate of the principal components of the latent matrix

_{m}*Z*is then simply

We present results based on these new PCA-based metrics for both NBA and NHL statistics. In Figure 6 we list three PCA-based metrics for the NBA and the corresponding original NBA metrics which load most heavily onto them. We also rank the top ten players across seasons according to

Although the meaning of these metrics can be harder to determine, they can provide a useful aggregation of high-dimensional measurements of player skill that facilitate fairer comparisons of players.

In Figure 7 we provide two PCA-based metrics for NHL statistics. We again list the metrics that have the highest loadings on two principal component along with the top ten players (in any season) by component. The first principal component largely reflects variation in offensive skill and easily picks up many of the offensive greats, including Ovechkin and Crosby. For comparison, we include another component, which corresponds to valuable defensive players who make little offensive contribution. This component loads positively on defensive point shares (DPS) and blocks (BLK), but negatively on shots and goals (S, G).

## 6 Discussion

Uncertainty quantification, a hallmark of statistical sciences, has so far been under-appreciated in sports analytics. Our work demonstrates the importance of understanding sources of variation and provides a method to quantify how different metrics reflect this variation. Specifically, we explore three different “meta-metrics” for evaluating the reliability of metrics in any sport: discrimination, stability and independence. Our results show that we can use meta-metrics to characterize the most discriminative and stable summaries amongst a set of omnibus metrics (like win shares, BPM and PER for the NBA), which can in turn help decision-makers identify the metrics that meet necessary requirements to be useful for a given task. First, highly non-discriminative metrics should not be used by decision makers to compare players since differences in these metrics reflect chance variation, not variation in ability. Second, metrics that are highly unstable generally should be considered questionable as criteria for player acquisition since these metrics are likely to change significantly over time either in response to a changing context or fluctuating player ability.

Along these lines, our meta-metrics can elucidate precisely which aspects of player skill are transferable and help analysts partition relevant sources of variability. In Section 4.1 we showed how “minutes played” is discriminative and stable and how this can translate to other omnibus metrics which incorporate minutes played. After controlling for minutes played, these metrics may provide little additional stable quantitative value. As another example, three point percentage in basketball is likely sensitive to the “openness” of the shooter, which is in turn a function of teammate ability and strategy.

With this in mind, decision makers should make an effort to use metrics that condition on as much relevant information as possible (like “openness” for the three point example) and then verify that these new metrics are more stable or discriminative and thus more useful for player acquisition decisions. In this sense, meta-metrics can be used as a benchmark for evaluating the improvement of new estimators. In Section 5.1 we provided one example, in which we demonstrate that an estimate based on a simple hierarchical model can improve the stability and discrimination of standard boxscore statistics by reducing chance variability.

Our methods also demonstrate how decision makers can synthesize information across multiple metrics using the independence criterion. Analysts should work with metrics that are roughly independent to avoid falsely interpreting multiple redundant metrics as additional evidence of player ability. We show how to identify the most independent existing metrics as well as demonstrate how to create new metrics that are all mutually independent (Section 5.2).

Finally, in this paper, we focused on reliability and dependence of metrics for *all players in the league* but the meta-metrics can easily be recalculated for relevant subsets of players. This is important because, as shown, in this context the most reliable metrics are often the metrics that distinguish between player types (e.g., blocks and rebounds in basketball). This may be irrelevant when making decisions involving a specific group of players (e.g., which NBA center to acquire). When using metrics to evaluate players of a certain type, we should compute the meta-metrics conditional on this player type. For instance, there is less variation in the number of blocks and rebounds by NBA centers, and as such, these metrics are less discriminative and stable than they are for the league as a whole. Moreover, the dependence between blocks and rebounds is largely driven by height, and thus the conditional dependence between blocks and rebounds given height is much smaller. Further, in our analyses we used sample size restrictions to eliminate players with very few minutes played (or in the case of 3P%, few attempts). Without these restrictions, the meta-metric scores may have largely been driven by differences between high volume and low volume players.

For this reason, it is important that the meta-metrics are always interpreted in the context of the appropriate group of players. In light of this point, it is notable that the meta-metrics that we present in this paper are stated in terms of expectations and variances, so that estimation of conditional meta-metrics simply requires replacing marginal expectations and variances with their conditional counterparts.

Another important consideration is that our meta-metrics only measure the internal quality of a metric. The meta-metrics are not designed to provide any information about how relevant the metrics are for the sport of interest. For instance, although we identified Corsi-based metrics as more discriminative and stable than the related Fenwick-based metrics, it is still possible that Fenwick metrics are more predictive of team performance. As a more extreme example, an athlete’s birthplace zip code would be perfectly discriminative, stable and independent from all other metrics, but is clearly irrelevant for determining a player’s value to the team. This suggests that in practice coaches and analysts should consider a fourth meta-metric: “relevance”. Relevance could simply be a qualitative description of the metric’s meaning and value as determined by domain experts or it could be a quantitative summary of the causal or predictive relationship between the metric and an outcome of interest, like wins or revenue generated. Nevertheless, the methods presented here provide a useful characterization of the reliability of existing metrics. We believe that future iterations of the meta-metrics outlined in this paper can become a standard analytical tool that will improve the decisions made and information gleaned from new and old metrics alike.

This work was partially supported by the Washington Research Foundation Fund for Innovation in Data-Intensive Discovery, the Moore/Sloan Data Science Environments Project at the University of Washington and New York University, U.S. National Science Foundation grants 1461435, by DARPA under Grant No. FA8750-14-2-0117, by ARO under Grant No. W911NF-15-1-0172, by Amazon, and by NSERC. The authors are grateful to Andrew Miller (Department of Computer Science, Harvard University), and Kirk Goldsberry for sharing data and ideas which contributed to the framing of this paper.

## Proof of 0 ≤ *S*_{m} ≤ 1

We calculate stability for metric *m* (4) as

To show 0 ≤ *S*_{m} ≤ 1, it suffices to show both

*E*[_{m}*V*[_{pm}*X*] −*V*[_{spm}*X*]] ≥ 0*V*[_{m}*X*] −*E*[_{m}*V*[_{spm}*X*]] −*E*[_{m}*V*[_{pm}*X*] −*V*[_{spm}*X*]] ≥ 0.

To verify (A), we can write

To check (B), note that

## Glossary of metrics

Glossary of NBA metrics used. All stats are per 36 minutes unless otherwise noted. See (Sports Reference LLC, 2016a) for more detail.

Metric | Description |
---|---|

MP | Minutes played |

FGA | Field goal attempts |

FG% | Field goal percentage |

3PA | 3 point attempts |

3P% | 3 point percentage |

2PA | 2 point attempts |

2P% | 2 point percentage |

FTA | Free throw attempts |

FT% | Free throw percentage |

PF | Personal fouls |

PTS | Points |

PER | Personal efficiency rating |

TS% | True shooting percentage |

3PAr | Three point attempt rate |

FTr | Free throw attempt rate |

ORB | Offensive rebounds |

DRB | Defensive rebounds |

TRB | Total rebounds |

AST | Assists |

STL | Steals |

BLK | Blocks |

TOV% | Turnover percentage (per possession) |

USG% | Usage per |

OWS | Offensive win shares |

DWS | Defensive win shares |

WS | Win shares |

WS/48 | Win shares per 48 minutes |

OBPM | Offensive box plus minus |

DBPM | Defensive box plus minus |

BPM | Box plus minus |

VORP | Value over replacement |

ORtg | Offensive rating |

DRtg | Defensive rating |

Avg Shot Dist | Average shot distance |

NBA Glossary cont.

Metric | Description |
---|---|

%FGA 2P | Percentage of field goal attempts that are 2 pointers |

%FGA 0–3 | Percentage of field goal attempts within 0–3 feet |

%FGA 3–10 | Percentage of field goal attempts within 3–10 feet |

%FGA 10–16 | Percentage of field goal attempts within 10–16 feet |

%FGA 16 < 3 | Percentage of field goal attempts between 16 feet and the 3 point line |

%FGA 3P | Percentage of field goal attempts that are 3 pointers |

%FG 2P | Percentage of made field goals that are 2 pointers |

%FG 0–3 | Percentage of made field goals within 0–3 feet |

%FG 3–10 | Percentage of made field goals within 3–10 feet |

%FG 10–16 | Percentage of made field goals within 10–16 feet |

%FG 16 < 3 | Percentage of made field goals between 16 feet and the 3 point line |

%FG 3P | Percentage of made field goals that are 3 pointers |

%Ast’d 2P | Percentage of made 2 point field goals that are assisted |

%FGA Dunks | Percentage of field goal attempts that are dunks |

Made Dunks | Made dunks (per 36 MP) |

%Ast’d 3P | Percentage of made 3 point field goals that are assisted |

%3PA - Corner | Percentage of 3 point field goal attempts taken from the corner |

3P% - Corner | 3 point field goal percentage from the corner |

OnCourt | Plus/minus per 100 possessions |

On-Off | Plus/minus net per 100 possession |

TOV - BadPass | Turnovers from bad passes |

TOV - LostBall | Turnovers due to lost ball |

TOV - Other | All other turnovers (traveling, out of bounds, etc) |

PF - Shooting | Shooting fouls committed |

PF - Blocking | Blocking fouls committed |

PF - Offensive | Offensive fouls committed |

PF - Take | Take fouls committed |

PGA | Points generated by assists |

SfDrawn | Shooting fouls drawn |

And1 | shots made on fouls drawn |

Blkd | Field goal attempts that are blocked |

Glossary of hockey metrics used.

Metric | Description |
---|---|

G | Goals |

A | Assists |

PTS | Points |

± | Plus / minus |

PIM | Penalties in minutes |

EV | Even strength goals |

PP | Power play goals |

SH | Short handed goals |

GW | Game winning goals |

S | Shots on goal |

S% | Shooting percentage |

TSA | Total shots attempted |

TOI | Time on ice |

FO% | Face off win percentage |

HIT | Hits at even strength |

BLK | Blocks at even strength |

TK | Takeways |

GV | Giveaways |

GC | Goals created |

TGF | Total goals for (while player was on the ice) |

PGF | Power player goals for (while player was on the ice) |

TGA | Total goals against (while player was on the ice) |

PGA | Power player goals against (while player was on the ice) |

OPS | Offensive point shares |

DPS | Defensive point shares |

PS | Total point shares |

CF | Corsi for (on ice shots + blocks + misses) |

CA | Corsi against (on ice shots + blocks + misses) |

CF% | Corsi for percentage: CF / (CF + CA) |

CF% rel | Relative Corsi for (on ice CF% − off ice CF%) |

FF | Fenwick for (shots + blocks + misses) |

FA | Fenwick against (shots + blocks + misses) |

FF% | Fenwick for percentage: FF / (FF + FA) |

FF% rel | Relative Fenwick for (on ice FF% − off ice FF%) |

oiSH% | Team on ice shooting percentage while player on the ice |

oiSV% | Team on ice save percentage while player on the ice |

PDO | Shooting percentage plus save percentage |

oZS% | Percentage of offensive zone starts while on the ice |

dZS% | Percentage of defensive zone starts while on the ice |

All metrics are normalized by total time on ice (TOI) unless otherwise noted.

## References

Arthur, R. 2015. “Stats Can’t Tell Us Whether Mike Trout or Josh Donaldson Should Be MVP.” http://fivethirtyeight.com/features/stats-cant-tell-us-whether-mike-trout-or-josh-donaldson-should-be-mvp/. Accessed on September 30, 2015.

Baumer, B. S., S. T. Jensen, and G. J. Matthews. 2015. “openwar: An Open Source System for Evaluating Overall Player Performance in Major League Baseball.” Journal of Quantitative Analysis in Sports 11:69–84.

Blackport, D. 2014. “How Long Does It Take for Three Point Shooting to Stabilize?” http://nyloncalculus.com/2014/08/29/long-take-three-point-shooting-stabilize/. Accessed on September 30, 2015.

Cervone, D., A. D’Amour, L. Bornn, and K. Goldsberry. 2016. “A Multiresolution Stochastic Process Model for Predicting Basketball Possession Outcomes.” Journal of the American Statistical Association 111:585–599.

Cronbach, L. J. 1951. “Coefficient Alpha and the Internal Structure of Tests.” Psychometrika 16:297–334.

Efron, B. 2015. “Frequentist Accuracy of Bayesian Estimates.” Journal of the Royal Statistical Society: Series B (Statistical Methodology) 77:617–646.

Efron, B. and C. Morris. 1975. “Data Analysis Using Stein’s Estimator and Its Generalizations.” Journal of the American Statistical Association 70:311–319.

Efron, B. and R. Tibshirani. 1986. “Bootstrap Methods for Standard Errors, Confidence Intervals, and Other Measures of Statistical Accuracy.” Statistical Science 1:54–75.

Fisher, R. A. 1925. Statistical Methods for Research Workers. Guildford: Genesis Publishing Pvt Ltd.

Franks, A., A. Miller, L. Bornn, and K. Goldsberry. 2015. “Counterpoints: Advanced Defensive Metrics for NBA Basketball,” in Proceedings of the 2015 MIT Sloan Sports Analytics Conference, MIT Sloan Sports Analytics Conference. Boston, MA.

Hoff, P. 2012. sbgcop: Semiparametric Bayesian Gaussian Copula Estimation and Imputation. URL http://CRAN.R-project.org/package=sbgcop, r package version 0.975.

Hoff, P. D. 2007. “Extending the Rank Likelihood for Semiparametric Copula Estimation.” The Annals of Applied Statistics 1:265–283.

Kuder, G. F. and M. W. Richardson. 1937. “The Theory of the Estimation of Test Reliability.” Psychometrika 2:151–160.

Lewis, M. 2004. Moneyball: The Art of Winning an Unfair Game. New York, NY: WW Norton & Company.

Mardia, K. V., J. T. Kent, and J. M. Bibby. 1980. Multivariate Analysis. San Diego: Academic Press.

Oliver, D. 2004. Basketball on Paper: Rules and Tools for Performance Analysis. Lincoln, NE: Potomac Books, Inc.

Peterson, M. 2014. “Corsi vs. Fenwick: How are They Different and When Do I Use Them?” http://faceoffviolation.com/dekestodangles/2014/11/19/corsi-vs-fenwick-different-use/. Accessed on September 6, 2016.

Sill, J. 2010. “Improved NBA adjusted plus-minus using regularization and out-of-sample testing,” in Proceedings of the 2010 MIT Sloan Sports Analytics Conference.

Sports Reference LLC. 2016a. “Basketball-Reference.com - Basketball Statistics and History.” http://www.basketball-reference.com/.

Sports Reference LLC. 2016b. “Hockey-Reference.com - Hockey Statistics and History.” http://www.hockey-reference.com/.

Sprigings, D. 2014. “donttellmeaboutheart.blogspot.com/ - How Long Does It Take for a Forward’s Shooting to Stabilize?” http://donttellmeaboutheart.blogspot.com/2014/12/how-long-does-it-take-for-forwards.html. Accessed on September 30, 2015.

Tak, H., J. Kelly, and C. N. Morris. 2016. “Rgbp: An R Package for Gaussian, Poisson, and Binomial Random Effects Models with Frequency Coverage Evaluations.” arXiv preprint arXiv:1612.01595.

Tango, T. M., M. G. Lichtman, and A. E. Dolphin. 2007. The Book: Playing the Percentages in Baseball. Lincoln, NE: Potomac Books, Inc.