 Research
 Open Access
 Published:
A relationship between the incremental values of area under the ROC curve and of area under the precisionrecall curve
Diagnostic and Prognostic Research volume 5, Article number: 13 (2021)
Abstract
Background
Incremental value (IncV) evaluates the performance change between an existing risk model and a new model. Different IncV metrics do not always agree with each other. For example, compared with a prescribeddose model, an ovariandose model for predicting acute ovarian failure has a slightly lower area under the receiver operating characteristic curve (AUC) but increases the area under the precisionrecall curve (AP) by 48%. This phenomenon of disagreement is not uncommon, and can create confusion when assessing whether the added information improves the model prediction accuracy.
Methods
In this article, we examine the analytical connections and differences between the AUC IncV (ΔAUC) and AP IncV (ΔAP). We also compare the true values of these two IncV metrics in a numerical study. Additionally, as both are semiproper scoring rules, we compare them with a strictly proper scoring rule: the IncV of the scaled Brier score (ΔsBrS) in the numerical study.
Results
We demonstrate that ΔAUC and ΔAP are both weighted averages of the changes (from the existing model to the new one) in separating the risk score distributions between events and nonevents. However, ΔAP assigns heavier weights to the changes in higherrisk regions, whereas ΔAUC weights the changes equally. Due to this difference, the two IncV metrics can disagree, and the numerical study shows that their disagreement becomes more pronounced as the event rate decreases. In the numerical study, we also find that ΔAP has a wide range, from negative to positive, but the range of ΔAUC is much smaller. In addition, ΔAP and ΔsBrS are highly consistent, but ΔAUC is negatively correlated with ΔsBrS and ΔAP when the event rate is low.
Conclusions
ΔAUC treats the wins and losses of a new risk model equally across different risk regions. When neither the existing or new model is the true model, this equality could attenuate a superior performance of the new model for a subregion. In contrast, ΔAP accentuates the change in the prediction accuracy for higherrisk regions.
Introduction
Risk prediction is crucial in many medical decisionmaking settings, such as managing disease prognosis. Numerous research has been dedicated to continually updating risk models for better prediction accuracy. For example, several papers have investigated the improvement in predicting the risk of cardiovascular disease by adding new biomarkers to the existing Framingham risk model, such as the Creactive protein [1, 2], and more recently, a polygenic risk score [3, 4].
In some applications, an existing marker is replaced with a new marker that provides more precise information. For example, cancer treatment such as radiation can have significant longterm health consequences for cancer survivors. Prescribed radiation doses to body regions, such as the abdomen and chest, are routinely available in medical charts. But to predict the risk of an organspecific outcome, e.g., secondary lung cancer or ovarian failure, a more precise measurement of the radiation exposure to specific organs provides better information. Radiation oncologists developed and applied algorithms to estimate these organspecific exposures [5].
The measurement of a new marker or the more precise measurement of a known risk factor is often costly and timeconsuming. Thus, it is important to verify that the new model indeed has a measurable and better prediction performance than the existing one, and thus, worth the extra resources. A number of metrics have been proposed to evaluate the incremental value (IncV) of the risk model that incorporates the new information. The IncV has primarily been discussed in settings where new markers are added to the existing risk profile [6, 7]. In this paper, the term IncV refers to the change of the prediction performance whenever an existing risk model is compared with a new one.
In medical research, the receiver operating characteristic (ROC) curve has been and remains the most popular tool for evaluating the prediction accuracy of a risk model, dating back to the 1960s when it was applied in diagnostic radiology and imaging systems [8, 9]. The area under the ROC curve (AUC) captures the discriminatory ability of a model, i.e., how well a model separates events (subjects who experience the event of interest) from nonevents (subjects who are eventfree). Recently, the precisionrecall (PR) curve is gaining popularity [10–13]. Originated from the information retrieval community in the 1980s [14, 15], it is a relatively new tool in medical research. The area under the PR curve is called the average positive predictive value or the average precision (AP) [16–18]. Several papers suggest that the PR curve and AP are more informative than the ROC curve and AUC for evaluating the risk model’s prediction performance for an unbalanced outcome, i.e., when the event rate is low ([16, 19, 20]).
Davis and Goadrich established a onetoone correspondence between an ROC curve and a PR curve [21]. When comparing the prediction performance of two risk models, e.g., new versus existing, the ROC curve of the new model dominates the ROC curve of the existing model if and only if the PR curve of the new model dominates the PR curve of the existing model. However, when the ROC and PR curves of the two models cross, it is not uncommon that the IncVs of AUC and AP contradict each other. Clark et al. [22] investigated two models for predicting acute ovarian failure among female childhood cancer survivors. The ovariandose model has a slightly lower AUC but an increased AP by about 48%, compared to the prescribeddose model. The disagreement creates confusion in determining whether the updated risk score improves the prediction accuracy.
In this article, we investigate the analytical connection and difference between the IncVs of AUC and of AP with respect to their true values derived from the underlying data generating mechanism. Unlike previous works investigating the inconsistency between the AUC and AP mainly via simulation studies, our numerical study focuses on the true values, not estimates, of these two IncV metrics. In addition, we examine the effect of the event rate on their (dis)agreement, both analytically and numerically.
Method
Notation and definitions
First, we lay out the notations and define concepts that are used throughout this article. Let D= 0 or 1 denote a binary outcome. For studies with an event time T, define D=I(T≤τ) for a given prediction time period τ, which indicates that the outcome is timedependent. In this article, we refer to subjects with D=1 as the events and those with D=0 as the nonevents. Let π=Pr(D=1) denote the event rate.
Risk model and risk score
A risk model is a function of a set of predictors X=(X_{1},⋯,X_{k−1}), which might include interaction terms and polynomial terms, to obtain the probability of D=1. Usually, we write this model as a regression model:
where g(·) is a smooth and monotonic link function, such as a logit link. For the censored event time outcomes, a risk model could be Cox’s proportional hazards model [23] or the timespecific generalized linear model [24]; both models can be expressed in the general form of Eq. (1) with modifications.
In practice, the underlying data generating mechanism is often complicated, and our working risk model in Eq. (1) is usually misspecified. Let π(X)=Pr(D=1∣X) denote the true probability of D=1, which is determined by the underlying distribution of D given X. Here, we refer to π(X) as the true risk and p(X) as the working risk from a working risk model. When the working risk model in Eq. (1) is misspecified, p(X)≠π(X).
The working risk p(X) can be regarded as a risk score and used to classify subjects into different risk categories. For example, given a cutoff value c, subjects with p(X)≤c are classified into the lowrisk group, whereas the highrisk group consists of subjects with p(X)>c. In general, a risk score, denoted as r(X), can be any function of X that reflects how likely a subject is an event. Thus, r(X) can be a nondecreasing transformation of p(X), e.g., r(X)=g^{−1}(p(X))=β_{0}+β_{1}X_{1}+⋯+β_{k−1}X_{k−1}.
Remark 1
In practice, the working risk p(X) is estimated from a data sample. The estimated regression coefficients \(\widehat \beta _{j},j=0,1,\cdots,k1\), are the solution to an estimating equation: \(\boldsymbol {\Psi }\left (\beta _{0},\cdots,\beta _{k1}\right) = \sum _{i=1}^{n} \Psi \left (\beta _{0},\cdots,\beta _{k1};D_{i},\mathbf {X}_{i}\right)\). The estimated risk given X is \(\widehat {p}(\mathbf {X})=g(\widehat {\beta }_{0}+\widehat {\beta }_{1}X_{1} + \cdots + \widehat {\beta }_{k1}X_{k1})\), which is not of interest here. In this article, we investigate the predictive performance of the population working risk \(p(\mathbf {X})=g\left (\beta _{0}^{\ast }+\beta _{1}^{\ast }X_{1} + \cdots + \beta _{k1}^{\ast }X_{k1}\right)\) where \(\beta _{j}^{\ast }\)’s are the solution of E_{(D,X)}[Ψ(β_{0},⋯,β_{k−1})]=0 with the expectation taken under the true joint distribution of (D,X), and \(\beta _{j}^{\ast }={\lim }_{n\rightarrow \infty }\widehat {\beta }_{j}\).
Accuracy measures and IncV metrics
The AUC and AP can be defined on any risk score r(X) since they are rankbased. The ROC curve is a curve of the true positive rate (TPR) versus the false positive rate (FPR). Given a cutoff value c, the TPR and FPR are the proportions of higherrisk score r(X)>c among the events and nonevents respectively, i.e., TPR(c)=Pr[r(X)>c∣D=1] and FPR(c)=Pr[r(X)>c∣D=0]. The AUC can be interpreted as the conditional probability that given a pair of an event and a nonevent, the event is assigned with a higherrisk score, i.e., AUC=Pr[r(X_{i})>r(X_{j})∣D_{i}=1,D_{j}=0].
The PR curve is a curve of the positive predictive value (PPV) versus the TPR. The PPV is defined as PPV(c)=Pr[D=1∣r(X)>c], the proportion of subjects with higherrisk scores that are events. The AP can be expressed as AP=E[PPV(r_{1}(X))] [18], where r_{1}(X) denotes the risk score of an event, and the expectation is taken under the distribution of r_{1}(X). The AP is eventrate dependent [18]; in contrast, the AUC does not depend on π since it is conditional on the event status.
Let Ψ_{old} and Ψ_{new} denote an accuracy measure Ψ (e.g., AUC or AP) of the existing and new risk models, respectively. The IncV is defined as ΔΨ=Ψ_{new}−Ψ_{old}, which quantifies the change in Ψ when comparing the new model with the existing one.
Data example
Accurate ovarian failure (AOF) is a treatment associated complication caused by ovarian exposure to radiation and chemotherapy. It is defined as permanent loss of ovarian function within 5 years of a cancer diagnosis or no menarche after cancer treatment by age 18. About 6% of female childhood cancer survivors have AOF. We evaluate and compare two recently published risk models [22] that predict AOF on an external validation dataset, the St. Jude Lifetime Cohort [25], which consists of 875 survivors with 50 AOF events.
Both models include the following risk factors: age at cancer diagnosis, cumulative dose of alkylating drugs measured using the cyclophosphamideequivalent dose, hematopoietic stemcell transplant, and radiation exposure. The difference between the two models is in the measurement of radiation exposure. The prescribeddose model uses the prescribed radiation doses to the abdominal and pelvic regions, which are routinely available in medical charts. The ovariandose model uses the minimum of the organspecific radiation exposure for both ovaries estimated by radiation oncologists. The equation for calculating the AOF risk using each model is developed using the Childhood Cancer Survivors Study and given in the supplementary material of Clark et al. (2020) [22].
Figure 1a and b show the ROC curves and PR curves of these two models. The estimated AUC is 0.96 for the prescribeddose model and 0.94 for the ovariandose model; ΔAUC is estimated to be − 0.02. The estimated AP is 0.46 for the prescribeddose model and 0.68 for the ovariandose model. The estimated ΔAP is 0.22. The estimation procedure is explained in Appendix.
Based on the ΔAUC, we conclude that the more expensive ovary dosimetry does not improve the prediction accuracy at all. However, based on the ΔAP, the ovariandose model clearly outperforms the prescribeddose model. Why do these two metrics give conflicting conclusions?
Analytical comparisons between ΔAUC and ΔAP
To answer this question, we first investigate the connections and differences between the AUC and AP using the following three hypothetical risk scores: r_{1},r_{2}, and r_{3}. We assume that all the risk scores among nonevents follow a standard normal distribution, i.e., r_{j}∣D=0∼N(0,1), for j=1,2,3. However, their distributions among events are different: (i) r_{1}∣D=1∼N(1.8,2), (ii) r_{2}∣D=1∼N(1.5,1.5), and (iii) r_{3}∣D=1∼N(3,1.5).
Figure 2 presents the comparisons of these three risk scores under an event rate π=0.05. Figure 2a shows their density curves stratified by events and nonevents. Among them, the two density curves of r_{3} are the most separated. Thus, the ROC and PR curves of r_{3} dominate those of r_{1} and r_{2} (Fig. 2b and c), and consequently, r_{3} has the largest AUC and AP. In contrast, the ROC and PR curves of r_{1} and r_{2} cross: r_{2} has a slightly larger AUC with \(\text {AUC}_{r_{2}}  \text {AUC}_{r_{1}} = 0.007\), but r_{1} has a considerably larger AP with \(\text {AP}_{r_{1}}  \text {AP}_{r_{2}} = 0.096\). Figure 3 exhibits the comparisons between r_{1} and r_{2} for three different event rates π=0.2, 0.05, and 0.01.
Analytically, both the AUC and AP measure the separation of the risk score distributions between events and nonevents. Let F_{1}(·) and F_{0}(·) denote the cumulative distribution functions (CDFs) of a risk score r(X) conditional on D=1 (events) and D=0 (nonevents), respectively. Let \(q_{\alpha }=F_{1}^{1}(\alpha)\) denote the αth quantile for the distribution F_{1},0≤α≤1. As shown in Eqs. (7) and (8) of Appendix, the AUC and AP can be expressed as functions of F_{0}(q_{α}), the proportion of nonevents whose risk scores are below the αth quantile of the risk scores among events. The F_{0}(q_{α}) measures the separation of the two distributions F_{1} and F_{0}: the larger the F_{0}(q_{α}) is at a given α, the more nonevents having lowerrisk scores, indicating a further separation between these two distributions. For example, the F_{0}(q_{α}) curve of r_{3} dominates those of r_{1} and r_{2} (Fig. 2d), which is consistent with the fact that r_{3} has the best separation between events and nonevents (Fig. 2a).
Furthermore, we can express both ΔAUC and ΔAP as
where w_{Ψ}(α) is a weight function, and Δ(α)=F_{new,0}(q_{new,α})−F_{old,0}(q_{old,α}), capturing how much the new working risk model changes the separation of these two distributions at a given α. Note that Δ(α) is independent of π because it is conditional on the event outcome. Thus, ΔAUC and ΔAP are weighted averages of Δ(α), but their weights are different. For ΔAUC,w_{AUC}(α)≡1 for 0≤α≤1, i.e., Δ(α) is equally weighted. For ΔAP,w_{AP}(α) is a function of α and π (Eq. (9) of Appendix).
To visualize how w_{AP}(α) changes with α and π, we plot the w_{AP}(α) in a log scale against α for different π in Fig. 3a, in the context of comparing the hypothetical risk scores r_{1} and r_{2}. For any π,w_{AP}(α) increases with α. This tells us that ΔAP assigns heavier weights to the uppertail quantiles of the risk score, representing higherrisk regions, and lighter weights to lowertail quantiles, representing lowerrisk regions, i.e., ΔAP emphasizes the change of the separation in higherrisk regions. However, the change is equally weighted in ΔAUC since w_{AUC}(α)≡1.
Additionally, w_{AP} is affected by π. When π is smaller, w_{AP}(α) is larger for α values close to 1 but smaller for α values close to 0. This indicates that, at a lower event rate, if a risk model can better separate the two risk score distributions at the upper quantiles, it will be rewarded more; if it has a worse separation at lower quantiles, it will be penalized less.
Hypothetic risk scores r _{1} and r _{2} revisited
Assuming that r_{2} is from an existing risk model and r_{1} is from a new one, \(\Delta (\alpha)=F_{r_{1},0}\left (q_{r_{1},\alpha }\right)  F_{r_{2},0}\left (q_{r_{2},\alpha }\right),\Delta \text {AUC} = \text {AUC}_{r_{1}}  \text {AUC}_{r_{2}}\), and \(\Delta \text {AP} = \text {AP}_{r_{1}}\text {AP}_{r_{2}}\). As shown in Fig. 3b, Δ(α)>0 for large α, and Δ(α)<0 for small α. It indicates that compared to r_{2},r_{1} has a better separation for the upper quantiles of the risk score but worse for lower quantiles. With equal weighting, ΔAUC is equivalent to the area under Δ(α) curve over its entire range. Since the area above 0 is approximately the same as the area below 0, ΔAUC≈0. As mentioned earlier, ΔAUC is invariant for different π. Thus, ΔAUC=−0.007 (Fig. 2b) for all three π values.
For ΔAP, the r_{1}’s uppertail better performance is weighted more than its lowertail worse performance, which explains ΔAP is all positive for the three π values (Fig. 3c). Additionally, when π gets smaller, the better separation of r_{1} at the upper quantiles is rewarded more, and meanwhile, its worse separation at lower quantiles is penalized less. Thus, even though Δ(α) stays the same across different π,ΔAP increases as π decreases (Fig. 3c).
Data example revisited
Let Δ(α)=F_{ovarian,0}(q_{ovarian,α})−F_{prescribed,0}(q_{prescribed,α}). Figure 1c plots the estimated Δ(α),w_{AP}(α), and w_{AP}(α)Δ(α). It shows that the estimated Δ(α)>0 for α>10%, whereas the prescribeddose model performs better with the estimated Δ(α)<0 when α<10%. It suggests that compared to the prescribeddose model, the ovariandose model separates the events and nonevents better among individuals predicted to be at a higher risk. Overall, under the estimated Δ(α) curve, the area below zero is slightly larger than the area above zero. Thus, the estimated ΔAUC is negative but close to zero. This indicates that these two models have comparable performance in terms of ΔAUC.
However, the estimated ΔAP rewards the superior performance of the ovariandose model at the upper quantiles with large weights, and thus, it is positive and sizable. Clark et al. [22] created four risk groups: low (< 5%), mediumlow (5% to < 20%), medium (20% to < 50%), and high risk (≥ 50%). The ovariandose model classifies 37 individuals (out of 875) as high risk, among which 30 (81%) experienced AOF, while the prescribeddose model predicted 13 individuals at high risk, with 6 (46%) AOF events. This again confirms that the ovariandose model is better at identifying the AOF events.
Comparison with Brier score.
Since both the AUC and AP are rankbased, they are semiproper scoring rules: the true model has the maximum AUC and AP among all the models, but a misspecified risk model and the true model can have the same AUC and AP when they rank the subjects’ risks in the same order. We decide to compare these two metrics with the Brier score (BrS), the only strictly proper scoring rule.
The BrS is the expected squared difference between the binary outcome D and the working risk p(X), i.e., BrS=E_{(D,X)}{[D−p(X)]^{2}}. The BrS is minimized at the true model, i.e., p(X)=π(X). A noninformative model, assigning the event rate to every subject, i.e., p(X)≡π, leads to the maximum BrS value π(1−π). A scaled Brier score (sBrS) is defined as sBrS=1−BrS/[π(1−π)], ranging from 0 and 1, with larger values indicating better performance [26].
Remark 2
Although the BrS cannot be directly expressed as a function of F_{0}(q_{α}), it is closely related to the two distributions F_{1} and F_{0}. Specifically, it can be written as
The first expectation is the mean squared prediction error (MSPE) of the working risk p(X) for events, determined by the distribution F_{1}, whereas the second expectation is the MSPE for nonevents, determined by the distribution F_{0}. Both MSPEs can be expressed as the sum of the variance of p(X) and its squared bias from 1 for events and from 0 for nonevents. A smaller BrS can result from one, or a combination, of the following: (i) the mean of p(X)for events closer to 1, (ii) the mean of p(X)for nonevents closer to 0, (iii) less variation in p(X) for events or nonevents or both. All of these lead to a further separation of the two distributions: F_{1} and F_{0}.
Let ΔsBrS denoted the IncV of sBrS. The sBrS is estimated to be 0.23 for the prescribeddose model and 0.50 for the ovariandose model, and ΔsBrS is estimated to be 0.27. Thus, similar to ΔAP,ΔsBrS favors the ovariandose model.
Why are ΔsBrS and ΔAP consistent in this example? Figure S1 of the supplementary material shows the histogram of the predicted risk \(\widehat p_{i}\) from each model among the AOF and nonAOF individuals. For the nonAOF individuals, the risk score distributions of these two models are similar. Consequently, the mean and variance of \(\widehat p_{i}\) for both models are also similar: the mean is 0.033 for the ovariandose model and 0.042 for the prescribeddose model; their variances are both about 0.0053. The MSPE for the ovariandose model is 0.0064, slightly lower than 0.0071 for the prescribeddose model.
For the AOF events, the risk score distribution of the ovarian model has a heavier right tail. This indicates that the ovariandose model pushes more AOF events to the highrisk group. As a result, the mean of \(\widehat p_{i}\) for the ovariandose model is 0.48, much closer to 1 than 0.23 for the prescribeddose model. The variance is 0.10 for the ovariandose model and 0.023 for the prescribeddose model. The MSPE of the ovariandose model is 0.367, much smaller than 0.613 of the prescribeddose model. Combining the MSPEs for events and nonevents weighted by their respective proportions, the estimated BrS for the ovariandose model is 0.027, which is smaller than 0.042, the estimated BrS for the prescribeddose model.
This data example illustrates a comparison of the three IncV metrics: ΔAUC,ΔAP, and ΔsBrS. Next, we expand the comparison via a numerical study.
Numerical study
As we are interested in the true values of the IncV for the population working risk (described in Remark 1), not in the IncV estimates from a sample, we do not use simulation studies; there are no data or samples involved. The numerical study in this section evaluates the IncV of adding a marker, denoted by Y, to a model with an existing marker, denoted by X. The true value of each IncV metric is directly derived from the distributional assumptions described below.
Let the markers X and Y be independent standard normal random variables. Given the values of these two markers, a binary outcome D follows a Bernoulli distribution with the probability of D=1 via the following model:
where Φ(·) is the CDF of a standard normal distribution. Given X and Y, π(X,Y) is the true risk. The true model in Eq. (3) includes an interaction between X and Y, indicating the effect of X on the risk changes with the value of Y, and vice versa.
Typically, in practice, none of the working models are the true model. Having this in mind, we compare the following two misspecified working models: (i) onemarker model: p(X)=Φ(γ_{0}+γ_{1}X), and (ii) twomarker model: p(X,Y)=Φ(γ_{0}+γ_{1}X+γ_{2}Y).
Here, we consider different values of β_{1},β_{2},β_{3} and π: β_{1}=0.3,0.4,⋯,0.9,1,β_{2}=0.3,0.4,⋯,0.9,1,β_{3}=− 0.5,− 0.4,⋯,− 0.1,0.1,⋯,0.4,0.5 (excluding 0), and π=0.01, 0.05, 0.1, 0.2, 0.5. Each combination of (β_{1},β_{2},β_{3},π) values is referred to as a scenario. Given a scenario, the value of β_{0} can be derived. In the supplementary material, we explain how to obtain the value of β_{0} and calculate the true values of AUC, AP, and sBrS of the onemarker and twomarker models as well as the true values of the IncV metrics.
Results
We compare the three IncV metrics based on the following two aspects: (1) size and range, and (2) agreement. A desirable IncV metric should be sensitive to the change in the predictive performance. If a new model improves the prediction accuracy, an IncV should have a sizable positive value. It should also be able to reflect a performance deterioration with a sizable negative value. If an IncV is often close to 0, we might question its utility in supporting decisionmaking. As mentioned earlier, inconsistency among different accuracy metrics is often encountered. Thus, we are also interested in the agreement among the three IncV metrics.
Size and range
Figure 4 plots the summary statistics (minimum, 25% quantile, median, 75% quantile, and maximum) of the three IncV metrics under different event rates. ΔAP has the widest range, followed by ΔsBrS, and ΔAUC has the narrowest range. This difference between ΔAUC and ΔAP is more evident for a lower event rate. For example, under π=0.01, the interquartile range (IQR) and median of ΔAUC are both 0.07. In contrast, the IQR of ΔAP is much wider, with a range of about 0.41 and a median of 0.21.
In addition, ΔAUC is negative in less than 1% of the scenarios (29 out of 3200). Furthermore, when it is negative, the value is very close to 0, which indicates that ΔAUC cannot distinguish between a useless marker and a harmful marker [27]. On the other hand, ΔAP is negative in about 12% of the scenarios (389 out of 3200), with a much larger size.
As π changes, the range of ΔAP varies the most among the three IncV metrics, whereas the quartiles of ΔAUC remain almost constant. As π increases, the ranges of all the IncV metrics get narrower and closer to each other. When π=0.5, both ΔAUC and ΔAP range from 0.015 to 0.25 with a median of 0.089, and ΔsBrS ranges from 0.019 to 0.32 with a median of 0.12.
Agreement
Correlation
We calculate the Pearson correlation between each pair of the IncV metrics under each π (Table 1). ΔAP and ΔsBrS are highly correlated for all values of π. As π increases, their correlation decreases from about 1 (π=0.01) to 0.84 (π=0.5), but the correlations of ΔAUC with the other two IncV metrics increases with π. When π=0.01,ΔAUC and ΔsBrS are negatively correlated and their correlation − 0.11 is the smallest among the three pairs; when π=0.5, they are the highest positively correlated. We also show the scatter plots of each pair under different π in Figure S7 (supplementary material).
Concordance
The sign of an IncV metric is often used to decide whether the new model is more accurate than the existing one. Positive IncVs favor the new model, while negative or zero values favor the existing one. Here, we define a concordance measure, which quantifies the consistency of the conclusions reached by a pair of IncV metrics.
Take ΔAP and ΔsBrS as an example. Under a scenario, we call the pair concordant if both are > 0 or ≤ 0. If one is > 0 and the other is ≤ 0, the pair is discordant. The measure of concordance is defined as the proportion of scenarios where the pair is concordant minus the proportion of scenarios where it is discordant. For instance, when π=0.01,ΔAP and ΔsBrS are concordant in about 97% of the total 640 scenarios (i.e., all the combinations of β_{1},β_{2}, and β_{3} values at each π) and discordant in about 3%. Thus, the concordance measure is 0.93 with a roundoff error.
Table 1 reports the concordance for all three pairs of the IncV metrics under each π. The results are similar to those above for the Pearson correlation. When π is small, such as 0.01, 0.05, and 0.1, ΔAP and ΔsBrS are the most concordant; when π=0.2 or 0.5, ΔAUC and ΔsBrS are the most concordant. ΔAUC and ΔAP are the least concordant for all values of π.
When π is close to 0.5, the three IncV metrics tend to agree. Using any of them, we would very likely reach the same conclusion about whether the new model is more accurate. However, when the event rate is low, i.e., for a rare outcome, ΔAUC can be inconsistent with both ΔsBrS and ΔAP.
ΔAUC versus ΔAP in selected scenarios
Next, we single out four scenarios for an indepth comparison between ΔAUC and ΔAP at π=0.01. The first two scenarios have similar ΔAUC but different ΔAP (Fig. 5), whereas the next two have similar ΔAP but different ΔAUC (Fig. 6).
Similar ΔAUC but different ΔAP
The two scenarios are (i) β_{1}=1,β_{2}=0.8, and β_{3}=0.2, and (ii) β_{1}=1,β_{2}=0.8, and β_{3}=−0.5. In both cases, ΔAUC is around 0.06, but ΔAP is 0.33 for scenario (i) and − 0.072 for scenario (ii).
In scenario (i), both the ROC and PR curves of the twomarker model dominate those of the onemarker model, respectively. This indicates that the twomarker model is better at each point, and consequently, Δ(α) is positive throughout (Fig. 5c). In this case, both ΔAUC and ΔAP are positive. However, the size of ΔAP 0.33 is much larger than ΔAUC 0.06, due to the large weight w_{AP}(α) at the upper quantiles (Fig. 5c).
In scenario (ii), both the two ROC curves and PR curves cross, and Δ(α) is below zero for upper quantiles and above zero for lower quantiles (Fig. 5c). This implies that the twomarker model can better separate between events and nonevents for lowerrisk regions, but not for higherrisk regions. As a result, ΔAUC and ΔAP are conflicting. ΔAUC is positive because the area under Δ(α) curve above zero is larger than that below zero. However, ΔAP is negative, as it weights the belowzero Δ(α) heavily.
Similar ΔAP but different ΔAUC
The next two scenarios are (iii) β_{1}=0.7,β_{2}=0.3, and β_{3}=− 0.3, and (iv) β_{1}=0.6,β_{2}=0.7, and β_{3}=−0.4. In both cases, ΔAP values are almost 0, but ΔAUC is approximately 0 for scenario (iii) and 0.202 for scenario (iv).
In scenario (iii), the two ROC curves and the two PR curves are almost identical. This indicates that adding the new marker does not change the separation of the distributions of the risk score between events and nonevents. It is also reflected in Fig. 6c where the entire Δ(α) curve almost overlaps with the zero line. Thus, both ΔAUC and ΔAP are close to zero. This is an example of both metrics agreeing that the new marker is “useless.”
In scenario (iv), although the twomarker model makes poorer predictions for higherrisk regions, its prediction is significantly better for the rest. Thus, ΔAUC is positive and sizable. However, since ΔAP weights heavily on higherrisk regions, the improvement on the majority is offset by the worse performance at the upper quantiles, which leads to a closetozero ΔAP.
What if the twomarker model is the true model, i.e., β _{3} = 0?
Figures S8 and S9 in the supplementary material examine this question and show the scatter plots and plots of the summary statistics of ΔAUC,ΔAP, and ΔsBrS for different π. As expected, all the IncVs are positive. For a smaller π,ΔAP ranges wider than ΔAUC does. As π increases, these two metrics get closer to each other. When π=0.5,ΔsBrS has the widest range.
Since all the IncVs are positive, their concordance is all 1. Table S1 (supplementary material) lists the Pearson correlation between each pair of the IncV metrics, which are all positive. When π is small, ΔsBrS is more strongly correlated with ΔAP than with ΔAUC. As π increases, all three IncV metrics are strongly correlated with each other.
Discussion
Pepe et al. (2013) proved that, when one of the two working models is the true model, the hypothesis H_{0}:p(X,Y)=p(X) is equivalent to the hypotheses of no improvement in the accuracy measures such as the AUC, net reclassification index (NRI), or integrated discrimination improvement (IDI) [6]. In their setting, the ROC and PR curves never cross. However, our paper focuses on situations where neither working model is the true model, and the two curves might cross. When they cross each other, the above equivalence among the hypotheses does not hold, and it implies that one of the two models outperforms the other at nonoverlapping risk regions. This could lead to the disagreement between ΔAUC and ΔAP.
ΔAUC has been criticized for being insensitive to the contribution from an added marker [28]. According to our analysis, the insensitivity is likely a result of its equal treatment across different risk regions, and thus, it often fails to reflect the “local” improvement or deterioration of the new risk model. In the AOF example, the ovariandose model demonstrates its superiority in higherrisk regions. However, this advantage disappears in ΔAUC, which takes a simple average over the ovariandose model’s wins in higherrisk regions and its losses in lowerrisk regions.
Similarly, if we consider a curve of negative predictive values (NPV, the proportion of nonevents among subjects having a lowerrisk score than a cutoff value) versus specificity (1− FPR), following our derivation of AP, the area under this curve can be expressed as E[NPV(r_{0}(X))] where r_{0}(X) denotes the risk score of a nonevent subject. We can regard this quantity to be the average NPV. Similar to ΔAP, the IncV of the average NPV, ΔaNPV, can be expressed as a weighted average of the change in the separation of the risk score distributions between events and nonevents. However, its weight is larger for lowertail quantiles of the risk score, indicating the average NPV emphasizes on the accuracy of lowerrisk regions.
Assessing the change in prediction accuracy is important in investigating the potential of a new marker (or a new measurement for an existing marker) [29]. However, neither ΔAUC or ΔAP considers the cost and benefit associated with the clinical utility of risk prediction [29, 30]. Going back to the AOF example, should more expensive ovary dosimetry be used for predicting AOF because it identifies more AOF cases? Unfortunately, both ΔAP and ΔAUC are insufficient to answer this question. Vickers and Elkin [29] proposed a net benefit and decision curve analysis for evaluating the clinical value of a risk model. The net benefit is defined as \(NB(p_{t}) = \pi _{1} \text {TPR}(p_{t}) (1\pi _{1})\text {FPR}(p_{t})\frac {p_{t}}{1p_{t}}\), quantifying the net benefit for subjects who are treated based on the rule that the risk probability is above the threshold value p_{t}.
We can express the above net benefit as a function of PPV: \(NB(p_{t})=Pr\left [p(\mathbf {X})>p_{t}\right ] \frac {PPV(p_{t})p_{t}}{1p_{t}}\). Here, Pr[p(X)>p_{t}] is the proportion of subjects who receive the treatment among the population, and \(\frac {PPV(p_{t})p_{t}}{1p_{t}}\) quantifies the expected net benefit given that a subject is treated. The net benefit is regarded as the scaled “average benefit per prediction” [31, 32], and thus, \(\frac {PPV(p_{t})p_{t}}{1p_{t}}\) is the average benefit per treated subject. Thus, ΔNB(p_{t}) is determined by the change of the proportion of treated subjects between the two models and ΔPPV(p_{t}). The analytical relationship between ΔNB and other IncV metrics such as ΔPPV(p_{t}) and ΔAP is worth further investigation.
Because the ranges of AUC, AP, and sBrS are different, the domains of their IncV metrics are also different: ΔAUC∈[− 0.5,0.5],ΔAP∈[π−1,1−π], and ΔsBrS∈[− 1,1]. It may be worthwhile to consider rescaling these IncV metrics to range from − 1 and 1. Alternatively, an IncV metric can be defined as a ratio such as Ψ_{new}/Ψ_{old}.
The AUC is conditional on the binary outcome, and consequently, only depends on the respective risk score distributions among the events and nonevents. Thus, it can be estimated from either a prospective cohort study or a casecontrol study. In contrast, the AP is conditional on the risk score obtained at baseline. Besides the risk score distributions, the AP also depends on the event rate, and thus, it has previously only been possible to estimate from cohort studies, but not from casecontrol studies. However, if one can acquire the information on the event rate from a previous cohort study or from surveillance data, the AP can be estimated via combining an estimated or assumed event rate with the risk score distributions of events and nonevents estimated from the casecontrol study using the derived expression of the AP (see Eq. (8) in Appendix).
Conclusion
In this article, we investigated the disagreement between two IncV metrics ΔAUC and ΔAP when neither the existing nor the new risk model is the true model. We showed that they are intrinsically connected; both can be expressed as an average of Δ(α), a quantity characterizing the change in the separation of the risk score distributions between events and nonevents when comparing an existing risk model to a new one. However, ΔAP is a weighted average, with weights monotonically increasing as the risk score increases, whereas ΔAUC is a simple average of the change. Due to this difference, they do not always agree with each other; the lower the event rate is, the more these two metrics disagree. In addition, compared to ΔAUC,ΔAP has a wider range and is subsequently more sensitive to the contribution from new information added to the existing risk model. Via the numerical study, we also show that ΔAP and ΔsBrS are highly consistent, but the correlation of ΔAUC and ΔsBrS transitions from a positive correlation to a negative one as the event rate decreases.
Appendix
Estimation of AUC, AP, and sBrS for binary outcomes
Suppose that the data \(\mathfrak {D}=\{(D_{i},\mathbf {X}_{i}),i=1,\cdots,n\}\) is collected from n subjects. Let \(\widehat {p}_{i}\) denoted the estimated risk, described in Remark 1. Let \(\widehat {r}_{i}\) be a risk score, which is a nondecreasing transformation of \(\widehat {p}_{i}\). The AUC and AP are estimated using \(\widehat {r}_{i}\) by the following nonparametric estimators
and
The BrS can be estimated using \(\widehat {p}_{i}\) by \(\widehat {\text {BrS}} = n^{1}\sum _{i=1}^{n} \left (D_{i}\widehat {p}_{i}\right)^{2}\). The event rate is estimated as \(\widehat {\pi } = n^{1}\sum _{i=1}^{n} D_{i}\). Then the sBrS is estimated as \(\widehat {\text {sBrS}} = 1  \widehat {\text {BrS}}/\left [\widehat {\pi }\left (1\widehat {\pi }\right)\right ]\).
Derivation of AUC and AP
Let π=Pr(D=1) be the event rate, and r(X)=r_{X} be a risk score. Let F(c)=Pr(r_{X}≤c) denote its cumulative distribution function (CDF) for the entire population, and F_{1}(c)=Pr(r_{X}≤c∣D=1) and F_{0}(c)=Pr(r_{X}≤c∣D=0) denote its CDFs for events and nonevents, respectively.
The TPR, FPR, and PPV are
where 1−F(c)=π[1−F_{1}(c)]+(1−π)[1−F_{0}(c)].
AUC is the area under the ROC curve, which can be expressed as
because \(\int _{\infty }^{\infty } d \text {TPR}(c) = 1\). Using the expressions in Eqs. (4) and (5), we have
Let \(q_{\alpha }=F_{1}^{1}(\alpha)\) be the αth quantile of the F_{1} distribution, i.e., F_{1}(q_{α})=α. Thus, let c=q_{α}, and we have
AP is the area under the PR curve, which can be expressed as
Using the expressions in Eqs. (4) and (6), we have
Again, let c=q_{α}, we have
Weight w _{AP} in ΔAP
Let AP_{old} and α_{new} denote the AP of the existing and new models:
Thus, with arithmetic operations, the IncV of AP can be expressed as
where
It is a function of α and π. It also depends on F_{new,0}(q_{new,α}) and F_{old,0}(q_{new,α}). In general, F_{0}(q_{α})≥α because the density curve for nonevents is to the left of that for events. Thus, how the weight changes with α and π is mainly determined by the numerator (π^{−1}−1)/(1−α). However, when π and α are fixed, larger values of F_{old,0}(r_{old,α}) or F_{new,0}(r_{new,α}) or both, i.e., better performance of at least one model, lead to larger weights.
Availability of data and materials
The R code for the numerical study and analyzing the data example is available in https://github.com/michellezhou2009/IncVAUCAP.
Abbreviations
 AOF:

Acute ovarian failure
 AP:

Average positive predictive value or average precision
 AUC:

Area under the ROC curve
 BrS:

Brier score
 CDF:

Cumulative distribution function
 FPR:

False positive rate
 IDI:

Integrated discrimination improvement
 IncV:

Incremental value
 NB:

Net benefit
 NRI:

Net reclassification index
 PPV:

Positive predictive value
 PR:

Precisionrecall
 ROC:

Receiver operating characteristic
 sBrS:

Scaled Brier score
 TPR:

True positive rate
References
Cook NR, Buring JE, Ridker PM. The effect of including creactive protein in cardiovascular risk prediction models for women. Ann Intern Med. 2006; 145(1):21–9.
Buckley DI, Fu R, Freeman M, Rogers K, Helfand M. Creactive protein as a risk factor for coronary heart disease: a systematic review and metaanalyses for the us preventive services task force. Ann Intern Med. 2009; 151(7):483–95.
Mosley JD, Gupta DK, Tan J, Yao J, Wells QS, Shaffer CM, Kundu S, RobinsonCohen C, Psaty BM, Rich SS, et al. Predictive accuracy of a polygenic risk score compared with a clinical risk score for incident coronary heart disease. JAMA. 2020; 323(7):627–35.
Elliott J, Bodinier B, Bond TA, ChadeauHyam M, Evangelou E, Moons KG, Dehghan A, Muller DC, Elliott P, Tzoulaki I. Predictive accuracy of a polygenic risk score–enhanced prediction model vs a clinical risk score for coronary artery disease. JAMA. 2020; 323(7):636–45.
Howell RM, Smith SA, Weathers RE, Kry SF, Stovall M. Adaptations to a generalized radiation dose reconstruction methodology for use in epidemiologic studies: an update from the md anderson late effect group. Radiat Res. 2019; 192(2):169–88.
Pencina MJ, D’Agostino Sr RB, D’Agostino Jr RB, Vasan RS. Evaluating the added predictive ability of a new marker: from area under the roc curve to reclassification and beyond. Stat Med. 2008; 27(2):157–72.
Pepe MS, Kerr KF, Longton G, Wang Z. Testing for improvement in prediction model performance. Stat Med. 2013; 32(9):1467–82.
Zweig MH, Campbell G. Receiveroperating characteristic (roc) plots: a fundamental evaluation tool in clinical medicine. Clin Chem. 1993; 39(4):561–77.
Pepe MS. The Statistical Evaluation of Medical Tests for Classification and Prediction. Oxford: Oxford University Press; 2003.
Badawi O, Liu X, Hassan E, Amelung PJ, Swami S. Evaluation of icu risk models adapted for use as continuous markers of severity of illness throughout the icu stay. Crit Care Med. 2018; 46(3):361–7.
Chaudhury S, Brookes KJ, Patel T, Fallows A, GuettaBaranes T, Turton JC, Guerreiro R, Bras J, Hardy J, Francis PT, et al. Alzheimer’s disease polygenic risk score as a predictor of conversion from mildcognitive impairment. Transl Psychiatry. 2019; 9(1):1–7.
Tang M, Hu P, Wang CF, Yu CQ, Sheng J, Ma SJ. Prediction model of cardiac risk for dental extraction in elderly patients with cardiovascular diseases. Gerontology. 2019; 65(6):591–8.
Xiao J, Ding R, Xu X, Guan H, Feng X, Sun T, Zhu S, Ye Z. Comparison and development of machine learning tools in the prediction of chronic kidney disease progression. J Transl Med. 2019; 17(1):119.
Raghavan V, Bollmann P, Jung GS. A critical investigation of recall and precision as measures of retrieval system performance. ACM Trans Inf Syst (TOIS). 1989; 7(3):205–29.
Manning CD, Schütze H. Foundations of Statistical Natural Language Processing. USA: MIT Press; 1999.
Yuan Y, Su W, Zhu M. Thresholdfree measures for assessing the performance of medical screening tests. Front Public Health. 2015; 3:57.
Su W, Yuan Y, Zhu M. A relationship between the average precision and the area under the roc curve. In: Proceedings of the 2015 International Conference on The Theory of Information Retrieval. New York: ACM: 2015. p. 349–52.
Yuan Y, Zhou QM, Li B, Cai H, Chow EJ, Armstrong GT. A thresholdfree summary index of prediction accuracy for censored time to event data. Stat Med. 2018; 37(10):1671–81.
Ozenne B, Subtil F, MaucortBoulch D. The precision–recall curve overcame the optimism of the receiver operating characteristic curve in rare diseases. J Clin Epidemiol. 2015; 68(8):855–9.
Saito T, Rehmsmeier M. The precisionrecall plot is more informative than the roc plot when evaluating binary classifiers on imbalanced datasets. PloS One. 2015; 10(3):0118432.
Davis J, Goadrich M. The relationship between precisionrecall and roc curves. In: Proceedings of the 23rd International Conference on Machine Learning. ICML ’06. New York: ACM: 2006. p. 233–40.
Clark RA, MostoufiMoab S, Yasui Y, Vu NK, Sklar CA, Motan T, Brooke RJ, Gibson TM, Oeffinger KC, Howell RM, Smith SA, Lu Z, Robison LL, Chemaitilly W, Hudson MM, Armstrong GT, Nathan PC, Yuan Y. Predicting acute ovarian failure in female survivors of childhood cancer: a cohort study in the childhood cancer survivor study (ccss) and the st jude lifetime cohort (sjlife). Lancet Oncol. 2020; 21(3):436–45.
Cox DR, et al. Regression models and life tables. JR Stat Soc B. 1972; 34(2):187–220.
Uno H, Cai T, Tian L, Wei L. Evaluating prediction rules for tyear survivors with censored regression models. J Am Stat Assoc. 2007; 102:527–37.
Hudson MM, Ness KK, Nolan VG, Armstrong GT, Green DM, Morris EB, Spunt SL, Metzger ML, Krull KR, Klosky JL, et al. Prospective medical assessment of adults surviving childhood cancer: study design, cohort characteristics, and feasibility of the st. jude lifetime cohort study. Pediatr Blood Cancer. 2011; 56(5):825–36.
Steyerberg EW, Vickers AJ, Cook NR, Gerds T, Gonen M, Obuchowski N, Pencina MJ, Kattan MW. Assessing the performance of prediction models: a framework for some traditional and novel measures. Epidemiol (Cambridge, Mass). 2010; 21(1):128.
Kattan MW, Gerds TA. The index of prediction accuracy: an intuitive measure useful for evaluating risk prediction models. Diagn Prognostic Res. 2018; 2(1):7.
Pepe MS, Janes H, Longton G, Leisenring W, Newcomb P. Limitations of the odds ratio in gauging the performance of a diagnostic, prognostic, or screening marker. Am J Epidemiol. 2004; 159(9):882–90.
Vickers AJ, Elkin EB. Decision curve analysis: a novel method for evaluating prediction models. Med Dec Making. 2006; 26(6):565–74.
Shah NH, Milstein A, Bagley SC. Making machine learning models clinically useful. JAMA. 2019; 322(14):1351–2.
Baker SG, Kramer BS. Peirce, youden, and receiver operating characteristic curves. Am Stat. 2007; 61(4):343–6.
Baker SG, Schuit E, Steyerberg EW, Pencina MJ, Vickers A, Moons KG, Mol BW, Lindeman KS. How to interpret a small increase in auc with an additional risk prediction marker: decision analysis comes through. Stat Med. 2014; 33(22):3946–59.
Acknowledgements
We thank the Associate Editor and two reviewers for the constructive comments and suggestions. The data example is from the St. Jude Lifetime cohort study, supported by National Cancer Institute grant U01CA195547 (PIs Hudson MM and Robinson LL). We thank the St. Jude Lifetime cohort study and Childhood Cancer Survivors Study participants and their families for providing the time and effort for participation and the internet team at St. Jude Children’s Research Hospital for the development of the web application of the risk prediction models.
Funding
Dr. Yuan’s research is supported by the Natural Sciences and Engineering Research Council of Canada (RGPIN201904862).
Author information
Authors and Affiliations
Contributions
Authors’ contributions
QZ and YY developed the concept, designed the analytical and numerical studies and drafted the manuscript. LZ conducted the numerical study and analyzed the data example. RB and MH prepared the data from the St. Jude Lifetime cohort Study. LZ, RB, and MH revised the manuscript. All the authors read and approved the final manuscript.
Authors’ information
Department of Mathematics and Statistics, Mississippi State University, Mississippi State, MS, 39762, USA
Qian M. Zhou
School of Public Health, University of Alberta, Edmonton, AB, T6G 2R3, Canada
Zhe Lu, Yan Yuan
St. Jude Children’s Research, Memphis, TN, 38105, USA
Russell J. Brook, Melissa M. Hudson
Corresponding author
Ethics declarations
Ethics approval and consent to participate
Not applicable.
Consent for publication
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Additional file 1
Supplementary material. The supplementary material includes (i) the histograms of the predicted AOF risk obtained from the prescribeddose model and ovariandose model for individuals with and without AOF, respectively, (ii) the procedure of obtaining the true values of the IncV metrics under the distributional assumptions of the numerical study, (iii) the results of the numerical study scenarios in which neither of the working risk models is the true model, including plots of the values of each IncV metric for all the scenarios under different event rates, and the scatter plots of each pair of the IncV metrics, and (iv) the results for the scenarios where the twomarker model is the true model, including plots of the values of each IncV metric for all the scenarios under different event rates, plots of their summary statistics, and a table listing the Pearson correlation of each pair of the IncV metrics. (PDF file)
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Zhou, Q.M., Zhe, L., Brooke, R.J. et al. A relationship between the incremental values of area under the ROC curve and of area under the precisionrecall curve. Diagn Progn Res 5, 13 (2021). https://doi.org/10.1186/s4151202100102w
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s4151202100102w
Keywords
 Prediction performance
 AUC
 Area under precisionrecall curve
 Brier score
 Proper scoring rules
 Rare outcome