Abstract
The Alda score is commonly used to quantify lithium responsiveness in bipolar disorder. Most often, this score is dichotomized into “responder” and “non-responder” categories, respectively. This practice is often criticized as inappropriate, since continuous variables are thought to invariably be “more informative” than their dichotomizations. We therefore investigated the degree of informativeness across raw and dichotomized versions of the Alda score, using data from a published study of the scale’s inter-rater reliability (n=59 raters of 12 standardized vignettes each). After learning a generative model for the relationship between observed and ground truth scores (the latter defined by a consensus rating of the 12 vignettes), we show that the dichotomized scale is more robust to inter-rater disagreement than the raw 0-10 scale. Further theoretical analysis shows that when a measure’s reliability is stronger at one extreme of the continuum—a scenario which has received little-to-no statistical attention, but which likely occurs for the Alda score ≥ 7—dichotomization of a continuous variable may be more informative concerning its ground truth value, particularly in the presence of noise. Our study suggests that research employing the Alda score of lithium responsiveness should continue using the dichotomous definition, particularly when data are sampled across multiple raters.
Introduction
The Alda score is a validated index of lithium responsiveness commonly used in bipolar disorder (BD) research [1]. This scale has two components. The first is the “A” subscale that provides an ordinal score (from 0 to 10, inclusive) of the overall “response” in a therapeutic trial of lithium. The second component is the “B” subscale that attempts to qualify the degree to which any improvement was causally related to lithium. The total Alda score is computed based on these two subscale scores, and takes integer values between 0 and 10. Many studies that employ the Alda score as a target variable dichotomize it, such that individuals with scores ≥ 7 are classified as “responders,” and those with scores < 7 are “non-responders.”
A common criticism that arises from this practice is that continuous variables should not be discretized by virtue of “information loss.” Indeed, discretizing continuous variables is widely viewed as an inappropriate practice [2–12]. However, the practice remains common across many areas of research, including our group’s work on lithium responsiveness in BD [13]. The primary justification for using the dichotomized Alda score as the lithium responsiveness definition has been based on the inter-rater reliability study by Manchia et al. [1], who showed that a cut-off of 7 had strong inter-rater agreement (weighted kappa 0.66). Furthermore, using mixture modeling, they also found that the empirical distribution of Alda scores supports the discretized definition. Therefore, there exist competing arguments regarding the appropriateness of dichotomizing lithium response. Resolving this dispute is critical, since the operational definition of lithium responsiveness is a concept upon which a large body of research will depend.
Although the Manchia et al. [1] analysis provides some justification for using a dichotomous lithium response definition, it does not dispel the argument of discretization-induced information loss entirely. However, there is some intuitive reason to believe that discretization is, at least pragmatically, the best approach to defining lithium response using the Alda score. First, the Alda score remains inherently subjective to some degree and is not based on precise biological measurements; an individual whose “true” Alda score is 6, for example, could have observed scores that vary widely across raters. Second, it is possible that responders may be more reliably identified than non-responders. For example, unambiguously “excellent” lithium response is a phenomenon that undoubtedly exists in naturalistic settings [14, 15], and for which the space of possible Alda scores is substantially smaller than for non-responders; that is, an Alda score of 8 can be obtained in far fewer ways than an Alda score of 5. As such, we hypothesize that agreement on the Alda score is higher at the upper end of the score range, and that this asymmetric agreement is a scenario in which dichotomization of the score is more informative than the raw measure. To evaluate this, we present both empirical re-analysis of the ConLiGen study by Manchia et al. [1]. and analyses of simulated data with varying levels of asymmetrical inter-rater reliability.
Materials and methods
Data
Detailed description of data and collection procedures is found in Manchia et al. [1]. Samples included in our analysis are detailed in Table 1, including the number of raters included across sites, and the average ratings obtained at each of those sites across the 12 assessment vignettes. As a gold standard, we used ratings that were assigned to each case vignette using a consensus process at the Halifax site (scores are noted in the first row of Table 1). The lithium responsiveness inter-rater reliability data are available in S1 File (total Alda score), and S2 File (Alda A-score).
Empirical Analysis of the Alda Score
In this analysis, we seek to evaluate whether discretization of the Alda score under the existing inter-rater reliability values preserves more mutual information (MI) between the observed and ground truth labels than does the raw scale representation. To accomplish this, we first develop a probabilistic formulation of raters’ score assignments based on a Multinomial-Dirichlet model, which we describe here.
Let denote the number of raters who assigned an Alda score , with to an individual whose gold standard Alda score is . The vector of rating counts for the gold standard score k is is . The probability of n(k) is multinomial with parameter vector , which is itself Dirichlet distributed θ(k) ∼ Dir(θ|α), where α is a pseudocount denoting the prior expectation of the number of ratings received for each score . In the present analysis, we assume that α is equal across all scores in , and thus we denote it simply as a scalar α = α; this has the effect of increasing the uncertainty of θ(k) (i.e. the ratings become more “noisy”).
The posterior of θ(k) given n(k) and α is Dirichlet with parameters and its maximum a posteriori (MAP) estimate is which can be viewed as the conditional distribution over scores for any given rater when the gold standard is k. In cases where no assessment vignette had a gold standard rating of k, we assumed that
The dichotomized Alda scores are defined as , where τ is the dichotomization threshold (set at τ = 7 for the Alda score), and where δ[·] is an indicator function that evaluates to 1 if the argument is true, and 0 otherwise. Given threshold τ (Responders ≥ τ and Non-responders < τ), the dichotomous counts are represented as follows with c(k) ~ Multinomial(ϕk), and ϕk ~ Dir(ϕ|ξ), where ξ is a pseudocount for the number of dichotomized ratings assigned to each of non-responders and responders. We can thus estimate the conditional distribution over observed dichotomized response ratings as
Mutual Information of Raw and Dichotomized Alda Score Representations
Let denote a given observed raw Alda score assigned to a case with ground truth score of . Given uniform priors on the true classes, the joint distribution is
For the binarized classes, we have a prior of , and the joint distribution is thus
The MI for these distributions can be computed as functions of the prior pseudocounts α and ξ: for the raw and dichotomized Alda scores, respectively. We can express the MI of the raw and dichotomized Alda score distributions both in terms of α, such that both distributions have an equivalent total “concentration:” ξ = α when ξ = 11α/2. This is equivalent to saying that our prior assumption about the uncertainty of the raw and dichotomized distributions assumes the same number of a priori ratings.
Our primary hypothesis—that the dichotomized Alda score is more informative with greater observation uncertainty—is evaluated by determining whether Iξ[yo||y∗] exceeds Iα[xo||x∗] as we increase the a priori observation noise (α and ξ).
Theoretical Modeling of Dichotomization under Asymmetrical Reliability
The previous experiment regarding dichotomization of the raw Alda score did not fully capture the effect of dichotomization of a continuous variable, since the raw Alda score is still discrete (albeit with a larger domain of support). Thus, we sought to investigate whether dichotomization of a truly continuous, though asymmetrically reliable, variable would show a similar pattern of preserving MI and statistical power under higher levels of observation noise and agreement asymmetry.
Synthetic Datasets
The simplest synthetic dataset generated was merely a sample of regularly spaced points across the [0,10] interval in both the x and y directions. This dataset was merely used to conduct a “sanity check” that our methods for computing MI correctly identified a value of 0. This was necessary since data with uniform random noise over the same interval will only yield MI of 0 in the limit of large sample sizes.
The main synthetic dataset accepted “ground truth” values x ∈ [0, 10] and yielded “observed” values y ∈ [0, 10] based on the following formula for the ith sample: where 0 ≤ ω ≤ 1 is a parameter governing the degree to which observed values are coupled to the ground truth based on f (xi) (data are entirely uniform random noise when ω = 0, and come entirely from f (xi) when ω = 1). The function f (xi) governing the agreement between ground truth and observed is essentially a 1:1 correspondence between x and y to which we add noise along the diagonal based on a uniform random variate with width σ.
We simulated two forms of diagonal spread. The first is constant across all values x ∈ [0, 10], which we call the symmetrical case, and which is represented by a parameter β = 1. The other is an asymmetrical case (represented as β = 0), in which the agreement between x and y is not constant across the [0, 10] range. Overall, the function f (xi) is defined as where R(l, u)(⋅) is a function to ensure that all points remain within the [l, u] interval in both axes. In the asymmetrical case, R(l, u)(·) reflects points at the [0, 10] bounds. In the symmetrical case, the data are all simply rescaled to lie in the [0, 10] interval.
Demonstration of the simulated synthetic data are shown in Figure 1. Every synthetically generated dataset included 750 samples, and for notational simplicity, we denote the kth synthetic dataset (given parameters β, ω, σ) as .
Computation of Mutual Information for Continuous and Discrete Distributions
Mutual information was computed for both continuous and dichotomized probability distributions on the data. Mutual information for the continuous distribution was computed by first performing Gaussian kernel density estimation (using Scott’s method for bandwidth selection) on the simulated dataset, and then approximating the following integral using Markov chain Monte-Carlo sampling:
Conversely, discrete MI was computed by first creating a 2-dimensional histogram by binning data based on a dichotomization threshold τ. Data that lie below the dichotomization threshold are denoted 1, and those that lie above the threshold are represented as 0. Based on this joint distribution, the dichotomized MI is
Note that continuous MI will remain constant across τ.
Statistical Power of Classical Tests of Association
Association between the observed (y) and ground truth (x) data can be measured using Pearson’s correlation coefficient (ρ) when data are left as continuous, or using Fisher’s exact test when data are dichotomized. The statistical power of the hypothesis that ρ ≠ 0 given dataset with N(k) observations and two-tailed statistical significance threshold α—which here is not the same α used as a Dirichlet concentration in Empirical Evaluation of the Alda Score of Lithium Response—can be easily shown to equal where Φ(·) and Φ−1 (·) are the cumulative distribution function and quantile functions for a standard normal distribution, and ζ(·) is Fisher’s Z-transformation
Under a dichotomization of with threshold τ association between the ground truth and observed data can be evaluated using a (two-tailed) Fisher’s exact test, whose alternative hypothesis is that the odds ratio (η) of the dichotomized data does not equal 1. The null-hypothesis has a Fisher’s noncentral hypergeometric distribution, where N(k) is the total number of observations in sample k, and and are the number of ground truth and observed data, respectively, that fall below the dichotomization threshold τ. Under the alternative hypothesis, this distribution has an odds ratio parameter estimated from the data:
The statistical power of Fisher’s exact test under this setup and a two-tailed significance threshold of α is where and are the survival functions of the alternative hypothesis and the inverse survival function of the null hypothesis, respectively.
Evaluation of Mutual Information
The central aspect of this analysis is comparison of the dichotomized and continuous MI across values of the dichotomization threshold τ, global noise ω, asymmetry parameter β, and diagonal spread σ. Under all cases, we expect that increases in the global noise parameter ω will reduce the MI. We also expect that with symmetrical reliability (i.e. β = 0), the dichotomized MI will be lower than the continuous MI across all thresholds. However, as the degree of asymmetry in the reliability increases, we expect the dichotomized MI to exceed the continuous MI (i.e. as σ increases when β = 1). Finally, as a sanity check, we expect that both continuous and dichotomized MI will be approximately 0 when applied to a grid of points regularly spaced over the [0,10] interval.
Evaluation of Effects on Statistical Power of Classical Tests of Association
Statistical power of the Pearson correlation coefficient and Fisher’s exact test were computed across symmetrical (β = 0) and asymmetrical (β = 1) conditions of the synthetic dataset described above. Owing to the greater computational efficiency of these calculations (compared to the MI), the diagonal spread parameter was varied more densely (σ ∈ {1, 2, …, 20}). The power of Fisher’s exact test was evaluated at two dichotomization thresholds: a median split at τ = 5 and a “tail split” at τ = 3. We evaluated three global noise settings (ω ∈ {0.3, 0.5, 0.7}). At each experimental setting, we computed the aforementioned power levels for 100 independent synthetic datasets. Results are presented using the mean and 95% confidence intervals of the power estimates over the 100 runs under each condition. We expect that the Fisher’s exact test under a “tail split” dichotomization (not a median split) will yield greater statistical power in the presence of asymmetrical reliability, greater diagonal spread, and higher global noise. However, under the symmetrically reliable condition, we expect that the statistical power will be greater for the continuous test of association.
Materials
Mutual information experiments were conducted in Mathematica v. 12.0.0 (Wolfram Research, Inc.; Champaign, IL). Experiments evaluating the statistical power under classical tests of continuous and dichotomous association were conducted in the Python programming language. Code for analyses are also provided in S3 File (Alda score analyses), S4 File (theoretical analysis of MI under asymmetrical reliability), S5 File (theoretical evaluation of classical associative tests). The Mathematica notebooks are also available in PDF form in S6 File and S7 File.
Results
Empirical Evaluation of the Alda Score of Lithium Response
Histograms of the observed Alda scores for each of the gold standard vignette values are depicted in Fig 2. Resulting joint distributions of the gold standard vs. observed Alda scores (in both the raw or dichotomized representations) are shown in Fig 3 (Panels A-C) across varying levels of observation noise. Fig 3D plots the MI for the raw and dichotomized Alda scores across increasing levels of the observation noise parameter α (recalling that ξ = 11α/2). Beyond an observation noise of approximately α > 3.52, one can see that the dichotomized lithium response definition retains greater MI between the true and observed labels, compared to the raw representation.
Discrete vs. Continuous Mutual Information in Asymmetrically Reliable Data
Fig 4 shows the results of the experiment on synthetic data. Under agreement levels that are constant across the (x, y) domains, one can observe that MI of dichotomized representations of the variables are generally lower than their continuous counterparts. However, under asymmetrical reliability (i.e. where agreement between x and y decreases as x increases), we see that MI is higher for the dichotomized, rather than the continuous, representations. In particular, as the level of agreement asymmetry increased (i.e. for higher values of σ), the best dichotomization thresholds decreased.
Statistical Power of Classical Associative Tests
Fig 5 plots the statistical power of null-hypothesis tests using continuous and dichotomized representations of the synthetic dataset. As expected, under conditions of symmetrical reliability, the continuous test of association (Pearson correlation) retains greater statistical power as the degree of diagonal spread increases, although this difference lessens at very high levels of diagonal spread or overall (uniform) noise. However, under conditions of asymmetrical reliability, dichotomizing data according to a “tail split” (here a threshold of τ = 3) preserves greater statistical power than either a median split (τ = 5) or continuous representation; this relationship was present even at high levels of diagonal spread and overall uniform noise.
Discussion
The present study makes two important contributions. First, using a sample of 59 ratings obtained using standardized vignettes compared to a consensus-defined gold standard [1], we showed that the dichotomized Alda score has a higher MI between the observed and gold standard ratings than does the raw scale (which ranges from 0-10). Those data suggested that the Alda score’s reliability is asymmetrical, with greater inter-rater agreement at the upper extreme. Secondly, therefore, using synthetic experiments we showed that asymmetrical inter-rater reliability in a score’s range is the likely cause of this relationship. Our results do not argue that lithium response is itself a categorical natural phenomenon. Rather, using the dichotomous definition as a target variable in supervised learning problems likely confers greater robustness to noise in the observed ratings.
Some have argued that the existence of categorical structure in one’s data [9], or evidence of improved reliability under a dichotomized structure [16], are potentially justifiable rationales for dichotomization of continuous variables. These claims are generally stated only briefly, and with less quantitative support than the more numerous mathematical treatments of the problems with dichotomization [9, 10, 16, 17]. However, these more rigorous quantitative analyses typically involve assumptions of symmetrical or Gaussian distributions of the underlying variables in the context of generalized linear modeling (although Irwin & McClelland [10] demonstrated that median splits of asymmetric and bimodal beta distributions is also deleterious). These analyses have led to vigorous generalized denunciation of variable dichotomization across several disciplines, but our current work offers important counterexamples to this narrative [10, 11].
The Alda score is more broadly used as a target variable in both predictive and associative analyses, and not as a predictor variable, which is an important departure from most analyses against dichotomization. Since there is no valid and reliable biomarker of lithium response, these cases must rely on the Alda score-based definition of lithium response as a “ground truth” target variable. In the case of predicting lithium response, where these ground truth labels are collected from multiple raters across different international sites, variation in lithium response scoring patterns across centres might further accentuate the extant between-site heterogeneity.
To this end, inter-individual differences in subjective rating scales may be more informative about the raters than the subjects, and one may wish to use dichotomization to discard this nuisance variance [8, 9, 16]. Doing so means that one turns regression supervised by a dubious target into classification with a more reliable (although coarser) target. Appropriately balancing these considerations may require more thought than adopting a blanket prohibition on dichotomization or some other form of preprocessing.
An important criticism of continuous variable dichotomization is that it may impede comparability of results across studies, both in terms of diminishing power and inflating heterogeneity [17]. However, this is more likely a problem when dichotomization thresholds are established on a study-by-study basis, without considering generalizability from the outset. These arguments do not necessarily apply to the Alda score, since the threshold of 7 has been established across a large consortium with support from both reliability and discrete mixture analysis [1], and is the effective standard split point for this scale [18].
Our study thus provides a unique point of support for the dichotomized Alda score insofar as we show that the retention of MI and frequentist statistical power is likely due to asymmetrical reliability across the range of scores. Our analyses show that there is a range of Alda scores (those identifying good lithium responders; scores ≥7) for which scores correspond more tightly to a consensus-defined gold standard in a large scale international consortium. In particular, we showed that this dichotomization will be more robust to increases in the prior uncertainty (i.e. the overall level of background “noise” in the relationship between true/observed scores). This feature is important since the sample of raters included in the Alda score’s calibration study [1] was relatively small and consisted of individuals involved in ConLiGen centres. It is reasonable to suspect that assessment of Alda score reliability in broader research and clinical settings would add further disagreement-based noise to the inter-rater reliability data. At present, use of the dichotomized scale could confer some robustness to that uncertainty.
More generally our study showed that if reliability of a measure is particularly high at one tail of its range, then a “tail split” dichotomization can outperform even the continuous representation of the variable. This presents an important counterexample to previous authors, such as Cohen [5], Irwin & McClelland [10], and MacCallum et al. [9] who argued that “tail splits” are still worse than median splits. While our study reaffirms these claims in the case of measures whose reliability is constant over the domain (see Figure 4B and the upper row of Figure 5), our analysis of the asymmetrically reliable scenario yields opposite conclusions, favouring a “tail split” dichotomization over both median splits and continuous representations. Tail split dichotomization was particularly robust when data were affected by both asymmetrical reliability and high degrees of uniform noise over the variable’s range. Together, these results suggest that dichotomization/categorization of a continuous measurement may be justifiable when its relationship to the underlying ground truth variable is noisy everywhere except at an extreme.
Our study has several limitations. First, our sample size for the re-analysis of the Alda score reliability was relatively small, and sourced from highly specialized raters involved in lithium-specific research. However, one may consider this sample as representative of the “best case scenario” for the Alda score’s reliability. It is likely that further expansion of the subject population would introduce more noise into the relationship between ground truth and observed Alda scores. It is likely that most of this additional disagreement would be observed for lower Alda scores, since (A) there are simply more potential item combinations that can yield an Alda score of 5 than an Alda score of 9, for example, and (B) unambiguously excellent lithium response is a phenomenon so distinct that some question whether lithium responsive BD may constitute a unique diagnostic entity [19, 20]. Thus, we believe that our sample size for the reliability analysis is likely sufficient to yield the present study’s conclusions.
Our study is also limited by the fact that theoretical analysis was largely simulation-based, and thus cannot offer the degree of generalizability obtained through rigorous mathematical proof. Nonetheless, our study offers sufficient evidence—in the form of a counterexample—to show that there exist scenarios in which dichotomization is statistically superior to preserving a variable’s continuous representation. Furthermore, we used well controlled experiments to isolate asymmetrical reliability as the cause of dichotomization’s superiority across simulated conditions.
Conclusion
In conclusion, we have shown that a dichotomous representation of the Alda score for lithium responsiveness is more robust to noise arising from inter-rater disagreement. The dichotomous Alda score is therefore likely a better representation of lithium responsiveness for multi-site studies in which lithium response is a target or dependent variable. Through both re-analysis of the Alda score’s real-world inter-rater reliability data and careful theoretical simulations, we were able to show that asymmetrical reliability across the score’s domain was the likely cause for superiority of the dichotomous definition. Our study is not only important for future research on lithium response, but other studies using subjective and potentially unreliable measures as dependent variables. Practically speaking, our results suggest that it might be better to classify something we can all agree upon than to regress something upon which we can not.
Supporting information
S1 Fig. A-Score reliability histograms. Histograms of ratings for each value of the ground truth Alda A-score. This figure was generated identically to Fig 2, but using the A-score data only.
S2 Fig. A-Score mutual information results. Mutual information between gold standard and observed Alda A-scores in relation to observation noise and the scale’s “raw” or dichotomized form. This figure was generated identically to Fig 3, but using the A-score data only.
S1 File. Total Alda score ratings. Inter-rater reliability data for the total Alda score.
S2 File. Alda A-score ratings. Inter-rater reliability data for the Alda A-score.
S3 File. Alda score analysis code. Mathematica notebook containing the empirical evaluation of the Alda Score of Lithium response. This notebook also contains additional analysis of the A-score alone.
S4 File. Theoretical analysis code. Mathematica notebook containing the theoretical analyses of discrete vs. continuous mutual information in asymmetrically reliable data.
S5 File. Code for statistical power tests. Jupyter notebook containing the theoretical analyses of the statistical power of classical associative tests under asymmetrically reliable data.
S6 File. Alda score analysis code (PDF Version). PDF version of S3 File for those without Mathematica license.
S7 File. Theoretical analysis code (PDF Version). PDF version of S4 File for those without Mathematica license.
Footnotes
↵¤ 5909 Veteran’s Memorial Lane Abbie J. Lane Memorial Building, QEII Health Sciences Centre, Halifax, Nova Scotia, Canada