Skip to main content

Using joint probability density to create most informative unidimensional indices: a new method using pain and psychiatric severity as examples

Abstract

Background

Dimension reduction methods do not always reduce their underlying indicators to a single composite score. Furthermore, such methods are usually based on optimality criteria that require discarding some information. We suggest, under some conditions, to use the joint probability density function (joint pdf or JPD) of p-dimensional random variable (the p indicators), as an index or a composite score. It is proved that this index is more informative than any alternative composite score. In two examples, we compare the JPD index with some alternatives constructed from traditional methods.

Methods

We develop a probabilistic unsupervised dimension reduction method based on the probability density of multivariate data. We show that the conditional distribution of the variables given JPD is uniform, implying that the JPD is the most informative scalar summary under the most common notions of information. B. We show under some widely plausible conditions, JPD can be used as an index. To use JPD as an index, in addition to having a plausible interpretation, all the random variables should have approximately the same direction(unidirectionality) as the density values (codirectionality). We applied these ideas to two data sets: first, on the 7 Brief Pain Inventory Interference scale (BPI-I) items obtained from 8,889 US Veterans with chronic pain and, second, on a novel measure based on administrative data for 912 US Veterans. To estimate the JPD in both examples, among the available JPD estimation methods, we used its conditional specifications, identified a well-fitted parametric model for each factored conditional (regression) specification, and, by maximizing the corresponding likelihoods, estimated their parameters. Due to the non-uniqueness of conditional specification, the average of all estimated conditional specifications was used as the final estimate. Since a prevalent common use of indices is ranking, we used measures of monotone dependence [e.g., Spearman’s rank correlation (rho)] to assess the strength of unidirectionality and co-directionality. Finally, we cross-validate the JPD score against variance–covariance-based scores (factor scores in unidimensional models), and the “person’s parameter” estimates of (Generalized) Partial Credit and Graded Response IRT models. We used Pearson Divergence as a measure of information and Shannon’s entropy to compare uncertainties (informativeness) in these alternative scores.

Results

An unsupervised dimension reduction was developed based on the joint probability density (JPD) of the multi-dimensional data. The JPD, under regularity conditions, may be used as an index. For the well-established Brief Pain Interference Inventory (BPI-I (the short form with 7 Items) and for a new mental health severity index (MoPSI) with 6 indicators, we estimated the JPD scoring. We compared, assuming unidimensionality, factor scores, Person’s scores of the Partial Credit model, the Generalized Partial Credit model, and the Graded Response model with JPD scoring. As expected, all scores’ rankings in both examples were monotonically dependent with various strengths. Shannon entropy was the smallest for JPD scores. Pearson Divergence of the estimated densities of different indices against uniform distribution was maximum for JPD scoring.

Conclusions

An unsupervised probabilistic dimension reduction is possible. When appropriate, the joint probability density function can be used as the most informative index. Model specification and estimation and steps to implement the scoring were demonstrated. As expected, when the required assumption in factor analysis and IRT models are satisfied, JPD scoring agrees with these established scores. However, when these assumptions are violated, JPD scores preserve all the information in the indicators with minimal assumption.

Peer Review reports

Introduction

In this paper we introduce an unsupervised probabilistic method for dimension reduction (DR) particularly when it results in a single composite score or an index.

Sufficient dimension reduction (SDR), as a supervised probabilistic approach, is based on the conditional densities [1,2,3,4] to construct the q summaries. It requires:

\(f\left(y|{x}_{1},{x}_{2},\dots ,{x}_{p}\right)=f\left(y|{s}_{1},{s}_{2},\dots ,{s}_{q}\right)\), where \({{s}_{k}={h}_{k}(x}_{1},{x}_{2},\dots ,{x}_{p}) , k=\text{1,2},..q.\)

To extend this idea to unsupervised situations, we use the joint density of predictors and replace the sufficiency of summaries, i.e., being the most informative for the response, with the notion of most informativeness. Thus, formally we define:

Definition

The function \({\varvec{s}}:{{\varvec{R}}}^{{\varvec{p}}}\to {{\varvec{R}}}^{{\varvec{q}}},\boldsymbol{ }{{s}_{k}={h}_{k}(x}_{1},{x}_{2},\dots ,{x}_{p}) , k=\text{1,2},..q.\) is a most informative summary if:

\(f\left({x}_{1},{x}_{2},\dots ,{x}_{p}|{s}_{1},{s}_{2},\dots ,{s}_{q}\right)\) is a uniform distribution, for each

$${\varvec{s}}=\left(s_1,s_2,\dots,s_q\right).$$
(1)

In other words, given the summary score, the conditional distribution of the indicators should have maximum entropy and hence minimum information. Note that this means \({\varvec{s}}\) contains all the information in the indicators. Note that \(f\left({x}_{1},{x}_{2},\dots ,{x}_{p}\right)\), the joint probability density (JPD) function, satisfies (1) (see the appendix for a proof).

The following set of necessary conditions are required for JPD summary to be an index:

First, \({x}_{1}, {x}_{2},\dots ,{x}_{p}\), must be at least ordinal be conceptually related, and undergird a broader concept, such as pain experience, mental health status, volatility of the stock market, and as such. We refer to this property as relevance. In factor analysis language, the latent space of \({x}_{1}, {x}_{2},\dots ,{x}_{p}\) indicators should be unidimensional. Second, all indicators must be recorded or coded such that their order (higher or lower) values agree, and their pairwise monotone correlations be strong and in the same direction (sign). We call this property the unidirectionality of the indicators. Third, the index, \(s,\) should be monotonically correlated with each of the indicators (variables) and with the same sign. We refer to this property as the codirectionality of the summary with the variables. Strength of relevance and unidirectionality can be achieved at the study design stage where one decides what indicators should be collected or how they should be measured. The strength of codirectionality depends on actual frequencies and how the density values change by changes in indicators’ value patterns.

Many unsupervised techniques optimize some aspect of variables’ joint probability density or joint probability mass function. For example, Principal Components Analysis [5], its kernelized versions [6, 7] and Factor Analysis are based on the variances and the covariances of the joint distribution, respectively. Item Response Theory (IRT) uses the joint pdf of indicators; however IRT benefits from four assumptions: local independence, correctness of functional form, usually a linear decomposition of the model parameters, the assumption that a unique parameter, called ability that measures subjects differences, and that this latent variable has a normal distribution in the population [though see multi-dimensional IRT [8] and nonparametric IRT [9] for some relaxations]. The local independence assumption permits one to write the joint pdf as a product of the univariates indicators’ (items’) pdfs given the parameter values.

Using the joint probability density, or a monotone transformation of it, as an index offers many advantages. Variables(indicators) do not need to meet restrictive assumptions beyond having at least ordinal scale, unidirectionality, and co-directionality with the JPD. Variables may be highly skewed, dependent, and of different types (nominal, categories, ordinal, interval, count, or continuous). With a JPD index, items' dependencies are respected–not just the covariances, and the information in all moments of different orders is preserved. Note that indices need not necessarily designed to refer to a ‘latent dimension’. The Charlson comorbidity indices, Economics and most other indices are such summaries. In observational studies, propensity score [10] is another instance. Propensity score is indeed a dimension-reducing index that reduces several possible confounders into a single score.

In the present paper, we compare the JPD scores with (Confirmatory) Factor Analysis and IRT scorings in two unique datasets. We first present the alternative JPD Index for the 7-item Brief Pain Interface Inventory (BPI-I) [cite] and then construct the JPD index for ordering mental health patients based on 6 administratively collected indicators. For both datasets, we then compare i) the JPD index, ii) factor analytic score when used as an index, and Person’s parameter from the Partial Credit model (PC) [11], the Generalized Partial Credit model (GPC) [12], and the Graded Response model (GRM) [13].

Background

Dimensionality reduction (DR) of p-variate data to a smaller set of q statistics overcomes computational challenges in data analysis [14,15,16] and [17]. Several supervised and unsupervised methods, both probabilistic, geometric, or mathematical have been developed. In DR literature, dimension refers to the number of summary statistics constructable from the multivariate data that: 1) are linearly independent and have no overt and hopefully hidden non-linear dependency, and 2) satisfy some optimality criteria in preserving a notion of information. In some DR methods, such as factor analysis, composite score, and index construction, it is also desirable that the derived summary or summaries, 3) can be interpreted as quantifying some unmeasurable constructs. Implication of this 2- or 3-step procedure has been the source of ongoing DR research. The lack of unique optimality criteria has been a source of subjective ad hoc decisions in DR procedures. Also, the impossibility of ensuring the nonexistence of nonlinear dependence has resulted in more restricted research formulation, e.g., mostly restricting to linear DR. The current study is about constructing an index, a single composite score, possibly non-linear, that can replace the indicators.

A quick review of DR methods has been reported by [18]. Textbooks on dimensionality reduction (DR) provide a more detailed account of the theoretical basis of these methods [19,20,21].

When DR results in a single statistic with codirectionality property, the statistic can be used as an index. There is a vast literature on classical index construction independent of DR research [See for example: [22,23,24]. Also, The United Nations Statistics Division recommendations for index construction [25].

Another conceptually similar problem is the construction of composite scores, in the scoring of instruments and tests, where one is to reduce several item responses to a single statistic for assessment purposes. The scoring methods in Classical Test Theory [26], various Factor Analysis formulations, provided some notion of unidimensionality of factor structure holds, and the Person’s parameter of an appropriate Rasch and more general IRT models [27, 28], all also can be used to construct indices, although these methods have broader agenda then index construction. The DR property of FA and IRT, especially when some notion of unidimenionality holds, is the most relevant feature of the methods to index construction. For completeness, we have briefly introduced these two methodologies in Additional file 1. In the following we introduce the data and the detailed implementation of the methods that will be used in the two examples.

Materials and methods

Design

The cross- sectional baseline data collected are from two samples: the Effects of Prescription Opioid Changes (EPOCH) [29] and Long Term Outcomes in Veterans Requesting A Compensation (TRACTION) [30] studies. Analyses for EPOCH were post-hoc and pre-planned for TRACTION. The Minneapolis VA Health Care System’s Internal Review Board for Human Studies reviewed and approved the studies’ protocol (#4495-B and #4586-A).

Participants

EPOCH survey cohort participants were 9,253 randomly selected Veterans with chronic pain who were receiving long-term opioid treatment as part of their management plan who completed a baseline survey. Of baseline survey responders, 364 did not complete all the pain interference inventory items and were excluded from these analyses [29]. Participants from TRACTION were 960 representatively sampled, gender-stratified Veterans who had served during Operations Enduring Freedom, Iraqi Freedom, and New Dawn and had pending VA disability claims for posttraumatic stress disorder Forty-eight TRACTION members did not use any VA health care during the study and are excluded from these analyses [30].

Data sources

Data sources were self-report survey (EPOCH) and administrative data from the VA Corporate Data Warehouse data (TRACTION). Data were collected between November 2015 and April 2017.

Measures

Pain interference

The short-form Brief Pain Inventory (BPI) [31] is widely used to assess clinical pain [32, 33]. Factor analysis points to two latent pain dimensions: “severity,” with four items, and “interference,” with seven items. BPI interference items have 11 response options scored from 0 to 10, with 0 = “Does not interfere” and 10 = “Completely interferes.” Analyses here are confined to the 7 interference items included in the BPI interference (BPI-I) short scale that was collected from the baseline EPOCH questionnaire. Items one to seven are concerned with 1: mood, 2: work, 3: general activity, 4: walking, 5: relationships, 6: enjoyment of life, and 7: sleep.

Psychiatric severity

The Manifestations of Psychiatric Severity Index (MoPSI) is a measure of psychiatric severity we developed to assess non-response bias in a randomized trial of survey methods [34]. MoPSI is comprised of 6 variables: the number of emergency department visits, psychiatric hospitalizations, and mental health visits TRACTION participants made in a 6-month period as well any ICD-9 and ICD-10-CM codes pertaining to self-harm behaviors, including suicidality; any diagnosis of substance use; and any diagnosis of alcohol use in the 180 days prior to survey. The rationale for selecting these 6 variables is reported elsewhere [34].

Note that MoPSI items do not lend themselves well to traditional psychometric covariance decomposition techniques, such as factor analysis, because there is a mix of data types (e.g., counts and dichotomous) with extreme skewness. Also, local independence of indicators, needed in IRT modeling does not hold across some variables. For example, even for subjects with a fixed person’s parameter value, one expects self-harm behaviors to trigger a more intense health care response (e.g., more clinic visits, emergency department visits, or hospitalizations) than for someone without self-harm behaviors and the same fixed person parameter. See the details below.

Analysis

We used R 4.2.1 and SAS version 9.4 for all analyses.

BPI-I scoring

A. Variance–covariance based scoring

Per published recommendations [31] based on classical factor analysis, scores for the BPI Interference scale were obtained by averaging participants’ responses across the 7 items. The factor structure of the BPI-I has been thoroughly explored in prior studies, e.g., [32, 33]. Alphas showed good internal consistency was 0.89 to 0.92 for the seven interference items [32]. A reported value for test–retest reliability was 0.97 for Pain Interference in a study on 109 patients [35]. However, we reexamined the psychometric properties of the brief BPI-I for our cohort.

In the following we are using Pearson correlation matrix for reliability, suitability and dimensionality identification, and our final confirmatory factor analysis and construct validity assessment. A sample of 8,889 patients who had used opioid for pain reduction for at least last 2 months prior to the data collection endorsed the BPI-I. Cronbach Alpha for BPI-I scale was 0.91 and after deleting each item it stayed within 0.89 to 0.92 showing acceptable reliability for the items.

The Bartlett's test of sphericity was significant (χ2(21) = 41,465.36, p < 0.001). The overall KMO value for our data was 0.897. Both criteria suggesting that the data are probably suitable for factor analysis. To assess the dimensionality, Comparison Data method, Lower bound of RMSEA 90% confidence interval and Akaike Information Criterion, all suggested 3 latent factors, while Hull method with CAF and Parallel Analysis with SMC identified 2 factors. However, the majority of other criteria: Empirical Kaiser criterion, Hull method with CFI, and with RMSEA, Kaiser-Guttman criterion with PCA, and with SMC, and the more common Parallel Analysis with PCA, all suggested unidimensionality. The fit measures of the confirmatory factor analysis, assuming a single construct, were χ2(14) = 4800.91, p < 0.001, CFI = 0.885, with RMSEA = 0.196 and SRMR = 0.056. The CFI analyzes model fit by examining the discrepancy between the data and the hypothesized model. SRMR is an absolute measure of fit and is defined as the standardized difference between the observed correlation and the predicted correlation. Also, RMSEA, the average of the residual variance and covariance. With these criteria, unidimensional model is marginally acceptable. The widely accepted BPI-I scoring is based on the total sum of the items and assumes a unidimensional construct for pain interference. Note that only 66% of total variance was explained by this factor. Factor loadings of CFA for BPI-I1 through BPI-I7 were 1.83, 2.12, 1.87, 1.93, 2.32, 1.84, and 2.22, respectively. The Pearson correlation between factor scores and the total sum was 0.996, supporting the practice of using sum score as the factor score.

B. IRT modeling of the BPI-I items

In Item Response Theory the estimated subjects’ ability, can be used to rank order the subjects. Here we will refer to the “ability” as “vulnerability” (\({\theta }_{v})\) namely, vulnerability to pain-related functionality indicators. We used Masters’ [11] Partial Credit models(PC), Generalized PC (GPC), and Graded Response Models(GRM) to estimate participants’ vulnerabilities. With (Generalized) Partial Credit models and Graded Response Model, indicators may be recorded as categorical, ordinal, or even continuous [36, 37]. For the present analysis, we assumed each BPI-I item was an ordered response with 11-categories (0–10).

Specifically, the probability that person \(v\) choses category \(h=\text{0,1},...10\) of ith BPI-I item is defined as:

$$P\left({X}_{vih}=h\right)=\frac{\text{exp}\left[h{\theta }_{v}+{\beta }_{ih}\right]}{\sum_{l=0}^{l=10}\text{exp}\left[l{\theta }_{v}+{\beta }_{il}\right]}$$
(2)

If assumptions of Item Response Theory hold, we would expect higher vulnerability to be associated with greater probability of endorsing higher or severer response options in all seven items. To investigate these, we used ICC and Items maps. The item characteristic curves (ICC) for each BPI-I item, relating the probability of endorsing any value between 0 to 10 for each question against a person’s estimated vulnerability. Values with steep ICC indicate good discriminant power. Person-item map can be used to examine the association between participants’ vulnerabilities and their probability of endorsing higher (severer) options for each BPI-I item. The R packages mirt: A Multidimensional Item Response Theory Package [38] and eRm: Extended Rasch Modeling [39] and psych package: Procedures for Psychological, Psychometric, and Personality Research [40] were used for the analyses.

C. JPD method for BPI-I items

The joint density or probability function by its definition is an expression of frequencies in the presence of dependencies. In technical sense it is radon-Nikodym derivative of the underlying probability model usually with respect to Lebesgue or counting measure. As detailed in Additional File 2, Appendix, to specify the 7-dimensional joint probability distribution for the BPI-I, we used all 7! (= 5040) permutations of the chain rule factorization:  

$$f\left(x_1,x_2,\dots,x_7\right)=f\left(x_1\right)f\left(x_2\vert x_1\right)\dots f\left(x_7\vert x_2,\dots,x_6\right)$$
(3)

where x1 through x7 corresponds to the 7 BPI-I items.

To model the conditional dependencies, we assumed each conditional model \(f\left({x}_{i}|{x}_{{l}_{1}},\dots ,{x}_{{i}_{k}}\right)\) had a two-parameter Gamma distribution whose mean, \(\mu =E\left({X}_{i}|{x}_{{l}_{1}},\dots ,{x}_{{i}_{k}}\right)=\sum {\beta }_{j}{x}_{{i}_{j}}\), was expressed as a linear function of all the conditioning variables and \(b=\frac{1}{s}\), where \(b\) is the Gamma distribution’s rate parameter, and \(s,\) the scale parameter. We used the log transform of \(f\left({x}_{1}, {x}_{2},\dots ,{x}_{7}\right)\) and then ML estimation of the parameters via generalized linear model (see Additional File 1 for code) estimates the log density= \(\widehat{b}\left(\frac{{-x}_{i}}{\widehat{\mu }}-\text{log}\left(\widehat{\mu }\right)\right)-\left(\widehat{b}-1\right)\text{log}\left({x}_{i}\right)\) for each conditional model \(f\left({x}_{i}|{x}_{{l}_{1}},\dots ,{x}_{{i}_{k}}\right).\) The average of the 5040 permutations;

$$log\left(f({x}_{1}, {x}_{2}, . . ., {x}_{7}\right))= \sum_{7!}\frac{\text{log}\left(f\left({x}_{{i}_{1}}, {x}_{{i}_{2}}, . . ., {x}_{{i}_{7}}\right)\right)}{7!}$$
(4)

Or any monotone transformation of that is the participant’s BPI-I score. Note that each of the 5040 conditional specifications provide estimates of the same JPD. The variations between the estimates:

$$\sum_{7!}\frac{{(\text{log}\left(f\left({x}_{{i}_{1}}, {x}_{{i}_{2}}, . . ., {x}_{{i}_{7}}\right)-log\left(f({x}_{1}, {x}_{2}, . . ., {x}_{7}\right))\right)}^{2}}{7!-1}$$

provides a measure of accuracy of the estimators. To assess the sampling error of each person’s JPD estimate, we used 50 bootstrapped samples (with replacement) constructed within subject standard deviation of the bootstrapped JPD estimates. The estimated standard errors for 8889 subjects in the pain study were between 0.19 to 1.89 when JPD was scaled to 0–10. For each subject our estimation method produces 5040(7!) JPD estimates, the variation between these estimates coming from alternative specifications provides a measure of accuracy of reported JPD which is an average of these scores. The between specifications standard deviations for each subject BPI_I JPD scores were between 0 and 0.252. These shows a rather robust JPD subjects’ estimates with respect to the conditional factorization of the joint pdf.

MoPSI Scoring

A. Variance–covariance based scoring

To use the factor analysis method, one needs to construct a matrix of “correlations” between the indicators (items). Three items (Mental health clinic, hospital, and emergency room visits) are count data where their larger values indicate more severe cases of mental health. Substance use (SUD), suicidality, and ETOH are ordered binary indicators, where (1) indicates a worse case compared to (0). The first challenge in EFA and subsequent confirmatory modeling is to decide how to quantify their associations (correlations). Spearman correlations (and several other measures of monotone dependence) may be used to construct a symmetric matrix. Instead of modeling dependence via ranks, one may construct a matrix of heterogeneous correlations consisting of Pearson product-moment correlations between numeric variables, polyserial correlations between numeric and ordinal variables, and Polychoric correlations between ordinal variables. Alternatively, we may ordinally categorize all indicators and use Polychoric correlation uniformly for all the p(p + 1)/2 pairs of variables. Hopefully, these symmetric matrices will be positive definite to qualify as a correlation matrix. The existence of multitude of possibilities has consequences in identifying the dimensionality of latent factor structure and especially the quality of the unidimensional index.

Specifically, here we explored the Spearman correlation (Scor), a Polychoric correlation matrix (Ocor), assuming the number of hospitalizations (0–4), and the number of emergency visits (0–3) being ordinal and, a matrix (Mcor) of mixture of correlations, where for non-categorical pairs Spearman’s rho, tetrachoric correlations for binary pairs, and Glass rank biserial for binary and number of visits are used. The Mcor matrix, though, is non-singular: it is not positive definite (the smallest eigenvalue is negative). This phenomenon often happens with these remedies when one tries factor analytic methods on data with varying scales. In the present example, this required ad hoc smoothing when performing dimension identification [41].

To develop confirmatory factor scores, we first identified if each of these correlations is suitable for factor extraction. The table below summarizes the results (Table 1).

Table 1 Suitability of data for FA

As it is seen, suitability of data for dimensionality identification depends on the subjective decisions involved in quantifying dependence. Unfortunately, there is also no unique optimal criteria for identifying the dimensionality. Table below lists the number of latent dimensions suggested by some more common criteria (Table 2).

Table 2 Number of latent dimensions

Next, for each of the three correlation matrices, we used EFA model averaging to estimate the one- and two-dimensional factors. The averaging was over estimation methods (Principal Axis, MLE and unweighted least squares), initial values and criterion type. Averaging performed via mean, with no trimming, across 108 EFAs for 2-factor model and 9 for one-factor model. The error rate was zero; all the solutions converged except for the mixed correlation matrix, where only 78% of alternative scenarios converged. All the solutions were admissible with no Heywood cases [cite] for all three matrices.

For reliability of the unidimensional CFA scores, assuming the ordinal scales, after categorizing number of mental health visits into eight categories, with Polychoric correlations, have ordinal alpha [42] of 0.923. The Omega, that is the ratio of variance of loadings times the latent factor over observed variance, was 0.749, however, the modified Omega (denominator variance being the CFA-implied variance) was 0.753. For developing an insight about construct validity of the unidimensional and two factor CFA models, fit indices are listed in the table below: [Table 3].

Table 3 Fitness measures for Psych Severity Indicators (M (SD) [Min; Max])

For one dimensional model, all indices indicate that the polychoric correlation structure fits better than the Spearman and the mixture of correlations. As always, the two-dimensional model indices indicate closer fit. For our comparisons we will use the factor scores of one-factor model in the corresponding confirmatory factor analysis.

B. Item response modeling of MoPSI items

We will use PC, GPC and GR models to estimate the corresponding participants’ vulnerability. We treated the number of mental health visits, emergency department visits, and hospitalizations as ordinal variables. However, to assess the quality of these indices (estimated Person’s paramter: vulnerabilities) a person-item map was graphed. This visually showed codirectionality between participants’ vulnerabilities (the indices) and their probability of endorsing higher (severer) options for each MoPSI items.

C. JPD method for the MoPSI

Consistent with our approach to estimating the BPI-I’s joint probability density, detailed above, we specified the MoPSI’s 6-dimensional joint probability distribution by using all 6! (= 720) permutations of the chain rule factorization:

$$f\left(x_1,x_2,\dots,x_6\right)=f\left(x_1\right)f\left(x_2\vert x_1\right)\dots f\left(x_6\vert x_2,\dots,x_5\right)$$
(5)

where x1 = number of mental health clinic visits, x2 = number of emergency department visits for mental health concerns, x3 = number of psychiatric hospitalizations, x4 = self-harm diagnoses, x5 = substance use diagnoses, and x6 = alcohol use diagnoses.

To model the conditional dependencies, we used logistic regression for the 3 binary diagnosis variables (self-harm, alcohol use, and substance use) and linear regression for the 3 count variables (number of mental health clinic visits, emergency department visits, and hospitalizations). Based on Bayesian information criterion (BIC) and Akaike information criterion (AIC), we found that the Poisson-Inverse Gaussian (PIG) distribution for the count variables had better fit indices than the Poisson and Negative Binomial and their zero-inflated versions. Each permutation of the factorization yields 720 new conditionally specified models that capture slightly different dependency structures. These terms each provide a single estimate of the same joint model, \(f\left({x}_{1}, {x}_{2},\dots ,{x}_{6}\right)\) or, after taking the log monotone transformation, of \(f\left({x}_{1}, {x}_{2},\dots ,{x}_{6}\right)\). Therefore, as in (3), to utilize all the information captured by the 6! models, we used the average of 720 estimates of \(log\left(f({x}_{1}, {x}_{2}, . . ., {x}_{6}\right)),\) asthe participant’s JPD MoPSI score. To estimate standard error of the estimated score per person, we calculated standard deviations of 40 bootstrapped estimates of JPDs for each person. The bootstrapped estimated sampling standard errors for all subjects were between (0.001, 2.433) with mean boosted SE for all subjects being 0.792. The standard deviations of each person estimated JPD between 720 specifications, for all subjects were between 0.504 and 3.382, with an average SD of 1.106.

Note that this information probably shows that we might not need to run all the possible permutations. A random sample of permutations might suffice.

The R code used to estimate the joint probability density of the MoPSI components is included in Additional File 1.

Unidirectionality and co-directionality

Assumptions of unidirectionality and co-directionality are the minimum required to conceptually label any summary as an index. Classical measures of monotone dependencies, such as rank correlations, may be used to assess for the strength of unidirectionality and should be positive. To account for skew in the BPI-I items, we used Spearman’s rho and Kendall’s tau to examine the pair-wise rank correlations of BPI-I items and to compare the correlation between each item and the 3 composite BPI-I scores obtained via standard scoring, Item Response Theory modeling, or JPD method. For the MoPSI, we examined pair-wise rank correlations using Spearman’s rho for non-categorical pairs, tetrachoric correlations for pairs of binaries, and the Glass rank biserial for binary/continuous pairs.

For co-directionality, subjects’ rankings provided by the summary or composite score should agree with each indicator’s ranking. The strength of this monotone agreement may also be assessed by rank correlations and should be positive. We used Spearman’s rank correlation and Kendall’s tau to correlate individual items to the MoPSI summary scores obtained through IRT modeling and the JPD method.

Comparing scoring approaches

We used simple scatterplots and Spearman’s rank correlation to compare the three BPI-I scoring approaches and to compare the two MoPSI scoring approaches, where 0 = no monotonic relationship and 1.00 indicates perfect, positive monotone relationship. Note that ranks stay invariant with respect to monotone transformations.

To more formally compare the approaches’ informativeness, we used maximum likelihood to estimate Shannon’s entropy [43]. Lower entropy values indicate that more information is contained within a summary or composite score. BIC and AIC indices indicated that a Weibull distribution was a better fit for the BPI-I and a two-parameter Gamma distribution, for the MoPSI. Shannon’s entropy \(E\) for Weibull distributions is calculated as follows:

$$E=\gamma \left(1-\frac{1}{k}\right)+\text{ln}\left(\frac{k}{l}\right)+1$$
(6)

where \(\gamma\) is Euler’s constant (approximately 0.57712), \(k\) = shape and \(l\) = scale. “Shape” refers to the actual shape or form of the distribution (e.g., is it peaked, rounded, flat?) and “scale” refers to the distribution’s dispersion. For two-parameter Gamma distribution, the entropy is:

$$E=a-\text{log}\left(b\right)+\text{log}\left(\Gamma\left(a\right)\right)+\left(1-a\right)*\frac{\text{d}\text{log}\left(\Gamma\left(a\right)\right)}{\text{d}a}$$
(7)

where \(a\) = shape and \(b\) = rate, where the rate is the inverse of scale.

For each BPI-I and MoPSI scoring approach, we also examined the number of distinct values (person-scores) returned by each scoring method. A higher number of distinct values signifies greater granularity and informativeness.

Outcomes

The main outcome is development of an unsupervised dimension reduction method based on joint pdf and noticing that the joint probability function under some plausible conditions may be used as an index. This composite score is the most informative unidimensional reduction of multivariate data. The BPI-I and MoPSI JPD-scores were developed as two applications of these results.

Power

This research does not test any hypothesis and hence power analysis is not relevant.

Results

Table 4 shows the summary statistics of the different BPI-I and MoPSI scoring approaches.

Table 4 Distribution statistics from scoring methods of the brief pain inventory-interference scale and scoring methods of the manifestations of psychiatric severity index

Table 5 shows the inter-item correlations of the seven BPI-I items and the item-to-score rank-correlations. As can be seen in the Table, all seven items are positively correlated and hence, unidirectional. Rank-correlations of the 7 items with the summary (composite) scores range from 0.605 to 0.897, suggesting moderate to strong co-directionality of all scorings with the items. The high rank-correlation between the standard, CFA, IRT scores, and JPD summary indices for BPI-I scoring suggests the JPD method produces scores with ranks highly concordant with the other methods. This demonstrates that JPD scoring encapsulates all the information captured by covariance structure decomposition (CFA), Persons’ vulnerability scores in IRT sub-models restricted by local independence assumption. Note that high rank correlations between IRT scorings and CFA confirms that the broader IRT models agree with CFA, provided the covariance matrix is the only set of parameters identifying the joint probability distribution. For centered set of indicators with a spherically symmetric distribution, like a multivariate normal distribution, CFA scoring then is equivalent to an appropriate IRT and more generally, the JPD scoring.

Table 5 Spearman correlations between brief pain inventory interference items and composite scores

Figure 1 shows the BPI-I’s 7 items’ information curves for Graded Response model. As expected, the items are more informative (Fisher information) about participants with middle size vulnerabilities. The items specially items 1 (mood), 4 (walking) and), 7(sleep) are more discriminating items for these vulnerabilities Generalized Partial Credit and Partial Credit models’ information curves, which we have not reported here, demonstrated similar patterns.

Fig. 1
figure 1

Item characteristic curves for the 7 Brief Pain Inventory Interference items. BPI-I = Brief Pain Inventory Interference scale

Figure 2 depicts items’ information curves for MoPSI indicators. Number of mental health Clinic visits (item 1) provides information over a wider range, including the less sever patients, while number of emergency department visits for mental health concerns (item 2) is more informative, almost non-overlapping about more severe cases (higher Person’s values). Number of psychiatric hospitalizations is the most informative (item 3) to discriminate between subjects with mainly above average severity. The item shows an erratic behavior, however even in its least informative situations is much more information than all other items (4–6): self-harm, substance, and alcohol use. These items are less-discriminating items provided less information but over a wider range of more severe cases.

Fig. 2
figure 2

Person-Item map for the 7 Brief Pain Inventory Interference items. Solid circles present the spread of item difficulties and hollow circles, response thresholds across the latent trait (vulnerability). Items are sorted along the y-axis by difficulty. Vul. Dist. = Vulnerability Distribution

Table 6 shows the inter-item correlations of the six MoPSI indicators and the item-to-total-score correlations for the 2 scoring versions. As Table 6 shows, the six indicators are positively correlated and thus unidirectional. Correlations of the indicators with the summary scores are positive, again supporting evidence of co-directionality. The indicators’ higher correlations with the JPD method compared to the IRT method suggests that the JPD method generally ranks patients closer to their observed indicators’ rankings. As with the BPI-I, there is a high rank correlation between scores obtained via alternative models.

Table 6 Spearman rank correlations between manifestation of psychiatric severity items and composite scores

BPI-I is a well-established items and their properties have been reported by several authors. To understand the MoPSI items behavior more closely, Fig. 3 shows the MoPSI’s 6 ICC plots, and Fig. 4, the PC model person-item map. As can be seen from Fig. 3, the 3 ICCs for binary items (suicidality, alcohol use, substance use) show that patient with higher severities have higher probabilities of endorsing them and low sever patients most likely will report no use or thoughts of suicidality are steep and thus can distinguish between more and less vulnerable participants. As Fig. 3, Panels A, B and F show higher values are reported by patients with higher severity while lower values are reported with high probability by less sever patients. ICCs for mental health care visits between 0 and 45 assessed a wider range of vulnerability, which is desirable if one wishes to rank-order individuals along a broad continuum, rather than identify only high (low) vulnerability. This is echoed in Fig. 4, where it shows that the number of mental health clinic visits is most important in covering the lower range of vulnerability, while the number of psychiatric hospitalizations, self-harm diagnoses, and the number of emergency department visits for psychiatric reasons are more important in capturing the higher range of vulnerability.

Fig. 3
figure 3

Item characteristic curves for the 6 manifestations of psychiatric severity index items

Fig. 4
figure 4

Person-Item map for the 6 Manifestations of Psychiatric Severity Index items. Solid circles present the spread of item difficulties and hollow circles, response thresholds across the latent trait (vulnerability). Items are sorted along the y-axis by difficulty. Vul. Dist. = Vulnerability Distribution. MH = Mental health. Dx = diagnosis. ED = Emergency department

Figure 5 shows the density plots of the BPI-I scorings, and Fig. 6, the density plot of the psych severity indices. As can be seen from Fig. 5, the density of the JPD score has highest Pearson Divergence from the uniform distribution over the interval [0,10]. Indeed, the Divergences for BPI-I scoring densities are proportional to: 0.004(BPI-I score), 0.035(JPD),0.007(GRD),0.007(GPC), 0.007(PC) and 0.006(CFA). The same observation holds for the densities of the psych severity scorings:

Fig. 5
figure 5

Density plots for brief pain inventory interference scorings. BPI-I = Brief Pain Inventory Interference scale. JPD = joint probability density

Fig. 6
figure 6

Density plots of manifestations of psychiatric severity index scorings. MoPSI = Manifestations of Psychiatric Severity Index. JPD = joint probability density

Shannon’s entropy was consistently lower for the JPD method than for the other methods. Pearson Divergence of the alternaive indices’ densities from the uniform density over the common 0–10 range. The JPD Divergence for BPI-I, is 5 to 8 times bigger than the other indices’ Divergences. The Pearson’s Divergence of distributions of MoPSI scorings from the uniform distribution are proportional to: 0.432, 0.006, 0.034, 0.015, 0.004, 0.006, 0.005, respectively for JPD, PC, CFA (Spearman), CFA (Polychoric), CFA (Mixture of measures of associations), GR and GPC models. As expected, The JPD Divergence is again much larger than the alternatives.

For psych severity scorings, there were no significant difference between the Graded Response model and Generalized Partial Credit model. Both fitted the data. However, all these three models, across different fit indices received a moderate support by data. See Table 7 below for the details:

Table 7 IRT Model Comparisons

The JPD, GRM, GPC, CFA methods for BPI-I returned 7,778 distinct ranks compared to 480 distinct values for the conventional and PC methods. Likewise, the MoPSI’s JPD method produced 121 distinct values (person-scores) compared to 40 values produced by PC modeling. Both Graded Response and Generalized Partial Credit models ranked the subjects into 68 rank-equivalent groups, indicating that the JPD method generally is more granular and can produce finer separations in rank-order.

Discussion

We showed that using p-dimensional joint probability density function and the notion of entropy one can produce q (q < p) statistics which are most informative. This unsupervised DR theory does not require a notion of sufficiency as is common in regression DR literature. In addition, it is always possible to reduce the p-dimensional variable into a single statistic being most informative. This single statistic necessarily may not be interpretable as a latent construct or plausible summary useful to encapsulate a broader concept or characteristic. We provided the two conditions of unidirectionality and codirectionality, under which the summary measure (the index, the composite score) will rank the subjects(units) approximately the same way as the variables (the indicators) will rank. The closeness the index ranking with p rankings depends on how strong the codirectionality and unidirectionaly hold. Factor Analysis, which is based on the second moments and cross moments of the joint pdf in the case that the variance-covariances exactly or at least approximately identify the joint pdf, and when unidimensionality of the indicators, in some sense holds, the unidiemensional factor scoring and JPD will produce similar ranking. Similarly, IRT models, which are based on the joint pdf of items satisfying a set of plausible conditions, most importantly: local independence and unidimensionality, naturally should result the indices similar to the estimated JPD index. We used this result to validate our estimated index against these more established procedures. A BPI-I estimated JPD index summarizing 7 items had very good Spearman rank correlations with those produced by confirmatory factor analysis Credit and Partial credit models. the same way the variable will rank. See Fig. 7 below:

Fig. 7
figure 7

Scatterplots of JPD against the IRT models BPI-I scoring. JPD = joint probability density

Very similar to the BPI-I indices, the MoPSI JPD estimator of 6 indicators with varying scale of measurements produced a psychiatric severity index similar to these alternative methodologies, though the strength of rank dependencies were moderate. In both cases, however, the JPD version proved more informative and granular. Conditional specification of the joint probability density has recently gained momentum in machine learning and anomaly and outlier detection, e.g., [44,45,46,47]. Results here suggest that estimating indicators’ joint probability density may also be useful for creating indices, particularly when dealing with set of multivariate data where higher moments, bigger than 2, are more informative or when IRT or classical test theory assumptions (e.g. linear dimension reduction) are not met.

BPI users should be reassured by yet another entry supporting the robustness of the BPI interference scale. While all the items appeared to be well-calibrated. Although we did not pursue it in the present paper, the JPD summary as the most informative non-linear summary might be a desirable second step in scoring when more common techniques, such as factor analysis, which is based on the linear decomposition of covariance structure, produce more than one meaningful (linear) factor. In such multidimensional (linear) latent structures, one might find lower dimensional latent construct that are non-linear. As we mentioned earlier, the single non-linear JPD index might not be interpretable as a latent trait, but when it does, it will be the best (most informative) representation of that latent construct. Were the BPI Severity and Interference subscales integrated into a single, fully informative scalar using the JPD method, investigators and clinicians could rank patients along a single pain experience continuum. This could be particularly helpful in addressing problems of rank-order when patients report low severity scores but high interference scores or vice versa.

This paper also produced a brief, novel measure of psychiatric severity based on administrative data. IRT analysis showed that emergency department visits and self-harm diagnoses best discriminated between individuals with high and low vulnerability, while the number of mental health visits varied over a much wider range of latent traits. Taken together, the six MoPSI items can discriminate individuals over both a wide and narrow range of vulnerability and have adequate measurement validity. The MoPSI’s concurrent, construct, and predictive validity have been reported elsewhere, and appear promising [34].

Limitations

To our knowledge, using joint probability density estimation to create indices has not been previously explored. The R code, included in Additional File 1, is customizable for others’ use by substituting their variable names (and data) for ours.

Estimating multivariate densities is an old, ongoing challenge in statistics [48, 49]. Even with parametric models, higher dimensional multivariate densities have intrinsic challenges [23] and non-parametric models, even more. We used generalized linear modelling of the mean(proportion) r with simple linear combinations of indicators (no interaction, higher orders or non-linearity in the parameters) to approximate the dependencies and then estimate the node probabilities. While this choice is not necessarily a limitation, as it was confirmed by alternative scoring methods, other more elaborate multivariate density estimations can also be used [48] or even combined via a super learning recipe. Our averaging of all p! individual factorizations is equivalent to constructing a simple super-learner combining the estimators constructed from different methoeds.

Using Bayesian Network (BN) analysis, chain decomposition smaller than p! could be possible when there is specific knowledge about dependencies [50]. If lacking content knowledge, one might try to discover the conditional independencies between the indicators, then reduce the joint model to products of separate conditional probabilities. Unfortunately, BN structure discovery depends on several factors, including the optimality search criteria used, the testing methods and employed models, and search algorithms [51] and [52]. As we showed in our example in Additional File 2, Appendix Table, errors in any one of these factors can result in an unreliable model. In the present analysis, we avoided specifying the plausible dependency structures by using all possible conditional factor models, albeit at the expense of high computational cost. Fortunately, the small and large sample statistical optimality properties of JPD index, as a multivariate density estimation problem is an ongoing research agenda. However, its use under various scenarios, e.g., when a set of categorical variables (such as gender, or, race) exist, needs further studies. The use of this unsupervised DR method at the present does not, of course, addresses the variable selection questions, its applications in (time-varying) high-dimensional models, where even for high-dimensional multivariate Gaussian model surprising results are possible [53]. A particular important problem we have not addressed is the question of constructing the most informative q-dimensional linear summaries (q < p) where the conditional distribution of the p-dimensional indicators given the q linear summaries is Uniform. Such linear reductions are possible when the indicators have an elliptical distribution [54].

Conclusions

An unsupervised marginal probabilistic dimensionality reduction method is possible. Under conditions of ordinality of the indicators, and their unidirectionality, and co-directionality, using the joint probability density one may reduce several indicators to a single scalar. The JPD method does not discard information and can handle challenging data types that may not be well-accommodated using more traditional techniques, provided an estimable multivariate density exist.

Availability of data and materials

The data that support the findings of this study are available from the corresponding author (SN) upon reasonable request and with local Minneapolis VA Health Care System IRB permission.

Abbreviations

AIC:

Akaike information criterion

BN:

Bayesian network

BIC:

Bayesian information criterion

BMI:

Body mass index

BPI:

Short-form Brief Pain Inventory

BPI-I:

Brief Pain Inventory-Interference

Dx:

Diagnosis

ED:

Emergency department

EPOCH:

Effects of prescription opioid changes study

eRm:

Extended rasch modeling R package

glm:

Generalized linear model R package

GPC:

Generalized partial credit model

GRD:

Graded response model

ICC:

Item characteristic curves

ICD-9-CM:

International Classification of Diseases, Clinical Modification, 9th revision

ICD-10-CM:

International Classification of Diseases, Clinical Modification, 10t revision

IQR:

Interquartile range

JPD:

Joint probability density

MH:

Mental health

MI:

Mutual information

MoPSI:

Manifestations of psychiatric severity index

PC:

Partial credit model

Pdf:

Probability density function

PIG:

Poisson inverse gaussian

TRACTION:

Long Term Outcomes in Veterans Requesting a Compensation Study

SAS:

Statistical analysis system

SD:

Standard deviation

Vul. Dist:

Vulnerability distribution

References

  1. Li K. Sliced inverse regression for dimension reduction, with discussions. J Am Statist Assoc. 1991;86:316–42.

    Article  Google Scholar 

  2. Li K. On principal Hessian directions for data visualization and dimension reduction: another application of Stein’s lemma J. Am Statist Assoc. 1992;87:1025–39.

    Article  Google Scholar 

  3. Cook R. On the interpretation of regression plots. J Am Statist Assoc. 1994;89:177–90.

    Article  Google Scholar 

  4. Bing L. Sufficient dimension reduction: methods and applications with R.  i-304pp,  Edition 1. Boca Raton: Chapman & Hall/CRC Press; 2018. https://doi.org/10.1201/9781315119427.

  5. Hotelling H. Analysis of a complex of statistical variables into principal components. J Educ Psychol. 1933;24(6):417–41.

    Article  Google Scholar 

  6. Hofmann T, Schoellkopf B, Smola A. Kernel methods in machine learning. Ann Stat. 2008;36(3):1171–220.

    Article  Google Scholar 

  7. Schoellkopf B, Smola A, Muller K-R. Kernel principal component analysis. In: Artificial Neural Networks — ICANN ’97 7th International Conference Lausanne, Switzerland, October 8–10, 1997 Proceedings. Edited by Gerstner W, Germond A, Hasler M, Nicoud J-D. Berlin: Springer; 1997. p. 583–588. https://doi.org/10.1007/BFb0020217.

  8. Reckase MD. Multidimensional item response theory, statistics for social and behavioral sciences series. X, 354pp, Edition 1. NY: Springer New York; 2009. https://doi.org/10.1007/978-0-387-89976-3.

  9. Sijtsma Klaas, Molenaar Ivo W. (editors). Introduction to nonparametric item response theory. i-168 pp, Edition 5. Thousand Oaks: Sage Publications Inc; 2002. https://doi.org/10.4135/9781412984676.

  10. Rosenbaum P, Rubin D. The central role of the propensity score in observational studies for causal effects. Biometrika. 1983;70(1):41–55.

    Article  Google Scholar 

  11. Masters G. A Rasch model for partial credit scoring. Psychometrika. 1982;47:149–74.

    Article  Google Scholar 

  12. Muraki E. A generalized partial credit model: application of an EM algorithm. Appl Psychol Meas. 1992;16:159–76.

    Article  Google Scholar 

  13. Samejima F. Estimation of latent ability using a response pattern of graded scores. Psychometrika Monograph Suppl. 1969;34(4 Pt. 2):100.

    Google Scholar 

  14. Cheng D, Li J, Liu L, Le T, Liu J, Yu K. Sufficient dimension reduction for average causal effect estimation. Data Min Knowl Disc. 2022;36(3):1174–96.

    Article  Google Scholar 

  15. Hall P, Miller H. Using generalized correlation to effect variable selection in very high dimensional problems J. Comput Graph Stat. 2012;18(3):533–50.

    Article  Google Scholar 

  16. Johnstone I, Titterington M. Statistical challenges of high-dimensional data. Phil Trans R Soc A. 2009;367:4237–53.

    Article  PubMed  PubMed Central  Google Scholar 

  17. Liliana Forzani and R. Dennis Cook. Partial Least Squares Regression: and Related Dimension Reduction Methods. CRC Press, 412pp, 2024.

  18. Chao, Guoqing ; Luo, Yuan ; Ding, Weiping. Recent Advances in Supervised Dimension Reduction: A Surve.y Machine learning and knowledge extraction, 2019–01, Vol. (1), p.341–358. Multidisciplinary Digital Publishing Institute.

  19. Lee A, Verleysen M. Nonlinear dimensionality reduction. XVII, 309 pp springer science & business media, springer-verlag New York, 2007. https://doi.org/10.1007/978-0-387-39351-3.

  20. Christopher J.C. Burges. Dimension reduction: A guided tour. i,104 pp, Edition 1. Boston: Now Foundations and Trend Publishers Inc.; 2010. https://doi.org/10.1561/2200000002.

  21. Benyamin Ghojogh, Mark Crowley, Fakhri Karray, Ali Ghodsi. Elements of dimensionality reduction and manifold learning, springer textbooks, XXVIII, 606, Edition 1. Switzerland: Springer International Publishing AG - Cham; 2023. https://doi.org/10.1007/978-3-031-10602-6.

  22. Chakrabartty SN. Scoring and analysis of likert scale: few approaches. J Knowledge Manag Inform Technol.  2014;1(2):3, 1–44. https://www.researchgate.net/profile/Sn-Chakrabartty/publication/321268871_Scoring_and_Analysis_of_Likert_Scale_Few_Approaches/links/5d5e623392851c37637173ba/Scoring-and-Analysis-of-Likert-Scale-Few-Approaches.pdf.

  23. Sava AM. Weighting Method for Developing Composite Indices: Application for Measuring Sectoral Specialization. Journal of Applied Quantitative Methods. 2016;II(3):77–84.

    Google Scholar 

  24. Nardo M., Saisana, M., Saltelli, A., Tarantola, S. (2005) “Tools for Composite OECD. Handbook on Constructing Composite Indicators: Methodology and User Guide. European Union, European Commission, Joint Research Centre OECD Publishing, Aug 22, 2008 - 162 pages

  25. Abeyasekera S. Multivariate methods for index construction. Household sample surveys in developing and transition countries department of economic and social affairs. Statistics division, studies in methods chapter 18. Series F, No. 96. New Yor: United Nations; 2015.

  26. Lord FM, Novick MR. Statistical theories of mental test scores. 1-568 pp, Edition 1. Reading MA: Addison-Wesley Publishing; 1969.

  27. Rasch G. Probabilistic models for some intelligence and attainment tests, xxiii-199 pp, Edition 1. Chicago: The University of Chicago Press; 1980. [Note: Also published in: The SAGE Encyclopedia of Research Design (1981)].

  28. Bock RD, Gibbons RD. Item response theory. Hoboken: John Wiley & Sons; 2021. p. xvii–366.

  29. Krebs E, Clothier B, Nugent S, Jensen A, Martinson B, Goldsmith E. The evaluating prescription opioid changes in veterans (EPOCH) study: design, survey response, and baseline characteristics. PLoS ONE. 2020;15(4). https://doi.org/10.1371/journal.pone.0230751. https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0230751.

  30. Murdoch M, Clothier B, Beebe T, Bangerter A, Noorbaloochi S. Impact of different cover letter content and incentives on non-response biasin a sample of Veterans applying for Department of Veterans Affairs disability benefits: a randomized, 3x2x2 factorial trial. BMC Med Res  Methodol. 2022;22:308. https://doi.org/10.1186/s12874-022-01783-7.

  31. Cleeland C. The brief pain inventory. User guide. Houston: MD Anderson Center; 2009. p. 166. https://www.mdanderson.org/documents/Departments-and-Divisions/Symptom-Research/BPI_UserGuide.pdf.

  32. Keller S, Bann C, Dodd S, Schein J, Mendoza T, Cleeland C. Validity of the brief pain inventory for use in documenting the outcomes of patients with noncancer pain. Clin J Pain. 2004;20(5):309–18. https://doi.org/10.1097/00002508-200409000-0000538.

  33. Tan G, Jensen M. JI T, Shanti B: Validation of the Brief Pain Inventory for chronic nonmalignant pain. J Pain. 2004;5(2):133–7.

    Article  PubMed  Google Scholar 

  34. Murdoch M, Clothier B, Noorbaloochi S. A new measure of psychiatric severity using administrative data: the Manifestations of Psychiatric Severity Index (MoPSI). PsyArXiv; 2023. https://doi.org/10.31234/osf.io/s5bdnchttps://osf.io/preprints/psyarxiv/s5bdn.

  35. Radbruch L, Loick G, Kiencke P, Lindena G, Sabatowski R, Grond S, Lehmann KA, Cleeland CS. Validation of the German version of the Brief Pain Inventory. J Pain Symptom Manage. 1999;18(3):180–7.

    Article  CAS  PubMed  Google Scholar 

  36. Andrich D, Marais I. A Course in Rasch Measurement Theory. Singapore: Springer; 2019.

    Book  Google Scholar 

  37. Mueller H. A Rasch model for continuous ratings. Psychometrika. 1987;52:165–81.

    Article  Google Scholar 

  38. Chalmers RP. mirt: A Multidimensional item response theory package for the R environment. J Stat Soft. 2012;48(6):1–29. https://doi.org/10.18637/jss.v048.i06.

  39. Mair P, Hatzinger R. Extended Rasch modeling: The eRm package for the application of IRT models in R. J Statistical Software. 2007;20:1–20.

    Article  Google Scholar 

  40. William Revelle. psych: Procedures for Psychological, Psychometric, and Personality Research_. Northwestern University, Evanston, Illinois. 2023, R package version 2.3.12, <https://CRAN.R-project.org/package=psych>.

  41. Lorenzo-Seva U, Ferrando PJ. Not positive definite correlation matrices in exploratory item factor analysis: Causes, consequences and a proposed solution. Struct Equ Model. 2021;28(1):138–47.

    Article  Google Scholar 

  42. Zumbo B, Gadermann A, Zeisser C. Ordinal versions of coefficients alpha and theta for likert rating scales. JMASM. 2007;6:21–9.

    Article  Google Scholar 

  43. Shannon C. A Mathematical Theory of Communication. Bell Syst Tech J. 1948;27(3):379–423.

    Article  Google Scholar 

  44. Liu Q, Xu J, Jiang R, Wong W. Density estimation using deep generative neural networks. PNAS. 2021;118(15):e2101344118.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  45. Arnold B, Castillo E, Sarabia J. Conditional specification of statistical models. New York, NY: Springer-Verlag; 1999.

    Google Scholar 

  46. Nachman B, Shih D. Anomaly detection with density estimation. Physical Review D. 2020;101(7): 075042.

    Article  CAS  Google Scholar 

  47. Liu B, Tan P-N, Zhou J. Unsupervised anomaly detection by robust density estimation. Proceedings of the AAAI Conference on Artificial Intelligence. 2022;36(4):4101–8.

    Article  PubMed  Google Scholar 

  48. Scott, David W. Multivariate density estimation: theory, practice, and visualization. Second Edition, Wiley Series in Probability and Statistics. Hoboken: John Wiley & Sons, Inc.; 2015. https://doi.org/10.1002/9781118575574.

  49. Parzen E. On estimation of a probability density function and mode. Ann Math Statist. 1962;33(3):1065–76.

    Article  Google Scholar 

  50. Pearl J. Probabilistic reasoning in intelligent systems: networks of plausible inference. Edition 2. San Mateo: Morgan Kaufmann Publishers; 1988. p. 1–554.

    Google Scholar 

  51. Scutari M. Learning Bayesian networks with the Bnlearn R Package. J Stat Soft. 2010;35(3):1–22. https://doi.org/10.18637/jss.v035.i03.

    Article  Google Scholar 

  52. Scutari M, Denis J. Bayesian networks with examples in R. Edition 2. New York: Taylor & Francis Group; 2021. p. i–274. https://doi.org/10.1201/9780429347436.

    Book  Google Scholar 

  53. Pourahmadi M. High-dimensional covariance estimation: with high-dimensional data. Hoboken: John Wiley & Sons, Inc.; 2013. p. 1–204. https://doi.org/10.1002/9781118573617.

    Book  Google Scholar 

  54. Cambanis S, Huang S, Simons G. On the theory of elliptically contoured distributions. J Multivariate Analysis. 1981;11(3):368–85.

    Article  Google Scholar 

Download references

Acknowledgements

We thank Erin Krebs for access to the EPOCH data.

Disclaimer

Views expressed are solely those of the authors and do not reflect the opinion, views, policies, or position of the Department of Veterans Affairs.

Funding

The Center for Care Delivery Outcomes Research is a VA Health Services Research and Development (HSR&D) Service Center of Innovation (Center grant #HFP 98–001). This work was supported by the VA HSR&D Service (grant numbers IIR 14–295 and IIR-14–004) and National Institutes of Health (F30AT009162). The funders had no role in data analysis, manuscript preparation, or decision to publish.

Author information

Authors and Affiliations

Authors

Contributions

SN oversaw analysis and interpretation, contributed to funding acquisition and study design, and drafted the manuscript. BAC contributed to funding acquisition, study design, data analysis and interpretation. MM obtained funding; designed the study; oversaw data collection and contributed to analysis and interpretation. SN, BAC, and MM contributed to manuscript revisions and read and approved the final manuscript.

Corresponding author

Correspondence to Siamak Noorbaloochi.

Ethics declarations

Ethics approval and consent to participate

The Minneapolis VA Health Care System’s Internal Review Board (IRB) for Human Studies reviewed the two study protocols (#4495-B and #4586-A). All methods were carried out in aMcordance with relevant guidelines and regulations.

Consent for publication

Not applicable.

Competing interests

The authors declare no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Noorbaloochi, S., Clothier, B.A. & Murdoch, M. Using joint probability density to create most informative unidimensional indices: a new method using pain and psychiatric severity as examples. BMC Med Res Methodol 24, 171 (2024). https://doi.org/10.1186/s12874-024-02299-y

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12874-024-02299-y

Keywords