Abstract
Keywords
Risk assessment for crime and violence has grown at a very fast pace during the last 30 years, especially in North America. Following the media coverage of particularly sordid sexual aggressions and murders by recidivists, decision-makers from Canadian and American governments funded the development and implementation of reliable and valid risk assessment procedures and correctional interventions based on the risk-need-responsivity model (Andrews & Bonta, 2010; Brouillette-Alarie & Lussier, 2018).
The first wave of reliable risk assessment tools is known as static actuarial assessment, a method that succeeded unstructured clinical judgment with its lackluster inter-rater agreement and poor predictive validity (Dawes et al., 1989; Grove et al., 2000). Actuarial assessment reliably determines the level of risk by mechanically combining empirically validated predictors. This method is considered “atheoretical,” because the main inclusion criterion of an item in a scale is its statistical association with the outcome of interest, not its theoretical relevance. The first actuarial scales comprised static risk factors only, which strengthened the perception that these tools were largely atheoretical (Andrews & Bonta, 2010; Bonta, 1996). The second wave of actuarial scales was defined by its inclusion of dynamic risk factors, also known as criminogenic needs. Therefore, they were better positioned to follow the evolution of risk over time and suggest intervention targets than instruments that exclusively comprised static risk factors (Andrews & Bonta, 2010; Bonta, 1996, 2002; Gendreau et al., 1996; Hanson & Harris, 2000). The most recent generation of actuarial scales is case management risk/need tools (Andrews & Bonta, 2010). In addition to assessing static and dynamic risk factors, this generation provides clear guidelines to ensure that case management will be consistent with the results of risk assessment, according to the risk and need principles of effective correctional intervention. The Level of Service/Case Management Inventory (LS/CMI; Andrews et al., 2004) is a prime example of a case management risk/need tool that assesses general recidivism risk.
Having reliable and valid risk assessments has numerous advantages. Overestimation of risk can lead to the long-term imprisonment of individuals who could otherwise become productive members of society or actively impair their chances of reintegration upon release. Indeed, high-risk statuses, such as “sexually violent predator” are known to be significant obstacles to reintegration, limiting housing, and employment opportunities (Harris, 2014). Conversely, underestimation of risk can lead to the release of dangerous individuals and result in new victims. Therefore, precise risk assessments that are neither too high nor too low have been increasingly seen as a cornerstone of correctional practice since the early 1990s (Brouillette-Alarie & Lussier, 2018).
The Focus of Risk Tools on Predictive Validity and the Neglect of Other Psychometric Aspects
Because the primary objective of actuarial scales is to make the most accurate predictions to inform risk management and intervention efforts, studies on risk tools have traditionally focused on predictive validity rather than construct validity or other psychometric aspects (Helmus & Babchishin, 2017). Contrarily to psychometric tests from the field of psychology or ability tests from the field of education, criminological risk tools are more interested in making an accurate prediction about the risk of reoffending than determining the ability of an individual on a construct (e.g., extraversion/introversion, algebra skills). Therefore, risk tool validation studies have typically neglected relevant sources of evidence that are potentially necessary to justify the interpretation and uses of scores stemming from actuarial scales (Helmus & Babchishin, 2017; Messick, 1989). At the forefront of these sources of evidence lies construct validity, the degree to which a test measures what it claims to be measuring (Cronbach & Meehl, 1955).
Many authors have advocated for the integration of construct-oriented approaches in criminological risk assessment practice (Babchishin et al., 2016; Brouillette-Alarie et al., 2016, 2022; Mann et al., 2010). Clarifying the construct validity of risk tools has many potential advantages for the field. First, it offers insight into why certain scales predict certain outcomes better than others, as this is dependent on the constructs they assess and how each construct is weighted in these scales. This, in turn, can help evaluators integrate the potentially conflicting results of risk scales when multiple tools are available for the same population but arrive at different conclusions (Barbaree et al., 2009). Second, understanding the constructs implicit in risk tools can improve predictive accuracy. Specifically, when the constructs are known, it is possible to improve the reliability and validity of their assessment using standard psychometric methods and, therefore, improve the predictive accuracy of scales (Brouillette-Alarie et al., 2016). Finally, construct-level approaches maximize the clinical relevance of existing scales by focusing on psychological dimensions and their nomological network, facilitating the identification of the “source” of the risk. Evaluations that address psychological features are generally better received by clinicians, practitioners, and decision-makers than those that only delineate the level of risk (Mann et al., 2010).
Over the last 10 to 15 years, studies have increasingly considered the construct validity of risk tools by using standard psychometric methods, such as factor analysis and convergent/discriminant validity analyses (Babchishin & Hanson, 2020; Brouillette-Alarie et al., 2016, 2018; Brouillette-Alarie & Proulx, 2019; Gordon et al., 2015). For example, the latent constructs of the Static-99R and Static-2002R (Hanson & Thornton, 2000, 2003; Helmus et al., 2012)—static risk tools for persons convicted of a sexual offense—were studied in a research program that led to a three-factor model comprising sexual criminality, general criminality, and “youthful stranger aggression” (a factor comprising items related to youth and victim harm; Brouillette-Alarie et al., 2016). Then, the nomological network of these static dimensions was studied by linking them with psychologically relevant items/constructs and recidivism outcomes (Brouillette-Alarie et al., 2018; Brouillette-Alarie & Proulx, 2019).
General criminality risk tools, such as the LS/CMI have also been subjected to factor analyses, though less often than risk tools for individuals convicted of a sexual offense. Gordon et al. (2015) reported that there were few available factor analyses of the LS/CMI and referred readers to analyses of the Level of Service Inventory—Revised (LSI-R; Andrews & Bonta, 1995), its predecessor, for insight into the latent structure of risk tools for “general” convicted persons. Studies of the LSI-R’s factor structure are not numerous and have been criticized due to their methodological choices (e.g., Andrews & Robinson, 1984; Arens et al., 1996). Indeed, these studies mostly used principal component analysis over factor analysis and entered dimensions rather than items in the analysis, a surprising choice that seemed to be driven by the desire to obtain a single resulting factor (Hsu et al., 2011). To correct the aforementioned limitations, Hsu et al. (2011) conducted a factor analysis of LSI-R items and obtained five dimensions for men and four for women. The four dimensions common to both genders were static risk, employment, pro-criminal attitudes, and mental health. The fifth dimension, exclusively present in men, was protective companions.
The Rasch Model and Its Potential Relevance for Risk Tool Validation
Another important and relatively recent means to study construct validity has been item response theory (IRT) and Rasch models. Scholars from the field of education have long advocated the use of such models, as they are less sensitive than classical test theory models to circular dependency (i.e., dependence on the overall performance of the validation sample; de Ayala, 2009). IRT was introduced in the 1950s and 1960s (Lord, 1953; Rasch, 1960) to better assess item difficulty and discrimination and create sample-free measures (Osteen, 2010). The Rasch model can be seen as a continuous line on which individuals and items are both placed according to their respective skill level/difficulty. Applied to the field of criminology, the Rasch model could establish which items are more difficult to endorse for individuals whose “disposition toward crime” is under study; endorsing an item considered difficult could enhance an individual’s disposition to crime more than an item everyone is likely to endorse. When analyzing data through the Rasch model, fit statistics are calculated, showing which items are too predictable or too unstable to help the model fit the data. Misfitting items can then be kept, discarded, or reviewed by experts (Bohlig et al., 1998).
Therefore, IRT and Rasch models could offer interesting avenues for improving existing risk scales or making shorter versions of them by removing unfitting or redundant items. Even though complete assessments of risk and needs are generally preferable to screening versions, short versions of risk scales can be relevant for jurisdictions where time and resources are limited. Screening versions of the LSI-R (Andrews & Bonta, 1998) and Psychopathy Checklist-Revised (Hart et al., 1995) have been developed, and IRT techniques could help to review or contextualize decisions that were made in creating short versions of these scales.
Another potential advantage of IRT-based models concerns refined item weighting. As of now, most actuarial scales comprise items worth one point each that are summed to determine the total risk score. However, as it pertains to face validity, it is unlikely that items on actuarial scales are all equally difficult and, thus, equally risk relevant. Therefore, it is not impossible that a more refined weighting of items according to their difficulty could enable improvements in predictive validity (Brouillette-Alarie et al., 2022). There are debates on the tangible benefits of differentially weighting items, with some results indicating that complex combinations rarely outperform the simple summing of dichotomous items (Ghiselli et al., 1981; Grann & Långström, 2007; Silver et al., 2000). However, differential weighting has its greatest impact when there is a wide variation in weighting values, little intercorrelation between items, and only a few items (Ghiselli et al., 1981; Kline, 2005). Considering that actuarial scales usually comprise few nonredundant items, they could potentially benefit from differential weighting (Georgiou, 2019).
Another overarching benefit of Rasch and IRT models is that they bring the focus to test items rather than total scores. Even though classical test theory comprises techniques to assess the performance of individual items, for example, difficulty calculations and item-total correlations, these tests tend to be underreported in non-IRT articles in the criminological literature. As the next section will illustrate, some of the most heavily used actuarial scales in corrections report very few (if any) data on the performance of their individual items.
The LS/CMI
The LS/CMI is the evolution of the Level of Service Inventory-Revised (LSI-R) and is the most heavily used actuarial scale for the prediction of general recidivism internationally (Wormith, 2011). It relies on the substantial body of work by Don Andrews and James Bonta and their theory of the psychology of criminal conduct (Andrews & Bonta, 2010). The scale has been validated with men, women, adults, adolescents, incarcerated persons, and persons on parole (for reviews, see Andrews & Bonta, 2010; Olver et al., 2014). It is one of the few fourth-generation risk tools, as it integrates case management elements in addition to third-generation risk assessment procedures (Brouillette-Alarie & Lussier, 2018). Case management sections are based on the tried-and-true risk, need, and responsivity principles of effective correctional intervention (Andrews & Bonta, 2010).
The LS/CMI comprises eight risk domains that have all demonstrated predictive validity towards general recidivism (Andrews & Bonta, 2010; Andrews et al., 2011; Olver et al., 2014). In the Canadian context, the predictive validity of its total score rivals that of the best tools in the field, with area under the curve (AUCs) reaching .75 or more depending on the validation sample (Andrews et al., 2011; Olver et al., 2014). In samples from the United States, the predictive validity of the LS measures in general was found to be much lower (Olver et al., 2014). Despite the substantial literature on the dimensions and total scores of the LS/CMI, very few (if any) data are available on the performance of its individual items. We are aware of no studies that relate to the predictive validity of LS/CMI items and only two studies that report other parameters, such as item-total correlations, difficulty, and discrimination (Giguère et al., 2015; Giguère & Lussier, 2016). The first study looked at these parameters using classical test theory, and the second one used two-parameter IRT. In the latter, Giguère and Lussier (2016) found that many LS/CMI items were redundant or displayed problematic discrimination and/or difficulty values. After removing the problematic items, they found that the remaining items achieved a predictive validity that was very close to that of the full 43 items.
Another application of IRT for risk tool validation can be found in Huang et al. (2021), who investigated the generalizability of the Youth Level of Service/Case Management Inventory (YLS/CMI; Hoge & Andrews, 2011) using a sample of Indigenous and non-Indigenous youth. Differential item functioning analyses indicated that items from the education domain were less likely to be endorsed by Indigenous youth, while items from the substance abuse domain were more likely to be endorsed. Importantly, predictive validity analyses revealed that the YLS/CMI was not predictive of criminal recidivism for Indigenous youth.
Even though we commend the LS/CMI for its sound theoretical underpinnings and substantial validation as it relates to its dimensional and total scores, the lack of publicly available data on the performance of its items is a significant limitation that needs addressing. Data on the psychometric properties of individual LS/CMI items could enable the identification of problematic items, which could in turn lead to item deletion, reworking, or reweighting, and, hopefully, improvements in predictive validity. It could also lead the way for the development of a newer and shorter version of the LS/CMI, if problematic items were to be found.
Objectives
The current study aimed to address the lack of validation data on individual LS/CMI items using the Rasch model and predictive validity analyses. The first step of our examination was to enter the 43 LS/CMI items in the Rasch model and see which ones did not fit the model (or latent trait). Second, the difficulty parameter of each remaining item was computed. Third, a Wright map of item difficulty and person ability (Wright & Stone, 1979) was drawn to better visualize Rasch results. Fourth, the predictive validity of LS/CMI items, dimensions, and total score was tested in relation to recidivism with a 2-year follow-up. Finally, the predictive validity of items was correlated to their difficulty to verify whether difficulty in Rasch equates to disposition toward crime (predictive validity toward recidivism). Results of our analyses will be used to discuss potential improvement pathways for the LS/CMI and the relevance of IRT-based models for risk tool validation.
Method
Sample and Data Collection
In 2007, when the Act Respecting the Quebec (Canada) Correctional System came into effect, a computerized system was established to allow probation officers and prison counselors to compile information from the completed LS/CMIs of convicted individuals. For this study, the sample was taken from the
In the ERB database, the same individual could be found in multiple entries, as each of their contacts with the criminal justice system was entered in one row. To ensure that each individual would be counted once in the analyses, we kept only the most recent record of each convicted individual. The final sample (
Because LS/CMI norms are different for incarcerated persons and individuals under parole, we chose, for length and clarity purposes, to present data exclusively on incarcerated persons. Merging these groups together would have been at odds with official LS/CMI documentation (Andrews et al., 2004), and reporting results for both groups would have made this study excessively long, as a separate Rasch model would have been necessary for each group. Obtaining item-level data on individuals under parole is nevertheless an important endeavor that needs to be undertaken in the future.
Measures
The French Version of the LS/CMI
The Level of Service/Case Management Inventory (LS/CMI; Andrews et al., 2004) is an assessment and case management tool that measures the risk and need factors of late adolescent and adult convicted persons. The section of the LS/CMI relevant to our investigation is the “General Risk/Need Factors.” This section contains 43 items sorted under the following dimensions: Criminal History (8 items); Education/Employment (6 items); Family/Marital (4 items); Leisure/Recreation (2 items); Alcohol/Drug Problem (8 items); Procriminal Attitude/Orientation (4 items); and Antisocial Pattern (4 items). Each item is coded on a binary response scale (present or absent) by a probation officer or prison counselor who conducts an interview with the person and consults their criminal record. The total score thus ranges from 0 to 43 points. The total risk and dimensional scores can be used to guide surveillance, determine release conditions, plan and deliver appropriate interventions, and modulate intervention intensity. The French version of the LS/CMI was developed using a cross-cultural procedure. The translated version was translated back into English, and both versions were submitted to the developers of the LS/CMI for approval (Guay, 2016).
Criminal Recidivism
In this study, recidivism was considered to occur when an individual who had been previously convicted committed a new crime upon release. The follow-up period was of 2 years, implying that data were right censored (i.e., if recidivism happened after 2 years of follow-up, it would count as no recidivism in the analyses). Breach of conditions was not considered as a new conviction. In our sample, 29.8% of men who had been sentenced to detention reoffended in the 2-year follow-up period. In conformance with previous studies that examined the predictive value of risk evaluation instruments, only time at risk for criminal recidivism was considered (see Giguère & Lussier, 2016). This implies that the follow-up period began as soon as individuals were released from the detention center. Descriptive statistics of our study can be found in Table 1.
Descriptive Statistics (
LS/CMI items have been paraphrased to avoid copyright issues.
Analytical Strategy
Rasch Model
The classical Rasch model (Rasch, 1960) is a unidimensional measurement model that mathematically represents the relationship between item difficulty and a person’s ability to allow predictions based on the difference in logits between the two:
The basic principle is that one’s probability to succeed an item will be higher if their ability (ϴ) exceeds the item’s difficulty (
The Rasch model produces an asymptotic sigmoid curve named the item characteristic curve (ICC) that represents the probability of succeeding to an item of a given difficulty at different ability levels (ϴ). While other IRT models take into account other parameters, such as discrimination (
Monotonicity
Monotonicity implies that an individual with a high ability on a latent trait should have a greater probability to endorse an item measuring that same trait than an individual with a lower ability. This postulate may be checked by plotting the observed data on each ICC to see if the probability to endorse an item increases with greater thetas. Monotonicity was checked post-Rasch modeling to ensure that data respected this assumption. The empirical curve was plotted against the theoretical curve and, although they did not stack perfectly, they were parallel and, thus, showed that an increase in ability yielded an increase in both the probability of endorsement and the observed proportion of endorsement of items. Monotonicity graphs can be found in the Supplementary Materials.
Unidimensionality
When using a unidimensional model, such as the Rasch model, evidence must be provided that a single or dominant trait is being measured (e.g., by studying the eigenvalues produced by a factor analysis of the data). A dominant trait can be assumed if a significant drop is seen between the first and second eigenvalues. Even though it was not the main purpose of our study, a factor analysis of the 43 LS/CMI items was conducted using MPlus 6.12 to assess if the scale was “unidimensional enough” for Rasch modeling (Bertrand & Blais, 2004). This analysis was based on the guidelines for risk tool factor analysis suggested by Brouillette-Alarie et al. (2016): (a) use of tetrachoric correlation matrices, (b) weighted least squares means- and variance-adjusted extraction, and (c) oblique (geomin) rotation. The first factor had 14.53 eigenvalues and the second 3.77, which constituted a ratio of 3.85 between the first and second factor. Usually, for a scale to be considered unidimensional enough for Rasch modeling, a ratio of 3 or higher is recommended (Bertrand & Blais, 2004). Thus, for the purposes of the Rasch model, the LS/CMI was considered sufficiently unidimensional.
Local independence
This postulate assumes that the success/failure on an item is independent of the success/failure on other items, thus solely dependent on the latent trait. Correlations are expected between items because, ideally, they all measure the same trait, but beyond that, there should not be excessively high correlations within the residuals. Because collinearity between LS/CMI items is notoriously high, we opted to not remove any items
Model-data fit
Conducting a Rasch analysis yields a difficulty parameter for each item and an ability parameter for each individual, with a standard error specific to each. The process also yields fit statistics for items and individuals, namely infit and outfit in mean square format. The outfit is calculated as the average of the squares of the standardized residuals (the residuals are squared before the averaging operation) and is particularly sensitive to the unexpected responses of people whose location is far from the item. Infit is calculated by multiplying the square of each standardized residual by the variance of the expected score and is particularly sensitive to responses expected from people whose location is close to the item. If data fit perfectly with model specifications, the expected values for infit and outfit are 1. While infit is rather hard to detect and interpret, the outfit is usually prominent (Linacre, 2002).
Different ranges of reasonable fit values are available depending on the nature of the test under scrutiny (Bond & Fox, 2001). Considering that the tool under study was the LS/CMI, we opted for the values that Wright and Linacre (1994) suggested for clinical observations. Thus, only items with fit statistics between 0.5 and 1.7 were kept. While Rasch calibration is usually done iteratively, each time removing the most unfitting item until all the remaining items fall within the desired range, the calibration only needed to loop once, since only one item was deemed unfitting under Wright and Linacre’s (1994) guidelines. The fit statistics for individuals were not scrutinized because response patterns were expected to vary.
For the sake of exhaustiveness, we also tried to run the Rasch model under strict conditions, namely those suggested for high-stakes tests (fit statistics between 0.8 and 1.2; Wright & Linacre, 1994). Under these circumstances, only 20 LS/CMI items remained after removing the unfitting ones. Because the aim of this study was to obtain data on LS/CMI items, discarding more than 50% of them a priori seemed counterproductive. Therefore, we settled on the more lenient clinical observation fit thresholds. All Rasch analyses were done with Winsteps (Linacre, 2021).
Wright map
The Wright map (also referred to as the item-person map) is a useful way to visualize both items and individuals vertically on the same graphic along the continuum of the targeted unidimensional space (Wright & Stone, 1979). A Wright map makes use of the fact that the difficulty of test items can be computed, and those test-item difficulties are expressed using the same linear scale as for the person measures. A logit scale is used to express item difficulty on a linear scale that extends from negative infinity to positive infinity. Item difficulty typically ranges from −3 (very easy) to +3 logits (very difficult; Boone et al., 2013). The Wright map depicts items organized according to difficulty level on the left and individuals positioned according to ability level on the right. In the Rasch model, the scale is set to zero for the item mean.
Predictive Validity
The predictive validity of LS/CMI items, dimensions, and total scores toward criminal recidivism was assessed with the AUC of receiver operating characteristic curves. AUCs refer to the probability that a randomly selected recidivist will have a higher score than a randomly selected nonrecidivist. It is an ordinal statistic that can be compared across different scaling of predictors. Rice and Harris’s (2005) thresholds for interpreting the effect sizes of AUCs were used: .556 is equivalent to a small effect, .639 is equivalent to a moderate effect, and .714 is equivalent to a large effect. These thresholds correspond, respectively, to Cohen’s
Interface Between Difficulty and Predictive Validity
The link between the difficulty (
Results
Rasch Modeling and the Wright Map
The Rasch model took one iteration beyond the initial one to have all items fit within the specified range (between 0.5 and 1.7, inclusively). Only Item 24 ended up being discarded. Table 2 shows the item parameters and fit statistics of the first and last iterations.
Rasch Modeling and Predictive Validity of LS/CMI Items (
LS/CMI items have been paraphrased to avoid copyright issues.
AUCs are statistically significant when their 95% confidence interval does not include .50.
The 42 items were given a difficulty parameter, and the 15,961 incarcerated men were given an ability parameter through a joint maximum likelihood estimation. With all the item and person parameters estimated, the Wright map was drawn (see Figure 1), with the 15,961 participants on the left and the 42 items of the LS/CMI on the right.

Wright Map of LS/CMI Items and Participants
The participants’ “ability” (ϴ) curve seemed to follow a normal distribution, slightly skewed negatively (to the left). Item 6 was the easiest item to endorse, while Item 40 was the hardest. The Wright map indicated which items targeted specific ranges of individuals. For instance, Items 10 and 43 were very well aligned with the ability of the “average” incarcerated individual (ϴ ≈ 0), meaning that these items, when administered to these persons, generated the most desirable variance. Note that to avoid copyright issues, the names of LS/CMI items were paraphrased in the following sections.
The difficulty of individual items was ordered in the anticipated direction. For example, Item 3 (three prior convictions) was more difficult than Item 2 (two prior convictions), and the latter was more difficult than Item 1 (Any prior convictions). Item 30 (Currently has problems with alcohol consumption) was harder than Item 28 (Ever had problems with alcohol consumption). The Leisure/Recreation dimension was on average the easiest, followed by Criminal History. The two most difficult dimensions were Procriminal Attitude/Orientation and Antisocial Pattern. However, for the latter dimension, the mean difficulty was heavily influenced by Item 40, the most difficult item in our sample.
Predictive Validity
The most predictive items were Items 7, 8, and 43, with moderate effect sizes. Items 6, 12, 13, 22, and 24 were not predictive in our sample. The remaining 34 items had small effect sizes. The most predictive dimension was Criminal History, which had a large effect size. The remaining seven dimensions had moderate effect sizes. No dimension had a negligible or small effect size. The total LS/CMI score had an AUC of .761, surpassing all the individual dimensions and items (its predictive validity was, however, close to that of the Criminal History dimension).
Interface Between Difficulty and Predictive Validity
The difficulty of each item (except CO24: Links with criminalized individuals) was correlated with its predictive validity to verify if difficult items were more predictive of recidivism than easier items. The scatter plot of LS/CMI items can be found in the Supplementary Materials, with predictive validity on the X-axis and difficulty on the Y-axis. The Pearson correlation between item difficulty and predictive validity was negligible (
Discussion
The objectives of the current study were to obtain item-level data on the LS/CMI and study the interface between IRT difficulty and predictive validity. The analyses conducted provided validation data on the French version of the LS/CMI with the whole population of Quebec’s incarcerated men registered and evaluated between March 2008 and October 2015.
LS/CMI Items
Rasch model fit indices and predictive validity analyses highlighted potential problems in five items: Items 6, 12, 13, 22, and 24. For two of these items (6 and 24), sample characteristics were likely responsible for their poor performance. Indeed, according to the LS/CMI coding manual (Andrews et al., 2004), Item 6 (Ever incarcerated) and Item 24 (Links with criminalized individuals) must be endorsed for all individuals under custody. Because our sample exclusively comprised incarcerated persons, these items lacked variance, explaining their lack of predictive validity and, in the case of Item 24, exclusion from the Rasch model. If our sample had also comprised individuals under parole, the picture may have been different.
Items 12 (Did not achieve grade 10) and 13 (Did not achieve grade 12), both related to educational level, were not predictive of recidivism in our sample despite being adequately distributed. School dropout seemed like a better predictor than educational level, the same being true for Items 15, 16, and 17, which look, respectively, at performance, peer interactions, and authority interactions in schools (or jobs). It may be that problematic behaviors in school are better predictors of recidivism than educational level, which may confound multiple noncriminogenic characteristics, such as IQ, motivation, or learning style. A recent meta-analysis of risk factors for recidivism reached results similar to ours concerning the mediocre predictive validity of educational level (Goodley et al., 2022). As to Item 22 (No prosocial activities), there seemed to be no sample selection explanations for its lackluster predictive validity. We also found no studies specifically about the link between the absence of structured activities and general recidivism. Therefore, before making any conclusion concerning Item 22, the present findings would have to be replicated.
The two most predictive items were Items 7 and 8, which covered, respectively, institutional misconduct and breach of release conditions. The behaviors described by these items are known risk factors of general recidivism (Goodley et al., 2022) and figure in multiple criminological risk scales, such as the STABLE-2007 (Brankley et al., 2021; Hanson et al., 2007) and the PCL-R. In addition, an upcoming study based on machine learning algorithms concluded that these two items were the most predictive of general recidivism in a sample very similar to the one used in the current study (Arbour et al., 2022). However, this convergence of results may be partly explained by commonalities in the samples used.
The predictive validity of LS/CMI items was generally good in relation to field standards and what can be expected in predictive potency from single items. For comparison purposes, a meta-analysis of the predictive validity of Static-99R items toward sexual recidivism found that odds ratios varied between 1.22 and 2.47 (Helmus & Thornton, 2015). Odds ratios between 1.68 and 3.46 are considered small. Those between 3.47 and 6.70 are considered moderate, and those of 6.71+ are considered large (Chen et al., 2010). This would mean that for the Static-99R, the predictive validity of individual items varied between negligible and small effects. For the LS/CMI in our sample, few items had negligible effects, most had small effects, and three items reached moderate effects. Considering that two of the nonpredictive items were because of sample selection, the overall picture of the predictive validity of individual LS/CMI items appeared adequate and in line with field standards, or better. We would, however, strongly encourage replication of these results, as more studies of the characteristics of individual LS/CMI items need to be conducted.
LS/CMI Dimensions and Total Score
The most predictive dimension of the LS/CMI was Criminal History, a finding consistent with Olver et al.’s (2014) meta-analysis and the well-known reliability and predictive validity of static risk factors (e.g., Brouillette-Alarie & Lussier, 2018; Giguère & Lussier, 2016). The other dimensions all achieved moderate predictive validities that surpassed those of their constituents (items). These predictive validities were all superior to those reported by Olver et al. (2014) for the same dimensions.
The predictive validity of the LS/CMI total score for Quebec incarcerated men was in line with or better than effect sizes reported in other validation samples. In a literature review conducted by Olver et al. (2014), the predictive validity (
Rasch Difficulty and Its Relationship With Predictive Validity
Rasch modeling provided the difficulty of the 42 LS/CMI items retained in the model. The two most difficult items (
The relative difficulty of items was in the anticipated direction, especially for items under the same dimension (e.g., Item 3 was more difficult than Item 2, which was more difficult than Item 1). It was, however, harder to contrast the difficulties of items not under the same dimension. For example, dissatisfaction with one’s marital situation (FM18) was more difficult than breaching one’s conditions during supervision (CH8). If the assumptions of the Rasch model are to be believed, the former should thus be more criminogenic than the latter. Predictive validity analyses revealed a different picture. Despite its significantly lower difficulty, Item 8 was far more predictive of general recidivism than Item 18. Rasch difficulty and predictive validity were globally unrelated, which casts a doubt on some aspects of the usefulness of IRT techniques for risk tool validation. Despite the enthusiasm of some authors to completely discard CTT in favor of IRT to process criminogenic data (e.g., Osgood et al., 2002), the current study indicates that the difficulty parameter may not be useful to improve the predictive validity of actuarial scales. It may be that difficulty in mathematical tests from the field of education cannot be interpreted in the same way as difficulty in items from actuarial scales. For mathematical tests, success on a complicated item usually implies success on an easier item, as they rely on the same skill. However, in the context of risk scales, that assumption might not hold true. To echo the above example, it is unlikely that being dissatisfied with one’s marital situation (the harder item) “automatically comes” with breaching one’s parole conditions (the easier item), as these items relate to different dimensions or latent constructs.
The difficulty in interpreting LS/CMI item difficulties was exacerbated by the relative multidimensionality of the scale. Even though the LS/CMI was “unidimensional enough” to run Rasch models according to standards in the IRT field (Bertrand & Blais, 2004), it was not fully unidimensional in the factor analysis that was conducted. This finding aligns with more methodologically solid studies of the LS/CMI’s factor structure (e.g., Hsu et al., 2011), which also found numerous dimensions. The multidimensionality of recidivism risk was empirically demonstrated for persons convicted for a sexual offense (Brouillette-Alarie et al., 2016, 2018; Brouillette-Alarie & Hanson, 2015; Brouillette-Alarie & Proulx, 2019), and is implicit in the multiplicity of LS dimensions and the theoretical underpinnings of the psychology of criminal conduct (Andrews & Bonta, 2010). It may be that a higher-order “risk” construct encompasses the eight subscales of the LS/CMI, akin to the
Limitations
Even though the current study aimed to be thorough in its methods and prudent in its conclusions, it is not without limitations. First, the results reported here are limited to incarcerated men from Quebec’s provincial prison system, meaning that our sample, even though it was a population, is not necessarily representative of other populations. Specifically, it may not be generalizable to (a) potentially higher-risk persons from Canadian federal prisons that have received a sentence of 2 years or more; (b) individuals who have received a community sentence or are on parole; (c) women; (d) ethnoculturally diverse correctional populations; and (e) individuals involved in the US correctional system—especially in light of the lackluster predictive validity found in evaluation studies of the LS/CMI in the US.
Second, as mentioned above, the potential multidimensionality of the LS/CMI, as well as its high number of items, may have curtailed the usefulness of Rasch modeling in this study. However, rather than sweeping the issue under the rug, we thoroughly discussed it by plotting the difficulty of items by their predictive validity. As such, results and discussions from the current study may be of use to future scholars who try to apply IRT techniques to criminological risk scales. Third, the local independence assumption could not be fully met as removing or merging collinear items from the LS/CMI would have resulted in merging nearly half of the items, which would have deprived readers of valuable data on LS/CMI items. In future studies, especially for analyses highly sensitive to collinearity (e.g., factor analysis), more thorough item preparation may be necessary. Fourth, the follow-up period of the current study was limited to 2 years, which may leave “little time” for individuals to reoffend, especially for individuals who are on the lower end of the risk spectrum. Finally, although we do not anticipate that such a limitation would have significant effects on our results, it is worthwhile to mention that the conclusions of the current study are based on the French version of the LS/CMI and may thus not be applicable to the English version of the scale.
Implications for Research and Practice
Taken together, results of the current study attest to the use of the LS/CMI to assess the risk of general recidivism in incarcerated men from the Quebec population. The predictive validity of LS/CMI items, dimensions, and total scores was very good and among the best of what the field has to offer. Even under Rasch scrutiny, items performed relatively well and covered the whole spectrum of the risk continuum (difficulties ranging from −2.94 to 2.66). Apart from items that suffered from sample selection, few items had lackluster effect sizes toward general recidivism.
The current study offers some interesting avenues for future developments of the LS/CMI in relation to data that was obtained concerning its items. First, the substantial variation in the predictive validity and difficulty of items challenges the equal weight (one point) attributed to all LS/CMI items. However, contrary to our initial hypotheses, item difficulty was thoroughly unrelated to predictive validity. As such, results from the current study indicate that the predictive validity of items may be a better basis upon which to re-weight items than Rasch difficulty. Second, should the lack of predictive validity of items related to educational level be replicated in other samples, it might be worthwhile to rethink their presence in the scale, especially in light of meta-analyses that challenged the association between educational level and delinquency (Goodley et al., 2022). Third, predictive validity analyses revealed that some items (7 and 8) were particularly indicative of recidivism potential. Because these items refer, respectively, to institutional misconduct and parole breach, correctional staff should be particularly wary of these behaviors as they may indicate relapse. Even though the LS/CMI does not explicitly integrate the stable/acute distinction, we anticipate that Items 7 and 8 would be prime candidates for acute risk factors of general recidivism. Fourth, the factor analysis that was made to ensure “sufficient unidimensionality” challenged the unidimensionality of the LS/CMI put forward by many authors (see Hsu et al., 2011). We think that robust factor analytic studies of such a widely used scale are overdue and could enable a comparison between the eight theoretical dimensions of the scale and its empirical latent structure.
Finally, as it relates to fundamental research, the current study highlighted some limitations of applying Rasch models from the field of education to risk assessment scales from the field of criminology. Rasch modeling has shown that items deemed difficult to endorse were not necessarily risk relevant and did not systematically show good predictive validity. Even though this could damper the enthusiasm of researchers toward IRT-based models, IRT does offer some underexplored options for the psychometric validation of risk scales. Namely, IRT techniques could enable the investigation of the usefulness of the discrimination parameter (instead of difficulty) to improve predictive validity. In addition, using differential item functioning could help to determine if items perform equally well for different groups of convicted persons (e.g., men vs. women). Importantly, this technique could elucidate other important psychometric properties of risk scales for Indigenous populations, which have been the focus of increased legal challenge and scrutiny in recent years (Gutierrez et al., 2017; Huang et al., 2021).
Supplemental Material
sj-docx-1-cjb-10.1177_00938548221131956 – Supplemental material for A Look at the Difficulty and Predictive Validity of LS/CMI Items With Rasch Modeling
Supplemental material, sj-docx-1-cjb-10.1177_00938548221131956 for A Look at the Difficulty and Predictive Validity of LS/CMI Items With Rasch Modeling by Guy Giguère, Sébastien Brouillette-Alarie and Christian Bourassa in Criminal Justice and Behavior
Footnotes
References
Supplementary Material
Please find the following supplemental material available below.
For Open Access articles published under a Creative Commons License, all supplemental material carries the same license as the article it is associated with.
For non-Open Access articles published, all supplemental material carries a non-exclusive license, and permission requests for re-use of supplemental material or any part of supplemental material shall be sent directly to the copyright owner as specified in the copyright notice associated with the article.
