PLoS Med. 2016 (May 17); 13 (5): e1002019 ~ FULL TEXT
Adrian C. Traeger , Nicholas Henschke, Markus Hübscher , Christopher M. Williams,
Steven J. Kamper, Christopher G. Maher, G. Lorimer Moseley, James H. McAuley
Prince of Wales Clinical School,
University of New South Wales,
Sydney, New South Wales, Australia.
BACKGROUND: Low back pain (LBP) is a major health problem. Globally it is responsible for the most years lived with disability. The most problematic type of LBP is chronic LBP (pain lasting longer than 3 mo); it has a poor prognosis and is costly, and interventions are only moderately effective. Targeting interventions according to risk profile is a promising approach to prevent the onset of chronic LBP. Developing accurate prognostic models is the first step. No validated prognostic models are available to accurately predict the onset of chronic LBP. The primary aim of this study was to develop and validate a prognostic model to estimate the risk of chronic LBP.
METHODS AND FINDINGS: We used the PROGRESS framework to specify a priori methods, which we published in a study protocol. Data from 2,758 patients with acute LBP attending primary care in Australia between 5 November 2003 and 15 July 2005 (development sample, n = 1,230) and between 10 November 2009 and 5 February 2013 (external validation sample, n = 1,528) were used to develop and externally validate the model. The primary outcome was chronic LBP (ongoing pain at 3 mo). In all, 30% of the development sample and 19% of the external validation sample developed chronic LBP. In the external validation sample, the primary model (PICKUP) discriminated between those who did and did not develop chronic LBP with acceptable performance (area under the receiver operating characteristic curve 0.66 [95% CI 0.63 to 0.69]). Although model calibration was also acceptable in the external validation sample (intercept = -0.55, slope = 0.89), some miscalibration was observed for high-risk groups. The decision curve analysis estimated that, if decisions to recommend further intervention were based on risk scores, screening could lead to a net reduction of 40 unnecessary interventions for every 100 patients presenting to primary care compared to a "treat all" approach. Limitations of the method include the model being restricted to using prognostic factors measured in existing studies and using stepwise methods to specify the model. Limitations of the model include modest discrimination performance. The model also requires recalibration for local settings.
CONCLUSIONS: Based on its performance in these cohorts, this five-item prognostic model for patients with acute LBP may be a useful tool for estimating risk of chronic LBP. Further validation is required to determine whether screening with this model leads to a net reduction in unnecessary interventions provided to low-risk patients.
From the FULL TEXT Article:
Why Was This Study Done?
A minority of patients who experience an episode of low back pain
develop persistent (chronic) pain.
Offering tests and treatments to all these patients exposes high
numbers of low-risk patients to unnecessary intervention,
which is very costly and potentially harmful.
A tool to help healthcare practitioners accurately predict whether
a patient with a recent episode of low back pain will develop
persistent pain stands to greatly reduce the burden of low
back pain on the health system and on patients.
What Did the Researchers Do and Find?
We developed a five-item screening questionnaire using study data
from 1,230 patients with a recent episode of low back pain.
We tested how well this screening questionnaire could predict the
onset of persistent pain in a separate sample of 1,528 patients.
We found that the screening questionnaire could predict the onset of
persistent pain with acceptable levels accuracy (area under the
receiver operating characteristic curve = 0.66 [95% CI 0.63 to 0.69];
intercept = 0.55, slope = 0.89).
What Do These Findings Mean?
This brief, easy-to-use screening questionnaire could help healthcare
practitioners and researchers make an early estimate of a patient’s
risk of persistent low back pain.
The screening questionnaire predicted outcome more accurately in patients
with low risk scores than in those with high risk scores.
Screening patients with a recent episode of low back pain could reduce
the number of unnecessary interventions provided to low-risk patients.
Low back pain (LBP) is a major global health problem that, compared to all other diseases and health conditions, is responsible for the most years lived with disability, an estimated 80 million years lived with disability in 2010.  The costs of care, investigations, and lost productivity associated with LBP are a significant economic burden for industrialized nations.  For example, estimates for treatments alone are US$50 billion per annum in the United States  and US$4 billion in the United Kingdom.  The impact of LBP can be profound; in Australia, LBP is the leading cause of early retirement  and of income poverty in older adults. 
Although most people with a new episode, or acute, LBP recover in a few weeks or months, around one-quarter of patients who present to primary care develop chronic LBP (pain lasting for longer than 3 mo).  Chronic LBP is the most problematic type of LBP; its prognosis is poor , and it accounts for the majority of costs. [4, 9] Between 1992 and 2006, the prevalence of chronic LBP in the United States more than doubled.  Managing patients with chronic LBP is difficult, and the effects of contemporary interventions are modest at best.  An alternative to costly and ineffective management of these patients is secondary prevention, where the goal is to prevent the onset of chronic LBP. 
An important first step in secondary prevention is to estimate an individual patient’s risk of developing chronic LBP. The Prognosis Research Strategy (PROGRESS) group recently provided a framework for this step, which involves developing and validating prognostic models to determine risk profiles. For these models to be considered clinically useful, they must be easy to use, be able predict outcome with acceptable accuracy, and be validated in external samples.
Risk estimates should be well matched to actual outcomes (calibration), higher for individuals who have a poor outcome than for those who do not (discrimination), and informative enough to justify screening compared to “treat all” or “treat none” approaches (net benefit). Estimates from validated models can add valuable information to the clinical decision-making process. 
Early, accurate prognostic information also provides the opportunity for practitioners to counsel their patients on the necessity of further treatment.  Offering tests and treatments to all patients with acute LBP (“treat all” approach) is expensive and risks exposing high numbers of low-risk patients to unnecessary intervention.  Overtreatment of conditions such as LBP overburdens healthcare systems and diverts scarce resources away from where they are most needed.  Undertreatment of high-risk patients with acute LBP may also be harmful. A “treat none” approach to acute LBP guarantees that a significant proportion will develop chronic LBP and its long-term consequences , and wastes an opportunity to intervene early in primary care.
Targeting early intervention according to risk profile has been shown to be effective in breast cancer  and cardiovascular disease  and has been identified as a research priority for managing LBP.  There is preliminary evidence that a stratified approach improves disability in samples with predominantly chronic LBP , but it remains unknown whether such a prognostic approach can prevent the onset of chronic LBP. The absence of a valid prognostic model to inform risk-stratified management of acute LBP is therefore an important area of uncertainty.  Not having a validated prognostic model for acute LBP is also problematic for secondary prevention trials that are designed to target pain [22, 23]; treat all approaches are unlikely to be efficient if the majority of included participants are at low risk of chronic LBP. [24, 25]
None of the commonly used screening tools in LBP are suited to this purpose. Tools such as the Start Back Tool (SBT) and the Orebro Musculoskeletal Pain Questionnaire (OMPQ) were either developed in samples that included patients with chronic LBP  or used to predict disability [26, 27] or return to work  outcomes. When these tools were subsequently tested in acute LBP samples, they predicted chronic LBP with modest accuracy at best. [27, 29–31]
The primary aim of this study was to develop and validate a prognostic model to identify risk of chronic LBP in patients with acute LBP. Specifically, we aimed to develop a model that can provide an estimate for an individual patient’s risk of chronic LBP with acceptable levels of accuracy (calibration, discrimination, and net benefit). A secondary aim was to determine whether prognostic models varied by how chronic LBP was defined. Specifically, we aimed to develop two additional prognostic models using outcomes of high pain and chronic disability.
The protocol for this study has been published. 
We used patient data from a prospective cohort study to develop the model (development sample) and patient data from a randomized trial to externally validate the model (external validation sample). Full details of these two studies have been published [23, 33], and their key differences are summarized in Table 1. Both studies were conducted in Sydney, Australia. In short, the cohort study recruited consecutive patients with acute LBP presenting to their primary care provider (general practitioner, physiotherapist, chiropractor) between 5 November 2003 and 15 July 2005. The randomized trial recruited consecutive patients with acute LBP presenting to their primary care practitioner between 10 November 2009 and 5 February 2013 to test the effect of paracetamol on recovery. There was no difference in treatment effects between groups. Both studies followed a published protocol [34, 35], and the trial was prospectively registered.
Predictors and Model Outcomes
Baseline data were available on 20 predictors in six broad groups of putative prognostic factors that have been identified in previous studies [36-38]: sociodemographic factors, general health, work factors, current LBP characteristics, past LBP history, and psychological factors. Primary care clinicians collected these data at the first consultation. A full list of individual candidate predictors is provided in Table 2.
To develop the primary model, PICKUP (Predicting the Inception of Chronic Pain), we defined the main outcome as whether or not patients had chronic LBP, that is, ongoing LBP 3 mo after the initial consultation. In the development study, pain intensity was measured with a six-point Likert scale.  We classified patients as having “chronic LBP” if they reported greater than “mild” (2 on the Likert scale) pain intensity at 3-mo follow-up and had no periods of recovery. 
To develop two secondary prognostic models (Models 2a and 2b), we used additional criteria to define chronic LBP. These secondary models allowed comparison of model performance to published models and to our primary prognostic model. Patients were classified as having “chronic LBP high pain” if they reported greater than “moderate” (3 on the Likert scale) pain intensity  at 3-mo follow-up (Model 2a). Patients were classified as having “chronic LBP disability” if they reported a score of 2 or more on a five-point Likert scale for disability  at 3-mo follow-up (Model 2b). Thresholds to define outcomes for all three models were determined a priori. 
In the external validation sample, pain and disability scores were converted from an 11-point scale used to measure pain intensity and a 24-item scale used to measure disability to the six-point and five-point scales, respectively, used in the development sample. Both of the original studies assessed 3-mo outcomes over the phone, an approach that yields comparable results to in person assessment on pain-related outcomes. 
The statistical analysis plan for this study was informed by recommendations from the PROGRESS group.  All preplanned analyses are outlined in our protocol published a priori. 
We planned a complete case analysis if less than 5% of predictor values were missing. If more than 5% of predictor values were missing, we planned to impute the missing values. Because PROGRESS does not recommend a complete case analysis, we performed a post hoc sensitivity analysis using the Expectation Maximization algorithm in SPSS to impute missing values and to test the robustness of our approach. We did not impute missing outcome values. 
To identify predictors in the development sample, we performed a forward stepwise logistic regression analysis. We set the significance level for variable selection at p < 0.10. To specify the model, age, sex, and duration of the pain episode were forced into the first block, and the remaining candidate predictors (Table 2) were selected using an automated stepwise procedure in the second block. Only those predictors identified in the second block using the stepwise procedure were included in the final models for external validation. We examined the linearity of continuous predictor variables using scatter plots and Box–Tidwell transformations. 
We assessed the predictive performance of the regression model by examining measures of discrimination, calibration, and overall performance. Discrimination refers to how likely the model is to allocate higher predicted risks to patients who develop chronic LBP during the study period and lower predicted risks to those who do not. We assessed discrimination by calculating the area under the receiver operating characteristic curve (AUC).  With this statistic, a value of 0.5 indicates that the model discriminates no better than chance and a value of 1 indicates that the model discriminates perfectly.  We further assessed discrimination by calculating the discrimination slope (the absolute difference in mean predicted risk in those who developed chronic LBP and those who did not)  and risk-stratified likelihood ratios.
Calibration refers to the agreement of predicted risks and actual outcomes. In both samples, we constructed calibration graphs that plotted predicted risks produced from the prognostic model versus observed proportions of chronic LBP in ten groups separated by decile of risk. We fitted a smoothed line to the calibration graph to calculate the calibration slope and intercept; values around 1 for the slope and 0 for the intercept represent correct calibration. 
Overall performance and model fit indices combine aspects of discrimination and calibration. We tested overall performance using the Nagelkerke R2 statistic and the Brier score. Nagelkerke R2 measures the additional variation in chronic LBP that is explained by the model compared to an intercept-only logistic model. A large difference in Nagelkerke R2 between the development and external validation samples indicates overfitting and poor generalizability.  The Brier score quantifies the average prediction error and ranges from 0 to 0.25; values close to 0 represent informative models, while values close to 0.25 represent non-informative models. 
Internal validation (development sample)
Prediction models tend to perform optimistically (i.e., overestimate performance) in the sample in which they are developed. To provide a robust estimate of model performance in the development sample, we bootstrapped all of the performance estimates according to Harrell et al.  Bootstrapping is the most efficient method of internally validating performance estimates in a development sample.  In brief, this procedure creates bootstrap samples by drawing random samples with replacement from the development sample (200 replications) and then tests model performance in the newly created sample. This allows performance estimates in the development sample to be adjusted for optimism. Although we planned to use SPSS to perform the bootstrap procedure , we found it to be easier to perform using R software with the syntax provided by Steyerberg. 
External validation (external validation sample)
To externally validate the model, we tested model predictions in the external validation sample and calculated the performance statistics described above. To update the model, we examined whether including a recently identified prognostic factor — sleep quality [50, 51] — added significantly (p < 0.10) to the model. All models were recalibrated according to the method of Steyerberg , which involved updating the logistic equation using the calibration slope and intercept obtained in the external validation procedure.
We assessed the potential clinical utility of the model by selecting cutoffs based on quartiles of predicted risk in the development sample. Predicted risk, or predicted probability, is calculated using the regression equation and produces a number between 0 and 1. For example, a predicted probability of 0.2 signifies a 20% (absolute) predicted risk of developing chronic LBP. Those in the highest quartile of predicted risk were classified as high risk, those in the middle two quartiles as medium risk, and those in the lowest quartile as low risk. Using these cutoffs, we calculated posterior probabilities and likelihood ratios with 95% confidence intervals.
To further explore clinical utility, we performed a decision curve analysis. This analysis allowed us to assess whether using a prognostic model to screen patients could be a superior decision-making approach to simply intervening with all patients (treat all approach) or intervening with none (treat none approach). The decision curve analysis calculates the net benefit of a particular decision-making approach across a range of risk thresholds where patients and their physicians might opt for further intervention. For example, a physician may decide to recommend further intervention (e.g., a course of physiotherapy) for cases with a greater than 30% risk of chronic LBP. A decision curve analysis estimates whether this approach would provide a net increase in the proportion of patients treated appropriately (i.e., patients with a poor prognosis are recommended further intervention, those with a good prognosis are not). Specifically, the net benefit is the difference in proportions of true positives and false positives when false positives are weighted by the odds of the selected cutoff (net benefit = TP–wFP/N, where TP is the number of true positive decisions, wFP is the number of false positive decisions × the odds of a given cutoff, and N is the total number of patients).
We also calculated the net number of unnecessary interventions avoided using the equation suggested by Vickers and Elkin :
(net benefit of the model – net benefit of treat all)/(threshold probability/
[1 – threshold probability]) × 100
This equation estimates the net number of unnecessary interventions that would be avoided if clinicians were to base their decision to recommend further intervention on predicted risks. For example, compared to treat all, what is the net number of unnecessary interventions that would be avoided if only those with 30% or higher risk of chronic LBP were recommended further intervention? We calculated, across a range of potential cutoff scores, the net reduction in the number of patients with good outcomes who would receive unnecessary interventions using a treat all strategy.
Statistical analyses were carried out in SPSS Statistics for Windows version 22.0 (IBM Corp) and R version 3.1.2 .
Both of the original studies were approved by the Human Research Ethics Committee of the University of Sydney (ref 11-2002/3/3144 and ref 11638). All participants provided written informed consent. Because we analyzed an existing non-identifiable dataset, the Human Research Ethics Committee did not require a separate ethics application for the current study. A non-identifiable dataset is provided in S1 and S2 Data.
Flow of patients in the development and external validation samples is shown in Figure 1. Eighteen patients (1.4%) in the cohort study (development sample) and 46 patients (2.7%) in the randomized trial (external validation sample) were un-contactable at 3-mo follow-up. Some patients were excluded from the external validation sample because they were not assessed for pain intensity (65 patients; 3.9%) or disability (87 patients; 5.2%) at 3-mo follow-up.
There were five missing predictor values in the development sample and 44 missing predictor values in the external validation sample. We found evidence against the hypothesis that predictor values were not missing completely at random (Little’s test, p > 0.05), and, because the number of missing values was small (<1%), we removed these cases from the primary analysis as per our protocol.  Imputing missing predictor values in the sensitivity analysis did not affect the results (S1 and S2 Tables).
Data were therefore available from 1,230 cases to develop the prognostic models. To externally validate the models, data were available from 1,528 complete cases to test PICKUP, 1,525 complete cases to test Model 2a, and 1,504 complete cases to test Model 2b.
Table 3 shows the characteristics of patients in the development and external validation samples. Patients were similar at baseline except for the proportion receiving disability compensation, which was higher in the development sample (18%) than in the external validation sample (7%).
Model Development and Internal Validation
At 3 mo, 30% of the patients in the development sample were classified as having chronic LBP. Table 4 shows predictors and regression coefficients for the primary model (PICKUP) and the two secondary models that were fitted in this sample. PICKUP contained five predictors. We did not detect significant non-linearity in any continuous predictor variables. Estimates for the predictive performance of each prognostic model in the development sample can be found in S2 Table. Recruitment setting (general practice, physiotherapy, chiropractic) did not affect performance estimates (S3 Table).
Table 5 summarizes the predictive performance of the prognostic models in the external sample.
At 3 mo, 19% of the patients in the external validation sample were classified as having chronic LBP. The Nagelkerke R2 value was 7.7%, compared to 10.9% in the development sample, and the Brier score was 0.15, indicating a similar overall model fit. S2 Table shows the full results of performance testing for each prognostic model in the development and external validation samples. Discrimination performance for PICKUP fell within our prespecified acceptable range: the AUC was 0.66 (95% CI 0.63 to 0.69), the likelihood ratio in the high-risk group was 2.99 (95% CI 2.81 to 3.18), and the 95% confidence intervals did not overlap with between risk groups (S4 Table).
higher in the development sample (18%) than in the external validation sample (7%).
All models showed some miscalibration in the external validation sample (Figure 2). PICKUP demonstrated the best calibration and fell within our prespecified acceptable range in the lower seven of the ten risk groups, that is, predictions were within 5% of actual proportions of chronic LBP. In all three models, calibration was better for the low-risk patients than it was for the high-risk patients. After recalibration, slope and intercept estimates for each model were close to 1 and 0, respectively, which indicates near perfect calibration (S1–S3 Figs). Updating PICKUP with an additional prognostic factor (sleep quality) did not add significantly to the model (p > 0.10).
Figure 3 shows the results of the decision curve analysis. Treat all strategies assume that if all patients are treated, none will develop an unfavorable outcome. This may or may not be a reasonable assumption in LBP. Although there are effective treatments for acute LBP , evidence-based interventions to prevent the onset of chronic LBP are not yet available. The assumed outcome from treating all patients with acute LBP is that all high-risk patients are offered further intervention that could reduce their risk of chronic LBP. The assumed outcome from treating no patients with acute LBP is that all high-risk patients will develop an unfavorable outcome. In our external validation cohort, for example, if no high-risk patients were offered further intervention, one in five would develop chronic LBP.
Treat all strategies demonstrated the highest net benefit at threshold probabilities between 0% and 10%. At thresholds above the population risk (incidence rates were 19% for chronic LBP, 10% for chronic LBP with high pain, and 14% for chronic LBP with disability), the net benefit of treating all became negative (Figure 3). The net benefit of treating none was always assumed to be zero.
All prognostic models showed equal or higher net benefit than the treat all and treat none strategies. Using PICKUP and a cutoff set at 19% (i.e., only patients with a predicted risk higher than the population risk of 19% are recommended further intervention), the net number of cases of chronic LBP that would be detected through screening, without any increase in the number of patients unnecessarily recommended further intervention, would be four in every 100 patients.
Figure 4 shows the estimated net number of unnecessary interventions avoided through screening. Using PICKUP and a cutoff set at 30% (i.e., only patients with a predicted risk of 30% or higher are recommended further intervention) would lead to a net reduction of around 40 unnecessary interventions per 100 patients.
Calculation of an Individual Predicted Risk
An individual score (ScoreCLBP) can be derived using the recalibrated logistic regression equation from PICKUP:
ScoreCLBP = –0.55 + 0.89*(–2.82 + [0.21*Pain + 0.44*Leg + 0.50*Comp + 0.06*Depress + 0.13*Risk])
where Pain = “How much low back pain have you had during the past week?”
1 = none, 2 = very mild, 3 = mild, 4 = moderate, 5 = severe, 6 = very severe
Leg = “Do you have leg pain?”
0 = no, 1 = yes
Comp = “Is your back pain compensable, e.g., through worker’s compensation or third party insurance?”
0 = no, 1 = yes
Depress = “How much have you been bothered by feeling depressed in the past week (0–10 scale)?”
0 = not at all, 10 = extremely
Risk = “In your view, how large is the risk that your current pain may become persistent (0–10 scale)?”
0 = none, 10 = extreme.
The predicted risk of developing chronic LBP (ProbCLBP) can then be calculated using the score and the following equation:
ProbCLBP = exp(ScoreCLBP)/(1 + exp[ScoreCLBP])
We have developed and tested the external validity of a prognostic model to identify the risk of chronic LBP in individuals with acute LBP. Values for discrimination and calibration fell within a prespecified  range of what we subjectively determined to be informative. Although the AUC values are modest (between 0.66 and 0.69), they suggest better predictive accuracy for pain outcomes than recently published values based on either clinician judgment alone (between 0.50 and 0.60)  or popular tools such as the SBT and OMPQ. [29-31] The results of our decision curve analysis indicate that, compared to treat all and treat none strategies, our model has the potential to substantially reduce harms associated with undertreating high-risk patients and overtreating low-risk patients with acute LBP.
The major strengths of this study are its preplanned methods, the use of large, high-quality datasets, and transparent reporting. To our knowledge, this is the largest “Type 3” study in LBP to have — in line with the PROGRESS initiative  — published a statistical analysis plan and reported results using the TRIPOD statement (see S1 TRIPOD Checklist). Type 3 studies build on foundational prognostic factor research (Type 1 and 2 studies)  by constructing prognostic models. Constructing accurate prognostic models is an essential step towards improving patient outcomes through stratified care (Type 4 studies).  We used large samples of patients with acute LBP to develop and externally validate the models. The samples had a number of differences (Table 1), not least of which was the overall risk of developing chronic LBP (30% in the development sample versus 19% in the external validation sample). Despite these differences, the models made informative predictions in the external sample, which indicates favorable generalizability and suggests that further testing in additional samples is warranted. We have reported different aspects of model performance that can be interpreted for clinical and research applications.
This study has some limitations. First, we were restricted to the use of predictor variables measured in the original studies. We were therefore not able to directly compare our model or update existing models in this study, as is recommended by the PROGRESS framework (Recommendation 21).  Moreover, we may not have included important prognostic variables in our models because they were not measured in the original studies. We attempted to overcome this limitation by updating the model at the external validation stage. Interestingly, when we updated the model with a recently identified prognostic factor, sleep quality [50, 51], there was no improvement in any of our indices of predictive performance. Second, we used an automated stepwise approach to specify the models, principally because it is objective and generally results in smaller, clinically applicable models , but stepwise methods have well-known limitations such as unstable variable selection  and biased coefficient estimation.  It is therefore conceivable that our choice to use stepwise selection may have reduced the predictive performance of the models. Third, the overall model fit statistics indicate that the variance explained by our prediction models is modest. Perhaps some factors that are yet to be tested thoroughly in LBP, for example, structural pathology shown on imaging , explain additional variance in chronic LBP. However, tests involving imaging are onerous, costly, and potentially harmful for patients with acute nonspecific LBP.  Fourth, by prespecifying in our protocol that we would impute missing predictor values only if they were missing in more than 5% of the sample, we did not strictly adhere to the PROGRESS recommendation to impute values where reasonable (Recommendation 20). The complete case approach that we used in our primary analysis can be inefficient and is known to produce bias in prediction research.  However, the number of missing predictor values was small (<2%), and our post hoc sensitivity analysis showed no major differences in results when a post hoc imputation procedure was performed (S1 and S2 Tables). This suggests that our a priori decision to remove cases with missing predictor values did not bias the results. Finally, because our prognostic model is in the form of a logistic regression equation, this limits its ease of use. To address this limitation, we developed a calculator (based on the recalibrated PICKUP) that is freely available online at http://pickuptool.neura.edu.au/.
Deciding whether a model is useful or not depends both on its performance and its purpose. In the research setting, discrimination is an important consideration. When such a large number of patients recover with minimal or no intervention, treat all approaches to preventing chronic LBP are inevitably going to be inefficient. Some treatments for LBP, if applied to low-risk patients, may even be harmful. Our models can help discriminate between patients who experience poor outcomes and patients who experience good outcomes, with acceptable performance (AUC > 0.6, likelihood ratios not overlapping). In the external validation sample, patients allocated to the high-risk group (i.e., in the highest quartile of predicted risk) were three times more likely to develop chronic LBP than their medium- or low-risk counterparts (in the middle two and lowest quartiles of predicted risk, respectively). Including only patients with a predicted risk above a 30% in a secondary prevention trial would lead to a net reduction of 40 unnecessary episodes of care (for patients with good outcomes) per 100 patients (Figure 4).
In the clinical setting, calibration is important for providing accurate risk estimates to patients. Our primary prognostic model (PICKUP) demonstrated acceptable calibration (<5% difference between predicted risks and observed proportions of chronic LBP) in seven out of ten risk strata. However, we did observe some miscalibration in the higher risk strata—as predicted risk increased, accuracy decreased and the model overestimated risk (Figure 2). This, along with our negative predictive values above 90% (S4 Table), means that people with lower risk estimates are very unlikely to develop chronic pain, but those with higher risk estimates may still recover quickly. That is, the models are better at ruling out future chronic LBP than ruling it in. However, after recalibration the estimates were almost perfectly calibrated (S1–S3 Figs). With further testing and recalibration, these models have potential to be useful in other clinical settings.
Our decision curve analysis suggested that the primary model is likely to be useful for patients whose decision to pursue further intervention is based on a predicted risk between 12% and 35%. The question that remains is whether these thresholds are clinically relevant. For a range of thresholds under 50% to be considered clinically relevant, the assumption is that patients place more value on detecting an imminent problem (true positive rate) than undergoing unnecessary treatment (false positive rate).  We would suggest that most patients with acute LBP would fall into this category: the consequences of undergoing, for example, an unnecessary course of physiotherapy, are outweighed by the prospect of missing a chance at preventing a long-term problem. However, this assumption rests on the nature of the treatment proposed. If the patient and their physician are considering invasive treatments such as spinal surgery, the patient might weigh the false positive rate more heavily, due to the higher risk of adverse events. In this case, a screening tool would need to yield a net benefit across a range of predicted risk cutoffs higher than 50%, and our model would not be considered useful.  We therefore speculate that our models are likely to be useful only for informing the choice between a wait-and-see approach and a course of conservative intervention.
Although several models have been developed in LBP, few have been externally validated , and none have been designed to predict the onset of chronic LBP. Pain is arguably the most important outcome to predict in LBP; it is clearly the most important issue for patients , and it is the slowest to recover.  The three available tools that have been tested in external samples of patients with acute LBP appear to predict pain outcomes at 3 and 6 mo with modest accuracy at best. Grotle et al.  tested the OMPQ in an acute LBP sample and reported an AUC for predicting pain at 6 mo of 0.62 (95% CI 0.51 to 0.73). Recent evaluations of SBT score in predicting ongoing pain at 6 mo in acute LBP samples reported AUC values of 0.50  and 0.54.  Williams et al.  reported an AUC of 0.60 (95% CI 0.56 to 0.64) for predicting recovery from pain (0 or 1/10 pain sustained for 7 d) at 3 mo. PICKUP appears to discriminate medium-term pain outcomes in patients with acute LBP more accurately than other validated models, and may be particularly useful for secondary prevention trials that target pain reduction. Because calibration performance has not been widely reported, we were unable to compare our model to others in these terms. Williams et al.  reported acceptable calibration for their model predicting outcomes in the first 2 wk but relatively poor calibration (more than 10% difference between predicted risks and observed proportions) for predicting pain outcomes at 3 mo. As suggested by PROGRESS, a formal comparison of our tool with other validated tools, for example, using a decision curve analysis, is a logical next step.
Based on its performance in these cohorts, this five-item prognostic model for patients with acute LBP may be a useful tool for estimating risk of chronic LBP. Further validation is required to determine whether screening with this model leads to a net reduction in unnecessary interventions provided to low-risk patients.
Model specification results using two missing data strategies
Compares analysis where cases missing predictor variables were removed to analysis where predictor values were imputed.
Comprehensive model performance results
Reports all performance indices measured in the development and validation samples. Prespecified “acceptable” levels were published in our protocol .
PICKUP performance in different clinical settings (development sample)
Sensitivity analysis examining model performance in patients seen in physiotherapy, general practice, and chiropractic settings.
Likelihood ratios and posterior probability estimates in external validation sample
Additional measures of predictive performance in the external validation sample.
S1 TRIPOD Checklist.
Adherence to TRIPOD reporting criteria for studies of prediction model development and validation
The authors express their gratitude to the study participants and the principal investigators involved in the trials whose work made this study possible.
AUC = area under the receiver operating characteristic curve
LBP = low back pain
OMPQ = Orebro Musculoskeletal Pain Questionnaire
PROGRESS = Prognosis Research Strategy
SBT = Start Back Tool
Global Burden of Disease Study 2013 Collaborators.
Global, regional, and national incidence, prevalence, and years lived with disability for
301 acute and chronic diseases and injuries in 188 countries, 1990–2013:
a systematic analysis for the Global Burden of Disease Study 2013.
Martin BI, Turner JA, Mirza SK, Lee MJ, Comstock BA, Deyo RA.
Trends in Health Care Expenditures, Utilization, and Health Status Among
US Adults With Spine Problems, 1997-2006
Spine (Phila Pa 1976). 2009 (Sep 1); 34 (19): 2077–2084
Deyo RA, Weinstein JN.
Low back pain.
N Engl J Med. 2001;344:363–370.
Hong J, Reed C, Novick D, Happich M.
Costs associated with treatment of chronic low back pain:
an analysis of the UK General Practice Research Database.
Spine (Phila Pa 1976). 2013;38:75–82
Schofield DJ, Shrestha RN, Passey ME, Earnest A, Fletcher SL.
Chronic disease and labour force participation among older Australians.
Med J Aust. 2008;189:447
Schofield DJ, Callander EJ, Shrestha RN, Percival R, Kelly SJ.
Labor force participation and the influence of having back problems on
income poverty in Australia.
Spine (Phila Pa 1976). 2012;37:1156–1163
Chou R, Shekelle P.
Will this patient develop persistent disabling low back pain?
Menezes Costa LC, Maher CG, Hancock MJ, McAuley JH.
The prognosis of acute and persistent low-back pain: a meta-analysis.
Gore M, Sadosky A, Stacey BR, Tai KS, Leslie D.
The Burden of Chronic Low Back Pain: Clinical Comorbidities, Treatment Patterns,
and Health Care Costs in Usual Care Settings
Spine (Phila Pa 1976). 2012 (May 15); 37 (11): E668–677
Freburger JK, Holmes GM, Agans RP, Jackman AM, et al.
The rising prevalence of chronic low back pain.
Arch Intern Med. 2009;169:251–258
Kamper SJ, Apeldoorn A, Chiarotto A, Smeets R, et al.
Multidisciplinary biopsychosocial rehabilitation for chronic low back pain:
Cochrane systematic review and meta-analysis.
Early identification and intervention in the prevention of musculoskeletal pain.
Am J Ind Med. 2002;41:433–442
Steyerberg EW, Moons KGM, van der Windt DA, Hayden JA, et al.
Prognosis Research Strategy (PROGRESS) 3: prognostic model research.
PLoS Med. 2013;10:e100138
McCaffery KJ, Jansen J, Scherer LD, Thornton H, et al.
Walking the tightrope: communicating overdiagnosis in modern healthcare.
Glasziou P, Moynihan R, Richards T, Godlee F.
Too much medicine; too little care.
Berwick DM, Hackbarth AD.
Eliminating waste in US health care.
Trastuzumab—mechanism of action and use in clinical practice.
N Engl J Med. 2007;357:39–51
Hingorani AD, Hemingway H.
How should we balance individual and population benefits of statins for preventing
Koes BW, van Tulder MW, Thomas S.
Diagnosis and Treatment of Low Back Pain
British Medical Journal 2006 (Jun 17); 332 (7555): 1430–1434
Hill JC, Whitehurst DG, Lewis M, Bryan S, Dunn KM, Foster NE, et al.
Comparison of Stratified Primary Care Management For Low Back Pain With
Current Best Practice (STarT Back): A Randomised Controlled Trial
Lancet. 2011 (Oct 29); 378 (9802): 1560–1571
Haskins R, Osmotherly PG, Rivett DA.
Validation and impact analysis of prognostic clinical prediction rules for low back
pain is needed: a systematic review.
J Clin Epidemiol. 2015;68:821–832
Traeger AC, Moseley GL, Hubscher M, Lee H, et al.
Pain education to prevent chronic low back pain:
a study protocol for a randomised controlled trial.
BMJ Open. 2014;4:e005505
Williams CM, Maher CG, Latimer J, McLachlan AJ, et al.
Efficacy of paracetamol for acute low-back pain:
a double-blind, randomised controlled trial.
Lancet. 2014;384:1586–1596. doi: 10.1016/s0140-6736(14)60805-9
Vickers AJ, Kramer BS, Baker SG.
Selecting patients for randomized trials: a systematic approach based on risk group.
Trusheim MR, Berndt ER, Douglas FL.
Stratified medicine: strategic and economic implications of combining drugs
and clinical biomarkers.
Nat Rev Drug Discov. 2007;6:287–293
Hill JC, Dunn KM, Lewis M, et al.
A Primary Care Back Pain Screening Tool:
Identifying Patient Subgroups For Initial Treatment
(The STarT Back Screening Tool)
Arthritis and Rheumatism 2008 (May 15); 59 (5): 632–641
Beneciuk JM, Bishop MD, Fritz JM, Robinson ME, et al.
The STarT back screening tool and individual psychological measures:
evaluation of prognostic capabilities for low back pain clinical outcomes
in outpatient physical therapy settings.
Phys Ther. 2013;93:321–333
Hockings RL, McAuley JH, Maher CG.
A systematic review of the predictive ability of the Orebro Musculoskeletal Pain Questionnaire.
Spine (Phila Pa 1976). 2008;33:E494–E500
Kongsted A, Andersen CH, Hansen MM, Hestbaek L.
Prediction of Outcome in Patients with Low Back Pain--A Prospective Cohort Study
Comparing Clinicians' Predictions with those of the Start Back Tool
Man Ther. 2016 (Feb); 21: 120–127
Mehling WE, Avins AL, Acree MC, Carey TS, Hecht FM.
Can a back pain screening tool help classify patients with acute pain into risk levels
for chronic pain?
Eur J Pain. 2015;19:439–446
Grotle M, Vollestad NK, Brox JI.
Screening for yellow flags in first-time acute low back pain: reliability and validity
of a Norwegian version of the Acute Low Back Pain Screening Questionnaire.
Clin J Pain. 2006;22:458–467
Traeger A, Henschke N, Hubscher M, Williams CM, et al.
Development and validation of a screening tool to predict the risk of chronic low
back pain in patients presenting with acute low back pain: a study protocol.
BMJ open. 2015;5:e007916
Henschke N, Maher CG, Refshauge KM et al.
Prognosis in Patients with Recent Onset Low Back Pain in Australian Primary Care:
Inception Cohort Study
British Medical Journal 2008 (Jul 7); 337: a171
Williams CM, Maher CG, Latimer J, McLachlan AJ, et al.
PACE—the first placebo controlled trial of paracetamol for acute low back pain:
statistical analysis plan.
Henschke N, Maher CG, Refshauge KM, Herbert RD, et al.
Prognosis of acute low back pain: design of a prospective inception cohort study.
BMC Musculoskelet Disord. 2006;7:54
Steenstra I, Verbeek J, Heymans M, Bongers P.
Prognostic factors for duration of sick leave in patients sick listed with acute
low back pain: a systematic review of the literature.
Occup Environ Med. 2005;62:851–860
Bekkering GE, Hendriks HJ, van Tulder MW, Knol DL, et al.
Prognostic factors for low back pain in patients referred for physiotherapy:
comparing outcomes and varying modeling techniques.
Spine (Phila Pa 1976). 2005;30:1881–1886
Pincus T, Burton AK, Vogel S, Field AP.
A systematic review of psychological factors as predictors of chronicity/disability
in prospective cohorts of low back pain.
Spine (Phila Pa 1976). 2002;27:E109–E120
Ware JE Jr, Sherbourne CD.
The MOS 36-item short-form health survey (SF-36). I. Conceptual framework and item selection.
Med Care. 1992;30:473–483
de Vet HC, Heymans MW, Dunn KM, Pope DP, van der Beek AJ, et al.
Episodes of low back pain: a proposal for uniform definitions to be used in research.
Spine (Phila Pa 1976). 2002;27:2409–2416
Moseley GL, Herbert RD, Parsons T, Lucas S, Van Hilten JJ, Marinus J.
Intense pain soon after wrist fracture strongly predicts who will develop complex
regional pain syndrome: prospective cohort study.
J Pain. 2014;15:16–23
Clinical prediction models: a practical approach to development, validation, and updating.
New York: Springer Science & Business Media; 2008
Box GE, Tidwell PW.
Transformation of the independent variables.
Becker A, Held H, Redaelli M, Strauch K, Chenot JF, et al.
Low back pain in primary care: costs of care and prediction of future health care utilization.
Spine (Phila Pa 1976). 2010;35:1714–1720
Harrell FE Jr, Califf RM, Pryor DB, Lee KL, Rosati RA.
Evaluating the yield of medical tests.
Steyerberg EW, Vickers AJ, Cook NR, Gerds T, Gonen M, et al.
Assessing the performance of prediction models: a framework for some traditional
and novel measures.
Gerds TA, Cai T, Schumacher M.
The performance of risk prediction models.
Biom J. 2008;50:457–479
Harrell F, Lee KL, Mark DB.
Tutorial in biostatistics multivariable prognostic models: issues in developing models,
evaluating assumptions and adequacy, and measuring and reducing errors.
Stat Med. 1996;15:361–387
Steyerberg EW, Harrell FE, Borsboom GJ, Eijkemans M.
Internal validation of predictive models: efficiency of some procedures for logistic
J Clin Epidemiol. 2001;54:774–781
Alsaadi SM, McAuley JH, Hush JM, Lo S, Lin CWC, et al.
Poor sleep quality is strongly associated with subsequent pain intensity in patients
with acute low back pain.
Arthritis Rheum. 2014;66:1388–1394
Aili K, Nyman T, Hillert L, Svartengren M.
Sleep disturbances predict future sickness absence among individuals with lower back
or neck-shoulder pain: a 5-year prospective study.
Scand J Public Health. 2015;43:315–323
Vickers AJ, Elkin EB.
Decision curve analysis: a novel method for evaluating prediction models.
Med Decis Making. 2006;26:565–574
Team RC. R:
a language and environment for statistical computing.
Vienna: R Foundation for Statistical Computing; 2014
Koes BW, van Tulder M, Lin CW, Macedo LG, McAuley J, Maher C.
An Updated Overview of Clinical Guidelines for the Management of
Non-specific Low Back Pain in Primary Care
European Spine Journal 2010 (Dec); 19 (12): 2075–2094
Riley RD, Hayden JA, Steyerberg EW, Moons KGM, Abrams K, et al.
Prognosis Research Strategy (PROGRESS) 2: prognostic factor research.
PLoS Med. 2013;10:e1001380
Hingorani AD, Windt DA, Riley RD, Abrams K, et al.
Prognosis research strategy (PROGRESS) 4: stratified medicine research.
Steyerberg EW, Ewout WS.
Clinical prediction models: a practical approach to development,
validation, and updating.
Dordrecht: Springer; 2008
Austin PC, Tu JV.
Automated variable selection methods for logistic regression produced unstable
models for predicting acute myocardial infarction mortality.
J Clin Epidemiol. 2004;57:1138–1146
Hancock M, Maher C, Petocz P, Lin CC, Steffens D, et al.
Risk factors for a recurrence of low back pain.
Spine J. 2015;15:2360–2368
Chou R, Fu R, Carrino JA, Deyo RA.
Imaging strategies for low-back pain: systematic review and meta-analysis.
Janssen KJ, Donders ART, Harrell FE, Vergouwe Y, Chen Q, et al.
Missing covariate data in medical research: to impute is better than to ignore.
J Clin Epidemiol. 2010;63:721–727
Steyerberg EW, Vickers AJ.
Decision curve analysis: a discussion.
Med Decis Making. 2008;28:146
Hush JM, Refshauge K, Sullivan G, De Souza L, Maher CG, McAuley JH.
Recovery: what does this mean to patients with low back pain?
Arthritis Rheum. 2009;61:124–131
Williams CM, Hancock MJ, Maher CG, McAuley JH, Lin CWC, Latimer J.
Predicting rapid recovery from acute low back pain based on the intensity, duration
and history of pain: a validation study.
Eur J Pain. 2014;18:1182–1189
Return to LOW BACK PAIN