MANUAL EXAMINATION OF THE SPINE: A SYSTEMATIC CRITICAL LITERATURE REVIEW OF REPRODUCIBILITY
 
   

Manual Examination of the Spine: A Systematic
Critical Literature Review of Reproducibility

This section is compiled by Frank M. Painter, D.C.
Send all comments or additions to:
    Frankp@chiro.org
 
   

FROM:   J Manipulative Physiol Ther 2006 (Jul); 29 (6): 475–485 ~ FULL TEXT


Mette Jensen Stochkendahl, Henrik Wulff Christensen, Jan Hartvigsen, Werner Vach, Mitchell Haas, Lise Hestbaek, Alan Adams, Gert Bronfort

Nordic Institute of Chiropractic and Clinical Biomechanics Research Department,
Part of Clinical Locomotion Science,
Odense, Denmark



Objective:   Poor reproducibility of spinal palpation has been reported in previously published literature, and authors of recent reviews have posted criticism on study quality. This article critically analyzes the literature pertaining to the inter- and intraobserver reproducibility of spinal palpation to investigate the consistency of study results and assess the level of evidence for reproducibility.

Methods:   Systematic review and meta-analysis were performed on relevant literature published from 1965 to 2005, identified using the electronic databases MEDLINE, MANTIS, and CINAHL and checking of reference lists. Descriptive data from included articles were extracted independently by 2 reviewers. A 6–point scale was constructed to assess the methodological quality of original studies. A meta-analysis was conducted among the high-quality studies to investigate the consistency of data, separately on motion palpation, static palpation, osseous pain, soft tissue pain, soft tissue changes, and global assessment. A standardized method was used to determine the level of evidence.

Results:   The quality score of 48 included studies ranged from 0% to 100%. There was strong evidence that the interobserver reproducibility of osseous and soft tissue pain is clinically acceptable (kappa > or = 0.4) and that intraobserver reproducibility of soft tissue pain and global assessment are clinically acceptable. Other spinal procedures are either not reproducible or the evidence is conflicting or preliminary.

Key Indexing Terms   Reproducibility of Results, Palpation, Literature Review, Diagnostic Tests, Spine, Meta-Analysis



From the FULL TEXT Article:

Background

Biomechanical dysfunction is thought to be an important contributor to spinal pain, and manual palpation is a widely used procedure for the diagnosis of such dysfunctions among providers of manual medicine. [1–3] Contrary to the expectations of many clinicians, unacceptable levels of reproducibility have been shown in the majority of the previously published literature, and authors of newer reviews have questioned the utility of manual examination procedures in spinal diagnosis altogether. [4–7]

Severe criticism has been posted on the design of the original studies, including

the use of asymptomatic subjects, [4, 5]

inexperienced observers, [5]

parallel testing, [4]

unclear definitions of positive findings and rating scales, [4, 6]

weak description of study results, [4, 5, 7] and

the need for improvement in overall study quality. [4, 7]

Furthermore, the dependence of Cohen's κ (the most widely used statistical method in studies on reproducibility) on the prevalence of positive findings, and the composition of the study population has been the subject of discussion. [8, 9]

Unfortunately, these reviews themselves have important limitations. For instance, some deal with only a minority of manual examination procedures such as chiropractic procedures only, [4] 1 spinal region, [4, 6, 10] or motion palpation only. [5] In only 3 reviews were a predefined quality system applied to assess study quality, [4, 6, 7] and in none of the reviews were both the number of studies, the methodological quality, and the consistency of the outcomes considered, as recommended by van Tulder and others. [11–13] Finally, in none of these reviews was the impact of the predefined criteria on the conclusions tested. Therefore, the value of palpation as a diagnostic tool is, at present, still unknown and so are the abilities of practitioners of manual therapy to reliably diagnose spinal dysfunctions using palpation.

We therefore decided that another systematic review taking into account the above issues was warranted. Furthermore, a meta-analysis including comparable studies of adequate methodological standard and assessment of the consistency of study outcomes would be highly useful. The purpose of this paper is therefore to systematically review and critically assess the design and statistical methodology of the literature pertaining to reproducibility of spinal palpation adopting standardized criteria for judging diagnostic studies. A meta-analysis was conducted to evaluate consistency of study outcomes. Finally, the level of evidence for the reproducibility of spinal palpation was determined.



Discussion

      Summary of Results

After reviewing studies dealing with reproducibility of manual palpation of the entire spine, including the SI joints, we found strong evidence for clinically acceptable reproducibility both within and between observers for palpation of osseous and soft tissue pain (STP) and within the same observer for global assessment (GA). Strong evidence for clinically unacceptable levels of reproducibility for intra- and interobserver global assessment, motion palpation (MP) and soft tissue changes (STC) was found. Intraobserver reproducibility was consistently higher than interobserver reproducibility, and reproducibility of palpation for pain response was consistently higher than reproducibility of palpation for motion.

The most recent and comprehensive review evaluating the reproducibility of spinal palpation by Seffinger et al [7] applied different inclusion and general review criteria, and thus, only 27 of 44 articles and 9 of 19 high-quality articles included in this review were evaluated. Furthermore, we included several more recent publications and articles dealing with the SI joints, GA, and evaluated single results from multiple test regimens. Our conclusions are based on predefined criteria and an evaluation of consistency of high-quality studies, a method not previously applied, whereas the conclusions by Seffinger et al [7] were based on both high- and low-quality studies without an evaluation of consistency. The authors concluded that pain provocation tests are most reliable, and soft tissue paraspinal palpatory diagnostic test is not reliable. Among the 12 highest-quality articles, pain provocation, motion, and landmark location tests were reliable within the same observer, but not always among observers under similar conditions. Overall, examiner' discipline, experience level, consensus on procedures used, training, or the use of symptomatic subjects did not improve reliability. This is in agreement with our findings. Furthermore, we conclude that palpation of pain is reproducible both within and among observers, whereas MP may be reproducible within the same observer.

      Methodological and Clinical Considerations

The experimental design of reproducibility studies has been criticized in previous reviews, [4–7, 68–71] and we found that 26 of 48 articles were of low methodological quality, had invalid statistical methods, or insufficient reporting of palpation procedures or test results.

Comparability of the studies included in a review is the important requirement to ensure valid generalizations. We ensured comparability with respect to the palpation procedures used, but the studies were rather heterogeneous with respect to characteristics such as definition of positive findings, segmental region, standardization, occupation, experience, symptomatic status of test population, and parallel testing. However, our investigation showed that most study characteristics had little influence on the study results, with the exception of the application condition. Especially, standing palpation was associated with very low κ values. Among the reviewed studies, standing palpation is used solely in the “Gillet test” of SI biomechanical dysfunction, and only 2 studies reporting this condition were included in our analysis. [39, 59] However, both contributed to the evaluation of the inter- and intraobserver agreement of MP. If we remove these 2 studies, then the average κ for the interobserver agreement increases to 0.19 (0.13–0.26), and the intraobserver agreement increases to 0.44 (0.14–0.73), such that the intraobserver agreement of MP can be regarded as acceptable.

Poor reproducibility of MP may reflect the design of reproducibility studies, rather than the quality of the palpation procedure. [29, 30, 72] Greater reproducibility may be attained by allowing positive findings in a neighboring spinal segment to count in assessing agreement. [29] However, this implies that we define a new, different diagnostic test which, then, requires a clinical rationale of test meaningfulness, beyond just an increase in κ values. [8] Further, parallel testing (test regimens) seems to aid the observer in making the clinical decision, thus enhancing reproducibility; [30, 42] a tendency we could also observe in our data. The acceptable intraobserver reproducibility for GA is also in line with this finding. However, when evaluating a combination of tests, information is only given about the reproducibility of the single test as part of this exact combination of tests. [14, 73] Moreover, we must be aware that conclusions on a single test from a study involving several tests may be only valid if the test is applied as part of this exact combination of tests. From a clinical perspective, increased reproducibility with parallel testing indicates that at this point, clinicians should not base their diagnosis on a single clinical examination finding such as palpation but, rather, conduct a range of tests. It is, however, premature to make clinical guidelines on how to use palpation because many aspects of palpation, such as the validity, still need to be investigated.

The reproducibility of palpation for pain response is consistently higher than palpation for motion and, consistently, substantially higher within an observer than among different observers. However, both palpatory pain studies and intraobserver studies in general have inherent problems with blinding of observers. In intraobserver studies, conscious and unconscious cues may render blinding of the observers impossible, and the independence of measures can not be guaranteed. In palpatory pain studies, blinding of subjects is impossible. Both situations imply the risk of overestimating reproducibility. It should also be noted that intraobserver reproducibility is somewhat higher than interobserver reproducibility by definition (depending on the magnitude of observer by subject interaction). [74]

A dilemma between high internal validity and clinical applicability arises when designing studies of reproducibility. For example, training studies contrast maximal (ideal) reproducibility with actual reproducibility in practice. To enhance the internal validity, rigid testing conditions should be set up with considerations to blinding, randomization, standardization and training, and parallel testing. However, rigid enforcement of testing condition often diverges from the clinical situation and, hence, may reduce the external validity. In a clinical situation, a mix of both asymptomatic and symptomatic patients will most likely present to practitioners of manual medicine. Therefore, the study population should consist of a mix of both symptomatic and asymptomatic subjects so that the reproducibility of the testing procedure has a relation to the characteristics of the study population. [14] Finally, in spite of the use in every day clinical routines, test procedures do not always necessarily evaluate the clinical entity it is intended to evaluate, and it is therefore important to discuss the content of the test procedure. [14, 75]

      Statistical Considerations

κ is widely accepted as the statistical method of choice for evaluating agreement between 2 observers for a binary classification. [8] It is, however, not without problems to use κ as the sole measure of observer agreement because information is lost when a 4–fold table is summarized into 1 number. Consequently, we do not know whether it is due to a difference in prevalence estimates between observers, or whether observers lack agreement in spite of similar prevalence if a moderate κ value is obtained in a study of reproducibility.

κ has been criticized for its dependence on the prevalence of positive findings, which limits its usefulness in meta-analyses, because studies with varying prevalence are typically compared. However, the composition of the study population may have greater impact on κ than the prevalence of positive findings. [9] Both a binary outcome and a reported κ value were required for studies to be part of our meta-analysis. However, binary outcomes may vary according to the definition of positive findings (ie, prevalence is directly dependent on the definition of positive findings). For example, if the observer is asked to identify any hypomobile segment(s) in a spinal region, the prevalence can vary from 0% to 100%, depending on the study population. If the observer is to identify the most hypomobile segment, the overall prevalence of positive findings will be 100%, but at any particular segment under investigation, the prevalence of the most hypomobile can be 0% to 100%. However, we found no association between the prevalence of positive findings and κ values. This supports that the composition of the study populations is probably of greater importance than the prevalence of positive findings, as suggested by Vach. [9]

Different words and schemes have been used to evaluate the strength of reproducibility, but there are no definitive guidelines for interpreting good concordance. [8, 76] Moreover, little research has been done to establish minimal, clinically acceptable reproducibility, and perhaps more important than qualifying the strength of concordance, the quantitative reproducibility indices need to be evaluated in terms of their clinical application. [8]

      Limitations of this Review

Different methodologies have been advocated for systematic reviews of trials addressing therapeutic efficacy, [12] but little consensus exists when it comes to assessing the quality of reproducibility studies. We have chosen to evaluate the strength of evidence based on a best-evidence synthesis method, and this is one of the main differences between this review and previously published reviews on the same topic. Heterogeneity across studies, in terms of test procedures, inclusion criteria, study design and presentation of results, may be masked by the best-evidence approach. Considerable heterogeneity in study characteristics was noted across studies included in this review. However, despite this heterogeneity, the meta-analysis showed very consistent overall findings and only moderate impact of the specific design characteristics on the study outcomes.

The exclusion from the meta-analysis of studies that did not report a binary outcome is another important difference between this and previous reviews. To compare studies of reproducibility, the same type of outcome and method of statistics must be applied. On this account, we had to exclude 5 high-quality studies from the meta-analysis. Results from these studies are not directly comparable to the included studies, but all 5 articles show results with similar trends of low interobserver agreement on MP and higher interobserver agreement on evaluation of pain; they were included in the level of evidence assessment. The restricted number of articles causes the strength of evidence to be preliminary or nonexistent in 3 categories. In return, the power of the conclusions with respect to pain and motion testing is compelling. However, results were, in some categories, based on a relatively small number of original studies, making the conclusions very sensitive to just a few future high-quality studies with different results.

A κ value was reported in all high-quality studies using a binary classification. Hence, there was no need to calculate these from a published 4–fold table. No attempts were made to retrieve additional, original results or materials from the primary authors.

Although every effort was made to find all published reproducibility studies, selection bias may have occurred because we included only English-language articles. Publication bias may have resulted in an overestimation of test reproducibility because studies arriving at positive conclusions are more likely to get published. [77, 78] Furthermore, reviewer bias is also a possible limitation of this review. Reviewers were not blinded to the authors or the results of the individual trials when the methodological scoring was performed because of our familiarity with the literature.

Despite acceptable study quality according to our criteria, many trials still had methodological limitations or, at best, inadequate reporting of methods. Nonetheless, reproducibility of spinal manual palpation has been very thoroughly investigated and more than 40 original articles have been evaluated in this review. However, to shed light on the clinical usefulness of palpation, the validity needs to be investigated, and new innovative research that addresses the concomitant problems of selecting a golden standard in motion testing is warranted. Future research should also address the question of palpation in the overall assessment of neck and back pain patients and the importance of palpation as part of the complete clinical evaluation of patients.



Conclusions

Palpation for pain is reproducible at a clinically acceptable level, both within the same observer and among observers. Palpation for global assessment (GA) is reproducible within the same observer but not among different observers. The level of evidence to support these conclusions is strong. The reproducibility of motion palpation (MP), soft tissue changes (STC) and static palpation (SP) is not clinically acceptable.

The level of evidence is strong for interobserver reproducibility of MP and STC, whereas no evidence or conflicting evidence exists for SP and intraobserver reproducibility of STC. Results are overall robust with respect to the predefined levels of acceptable quality. However, the results are sensitive to changes in the preset level of clinically acceptable reproducibility and to the number of included studies.



REFERENCES:

  1. Bergmann TF Petersen DH
    Joint principles and procedures.
    in: Bergmann TF Petersen DH Lawrence DJ Chiropractic technique: principles and procedures.
    Churchill Livingstone Inc, New York1993: 51-121

  2. Schafer RC, Faye LJ
    Introduction To the Dynamic Chiropractic Paradigm
    in: Schafer RC Faye LJ Motion palpation and chiropractic technique. 1st ed.
    The motion palpation institute, Huntington Beach, CA: 1-41

  3. Maitland GD
    Vertebral manipulation.
    3rd ed. Butterworths, London 1977

  4. Hestbaek L Leboeuf-Yde C
    Are chiropractic tests for the lumbo-pelvic spine reliable
    and valid? A systematic critical literature review.
    J Manipulative Physiol Ther. 2000; 23: 258-275

  5. Huijbregts PA
    Spinal motion palpation: a review of reliability studies.
    J Man Manip Ther. 2002; 10: 24-39

  6. van der Wurff P Hagmeijer RH Meyne W
    Clinical tests of the sacroiliac joint. A systemic methodological review.
    Part 1: reliability.
    Man Ther. 2000; 5: 30-36

  7. Seffinger MA Najm WI Mishra SI Adams A Dickerson VM Murphy LS et al.
    Reliability of spinal palpation for diagnosis of back and neck pain:
    a systematic review of the literature.
    Spine. 2004; 29: E413-E425

  8. Haas M
    Statistical methodology for reliability studies.
    J Manipulative Physiol Ther. 1991; 14: 119-132

  9. Vach W
    The dependence of Cohen's kappa on the prevalence does not matter.
    J Clin Epidemiol. 2005; 58: 655-661

  10. Vaughan B
    Inter-examiner reliability in detecting cervical spine dysfunction: a short review.
    J Osteopath Med. 2002; 5: 24-27

  11. van Tulder MW Assendelft WJ Koes BW et al.
    Method guidelines for systematic reviews in the
    Cochrane collaboration back review group for spinal disorders.
    Spine. 1997; 22: 2323-2330

  12. Clarke M Oxmann AD
    Cochrane reviewers' handbook 4.2.0.
    Cochrane Collaboration, Oxford 2003 (cited 2004 Jun 1)

  13. Hoogendoorn WE van Poppel MN Bongers PM Koes BW Bouter LM
    Systematic review of psychosocial factors at work and private life
    as risk factors for back pain.
    Spine. 2000; 25: 2114-2125

  14. Patijn J
    Reproducibility and validity studies of diagnostic procedures in manual/
    musculoskeletal medicine.
    in: International Federation for Manual/Musculoskeletal Medicine Scientific committee.
    Protocol Formats 2004

  15. Deeks JJ
    Systematic reviews in health care:
    systematic reviews of evaluations of diagnostic and screening tests.
    BMJ. 2001; 323: 157-162

  16. Irwig L Macaskill P Glasziou P et al.
    Meta-analytic methods for diagnostic test accuracy.
    J Clin Epidemiol. 1995; 48: 119-130

  17. Altman DG
    Some common problems in medical research.
    in: Altman DG Practical statistics for medical research.
    Chapman & Hall, London1991: 396-439

  18. Bigos S, Bower O, Braen G, et al.
    Acute Lower Back Problems in Adults. Clinical Practice Guideline No. 14.
    Rockville, MD: Agency for Health Care Policy and Research,
    Public Health Service, U.S. Department of Health and Human Services; 1994

  19. Hartvigsen J Lings S Leboeuf-Yde C Bakketeig L
    Psychosocial factors at work in relation to low back pain and consequences
    of low back pain; a systematic, critical review of prospective cohort studies.
    Occup Environ Med. 2004; 61: e2

  20. Pool JJ Hoving JL De Vet HC van Mameren H Bouter LM
    The interexaminer reproducibility of physical examination of the cervical spine.
    J Manipulative Physiol Ther. 2004; 27: 84-90

  21. Fjellner A Bexander C Faleij R Strender LE
    Interexaminer reliability in physical examination of the cervical spine.
    J Manipulative Physiol Ther. 1999; 22: 511-516

  22. Strender LE Lundin M Nell K
    Interexaminer reliability in physical examination of the neck.
    J Manipulative Physiol Ther. 1997; 20: 516-520

  23. Strender LE Sjoblom A Sundell K Ludwig R Taube A
    Interexaminer reliability in physical examination of patients with low back pain.
    Spine. 1997; 22: 814-820

  24. Keating JC Bergmann TF Jacobs GE Finer BA Larson K
    Interexaminer reliability of eight evaluative dimensions of lumbar segmental abnormality.
    J Manipulative Physiol Ther. 1990; 13: 463-470

  25. Viikari-Juntura E
    Interexaminer reliability of observations in physical examinations of the neck.
    Phys Ther. 1987; 67: 1526-1532

  26. Sebastian D Chovvath R
    Reliability of palpation assessment in non-neutral dysfunctions of the lumbar spine.
    Orthop Phys Ther Pract. 2004; 16: 23-26

  27. Hicks GE Fritz JM Delitto A Mishock J
    Interrater reliability of clinical examination measures for
    identification of lumbar segmental instability.
    Arch Phys Med Rehabil. 2003; 84: 1858-1864

  28. Downey B Nicholas T Niere K
    Can manipulative physiotherapists agree on which lumbar level to treat based on palpation?
    Physiotherapy. 2003; 89: 74-81

  29. Christensen HW Vach W Manniche C Haghfelt T Hartvigsen L Høilund-Carlsen PF
    Palpation of the upper thoracic spine—an observer reliability study.
    J Manipulative Physiol Ther. 2002; 25: 285-292

  30. Horneij E Hemborg B Johnsson B Ekdahl C
    Clinical tests on impairment level related to low back pain: a study of test reliability.
    J Rehabil Med. 2002; 34: 176-182

  31. Marcotte J Normand MC Black P
    The kinematics of motion palpation and its effect on
    the reliability for cervical spine rotation.
    J Manipulative Physiol Ther. 2002; 25: E7

  32. Comeaux Z Eland D Chila A Pheley A Tate M
    Measurement challenges in physical diagnosis:
    refining interrater palpation, perception and comminication.
    J Bodyw Mov Ther. 2001; 5: 245-253

  33. Ghoukassian M Nicholls B McLaughlin P
    Inter-examiner reliability of the Johnson and Friedman percussion scan of the thoracic spine.
    J Osteopath Med. 2001; 4: 15-20

  34. French SD Green S Forbes A
    Reliability of chiropractic methods commonly used to detect
    manipulable lesions in patients with chronic low-back pain.
    J Manipulative Physiol Ther. 2000; 23: 231-238

  35. Smedmark V Wallin M
    Inter-examiner reliability in assessing passive intervertebral motion of the cervical spine.
    Man Ther. 2000; 5: 97-101

  36. van Suijlekom HA de Vet HC van den Berg SG Weber WE
    Interobserver reliability in physical examination of
    the cervical spine in patients with headache.
    Headache. 2000; 40: 581-586

  37. Vincent-Smith B Gibbons P
    Inter-examiner and intra-examiner reliability of standing flexion test.
    Man Ther. 1999; 4: 87-93

  38. Hawk C Phongphua C Bleecker J Swank L Lopez D Rubley T
    Preliminary study of the reliability of assessment procedures for
    indications for chiropractic adjustments of the lumbar spine.
    J Manipulative Physiol Ther. 1999; 22: 382-389

  39. Meijne W van Neerbos K Aufdemkampe G van der Wurff P
    Intraexaminer and interexaminer reliability of the Gillet test.
    J Manipulative Physiol Ther. 1999; 22: 4-9

  40. Lundberg G Gerdle B
    The relationships between spinal sagittal configuration, joint mobility,
    general low back mobility and segmental mobility in female homecare personnel.
    Scand J Rehabil Med. 1999; 31: 197-206

  41. Cattrysse E Swinkels RAH Oostendorp RAB Duquet W
    Upper cervical instability: are clinical tests reliable?
    Man Ther. 1997; 2: 91-97

  42. Jull G Zito G
    Inter-examiner reliability to detect painful upper cervical joint dysfunction.
    Aust J Physiother. 1997; 43: 125-129

  43. McPartland JM Goodridge JP
    Counterstrain and traditional osteopathic examination
    of the cervical spine compared.
    J Bodyw Mov Ther. 1997; 1: 173-178

  44. Tuchin P Hart J Colman R Johnson C Gee A Edwards I et al.
    Interexaminer reliability of chiropractic evaluation for
    cervical spine problems—a pilot study.
    Chiropr J Aust. 1996; 5: 23-29

  45. Haas M
    Reliability of manual end-play palpation of the thoracic spine.
    Chiropr Tech. 1995; 7: 120-124

  46. Lindsay DM
    Interrater reliability of manual therapy assessment techniques.
    Phys Ther Can. 1995; 47: 173-180

  47. Binkley J Stratford PW Gill C
    Interrater reliability of lumbar accessory motion mobility testing.
    Phys Ther. 1995; 75: 786-792

  48. Inscoe EL Witt PL Gross MT Mitchell RU
    Reliability in evaluating passive intervertebral motion of the lumbar spine.
    J Man Manip Ther. 1995; 3: 135-143

  49. Maher C Adams R
    Reliability of pain and stiffness assessments in clinical manual lumbar spine examination.
    Phys Ther. 1994; 74: 801-809

  50. Hubka MJ Phelan SP
    Interexaminer reliability of palpation for cervical spine tenderness.
    J Manip Physiol Ther. 1994; 17: 591-595

  51. Paydar D Thiel H Gemmell H
    Intra- and interexaminer reliability of certain pelvic palpatory procedures
    and the sitting flexion test for sacroiliac joint mobility and dysfunction.
    J Neuromusculoskel Syst. 1994; 2: 65-69

  52. Boline PD Haas M Meyer JJ Kassak K Nelson C Keating JC
    Interexaminer reliability of eight evaluative dimensions of lumbar segmental abnormality: part II.
    J Manipulative Physiol Ther. 1993; 16: 363-374

  53. Mior SA McGregor M Schut B
    The role of experience in clinical accuracy.
    J Manipulative Physiol Ther. 1990; 13: 68-71

  54. Leboeuf C
    Chiropractic examination procedures: a reliability and consistency study.
    J Aust Chiropr Assoc. 1989; 19: 101-104

  55. Herzog W Read LJ Conway PJ Shaw LD McEwen MC
    Reliability of motion palpation procedures to detect sacroiliac joint fixations.
    J Manipulative Physiol Ther. 1989; 12: 86-92

  56. Nansel DD Peneff AL Jansen RD Cooperstein R
    Interexaminer concordance in detecting joint-play asymmetries in
    the cervical spines of otherwise asymptomatic subjects.
    J Manipulative Physiol Ther. 1989; 12: 428-433

  57. Mootz RD Keating JC Kontz HP Milus TB Jacobs GE
    Intra- and interobserver reliability of passive motion palpation of the lumbar spine.
    J Manipulative Physiol Ther. 1989; 12: 440-445

  58. Boline PD
    Interexaminer reliability of palpatory evaluations of the lumbar spine.
    Am J Chiropr Med. 1988; 1: 5-11

  59. Carmichael JP
    Inter- and intra-examiner reliability of palpation for sacroiliac joint dysfunction.
    J Manipulative Physiol Ther. 1987; 10: 164-171

  60. Love RM Brodeur RR
    Inter- and intra-examiner reliability of motion palpation for the thoracolumbar spine.
    J Manipulative Physiol Ther. 1987; 10: 1-4

  61. Bergstrøm E Courtis G
    An inter- and intra-examiner reliability study of motion palpation
    of the lumbar spine in lateral flexion in the seated position.
    Eur J Chiropr. 1986; 34: 121-141

  62. Mior SA King R
    Intra and interexaminer reliability of motion palpation in the cervical spine.
    J Can Chiropr Assoc. 1985; 29: 195-199

  63. Deboer KF Harmon R Tuttle CD Wallace H
    Reliability study of detection of somatic dysfunctions in the cervical spine.
    J Manipulative Physiol Ther. 1985; 8: 9-16

  64. Potter NA Rothstein JM
    Intertester reliability for selected clinical tests of the sacroiliac joint.
    Phys Ther. 1985; 65: 1671-1675

  65. Johnston WL Allan BR Hendra JL Neff DR Rosen ME Sills LD et al.
    Interexaminer study of palpation in detecting location of spinal segmental dysfunction.
    J Am Osteopath Assoc. 1983; 82: 839-845

  66. Gonella C Paris SV Kutner M
    Reliability in evaluating passive intervertebral motion.
    Phys Ther. 1982; 62: 436-444

  67. Wiles MR
    Reproducibility and interexaminer correlation of motion palpation
    findings of the sacroiliac joints.
    J Can Chiropr Assoc. 1980; 24: 59-69

  68. Oldreive WL
    Manual therapy rounds. A critical review of the literature on tests of the sacroiliac joint.
    J Man Manip Ther. 1995; 3: 157-161

  69. Keating JC
    Inter-examiner reliability of motion palpation of the lumbar spine:
    a review of quantitative literature.
    Am J Chiropr Med. 1989; 2: 107-110

  70. Panzer DM
    The reliability of lumbar motion palpation.
    J Manipulative Physiol Ther. 1992; 15: 518-524

  71. Haas M
    The reliability of reliability.
    J Manipulative Physiol Ther. 1991; 14: 199-208

  72. Humphreys BK, Delahaye M, Peterson CK:
    An Investigation into the Validity of Cervical Spine Motion Palpation Using
    Subjects with Congenital Block Vertebrae as a 'Gold Standard'

    BMC Musculoskelet Disord 2004 (Jun 15); 5 (1): 19

  73. van Deursen L Patijn J Ockhuysen A Vortman BJ
    The value of some clinical tests of the sacro-iliac joint.
    Man Med. 1990; 5: 96-99

  74. Feldt LS McKee ME
    Estimation of the reliability of skill tests.
    Res Q. 1958; 29: 279-293

  75. Haas M, Groupp E, Panzer D, Partna L, Lumsden S, Aickin M
    Efficacy of Cervical Endplay Assessment
    as an Indicator for Spinal Manipulation

    Spine (Phila Pa 1976) 2003 (Jun 1); 28 (11): 1091–1096

  76. Landis JR Koch GC
    The measurement of observer agreement for categorical data.
    Biometrics. 1977; 33: 159-174

  77. Huxley R Neil A Collins R
    Unravelling the fetal origins hypothesis: is there really an
    inverse association between birthweight and subsequent blood pressure?
    Lancet. 2002; 360: 659-665

  78. Chan AW Hrobjartsson A Haahr MT Gotzsche PC Altman DG
    Empirical evidence for selective reporting of outcomes in
    randomized trials: comparison of protocols to published articles.
    JAMA. 2004; 291: 2457-2465

Return to SPINAL PALPATION

Return to LOCATING SUBLUXATIONS

Since 2-21-2022

© 1995–2025 ~ The Chiropractic Resource Organization ~ All Rights Reserved