Next Article in Journal
Special Issue “Diagnosis and Treatment of Rare Diseases”
Previous Article in Journal
Contemporary Prevalence of Oral Clefts in the US: Geographic and Socioeconomic Considerations
Previous Article in Special Issue
Unveiling Timetable for Physical Therapy after Single-Level Lumbar Surgery for Degenerative Disc Disease: Insights from a Systematic Review and Meta-Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Intra-Examiner Reliability and Validity of Sagittal Cervical Spine Mensuration Methods Using Deep Convolutional Neural Networks

by
Mohammad Mehdi Hosseini
1,
Mohammad H. Mahoor
1,2,
Jason W. Haas
3,
Joseph R. Ferrantelli
3,4,
Anne-Lise Dupuis
4,
Jason O. Jaeger
5 and
Deed E. Harrison
3,*
1
Ritchie School of Engineering and Computer Science, University of Denver, Denver, CO 80208, USA
2
Dreamface Technologies LLC, Centennial, CO 80111, USA
3
CBP Non-Profit, Inc., Eagle, ID 83616, USA
4
PostureCo, Inc., Trinity, FL 34655, USA
5
Community Based Internship Program, Associate Faculty, Southern California University of Health Sciences, Whittier, CA 90604, USA
*
Author to whom correspondence should be addressed.
Submission received: 19 January 2024 / Revised: 20 April 2024 / Accepted: 24 April 2024 / Published: 27 April 2024

Abstract

:
Background: The biomechanical analysis of spine and postural misalignments is important for surgical and non-surgical treatment of spinal pain. We investigated the examiner reliability of sagittal cervical alignment variables compared to the reliability and concurrent validity of computer vision algorithms used in the PostureRay® software 2024. Methods: A retrospective database of 254 lateral cervical radiographs of patients between the ages of 11 and 86 is studied. The radiographs include clearly visualized C1–C7 vertebrae that were evaluated by a human using the software. To evaluate examiner reliability and the concurrent validity of the trained CNN performance, two blinded trials of radiographic digitization were performed by an extensively trained expert user (US) clinician with a two-week interval between trials. Then, the same clinician used the trained CNN twice to reproduce the same measures within a 2-week interval on the same 254 radiographs. Measured variables included segmental angles as relative rotation angles (RRA) C1–C7, Cobb angles C2–C7, relative segmental translations (RT) C1–C7, anterior translation C2–C7, and absolute rotation angle (ARA) C2–C7. Data were remotely extracted from the examiner’s PostureRay® system for data collection and sorted based on gender and stratification of degenerative changes. Reliability was assessed via intra-class correlations (ICC), root mean squared error (RMSE), and R2 values. Results: In comparing repeated measures of the CNN network to itself, perfect reliability was found for the ICC (1.0), RMSE (0), and R2 (1). The reliability of the trained expert US was in the excellent range for all variables, where 12/18 variables had ICCs ≥ 0.9 and 6/18 variables were 0.84 ≤ ICCs ≤ 0.89. Similarly, for the expert US, all R2 values were in the excellent range (R2 ≥ 0.7), and all RMSEs were small, being 0.42 ≤ RMSEs ≤ 3.27. Construct validity between the expert US and the CNN network was found to be in the excellent range with 18/18 ICCs in the excellent range (ICCs ≥ 0.8), 16/18 R2 values in the strong to excellent range (R2 ≥ 0.7), and 2/18 in the good to moderate range (R2 RT C6/C7 = 0.57 and R2 Cobb C6/C7 = 0.64. The RMSEs for expert US vs. the CNN network were small, being 0.37 ≤ RMSEs ≤ 2.89. Conclusions: A comparison of repeated measures within the computer vision CNN network and expert human found exceptional reliability and excellent construct validity when comparing the computer vision to the human observer.

1. Introduction

The burden of spine injuries and chronic spine pain for patients and society is tremendous and a growing global concern. Diagnosis, treatment, and long-term consequences of spine conditions are the single greatest cause of disability due to musculoskeletal disorders globally [1]. Diagnosis and intervention for spine conditions vary greatly due to socio-economic conditions, access to current treatment methods and techniques, presence of prior and concomitant conditions, and past interventions, whether conservative, therapeutic, or surgical [1,2,3]. Interventions for spine conditions each have varying potential benefits and costs for the patient and society [4,5,6]. Finding efficacious, validated, repeatable, and reliable diagnostic and therapeutic or surgical methods to resolve and improve spine pain is of great benefit to the individual as well as global populations [7,8,9,10].
Technological advances in spine pain diagnosis and treatment are necessary to reduce cost, improve outcomes, increase efficiency for the facilities and clinicians as well as reduce poor outcomes that may require additional care [11,12,13]. Technology can make the process of spine condition diagnosis more efficacious [14,15,16]. For example, previously, we presented a machine learning deep convoluted neural network (DCNN or CNN) that demonstrated that computer vision (CV) is superior to human measurement of spine displacements [17]. CNNs or DCNNs evolved from traditional artificial neural networks, having their origins based on the understanding of the visual cortex of animals, and are commonly used to identify imaging and video patterns. In this investigation [17], following thousands of evaluations, the program model and software found perfect reliability via intra-class correlation (ICC) and linear regression R2 values. Further, the root mean squared error (RMSE), which measures the average difference between a statistical model’s predicted values and the actual values, was zero. To our knowledge, no prior program, study, or software has demonstrated this perfect accuracy and repeatability for spine mensuration. This technology could prove critical for improving the biomechanical analysis of normal and abnormal spinal configurations and could significantly alter treatment for many treating physicians.
This current investigation is a continuation of a previous investigation that we performed, where we provided comparisons with the current CNN model to other CNN models [17]. Herein, we present unique findings from intra-examiner measurement reliability with a repeated measures design using a highly skilled examiner (human) on an original retrospective database of lateral cervical radiographs. Secondly, we present the machine learning CNN system using the same repeated measures design on the same X-ray images to simultaneously investigate reliability with concurrent construct validity against the human study to determine abnormal cervical sagittal spine configuration using intersegmental, regional, and global analyses. We hypothesize that while both the human and CNN systems will have excellent intra-examiner reliability, the CNN will be perfect to near perfect and the construct validity will be high.

2. Materials and Methods

2.1. Radiographic Image Selection Inclusion Criteria

This study retrospectively obtained 254 consecutive lateral cervical radiographs from a clinical chiropractic practice that required radiographic examination of presenting patients between 1 January 2021 and 10 September 2021. Due to the retrospective nature of our collected material, our design is exempt from IRB approval under section 45 CFR 46.101(b)(4). See https://www.hhs.gov/ohrp/regulations-and-policy/decision-charts-pre-2018/index.html#c5 (accessed on 15 April 2024). The patient ages at the time images were obtained ranged from 11 years old to 86 years old, with 108 males and 146 females. To mimic “real-world clinical practice” and better test reliability and validity of a human and the CNN system, we included all radiographs only if the C1–C7 region was visible. The radiographic images were obtained retrospectively using the PostureRay® radiographic documentation system (PostureCo, Inc., Trinity, FL, USA). This patented software system is commonly used by clinicians to streamline radiographic spinal alignment documentation workflows.

2.2. Radiographic Image Selection Exclusion Criteria

Radiographic image exclusion criteria were based on two criteria: (1) if C1–C7 were not clearly visible to the eye or were cut off on the image, and (2) if surgical devices or other obvious artifacts were visible on the X-ray image. No other exclusion criteria were used, and all types of spine degenerative changes and altered sagittal alignment were allowed. As a result, 254 radiographic images were included in this study sample. These retrospective X-ray images were not part of the original dataset used to train the neural networks, representing the first exposure of the trained network to these data. The original Deep CNN was trained and evaluated on 24,419 annotated unique patients’ lateral cervical X-rays, digitized by an expert clinician. It is notable that 95 percent of these data were used for training and 5 percent for validation. For more details, see [17].

2.3. Intra-Examiner Reliability and Construct Validity Design

To evaluate intra-examiner reliability and construct validity of the trained network’s performance, any prior digitization annotation markings and measurements were cleaned from the images prior to clinician processing. Two blind trials of digitization were then performed by a trained clinician (JRF) with a two-week interval between trials. The anatomical digitization points used were as follows:
  • Three points on C1: anterior tubercle, midpoint C1 at the posterior margin of the dens, and midpoint of the posterior spinal laminar line.
  • C2–C7 digitization consisted of four points per vertebra: anterior superior, posterior superior, anterior inferior, and posterior inferior vertebral body margins.
In this current study, lateral cervical measurements obtained in the PostureRay® software 2024 were derived from the following anatomical digitization points:
  • Atlas plane relative to horizontal;
  • Segmental posterior body tangent relative rotational angles (RRAs);
  • Cobb analysis using vertebral endplate angles;
  • Segmental relative linear translation distances;
  • Global posterior tangent absolute rotational angle from C2 to C7;
  • Global sagittal horizontal translation alignment of C2 relative to C7.
After each trial, the data were remotely extracted from the examiner’s PostureRay® 2024 system for data collection and sorted based on gender and stratification of degenerative changes. The clinician responsible for digitization did not have access to the raw data, nor were they involved in interpretation of statistical analysis at any point. Figure 1 shows several images with the landmark points used in this investigation.

2.4. Measurement Variables Reported

In the following section, we describe our methodology in more detail and the parameters and measures used for evaluating the model’s reliability. We originally trained a CNN-based deep neural network model using more than 24 K sagittal cervical spine X-ray images and the provided anatomical landmark points. The landmark points were labeled by expert humans and utilized to train a robust model. The images included three types of poses: poses including normal neutral lateral cervical, lateral cervical extension, and lateral cervical flexion. The details of our model design and implementation can be found elsewhere [17]. In the current study, we randomly selected 254 consecutive images meeting the above inclusion criteria, not included in the training set, to automatically predict the landmarks. In the next step, an expert human corrected the location of the falsely predicted landmark points. It is noteworthy that we repeated this experiment at two different times (two weeks apart) to be able to analyze the intra-examiner agreement during the first and second rounds of landmark corrections.
To evaluate the agreement between the model and the expert human, and the expert’s agreement in the first and second experiments, 18 translational and rotational measurement variables are extracted. The variables are as follows:
  • ARA (°): Absolute rotational angle refers to the overall curve of the cervical lordosis. It is computed as the angle between the vertebrae C2 and C7. It is the angle between two straight lines, where they intersect each other. The first line passes through the posterior inferior and posterior superior body corners of vertebra C2, and the second line is the line that intersects the posterior inferior and posterior superior vertebral body corners of C7.
  • RRA (°): Relative rotational angle is the angle between two consecutive vertebrae. To calculate this angle, we draw the lines passing through the posterior superior and posterior inferior of any vertebral body corners and then calculate the angle where they cross each other. Thus, creating the slope or the first derivative of the curve when expanded across the vertebral column.
  • KA (°): This represents endplate cross-sectional angle, where for two adjacent vertebrae, we draw the lines that pass the anterior inferior and posterior inferior body corner of each vertebra body as well as the anterior superior and posterior superior body corner and then calculate the angle of their intersection. This measurement is considered less reliable due to the nature of degenerative change at the endplate, which can make two like points difficult to assess.
  • ST (mm): Denotes segmental translations. Like RR and KA features, it is calculated for any pair of adjacent vertebrae and determines the forward or backward translation along the z-axis between two neighboring vertebrae. Positive value means anterior translation, and a negative value means posterior translation relative to the adjacent segment.
  • C1H (°): Demonstrates the atlas plane angle relative to true horizontal and is measured as an angle between a horizontal line and vertebra C1.
  • TR (mm): The translational distance of the C2 posterior superior body corner relative to a vertical line drawn superiorly from the C7 posterior inferior body corner is considered as the translation measure in millimeters.
It is notable that the variables RRA (a.k.a. RR), KA, and ST are calculated for any two consecutive pairs of cervical vertebrae from C2 to C7 and, thus, provide a segmental stability analysis for both rotations and translations.

2.5. Statistical Analysis

Using the Python (3.8.10) libraries, including NumPy (1.23.4), Pandas (1.5.3), Scikit-learn (1.2.1), SciPy (1.10.0), and Pingouin (0.5.3), statistical analysis of human intra-examiner and CNN reliability was performed on both trials to assess reliability data as well as to compare the CNN measurements vs. the clinicians. Additionally, real-world construct validity was evaluated by the clinician after the network automatically predicted digitization localizations. In this process, the clinician adjusted the anatomical predictions when necessary, ensuring correct anatomical locations. This allowed tracking of rotations and translations of the computer-predicted digitized locations compared to the ground truths determined by the clinician. PostureRay® calculated rotations and translations of clinical lines of mensuration based on these digitization points, and statistical analysis was performed on these measurements.
As a detailed reliability assessment of the analytical measures, in addition to the mean error and standard deviation of errors of measurement, we report the root mean squared error (RMSE), intraclass correlations (ICC), and linear regression R2 measures in this study. Note: (1) the RMSE measures the average difference between a statistical model’s predicted values and the actual values. (2) The intraclass correlation (ICC) is a descriptive statistic of reliability between 2 or more datasets where quantitative measurements are made on units that are organized into their respective groups. The ICC ranges from 0 to 1 and describes how strongly units in the same group compare to one another, where 1 is perfect. (3) Finally, the R2 linear regression analysis was used to compare the two measured variable sets for human vs. human measures, CNN vs. CNN measures, and human vs. CNN measures in order to determine the statistical fit and percentage variation between the two measurements for within and between each of the methods. In general, interpreting the relative strength of a relationship based on its R2 value is the following: (1) none or very weak effect size R2 < 0.3; (2) a weak effect size 0.3 < R2 < 0.5; (3) a moderate effect size is 0.5 < R2 < 0.7; (4) a strong effect size is given by R2 > 0.7; and (5) R2 = 1.0 is perfect agreement [18].

3. Results

We extracted 18 variables for all our data sampling and analyzed them using three measures: RMSE, ICC, and R2. Figure 2 illustrates these values on the model (CV) with respect to the expert human (US) over three general features: ARA, C1H, and TR. On the other hand, Figure 3, Figure 4, and Figure 5 show the same measures for the segmental features, KA, RRA, and ST, respectively.
In Figure 2, Figure 3, Figure 4 and Figure 5, the line shows zero error, whereas the points with more distance from the line indicate more error. There is a direct relation between lower error and the measures ICC and R2, while the higher RMSE indicates more error. To distinguish the accuracy of the model on the different variables (features), we classify the features into three groups based on their R2 value. As shown in the figures, the R2 value for the features ARA, KA23, KA34, KA45, KA56, and TR is more than 0.90, so we classify them as the super-clean group of the features. However, R2 is between 0.75 and 0.90 for the features C1H, RR23, RR34, RR45, RR56, ST23, and ST34, which constitute the clean group of features. Finally, since the R2 score of the features KA67, RR67, ST45, ST56, and ST67 is between 0.5 and 0.75, they are considered semi-clean features.
The reported R2, ICC, and RMSE in Figure 2, Figure 3, Figure 4 and Figure 5 show that the error rate between the model (CV) and the expert human (US) annotator is not significant. While the R2 score is in the range of [0.57, 0.99], the ICC varies in the boundary of [0.80, 0.99]. These numbers are accompanied by the acceptable RMSE for all the proposed features. Based on the information shown in these figures, the error rate between the model and the expert human is negligible; therefore, the model is reliable. This reliability is also assessed by repeating the experiment of the annotation by an expert human twice.
To assess the reliability of the model, every image was annotated two times by the same expert human at different times to study the intra-annotator error rate. Figure 6 reveals the error rate of the expert human on some of the randomly selected features. A comparison between the similar features of Figure 2, Figure 3, Figure 4 and Figure 5 and Figure 6 depicts that the error value between the model and the expert human is in the range of the error between the two experiments conducted by the expert human. For example, for the feature KA23, the R2 value for the model analysis is 0.93, while this measure for the expert analysis is 0.92. Studying the features in Figure 6 shows that among the six studied features, in the features KA23, KA56, and ST23, the agreement between the model and expert is even more than two experiments held by the expert human. In addition, Table 1 presents the mean and standard deviation of the error, as well as the root mean squared error, ICC, and R2 for both experiments between the model and human and human against human. This experimental result highlights the fact that the accuracy of the deep neural network model is comparable to the real expert human, and medical landmark detection issues can be trusted.
Moreover, Figure 7 reveals similar discrepancies over different features in CV versus US and US versus US experiments. This similarity determines that the model’s variation is like the human annotator’s variation. This information indicates that the trained model is as reliable as the human expert annotator. It is notable that in the CV versus CV analysis, there is no error, and the ICC is maximum at 1.0. The reason is that the trained model, like all the other CNN models, is deterministic and always generates the same results for a specific input. To find some comparison between the baseline CNN model and the other models (CV vs. the other CVs), see [17].

4. Discussion

Deep learning models are widely used in automated medical image processing tasks, including image segmentation, tumor detection, and anatomical landmark point detection. In a recent study, we developed a Convolutional Neural Networks (CNNs) based model to automatically detect spine landmark points in the sagittal cervical spine on X-ray images [17]. Our developed model has been shown to be very accurate in predicting radiographic anatomical landmark vertebral points [17]. This prior investigation demonstrated that computer vision (CV) is superior to human measurement of spine displacements [17]. Our current investigation is a continuation of our previous study, where we provided comparisons with the current CNN model to other CNN models that exist in the literature, and the reader is referred to this background information [17].
In our current investigation, to further investigate the reliability and real-world validity of the model against a human annotator over time with repeated measures, we first collected reference data, annotated by an expert human annotator, and then compared the model’s predictions with these reference points. To further evaluate the annotation process, we also asked the annotator to repeat the annotations a second time two weeks later to further measure the annotator’s reliability. This allowed us to compare the trustworthiness of the model concerning the annotator’s reliability. Using 18 standard rotational and translational variable measurements for the sagittal cervical spine, our reliability results indicate good to excellent intra-class correlation coefficients (ICCs), small root mean squared errors (RMSE), and good to excellent to perfect R2 values, depending upon the variable assessed. Furthermore, our findings that the error rate between the two human user and computer vision experiments is very similar indicate that the computer vision model expectancy outcomes are the same as the human annotator. Furthermore, there is no difference between the CV model output under the two tested experiments, so the error rate is zero, the ICC is maximum, and the R2 value is perfect. Thus, both of our study’s hypotheses are validated in as much as the human and the CNN system have excellent intra-examiner reliability, and the CNN model has high construct validity compared to the experienced human.
Neck pain is a major contributor to the global burden of disease and is rated as the fourth greatest contributor to global disability [19]. Chronic neck pain is associated with reduced productivity and increased healthcare utilization and can lead to functional impairment and psychological distress, both of which can compromise overall quality of life [20]. There is a growing interest concerning the understanding of the biomechanics of the sagittal configuration of the cervical spine [21]. Importantly, in the past two decades, cervical sagittal alignment has gained more attention as an important clinical outcome in healthcare. It has been demonstrated that abnormal cervical sagittal alignment significantly influences human health and well-being, as it has been shown to be associated with pain [22], disability [23], overall functional performance [24], and quality of life [25]. Despite modern advances in technology related to imaging leading to improved diagnosis and treatment, billions of humans continue to suffer from daily spine and musculoskeletal pain and disability [11,26,27,28,29]. Physiotherapy, spinal manipulation, and exercise therapy have all been discussed as possible treatments for spine pain. However, these interventions typically do not have high-quality, long-term studies demonstrating successful improvements in HRQoL or patient-reported outcomes. Physical medicine and rehabilitation investigations have reported some positive pain outcomes but do not often report improvements in coronal and sagittal postural and spine balance parameters with the long-term stability of the successful intervention [30,31,32,33,34,35].
The diagnosis and treatment of spine pain and spinal trauma to determine the necessity for more invasive methods have been reported for many decades. Clinically, the use of X-ray for simple images of structure and tissues has been a consistently relied upon tool for spinal conditions causing pain. Reliable, repeatable, valid, and economical methods are necessary for the proper diagnosis of spine pain and associated conditions. Safe and efficacious treatment of spine conditions is a desirable clinical outcome for astute clinicians, physicians, surgeons, and therapists [36,37,38]. Cervical spine radiography provides physicians with a simple and repeatable method to determine sagittal and coronal balance, intersegmental spine misalignment, and differential diagnosis and frequently changes treatment options and approaches [36,37,38,39,40,41]. Specific spine rehabilitation protocols (based on radiographic measured variables) designed to lessen abnormal tissue loads via specific opposite posture exercises, spine extension traction, and spine manipulative therapy show potential for the treatment of spine pain and associated conditions using conservative and safe, repeatable, and efficacious methods [37,38,39,40]. These postural and structural rehabilitation investigation methods studied the sagittal spine configuration and developed average and ideal models for spine clinicians to use to make proper diagnosis and treatment recommendations based on the measurements [36,37,38,40,41].
The diagnosis and treatment of spine conditions have advanced with modern technology, and this technology has enabled advances in options for care. Digital radiography, computerized mensuration programs, and precision digitization tools are necessary to aid and reduce human error from both interventions and spine alignment diagnosis [14,15,16,17]. It has previously been shown that radiography mensuration techniques using analog tools such as pencils and protractors are repeatable, reliable, and valid with multiple investigations that show good inter- and intra-examinator agreement [42,43,44,45]. Radiographic measures of total cervical curvature (absolute rotation angle, ARA C2-C7, and Cobb angles) have previously been shown to have excellent examiner reliability [42,43]. For example, a recent meta-analysis identified that the Cobb method (inferior C2–inferior C7), the Cobb method (middle C1–inferior C7), and the absolute rotation angle (C2–C7) all have very high inter-rater reliability [42]. Similarly, relative rotation angles (RRA’s) for measurement of segmental cervical lordosis have been found to have excellent examiner reliability [44,45]. Finally, the measurement of anterior head translation (AHT) using the horizontal offset of C2 relative to a vertical line originating at the posterior inferior body of C7 has been found to have excellent reliability [44,45].
This current CNN model shows far superior accuracy to previously reported reliability investigations in as much as our results demonstrate a perfect R2 analysis, which is not reproducible with human evaluators even with great experience [42,43,44,45]. Likewise, when comparing the CNN model to itself in the repeated measures, the root mean squared errors (RMSE) were zero, and the ICCs were maximum (1.0), indicating perfect agreement with itself. To our knowledge, it has not been previously demonstrated that a cervical spine radiography CNN alignment tool can demonstrate such precision in the measurement of all 18 of the measured sagittal plane variables as performed herein. This is especially important when understanding that computer vision recognizes the lateral cervical radiograph every time and with exacting reliability and has demonstrated the ability to measure the structural abnormalities every time with an error of zero. There were no other programs in the literature that have computer vision networks that can recognize a lateral cervical radiograph every time and measure repeatably with such precision. Of note, the program appears to be learning much in the way that the human measurements improve over time. The clinical application of a tool such as this software should provide clinicians with much more certainty in their accurate diagnosis of spine abnormalities and likely improve the outcome of treatment due to less human error in the assessment and application.
In practical application, our original CNN model (and other CCN models) is a more accurate method of assessing anatomical and biomechanical positions of the cervical spine in the analysis of radiographic images as compared to a trained and experienced human user. These computerized analytical models have clear advantages over human capabilities. However, caution is advisable in this regard as the spine alignment data that these models derive and report are only one part of the healthcare physician’s basis in the formulation of conclusions and proper diagnoses for a given patient; the findings must be taken in the larger context of the full and comprehensive patient examination. It is the combination of medical knowledge and experience of the treating healthcare provider combined with image analysis using sophisticated CNN models that will result in a well-planned and executed treatment plan and procedures. Despite the recognition of sophisticated computerized examination methods as being more objective, with more precise measurements, the decisive variable in diagnosis and therapy application is the unique clinical presentation of the patient; thus, these CNN models and their enhanced measurements must be considered as an auxiliary tool for and not to replace the physician. Therefore, it is worth remembering that the human aspect of medicine has not lost and must not lose its importance.

4.1. Limitations

The limitations of this study are the fact that it is the first report of perfect accuracy with computer vision spine biomechanics mensuration, and repeated studies are necessary for firm conclusions. Further, larger studies are necessary to make absolute statements confirming the perfect accuracy of the program across various external datasets encompassing multiple spine conditions and surgical instrumented or fused segments. Larger studies incorporating radiography of other views of the spine and multiple spinal regions (full spine films) need to be performed as well, and investigations involving degeneration, congenital and morphological anomalies, as well as the consequences of single or multiple traumatic spine injuries should be performed [46]. Accordingly, larger studies are planned to involve more physicians’ images across multiple conditions using the software and CNN model. Studies involving patients with and without the use of PostureRay® could further illuminate the necessity of precise alignment diagnosis before surgical and non-surgical interventions. However, it is noteworthy that the baseline model is trained over a set of images cropped with 5–10 percent boundaries around all the spines; therefore, cropping may affect the model performance. This is inevitable in machine learning tasks. The easiest solution to tackle this source of error is training a model to automatically crop a fixed area around the spine boundary.

4.2. Conclusions

A machine learning tool (which is part of the PostureRay® software 2024) is simple, economical, valid, and repeatable as an instrument to aid in the measurement of the sagittal cervical spine alignment. Our investigation demonstrated the machine learning computer vision tool has a perfect R2 statistical analysis, a zero root mean square error, and an ICC of 1.0 (perfect reliability) when tested against itself with a repeated measure design. Additionally, the construct validity of the CNN software 2024 compared to an expert annotator was in the excellent range. This easy-to-use tool is far superior with regards to reliability when compared to analysis by a human clinician, even with many years of radiographic mensuration experience both manually and digitally and the tool appears to be perfect relative to itself every time, unlike the human. To our knowledge, this excellent reliability and validity has not been previously reported in the machine learning literature. Additional research is warranted to determine full spine condition implications for this technology.

Author Contributions

J.R.F. and D.E.H. conceived the research idea and participated in its design; J.R.F. performed the radiographic mensuration. M.M.H., M.H.M. and D.E.H. contributed to the statistical analysis and interpretation; A.-L.D. participated in the data collection and software programming.; M.M.H., M.H.M., J.W.H., J.R.F., J.O.J. and D.E.H. contributed to the interpretation of the results and wrote varying drafts. All authors have read and agreed to the published version of the manuscript.

Funding

We acknowledge CBP Non-Profit, Inc. (Eagle, ID, USA) for funding support.

Institutional Review Board Statement

This article is a retrospective review of clinical records and is exempt from IRB approval under section 45 CFR 46.101(b)(4). See https://www.hhs.gov/ohrp/regulations-and-policy/decision-charts-pre-2018/index.html#c5, accessed on 11 February 2024.

Informed Consent Statement

All data collection was retrospective in nature, and informed consent was not necessary for analysis.

Data Availability Statement

The datasets analyzed in the current study are available from the corresponding authors upon reasonable request.

Conflicts of Interest

Authors M.M.H., M.H.M. and J.J. declare no competing interests. A.-L.D. is a compensated programmer for PostureCo, Inc. J.R.F. is the CEO of PostureCo, Inc. and receives compensation for the sale of PostureRay® software. J.W.H. is a compensated researcher for CBP Non-Profit, Inc. D.E.H. is a consultant for PostureCo and is the CEO of Chiropractic BioPhysics® (CBP®) and provides post-graduate education to healthcare providers and physicians. Spine rehabilitation devices are distributed through his company. D.E.H. is the president of CBP Non-Profit, Inc., a not-for-profit spine research foundation.

References

  1. Fatoye, F.; Gebrye, T.; Ryan, C.G.; Useh, U.; Mbada, C. Global and regional estimates of clinical and economic burden of low back pain in high-income countries: A systematic review and meta-analysis. Front. Public Health 2023, 11, 1098100. [Google Scholar] [CrossRef] [PubMed]
  2. Casiano, V.E.; Sarwan, G.; Dydyk, A.M.; Varacallo, M. Back Pain; StatPearls Publishing: Treasure Island, FL, USA, 2023. [Google Scholar]
  3. Ferreira, M.L.; de Luca, K.; Haile, L.; Steinmetz, J.; Culbreth, G.; Cross, M.; Kopec, J.; Ferreira, P.H.; Blyth, F.; Buchbinder, R.; et al. Global, regional, and national burden of low back pain, 1990–2020, its attributable risk factors, and projections to 2050: A systematic Analysis of the global burden of disease study 2021. Lancet Rheumatol. 2023, 5, e316–e329. [Google Scholar] [CrossRef]
  4. Zhu, K.; Devine, A.; Dick, I.M.; Prince, R.L. Association of back pain frequency with mortality, coronary heart events, mobility, and quality of life in elderly women. Spine 2007, 32, 2012–2018. [Google Scholar] [CrossRef] [PubMed]
  5. Roseen, E.J.; LaValley, M.P.; Li, S.; Saper, R.B.; Felson, D.T.; Fredman, L. Association of back pain with all-cause and cause-specific mortality among older women: A cohort study. J. Gen. Intern. Med. 2019, 34, 90–97. [Google Scholar] [CrossRef] [PubMed]
  6. Williams, A.; Kamper, S.J.; Wiggers, J.H.; O’brien, K.M.; Lee, H.; Wolfenden, L.; Yoong, S.L.; Robson, E.; McAuley, J.H.; Hartvigsen, J.; et al. Musculoskeletal conditions may increase the risk of chronic disease: A systematic review and meta-analysis of cohort studies. BMC Med. 2018, 16, 167. [Google Scholar] [CrossRef] [PubMed]
  7. Champain, S.; Benchikh, K.; Nogier, A.; Mazel, C.; Guise, J.; De Skalli, W. Validation of and new clinical quantitative analysis software applicable in spine orthopedic studies. Eur. Spine J. 2006, 15, 981–991. [Google Scholar] [CrossRef] [PubMed]
  8. Schwartz, J.T.; Cho, B.H.; Tang, P.; Schefflein, J.; Arvind, V.; Kim, J.S.; Doshi, A.H.; Cho, S.K. Deep learning automates measurement of spinopelvic parameters on lateral lumbar radiographs. Spine 2021, 46, E671–E678. [Google Scholar] [CrossRef] [PubMed]
  9. Yang, S.; Zhong, S.; Fan, Y.; Zhu, Y.; Xu, N.; Liao, Y.; Fan, G.; Liao, X.; He, S. Research hotspots and trends on spinal cord stimulation for pain treatment: A two-decade bibliometric analysis. Front. Neurosci. 2023, 17, 1158712. [Google Scholar] [CrossRef] [PubMed]
  10. Kristjansson, E.; Leivseth, G.; Brinckmann, P.; Frobin, W. Increased sagittal plane segmental motion in the lower cervical spine in women with chronic whiplash-associated disorders, grades I–II: A case-control study using a new measurement protocol. Spine 2003, 28, 2215–2221. [Google Scholar] [CrossRef]
  11. Yagi, M.; Yamanouchi, K.; Fujita, N.; Funao, H.; Ebata, S. Revolutionizing spinal care: Current applications and future directions of artificail intelligence and machine learning. J. Clin. Med. 2023, 12, 4188. [Google Scholar] [CrossRef]
  12. Orhurhu, V.J.; Chu, R.; Gill, J. Failed Back Surgery Syndrome; StatPearls Publishing: Treasure Island, FL, USA, 2023. [Google Scholar]
  13. Ju, C.I.; Lee, S.M. Complications and Management of Endoscopic Spinal Surgery. Neurospine 2023, 20, 56–77. [Google Scholar] [CrossRef] [PubMed] [PubMed Central]
  14. Tangsrivimol, J.A.; Schonfeld, E.; Zhang, M.; Veeravagu, A.; Smith, T.R.; Härtl, R.; Lawton, M.T.; El-Sherbini, A.H.; Prevedello, D.M.; Glicksberg, B.S.; et al. Artificial intelligence in neurosurgery: A state-of-the-art review from past to future. Diagnostics 2023, 13, 2429. [Google Scholar] [CrossRef] [PubMed] [PubMed Central]
  15. Wang, F.; Preininger, A. AI in health: State of the art, challenges and future directions. Yearb. Med. Inform. 2019, 28, 16–26. [Google Scholar] [CrossRef] [PubMed]
  16. Zhou, S.; Zhou, F.; Sun, Y.; Chen, X.; Diao, Y.; Zhao, Y.; Huang, H.; Fan, X.; Zhang, G.; Li, X. The application of artificial intelligence in spine surgery. Front. Surg. 2022, 9, 885599. [Google Scholar] [CrossRef] [PubMed]
  17. Fard, A.P.; Ferrantelli, J.; Dupuis, A.-L.; Mahoor, M.H. Sagittal cervical spine landmark point detection in x-ray using deep convoluted neural networks. Sci. Rep. 2021, 11, 7618. [Google Scholar] [CrossRef]
  18. Moore, D.S.; Notz, W.I.; Flinger, M.A. The Basic Practice of Statistics, 6th ed.; W. H. Freeman and Company: New York, NY, USA, 2013; p. 138. [Google Scholar]
  19. Shin, D.W.; Shin, J.I.; Koyanagi, A.; Jacob, L.; Smith, L.; Lee, H.; Chang, Y.; Song, T.J. Global, regional, and national neck pain burden in the general population, 1990–2019: An analysis of the global burden of disease study 2019. Front Neurol. 2022, 13, 955367. [Google Scholar] [CrossRef] [PubMed]
  20. Hush, J.M.; Michaleff, Z.; Maher, C.G.; Refshauge, K. Individual, physical and psychological risk factors for neck pain in Australian office workers: A 1-year longitudinal study. Eur. Spine J. 2009, 18, 1532–1540. [Google Scholar] [CrossRef] [PubMed]
  21. Ling, F.P.; Chevillotte, T.; Leglise, A.; Thompson, W.; Bouthors, C.; Le Huec, J.C. Which parameters are relevant in sagittal balance analysis of the cervical spine? A literature review. Eur. Spine J. 2018, 27 (Suppl. S1), 8–15. [Google Scholar] [CrossRef] [PubMed]
  22. Mahmoud, N.F.; Hassan, K.A.; Abdelmajeed, S.F.; Moustafa, I.M.; Silva, A.G. The Relationship between Forward Head Posture and Neck Pain: A Systematic Review and Meta-Analysis. Curr. Rev. Musculoskelet. Med. 2019, 12, 562–577. [Google Scholar] [CrossRef] [PubMed]
  23. Jackson-Fowl, B.; Hockley, A.; Naessig, S.; Ahmad, W.; Pierce, K.; Smith, J.S.; Ames, C.; Shaffrey, C.; Bennett-Caso, C.; Williamson, T.K.; et al. Adult cervical spine deformity: A state-of-the-art review. Spine Deform. 2024, 12, 3–23. [Google Scholar] [CrossRef]
  24. Saad, N.; Moustafa, I.M.; Ahbouch, A.; Alsaafin, N.M.; Oakley, P.A.; Harrison, D.E. Are Rotations and Translations of Head Posture Related to Gait and Jump Parameters? J. Clin. Med. 2023, 12, 6211. [Google Scholar] [CrossRef] [PubMed]
  25. Aafreen, A.; Khan, A.R.; Khan, A.; Ahmad, A.; Alzahrani, A.H.; Alhusayni, A.I.; Alameer, A.H.; Alajam, R.A.; Ganesan, B.B.M.; Shaphe, M.A. Neck Health Metrics and Quality of Life: A Comparative Study in Bike Drivers with and without Neck Pain. J. Multidiscip. Healthc. 2023, 16, 3575–3584. [Google Scholar] [CrossRef] [PubMed]
  26. Browd, S.R.; Park, C.; Donoho, D.A. Potential Applications of Artificial Intelligence and Machine Learning in Spine Surgery Across the Continuum of Care. Int. J. Spine Surg. 2023, 17 (Suppl. S1), S26–S33. [Google Scholar] [CrossRef]
  27. Hornung, A.L.; Hornung, C.M.; Mallow, G.M.; Barajas, J.N.; Orías, A.A.E.; Galbusera, F.; Wilke, H.-J.; Colman, M.; Phillips, F.M.; An, H.S.; et al. Artificial intelligence and spine imaging: Limitations, regulatory issues and future direction. Eur. Spine J. 2022, 31, 2007–2021. [Google Scholar] [CrossRef] [PubMed]
  28. Broida, S.E.; Schrum, M.L.; Yoon, E.; Sweeney, A.P.; Dhruv, N.N.; Gombolay, M.C.; Yoon, S.T. Improving surgical triage in spine clinic: Predicting likelihood of surgery using machine learning. World Neurosurg. 2022, 163, e192–e198. [Google Scholar] [CrossRef]
  29. Ames, C.P.; Smith, J.S.; Pellisé, F.; Kelly, M.; Alanay, A.; Acaroglu, E.; Pérez-Grueso, F.J.S.; Kleinstück, F.; Obeid, I.; Vila-Casademunt, A.; et al. Artificial intelligence based hierarchical clustering of patient types and intervention categories in adult spinal deformity surgery: Towards a new classification scheme that predicts quality and value. Spine 2019, 44, 915–926. [Google Scholar] [CrossRef] [PubMed]
  30. Cunha, A.C.; Burke, T.N.; França, F.J.; Marques, A.P. Effect of global posture reeducation and of static stretching on pain, range of motion, and quality of life in women with chronic neck pain: A randomized clinical trial. Clinics 2008, 63, 763–770. [Google Scholar] [CrossRef] [PubMed]
  31. Falla, D.; Jull, G.; Russell, T.; Vicenzino, B.; Hodges, P. Effect of neck exercise on sitting posture in patients with chronic neck pain. Phys. Ther. 2007, 87, 408–417. [Google Scholar] [CrossRef] [PubMed]
  32. Özer, K.D.; Toprak, Ç.Ş. Effectiveness of relaxation training in addition to stabilization exercises in chronic neck pain: A randomized clinical trial. Turk. J. Physiother. Rehabil. 2019, 30, 145–153. [Google Scholar] [CrossRef]
  33. Lwin, N.N.; Myint, T.; Oo, W.M.; San, H.H.; Tun, M.T. Efficacy on pressure-biofeedback guided craniocervical flexion exercise in neck pain: A randomized controlled trial. J. Musculoskelet. Res. 2021. [Google Scholar] [CrossRef]
  34. Kim, J.Y.; Kwag, K.I. Clinical effects of deep cervical flexor muscle activation in patients with chronic neck pain. J. Phys. Ther. Sci. 2016, 28, 269–273. [Google Scholar] [CrossRef] [PubMed]
  35. Shah, N.; Gill, M.A.; Singal, S.K.; Payla, M. Effect of core stability exercise in patients with neck pain. Indian. J. Physiother. Occup. Ther. 2020, 14, 102–107. [Google Scholar] [CrossRef]
  36. Oakley, P.A.; Ehsani, N.N.; Moustafa, I.M.; Harrison, D.E. Restoring cervical lordosis by cervical extension traction meth-ods in the treatment of cervical spine disorders: A systematic review of controlled trials. J. Phys. Ther. Sci. 2021, 33, 784–794. [Google Scholar] [CrossRef] [PubMed]
  37. Katz, E.A.; Katz, S.B.; Freeman, M.D. Non-surgical management of upper cervical instability via improved cervical lordosis: A case series of adult patients. J. Clin. Med. 2023, 12, 1797. [Google Scholar] [CrossRef] [PubMed]
  38. Oakley, P.A.; Harrison, D.D.; Harrison, D.E.; Haas, J.W. Evidence-based protocol for structural rehabilitation of the spine and posture: Review of clinical biomechanics of posture (CBP) publications. J. Can. Chiropr. Assoc. 2005, 49, 270–296. [Google Scholar] [PubMed]
  39. Fortner, M.O.; Oakley, P.A.; Harrison, D.E. Non-surgical improvement of cervical lordosis is possible in advanced spinal osteoarthritis: A CBP® case report. J. Phys. Ther. Sci. 2018, 30, 108–112. [Google Scholar] [CrossRef] [PubMed]
  40. Oakley, P.A.; Harrison, D.E. Reducing thoracic hyperkyphosis subluxation deformity: A systematic review of chiropractic biophysics® methods employed in its structural improvement. J. Contemp. Chiropr. 2018, 1, 59–66. [Google Scholar]
  41. Harrison, D.D.; Harrison, D.E.; Janik, T.J.; Cailliet, R.; Ferrantelli, J.R.; Haas, J.W.; Holland, B. Modeling of the sagittal cervical spine as a method to discriminate hypo-lordosis: Results of elliptical and circular modeling in 72 asymptomatic subjects, 52 acute neck pain subjects, and 70 chronic neck pain subjects. Spine 2004, 29, 2485–2492. [Google Scholar] [CrossRef]
  42. Pivotto, L.R.; Navarro, I.J.R.L.; Candotti, C.T. Radiography and photogrammetry-based methods of assessing cervical spine posture in the sagittal plane: A systematic review with meta-analysis. Gait Posture 2021, 84, 357–367. [Google Scholar] [CrossRef] [PubMed]
  43. Ferracini, G.N.; Chaves, T.C.; Dach, F.; Bevilaqua-Grossi, D.; Fernández-de-Las-Peñas, C.; Speciali, J.G. Analysis of the cranio-cervical curvatures in subjects with migraine with and without neck pain. Physiotherapy 2017, 103, 392–399. [Google Scholar] [CrossRef]
  44. Harrison, D.E.; Holland, B.; Harrison, D.D.; Janik, T.J. Further reliability analysis of the Harrison radiographic line-drawing methods: Crossed ICCs for lateral posterior tangents and modified Risser-Ferguson method on AP views. J. Manip. Physiol. Ther. 2002, 25, 93–98. [Google Scholar] [CrossRef] [PubMed]
  45. Harrison, D.E.; Harrison, D.D.; Cailliet, R.; Troyanovich, S.J.; Janik, T.J.; Holland, B. Cobb method or Harrison posterior tangent method: Which to choose for lateral cervical radiographic analysis. Spine 2000, 25, 2072–2078. [Google Scholar] [CrossRef] [PubMed]
  46. Rydman, E.; Elkan, P.; Eneqvist, T.; Ekman, P.; Järnbert-Pettersson, H. The significance of cervical sagittal alignment for nonrecovery after whiplash injury. Spine J. 2020, 20, 1229–1238. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Four examples of model prediction (yellow points) versus human annotation (red points). The accuracy of the model in predicting the landmark points is comparable with the human annotator.
Figure 1. Four examples of model prediction (yellow points) versus human annotation (red points). The accuracy of the model in predicting the landmark points is comparable with the human annotator.
Jcm 13 02573 g001
Figure 2. Error analysis between human expert (US) and the model (CV) over three general features, including ARA, C1H, and TR. While x-axis shows the feature value calculated based on the expert’s annotation, y-axis determines the value based on the model’s prediction. The points on the line have zero errors.
Figure 2. Error analysis between human expert (US) and the model (CV) over three general features, including ARA, C1H, and TR. While x-axis shows the feature value calculated based on the expert’s annotation, y-axis determines the value based on the model’s prediction. The points on the line have zero errors.
Jcm 13 02573 g002
Figure 3. Error analysis between human expert (US) and the model (CV) over KA feature. While x-axis shows the feature value calculated based on the expert’s annotation, y-axis determines the value based on the model’s prediction. The points on the line have zero errors.
Figure 3. Error analysis between human expert (US) and the model (CV) over KA feature. While x-axis shows the feature value calculated based on the expert’s annotation, y-axis determines the value based on the model’s prediction. The points on the line have zero errors.
Jcm 13 02573 g003
Figure 4. Error analysis between human expert (US) and the model (CV) over RRA (RR) feature. While x-axis shows the feature value calculated based on the expert’s annotation, y-axis determines the value based on the model’s prediction. The points on the line have zero errors.
Figure 4. Error analysis between human expert (US) and the model (CV) over RRA (RR) feature. While x-axis shows the feature value calculated based on the expert’s annotation, y-axis determines the value based on the model’s prediction. The points on the line have zero errors.
Jcm 13 02573 g004
Figure 5. Error analysis between human expert (US) and the model (CV) over ST feature. While x-axis shows the feature value calculated based on the expert’s annotation, y-axis determines the value based on the model’s prediction. The points on the line have zero errors.
Figure 5. Error analysis between human expert (US) and the model (CV) over ST feature. While x-axis shows the feature value calculated based on the expert’s annotation, y-axis determines the value based on the model’s prediction. The points on the line have zero errors.
Jcm 13 02573 g005
Figure 6. Error on two experiments by an expert human (US). While the x-axis shows the feature value calculated based on the expert’s first experiment, the y-axis determines the feature value based on the expert’s second experiment. The points on the line have zero errors. See the Methods section for description of the variables.
Figure 6. Error on two experiments by an expert human (US). While the x-axis shows the feature value calculated based on the expert’s first experiment, the y-axis determines the feature value based on the expert’s second experiment. The points on the line have zero errors. See the Methods section for description of the variables.
Jcm 13 02573 g006
Figure 7. Diagram of the ICC, mean, and standard deviation of the error over different experiments. Comparing the diagrams related to US vs. CV and US vs. US shows that the error rate between these two experiments is very similar. Therefore, the model expectancy is the same as the human annotator. There is no difference between the model output in different experiments (CV vs. CV), so the error rate is zero, and ICC is maximum (ICC = 1).
Figure 7. Diagram of the ICC, mean, and standard deviation of the error over different experiments. Comparing the diagrams related to US vs. CV and US vs. US shows that the error rate between these two experiments is very similar. Therefore, the model expectancy is the same as the human annotator. There is no difference between the model output in different experiments (CV vs. CV), so the error rate is zero, and ICC is maximum (ICC = 1).
Jcm 13 02573 g007
Table 1. Details of different measures on various measurements reported. Mean: mean; Std: standard deviation of the error; RMSE: root mean squared error; ICC: intraclass correlation coefficients. Features (variables) include ARA: absolute rotation angle; C1H: atlas plane line to horizontal; KA: Cobb angle endplate lines from C2/C3 inclusive to C6/C7; RRA: posterior body tangent lines from C2/C3 inclusive to C6/C7; ST: intersegmental sagittal translation distance from C2/C3 inclusive to C6/C7; TR: global sagittal translation distance of C2 relative to C7.
Table 1. Details of different measures on various measurements reported. Mean: mean; Std: standard deviation of the error; RMSE: root mean squared error; ICC: intraclass correlation coefficients. Features (variables) include ARA: absolute rotation angle; C1H: atlas plane line to horizontal; KA: Cobb angle endplate lines from C2/C3 inclusive to C6/C7; RRA: posterior body tangent lines from C2/C3 inclusive to C6/C7; ST: intersegmental sagittal translation distance from C2/C3 inclusive to C6/C7; TR: global sagittal translation distance of C2 relative to C7.
Computer Vision (CV) vs. Expert Human (US)Expert Human (US) vs. Expert Human (US)
MeanStdRMSEICCR2MeanStdRMSEICCR2
ARA (°)1.130.0142.890.970.940.890.0091.970.980.97
C1H (°)1.480.0182.480.920.860.930.0121.660.970.95
KA23 (°)0.550.0151.480.960.930.590.0171.650.960.92
KA34 (°)0.930.0172.140.950.911.050.0192.460.940.89
KA45 (°)0.800.0171.990.950.911.000.0192.250.940.89
KA56 (°)0.850.0181.980.950.910.980.0202.200.940.89
KA67 (°)1.740.0374.180.840.641.590.0283.270.890.78
RRA23 (°)0.660.0221.760.930.860.750.0221.710.930.87
RRA34 (°)0.720.0211.690.940.880.830.0211.760.930.86
RRA45 (°)0.800.0232.000.930.860.950.0201.790.930.88
RRA56 (°)0.790.0222.110.930.860.940.0191.780.940.89
RRA67 (°)1.090.0332.530.860.700.970.0251.930.900.82
ST23 (mm)0.160.0230.370.920.840.230.0280.450.870.77
ST34 (mm)0.160.0250.410.910.820.220.0280.460.880.77
ST45 (mm)0.240.0310.480.840.700.240.0310.490.860.72
ST56 (mm)0.210.0310.460.860.730.220.0300.450.870.74
ST67 (mm)0.230.0390.520.800.570.220.0320.420.840.70
TR (mm)0.450.0030.960.990.990.440.0030.870.990.99
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hosseini, M.M.; Mahoor, M.H.; Haas, J.W.; Ferrantelli, J.R.; Dupuis, A.-L.; Jaeger, J.O.; Harrison, D.E. Intra-Examiner Reliability and Validity of Sagittal Cervical Spine Mensuration Methods Using Deep Convolutional Neural Networks. J. Clin. Med. 2024, 13, 2573. https://0-doi-org.brum.beds.ac.uk/10.3390/jcm13092573

AMA Style

Hosseini MM, Mahoor MH, Haas JW, Ferrantelli JR, Dupuis A-L, Jaeger JO, Harrison DE. Intra-Examiner Reliability and Validity of Sagittal Cervical Spine Mensuration Methods Using Deep Convolutional Neural Networks. Journal of Clinical Medicine. 2024; 13(9):2573. https://0-doi-org.brum.beds.ac.uk/10.3390/jcm13092573

Chicago/Turabian Style

Hosseini, Mohammad Mehdi, Mohammad H. Mahoor, Jason W. Haas, Joseph R. Ferrantelli, Anne-Lise Dupuis, Jason O. Jaeger, and Deed E. Harrison. 2024. "Intra-Examiner Reliability and Validity of Sagittal Cervical Spine Mensuration Methods Using Deep Convolutional Neural Networks" Journal of Clinical Medicine 13, no. 9: 2573. https://0-doi-org.brum.beds.ac.uk/10.3390/jcm13092573

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop