Vision-based Assessment of Joint Extensibility in Ehlers Danlos Syndrome
Assessing the Feasibility of a Smartphone-based, Machine Learning Visual Imaging Application for Assessment of Hyperextensibility of Peripheral Joints in Ehlers Danlos Syndrome
1 other identifier
observational
225
1 country
1
Brief Summary
Ehlers Danlos Syndrome (EDS) is a heterogenous group of genetic disorders with 13 identified subtypes. Hypermobile EDS (hEDS), although the most common subtype of EDS, does not yet have an identified genetic mutation for diagnostic confirmation. Generalized joint hypermobility (GJH) is one of the hallmark features of hEDS. The scoring system used in measurement of GJH was described by Beighton. The Beighton score is calculated using a dichotomous scoring system to assess the extensibility of nine joints. Each joint is scored as either hypermobile (score = 1) or not hypermobile (score = 0). The total score (Beighton score) can vary between a minimum of 0 and a maximum of 9, with higher scores indicating greater joint laxity. While there is moderate validity and inter-rater variability in using the Beighton score, there continue to be several challenges with its widespread and consistent application by clinicians. Some of the barriers reported in the literature include: i) In open, non-standardized systems there can be significant variation in the method to perform these joint extensibility tests including assessing baseline measurements, ii) Determining consistent and standard measurement tools/methodology e.g. goniometer use can vary widely iii) Assessing the reliability of the cut off values and, iv) Performing full assessment prior to informing patients of possible classification of GJH positivity (low specificity and low positive predictive). Inappropriate implementation of tests to assess GJH results in inaccurate identification of GJH and potentially unintended negative consequences of making the wrong diagnosis of EDS. The objective of this study is to create a more robust and valid method of joint mobility measurement and reduce error in the screening of EDS through use of a smartphone-based machine learning application systems for measurement of joint extensibility. The project will: i) Create a smart-phone enabled visual imaging app to assess the measurement of joint extensibility, ii) Assess the feasibility of using the smart-phone app in a clinical setting to screen potential EDS patients, iii) Determine the validity of the application in comparison to in person clinical assessment in a tertiary care academic EDS program. If successful, the smart-phone application could help standardize the care of potential EDS patients in an efficient and cost-effective manner.
Trial Health
Trial Health Score
Automated assessment based on enrollment pace, timeline, and geographic reach
participants targeted
Target at P75+ for all trials
Started Apr 2022
Typical duration for all trials
1 active site
Health score is calculated from publicly available data and should be used for screening purposes only.
Trial Relationships
Click on a node to explore related trials.
Study Timeline
Key milestones and dates
Study Start
First participant enrolled
April 26, 2022
CompletedFirst Submitted
Initial submission to the registry
May 4, 2022
CompletedFirst Posted
Study publicly available on registry
May 9, 2022
CompletedPrimary Completion
Last participant's last visit for primary outcome
June 1, 2024
CompletedStudy Completion
Last participant's last visit for all outcomes
December 1, 2024
CompletedNovember 28, 2023
November 1, 2023
2.1 years
May 4, 2022
November 27, 2023
Conditions
Keywords
Outcome Measures
Primary Outcomes (3)
Comparison of agreement in predicted angle by pose-estimation library
The performance of the developed machine learning models for predicting the range of motion will be analyzed by the pose-estimation library used. This analysis will be performed on the subset of the data collected during the first 2 months of data collection. This information will be used to select the pose-estimation libraries to proceed with when refining the machine learning models.
4 months
Comparison of agreement in predicted angle by joint
The performance of the developed machine learning models for predicting the range of motion at each joint (spine, knee, ankle, elbow, shoulder, thumb, fifth finger) will be analyzed independently for each joint. This will provide insight with respect to which joints the system is more accurate at predicting from video.
1 year
Assess the accuracy of range of motion prediction using vision-based data
Machine learning models trained on videos of individuals performing the joint hypermobility maneuvers will be developed. Their performance will be compared to the range of motion measured by an expert clinician using a goniometer.
1 year
Study Arms (1)
New patients at the GoodHope EDS clinic at Toronto General Hospital
All patients seen in the EDS clinic are eligible for inclusion, regardless of their presenting diagnosis or the results of their assessments.
Interventions
No intervention will be used. Consenting participants will have video recordings taken during their exam of joint hypermobility which will be analyzed at a later time
Eligibility Criteria
The population being studied includes all patients referred to or seen in the GoodHope EDS clinic. The clinic accepts referrals from symptomatic adult patients (age \> 18 years), with EDS, or suspected EDS. EDS is a connective tissue disorder with 100% penetrance, but variable in phenotypic expression, suspected cases of EDS or G-HSD may therefore include other hereditary or acquired connective tissue diseases/disorder, and/or complex chronic illnesses characterized by, or that feature, joint hypermobility, pain, and fatigue.
You may not qualify if:
- Patients who do not consent to participate will not be included (participants may withdraw consent at any time)
Contact the study team to confirm eligibility.
Sponsors & Collaborators
Study Sites (1)
GoodHope EDS - Toronto General Hospital
Toronto, Ontario, M5G 2C4, Canada
Related Publications (21)
Critical Care Services Ontario, Ehlers-Danlos Syndrome Expert Panel Report, 2016. https://www.health.gov.on.ca/en/common/ministry/publications/reports/eds/Default.aspx.
BACKGROUNDCahill SV, Sharkey MS, Carter CW. Clinical assessment of generalized ligamentous laxity using a single test: is thumb-to-forearm apposition enough? J Pediatr Orthop B. 2021 May 1;30(3):296-300. doi: 10.1097/BPB.0000000000000732.
PMID: 32301823BACKGROUNDHe K, Gkioxari G, Dollár P, Girshick R. Mask R-CNN. In: Proceedings of the IEEE international conference on computer vision. 2017. p. 2961-9.
BACKGROUNDCao Z, Hidalgo G, Simon T, Wei SE, Sheikh Y. OpenPose: Realtime Multi-Person 2D Pose Estimation Using Part Affinity Fields. IEEE Trans Pattern Anal Mach Intell. 2021 Jan;43(1):172-186. doi: 10.1109/TPAMI.2019.2929257. Epub 2020 Dec 4.
PMID: 31331883BACKGROUNDFang H-S, Xie S, Tai Y-W, Lu C. RMPE: Regional Multi-person Pose Estimation. 2016 Nov 30; Available from: http://arxiv.org/abs/1612.00137
BACKGROUNDLin T-Y, Maire M, Belongie S, Bourdev L, Girshick R, Hays J, et al. Microsoft COCO: Common Objects in Context. 2014 May 1; Available from: http://arxiv.org/abs/1405.0312
BACKGROUNDAndriluka M, Pishchulin L, Gehler P, Schiele B. 2D Human Pose Estimation: New Benchmark and State of the Art Analysis. In: 2014 IEEE Conference on Computer Vision and Pattern Recognition. 2014. p. 3686-93
BACKGROUNDLugaresi C, Tang J, Nash H, McClanahan C, Uboweja E, Hays M, et al. MediaPipe: A Framework for Building Perception Pipelines. 2019 Jun 14; Available from: https://arxiv.org/abs/1906.08172
BACKGROUNDZhang F, Bazarevsky V, Vakunov A, Tkachenka A, Sung G, Chang C-L, et al. MediaPipe Hands: On-device Real-time Hand Tracking. 2020 Jun 17; Available from: http://arxiv.org/abs/2006.10214
BACKGROUNDMehdizadeh S, Nabavi H, Sabo A, Arora T, Iaboni A, Taati B. Concurrent validity of human pose tracking in video for measuring gait parameters in older adults: a preliminary analysis with multiple trackers, viewing angles, and walking directions. J Neuroeng Rehabil. 2021 Sep 15;18(1):139. doi: 10.1186/s12984-021-00933-0.
PMID: 34526074BACKGROUNDSabo A, Mehdizadeh S, Ng KD, Iaboni A, Taati B. Assessment of Parkinsonian gait in older adults with dementia via human pose tracking in video data. J Neuroeng Rehabil. 2020 Jul 14;17(1):97. doi: 10.1186/s12984-020-00728-9.
PMID: 32664973BACKGROUNDLu M, Zhao Q, Poston KL, Sullivan EV, Pfefferbaum A, Shahid M, Katz M, Montaser-Kouhsari L, Schulman K, Milstein A, Niebles JC, Henderson VW, Fei-Fei L, Pohl KM, Adeli E. Quantifying Parkinson's disease motor severity under uncertainty using MDS-UPDRS videos. Med Image Anal. 2021 Oct;73:102179. doi: 10.1016/j.media.2021.102179. Epub 2021 Jul 21.
PMID: 34340101BACKGROUNDWilliams S, Zhao Z, Hafeez A, Wong DC, Relton SD, Fang H, Alty JE. The discerning eye of computer vision: Can it measure Parkinson's finger tap bradykinesia? J Neurol Sci. 2020 Sep 15;416:117003. doi: 10.1016/j.jns.2020.117003. Epub 2020 Jun 30.
PMID: 32645513BACKGROUNDOta M, Tateuchi H, Hashiguchi T, Kato T, Ogino Y, Yamagata M, Ichihashi N. Verification of reliability and validity of motion analysis systems during bilateral squat using human pose tracking algorithm. Gait Posture. 2020 Jul;80:62-67. doi: 10.1016/j.gaitpost.2020.05.027. Epub 2020 May 25.
PMID: 32485426BACKGROUNDSlembrouck M, Luong H, Gerlo J, Schütte K, Van Cauwelaert D, De Clercq D, et al. Multiview 3d Markerless Human Pose Estimation from Openpose Skeletons. In: International Conference on Advanced Concepts for Intelligent Vision Systems. Springer; 2020. p. 166-78.
BACKGROUNDWang H, Xie Z, Lu L, Li L, Xu X. A computer-vision method to estimate joint angles and L5/S1 moments during lifting tasks through a single camera. J Biomech. 2021 Dec 2;129:110860. doi: 10.1016/j.jbiomech.2021.110860. Epub 2021 Nov 8.
PMID: 34794041BACKGROUNDYahya M, Shah JA, Warsi A, Kadir K, Khan S, Izani M. Real time elbow angle estimation using single RGB camera. 2018 Aug 21; Available from: https://arxiv.org/abs/1808.07017
BACKGROUNDShi B, Brentari D, Shakhnarovich G, Livescu K. Fingerspelling Detection in American Sign Language. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2021. p. 4166-75
BACKGROUNDKim I-H, Jung I-H. A Study on Korea Sign Language Motion Recognition Using OpenPose Based on Deep Learning. 디지털콘텐츠학회논문지 (Journal of Digital Contents Society). 2021;22(4):681-7.
BACKGROUNDHakim A. Hypermobile Ehlers-Danlos Syndrome. 2004 Oct 22 [updated 2024 Feb 22]. In: Adam MP, Bick S, Mirzaa GM, Pagon RA, Wallace SE, Amemiya A, editors. GeneReviews(R) [Internet]. Seattle (WA): University of Washington, Seattle; 1993-2026. Available from http://www.ncbi.nlm.nih.gov/books/NBK1279/
PMID: 20301456BACKGROUNDMittal N, Sabo A, Deshpande A, Clarke H, Taati B. Feasibility of video-based joint hypermobility assessment in individuals with suspected Ehlers-Danlos syndromes/generalised hypermobility spectrum disorders: a single-site observational study protocol. BMJ Open. 2022 Dec 16;12(12):e068098. doi: 10.1136/bmjopen-2022-068098.
PMID: 36526308DERIVED
MeSH Terms
Conditions
Condition Hierarchy (Ancestors)
Study Officials
- PRINCIPAL INVESTIGATOR
Nimish Mittal, MD
GoodHope Ehlers Danlos Syndrome Clinic, Toronto General Hospital
Study Design
- Study Type
- observational
- Observational Model
- CASE ONLY
- Time Perspective
- CROSS SECTIONAL
- Sponsor Type
- OTHER
- Responsible Party
- PRINCIPAL INVESTIGATOR
- PI Title
- Medical Director - GoodHope Ehlers Danlos Syndrome Clinic
Study Record Dates
First Submitted
May 4, 2022
First Posted
May 9, 2022
Study Start
April 26, 2022
Primary Completion
June 1, 2024
Study Completion
December 1, 2024
Last Updated
November 28, 2023
Record last verified: 2023-11
Data Sharing
- IPD Sharing
- Will not share
No IPD will be shared with other researchers.