Autoencoder as a New Method for Maintaining Data Privacy While Analyzing Videos of Patients With Motor Dysfunction: Proof-of-Concept Study

Abstract
Journal of Medical Internet Research - International Scientific Journal for Medical Research, Information and Communication on the Internet #Preprint #PeerReviewMe: Warning: This is a unreviewed preprint. Readers are warned that the document has not been peer-reviewed by expert/patient reviewers or an academic editor, may contain misleading claims, and is likely to undergo changes before final publication, if accepted, or may have been rejected/withdrawn. Readers with interest and expertise are encouraged to sign up as peer-reviewer, if the paper is within an open peer-review period. Please cite this preprint only for review purposes or for grant applications and CVs (if you are the author). Background: In chronical neurological diseases, especially in multiple sclerosis (MS), clinical assessment of motor dysfunction is crucial to monitor patient’s disease. Traditional scales such as the Expanded Disability Status Scale (EDSS) are not sensitive enough to detect slight changes in motor performance. Video recordings of patient performance are more accurate and increase reliability in severity ratings. They allow an automated, quantitative, machine learning algorithms-based analysis of patient’s motor performance. Creation of these algorithms usually involves non-healthcare professionals, which is a challenge regarding data-privacy. Autoencoders embed visual information into a lower-dimensional latent space which preserves information needed for algorithm development but is not visually interpretable by humans. They consist of an encoder that creates encodes videos (creating a sequence of coded frame vectors) and a paired decoder that transforms the coded frame vectors into the original video. Videos encoded in this way can be shared with non-medical collaborators. Objective: The aim of this proof of concept study was to test whether coded frame vectors of autoencoders contain relevant information to analyse videos of patient movements whilst preserving data privacy. Methods: In this study, 20 pre-rated videos of patients performing the finger-to-nose test were recorded. An autoencoder created encoded frame vectors from the original videos and decoded the videos again. Original and decoded videos were shown to 10 neurologists of an academic MS centre in Basel, Switzerland. Neurologists tested whether these 200 videos in total were human-readable and rated the severity grade of each video according to the Neurostatus-EDSS definitions of limb ataxia. Furthermore, the neurologists tested whether ratings were equivalent between original and decoded videos. Results: From 200 presented videos, and after decoding of the video data, 172 (86%) were evidently rated. The intra-rater agreement between the original and decoded videos was 0.317 (kappa = Cohen’s weighted kappa), with an average difference of 0.26 (original is rated as more severe). The inter-rater agreement before coding of the videos was 0.459 (kappa) and 0.302 (kappa) after decoding. Conclusions: Although larger studies and a further improvement of the autoencoder are needed, this proof of concept study is a first step for a promising method that enables the use of patient videos while preserving data privacy, especially when non-healthcare professionals are involved. Our findings emphasise that autoencoder provides a similar level of security to normal encryption - assuming that the decoder is not shared – especially for the use of automated machine-learning algorithm-based analysis of patient videos.