Hostname: page-component-f554764f5-fnl2l Total loading time: 0 Render date: 2025-04-20T18:02:28.347Z Has data issue: false hasContentIssue false

361 Automated assessment of facial nerve function using multimodal machine learning

Published online by Cambridge University Press:  11 April 2025

Oren Wei
Affiliation:
Johns Hopkins University School of Medicine
Diana Lopez
Affiliation:
Johns Hopkins University School of Medicine, Departmentof Otolaryngology-Head & Neck Surgery
Ioan Lina
Affiliation:
Vanderbilt University Medical Center, Department of Otolaryngology-Head & Neck Surgery
Kofi Boahene
Affiliation:
Johns Hopkins University School of Medicine, Departmentof Otolaryngology-Head & Neck Surgery
Rights & Permissions [Opens in a new window]

Abstract

Core share and HTML view are not available for this content. However, as you have access to this content, a full PDF is available via the ‘Save PDF’ action button.

Objectives/Goals: Current popular scoring systems for evaluating facial nerve function are subjective and imprecise. This study aims to quantify speech and facial motor changes in patients suffering from facial palsy after cerebellopontine angle (CPA) tumor resection to lay the foundation for a scoring algorithm that is higher resolution and more objective. Methods/Study Population: We will obtain audio and video recordings from 20 adult patients prior to and after surgical resection of unilateral CPA tumors between October 2024 and February 2025. We will obtain preoperative recordings within two weeks prior to surgery and postoperative recordings following a preset schedule starting from the day after surgery up to one year. Audio recordings entail patient readings of standardized passages and phonations while video recordings entail patient performance of standardized facial expressions. We will analyze video data for key distance measurements, such as eye opening and wrinkle size, using DynaFace. We will process audio data using VoiceLab to extract metrics such as prominence and tonality. We will perform statistical tests such as t-tests and ANOVA to elucidate changes across time. Results/Anticipated Results: I expect to obtain 9 sets of audio and video recordings from each of the 20 participants. In terms of speech, I expect average speech duration to increase postoperatively. Similarly, I expect to find increases in time spent breathing, number of breaths taken, and mean breathing duration. In terms of facial movement, I expect nasolabial fold length to decrease postoperatively, as well as eye opening size and left-right symmetry at rest. For both audio and video, I expect these changes to revert towards their preoperative baseline as time passes. I also expect average House-Brackmann and Sunnybrook facial grading scores to increase postoperatively and then decrease with time, correlating strongly with the video and audio findings. I will use trajectory analysis and time point matching to handle any missing data. Discussion/Significance of Impact: This study will validate our analysis platform’s ability to automatically quantify measurable changes that occur to speech and facial movement which correlate strongly with existing scoring systems. Future work will synthesize these data streams to move towards establishing biomarkers for facial nerve function that aid clinical decision-making.

Type
Informatics, AI and Data Science
Creative Commons
Creative Common License - CCCreative Common License - BYCreative Common License - NCCreative Common License - ND
This is an Open Access article, distributed under the terms of the Creative Commons Attribution-NonCommercial-NoDerivatives licence (https://creativecommons.org/licenses/by-nc-nd/4.0/), which permits non-commercial re-use, distribution, and reproduction in any medium, provided the original work is unaltered and is properly cited. The written permission of Cambridge University Press must be obtained for commercial re-use or in order to create a derivative work.
Copyright
© The Author(s), 2025. The Association for Clinical and Translational Science