Dizziness affects around 7.4% of adults, presenting diagnostic challenges, particularly in early differentiation of benign (inner ear disease) vs dangerous (stroke) causes. The ability to detect subtle changes in the otolith can provide clues regarding central and peripheral vestibular function. The degree and direction of dynamic ocular torsion obtained using the beside vOCR test is a quantitative measure of vestibular function. Recent advances in deep learning-based dynamic torsional detection may make it possible to automate the vOCR, increasing accessibility to non-expert providers on the frontline.
36 participants were enrolled, including 18 healthy controls and 18 with vestibular loss. Each participant underwent the vOCR test, involving 24 tilts consisting of both neck and trunk movements to stimulate otolith function. Torsional waveforms were recorded, and beats of torsion, each consisting of a fast and slow phase, were extracted. Clips of 500ms, containing at least one full beat of torsion, were isolated. Several Machine Learning Models (MLMs) were trained on the extracted waveform features. For image-based analysis, a filtered image was generated by summing the differences between consecutive video frames, creating a visualization that captured the dynamic motion in each clip. A simplified 2D ResNet18 model was then trained on these filtered images.