Department of Otolaryngology, Head and Neck Surgery, Beijing Tsinghua Changgung Hospital, School of Clinical Medicine, Tsinghua University, Beijing, China.
Institute of Integrated Circuit, Tsinghua University, Beijing, China.
Otolaryngol Head Neck Surg. 2024 Oct;171(4):1165-1171. doi: 10.1002/ohn.840. Epub 2024 Jun 1.
Recognition of auditory brainstem response (ABR) waveforms may be challenging, particularly for older individuals or those with hearing loss. This study aimed to investigate deep learning frameworks to improve the automatic recognition of ABR waveforms in participants with varying ages and hearing levels.
The research used a descriptive study design to collect and analyze pure tone audiometry and ABR data from 100 participants.
The research was conducted at a tertiary academic medical center, specifically at the Clinical Audiology Center of Tsinghua Chang Gung Hospital (Beijing, China).
Data from 100 participants were collected and categorized into four groups based on age and hearing level. Features from both time-domain and frequency-domain ABR signals were extracted and combined with demographic factors, such as age, sex, pure-tone thresholds, stimulus intensity, and original signal sequences to generate feature vectors. An enhanced Wide&Deep model was utilized, incorporating the Light-multi-layer perceptron (MLP) model to train the recognition of ABR waveforms. The recognition accuracy (ACC) of each model was calculated for the overall data set and each group.
The ACC rates of the Light-MLP model were 97.8%, 97.2%, 93.8%, and 92.0% for Groups 1 to 4, respectively, with a weighted average ACC rate of 95.4%. For the Wide&Deep model, the ACC rates were 93.4%, 90.8%, 92.0%, and 88.3% for Groups 1 to 4, respectively, with a weighted average ACC rate of 91.0%.
Both the Light-MLP model and the Wide&Deep model demonstrated excellent ACC in automatic recognition of ABR waveforms across participants with diverse ages and hearing levels. While the Wide&Deep model's performance was slightly poorer than that of the Light-MLP model, particularly due to the limited sample size, it is anticipated that with an expanded data set, the performance of Wide&Deep model may be further improved.
识别听觉脑干反应(ABR)波形可能具有挑战性,特别是对于年龄较大或有听力损失的个体。本研究旨在调查深度学习框架,以提高对不同年龄和听力水平的参与者自动识别 ABR 波形的能力。
本研究采用描述性研究设计,从 100 名参与者中收集和分析纯音测听和 ABR 数据。
研究在一家三级学术医学中心进行,具体地点是清华大学长庚医院临床听力学中心(北京,中国)。
从 100 名参与者中收集数据,并根据年龄和听力水平将其分为四组。从时域和频域 ABR 信号中提取特征,并结合人口统计学因素,如年龄、性别、纯音阈值、刺激强度和原始信号序列,生成特征向量。使用增强型 Wide&Deep 模型,结合轻量级多层感知机(MLP)模型,对 ABR 波形的识别进行训练。计算每个模型在整个数据集和每个组中的识别准确率(ACC)。
Light-MLP 模型在第 1 至 4 组的 ACC 率分别为 97.8%、97.2%、93.8%和 92.0%,加权平均 ACC 率为 95.4%。对于 Wide&Deep 模型,在第 1 至 4 组的 ACC 率分别为 93.4%、90.8%、92.0%和 88.3%,加权平均 ACC 率为 91.0%。
Light-MLP 模型和 Wide&Deep 模型在识别不同年龄和听力水平的参与者的 ABR 波形方面均表现出出色的 ACC。虽然 Wide&Deep 模型的性能略逊于 Light-MLP 模型,特别是由于样本量有限,但预计随着数据集的扩大,Wide&Deep 模型的性能可能会进一步提高。