中文版 | English
题名

A Speech-Level-Based Segmented Model to Decode the Dynamic Auditory Attention States in the Competing Speaker Scenes

作者
通讯作者Chen, Fei
发表日期
2022-02-10
DOI
发表期刊
EISSN
1662-453X
卷号15
摘要
In the competing speaker environments, human listeners need to focus or switch their auditory attention according to dynamic intentions. The reliable cortical tracking ability to the speech envelope is an effective feature for decoding the target speech from the neural signals. Moreover, previous studies revealed that the root mean square (RMS)-level-based speech segmentation made a great contribution to the target speech perception with the modulation of sustained auditory attention. This study further investigated the effect of the RMS-level-based speech segmentation on the auditory attention decoding (AAD) performance with both sustained and switched attention in the competing speaker auditory scenes. Objective biomarkers derived from the cortical activities were also developed to index the dynamic auditory attention states. In the current study, subjects were asked to concentrate or switch their attention between two competing speaker streams. The neural responses to the higher- and lower-RMS-level speech segments were analyzed via the linear temporal response function (TRF) before and after the attention switching from one to the other speaker stream. Furthermore, the AAD performance decoded by the unified TRF decoding model was compared to that by the speech-RMS-level-based segmented decoding model with the dynamic change of the auditory attention states. The results showed that the weight of the typical TRF component approximately 100-ms time lag was sensitive to the switching of the auditory attention. Compared to the unified AAD model, the segmented AAD model improved attention decoding performance under both the sustained and switched auditory attention modulations in a wide range of signal-to-masker ratios (SMRs). In the competing speaker scenes, the TRF weight and AAD accuracy could be used as effective indicators to detect the changes of the auditory attention. In addition, with a wide range of SMRs (i.e., from 6 to -6 dB in this study), the segmented AAD model showed the robust decoding performance even with short decision window length, suggesting that this speech-RMS-level-based model has the potential to decode dynamic attention states in the realistic auditory scenarios.
关键词
相关链接[来源记录]
收录类别
语种
英语
学校署名
第一 ; 通讯
资助项目
National Natural Science Foundation of China[61971212] ; Shenzhen Sustainable Support Program for High-level University[20200925154002001,"G02236002"]
WOS研究方向
Neurosciences & Neurology
WOS类目
Neurosciences
WOS记录号
WOS:000761043500001
出版者
来源库
Web of Science
引用统计
被引频次[WOS]:1
成果类型期刊论文
条目标识符http://sustech.caswiz.com/handle/2SGJ60CL/291074
专题工学院_电子与电气工程系
作者单位
1.Southern Univ Sci & Technol, Dept Elect & Elect Engn, Shenzhen, Peoples R China
2.Univ Hong Kong, Dept Elect & Elect Engn, Pokfulam, Hong Kong, Peoples R China
第一作者单位电子与电气工程系
通讯作者单位电子与电气工程系
第一作者的第一单位电子与电气工程系
推荐引用方式
GB/T 7714
Wang, Lei,Wang, Yihan,Liu, Zhixing,et al. A Speech-Level-Based Segmented Model to Decode the Dynamic Auditory Attention States in the Competing Speaker Scenes[J]. FRONTIERS IN NEUROSCIENCE,2022,15.
APA
Wang, Lei,Wang, Yihan,Liu, Zhixing,Wu, Ed X.,&Chen, Fei.(2022).A Speech-Level-Based Segmented Model to Decode the Dynamic Auditory Attention States in the Competing Speaker Scenes.FRONTIERS IN NEUROSCIENCE,15.
MLA
Wang, Lei,et al."A Speech-Level-Based Segmented Model to Decode the Dynamic Auditory Attention States in the Competing Speaker Scenes".FRONTIERS IN NEUROSCIENCE 15(2022).
条目包含的文件
条目无相关文件。
个性服务
原文链接
推荐该条目
保存到收藏夹
查看访问统计
导出为Endnote文件
导出为Excel格式
导出为Csv格式
Altmetrics Score
谷歌学术
谷歌学术中相似的文章
[Wang, Lei]的文章
[Wang, Yihan]的文章
[Liu, Zhixing]的文章
百度学术
百度学术中相似的文章
[Wang, Lei]的文章
[Wang, Yihan]的文章
[Liu, Zhixing]的文章
必应学术
必应学术中相似的文章
[Wang, Lei]的文章
[Wang, Yihan]的文章
[Liu, Zhixing]的文章
相关权益政策
暂无数据
收藏/分享
所有评论 (0)
[发表评论/异议/意见]
暂无评论

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。