题名 | EEG-based auditory attention decoding using speech-level-based segmented computational models |
作者 | |
通讯作者 | Chen, Fei |
发表日期 | 2021-08-01
|
DOI | |
发表期刊 | |
ISSN | 1741-2560
|
EISSN | 1741-2552
|
卷号 | 18期号:4 |
摘要 | Objective. Auditory attention in complex scenarios can be decoded by electroencephalography (EEG)-based cortical speech-envelope tracking. The relative root-mean-square (RMS) intensity is a valuable cue for the decomposition of speech into distinct characteristic segments. To improve auditory attention decoding (AAD) performance, this work proposed a novel segmented AAD approach to decode target speech envelopes from different RMS-level-based speech segments. Approach. Speech was decomposed into higher- and lower-RMS-level speech segments with a threshold of -10 dB relative RMS level. A support vector machine classifier was designed to identify higher- and lower-RMS-level speech segments, using clean target and mixed speech as reference signals based on corresponding EEG signals recorded when subjects listened to target auditory streams in competing two-speaker auditory scenes. Segmented computational models were developed with the classification results of higher- and lower-RMS-level speech segments. Speech envelopes were reconstructed based on segmented decoding models for either higher- or lower-RMS-level speech segments. AAD accuracies were calculated according to the correlations between actual and reconstructed speech envelopes. The performance of the proposed segmented AAD computational model was compared to those of traditional AAD methods with unified decoding functions. Main results. Higher- and lower-RMS-level speech segments in continuous sentences could be identified robustly with classification accuracies that approximated or exceeded 80% based on corresponding EEG signals at 6 dB, 3 dB, 0 dB, -3 dB and -6 dB signal-to-mask ratios (SMRs). Compared with unified AAD decoding methods, the proposed segmented AAD approach achieved more accurate results in the reconstruction of target speech envelopes and in the detection of attentional directions. Moreover, the proposed segmented decoding method had higher information transfer rates (ITRs) and shorter minimum expected switch times compared with the unified decoder. Significance. This study revealed that EEG signals may be used to classify higher- and lower-RMS-level-based speech segments across a wide range of SMR conditions (from 6 dB to -6 dB). A novel finding was that the specific information in different RMS-level-based speech segments facilitated EEG-based decoding of auditory attention. The significantly improved AAD accuracies and ITRs of the segmented decoding method suggests that this proposed computational model may be an effective method for the application of neuro-controlled brain-computer interfaces in complex auditory scenes. |
关键词 | |
相关链接 | [来源记录] |
收录类别 | |
语种 | 英语
|
学校署名 | 第一
; 通讯
|
资助项目 | National Natural Science Foundation of China[61971212]
; Basic Research Foundation of Shenzhen["KQJSCX20180319114453986","GJHZ20180928155002157"]
; High-level University Fund of Southern University of Science and Technology[G02236002]
|
WOS研究方向 | Engineering
; Neurosciences & Neurology
|
WOS类目 | Engineering, Biomedical
; Neurosciences
|
WOS记录号 | WOS:000655389700001
|
出版者 | |
EI入藏号 | 20212310450283
|
EI主题词 | Biomedical signal processing
; Brain computer interface
; Computation theory
; Computational methods
; Decoding
; Electroencephalography
; Electrophysiology
; Speech
; Speech analysis
; Speech communication
; Support vector machines
|
EI分类号 | Biomedical Engineering:461.1
; Medicine and Pharmacology:461.6
; Information Theory and Signal Processing:716.1
; Computer Theory, Includes Formal Logic, Automata Theory, Switching Theory, Programming Theory:721.1
; Computer Peripheral Equipment:722.2
; Computer Software, Data Handling and Applications:723
; Data Processing and Image Processing:723.2
; Speech:751.5
|
来源库 | Web of Science
|
引用统计 |
被引频次[WOS]:5
|
成果类型 | 期刊论文 |
条目标识符 | http://sustech.caswiz.com/handle/2SGJ60CL/229420 |
专题 | 工学院_电子与电气工程系 |
作者单位 | 1.Southern Univ Sci & Technol, Dept Elect & Elect Engn, Shenzhen, Peoples R China 2.Univ Hong Kong, Dept Elect & Elect Engn, Hong Kong, Peoples R China |
第一作者单位 | 电子与电气工程系 |
通讯作者单位 | 电子与电气工程系 |
第一作者的第一单位 | 电子与电气工程系 |
推荐引用方式 GB/T 7714 |
Wang, Lei,Wu, Ed X.,Chen, Fei. EEG-based auditory attention decoding using speech-level-based segmented computational models[J]. Journal of Neural Engineering,2021,18(4).
|
APA |
Wang, Lei,Wu, Ed X.,&Chen, Fei.(2021).EEG-based auditory attention decoding using speech-level-based segmented computational models.Journal of Neural Engineering,18(4).
|
MLA |
Wang, Lei,et al."EEG-based auditory attention decoding using speech-level-based segmented computational models".Journal of Neural Engineering 18.4(2021).
|
条目包含的文件 | 条目无相关文件。 |
|
除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。
修改评论