Abstract
Auditory attention is a cognitive process that involves listeners to give their attention, which the listener wants to focus on. Consequently, the auditory attention processes their stimulus of interest and decodes it as a response of individuals. This study is aimed to decode the listener’s auditory attention state at continuous hearing complications in different (e.g., by quiet and hypothetical noisy background) environments. There have numerous ways to decode the auditory attention state. Here, we approach the difficulty of auditory attention to the auditory oddball paradigm based on electroencephalogram (EEG) signal. We have carried two experiments and having 840 trials for each listening state (attended and unattended). We have applied a stimulus for the attention detection model and it was verified for decoding the auditory listening attention successfully to listening. Finally, we present the decoding state of acoustic listening attention from electroencephalogram (EEG) signals through a bidirectional long short-term memory (bLSTM) and a support vector machine (SVM) model. The performance of the two models has been compared, and it reveals that these two models perform at a significant accuracy to detect the auditory listening attention state. In the frontal lobe, which is associated with attention, the SVM model has achieved a statistically significant accuracy of 74% at a 50 ms time window.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Wild, C.J., Yusuf, A., Wilson, D.E., Peelle, J.E., Davis, M.H., Johnsrude, I.S.: Effortful listening: the processing of degraded speech depends critically on attention. J. Neurosci. 32(40), 14010–14021 (2012)
Powell, P.S., Strunk, J., James, T., Polyn, S.M., Duarte, A.: Decoding selective attention to context memory: an aging study. Neuroimage 181, 95–107 (2018)
Landry, M., Lifshitz, M., Raz, A.: Brain correlates of hypnosis: a systematic review and meta-analytic exploration. Neurosci. Biobehav. Rev. 81(A), 75–98 (2017)
Todorovic, A., Schoffelen, J.M., van Ede, F., Maris, E., de Lange, F.P.: Temporal expectation and attention jointly modulate auditory oscillatory activity in the beta band. PLoS ONE 10(3), e0120288 (2015)
Morash, V., Bai, O., Furlani, S., Lin, P., Hallett, M.: Classifying EEG signals preceding right hand, left hand, tongue, and right foot movements and motor imageries. Clin. Neurophysiol. 119, 2570–2578 (2008)
Sumit, S.S., Watada, J., Nasrin, F., Ahmed, N.I., Rambli, D.R.A.: Imputing missing values: reinforcement bayesian regression and random forest. In: Kreinovich, V., Hoang Phuong, N. (eds.) Soft Computing for Biomedical Applications and Related Topics. Studies in Computational Intelligence, vol. 899. Springer, Cham (2021)
Mahmud, M., Shamim Kaiser, M., Hussain, A.: Deep Learning in Mining Biological Data (2020). arXiv preprint arXiv:2003.00108v1
Strauss, D.J., Francis, A.L.: Toward a taxonomic model of attention in effortful lis-tening. Cogn. Affect. Behav. Neurosci. 17(4), 809–825 (2017)
De Pascalis, V., Bellusci, A., Gallo, C., Magurano, M.R., Chen, A.C.: Pain reduction strategies in hypnotic context and hypnosis: ERPs and SCRs during a secondary auditory task. Int. J. Clin. Exp. Hypn. 52(4), 343–363 (2004)
Yann Cojan, D., Piguet, C., Vuilleumier, P.: What makes your brain suggestible? Hypnotizability is associated with differential brain activity during attention out-side hypnosis. Neuroimage 117, 367–374 (2015)
Gruzelier, J., Gray, M., Horn, P.: The involvement of frontally modulated attention in hypnosis and hypnotic susceptibility: cortical evoked potential evidence. Contemp. Hypn. 19(4), 179–189 (2002)
Kirenskaya, A.V., Storozheva, Z.I., Solntseva, S.V., Novototsky-Vlasov, V.Y., Gordeev, M.N.: Auditory evoked potentials evidence for differences in information processing between high and low hypnotizable subjects. Int. J. Clin. Exp. Hypn. 67(1), 81–103 (2019)
Wong, D.D.E., Fuglsang, S.A., Hjortkjær, J., Ceolini, E., Slaney, M., de Cheveigné, A.: A comparison of regularization methods in forward and backward models for auditory attention decoding. Front. Neurosci. 12, 531 (2018)
Haghighi, M., Moghadamfalahi, M., Akcakaya, M., Erdogmus, D.: EEG-assisted modulation of sound sources in the auditory scene. Biomed. Signal Process. Control 39, 263–270 (2018)
Siripornpanich, V., Rachiwong, S., Ajjimaporn, A.: A pilot study on salivary cortisol secretion and auditory P300 event-related potential in patients with physical disability-related stress. Int. J. Neurosci. 130(2), 170–175 (2020)
Bednar, A., Lalor, E.C.: Where is the cocktail party? Decoding locations of attended and unattended moving sound sources using EEG. Neuroimage 205 (2020)
Brown, K.J., Gonsalvez, C.J., Harris, A.W., Williams, L.M., Gordon, E.: Target and non-target ERP disturbances in first episode vs. chronic schizophrenia. Clin. Neurophysiol. 113(11), 1754–1763 (2002)
Ciccarelli, G., Nolan, M., Perricone, J., et al.: Comparison of two-talker attention decoding from EEG with nonlinear neural networks and linear methods. Sci. Rep. 9(1), 11538 (2019)
Deckers, L., Das, N., Hossein Ansari, A., Bertrand, A., Francart, T.: EEG-based detection of the attended speaker and the locus of auditory attention with convolutional neural networks (2018). bioRxiv preprint bioRxiv: 475673
Lu, J., Yan, H., Chang, C., Wang, N.: Comparison of machine learning and deep learning approaches for decoding brain computer interface: an fNIRS Study. In: Shi, Z., Vadera, S., Chang, E. (eds.) Intelligent Information Processing X. IIP 2020. IFIP Advances in Information and Communication Technology, vol. 581. Springer, Cham (2020)
Skomrock, N.D., et al.: A characterization of brain-computer interface performance trade-offs using support vector machines and deep neural networks to decode movement intent. Front. Neurosci. 12, 763 (2018)
Mahmud, M., Kaiser, M.S., Rahman, M.M., Rahman, M.A., Shabut, A., Al-Mamun, S., Hussain, A.: A brain-inspired trust management model to assure security in a cloud based IoT framework for neuroscience applications. Cogn. Comput. 10(5), 864–873 (2018)
Jasper, H.: Report of the committee on methods of clinical examination in electroen-cephalography. Electroencephalogr. Clin. Neurophysiol. 10, 370–375 (1958)
Delorme, A., Makeig, S.: EEGLAB: an open source toolbox for analysis of single-trial EEG dynamics including independent component analysis. J. Neurosci. Methods 134, 9–21 (2004)
Graves, A., Schmidhuber, J.: Framewise phoneme classification with bidirectional LSTM and other neural network architectures. Neural Networks 18(5–6), 602–610 (2005)
Chang, C.C., Lin, C.J.: LIBSVM: a library for support vector machines. ACM Trans. Intell. Syst. Technol. 2(3), 1–27 (2011)
Rahman, M.A.: Gaussian process in computational biology: covariance functions for transcriptomics. Ph.D. Thesis, The University of Sheffield, UK (2018)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Nasrin, F., Ahmed, N.I., Rahman, M.A. (2021). Auditory Attention State Decoding for the Quiet and Hypothetical Environment: A Comparison Between bLSTM and SVM. In: Kaiser, M.S., Bandyopadhyay, A., Mahmud, M., Ray, K. (eds) Proceedings of International Conference on Trends in Computational and Cognitive Engineering. Advances in Intelligent Systems and Computing, vol 1309. Springer, Singapore. https://doi.org/10.1007/978-981-33-4673-4_23
Download citation
DOI: https://doi.org/10.1007/978-981-33-4673-4_23
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-33-4672-7
Online ISBN: 978-981-33-4673-4
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)