Abstract
In this study, the visual user interface elements and the auditory user interface information elements are analyzed. From the information elements of the visual interface and the speech elements of the auditory interface as the entry point, it is hoped to construct a more efficient audio-visual human-computer interaction interface. Taking the alarm information as an example, this paper studied the cognitive rules of the user on the audiovisual interface in the quiet-noise environment through the behavioral performance evaluation and the measurement method of the brainwave physiological index evaluation, and establishes the mapping relationship between “audiovisual-cognition”.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Lebib, R., Papo, D., Bode, S.D., et al.: Evidence of a visual-to-auditory cross-modal sensory gating phenomenon as reflected by the human P50 event-related brain potential modulation. Neurosci. Lett. 341(3), 185–188 (2003)
Shams, L., Kamitani, Y., Thompson, S., et al.: Sound alters visual evoked potentials in humans. Neuroreport 12(17), 3849 (2005)
Qiao, X., Peng, J.: Audiovisual EEG feature extraction and brain cognitive mechanism analysis. J. Test Meas. Technol. 27(6), 509–515 (2013)
Sinke, C., Neufeld, J., Wiswede, D., et al.: N1 enhancement in synesthesia during visual and audio–visual perception in semantic cross-modal conflict situations: an ERP study. Front. Hum. Neurosci. 8(2), 21 (2014)
Kaganovich, N., Schumaker, J., Rowland, C.: Matching heard and seen speech: an ERP study of audiovisual word recognition. Brain Lang. 157–158, 14–24 (2016)
Metzger, U., Parasuraman, R.: Effects of automated conflict cuing and traffic density on air traffic controller performance and visual attention in a datalink environment. Int. J. Aviat. Psychol. 16(4), 343–362 (2006)
Acknowledgments
The authors would like to gratefully acknowledge the reviewers’ comments. This work was supported jointly by National Natural Science Foundation of China (No. 71871056, 71471037), Equipment Pre research & Ministry of education of China Joint fund.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Hu, L., Xue, C., Shao, J. (2020). A Variety of Visual-Speech Matching ERP Studies in Quiet-Noise Scenarios. In: Ahram, T., Karwowski, W., Vergnano, A., Leali, F., Taiar, R. (eds) Intelligent Human Systems Integration 2020. IHSI 2020. Advances in Intelligent Systems and Computing, vol 1131. Springer, Cham. https://doi.org/10.1007/978-3-030-39512-4_50
Download citation
DOI: https://doi.org/10.1007/978-3-030-39512-4_50
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-39511-7
Online ISBN: 978-3-030-39512-4
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)