Publication: A saliency-based attention LSTM model for cognitive load classification from speech
dc.affiliation.dpto | UC3M. Departamento de Teoría de la Señal y Comunicaciones | es |
dc.affiliation.grupoinv | UC3M. Grupo de Investigación: Procesado Multimedia | es |
dc.contributor.author | Gallardo Antolín, Ascensión | |
dc.contributor.author | Montero Martínez, Juan Manuel | |
dc.contributor.funder | Ministerio de Economía y Competitividad (España) | es |
dc.date.accessioned | 2020-12-22T12:49:53Z | |
dc.date.available | 2020-12-22T12:49:53Z | |
dc.date.issued | 2019 | |
dc.description | Proceeding of: Interspeech 2019, 20th Annual Conference of the International Speech Communication Association, Graz, Austria, 15-19 September 2019 | en |
dc.description.abstract | Cognitive Load (CL) refers to the amount of mental demand that a given task imposes on an individual's cognitive system and it can affect his/her productivity in very high load situations. In this paper, we propose an automatic system capable of classifying the CL level of a speaker by analyzing his/her voice. Our research on this topic goes into two main directions. In the first one, we focus on the use of Long Short-Term Memory (LSTM) networks with different weighted pooling strategies for CL level classification. In the second contribution, for overcoming the need of a large amount of training data, we propose a novel attention mechanism that uses the Kalinli's auditory saliency model. Experiments show that our proposal outperforms significantly both, a baseline system based on Support Vector Machines (SVM) and a LSTM-based system with logistic regression attention model. | en |
dc.description.sponsorship | The work leading to these results has been partly supported by Spanish Government grants TEC2017-84395-P and TEC2017-84593-C2-1-R. | en |
dc.format.extent | 5 | es |
dc.identifier.bibliographicCitation | Interspeech 2019, 20th Annual Conference of the International Speech Communication Association, Graz, Austria, 15-19 September 2019 [Proceedings]. ISCA, 2019, Pp. 216-220 | en |
dc.identifier.doi | https://doi.org/10.21437/Interspeech.2019-1603 | |
dc.identifier.publicationfirstpage | 216 | es |
dc.identifier.publicationlastpage | 220 | es |
dc.identifier.publicationtitle | Interspeech 2019, 20th Annual Conference of the International Speech Communication Association, Graz, Austria, 15-19 September 2019 [Proceedings] | en |
dc.identifier.uri | https://hdl.handle.net/10016/31660 | |
dc.identifier.uxxi | CC/0000030601 | |
dc.language.iso | eng | en |
dc.publisher | International Speech Communication Association (ISCA) | en |
dc.relation.eventdate | 2019-09-15 | es |
dc.relation.eventplace | Graz, AUSTRIA | en |
dc.relation.eventtitle | 20th Annual Conference of the International Speech Communication Association (INTERSPEECH 2019) | en |
dc.relation.projectID | Gobierno de España. TEC2017-84395-P | es |
dc.relation.projectID | Gobierno de España. TEC2017-84593-C2-1-R | es |
dc.rights | © 2019 ISCA | en |
dc.rights.accessRights | open access | es |
dc.subject.eciencia | Electrónica | es |
dc.subject.eciencia | Telecomunicaciones | es |
dc.subject.other | Cognitive load | en |
dc.subject.other | Speech | en |
dc.subject.other | LSTM | en |
dc.subject.other | Weigthed pooling | en |
dc.subject.other | Auditory saliency | en |
dc.subject.other | Attention model | en |
dc.title | A saliency-based attention LSTM model for cognitive load classification from speech | en |
dc.type | conference paper | * |
dc.type.hasVersion | VoR | * |
dspace.entity.type | Publication |
Files
Original bundle
1 - 1 of 1
Loading...
- Name:
- saliency_INTERSPEECH_2019.pdf
- Size:
- 289.22 KB
- Format:
- Adobe Portable Document Format