dc.contributor.author | O'Dwyer, Jonny | |
dc.contributor.author | Murray, Niall | |
dc.contributor.author | Flynn, Ronan | |
dc.date.accessioned | 2019-05-09T09:01:34Z | |
dc.date.available | 2019-05-09T09:01:34Z | |
dc.date.copyright | 2018-02 | |
dc.date.issued | 2018 | |
dc.identifier.citation | O'Dwyer, J., Murray, N., Flynn, R. (2018). Affective computing using speech and eye gaze: a review and bimodal system proposal for continuous affect prediction. In - arXiv preprint arXiv:1805.06652, 2018. | en_US |
dc.identifier.other | Software Research Institute - Articles | en_US |
dc.identifier.uri | https://research.thea.ie/handle/20.500.12065/2680 | |
dc.description.abstract | Speech has been a widely used modality in the field of affective computing. Recently however, there has been a growing interest in the use of multi-modal affective computing systems. These multi-modal systems incorporate both verbal
and non-verbal features for affective computing tasks. Such multi-modal affective computing systems are advantageous for emotion assessment of individuals
in audio-video communication environments such as teleconferencing, healthcare, and education. From a review of the literature, the use of eye gaze features extracted from video is a modality that has remained largely unexploited
for continuous affect prediction. This work presents a review of the literature
within the emotion classification and continuous affect prediction sub-fields of
affective computing for both speech and eye gaze modalities. Additionally, continuous affect prediction experiments using speech and eye gaze modalities are
presented. A baseline system is proposed using open source software, the performance of which is assessed on a publicly available audio-visual corpus. Further
system performance is assessed in a cross-corpus and cross-lingual experiment.
The experimental results suggest that eye gaze is an effective supportive modality for speech when used in a bimodal continuous affect prediction system. The
addition of eye gaze to speech in a simple feature fusion framework yields a
prediction improvement of 6.13% for valence and 1.62% for arousal. | en_US |
dc.format | PDF | en_US |
dc.language.iso | en | en_US |
dc.rights | Attribution-NonCommercial-NoDerivs 3.0 Ireland | * |
dc.rights.uri | http://creativecommons.org/licenses/by-nc-nd/3.0/ie/ | * |
dc.subject | Human-computer interaction | en_US |
dc.subject | Artificial intelligence | en_US |
dc.subject | User interfaces (Computer systems) | en_US |
dc.title | Affective computing using speech and eye gaze: a review and bimodal system proposal for continuous affect prediction. | en_US |
dc.type | Article | en_US |
dc.description.peerreview | yes | en_US |
dc.identifier.orcid | https://orcid.org/0000-0002-5919-0596 | |
dc.rights.access | Open Access | en_US |
dc.subject.department | Software Research Institute AIT | en_US |