Abstract
Emotion evoked by an advertisement plays a key role in influencing brand recall and eventual consumer choices. Automatic ad affect recognition has several useful applications. However, the use of content-based feature representations does not give insights into how affect is modulated by aspects such as the ad scene setting, salient object attributes and their interactions. Neither do such approaches inform us on how humans prioritize visual information for ad understanding. Our work addresses these lacunae by decomposing video content into detected objects, coarse scene structure, object statistics and actively attended objects identified via eye-gaze. We measure the importance of each of these information channels by systematically incorporating related information into ad affect prediction models. Contrary to the popular notion that ad affect hinges on the narrative and the clever use of linguistic and social cues, we find that actively attended objects and the coarse scene structure better encode affective information as compared to individual scene objects or conspicuous background elements.
Original language | English |
---|---|
Title of host publication | Proceedings of the 20th ACM International Conference on Multimodal Interaction (ICMI 2018) |
Editors | Sidney K. D’Mello, Panayiotis (Panos) Georgiou, Stefan Scherer, Emily Mower Provost, Mohammad Soleymani, Marcelo Worsley |
Place of Publication | United States |
Publisher | Association for Computing Machinery (ACM) |
Pages | 210-219 |
Number of pages | 10 |
ISBN (Print) | 9781450356923 |
DOIs | |
Publication status | Published - 2 Oct 2018 |
Externally published | Yes |
Event | 20th ACM International Conference on Multimodal Interaction, ICMI 2018 - Boulder, United States Duration: 16 Oct 2018 → 20 Oct 2018 |
Conference
Conference | 20th ACM International Conference on Multimodal Interaction, ICMI 2018 |
---|---|
Country/Territory | United States |
City | Boulder |
Period | 16/10/18 → 20/10/18 |