SAFETYLIT WEEKLY UPDATE

We compile citations and summaries of about 400 new articles every week.
RSS Feed

HELP: Tutorials | FAQ
CONTACT US: Contact info

Search Results

Journal Article

Citation

Knoeferle KM, Knoeferle P, Velasco C, Spence C. J. Exp. Psychol. Appl. 2016; 22(2): 196-210.

Affiliation

Crossmodal Research Laboratory, Department of Experimental Psychology, University of Oxford.

Copyright

(Copyright © 2016, American Psychological Association)

DOI

10.1037/xap0000084

PMID

27295466

Abstract

Building on models of crossmodal attention, the present research proposes that brand search is inherently multisensory, in that the consumers' visual search for a specific brand can be facilitated by semantically related stimuli that are presented in another sensory modality. A series of 5 experiments demonstrates that the presentation of spatially nonpredictive auditory stimuli associated with products (e.g., usage sounds or product-related jingles) can crossmodally facilitate consumers' visual search for, and selection of, products. Eye-tracking data (Experiment 2) revealed that the crossmodal effect of auditory cues on visual search manifested itself not only in RTs, but also in the earliest stages of visual attentional processing, thus suggesting that the semantic information embedded within sounds can modulate the perceptual saliency of the target products' visual representations. Crossmodal facilitation was even observed for newly learnt associations between unfamiliar brands and sonic logos, implicating multisensory short-term learning in establishing audiovisual semantic associations. The facilitation effect was stronger when searching complex rather than simple visual displays, thus suggesting a modulatory role of perceptual load. (PsycINFO Database Record

(c) 2016 APA, all rights reserved).


Language: en

NEW SEARCH


All SafetyLit records are available for automatic download to Zotero & Mendeley
Print