Comparing identification of vocal imitations and computational sketches of everyday soundstextconferencePaperLemaitreGuillaumeautHouixOlivierautVoisinFrédéricautMisdariisNicolasautSusiniPatrickautAll rights reserved10.1121/1.4970854https://hal.archives-ouvertes.fr/hal-01448968Sounds are notably difficult to describe. It is thus not surprising that human speakers often use many imitative vocalizations to communicate about sounds. In practice,vocal imitations of non-speech everyday sounds (e.g. the sound of a car passing by) arevery effective: listeners identify sounds better with vocal imitations than with verbal descriptions, despite the fact that vocal imitations are often inaccurate, constrained by the human vocal apparatus. The present study investigated the semantic representations evoked by vocal imitations by experimentally quantifying how well listeners could match sounds to category labels. Itcompared two different types of sounds: human vocal imitations, and computational auditory sketches (created by algorithmic computations), both based on easily identifiable sounds (sounds of human actions and manufactured products). The results show that performance with the best vocal imitations was similar to the best auditory sketches for most categories of sounds.More detailed analyses showed that the acoustic distance between vocal imitations and referent sounds is not sufficient to account for such performance. They suggested that instead of reproducing the acoustic properties of the referent sound as accurately as vocally possible, vocal imitations focus on a few important features dependent on each particular sound category.HAL Archives Ouvertesconference publication1403390 - 3390Honolulu, United StatesNovember 2016monographicMeeting of the Acoustical Society of AmericaThe Journal of the Acoustical Society of America