Computational audio analysis (report from Dagstuhl Seminar 13451)

  • Compared to traditional speech, music, or sound processing, the computational analysis of general audio data has a relatively young research history. In particular, the extraction of affective information (i.e., information that does not deal with the 'immediate' nature of the content such as the spoken words or note events) from audio signals has become an important research strand with a huge increase of interest in academia and industry. At an early stage of this novel research direction, many analysis techniques and representations were simply transferred from the speech domain to other audio domains. However, general audio signals (including their affective aspects) typically possess acoustic and structural characteristics that distinguish them from spoken language or isolated `controlled' music or sound events. In the Dagstuhl Seminar 13451 titled "Computational Audio Analysis" we discussed the development of novel machine learning as well as signal processing techniques that areCompared to traditional speech, music, or sound processing, the computational analysis of general audio data has a relatively young research history. In particular, the extraction of affective information (i.e., information that does not deal with the 'immediate' nature of the content such as the spoken words or note events) from audio signals has become an important research strand with a huge increase of interest in academia and industry. At an early stage of this novel research direction, many analysis techniques and representations were simply transferred from the speech domain to other audio domains. However, general audio signals (including their affective aspects) typically possess acoustic and structural characteristics that distinguish them from spoken language or isolated `controlled' music or sound events. In the Dagstuhl Seminar 13451 titled "Computational Audio Analysis" we discussed the development of novel machine learning as well as signal processing techniques that are applicable for a wide range of audio signals and analysis tasks. In particular, we looked at a variety of sounds besides speech such as music recordings, animal sounds, environmental sounds, and mixtures thereof. In this report, we give an overview of the various contributions and results of the seminar. We start with an executive summary, which describes the main topics, goals, and group activities. Then, one finds a list of abstracts giving a more detailed overview of the participants' contributions as well as of the ideas and results discussed in the group meetings of our seminar. To conclude, an attempt is made to define the field as given by the views of the participants.show moreshow less

Download full text files

Export metadata

Statistics

Number of document requests

Additional Services

Share in Twitter Search Google Scholar
Metadaten
Author:Meinard Müller, Shrikanth S. Narayanan, Björn SchullerORCiDGND
URN:urn:nbn:de:bvb:384-opus4-770182
Frontdoor URLhttps://opus.bibliothek.uni-augsburg.de/opus4/77018
ISSN:2192-5283OPAC
Parent Title (German):Dagstuhl Reports
Publisher:Schloss Dagstuhl
Place of publication:Dagstuhl
Type:Article
Language:English
Year of first Publication:2014
Publishing Institution:Universität Augsburg
Release Date:2020/06/17
Volume:3
Issue:11
First Page:1
Last Page:28
DOI:https://doi.org/10.4230/DagRep.3.11.1
Institutes:Fakultät für Angewandte Informatik
Fakultät für Angewandte Informatik / Institut für Informatik
Fakultät für Angewandte Informatik / Institut für Informatik / Lehrstuhl für Embedded Intelligence for Health Care and Wellbeing
Dewey Decimal Classification:0 Informatik, Informationswissenschaft, allgemeine Werke / 00 Informatik, Wissen, Systeme / 004 Datenverarbeitung; Informatik
Licence (German):CC-BY 3.0: Creative Commons - Namensnennung (mit Print on Demand)