Detecting somatisation disorder via speech: introducing the Shenzhen Somatisation Speech Corpus

  • Objective Speech recognition technology is widely used as a mature technical approach in many fields. In the study of depression recognition, speech signals are commonly used due to their convenience and ease of acquisition. Though speech recognition is popular in the research field of depression recognition, it has been little studied in somatisation disorder recognition. The reason for this is the lack of a publicly accessible database of relevant speech and benchmark studies. To this end, we introduce our somatisation disorder speech database and give benchmark results. Methods By collecting speech samples of somatisation disorder patients, in cooperation with the Shenzhen University General Hospital, we introduce our somatisation disorder speech database, the Shenzhen Somatisation Speech Corpus (SSSC). Moreover, a benchmark for SSSC using classic acoustic features and a machine learning model is proposed in our work. Results To obtain a more scientific benchmark, we haveObjective Speech recognition technology is widely used as a mature technical approach in many fields. In the study of depression recognition, speech signals are commonly used due to their convenience and ease of acquisition. Though speech recognition is popular in the research field of depression recognition, it has been little studied in somatisation disorder recognition. The reason for this is the lack of a publicly accessible database of relevant speech and benchmark studies. To this end, we introduce our somatisation disorder speech database and give benchmark results. Methods By collecting speech samples of somatisation disorder patients, in cooperation with the Shenzhen University General Hospital, we introduce our somatisation disorder speech database, the Shenzhen Somatisation Speech Corpus (SSSC). Moreover, a benchmark for SSSC using classic acoustic features and a machine learning model is proposed in our work. Results To obtain a more scientific benchmark, we have compared and analysed the performance of different acoustic features, i. e., the full ComParE feature set, or only MFCCs, fundamental frequency (F0), and frequency and bandwidth of the formants (F1-F3). By comparison. the best result of our benchmark is the 76.0 % unweighted average recall achieved by a support vector machine with formants F1–F3. Conclusion The proposal of SSSC bridges a research gap in somatisation disorder, providing researchers with a publicly accessible speech database. In addition, the results of the benchmark show the scientific validity and feasibility of computer audition for speech recognition in somatization disorders.show moreshow less

Download full text files

Export metadata

Statistics

Number of document requests

Additional Services

Share in Twitter Search Google Scholar
Metadaten
Author:Kun Qian, Ruolan Huang, Zhihao Bao, Yang Tan, Zhonghao Zhao, Mengkai Sun, Bin Hu, Björn W. SchullerORCiDGND, Yoshiharu Yamamoto
Frontdoor URLhttps://opus.bibliothek.uni-augsburg.de/opus4/104187
ISSN:2667-1026OPAC
Parent Title (English):Intelligent Medicine
Publisher:Elsevier BV
Place of publication:Amsterdam
Type:Article
Language:English
Year of first Publication:2023
Publishing Institution:Universität Augsburg
Release Date:2023/05/09
Tag:Health Informatics; Medicine (miscellaneous); Biomedical Engineering; Artificial Intelligence
DOI:https://doi.org/10.1016/j.imed.2023.03.001
Institutes:Fakultät für Angewandte Informatik
Fakultät für Angewandte Informatik / Institut für Informatik
Fakultät für Angewandte Informatik / Institut für Informatik / Lehrstuhl für Embedded Intelligence for Health Care and Wellbeing
Dewey Decimal Classification:0 Informatik, Informationswissenschaft, allgemeine Werke / 00 Informatik, Wissen, Systeme / 004 Datenverarbeitung; Informatik
Latest Publications (not yet published in print):Aktuelle Publikationen (noch nicht gedruckt erschienen)
Licence (German):CC-BY-NC-ND 4.0: Creative Commons: Namensnennung - Nicht kommerziell - Keine Bearbeitung (mit Print on Demand)