Academic Journal

Interactive extraction of diverse vocal units from a planar embedding without the need for prior sound segmentation

التفاصيل البيبلوغرافية
العنوان: Interactive extraction of diverse vocal units from a planar embedding without the need for prior sound segmentation
المؤلفون: Lorenz, Corinna, Hao, Xinyu, Tomka, Tomas, Rüttimann, Linus, Hahnloser, Richard H R
المصدر: Lorenz, Corinna; Hao, Xinyu; Tomka, Tomas; Rüttimann, Linus; Hahnloser, Richard H R (2023). Interactive extraction of diverse vocal units from a planar embedding without the need for prior sound segmentation. Frontiers in Bioinformatics, 2:966066.
بيانات النشر: Frontiers Research Foundation
سنة النشر: 2023
المجموعة: University of Zurich (UZH): ZORA (Zurich Open Repository and Archive
مصطلحات موضوعية: Institute of Neuroinformatics, 570 Life sciences, biology, General Medicine
الوصف: Annotating and proofreading data sets of complex natural behaviors such as vocalizations are tedious tasks because instances of a given behavior need to be correctly segmented from background noise and must be classified with minimal false positive error rate. Low-dimensional embeddings have proven very useful for this task because they can provide a visual overview of a data set in which distinct behaviors appear in different clusters. However, low-dimensional embeddings introduce errors because they fail to preserve distances; and embeddings represent only objects of fixed dimensionality, which conflicts with vocalizations that have variable dimensions stemming from their variable durations. To mitigate these issues, we introduce a semi-supervised, analytical method for simultaneous segmentation and clustering of vocalizations. We define a given vocalization type by specifying pairs of high-density regions in the embedding plane of sound spectrograms, one region associated with vocalization onsets and the other with offsets. We demonstrate our two-neighborhood (2N) extraction method on the task of clustering adult zebra finch vocalizations embedded with UMAP. We show that 2N extraction allows the identification of short and long vocal renditions from continuous data streams without initially committing to a particular segmentation of the data. Also, 2N extraction achieves much lower false positive error rate than comparable approaches based on a single defining region. Along with our method, we present a graphical user interface (GUI) for visualizing and annotating data.
نوع الوثيقة: article in journal/newspaper
وصف الملف: application/pdf
اللغة: English
تدمد: 2673-7647
Relation: https://www.zora.uzh.ch/id/eprint/254202/1/ZORA_pdf.pdf; info:pmid/36710910; urn:issn:2673-7647
DOI: 10.3389/fbinf.2022.966066
الاتاحة: https://www.zora.uzh.ch/id/eprint/254202/
https://www.zora.uzh.ch/id/eprint/254202/1/ZORA_pdf.pdf
https://doi.org/10.3389/fbinf.2022.966066
Rights: info:eu-repo/semantics/openAccess ; Creative Commons: Attribution 4.0 International (CC BY 4.0) ; http://creativecommons.org/licenses/by/4.0/
رقم الانضمام: edsbas.D6456BDF
قاعدة البيانات: BASE
الوصف
تدمد:26737647
DOI:10.3389/fbinf.2022.966066