Dipl.-Ing. Christian Kehling

Wissenschaftlicher Mitarbeiter sowie Doktorand
in Kooperation mit dem Fraunhofer Institute for Digital Media Technology IDMT Webpräsenz Fraunhofer Institute for Digital Media Technology IDMT

Kontakt am Fachgebiet EMT
Helmholtzbau, Raum H 3516
+49 3677 69 2602
christian.kehling@tu-ilmenau.de

Kontakt am Fraunhofer IDMT
Ehrenbergstraße 31, Raum 1.45
+49 3677 467-218
christian.kehling@idmt.fraunhofer.de

Forschungsgebiete

  • Domain-agnostic Machine Listening
  • Deep Music Information Retrieval
  • Neural Network Compression
  • Explainable AI (XAI) in Room Acoustics
  • Audio Signal Enhancement using AI
     

Lehre

Publikationen

  • Knowledge Transfer from Neural Networks for Speech Music Classification, 15th International Symposium on Computer Music Multidisciplinary Research, Tokyo Japan, 2021
  • Analyzing the Potential of Pre-Trained Embeddings for Audio Classification Tasks, 27th European Signal Processing Conference (EUSIPCO), Amsterdam, The Netherlands, 2020
  • Springer Handbook for Systematic Musicology, Chapter “Music Technology and Music Education”, Springer Verlag, 2018
  • Parameter Extraction for Bass Guitar Sound Models Including Playing Styles, Proceedings of the 40th International Conference on Acoustics, Speech, and Signal Processing (ICASSP-15), 2015
  • Automatic Tablature Transcription of Electric Guitar Recordings by Estimation of Score- and Instrument-Related Parameters, Proceedings of the 17th International Conference on - Digital Audio Effects (DAFx-14), 2014
  • Parametrische Audiokodierung am Beispiel der E-Gitarre: Analyse, Transkription und Synthese, Akademiker Verlag, 2014

Literaturliste

Anzahl der Treffer: 4
Erstellt: Wed, 01 May 2024 23:14:27 +0200 in 0.0487 sec


Kehling, Christian; Cano, Estefanía
Knowledge transfer from neural networks for speech music classification. - In: Music in the AI era, (2023), S. 202-213

A frequent problem when dealing with audio classification tasks is the scarcity of suitable training data. This work investigates ways of mitigating this problem by applying transfer learning techniques to neural network architectures for several classification tasks from the field of Music Information Retrieval (MIR). First, three state-of-the-art architectures are trained and evaluated with several datasets for the task of speech/music classification. Second, feature representations or embeddings are extracted from the trained networks to classify new tasks with unseen data. The effect of pre-training with respect to the similarity of the source and target tasks are investigated in the context of transfer learning, as well as different fine-tuning strategies.



Grollmisch, Sascha; Cano, Estefanía; Kehling, Christian; Taenzer, Michael
Analyzing the potential of pre-trained embeddings for audio classification tasks. - In: 28th European Signal Processing Conference (EUSIPCO 2020), (2020), S. 790-794

In the context of deep learning, the availability of large amounts of training data can play a critical role in a models performance. Recently, several models for audio classification have been pre-trained in a supervised or self-supervised fashion on large datasets to learn complex feature representations, socalled embeddings. These embeddings can then be extracted from smaller datasets and used to train subsequent classifiers. In the field of audio event detection (AED) for example, classifiers using these features have achieved high accuracy without the need of additional domain knowledge. This paper evaluates three state-of-the-art embeddings on six audio classification tasks from the fields of music information retrieval and industrial sound analysis. The embeddings are systematically evaluated by analyzing the influence on classification accuracy of classifier architecture, fusion methods for file-wise predictions, amount of training data, and initial training domain of the embeddings. To better understand the impact of the pre-training step, results are also compared with those acquired with models trained from scratch. On average, the OpenL3 embeddings performed best with a linear SVM classifier. For a reduced amount of training examples, OpenL3 outperforms the initial baseline.



https://doi.org/10.23919/Eusipco47968.2020.9287743
Schuller, Gerald; Abeßer, Jakob; Kehling, Christian
Parameter extraction for bass guitar sound models including playing styles. - In: 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), ISBN 978-1-4673-6998-5, (2015), S. 404-408

https://doi.org/10.1109/ICASSP.2015.7178000
Kehling, Christian;
Entwicklung eines parametrischen Instrumentencoders basierend auf Analyse und Re-Synthese von Gitarrenaufnahmen. - ca 130 S. : Ilmenau, Techn. Univ., Diplomarbeit, 2013

In dieser Diplomarbeit wird ein Algorithmus vorgestellt, der eine parametrische Audiocodierung von monotimbralen Gitarrensignalen ermöglicht. Als Grundlage dient dabei einzig das Ausgangssignal einer handelsüblichen Elektrogitarre. Anhand des digitalisierten Audiosignals werden Parameter extrahiert, die zum einen die automatische Notation des gespielten Stückes realisieren und zum anderen eine Synthese des notierten Stückes mittels eines Physical Modeling Verfahrens ermöglichen. Betrachtet werden sowohl monophone als auch polyphone Stücke sowie gängige Spieltechniken auf einer Gitarre. Weitere Bestandteile der Arbeit sind die Aufnahme und Annotation eines für diese Arbeit benötigten Datensatzes aus 261 Audiofiles sowie die anschließende Evaluation des Codierungsalgorithmus mittels eines einfachen MUSHRA Hörtestverfahrens zur Beurteilung der generierten Ergebnisse. Außerdem wird eine Lösung für ein mögliches Datenformat zur speicherschonenden Notation und Archivierung der Parameter vorgestellt.