Session A, Session B, Session C, Session D,
Jeong Choi; Jongpil Lee; Jiyoung Park; Juhan Nam
"Investigated the paradigm of zero-shot learning applied to music domain. Organized 2 side information setups for music calssification task. Proposed a data split scheme and associated evaluation settings for the multi-label zero-shot learning."
"We find extremely valuable to compare playlist datasets generated in different contexts, as it allows to understand how changes in the listening experience are affecting playlist creation strategies."
[A-13] Conditioned-U-Net: Introducing a Control Mechanism in the U-Net for Multiple Source Separations | paper | code
Gabriel Meseguer-Brocal; Geoffroy Peeters
"In this paper, we apply conditioning learning to source separation and introduce a control mechanism to the standard U-Net architecture. The control mechanism allows multiple instrument separations with just one model without losing performance."
"DrummerNet is a drum transcriber trained in an unsupervised fashion. DrummerNet learns to transcribe by learning to reconstruct the audio with the transcription estimate. Unsupervised learning + a large dataset allow DrummerNet to be less-biased."
[B-09] Towards Explainable Emotion Recognition in Music: The Route via Mid-level Features | paper | demo
Shreyan Chowdhury, Andreu Vall Portabella, Verena Haunschmid, Gerhard Widmer
"Explainable predictions of emotion from music can be obtained by introducing an intermediate representation of mid-level perceptual features in the predictor deep neural network."
[C-01] Learning a Joint Embedding Space of Monophonic and Mixed Music Signals for Singing Voice | paper | code | mashup example
"The paper introduces a new method of obtaining a consistent singing voice representation from both monophonic and mixed music signals. Also, it presents a simple music mashup pipeline to create a large synthetic singer dataset"
Eva Zangerle; Michael Vötter; Ramona Huber; Yi-Hsuan Yang
"We show that for predicting the potential success of a song, both low- and high-level audio features are important. We use a deep and wide neural network to model these features and perform a regression task on the track’s rank in the charts."
[C-07] Learning to Traverse Latent Spaces for Musical Score Inpainting | paper | code | audio-examples
Ashis Pati, Alexander Lerch, Gaëtan Hadjeres
"Recurrent Neural Networks can be trained using latent embeddings of a Variational Auto-Encoder-based model to to perform interactive music generation tasks such as inpainting."
[C-09] The AcousticBrainz Genre Dataset: Multi-Source, Multi-Level, Multi-Label, and Large-Scale | paper | dataset
Dmitry Bogdanov, Alastair Porter, Hendrik Schreiber, Julián Urbano, Sergio Oramas
"The AcousticBrainz Genre Dataset allows researchers to explore how the same music pieces are annotated differently by different communities following their own genre taxonomies, and how these differences can be addressed by genre recognition systems."
Saebyul Park; Taegyun Kwon; Jongpil Lee; Jeounghoon Kim; Juhan Nam
"We propose a cross-scape plot representation to visualize multi-scaled melody similarity between two symbolic music. We evaluate its effectiveness on examples from folk music collections with similarity-based categories and plagiarism cases."
[D-04] A Dataset of Rhytmic Pattern Reproductions and Baseline Automatic Assessment System | paper | code | MAST rhythm dataset | re-annotated dataset
Felipe Falcão, Baris Bozkurt, Xavier Serra, Nazareno Andrade, Ozan Baysal
"This present work is an effort to address the shortage of music datasets designed for rhythmic assessment. A new dataset and baseline rhythmic assessment system are provided in order to support comparative studies about rhythmic assessment."
[D-06] Blending Acoustic and Language Model Predictions for Automatic Music Transcription | paper | code | supplementary-material
Adrien Ycart, Andrew McLeod, Emmanouil Benetos, Kazuyoshi Yoshii
"Dynamically integrating predictions from an acoustic and a language model with a blending model improves automatic music transcription performance on the MAPS dataset. Results are further improved by operating on 16th-note timesteps rather than 40ms."
[D-08] A Comparative Study of Neural Models for Polyphonic Music Sequence Transduction | paper
Adrien Ycart, Daniel Stoller, Emmanouil Benetos
"A systematic study using various neural models and automatic music transcription systems shows that a cross-entropy-loss CNN improves transduction performance, while an LSTM does not. Using an adversarial set-up also does not yield improvement."