Convert Figma logo to code with AI

ybayle logoawesome-deep-learning-music

List of articles related to deep learning applied to music

2,805
339
2,805
2

Top Related Projects

C++ library for audio and music analysis, description and synthesis, including Python bindings

Quick Overview

The "awesome-deep-learning-music" repository is a curated list of scientific research papers about deep learning techniques applied to music. It covers various aspects of music information retrieval, generation, and analysis using deep learning methods. The repository serves as a comprehensive resource for researchers and practitioners in the field of AI and music.

Pros

  • Extensive collection of research papers covering diverse topics in music and deep learning
  • Well-organized structure with papers categorized by specific areas of focus
  • Regularly updated with new publications and contributions from the community
  • Includes links to code implementations when available, enhancing reproducibility

Cons

  • Lacks detailed explanations or summaries of the listed papers
  • May be overwhelming for beginners due to the large number of papers
  • Some links to papers or code implementations may become outdated over time
  • Does not provide a standardized way to compare or evaluate the different approaches

Note: As this is not a code library but a curated list of research papers, there are no code examples or getting started instructions to provide.

Competitor Comparisons

C++ library for audio and music analysis, description and synthesis, including Python bindings

Pros of essentia

  • Comprehensive C++ library for audio analysis and music information retrieval
  • Extensive set of audio features and algorithms implemented
  • Well-documented and actively maintained by the Music Technology Group

Cons of essentia

  • Steeper learning curve due to C++ implementation
  • Focused on audio analysis rather than deep learning specifically
  • Requires compilation and setup, which may be challenging for beginners

Code comparison

essentia:

#include <essentia/algorithmfactory.h>
#include <essentia/essentiamath.h>

AlgorithmFactory& factory = AlgorithmFactory::instance();
Algorithm* mfcc = factory.create("MFCC");

awesome-deep-learning-music:

# No direct code comparison available, as it's a curated list of resources
# Rather than a code library

Additional notes

awesome-deep-learning-music is a curated list of resources for deep learning in music, while essentia is a comprehensive audio analysis library. The former provides links to various tools, papers, and projects, while the latter offers a concrete implementation of audio processing algorithms.

essentia is more suitable for developers looking to implement audio analysis features directly in their applications, while awesome-deep-learning-music serves as a valuable reference for researchers and developers exploring deep learning applications in music.

Convert Figma logo designs to code with AI

Visual Copilot

Introducing Visual Copilot: A new AI model to turn Figma designs to high quality code using your components.

Try Visual Copilot

README

⚠️ This repo is unmaintained. While the info are still relevant, contributions to keep it up to date are welcome! A good starting point are the articles referenced here: https://github.com/ybayle/awesome-deep-learning-music/issues/5

Deep Learning for Music (DL4M) Awesome

By Yann Bayle (Website, GitHub) from LaBRI (Website, Twitter), Univ. Bordeaux (Website, Twitter), CNRS (Website, Twitter) and SCRIME (Website).

TL;DR Non-exhaustive list of scientific articles on deep learning for music: summary (Article title, pdf link and code), details (table - more info), details (bib - all info)

The role of this curated list is to gather scientific articles, thesis and reports that use deep learning approaches applied to music. The list is currently under construction but feel free to contribute to the missing fields and to add other resources! To do so, please refer to the How To Contribute section. The resources provided here come from my review of the state-of-the-art for my PhD Thesis for which an article is being written. There are already surveys on deep learning for music generation, speech separation and speaker identification. However, these surveys do not cover music information retrieval tasks that are included in this repository.

Table of contents

DL4M summary

 YearArticles, Thesis and ReportsCode
1988Neural net modeling of musicNo
1988Creation by refinement: A creativity paradigm for gradient descent learning networksNo
1988A sequential network design for musical applicationsNo
1989The representation of pitch in a neural net model of chord classificationNo
1989Algorithms for music composition by neural nets: Improved CBR paradigmsNo
1989A connectionist approach to algorithmic compositionNo
1994Neural network music composition by prediction: Exploring the benefits of psychoacoustic constraints and multi-scale processingNo
1995Automatic source identification of monophonic musical instrument soundsNo
1995Neural network based model for classification of music typeNo
1997A machine learning approach to musical style recognitionNo
1998Recognition of music typesNo
1999Musical networks: Parallel distributed perception and performanceNo
2001Multi-phase learning for jazz improvisation and interactionNo
2002A supervised learning approach to musical style recognitionNo
2002Finding temporal structure in music: Blues improvisation with LSTM recurrent networksNo
2002Neural networks for note onset detection in piano musicNo
2004A convolutional-kernel based approach for note onset detection in piano-solo audio signalsNo
2009Unsupervised feature learning for audio classification using convolutional deep belief networksNo
2010Audio musical genre classification using convolutional neural networks and pitch and tempo transformationsNo
2010Automatic musical pattern feature extraction using convolutional neural networkNo
2011Audio-based music classification with a pretrained convolutional networkNo
2012Rethinking automatic chord recognition with convolutional neural networksNo
2012Moving beyond feature design: Deep architectures and automatic feature learning in music informaticsNo
2012Local-feature-map integration using convolutional neural networks for music genre classificationNo
2012Learning sparse feature representations for music annotation and retrievalNo
2012Unsupervised learning of local features for music classificationNo
2013Multiscale approaches to music audio feature learningNo
2013Musical onset detection with convolutional neural networksNo
2013Deep content-based music recommendationNo
2014The munich LSTM-RNN approach to the MediaEval 2014 Emotion In Music taskNo
2014End-to-end learning for music audioNo
2014Deep learning for music genre classificationNo
2014Recognition of acoustic events using deep neural networksNo
2014Deep image features in music information retrievalNo
2014From music audio to chord tablature: Teaching deep convolutional networks to play guitarNo
2014Improved musical onset detection with convolutional neural networksNo
2014Boundary detection in music structure analysis using convolutional neural networksNo
2014Improving content-based and hybrid music recommendation using deep learningNo
2014A deep representation for invariance and music classificationNo
2015Auralisation of deep convolutional neural networks: Listening to learned featuresGitHub
2015Downbeat tracking with multiple features and deep neural networksNo
2015Music boundary detection using neural networks on spectrograms and self-similarity lag matricesNo
2015Classification of spatial audio location and content using convolutional neural networksNo
2015Deep learning, audio adversaries, and music content analysisNo
2015Deep learning and music adversariesGitHub
2015Singing voice detection with deep recurrent neural networksNo
2015Automatic instrument recognition in polyphonic music using convolutional neural networksNo
2015A software framework for musical data augmentationNo
2015A deep bag-of-features model for music auto-taggingNo
2015Music-noise segmentation in spectrotemporal domain using convolutional neural networksNo
2015Musical instrument sound classification with deep convolutional neural network using feature fusion approachNo
2015Environmental sound classification with convolutional neural networksNo
2015Exploring data augmentation for improved singing voice detection with neural networksGitHub
2015Singer traits identification using deep neural networkNo
2015A hybrid recurrent neural network for music transcriptionNo
2015An end-to-end neural network for polyphonic music transcriptionNo
2015Deep karaoke: Extracting vocals from musical mixtures using a convolutional deep neural networkNo
2015Folk music style modelling by recurrent neural networks with long short term memory unitsGitHub
2015Deep neural network based instrument extraction from musicNo
2015A deep neural network for modeling musicNo
2016An efficient approach for segmentation, feature extraction and classification of audio signalsNo
2016Text-based LSTM networks for automatic music compositionNo
2016Towards playlist generation algorithms using RNNs trained on within-track transitionsNo
2016Automatic tagging using deep convolutional neural networksNo
2016Automatic chord estimation on seventhsbass chord vocabulary using deep neural networkNo
2016DeepBach: A steerable model for Bach chorales generationGitHub
2016Bayesian meter tracking on learned signal representationsNo
2016Deep learning for musicNo
2016Learning temporal features using a deep neural network and its application to music genre classificationNo
2016On the potential of simple framewise approaches to piano transcriptionNo
2016Feature learning for chord recognition: The deep chroma extractorGitHub
2016A fully convolutional deep auditory model for musical chord recognitionNo
2016A deep bidirectional long short-term memory based multi-scale approach for music dynamic emotion predictionNo
2016Event localization in music auto-taggingGitHub
2016Deep convolutional networks on the pitch spiral for musical instrument recognitionGitHub
2016SampleRNN: An unconditional end-to-end neural audio generation modelGitHub
2016Robust audio event recognition with 1-max pooling convolutional neural networksNo
2016Experimenting with musically motivated convolutional neural networksGitHub
2016Singing voice melody transcription using deep neural networksNo
2016Singing voice separation using deep neural networks and F0 estimationWebsite
2016Learning to pinpoint singing voice from weakly labeled examplesNo
2016Analysis of time-frequency representations for musical onset detection with convolutional neural networkNo
2016Note onset detection in musical signals via neural-network-based multi-ODF fusionNo
2016Music transcription modelling and composition using deep learningGitHub
2016Convolutional neural network for robust pitch determinationNo
2016Deep convolutional neural networks and data augmentation for acoustic event detectionWebsite
2017Gabor frames and deep scattering networks in audio processingNo
2017Vision-based detection of acoustic timed events: A case study on clarinet note onsetsNo
2017Deep learning techniques for music generation - A surveyNo
2017JamBot: Music theory aware chord based generation of polyphonic music with LSTMsGitHub
2017XFlow: 1D <-> 2D cross-modal deep neural networks for audiovisual classificationNo
2017Machine listening intelligenceNo
2017Monoaural audio source separation using deep convolutional neural networksGitHub
2017Deep multimodal network for multi-label classificationNo
2017A tutorial on deep learning for music information retrievalGitHub
2017A comparison on audio signal preprocessing methods for deep neural networks on music taggingGitHub
2017Transfer learning for music classification and regression tasksGitHub
2017Convolutional recurrent neural networks for music classificationGitHub
2017An evaluation of convolutional neural networks for music classification using spectrogramsNo
2017Large vocabulary automatic chord estimation using deep neural nets: Design framework, system variations and limitationsNo
2017Basic filters for convolutional neural networks: Training or design?No
2017Ensemble Of Deep Neural Networks For Acoustic Scene ClassificationNo
2017Robust downbeat tracking using an ensemble of convolutional networksNo
2017Music signal processing using vector product neural networksNo
2017Transforming musical signals through a genre classifying convolutional neural networkNo
2017Audio to score matching by combining phonetic and duration informationGitHub
2017Interactive music generation with positional constraints using anticipation-RNNsNo
2017Deep rank-based transposition-invariant distances on musical sequencesNo
2017GLSR-VAE: Geodesic latent space regularization for variational autoencoder architecturesNo
2017Deep convolutional neural networks for predominant instrument recognition in polyphonic musicNo
2017CNN architectures for large-scale audio classificationNo
2017DeepSheet: A sheet music generator based on deep learningNo
2017Talking Drums: Generating drum grooves with neural networksNo
2017Singing voice separation with deep U-Net convolutional networksGitHub
2017Music emotion recognition via end-to-end multimodal neural networksNo
2017Chord label personalization through deep learning of integrated harmonic interval-based representationsNo
2017End-to-end musical key estimation using a convolutional neural networkNo
2017MediaEval 2017 AcousticBrainz genre task: Multilayer perceptron approachNo
2017Classification-based singing melody extraction using deep convolutional neural networksNo
2017Multi-level and multi-scale feature aggregation using pre-trained convolutional neural networks for music auto-taggingNo
2017Multi-level and multi-scale feature aggregation using sample-level deep convolutional neural networks for music classificationGitHub
2017Sample-level deep convolutional neural networks for music auto-tagging using raw waveformsNo
2017A SeqGAN for Polyphonic Music GenerationGitHub
2017Harmonic and percussive source separation using a convolutional auto encoderNo
2017Stacked convolutional and recurrent neural networks for music emotion recognitionNo
2017A deep learning approach to source separation and remixing of hiphop musicNo
2017Music Genre Classification Using Masked Conditional Neural NetworksNo
2017Monaural Singing Voice Separation with Skip-Filtering Connections and Recurrent Inference of Time-Frequency MaskGitHub
2017Generating data to train convolutional neural networks for classical music source separationGitHub
2017Monaural score-informed source separation for classical music using convolutional neural networksGitHub
2017Multi-label music genre classification from audio, text, and images using deep featuresGitHub
2017A deep multimodal approach for cold-start music recommendationGitHub
2017Melody extraction and detection through LSTM-RNN with harmonic sum lossNo
2017Representation learning of music using artist labelsNo
2017Toward inverse control of physics-based sound synthesisWebsite
2017DNN and CNN with weighted and multi-task loss functions for audio event detectionNo
2017Score-informed syllable segmentation for a cappella singing voice with convolutional neural networksGitHub
2017End-to-end learning for music audio tagging at scaleGitHub
2017Designing efficient architectures for modeling temporal features with convolutional neural networksGitHub
2017Timbre analysis of music audio signals with convolutional neural networksGitHub
2017The MUSDB18 corpus for music separationGitHub
2017Deep learning and intelligent audio mixingNo
2017Deep learning for event detection, sequence labelling and similarity estimation in music signalsNo
2017Music feature maps with convolutional neural networks for music genre classificationNo
2017Automatic drum transcription for polyphonic recordings using soft attention mechanisms and convolutional neural networksGitHub
2017Adversarial semi-supervised audio source separation applied to singing voice extractionNo
2017Taking the models back to music practice: Evaluating generative transcription models built using deep learningGitHub
2017Generating nontrivial melodies for music as a serviceNo
2017Invariances and data augmentation for supervised music transcriptionGitHub
2017Lyrics-based music genre classification using a hierarchical attention networkGitHub
2017A hybrid DSP/deep learning approach to real-time full-band speech enhancementGitHub
2017Convolutional methods for music analysisNo
2017Extending temporal feature integration for semantic audio analysisNo
2017Recognition and retrieval of sound events using sparse coding convolutional neural networkNo
2017A two-stage approach to note-level transcription of a specific pianoNo
2017Reducing model complexity for DNN based large-scale audio classificationNo
2017Audio spectrogram representations for processing with convolutional neural networksWebsite
2017Unsupervised feature learning based on deep models for environmental audio taggingNo
2017Attention and localization based on a deep convolutional recurrent model for weakly supervised audio taggingGitHub
2017Surrey-CVSSP system for DCASE2017 challenge task4GitHub
2017A study on LSTM networks for polyphonic music sequence modellingWebsite
2018MuseGAN: Multi-track sequential generative adversarial networks for symbolic music generation and accompanimentGitHub
2018Music transformer: Generating music with long-term structureNo
2018Music theory inspired policy gradient method for piano music transcriptionNo
2019Enabling factorized piano music modeling and generation with the MAESTRO datasetGitHub
2019Generating Long Sequences with Sparse TransformersGitHub
2021DadaGP: a Dataset of Tokenized GuitarPro Songs for Sequence ModelsGitHub

Go back to top

DL4M details

A human-readable table summarized version if displayed in the file dl4m.tsv. All details for each article are stored in the corresponding bib entry in dl4m.bib. Each entry has the regular bib field:

  • author
  • year
  • title
  • journal or booktitle

Each entry in dl4m.bib also displays additional information:

  • link - HTML link to the PDF file
  • code - Link to the source code if available
  • archi - Neural network architecture
  • layer - Number of layers
  • task - The proposed tasks studied in the article
  • dataset - The names of the dataset used
  • dataaugmentation - The type of data augmentation technique used
  • time - The computation time
  • hardware - The hardware used
  • note - Additional notes and information
  • repro - Indication to what extent the experiments are reproducible

Go back to top

Code without articles

Go back to top

Statistics and visualisations

  • 167 papers referenced. See the details in dl4m.bib. There are more papers from 2017 than any other years combined. Number of articles per year: Number of articles per year
  • If you are applying DL to music, there are 364 other researchers in your field.
  • 34 tasks investigated. See the list of tasks. Tasks pie chart: Tasks pie chart
  • 55 datasets used. See the list of datasets. Datasets pie chart: Datasets pie chart
  • 30 architectures used. See the list of architectures. Architectures pie chart: Architectures pie chart
  • 9 frameworks used. See the list of frameworks. Frameworks pie chart: Frameworks pie chart
  • Only 47 articles (28%) provide their source code. Repeatability is the key to good science, so check out the list of useful resources on reproducibility for MIR and ML.

Go back to top

Advices for reviewers of dl4m articles

Please refer to the advice_review.md file.

How To Contribute

Contributions are welcome! Please refer to the CONTRIBUTING.md file.

Go back to top

FAQ

How are the articles sorted?

The articles are first sorted by decreasing year (to keep up with the latest news) and then alphabetically by the main author's family name.

Why are preprint from arXiv included in the list?

I want to have exhaustive research and the latest news on DL4M. However, one should take care of the information provided in the articles currently in review. If possible you should wait for the final accepted and peer-reviewed version before citing an arXiv paper. I regularly update the arXiv links to the corresponding published papers when available.

How much can I trust the results published in an article?

The list provided here does not guarantee the quality of the articles. You should either try to reproduce the experiments described or submit a request to ReScience. Use one article's conclusion at your own risks.

Go back to top

Acronyms used

A list of useful acronyms used in deep learning and music is stored in acronyms.md.

Go back to top

Sources

The list of conferences, journals and aggregators used to gather the proposed materials is stored in sources.md.

Go back to top

Contributors

Go back to top

Other useful related lists and resources

Audio

Go back to top

Music datasets

Go back to top

Deep learning

Go back to top

Cited by

If you use the information contained in this repository, please let us know! This repository is cited by:

Go back to top

License

You are free to copy, modify, and distribute Deep Learning for Music (DL4M) with attribution under the terms of the MIT license. See the LICENSE file for details. This project use another projects and you may refer to them for appropriate license information :

Go back to top