Giter Site home page Giter Site logo

ybayle / awesome-deep-learning-music Goto Github PK

View Code? Open in Web Editor NEW
2.7K 144.0 340.0 6.01 MB

List of articles related to deep learning applied to music

License: MIT License

TeX 84.24% Python 15.76%
awesome awesome-list unicorns list lists resources deeplearning deep-learning deep-neural-networks neural-network

awesome-deep-learning-music's Introduction

⚠️ This repo is unmaintained. While the info are still relevant, contributions to keep it up to date are welcome! A good starting point are the articles referenced here: #5

Deep Learning for Music (DL4M) Awesome

By Yann Bayle (Website, GitHub) from LaBRI (Website, Twitter), Univ. Bordeaux (Website, Twitter), CNRS (Website, Twitter) and SCRIME (Website).

TL;DR Non-exhaustive list of scientific articles on deep learning for music: summary (Article title, pdf link and code), details (table - more info), details (bib - all info)

The role of this curated list is to gather scientific articles, thesis and reports that use deep learning approaches applied to music. The list is currently under construction but feel free to contribute to the missing fields and to add other resources! To do so, please refer to the How To Contribute section. The resources provided here come from my review of the state-of-the-art for my PhD Thesis for which an article is being written. There are already surveys on deep learning for music generation, speech separation and speaker identification. However, these surveys do not cover music information retrieval tasks that are included in this repository.

Table of contents

DL4M summary

 Year Articles, Thesis and Reports Code
1988 Neural net modeling of music No
1988 Creation by refinement: A creativity paradigm for gradient descent learning networks No
1988 A sequential network design for musical applications No
1989 The representation of pitch in a neural net model of chord classification No
1989 Algorithms for music composition by neural nets: Improved CBR paradigms No
1989 A connectionist approach to algorithmic composition No
1994 Neural network music composition by prediction: Exploring the benefits of psychoacoustic constraints and multi-scale processing No
1995 Automatic source identification of monophonic musical instrument sounds No
1995 Neural network based model for classification of music type No
1997 A machine learning approach to musical style recognition No
1998 Recognition of music types No
1999 Musical networks: Parallel distributed perception and performance No
2001 Multi-phase learning for jazz improvisation and interaction No
2002 A supervised learning approach to musical style recognition No
2002 Finding temporal structure in music: Blues improvisation with LSTM recurrent networks No
2002 Neural networks for note onset detection in piano music No
2004 A convolutional-kernel based approach for note onset detection in piano-solo audio signals No
2009 Unsupervised feature learning for audio classification using convolutional deep belief networks No
2010 Audio musical genre classification using convolutional neural networks and pitch and tempo transformations No
2010 Automatic musical pattern feature extraction using convolutional neural network No
2011 Audio-based music classification with a pretrained convolutional network No
2012 Rethinking automatic chord recognition with convolutional neural networks No
2012 Moving beyond feature design: Deep architectures and automatic feature learning in music informatics No
2012 Local-feature-map integration using convolutional neural networks for music genre classification No
2012 Learning sparse feature representations for music annotation and retrieval No
2012 Unsupervised learning of local features for music classification No
2013 Multiscale approaches to music audio feature learning No
2013 Musical onset detection with convolutional neural networks No
2013 Deep content-based music recommendation No
2014 The munich LSTM-RNN approach to the MediaEval 2014 Emotion In Music task No
2014 End-to-end learning for music audio No
2014 Deep learning for music genre classification No
2014 Recognition of acoustic events using deep neural networks No
2014 Deep image features in music information retrieval No
2014 From music audio to chord tablature: Teaching deep convolutional networks to play guitar No
2014 Improved musical onset detection with convolutional neural networks No
2014 Boundary detection in music structure analysis using convolutional neural networks No
2014 Improving content-based and hybrid music recommendation using deep learning No
2014 A deep representation for invariance and music classification No
2015 Auralisation of deep convolutional neural networks: Listening to learned features GitHub
2015 Downbeat tracking with multiple features and deep neural networks No
2015 Music boundary detection using neural networks on spectrograms and self-similarity lag matrices No
2015 Classification of spatial audio location and content using convolutional neural networks No
2015 Deep learning, audio adversaries, and music content analysis No
2015 Deep learning and music adversaries GitHub
2015 Singing voice detection with deep recurrent neural networks No
2015 Automatic instrument recognition in polyphonic music using convolutional neural networks No
2015 A software framework for musical data augmentation No
2015 A deep bag-of-features model for music auto-tagging No
2015 Music-noise segmentation in spectrotemporal domain using convolutional neural networks No
2015 Musical instrument sound classification with deep convolutional neural network using feature fusion approach No
2015 Environmental sound classification with convolutional neural networks No
2015 Exploring data augmentation for improved singing voice detection with neural networks GitHub
2015 Singer traits identification using deep neural network No
2015 A hybrid recurrent neural network for music transcription No
2015 An end-to-end neural network for polyphonic music transcription No
2015 Deep karaoke: Extracting vocals from musical mixtures using a convolutional deep neural network No
2015 Folk music style modelling by recurrent neural networks with long short term memory units GitHub
2015 Deep neural network based instrument extraction from music No
2015 A deep neural network for modeling music No
2016 An efficient approach for segmentation, feature extraction and classification of audio signals No
2016 Text-based LSTM networks for automatic music composition No
2016 Towards playlist generation algorithms using RNNs trained on within-track transitions No
2016 Automatic tagging using deep convolutional neural networks No
2016 Automatic chord estimation on seventhsbass chord vocabulary using deep neural network No
2016 DeepBach: A steerable model for Bach chorales generation GitHub
2016 Bayesian meter tracking on learned signal representations No
2016 Deep learning for music No
2016 Learning temporal features using a deep neural network and its application to music genre classification No
2016 On the potential of simple framewise approaches to piano transcription No
2016 Feature learning for chord recognition: The deep chroma extractor GitHub
2016 A fully convolutional deep auditory model for musical chord recognition No
2016 A deep bidirectional long short-term memory based multi-scale approach for music dynamic emotion prediction No
2016 Event localization in music auto-tagging GitHub
2016 Deep convolutional networks on the pitch spiral for musical instrument recognition GitHub
2016 SampleRNN: An unconditional end-to-end neural audio generation model GitHub
2016 Robust audio event recognition with 1-max pooling convolutional neural networks No
2016 Experimenting with musically motivated convolutional neural networks GitHub
2016 Singing voice melody transcription using deep neural networks No
2016 Singing voice separation using deep neural networks and F0 estimation Website
2016 Learning to pinpoint singing voice from weakly labeled examples No
2016 Analysis of time-frequency representations for musical onset detection with convolutional neural network No
2016 Note onset detection in musical signals via neural-network-based multi-ODF fusion No
2016 Music transcription modelling and composition using deep learning GitHub
2016 Convolutional neural network for robust pitch determination No
2016 Deep convolutional neural networks and data augmentation for acoustic event detection Website
2017 Gabor frames and deep scattering networks in audio processing No
2017 Vision-based detection of acoustic timed events: A case study on clarinet note onsets No
2017 Deep learning techniques for music generation - A survey No
2017 JamBot: Music theory aware chord based generation of polyphonic music with LSTMs GitHub
2017 XFlow: 1D <-> 2D cross-modal deep neural networks for audiovisual classification No
2017 Machine listening intelligence No
2017 Monoaural audio source separation using deep convolutional neural networks GitHub
2017 Deep multimodal network for multi-label classification No
2017 A tutorial on deep learning for music information retrieval GitHub
2017 A comparison on audio signal preprocessing methods for deep neural networks on music tagging GitHub
2017 Transfer learning for music classification and regression tasks GitHub
2017 Convolutional recurrent neural networks for music classification GitHub
2017 An evaluation of convolutional neural networks for music classification using spectrograms No
2017 Large vocabulary automatic chord estimation using deep neural nets: Design framework, system variations and limitations No
2017 Basic filters for convolutional neural networks: Training or design? No
2017 Ensemble Of Deep Neural Networks For Acoustic Scene Classification No
2017 Robust downbeat tracking using an ensemble of convolutional networks No
2017 Music signal processing using vector product neural networks No
2017 Transforming musical signals through a genre classifying convolutional neural network No
2017 Audio to score matching by combining phonetic and duration information GitHub
2017 Interactive music generation with positional constraints using anticipation-RNNs No
2017 Deep rank-based transposition-invariant distances on musical sequences No
2017 GLSR-VAE: Geodesic latent space regularization for variational autoencoder architectures No
2017 Deep convolutional neural networks for predominant instrument recognition in polyphonic music No
2017 CNN architectures for large-scale audio classification No
2017 DeepSheet: A sheet music generator based on deep learning No
2017 Talking Drums: Generating drum grooves with neural networks No
2017 Singing voice separation with deep U-Net convolutional networks GitHub
2017 Music emotion recognition via end-to-end multimodal neural networks No
2017 Chord label personalization through deep learning of integrated harmonic interval-based representations No
2017 End-to-end musical key estimation using a convolutional neural network No
2017 MediaEval 2017 AcousticBrainz genre task: Multilayer perceptron approach No
2017 Classification-based singing melody extraction using deep convolutional neural networks No
2017 Multi-level and multi-scale feature aggregation using pre-trained convolutional neural networks for music auto-tagging No
2017 Multi-level and multi-scale feature aggregation using sample-level deep convolutional neural networks for music classification GitHub
2017 Sample-level deep convolutional neural networks for music auto-tagging using raw waveforms No
2017 A SeqGAN for Polyphonic Music Generation GitHub
2017 Harmonic and percussive source separation using a convolutional auto encoder No
2017 Stacked convolutional and recurrent neural networks for music emotion recognition No
2017 A deep learning approach to source separation and remixing of hiphop music No
2017 Music Genre Classification Using Masked Conditional Neural Networks No
2017 Monaural Singing Voice Separation with Skip-Filtering Connections and Recurrent Inference of Time-Frequency Mask GitHub
2017 Generating data to train convolutional neural networks for classical music source separation GitHub
2017 Monaural score-informed source separation for classical music using convolutional neural networks GitHub
2017 Multi-label music genre classification from audio, text, and images using deep features GitHub
2017 A deep multimodal approach for cold-start music recommendation GitHub
2017 Melody extraction and detection through LSTM-RNN with harmonic sum loss No
2017 Representation learning of music using artist labels No
2017 Toward inverse control of physics-based sound synthesis Website
2017 DNN and CNN with weighted and multi-task loss functions for audio event detection No
2017 Score-informed syllable segmentation for a cappella singing voice with convolutional neural networks GitHub
2017 End-to-end learning for music audio tagging at scale GitHub
2017 Designing efficient architectures for modeling temporal features with convolutional neural networks GitHub
2017 Timbre analysis of music audio signals with convolutional neural networks GitHub
2017 The MUSDB18 corpus for music separation GitHub
2017 Deep learning and intelligent audio mixing No
2017 Deep learning for event detection, sequence labelling and similarity estimation in music signals No
2017 Music feature maps with convolutional neural networks for music genre classification No
2017 Automatic drum transcription for polyphonic recordings using soft attention mechanisms and convolutional neural networks GitHub
2017 Adversarial semi-supervised audio source separation applied to singing voice extraction No
2017 Taking the models back to music practice: Evaluating generative transcription models built using deep learning GitHub
2017 Generating nontrivial melodies for music as a service No
2017 Invariances and data augmentation for supervised music transcription GitHub
2017 Lyrics-based music genre classification using a hierarchical attention network GitHub
2017 A hybrid DSP/deep learning approach to real-time full-band speech enhancement GitHub
2017 Convolutional methods for music analysis No
2017 Extending temporal feature integration for semantic audio analysis No
2017 Recognition and retrieval of sound events using sparse coding convolutional neural network No
2017 A two-stage approach to note-level transcription of a specific piano No
2017 Reducing model complexity for DNN based large-scale audio classification No
2017 Audio spectrogram representations for processing with convolutional neural networks Website
2017 Unsupervised feature learning based on deep models for environmental audio tagging No
2017 Attention and localization based on a deep convolutional recurrent model for weakly supervised audio tagging GitHub
2017 Surrey-CVSSP system for DCASE2017 challenge task4 GitHub
2017 A study on LSTM networks for polyphonic music sequence modelling Website
2018 MuseGAN: Multi-track sequential generative adversarial networks for symbolic music generation and accompaniment GitHub
2018 Music transformer: Generating music with long-term structure No
2018 Music theory inspired policy gradient method for piano music transcription No
2019 Enabling factorized piano music modeling and generation with the MAESTRO dataset GitHub
2019 Generating Long Sequences with Sparse Transformers GitHub
2021 DadaGP: a Dataset of Tokenized GuitarPro Songs for Sequence Models GitHub

Go back to top

DL4M details

A human-readable table summarized version if displayed in the file dl4m.tsv. All details for each article are stored in the corresponding bib entry in dl4m.bib. Each entry has the regular bib field:

  • author
  • year
  • title
  • journal or booktitle

Each entry in dl4m.bib also displays additional information:

  • link - HTML link to the PDF file
  • code - Link to the source code if available
  • archi - Neural network architecture
  • layer - Number of layers
  • task - The proposed tasks studied in the article
  • dataset - The names of the dataset used
  • dataaugmentation - The type of data augmentation technique used
  • time - The computation time
  • hardware - The hardware used
  • note - Additional notes and information
  • repro - Indication to what extent the experiments are reproducible

Go back to top

Code without articles

Go back to top

Statistics and visualisations

  • 167 papers referenced. See the details in dl4m.bib. There are more papers from 2017 than any other years combined. Number of articles per year: Number of articles per year
  • If you are applying DL to music, there are 364 other researchers in your field.
  • 34 tasks investigated. See the list of tasks. Tasks pie chart: Tasks pie chart
  • 55 datasets used. See the list of datasets. Datasets pie chart: Datasets pie chart
  • 30 architectures used. See the list of architectures. Architectures pie chart: Architectures pie chart
  • 9 frameworks used. See the list of frameworks. Frameworks pie chart: Frameworks pie chart
  • Only 47 articles (28%) provide their source code. Repeatability is the key to good science, so check out the list of useful resources on reproducibility for MIR and ML.

Go back to top

Advices for reviewers of dl4m articles

Please refer to the advice_review.md file.

How To Contribute

Contributions are welcome! Please refer to the CONTRIBUTING.md file.

Go back to top

FAQ

How are the articles sorted?

The articles are first sorted by decreasing year (to keep up with the latest news) and then alphabetically by the main author's family name.

Why are preprint from arXiv included in the list?

I want to have exhaustive research and the latest news on DL4M. However, one should take care of the information provided in the articles currently in review. If possible you should wait for the final accepted and peer-reviewed version before citing an arXiv paper. I regularly update the arXiv links to the corresponding published papers when available.

How much can I trust the results published in an article?

The list provided here does not guarantee the quality of the articles. You should either try to reproduce the experiments described or submit a request to ReScience. Use one article's conclusion at your own risks.

Go back to top

Acronyms used

A list of useful acronyms used in deep learning and music is stored in acronyms.md.

Go back to top

Sources

The list of conferences, journals and aggregators used to gather the proposed materials is stored in sources.md.

Go back to top

Contributors

Go back to top

Other useful related lists and resources

Audio

Go back to top

Music datasets

Go back to top

Deep learning

Go back to top

Cited by

If you use the information contained in this repository, please let us know! This repository is cited by:

Go back to top

License

You are free to copy, modify, and distribute Deep Learning for Music (DL4M) with attribution under the terms of the MIT license. See the LICENSE file for details. This project use another projects and you may refer to them for appropriate license information :

Go back to top

awesome-deep-learning-music's People

Contributors

bayle42 avatar cprakashagr avatar devn avatar faroit avatar ironjr avatar just1900 avatar knstmrd avatar otnemrasordep avatar sergiooramas avatar ssfrr avatar victolee0 avatar ybayle avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

awesome-deep-learning-music's Issues

Sort papers by category

Right now, the list of all academic papers on DL4M is currently listed in order of publish date.

While the collection of papers is great, I think the list may benefit by some sort of organization by topic? Some examples of topics to sort by might be genre classification (or more generally, audio tagging), music feature extraction, generative models, or recommendation models. It could also be nice to sort by model architecture type (e.g. DNN vs LSTM vs ConvNet).

If the owners of this repo are interested and can come to consensus on the paper topics, I'd be happy to supply a PR if needed 😄

Missing information

In dl4m.bib:

  • 2 missing pdf: Bharucha1988 and Todd1988
  • 40 papers are missing the task field
  • 82 papers are missing the architecture field
  • 84 papers are missing the dataset field

Visualisations:

Tips and tricks:
http://forums.fast.ai/t/30-best-practices/12344

Unsorted references waiting to be processed:
https://github.com/davidwfong/ViolinMelodyCNNs
https://www.researchgate.net/publication/325120491_Modeling_Music_Studies_of_Music_Transcription_Music_Perception_and_Music_Production
http://www.cs.dartmouth.edu/~sarroff/papers/sarroff2018a.pdf
https://www.cs.dartmouth.edu/~sarroff/pages/publications/
https://gitlab.com/rckennedy15/CAPSTONE_2017-2018
https://gitlab.com/kidaa/biaxial-rnn-music-composition
https://github.com/chrisdonahue/wavegan
https://www.tandfonline.com/doi/full/10.1080/09298215.2018.1458885?af=R
https://github.com/Veleslavia/ICMR2017
https://github.com/rupakvignesh/Singing-Voice-Separation
https://github.com/tae-jun/resemul
http://repository.ubn.ru.nl/bitstream/handle/2066/179506/179506.pdf?sequence=1
http://www.mdpi.com/2076-3417/8/1/150/htm
https://arxiv.org/pdf/1511.06939.pdf
https://link.springer.com/chapter/10.1007/978-3-319-73600-6_11
https://link.springer.com/chapter/10.1007/978-3-319-73603-7_44
https://arxiv.org/abs/1711.00927
https://arxiv.org/abs/1803.02421
https://arxiv.org/abs/1803.02353
http://jingxixu.com/files/deeplearning.pdf
https://arxiv.org/abs/1803.05428
https://www.sciencedirect.com/science/article/pii/S0925231218302431
https://arxiv.org/abs/1705.09792
http://www.apsipa.org/proceedings/2017/CONTENTS/papers2017/15DecFriday/FA-01/FA-01.4.pdf
https://arxiv.org/pdf/1611.06265.pdf
https://arxiv.org/abs/1802.09221
https://github.com/remyhuang/pop-music-highlighter
https://remyhuang.github.io/files/huang17ismir-lbd.pdf
https://remyhuang.github.io/files/huang17apsipa.pdf
https://github.com/markostam : multiple DL applied to CSI, ...
https://link.springer.com/article/10.1007/s11265-018-1334-2
https://www.researchgate.net/publication
https://qmro.qmul.ac.uk/xmlui/bitstream/handle/123456789/34703/Ycart%20Polyphonic%20Music%20Sequence%202018%20Accepted.pdf?sequence=3/323184729_BachProp_Learning_to_Compose_Music_in_Multiple_Styles
https://www.sciencedirect.com/science/article/pii/S1574954117302467
http://cs229.stanford.edu/proj2017/final-reports/5242716.pdf
https://github.com/keunwoochoi/LSTMetallica
https://arxiv.org/abs/1802.08370
http://cs229.stanford.edu/proj2017/final-reports/5241796.pdf
https://github.com/pawelpeksa/music_emotion_recognition_neuralnets
https://arxiv.org/abs/1802.06432
https://arxiv.org/abs/1705.10843
http://cs229.stanford.edu/proj2017/final-reports/5244969.pdf
https://github.com/tatsuyah/deep-improvisation
http://media.aau.dk/smc/ml4audio/
http://papers.nips.cc/paper/6146-soundnet-learning-sound-representations-from-unlabeled-video.pdf
https://github.com/deepsound-project/genre-recognition
https://github.com/umbrellabeach/music-generation-with-DL
https://github.com/corticph/MSTmodel
https://arxiv.org/pdf/1706.09588.pdf
https://arxiv.org/abs/1802.05162
https://link.springer.com/article/10.1007/s10844-018-0497-4
https://github.com/devicehive/devicehive-audio-analysis
https://arxiv.org/abs/1802.04051
https://arxiv.org/abs/1802.04208
https://magenta.tensorflow.org/onsets-frames
dblp.uni-trier.de/db/conf/icmc/icmc2002 (ctrl+f neural network)
http://tandfonline.com/doi/full/10.1080/09298215.2017.1367820?af=R&
https://arxiv.org/pdf/1801.01589.pdf
https://arxiv.org/abs/1802.03144
https://arxiv.org/pdf/1712.00866.pdf
https://www.linux.ime.usp.br/~iancarv/mac0499/tcc.pdf
https://arxiv.org/abs/1802.06182
https://arxiv.org/abs/1712.05119
https://arxiv.org/abs/1712.05274
https://github.com/jakobabesser/walking_bass_transcription_dnn
https://towardsdatascience.com/how-i-created-a-classifier-to-determine-the-potential-popularity-of-a-song-6d63093ba221
https://github.com/ds7711/music_genre_classification
https://arxiv.org/abs/1710.10451
https://arxiv.org/abs/1712.07799
https://arxiv.org/pdf/1712.08370.pdf
https://arxiv.org/abs/1710.10974
https://arxiv.org/abs/1802.05178
https://arxiv.org/pdf/1703.01789.pdf
https://arxiv.org/abs/1711.05772
https://arxiv.org/abs/1801.01589
https://arxiv.org/abs/1712.09668
https://github.com/unnati-xyz/music-generation
https://github.com/calclavia/DeepJ and https://arxiv.org/pdf/1801.00887.pdf
https://scholar.google.fr/scholar?hl=fr&as_sdt=0%2C5&q=Automatic+Programming+of+VST+Sound+Synthesizers+using+Deep+Networks+and+Other+Techniques+MJ+Yee-King%2C+L+Fedden%2C+M+d%27Inverno&btnG=
https://arxiv.org/ftp/arxiv/papers/1712/1712.01011.pdf
https://github.com/AI-ON/Few-Shot-Music-Generation
https://christophm.github.io/interpretable-ml-book/
https://github.com/dshieble/Music_RNN_RBM
https://github.com/feynmanliang/bachbot
https://github.com/awjuliani/sound-cnn
https://github.com/robbiebarrat/rapping-neural-network
https://www.researchgate.net/publication/322977005_Audio_Event_Detection_Using_Multiple-Input_Convolutional_Neural_Network
https://arxiv.org/abs/1712.04371
https://arxiv.org/abs/1712.01011
https://arxiv.org/abs/1707.09219
https://arxiv.org/abs/1712.05901
https://arxiv.org/abs/1712.06076
https://arxiv.org/abs/1712.02898
https://arxiv.org/abs/1712.03228
https://arxiv.org/abs/1712.04382
https://arxiv.org/abs/1712.01456
https://arxiv.org/abs/1712.03835
https://arxiv.org/abs/1712.00334
https://arxiv.org/abs/1712.00640
https://arxiv.org/abs/1712.00866
https://arxiv.org/abs/1712.00254
https://arxiv.org/pdf/1712.05119.pdf
https://arxiv.org/abs/1712.00166
https://arxiv.org/pdf/1711.11160.pdf
https://arxiv.org/pdf/1711.08976.pdf
https://github.com/drscotthawley/panotti
https://arxiv.org/abs/1703.10847
http://www.music-ir.org/mirex/abstracts/2017/LPNKK1.pdf
http://www.music-ir.org/mirex/abstracts/2017/PLNPH1.pdf
https://github.com/zhangqianhui/AdversarialNetsPapers
https://github.com/LqNoob/MelodyExtraction-MCDNN
https://github.com/EdwardLin2014/CNN-with-IBM-for-Singing-Voice-Separation
https://github.com/posenhuang/deeplearningsourceseparation
https://github.com/minzwon/kakao/blob/master/analyzing.ipynb
https://www.researchgate.net/publication/278662921_Deep_Image_Features_in_Music_Information_Retrieval
https://arxiv.org/abs/1611.09827v2
https://arxiv.org/abs/1711.08976
https://github.com/kkp15/kkp15.github.io
https://www.sciencedirect.com/science/article/pii/S0925231217317666
https://arxiv.org/pdf/1710.11428.pdf (http://mirlab.org:8080/demo/SVSGAN/)
www.karindressler.de/papers/dissertation_dressler.pdf
https://arxiv.org/abs/1705.09792
http://ieeexplore.ieee.org/abstract/document/8103116/
https://arxiv.org/abs/1709.04384
https://arxiv.org/abs/1711.05772
https://ismir2017.smcnus.org/lbds/Kim2017a.pdf
https://arxiv.org/pdf/1711.04845.pdf
https://ismir2017.smcnus.org/lbds/Schedl2017.pdf
https://lib.ugent.be/fulltxt/RUG01/002/367/502/RUG01-002367502_2017_0001_AC.pdf
https://arxiv.org/abs/1412.6596
https://arxiv.org/abs/1706.02361
https://github.com/jthickstun/thickstun2017learning
https://arxiv.org/abs/1707.09219
https://arxiv.org/abs/1711.01369
https://arxiv.org/abs/1710.11428
https://arxiv.org/abs/1706.02361
https://arxiv.org/abs/1711.04480
https://arxiv.org/abs/1706.06525
https://github.com/qiuqiangkong/ICASSP2018_audioset
https://arxiv.org/abs/1707.05589
https://www.researchgate.net/publication/320632662_Music_Genre_Classification_Using_Masked_Conditional_Neural_Networks
https://arxiv.org/pdf/1711.02209.pdf
https://arxiv.org/pdf/1711.01369.pdf
https://arxiv.org/pdf/1711.00927.pdf
https://arxiv.org/abs/1709.06298
https://arxiv.org/abs/1709.04384
https://ismir2017.smcnus.org/lbds/Suh2017.pdf
https://ismir2017.smcnus.org/lbds/Pons2017.pdf
https://link.springer.com/chapter/10.1007/978-3-319-69911-0_14
https://www.preprints.org/manuscript/201711.0027/v1
https://github.com/Js-Mim
https://www.researchgate.net/publication/320867112_Audio_Set_classification_with_attention_model_A_probabilistic_perspective
https://arxiv.org/abs/1711.00351
https://arxiv.org/pdf/1710.10451.pdf
https://arxiv.org/pdf/1710.11153.pdf
http://www.cs.tut.fi/sgn/arg/dcase2017/documents/challenge_technical_reports/DCASE2017_Piczak_208.pdf
http://www.cs.tut.fi/sgn/arg/dcase2017/documents/challenge_technical_reports/DCASE2017_Maka_203.pdf
https://arxiv.org/abs/1711.00913
https://arxiv.org/abs/1711.00927
https://arxiv.org/find/cs/1/au:+Oord_A/0/1/0/all/0/1 & https://avdnoord.github.io/homepage/vqvae/
https://github.com/qiuqiangkong/ICASSP2018_joint_separation_classification
https://www.researchgate.net/publication/320859133_SymCHM-An_Unsupervised_Approach_for_Pattern_Discovery_in_Symbolic_Music_with_a_Compositional_Hierarchical_Model
https://www.researchgate.net/publication/315570382_Single_Channel_Audio_Source_Separation_using_Convolutional_Denoising_Autoencoders
https://github.com/andabi/music-source-separation
https://github.com/andabi/deep-voice-conversion
https://arxiv.org/abs/1711.00229
https://arxiv.org/abs/1711.00048
https://arxiv.org/pdf/1710.11549.pdf
https://arxiv.org/abs/1711.02209
https://arxiv.org/abs/1710.11473
https://arxiv.org/abs/1710.11428
https://arxiv.org/abs/1710.11418
https://arxiv.org/abs/1710.11385
https://arxiv.org/abs/1710.11153
http://danetapi.com/chimera
https://arxiv.org/abs/1710.10451
https://github.com/lamtharnhantrakul/audio_kernels
https://github.com/Impro-Visor/lstmprovisor-python
https://github.com/hexahedria/biaxial-rnn-music-composition
https://github.com/rabitt/ismir2017-deepsalience
https://www.researchgate.net/publication/313895490_Comparing_Shallow_versus_Deep_Neural_Network_Architectures_for_Automatic_Music_Genre_Classification
https://github.com/marl/crepe
https://www.semanticscholar.org/search?year%5B%5D=1991&year%5B%5D=2017&q=deep%20learning%20music%20audio%20neural%20network&sort=relevance
http://rodrigob.github.io/are_we_there_yet/build/
https://github.com/syhw/wer_are_we
https://www.researchgate.net/publication/320589850_Masked_Conditional_Neural_Networks_for_Audio_Classification
https://arxiv.org/pdf/1606.04930.pdf
https://github.com/emilylawton/deep-learning-resources
https://www.audiolabs-erlangen.de/resources/MIR/2017-GI-Tutorial-Musik/2017_MuellerWeissBalke_GI_DeepLearningMIR.pdf
https://books.google.fr/books?hl=fr&lr=&id=1_06DwAAQBAJ&oi=fnd&pg=PA237&ots=QHQvylLIO7&sig=pSqGpvQxa9RUX601lf40mQBPDX8#v=onepage&q&f=false
http://cmmr2017.inesctec.pt/wp-content/uploads/2017/09/43_CMMR_2017_paper_31.pdf
https://ismir2017.smcnus.org/wp-content/uploads/2017/10/217_Paper.pdf
https://ismir2017.smcnus.org/wp-content/uploads/2017/10/77_Paper.pdf
https://ismir2017.smcnus.org/wp-content/uploads/2017/10/28_Paper.pdf
https://ismir2017.smcnus.org/wp-content/uploads/2017/10/91_Paper.pdf
https://ismir2017.smcnus.org/wp-content/uploads/2017/10/17_Paper.pdf
https://ismir2017.smcnus.org/wp-content/uploads/2017/10/123_Paper.pdf
https://ismir2017.smcnus.org/wp-content/uploads/2017/10/137_Paper.pdf
PDF: musicalmetacreation.org/buddydrive/file/smith/ & source : http://musicalmetacreation.org/proceedings__trashed/mume-2017/
https://www.researchgate.net/publication/320519760_Musical_Query-by-Semantic-Description_Based_on_Convolutional_Neural_Network
https://www.researchgate.net/publication/314382920_Inside_the_Spectrogram_Convolutional_Neural_Networks_in_Audio_Processing
http://ieeexplore.ieee.org/abstract/document/8073570/
https://ismir2017.smcnus.org/wp-content/uploads/2017/10/135_Paper.pdf
https://www.researchgate.net/publication/320488483_Acoustic_Scene_Classification_by_Combining_Autoencoder-Based_Dimensionality_Reduction_and_Convolutional_Neural_Networks
https://www.mendeley.com/research-papers/deep-multimodal-approach-coldstart-music-recommendation-1/?dgcid=raven_md_feed_email
https://www.mendeley.com/research-papers/classification-audio-signals-using-svm-rbfnn-1/?dgcid=raven_md_feed_email
https://ismir2017.smcnus.org/wp-content/uploads/2017/10/9_Paper.pdf
https://link.springer.com/chapter/10.1007/978-3-319-68121-4_18
https://ismir2017.smcnus.org/wp-content/uploads/2017/10/164_Paper.pdf
https://www.researchgate.net/publication/320416485_A_System_for_2017_DCASE_Challenge_Using_Deep_Sequenrial_Image_and_Wavelet_Features?discoverMore=1
https://www.researchgate.net/publication/315100151_Improving_music_source_separation_based_on_deep_neural_networks_through_data_augmentation_and_network_blending
https://www.researchgate.net/publication/320333553_Data_augmentation_for_deep_learning_source_separation_of_HipHop_songs
https://github.com/karoldvl/paper-2017-DCASE
https://repositori.upf.edu/bitstream/handle/10230/32919/Martel_2017.pdf?sequence=1&isAllowed=y
https://www.researchgate.net/publication/320333553_Data_augmentation_for_deep_learning_source_separation_of_HipHop_songs?discoverMore=1
https://arxiv.org/abs/1710.04288
http://www.cs.tut.fi/sgn/arg/dcase2017/documents/challenge_technical_reports/DCASE2017_Lee_201.pdf
http://www.cs.tut.fi/sgn/arg/dcase2017/documents/challenge_technical_reports/DCASE2017_Yu_188.pdf
http://www.semanticaudio.co.uk/wp-content/uploads/2017/09/WIMP2017_Martinez-RamirezReiss.pdf
https://arxiv.org/pdf/1609.04243.pdf
https://arxiv.org/abs/1711.01634
https://arxiv.org/pdf/1706.02361.pdf
https://github.com/RichardYang40148/MidiNet/tree/master/v1
https://ismir2017.smcnus.org/programschedule/
https://twitter.com/keunwoochoi/status/912341967648018435
http://ieeexplore.ieee.org/abstract/document/8049362/ A data-driven model of tonal chord sequence complexity Bruno Di Giorgi ; Simon Dixon ; Massimiliano Zanoni ; Augusto Sarti 2017
https://www.researchgate.net/publication/282997080_A_survey_Time_travel_in_deep_learning_space_An_introduction_to_deep_learning_models_and_how_deep_learning_models_evolved_from_the_initial_ideas
https://www.researchgate.net/publication/317265107_Attention_and_Localization_Based_on_a_Deep_Convolutional_Recurrent_Model_for_Weakly_Supervised_Audio_Tagging
https://www.researchgate.net/publication/319276246_A_Recurrent_Encoder-Decoder_Approach_With_Skip-Filtering_Connections_for_Monaural_Singing_Voice_Separation
https://www.researchgate.net/publication/296704118_Deep_Neural_Networks_for_Dynamic_Range_Compression_in_Mastering_Applications
http://c4dm.eecs.qmul.ac.uk/news/news.2016-11-25.C4DM_Seminar_-_Tian_Cheng_and_Siddharth_Sigtia_(Video_Available).html
https://arxiv.org/abs/1703.08019
http://slim-sig.irisa.fr/me17/Mediaeval_2017_paper_49.pdf
https://groups.csail.mit.edu/sls/publications/2017/YuZhang_PhD_Thesis.pdf
https://dl.gi.de/bitstream/handle/20.500.12116/3859/B1-9.pdf?sequence=1&isAllowed=y
https://www.meetup.com/fr-FR/Berlin-Music-Information-Retrieval-Meetup/events/243855597/?eventId=243855597
https://www.researchgate.net/publication/315570382_Single_Channel_Audio_Source_Separation_using_Convolutional_Denoising_Autoencoders
https://www.researchgate.net/publication/320409290_Wavelets_Revisited_for_the_Classification_of_Acoustic_Scenes
https://scholar.google.fr/citations?user=YOY2MFEAAAAJ&hl=fr&oi=sra
https://www.researchgate.net/publication/314382920_Inside_the_Spectrogram_Convolutional_Neural_Networks_in_Audio_Processing
http://benanne.github.io/2014/08/05/spotify-cnns.html
https://vaplab.ee.ncu.edu.tw/english/pcchang/pdf/j52.pdf
https://github.com/auDeep/auDeep
https://www.cs.tut.fi/sgn/arg/dcase2017/documents/challenge_technical_reports/DCASE2017_Amiriparian_173.pdf
https://qmro.qmul.ac.uk/xmlui/bitstream/handle/123456789/25936/QUINTON_Elio_Final_PhD_030817.pdf?sequence=1
https://csmc2017.wordpress.com/proceedings/
http://ofai.at/~jan.schlueter/
https://www.audiolabs-erlangen.de/fau/assistant/balke/publications Deep Learning for Jazz Walking Bass Transcription
https://link.springer.com/chapter/10.1007/978-3-319-63450-0_14
https://www.researchgate.net/publication/318030697_Multi-scale_Multi-band_DenseNets_for_Audio_Source_Separation
https://www.researchgate.net/publication/282001406_Deep_neural_network_based_instrument_extraction_from_music
https://www.researchgate.net/publication/315100151_Improving_music_source_separation_based_on_deep_neural_networks_through_data_augmentation_and_network_blending
http://ieeexplore.ieee.org/abstract/document/7994970/
http://www.semanticaudio.co.uk/wp-content/uploads/2017/09/WIMP2017_Martinez-RamirezReiss.pdf
https://github.com/search?utf8=%E2%9C%93&q=deep+learning+music&type=
https://www.researchgate.net/publication/318030697_Multi-scale_Multi-band_DenseNets_for_Audio_Source_Separation?_esc=Profile%3A%3AInterests&_iepl%5BviewId%5D=1VIp27Fb9rzMbMunG8OwuWAr&_iepl%5BprofilePublicationItemVariant%5D=default&_iepl%5Bcontexts%5D%5B0%5D=prfipi&_iepl%5BtargetEntityId%5D=PB%3A318030697&_iepl%5BinteractionType%5D=publicationTitle
https://arxiv.org/abs/1706.07162
https://github.com/oriolromani/MIRdeepLearning
https://link.springer.com/chapter/10.1007/978-3-319-68612-7_40
https://arxiv.org/abs/1703.09039
Convolution-based Classification of Audio and Symbolic Representations of Music. Gissel Velarde, Carlos Cancino Chacón, David Meredith, Tillman Weyde and Maarten Grachten. October 22, 2016 (unpublished)

DL4M 2018 articles (to be considered after dealing with 2017):
https://ieeexplore.ieee.org/stamp/stamp.jsp?arnumber=8332139
https://arxiv.org/abs/1711.00048 ICASSP2018
https://www.tandfonline.com/doi/full/10.1080/09298215.2018.1438476?af=R
https://arxiv.org/abs/1803.04357
https://arxiv.org/abs/1803.04030
https://arxiv.org/abs/1709.01674
https://arxiv.org/abs/1803.06841
https://arxiv.org/abs/1804.04053
https://arxiv.org/abs/1803.06841
https://arxiv.org/abs/1803.08629
https://arxiv.org/abs/1804.00047
https://arxiv.org/abs/1804.00525
https://arxiv.org/ftp/arxiv/papers/1804/1804.02918.pdf
https://arxiv.org/abs/1804.04212
http://www.mdpi.com/2076-3417/8/4/507/htm
https://arxiv.org/abs/1801.07141
https://arxiv.org/abs/1705.06979
https://arxiv.org/abs/1612.04742
https://www.researchgate.net/publication/322216935_Jazz_music_sub-genre_classification_using_deep_learning
https://www.researchgate.net/profile/Loris_Nanni/publication/323938467_Ensemble_of_deep_learning_visual_and_acoustic_features_for_music_genre_classification/links/5ab52e3745851515f599c5da/Ensemble-of-deep-learning-visual-and-acoustic-features-for-music-genre-classification.pdf
http://www.mdpi.com/2076-3417/8/4/606/htm
https://github.com/pkmital/time-domain-neural-audio-style-transfer
https://arxiv.org/abs/1804.07145
https://arxiv.org/abs/1804.07297
https://arxiv.org/abs/1803.01271
https://hal-lirmm.ccsd.cnrs.fr/lirmm-01766781/document
https://arxiv.org/abs/1804.07300
https://arxiv.org/abs/1804.07690
https://arxiv.org/abs/1804.08300
https://arxiv.org/abs/1804.08167
https://dl.acm.org/citation.cfm?id=3191822
https://dl.acm.org/citation.cfm?id=3191823
https://arxiv.org/abs/1709.00611
https://arxiv.org/abs/1804.09399
https://arxiv.org/abs/1804.02918
https://arxiv.org/abs/1804.09808
https://arxiv.org/abs/1804.07297
https://dspace.library.uvic.ca/bitstream/handle/1828/9264/Singh_Harpreet_MSc_2018.pdf?sequence=3&isAllowed=y
https://arxiv.org/pdf/1804.09202.pdf
https://arxiv.org/abs/1805.00237 with https://github.com/jordipons/elmarc
https://github.com/NarainKrishnamurthy/BeatGAN2.0
https://arxiv.org/abs/1804.09808
https://github.com/johnglover/sound-rnn
https://github.com/NadzeyaKadakova/Studies/blob/master/95-jazznet/Jazz%20Solo%20with%20an%20LSTM%20Network%20.ipynb
https://www.politesi.polimi.it/bitstream/10589/139073/1/tesi.pdf
https://arxiv.org/abs/1805.02043
https://arxiv.org/abs/1805.02603
https://arxiv.org/abs/1805.03647
https://github.com/gantheory/playlist-cleaning
https://arxiv.org/pdf/1805.02410.pdf
https://arxiv.org/abs/1803.01271
https://ieeexplore.ieee.org/abstract/document/8356323/
https://arxiv.org/abs/1805.05324
https://marl.smusic.nyu.edu/nieto/publications/TISMIR2018.pdf
http://www.aes.org/e-lib/browse.cfm?elib=19513
https://arxiv.org/abs/1805.07848
https://arxiv.org/abs/1805.08559
https://arxiv.org/abs/1805.08501
https://arxiv.org/abs/1805.10808
https://arxiv.org/abs/1804.00525
https://arxiv.org/abs/1805.10548
https://arxiv.org/abs/1805.12176
https://arxiv.org/abs/1806.00195
https://arxiv.org/abs/1801.10492
https://arxiv.org/abs/1806.00509
https://arxiv.org/abs/1806.00770
https://arxiv.org/abs/1806.01180
https://arxiv.org/abs/1805.08559 (https://github.com/sungheonpark/music_source_sepearation_SH_net)
https://arxiv.org/abs/1806.08724
https://arxiv.org/abs/1806.08686

Some speech articles:
https://arxiv.org/pdf/1710.09798.pdf
https://arxiv.org/abs/1804.02918
https://infoscience.epfl.ch/record/203464/files/Palaz_Idiap-RR-18-2014.pdf
https://link.springer.com/chapter/10.1007/978-3-319-66429-3_2
https://www.researchgate.net/profile/Cong-Thanh_Do/publication/319269623_Improved_Automatic_Speech_Recognition_Using_Subband_Temporal_Envelope_Features_and_Time-Delay_Neural_Network_Denoising_Autoencoder/links/599f388a4585151e3c6acdd8/Improved-Automatic-Speech-Recognition-Using-Subband-Temporal-Envelope-Features-and-Time-Delay-Neural-Network-Denoising-Autoencoder.pdf
https://arxiv.org/pdf/1708.08740.pdf
http://newiranians.ir/TASLP2339736-proof.pdf
https://asmp-eurasipjournals.springeropen.com/articles/most-recent/rss.xml
https://arxiv.org/pdf/1709.00308.pdf
https://www.researchgate.net/publication/312520074_A_review_on_Deep_Learning_approaches_in_Speaker_Identification
https://www.researchgate.net/publication/317711457_A_Hybrid_Approach_with_Multi-channel_I-Vectors_and_Convolutional_Neural_Networks_for_Acoustic_Scene_Classification
https://www.researchgate.net/publication/320180136_Large-scale_weakly_supervised_audio_classification_using_gated_convolutional_neural_network

Propose Logo

Hi. @ybayle I'm a graphic designer and an open source enthusiast. I wanna make contribution for your project, what do you think ?
42413020-36c0e0ce-8242-11e8-86c6-a2589d5bfcab-1-1-1
Best regard. Mirza Zulfan

Source code of the MASTERO paper (2019) is found here

As title.
The source code for "Enabling factorized piano music modeling and generation with the MAESTRO dataset" is listed as unknown.
It is now a part of Magenta and is found here.

There are also 2 relevant papers in the same repo but not listed here:
Onsets and Frames: Dual-Objective Piano Transcription, and
Improving Perceptual Quality of Drum Transcription with the Expanded Groove MIDI Dataset.

It would be great to add this repo and these 2 papers to the list.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.