Skip to content

klean2050/IC_Augment

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 

Repository files navigation

Augmentation Methods on Monophonic Audio for Instrument Classification in Polyphonic Music

Image

This is the official implementation code of the paper:

  • *A. Kratimenos, *K. Avramidis, C. Garoufis, A. Zlatintsi, & P. Maragos, "Augmentation Methods on Monophonic Audio for Instrument Classification in Polyphonic Music," 2020 28th European Signal Processing Conference (EUSIPCO), Amsterdam, 2021, pp. 156-160, doi: 10.23919/Eusipco47968.2020.9287745.

Abstract

Instrument classification is one of the fields in Music Information Retrieval (MIR) that has attracted a lot of research interest. However, the majority of that is dealing with monophonic music, while efforts on polyphonic material mainly focus on predominant instrument recognition or multi-instrument recognition for entire tracks. We present an approach for instrument classification in polyphonic music using monophonic training data that involves mixing-augmentation methods. Specifically, we experiment with pitch and tempo-based synchronization, as well as mixes of tracks with similar music genres. Further, a custom CNN model is proposed, that uses the augmented training data efficiently and a plethora of suitable evaluation metrics are discussed as well. The tempo-sync and genre techniques stand out, achieving an 81% label ranking average precision accuracy, detecting up to 9 instruments in over 2300 testing tracks.

Citation

If you find our paper useful in your research, please consider citing:

@inproceedings{kratimenos2020ic,
  title={Augmentation Methods on Monophonic Audio for Instrument Classification in Polyphonic Music},
  author={Kratimenos, A. and Avramidis, K. and Garoufis, C. and Zlatintsi, A. and Maragos, P.},
  booktitle={2020 28th European Signal Processing Conference (EUSIPCO)},
  pages={156-160},
  year={2020},
  doi={10.23919/Eusipco47968.2020.9287745}
}