Essential Maintenance: All Authorea-powered sites will be offline 4pm-6pm EDT Tuesday 28 May for essential maintenance.
We apologise for any inconvenience.

loading page

Dilated Convolutional Model for Melody Extraction
  • Xian Wang ,
  • Lingqiao Liu ,
  • Javen Shi
Xian Wang
The University of Adelaide

Corresponding Author:[email protected]

Author Profile
Lingqiao Liu
Author Profile
Javen Shi
Author Profile


Melody extraction is a challenging task in music information retrieval that enables many down-stream applications. In this paper we propose a simple dilated convolutional model for melody extraction. It takes variable-q transforms as inputs. It first uses consecutive layers of convolution to capture local temporal-frequency patterns. Afterward, it relies only a single layer of dilated convolution for capturing global frequency patterns formed by the pitches and harmonics of active notes. This model is effective in that it achieves the-state-of-the-art performance on most datasets, for both general and vocal melody extraction. In addition, it gets the best performance with the least training data.