Title
Timbre-enhanced Multi-modal Music Style Transfer with Domain Balance Loss
Abstract
Style transfer of the polyphonic music recordings has always been a challenging task due to the difficulty of learning representations for both domain invariant (i.e. content) and domain-variant (i.e. style) features of the music. Although there exists prior works which employ the Multi-modal Unsupervised Image-to-Image Translation (MUNIT) framework to perform the music style transfer in an unsupervised manner and successfully provide the promising results, the gap between the transferred music recordings and the real ones is still noticeable. In order to reduce such gap, we propose and experiment several techniques for improving the transferred results, including the domain balanced loss, up-sampling, content discriminator, recycle loss, and the data scaling. We conduct extensive experiments on the tasks of bilateral style transfer among four different genres, namely: piano solo, guitar solo, string quartet, and chiptune. In evaluation, an objective testing scheme is proposed to investigate the pros and cons of all our proposed techniques, while we also design a subjective testing method for making comparison among different approaches and show that our proposed method is able to provide superior performance with respect to the prior works.
Year
DOI
Venue
2020
10.1109/TAAI51410.2020.00027
2020 International Conference on Technologies and Applications of Artificial Intelligence (TAAI)
Keywords
DocType
ISSN
recycle loss,bilateral style transfer,timbre-enhanced Multimodal music style transfer,domain balance loss,polyphonic music recordings,domain invariant,Image-to-Image Translation framework,unsupervised manner,transferred music recordings,transferred results,domain balanced loss
Conference
2376-6816
ISBN
Citations 
PageRank 
978-1-6654-4737-9
0
0.34
References 
Authors
0
5
Name
Order
Citations
PageRank
Tsai-Jyun Fan100.34
Chien-Yu Lu212112.08
Wei-Chen Chiu301.01
Li Su400.34
Che-Rung Lee596.64