Extending music based on emotion and tonality via generative adversarial network

Bo Wei Tseng, Yih Liang Shen, Tai Shih Chi

研究成果: Conference contribution同行評審

3 引文 斯高帕斯(Scopus)

摘要

We propose a generative model for music extension in this paper. The model is composed of two classifiers, one for music emotion and one for music tonality, and a generative adversarial network (GAN). Therefore, it can generate symbolic music not only based on low level spectral and temporal characteristics, but also on high level emotion and tonality attributes of previously observed music pieces. The generative model works in a universal latent space constructed by the variational autoencoder (VAE) for representing music pieces. We conduct subjective listening tests and derive objective measures for performance evaluation. Experimental results show that the proposed model produces much smoother and more authentic music pieces than the baseline model in terms of all subjective and objective measures.

原文English
主出版物標題2021 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2021 - Proceedings
發行者Institute of Electrical and Electronics Engineers Inc.
頁面86-90
頁數5
ISBN(電子)9781728176055
DOIs
出版狀態Published - 6月 2021
事件2021 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2021 - Virtual, Toronto, 加拿大
持續時間: 6 6月 202111 6月 2021

出版系列

名字ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings
ISSN(列印)1520-6149

Conference

Conference2021 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2021
國家/地區加拿大
城市Virtual, Toronto
期間6/06/2111/06/21

指紋

深入研究「Extending music based on emotion and tonality via generative adversarial network」主題。共同形成了獨特的指紋。

引用此