TY - JOUR
T1 - Automatic violin synthesis using expressive musical term features
AU - Yang, Chih Hong
AU - Li, Pei Ching
AU - Su, Alvin W.Y.
AU - Su, Li
AU - Yang, Yi Hsuan
N1 - Funding Information:
The authors would like to thank the Ministry of Science and Technology of Taiwan for its financial support of this work, under contract MOST 103-2221-E-006-140-MY3.
PY - 2016
Y1 - 2016
N2 - The control of interpretational properties such as duration, vibrato, and dynamics is important in music performance. Musicians continuously manipulate such properties to achieve different expressive intentions. This paper presents a synthesis system that automatically converts a mechanical, deadpan interpretation to distinct expressions by controlling these expressive factors. Extending from a prior work on expressive musical term analysis, we derive a subset of essential features as the control parameters, such as the relative time position of the energy peak in a note and the mean temporal length of the notes. An algorithm is proposed to manipulate the energy contour (i.e. for dynamics) of a note. The intended expressions of the synthesized sounds are evaluated in terms of the ability of the machine model developed in the prior work. Ten musical expressions such as Risoluto and Maestoso are considered, and the evaluation is done using held-out music pieces. Our evaluations show that it is easier for the machine to recognize the expressions of the synthetic version, comparing to those of the real recordings of an amateur student. While a listening test is under construction as a next step for further performance validation, this work represents to our best knowledge a first attempt to build and quantitatively evaluate a system for EMT analysis/synthesis.
AB - The control of interpretational properties such as duration, vibrato, and dynamics is important in music performance. Musicians continuously manipulate such properties to achieve different expressive intentions. This paper presents a synthesis system that automatically converts a mechanical, deadpan interpretation to distinct expressions by controlling these expressive factors. Extending from a prior work on expressive musical term analysis, we derive a subset of essential features as the control parameters, such as the relative time position of the energy peak in a note and the mean temporal length of the notes. An algorithm is proposed to manipulate the energy contour (i.e. for dynamics) of a note. The intended expressions of the synthesized sounds are evaluated in terms of the ability of the machine model developed in the prior work. Ten musical expressions such as Risoluto and Maestoso are considered, and the evaluation is done using held-out music pieces. Our evaluations show that it is easier for the machine to recognize the expressions of the synthetic version, comparing to those of the real recordings of an amateur student. While a listening test is under construction as a next step for further performance validation, this work represents to our best knowledge a first attempt to build and quantitatively evaluate a system for EMT analysis/synthesis.
UR - http://www.scopus.com/inward/record.url?scp=85030245783&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85030245783&partnerID=8YFLogxK
M3 - Conference article
AN - SCOPUS:85030245783
SN - 2413-6700
SP - 209
EP - 215
JO - Proceedings of the International Conference on Digital Audio Effects, DAFx
JF - Proceedings of the International Conference on Digital Audio Effects, DAFx
T2 - 19th International Conference on Digital Audio Effects, DAFx 2016
Y2 - 5 September 2016 through 9 September 2016
ER -