dc.creatorPadilla, Victor
dc.creatorConklin, Darrell
dc.date.accessioned2022-02-08T09:57:27Z
dc.date.accessioned2023-03-07T19:34:37Z
dc.date.available2022-02-08T09:57:27Z
dc.date.available2023-03-07T19:34:37Z
dc.date.created2022-02-08T09:57:27Z
dc.identifier1989-1660
dc.identifierhttps://reunir.unir.net/handle/123456789/12408
dc.identifierhttp://doi.org/10.9781/ijimai.2018.10.002
dc.identifier.urihttps://repositorioslatinoamericanos.uchile.cl/handle/2250/5906704
dc.description.abstractGenerating new music based on rules of counterpoint has been deeply studied in music informatics. In this article, we try to go further, exploring a method for generating new music based on the style of Palestrina, based on combining statistical generation and pattern discovery. A template piece is used for pattern discovery, and the patterns are selected and organized according to a probabilistic distribution, using horizontal viewpoints to describe melodic properties of events. Once the template is covered with patterns, two-voice counterpoint in a florid style is generated into those patterns using a first-order Markov model. The template method solves the problem of coherence and imitation never addressed before in previous research in counterpoint music generation. For constructing the Markov model, vertical slices of pitch and rhythm are compiled over a large corpus of dyads from Palestrina masses. The template enforces different restrictions that filter the possible paths through the generation process. A double backtracking algorithm is implemented to handle cases where no solutions are found at some point within a generation path. Results are evaluated by both information content and listener evaluation, and the paper concludes with a proposed relationship between musical quality and information content. Part of this research has been presented at SMC 2016 in Hamburg, Germany.
dc.languageeng
dc.publisherInternational Journal of Interactive Multimedia and Artificial Intelligence (IJIMAI)
dc.relation;vol. 5, nº 3
dc.relationhttps://www.ijimai.org/journal/bibcite/reference/2692
dc.rightsopenAccess
dc.subjectartificial intelligence
dc.subjectmusic informatics
dc.subjectmusic generation
dc.subjectsequential pattern mining
dc.subjectstatistical models of music
dc.subjectIJIMAI
dc.titleGeneration of Two-Voice Imitative Counterpoint from Statistical Models
dc.typearticle


Este ítem pertenece a la siguiente institución