dc.creatorYsique-Neciosup, Jose
dc.creatorMercado-Chavez, Nilton
dc.creatorUgarte, Willy
dc.date.accessioned2022-09-08T12:17:28Z
dc.date.accessioned2024-05-07T03:13:41Z
dc.date.available2022-09-08T12:17:28Z
dc.date.available2024-05-07T03:13:41Z
dc.date.created2022-09-08T12:17:28Z
dc.date.issued2022-01-01
dc.identifier15464261
dc.identifier10.1002/cav.2110
dc.identifierhttp://hdl.handle.net/10757/660896
dc.identifier1546427X
dc.identifierComputer Animation and Virtual Worlds
dc.identifier2-s2.0-85136901947
dc.identifierSCOPUS_ID:85136901947
dc.identifier.urihttps://repositorioslatinoamericanos.uchile.cl/handle/2250/9329817
dc.description.abstractDeep learning models have shown that it is possible to train neural networks to dispense, to a lesser or greater extent, with the need for human intervention for the task of image animation, which helps to reduce not only the production time of these audiovisual pieces, but also presents benefits with respect to the economic investment they require to be made. However, these models suffer from two common problems: the animations they generate are of very low resolution and they require large amounts of training data to generate good results. To deal with these issues, this article introduces the architectural modification of a state-of-the-art image animation model integrated with a video super-resolution model to make the generated videos more visually pleasing to viewers. Although it is possible to train the animation models with higher resolution images, the time it would take to train them would be much longer, which does not necessarily benefit the quality of the animation, so it is more efficient to complement it with another model focused on improving the animation resolution of the generated video as we demonstrate in our results. We present the design and implementation of a convolutional neural network based on an state-of-art model focused on the image animation task, which is trained with a set of facial data from videos extracted from the YouTube platform. To determine which of all the modifications to the selected state-of-the-art model architecture is better, the results are compared with different metrics that evaluate the performance in image animation and video quality enhancement tasks. The results show that modifying the architecture of the model focused on the detection of characteristic points significantly helps to generate more anatomically and visually attractive videos. In addition, perceptual testing with users shows that using a super-resolution video model as a plugin helps generate more visually appealing videos.
dc.languageeng
dc.publisherJohn Wiley and Sons Ltd
dc.relationhttps://onlinelibrary.wiley.com/doi/10.1002/cav.2110
dc.rightsinfo:eu-repo/semantics/embargoedAccess
dc.sourceComputer Animation and Virtual Worlds
dc.subjectconvolutional neural network
dc.subjectimage animation
dc.subjectkeypoints
dc.subjectU-Net
dc.subjectvideo super-resolution
dc.titleDeepHistory: A convolutional neural network for automatic animation of museum paintings
dc.typeinfo:eu-repo/semantics/article


Este ítem pertenece a la siguiente institución