dc.contributorLopes, Heitor Silverio
dc.contributorhttps://orcid.org/0000-0003-3984-1432
dc.contributorhttp://lattes.cnpq.br/4045818083957064
dc.contributorGabardo, Ademir Cristiano
dc.contributorhttp://lattes.cnpq.br/9872210667199371
dc.contributorLopes, Heitor Silverio
dc.contributorhttps://orcid.org/0000-0003-3984-1432
dc.contributorhttp://lattes.cnpq.br/4045818083957064
dc.contributorDorini, Leyza Elmeri Baldo
dc.contributorhttps://orcid.org/0000-0002-0483-3435
dc.contributorhttp://lattes.cnpq.br/5726947194230379
dc.contributorRibeiro, Manasses
dc.contributorhttps://orcid.org/0000-0002-7526-5092
dc.contributorhttp://lattes.cnpq.br/6475893755893056
dc.creatorBerno, Brenda Cinthya Solari
dc.date.accessioned2021-07-06T00:35:41Z
dc.date.accessioned2022-12-06T14:54:37Z
dc.date.available2021-07-06T00:35:41Z
dc.date.available2022-12-06T14:54:37Z
dc.date.created2021-07-06T00:35:41Z
dc.date.issued2021-05-21
dc.identifierBERNO, Brenda Cinthya Solari. Sketch-Based multimodal image retrieval using deep learning. 2021. Dissertação (Mestrado em Engenharia Elétrica e Informática Industrial) - Universidade Tecnológica Federal do Paraná, Curitiba, 2021.
dc.identifierhttp://repositorio.utfpr.edu.br/jspui/handle/1/25496
dc.identifier.urihttps://repositorioslatinoamericanos.uchile.cl/handle/2250/5257859
dc.description.abstractThe constant growth of multimedia data generated every day makes it increasingly difficult to retrieve it. Google is known to do a good job of retrieving documents by searching for keyword matches. However, multimedia data hardly contain keywords that identify them. The main objective of this work is to retrieve a photographic image using another modality different from that of the photograph, such as a sketch. A sketch is different from the image since it is a set of hand-drawn lines and colors and texture is lost, when compared with a photograph that is a more complex visual representation representing the real world. The selected study case for this method is tattoo photograph retrieval using sketches. Due to the lack of appropriate data for this study, a new dataset of sketches and tattoo images was created. The proposed model consists of a Siamese neural network that receives as input visual features previously extracted from each modality to learn an optimal representation for photographs and sketches within an embedded space, where the image of a class is close to the sketch of the same class. Two cost functions were tested, and experiments showed that the contrastive loss function achieved better results than the triplet loss function in the retrieval of images. Despite having limited data, in the image retrieval experiments the average precision achieved 85% precision for our dataset at top-5 results and 85% precision for Sketchy at top-10 results. We observed that retrieval results depend on the quality and diversity of the data used for training, especially in sketch-based image retrieval, which, in turn, depends on the user’s ability to draw. Overall, the proposed methods are promising and results encourage further research. Future works include the extension of the dataset (both tattoo images and sketches) and, also, experiments with other modalities.
dc.publisherUniversidade Tecnológica Federal do Paraná
dc.publisherCuritiba
dc.publisherBrasil
dc.publisherPrograma de Pós-Graduação em Engenharia Elétrica e Informática Industrial
dc.publisherUTFPR
dc.rightshttp://creativecommons.org/licenses/by/4.0/
dc.rightsopenAccess
dc.subjectSistemas multimídia
dc.subjectRecuperação de dados (Computação)
dc.subjectSistemas de recuperação da informação
dc.subjectRedes neurais (Computação)
dc.subjectVisão Computacional
dc.subjectAprendizado do computador
dc.subjectTatuagem - Imagem
dc.subjectMultimedia systems
dc.subjectData recovery (Computer science)
dc.subjectInformation storage and retrieval systems
dc.subjectNeural networks (Computer science)
dc.subjectComputer vision
dc.subjectMachine learning
dc.subjectTattooing - Imaging
dc.titleSketch-Based multimodal image retrieval using deep learning
dc.typemasterThesis


Este ítem pertenece a la siguiente institución