dc.contributorUniversidade Estadual Paulista (Unesp)
dc.contributorUniversidade Federal de São Carlos (UFSCar)
dc.contributorUniversidade de São Paulo (USP)
dc.contributorFluminense Federal University (UFF)
dc.date.accessioned2018-12-11T17:13:31Z
dc.date.available2018-12-11T17:13:31Z
dc.date.created2018-12-11T17:13:31Z
dc.date.issued2017-07-01
dc.identifierJournal of Computer Science, v. 13, n. 6, p. 192-198, 2017.
dc.identifier1549-3636
dc.identifierhttp://hdl.handle.net/11449/174933
dc.identifier10.3844/jcssp.2017.192.198
dc.identifier2-s2.0-85025129320
dc.identifier2-s2.0-85025129320.pdf
dc.identifier4644812253875832
dc.identifier0000-0002-9325-3159
dc.description.abstractThe progressive growth in the volume of digital data has become a technological challenge of great interest in the field of computer science. That comes because, with the spread of personal computers and networks worldwide, content generation is taking larger proportions and very different formats from what had been usual until then. To analyze and extract relevant knowledge from these masses of complex and large volume data is particularly interesting, but before that, it is necessary to develop techniques to encourage their resilient storage. Very often, storage systems use a replication scheme for preserving the integrity of stored data. This involves generating copies of all information that, if lost by individual hardware failures inherent in any massive storage infrastructure, do not compromise access to what was stored. However, it was realized that accommodate such copies requires a real storage space often much greater than the information would originally occupy. Because of that, there is error correction codes, or erasure codes, which has been used with a mathematical approach considerably more refined than the simple replication, generating a smaller storage overhead than their predecessors techniques. The contribution of this work is a fully decentralized storage strategy that, on average, presents performance improvements of over 80%in access latency for both replicated and encoded data, while minimizing by 55% the overhead for a terabyte-sized dataset when encoded and compared to related works of the literature.
dc.languageeng
dc.relationJournal of Computer Science
dc.relation0,147
dc.rightsAcesso aberto
dc.sourceScopus
dc.subjectBig data
dc.subjectCache
dc.subjectData storage
dc.subjectErasure coding
dc.subjectObject storage
dc.titleA fast access big data approach for configurable and scalable object storage Enabling mixed fault-tolerance
dc.typeArtículos de revistas


Este ítem pertenece a la siguiente institución