Cloud storage reliability for Big Data applications : a state of the art survey

Research output: Contribution to journalArticlepeer-review

Abstract

Cloud storage systems are now mature enough to handle a massive volume of heterogeneous and rapidly changing data, which is known as Big Data. However, failures are inevitable in cloud storage systems as they are composed of large scale hardware components. Improving fault tolerance in cloud storage systems for Big Data applications is a significant challenge. Replication and Erasure coding are the most important data reliability techniques employed in cloud storage systems. Both techniques have their own trade-off in various parameters such as durability, availability, storage overhead, network bandwidth and traffic, energy consumption and recovery performance. This survey explores the challenges involved in employing both techniques in cloud storage systems for Big Data applications with respect to the aforementioned parameters. In this paper, we also introduce a conceptual hybrid technique to further improve reliability, latency, bandwidth usage, and storage efficiency of Big Data applications on cloud computing.
Original languageEnglish
Pages (from-to)35-47
Number of pages13
JournalJournal of Network and Computer Applications
Volume97
DOIs
Publication statusPublished - 2017

Keywords

  • big data
  • cloud computing
  • fault-tolerant computing
  • replication (experimental design)

Fingerprint

Dive into the research topics of 'Cloud storage reliability for Big Data applications : a state of the art survey'. Together they form a unique fingerprint.

Cite this