1 / 1

Copysets : Reducing the Frequency of Data Loss in Cloud Storage

Copysets : Reducing the Frequency of Data Loss in Cloud Storage. Asaf Cidon, Stephen Rumble, Ryan Stutsman Sachin Katti , John Ousterhout and Mendel Rosenblum Stanford University. Each Power Outage Causes Data Loss. Cloud storage systems use random replication

ira
Download Presentation

Copysets : Reducing the Frequency of Data Loss in Cloud Storage

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Copysets: Reducing the Frequency of Data Loss in Cloud Storage Asaf Cidon, Stephen Rumble, Ryan Stutsman SachinKatti, John Ousterhout and Mendel Rosenblum Stanford University Each Power Outage Causes Data Loss • Cloud storage systems use random replication • Random replication is vulnerable to power outages • ~1% of nodes fail to reboot after power outage • Each data loss event has a fixed cost: • Better to lose data infrequently at the expense of losing more data in each event Minimize Copysets Minimize Data Loss Events • Copyset: unique set of nodes that contain all replicas of a chunk of data • System loses data when nodes of at least 1 copyset fail simultaneously • Random replication creates too many copysets • Minimum Copysets: statically split nodes into copysets. Each node belongs to single copyset • Place first replica on random node • Place other replicas deterministically on first node’s copyset • On 5000 node cluster, data loss event every 625 years, each event loses data of entire node Copyset Replication • Problem: most systems need to scatter data across a number of nodes (scatter width) • Otherwise, we increase recovery time and impact load balancing • Copyset Replication: Given a scatter width, minimize the number of copysets: • Copyset Replication significantly reduces data loss • While preserving system’s scatter width and node recovery time • Implemented and evaluated on HDFS and RAMCloud • Minimal overhead on normal operations and recovery

More Related