1 / 34

Windows Azure Storage – A Highly Available Cloud Storage Service with Strong Consistency

Windows Azure Storage – A Highly Available Cloud Storage Service with Strong Consistency.

yolandas
Download Presentation

Windows Azure Storage – A Highly Available Cloud Storage Service with Strong Consistency

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Windows Azure Storage – A Highly Available Cloud Storage Service with Strong Consistency Brad Calder, Ju Wang, Aaron Ogus, Niranjan Nilakantan, Arild Skjolsvold, Sam McKelvie, Yikang Xu, Shashwat Srivastav, Jiesheng Wu, Huseyin Simitci, Jaidev Haridas, Chakravarthy Uddaraju, Hemal Khatri, Andrew Edwards, Vaman Bedekar, Shane Mainali, Rafay Abbasi, Arpit Agarwal, Mian Fahim ul Haq, Muhammad Ikram ul Haq, Deepali Bhardwaj, Sowmya Dayanand, Anitha Adusumilli, Marvin McNett, Sriram Sankaran, Kavitha Manivannan, Leonidas Rigas Microsoft Corporation

  2. Windows Azure Storage – Agenda • What it is and Data Abstractions • Architecture and How it Works • Storage Stamp • Partition Layer • Stream Layer • Design Choices and Lessons Learned

  3. Windows Azure Storage Geographically Distributed across 3 Regions Anywhere at Anytime Access to data >200 Petabytes of raw storage by December 2011

  4. Windows Azure Storage Data Abstractions • Blobs – File system in the cloud • Tables– Massively scalable structured storage • Queues – Reliable storage and delivery of messages • Drives – Durable NTFS volumes for Windows Azure applications

  5. Windows Azure Storage High Level Architecture

  6. Design Goals • Highly Available with Strong Consistency • Provide access to data in face of failures/partitioning • Durability • Replicate data several times within and across data centers • Scalability • Need to scale to exabytes and beyond • Provide a global namespace to access data around the world • Automatically load balance data to meet peak traffic demands

  7. Windows Azure Storage Stamps Access blob storage via the URL: http://<account>.blob.core.windows.net/ Storage Location Service Data access LB LB Storage Stamp Storage Stamp Front-Ends Front-Ends Partition Layer Partition Layer Inter-stamp (Geo) replication Stream Layer Stream Layer Intra-stamp replication Intra-stamp replication

  8. Storage Stamp Architecture – Stream Layer • Append-only distributed file system • All data from the Partition Layer is stored into files (extents) in the Stream layer • An extent is replicated 3 times across different fault and upgrade domains • With random selection for where to place replicas for fast MTTR • Checksum all stored data • Verified on every client read • Scrubbed every few days • Re-replicate on disk/node/rack failure or checksum mismatch M Stream Layer (Distributed File System) M Paxos M Extent Nodes (EN)

  9. Storage Stamp Architecture – Partition Layer • Provide transaction semantics and strong consistency for Blobs, Tables and Queues • Stores and reads the objects to/from extents in the Stream layer • Provides inter-stamp (geo) replication by shipping logs to other stamps • Scalable object index via partitioning Partition Master Lock Service Partition Layer Partition Server Partition Server Partition Server Partition Server M M Paxos Stream Layer M Extent Nodes (EN)

  10. Storage Stamp Architecture Front End Layer FE FE FE FE FE • Stateless Servers • Authentication + authorization • Request routing Partition Master Lock Service Partition Layer Partition Server Partition Server Partition Server Partition Server M M Paxos Stream Layer M Extent Nodes (EN)

  11. Storage Stamp Architecture Incoming Write Request Ack Front End Layer FE FE FE FE FE Partition Master Lock Service Partition Layer Partition Server Partition Server Partition Server Partition Server M M Paxos Stream Layer M Extent Nodes (EN)

  12. Partition Layer

  13. Partition Layer – Scalable Object Index • 100s of Billions of blobs, entities, messages across all accounts can be stored in a single stamp • Need to efficiently enumerate, query, get, and update them • Traffic pattern can be highly dynamic • Hot objects, peak load, traffic bursts, etc • Need a scalable index for the objects that can • Spread the index across 100s of servers • Dynamically load balance • Dynamically change what servers are serving each part of the index based on load

  14. Scalable Object Index via Partitioning • Partition Layer maintains an internal Object Index Table for each data abstraction • Blob Index: contains all blob objects for all accounts in a stamp • Table Entity Index: contains all entities for all accounts in a stamp • Queue Message Index: contains all messages for all accounts in a stamp • Scalability is provided for each Object Index • Monitor load to each part of the index to determine hot spots • Index is dynamically split into thousands of Index RangePartitions based on load • Index RangePartitions are automatically load balanced across servers to quickly adapt to changes in load

  15. Partition Layer – Index Range Partitioning • Split index into RangePartitions based on load • Split at PartitionKey boundaries • PartitionMap tracks Index RangePartition assignment to partition servers • Front-End caches the PartitionMap to route user requests • Each part of the index is assigned to only one Partition Server at a time Blob Index Storage Stamp Partition Master A-H: PS1 H’-R: PS2 R’-Z: PS3 A-H: PS1 H’-R: PS2 R’-Z: PS3 Partition Map Partition Server A-H Front-End Server PS 1 Partition Server Partition Server R’-Z H’-R PS 3 PS 2 Partition Map

  16. Each RangePartition – Log Structured Merge-Tree Writes Read/Query Memory Data Memory Table Row Cache Load Metrics Bloom Filters Index Cache Persistent Data (Stream Layer) Row Data Stream Checkpoint File Table Checkpoint File Table Checkpoint File Table Commit Log Stream Blob Data Stream Metadata log Stream Blob Data Blob Data Blob Data

  17. Stream Layer

  18. Stream Layer • Append-Only Distributed File System • Streams are very large files • Has file system like directory namespace • Stream Operations • Open, Close, Delete Streams • Rename Streams • Concatenate Streams together • Append for writing • Random reads

  19. Stream Layer Concepts Block • Min unit of write/read • Checksum • Up to N bytes (e.g. 4MB) Stream • Hierarchical namespace • Ordered list of pointers to extents • Append/Concatenate Extent • Unit of replication • Sequence of blocks • Size limit (e.g. 1GB) • Sealed/unsealed Stream //foo/myfile.data Ptr E1 Ptr E2 Ptr E3 Ptr E4 Block Block Block Block Block Block Block Block Block Block Block Block Block Block Block Extent E1 Extent E4 Extent E3 Extent E2 unsealed sealed unsealed unsealed sealed sealed

  20. Creating an Extent Paxos Create Stream/Extent SM Partition Layer SM Stream Master EN1 PrimaryEN2, EN3 Secondary Allocate Extent replica set EN 1 EN 2 EN 3 EN Primary Secondary A Secondary B

  21. Replication Flow Paxos Partition Layer SM EN1 PrimaryEN2, EN3 Secondary SM SM Append Ack EN 1 EN 2 EN 3 EN Primary Secondary A Secondary B

  22. Providing Bit-wise Identical Replicas • Want all replicas for an extent to be bit-wise the same, up to a committed length • Want to store pointers from the partition layer index to an extent+offset • Want to be able to read from any replica • Replication flow • All appends to an extent go to the Primary • Primary orders all incoming appends and picks the offset for the append in the extent • Primary then forwards offset and data to secondaries • Primary performs in-order acks back to clients for extent appends • Primary returns the offset of the append in the extent • An extent offset can commit back to the client once all replicas have written that offset and all prior offsets have also already been completely written • This represents the committed length of the extent

  23. Dealing with Write Failures Failure during append • Ack from primary lost when going back to partition layer • Retry from partition layer can cause multiple blocks to be appended (duplicate records) • Unresponsive/Unreachable Extent Node (EN) • Append will not be acked back to partition layer • Seal the failed extent • Allocate a new extent and append immediately Stream //foo/myfile.dat Ptr E1 Ptr E2 Ptr E3 Ptr E4 Ptr E5 ? Extent E5 Extent E1 Extent E2 Extent E3 Extent E4

  24. Extent Sealing (Scenario 1) Paxos Seal Extent Seal Extent Sealed at 120 Partition Layer SM SM Stream Master Append 120 Ask for current length 120 EN 1 EN 2 EN 3 EN 4 Primary Secondary A Secondary B

  25. Extent Sealing (Scenario 1) Paxos Seal Extent Sealed at 120 Partition Layer SM SM Stream Master Sync with SM 120 EN 1 EN 2 EN 4 EN 3 Primary Secondary A Secondary B

  26. Extent Sealing (Scenario 2) Paxos Seal Extent Seal Extent Sealed at 100 Partition Layer SM SM SM Append Ask for current length 120 100 EN 1 EN 2 EN 3 EN 4 Primary Secondary A Secondary B

  27. Extent Sealing (Scenario 2) Paxos Seal Extent Sealed at 100 Partition Layer SM SM SM 100 Sync with SM EN 4 EN 1 EN 2 EN 3 Primary Secondary A Secondary B

  28. Providing Consistency for Data Streams Partition Server • For Data Streams, Partition Layer only reads from offsets returned from successful appends • Committed on all replicas • Row and Blob Data Streams • Offset valid on any replica SM SM SM Safe to read from EN3 EN 1 EN 2 EN 3 • Network partition • PS can talk to EN3 • SM cannot talk to EN3 Primary Secondary A Secondary B

  29. Providing Consistency for Log Streams Check commit length Partition Server • Logs are used on partition load • Commit and Metadata log streams • Check commit length first • Only read from • Unsealed replica if all replicas have the same commit length • A sealed replica SM SM Use EN1, EN2 for loading SM Seal Extent Check commit length EN 1 EN 2 EN 3 • Network partition • PS can talk to EN3 • SM cannot talk to EN3 Primary Secondary A Secondary B

  30. Our Approach to the CAP Theorem • Layering and co-design provides extra flexibility to achieve “C” and “A” at same time while being partition/failure tolerant for our fault model • Stream Layer • Availability with Partition/failure tolerance • For Consistency, replicas are bit-wise identical up to the commit length • Partition Layer • Consistency with Partition/failure tolerance • For Availability, RangePartitions can be served by any partition server and are moved to available servers if a partition server fails • Designed for specific classes of partitioning/failures seen in practice • Process to Disk to Node to Rack failures/unresponsiveness • Node to Rack level network partitioning

  31. Design Choices and Lessons Learned

  32. Design Choices • Multi-Data Architecture • Use extra resources to serve mixed workload for incremental costs • Blob -> storage capacity • Table -> IOps • Queue -> memory • Drives -> storage capacity and IOps • Multiple data abstractions from a single stack • Improvements at lower layers help all data abstractions • Simplifies hardware management • Tradeoff: single stack is not optimized for specific workload pattern • Append-only System • Greatly simplifies replication protocol and failure handling • Consistent and identical replicas up to the extent’s commit length • Keep snapshots at no extra cost • Benefit for diagnosis and repair • Erasure Coding • Tradeoff: GC overhead • Scaling Compute Separate from Storage • Allows each to be scaled separately • Important for multitenant environment • Moving toward full bisection bandwidth between compute and storage • Tradeoff: Latency/BW to/from storage

  33. Lessons Learned • Automatic load balancing • Quickly adapt to various traffic conditions • Need to handle every type of workload thrown at the system • Built an easily tunable and extensible language to dynamically tune the load balancing rules • Need to tune based on many dimensions • CPU, Network, Memory, tps, GC load, Geo-Rep load, Size of partitions, etc • Achieving consistently low append latencies • Ended up using journaling • Efficient upgrade support • Pressure point testing

  34. Windows Azure Storage Summary • Highly Available Cloud Storage with Strong Consistency • Scalable data abstractions to build your applications • Blobs – Files and large objects • Tables – Massively scalable structured storage • Queues – Reliable delivery of messages • Drives – Durable NTFS volume for Windows Azure applications • More information • Windows Azure tutorial this Wednesday 26th, 17:00 at start of SOCC • http://blogs.msdn.com/windowsazurestorage/

More Related