SlideShare ist ein Scribd-Unternehmen logo
1 von 32
Downloaden Sie, um offline zu lesen
+




MyCassandra: A Cloud Storage Supporting
both Read Heavy and Write Heavy Workloads	
 
       Shunsuke Nakamura (Tokyo Institute of Technology, NHN Japan)
                       Kazuyuki Shudo (Tokyo Inistitute of Technology)	
 
                                                    Session 6 - Storage, SYSTOR 2012
                                                                 (Haifa, Israel, Jun 4-6)
+
    Cloud Storage	
 
Distributed data store processing large amount of data

        NoSQL, Key-Value Storage (KVS), Document-Oriented DB, GraphDB
                  Example: memcached, Google Bigtable, Amazon Dynamo, Amazon SimpleDB, Apache
                   Cassandra, Voldemort, Ringo, Vpork, MongoDB, CouchDB, Tokyo Tyrant, Flare, ROMA, kumofs,
                   Kai, Redis, LevelDB, Hadoop HBase, Hypertable,Yahoo! PNUTS, Scalaris, Dynomite, ThruDB,
                   Neo4j, IBM ObjectGrid, Giraph, Oracle Coherence, and the others. (> 100 products)


        Characteristics: “limited functions, massive volume, high performance”
             Data access only by primary key
             No luxury features such as join, global transaction
             Scalable to much larger data and number of nodes
+
    Design policies of cloud storages	
 
    There are many trade-offs. 	
 

        data model
             key/value, multi-dimensional map, document or graph

        performance - write vs. read

        latency vs. persistence
             latency – memory and disk utilization
             persistence – synchronous vs. asynchronous (snapshot)

        replication – synchronous vs. asynchronous

        consistency between replicas – strong vs. weak
        data partitioning – row vs. column

        distribution – master/slave vs. decentralized

Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
+
    MyCassandra focuses on
    performance trade-off	
 
        data model
             key/value vs. multi-dimensional map vs. document vs. graph

        performance - write vs. read

        latency vs. persistence
             latency – memory and disk utilization
             persistence – synchronous vs. asynchronous (snapshot)

        replication – synchronous vs. asynchronous

        consistency between replicas – strong vs. weak
        data partitioning – row vs. column

        distribution – master/slave vs. decentralized

Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
+
    Performance trade-off	
 
    Write-optimized vs. read-optimized	
 

        A cloud storage with persistence is designed to optimize either
         write or read workload.

        Storage engine determines which workload a cloud storage
         treats efficiently.

                                              Bigtable, Cassandra,            MySQL, Yahoo!
                                              HBase	
                         Sherpa
Indexing	
                                    Log-Structured                  B-Trees [R.Bayer ’70]
                                              Merge Tree [P. O’Neil ‘96]	
 
Write to disk	
                               append                          random reads, writes	
 
Read to disk	
                                random reads + merge            random read	
 
Performance	
                                 write-optimized                 read-optimized	
 
Storage engine                                Bigtable clone                  MySQL	
 
Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
+ Performance trade-off	
 
      - write-optimized vs. read-optimized -	
 
                                  Write latency for write-heavy workload	
 




                                                                                    Better	
 
                                      read-optimized	
 



                                                               write-optimized	
 
                                                                              6




                               Yahoo! Cloud Serving Benchmark, SOCC ’10	
 
                                                - mycassandra -
Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
+ Performance trade-off 	
 
      - write-optimized vs. read-optimized -	
 
                                    Read latency for read-heavy workload	
 



                                                      write-optimized	
 



                                                                                  Better	
 




                                                                     read-optimized	
 




                               Yahoo! Cloud Serving Benchmark, SOCC ’10	
 
                                                - mycassandra -
Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
+
     Research overview	
 
    Contribution:
     A technique to build a cloud storage performing well with both read and
     write workloads

    Steps:
     1.    MyCassandra: Storage engine enabled Apache Cassandra
     2.    MyCassandra Cluster: Heterogeneous cluster with different storage engines

                 1. MyCassandra	
                              2. MyCassandra Cluster	
 


       read-optimized	
 


                                                                read and write-optimized	
 
              select	
             write-optimized	
 



Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
+
    Apache Cassandra	
 
    Open-sourced by                                            in 2008              	
 
    A top-level project in	
 

        Features:
             Scalability up to hundreds of servers across multiple racks/datacenters
             High availability without SPOF by adopting a decentralized architecture
             Write-optimized	
 
                                                                         dc1	
                dc2	
 



                                                                Clustering across multiple racks/DCs
                                                                Replication strategy based on region	
 
                                                                                dc3	
 



Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
+
     Apache Cassandra	
 
     A decentralized cloud storage without SPOF	
 

         Consistent Hashing (a decentralized algorithm):
          Assign identifiers to both nodes and data on its circular ID space.

A-Z: hash value	
 
Num of replica := 3	
                                 ID space	
 

                 A	
                 F	
 
          Z	
                                                          Roles of each node
                                 secondary 1	
 
                                                                       •  Proxy, serving clients
                                    Q	
                                •  Primary/secondary data nodes
             V
             	
                                           N	
 
          primary	
                                   secondary 2	
 
                                        hash(key) = Q	
 
                                   key	
  values	
 

 Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
+
      Apache Cassandra	
 
      Write-optimized storage engine, a Bigtable clone	
 
    O(1) fast write operation
         Write an update to disk sequentially
          - Fast because of no random I/O to disk
          - Always writable because of no write-lock
                                                                                        memory	
 
                                                           sync	
              <k1, obj (v1+v2)>	
 async flush	
 
              write path	
                                                  Memtable	
 
1.  Append an update to
    CommitLog for persistence Only                             sequential write	
          disk	
 
2.  Update Memtable, a map in
                                                                              <k1, v1>, <k1, v2>
    memory, for quick reading
3.  Acknowledge a client                                                   CommitLog	
 
4.  Asynchronously flush Memtable                                                                  <k1,obj1>	
 
    to SSTable                                                               SSTable 1	
 
5.  Delete flushed data from                                                                       <k1,obj2>	
 
    CommitLog and Memtable	
                                                 SSTable 2	
 
                                                                                                   <k1,obj3>	
 
Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)	
 
                                                                             SSTable 3
+
      Apache Cassandra	
 
      Write-optimized storage engine, a Bigtable clone	
 
    Slow read operation
         Read data from Memtable and multiple SSTables, and merge them
          - Slow because of multiple random I/Os on disk

                                                                                                 memory	
 
                                                                          <k1,obj>	
 
                                                                                         Memtable	
 
                                                                                                        disk	
 

                                                                                        CommitLog	
 
                                                               merge	
 
                                                                          <k1,obj1>	
 
                                                                                         SSTable 1	
 
                    multiple random I/Os	
                                <k1,obj2>	
 
                                                                                         SSTable 2	
 
                                                                          <k1,obj3>	
 
                                                                                         SSTable 3	
 
Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
+ Performance of original Cassandra	
 
                           Write performance is much higher.
                              YCSB results show:
                                   Average: write is 9 x as fast as read.
                                   99.9%ile: write is 43.5 x as fast as read.

                                                        Better	
 
                                                                read
Number of operations	
 




                                                                                              write	
 
                                                         avg.   6.16 ms	
 
                                                                                              read	
 



                                                                                   Latency (ms)	
 
                                                                                   99.9 %ile	
 
                                              write                              write: 2.0 ms
                                          avg. 0.69 ms	
                         read: 86.9 ms	
 

                                                        Latency (ms)
1. Storage Engine Support	
 
+
                           1.MyCassandra	
 


               read-optimized	
 




                       select	
              write-optimized	
 



    Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
+
    MyCassandra: A modular cloud storage	
 
    Storage engines are supported	
 
          Storage engine feature inspired by MySQL
               An independent and pluggable component
               Perform disk I/O

          A cloud storage can be either write-optimized or read-optimized by
           selecting storage engine
          Keep Cassandra’s original distribution architecture and data model

                               Decentralized	
 

                          Consistent Hashing
            Bigtable	
    Gossip Protocol	
 



                                                     Bigtable	
  MySQL	
  Redis	
  …	
 
                                                             selectable	
 
    InnoDB	
  MyISAM	
 Memory	
 …	
 
          selectable	
                            Decentralized + Storage engine
       Storage engine
MyCassandra implementation	
 
                                         Cassandra’s original distribution arch.	
 




                                                   Storage Engine Interface introduced



                                                                                        Implement
                                                                                      Storage Engine
                                                        Storage Engine Interface	
       Interface	
 




Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
Performance of each storage engine	
 
  storage          engines
         Bigtable: write-optimized (original Casssandra 0.7.5)
         MySQL: read-optimized (MySQL 6.0 with InnoDB, JDBC API, stored procedure)
         Redis: in-memory KVS (Redis 2.2.8)	
 


                                                                            6 nodes
                                                                            -  Crucial’s SSD
                                                                            -  allocate 6GB mem in 8GB
                          x 11.79	
 
                                                                            1KB x 36 million data set	
 


                                                                x 9.87	
 




                                              workload	
 

 Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
2. Heterogeneous cluster of
               different storage engines	
 
       +
                              2.MyCassandra Cluster	
 




                                read and write-optimized	
 




           Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
•  W: write-optimized
 Basic idea	
                                                                     •  R: read-optimized
                                                                                  •  RW: in-memory


     Replicate data on different storage engine nodes
                                                                                          write query	
 
     Route a query to nodes processing it efficiently
          Synchronously route to nodes processing quickly                        sync	
              async	
 
          Asynchronously route to nodes processing slowly
           → Exploit each node’s advantage
                                                                                      W                R	
 
     Furthermore, maintain consistency between replicas as much as the
      original Cassandra

 Quorum Protocol:                        (write agreements) +   (read agreements) >     (num of replicas)

      = Guarantee retrieval of the latest data

                                                                                      write	
          read	
 
      Consequence: At least one node processes
      both read and write queries synchronously and quickly
      → In-memory nodes play this role.                                               W       RW	
      R	
 
Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
•  W: write-optimized
Cluster design	
                                                       •  R: read-optimized
                                                                       •  RW: in-memory
  Combine       nodes with different storage engines
       write-optimized (W), read-optimized (R), in-memory (RW)
  Disseminate        storage engine types of each nodes
       The type is attached to gossip messages
  Place   replicas on nodes with different storage engines
       Proxy (any node requested) selects the storing nodes
          1.    The primary node determined based on the queried key
          2.    N -1 secondary nodes with different storage engines
  Multiple     nodes share a single server for load balance
                                                                            Proxy (any node)	
 
        Cluster configuration (N=3)	
 

                  gossip	
 
                                            RW	
 


    W                                                   W           RW	
  RW	
        R	
 
                RW	
          R	
    W	
 
                                                    secondary2	
                   primary	
 
                                                                        secondary1	
 
           responsible nodes
•  W: write-optimized
        Process for a write access	
                                                      •  R: read-optimized
                                                                                          •  RW: in-memory

 •  Quorum parameters                                           Client	
 
      = 3, = = 2                                                         1)  A  proxy receives a write query
 •  Num. of replicas                      Write for a single record	
         from a client. The proxy routes
                                                                              to nodes storing the record.
    W:RW:R = 1:1:1	
                         Proxy	
 
                                 …
                                                          …              2)  The proxy waits ACKs. W, RW
               …
                                                                              nodes usually reply quickly.
                          Wait for two ACKs                        …

                          for write and return                           3-a) If  writing succeeds and the
RW	
                                                                          proxy receives ACKs, it
                                                       Async write
                                                                              returns a success message.
           R
                                                                 W       3-b) If  a data node fails to write,
                     W               RW	
               R	
                   the proxy waits for ACKs
                                                                              including R nodes and returns a
               Nodes storing the record	
                                     success message.

                                                                         4)   After returning, the proxy
         Write latency: max (W, RW)	
 
                                                                               asynchronously waits ACKs
                                                                               from the remaining nodes.
    Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
•  W: write-optimized
        Process for a read access	
                                                     •  R: read-optimized
                                                                                        •  RW: in-memory
 •  Quorum parameters
      = 3, = = 2                                                Client	
 1)  A proxy receives a read query
                                                                           and routes to storing nodes.
 •  Num. of replicas                      Read for a single record	
 
    W:RW:R = 1:1:1	
                                                    2)  Theproxy waits for ACKs. R
                                             Proxy	
                       and RW nodes reply quickly.
               …
                                 …
                                                          …
                                                                        3-a) If returned values are
                                                                    …
                                                                          consistent, the proxy returns it.
                   Async check
                                           Check consistentcy
RW	
               consistency                                          3-b) If the values are mismatched,
                                           and return result	
 
                                                                          the proxy waits for consistent
           R                                                    W         values including W nodes.
                     W               RW	
               R	
 
                                                                        4) After returning, the proxy waits
               Nodes storing the record	
                                 from the remaining nodes. If the
                                                                          proxy notices inconsistent
         Read latency: max (R, RW)	
                                      values, it asynchronously
                                                                          updates them to the consistent
                                                                          one. Cassandra’s feature
    Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)             ReadRepair does it.
+
     Performance Evaluation	
 
     Demonstrate that a heterogeneous cluster performs
     well with both read- and write-heavy workloads	
 
    Targets
          MyCassandra Cluster:                                3 different nodes/server x 6 servers
          Cassandra:                                          1 node/server            x 6 servers

    Quorum parameters
             = 3,       =      =2

    Storage Engine
          Bigtable (W), MySQL / InnoDB (R), Redis (RW)

    Yahoo! Cloud Serving Benchmark (YCSB) [SOCC ’10]
     1.     Load data (1KB record, 10 x 100bytes columns) from a YCSB client
     2.     Warm up
     3.     Run benchmark and measure response times from a client

Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
+
    YCSB workloads	
 

                  Workload	
    Application Operation                           Record
                                example	
   ratio                               selection	
 
                  Write-Only	
  Log	
       Read: 0%                            Zipfian
Write                                                          Write: 100%	
 
heavy	
           Write-Heavy	
  Session store	
  Read: 50%
                                                               Write: 50%
                  Read-Heavy	
  Photo                          Read: 95%
Read                                                           Write: 5%	
 
heavy	
 
                                tagging	
 
                  Read-Only	
  Cache	
                         Read: 100%
                                                               Write: 0%	
 
                     Zipfian distribution: the access frequency of each datum is
                     determined by its popularity, not by freshness.	
 



Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
Write/Read latency (Response time)	
 
     1.5
                             avg. write-latency                            Cassandra
                 + 0.57ms (max)	
                                          MyCassandra Cluster
       1
Better	
        + 42.5%	
            + 59.5%	
            + 69.5%	
           Performs well
     0.5                                                                           with
                                                             write:5%	
       MySQL + Redis	
 
                  write:100%	
             write:50%	
                            write:0%	
 
       0
     (ms)	
          max 90.4% lower in read-only workload	
 
      35
      30
                                                    avg. read-latency
                       - 26.5ms (max)	
 
      25
     20
Better	
                                                         - 88.8%	
              - 90.4%	
 
     15
                                     - 83.3%
      10
       5
                    read:0%	
            read:50%	
          read:95%	
           read:100%	
 
       0
      (ms)	
 
                  Write-Only           Write-Heavy          Read-Heavy            Read-Only
Throughput	
 
         25000                                                                      Cassandra
                                x 0.87	
            QPS for 40 clients              MyCassandra Cluster
         20000

         15000

Better	
 10000
                                                          x 2.16	
                        x 11.00	
 
                                                                        x 4.07	
 
           5000

                 0
                            [100:0]	
                   [50:50]	
        [5:95]	
         [0:100]	
  [write:read]	
 
    (query/sec)	
          Write-Only                 Write-Heavy     Read-Heavy         Read-Only

                                 Write heavy	
                               Read heavy	
 
                             •  11.0 times as high as Cassandra in Read-Only workload
                             •  Write performance is comparable with Cassandra.	
 



  Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
+
    Conclusion	
 

      A  cloud storage supporting both write-heavy and read-
        heavy workloads by combining different storage engine
        nodes.

            MyCassandra Cluster achieved better throughput than the
             original Cassandra on read heavy workload.


            With a read-heavy workload
               Read latency: 90.4 % lower at most

                 Throughput: 11.0 times at most



Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
+
    Related Work	
 
        Indexing algorithm whose goals include achieving both write and
         read performance
             FD-Tree: Tree Indexing on Flash Disks, VLDB ’10
             bLSM: A General Purpose Log Structured Merge Tree, SIGMOD ‘12
             Fractal-Tree: It’s implemented in TokuDB (MySQL storage engine)


        Modular data stores:
             MySQL
             Anvil, SOSP ’09
             Cloudy, VLDB ’10
             Dynamo, SOSP ’07
             Fractured Mirrors:
           MyCassandra,                       SYSTOR ‘12: read vs. write	
 


Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
Discussion 1. the slight higher
 +
   write latency	
 
      The cause is load balancing.	
 
              Cassandra
                   Write to any             nodes in N nodes
              MyCassandra Cluster
                   Write to the specified                       and   nodes
          However this cost well worths improving for read performance.

                                                                               MyCassandra
                                       Cassandra	
                               Cluster	
 


Sync operation is write	
                                 read	
          write	
         read	
  Sync operation is
equally distributed.	
                                                                            fixed.

                                                                          W       RW	
    R	
 

  Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
+
    Discussion 2. in-memory node	
 
    Q. Memory overflow

    A. In-memory node plays as LRU-like cache.
      The swapped data is recovered from the other persistent
      nodes by read repair.

    Q. Fault tolerance

    A.       1) Write to an alternative node, and if the node is recovered, it
             resolves inconsistency using values from the node.

             2) Asynchronous snapshot (Redis’s feature)


     Q. Whole in-memory nodes

     A. This case limits capacities in cluster with the memory’s capacity.

Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
+
    オープンソース化	
 




Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
+




Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)

Weitere ähnliche Inhalte

Was ist angesagt?

No Sql Introduction
No Sql IntroductionNo Sql Introduction
No Sql IntroductionDingding Ye
 
Apache HBase for Architects
Apache HBase for ArchitectsApache HBase for Architects
Apache HBase for ArchitectsNick Dimiduk
 
Nyc summit intro_to_cassandra
Nyc summit intro_to_cassandraNyc summit intro_to_cassandra
Nyc summit intro_to_cassandrazznate
 
Accumulo Summit 2014: Benchmarking Accumulo: How Fast Is Fast?
Accumulo Summit 2014: Benchmarking Accumulo: How Fast Is Fast?Accumulo Summit 2014: Benchmarking Accumulo: How Fast Is Fast?
Accumulo Summit 2014: Benchmarking Accumulo: How Fast Is Fast?Accumulo Summit
 
Introduction to hadoop and hdfs
Introduction to hadoop and hdfsIntroduction to hadoop and hdfs
Introduction to hadoop and hdfsTrendProgContest13
 
HBaseCon 2012 | HBase and HDFS: Past, Present, Future - Todd Lipcon, Cloudera
HBaseCon 2012 | HBase and HDFS: Past, Present, Future - Todd Lipcon, ClouderaHBaseCon 2012 | HBase and HDFS: Past, Present, Future - Todd Lipcon, Cloudera
HBaseCon 2012 | HBase and HDFS: Past, Present, Future - Todd Lipcon, ClouderaCloudera, Inc.
 
Design for a Distributed Name Node
Design for a Distributed Name NodeDesign for a Distributed Name Node
Design for a Distributed Name NodeAaron Cordova
 
Hug Hbase Presentation.
Hug Hbase Presentation.Hug Hbase Presentation.
Hug Hbase Presentation.Jack Levin
 
Millions of Regions in HBase: Size Matters
Millions of Regions in HBase: Size MattersMillions of Regions in HBase: Size Matters
Millions of Regions in HBase: Size MattersDataWorks Summit
 
Architecting the Future of Big Data & Search - Eric Baldeschwieler
Architecting the Future of Big Data & Search - Eric BaldeschwielerArchitecting the Future of Big Data & Search - Eric Baldeschwieler
Architecting the Future of Big Data & Search - Eric Baldeschwielerlucenerevolution
 
Hadoop Summit 2012 | HBase Consistency and Performance Improvements
Hadoop Summit 2012 | HBase Consistency and Performance ImprovementsHadoop Summit 2012 | HBase Consistency and Performance Improvements
Hadoop Summit 2012 | HBase Consistency and Performance ImprovementsCloudera, Inc.
 
Scaling HDFS to Manage Billions of Files
Scaling HDFS to Manage Billions of FilesScaling HDFS to Manage Billions of Files
Scaling HDFS to Manage Billions of FilesHaohui Mai
 
HBaseCon 2015: Elastic HBase on Mesos
HBaseCon 2015: Elastic HBase on MesosHBaseCon 2015: Elastic HBase on Mesos
HBaseCon 2015: Elastic HBase on MesosHBaseCon
 
Demystifying bigdata ashish1
Demystifying bigdata ashish1Demystifying bigdata ashish1
Demystifying bigdata ashish1Ashish singh
 
Automation of Hadoop cluster operations in Arm Treasure Data
Automation of Hadoop cluster operations in Arm Treasure DataAutomation of Hadoop cluster operations in Arm Treasure Data
Automation of Hadoop cluster operations in Arm Treasure DataYan Wang
 
SQLIO - measuring storage performance
SQLIO - measuring storage performanceSQLIO - measuring storage performance
SQLIO - measuring storage performancevalerian_ceaus
 
HBaseCon 2012 | HBase, the Use Case in eBay Cassini
HBaseCon 2012 | HBase, the Use Case in eBay Cassini HBaseCon 2012 | HBase, the Use Case in eBay Cassini
HBaseCon 2012 | HBase, the Use Case in eBay Cassini Cloudera, Inc.
 
Reducing Database Pain & Costs with Postgres
Reducing Database Pain & Costs with PostgresReducing Database Pain & Costs with Postgres
Reducing Database Pain & Costs with PostgresEDB
 
Ceph on Intel: Intel Storage Components, Benchmarks, and Contributions
Ceph on Intel: Intel Storage Components, Benchmarks, and ContributionsCeph on Intel: Intel Storage Components, Benchmarks, and Contributions
Ceph on Intel: Intel Storage Components, Benchmarks, and ContributionsColleen Corrice
 

Was ist angesagt? (20)

No Sql Introduction
No Sql IntroductionNo Sql Introduction
No Sql Introduction
 
Apache HBase for Architects
Apache HBase for ArchitectsApache HBase for Architects
Apache HBase for Architects
 
Nyc summit intro_to_cassandra
Nyc summit intro_to_cassandraNyc summit intro_to_cassandra
Nyc summit intro_to_cassandra
 
Accumulo Summit 2014: Benchmarking Accumulo: How Fast Is Fast?
Accumulo Summit 2014: Benchmarking Accumulo: How Fast Is Fast?Accumulo Summit 2014: Benchmarking Accumulo: How Fast Is Fast?
Accumulo Summit 2014: Benchmarking Accumulo: How Fast Is Fast?
 
Introduction to hadoop and hdfs
Introduction to hadoop and hdfsIntroduction to hadoop and hdfs
Introduction to hadoop and hdfs
 
HBaseCon 2012 | HBase and HDFS: Past, Present, Future - Todd Lipcon, Cloudera
HBaseCon 2012 | HBase and HDFS: Past, Present, Future - Todd Lipcon, ClouderaHBaseCon 2012 | HBase and HDFS: Past, Present, Future - Todd Lipcon, Cloudera
HBaseCon 2012 | HBase and HDFS: Past, Present, Future - Todd Lipcon, Cloudera
 
Design for a Distributed Name Node
Design for a Distributed Name NodeDesign for a Distributed Name Node
Design for a Distributed Name Node
 
Hug Hbase Presentation.
Hug Hbase Presentation.Hug Hbase Presentation.
Hug Hbase Presentation.
 
Millions of Regions in HBase: Size Matters
Millions of Regions in HBase: Size MattersMillions of Regions in HBase: Size Matters
Millions of Regions in HBase: Size Matters
 
HBase Storage Internals
HBase Storage InternalsHBase Storage Internals
HBase Storage Internals
 
Architecting the Future of Big Data & Search - Eric Baldeschwieler
Architecting the Future of Big Data & Search - Eric BaldeschwielerArchitecting the Future of Big Data & Search - Eric Baldeschwieler
Architecting the Future of Big Data & Search - Eric Baldeschwieler
 
Hadoop Summit 2012 | HBase Consistency and Performance Improvements
Hadoop Summit 2012 | HBase Consistency and Performance ImprovementsHadoop Summit 2012 | HBase Consistency and Performance Improvements
Hadoop Summit 2012 | HBase Consistency and Performance Improvements
 
Scaling HDFS to Manage Billions of Files
Scaling HDFS to Manage Billions of FilesScaling HDFS to Manage Billions of Files
Scaling HDFS to Manage Billions of Files
 
HBaseCon 2015: Elastic HBase on Mesos
HBaseCon 2015: Elastic HBase on MesosHBaseCon 2015: Elastic HBase on Mesos
HBaseCon 2015: Elastic HBase on Mesos
 
Demystifying bigdata ashish1
Demystifying bigdata ashish1Demystifying bigdata ashish1
Demystifying bigdata ashish1
 
Automation of Hadoop cluster operations in Arm Treasure Data
Automation of Hadoop cluster operations in Arm Treasure DataAutomation of Hadoop cluster operations in Arm Treasure Data
Automation of Hadoop cluster operations in Arm Treasure Data
 
SQLIO - measuring storage performance
SQLIO - measuring storage performanceSQLIO - measuring storage performance
SQLIO - measuring storage performance
 
HBaseCon 2012 | HBase, the Use Case in eBay Cassini
HBaseCon 2012 | HBase, the Use Case in eBay Cassini HBaseCon 2012 | HBase, the Use Case in eBay Cassini
HBaseCon 2012 | HBase, the Use Case in eBay Cassini
 
Reducing Database Pain & Costs with Postgres
Reducing Database Pain & Costs with PostgresReducing Database Pain & Costs with Postgres
Reducing Database Pain & Costs with Postgres
 
Ceph on Intel: Intel Storage Components, Benchmarks, and Contributions
Ceph on Intel: Intel Storage Components, Benchmarks, and ContributionsCeph on Intel: Intel Storage Components, Benchmarks, and Contributions
Ceph on Intel: Intel Storage Components, Benchmarks, and Contributions
 

Andere mochten auch

On Analyzing and Specifying Concerns for Data as a Service
On Analyzing and Specifying Concerns for Data as a ServiceOn Analyzing and Specifying Concerns for Data as a Service
On Analyzing and Specifying Concerns for Data as a ServiceHong-Linh Truong
 
TECHNICAL CASE STUDY: STORK
TECHNICAL CASE STUDY: STORKTECHNICAL CASE STUDY: STORK
TECHNICAL CASE STUDY: STORKForgeRock
 
How to Pick a Technology Vendor: Top 6 Questions to Ask
How to Pick a Technology Vendor: Top 6 Questions to AskHow to Pick a Technology Vendor: Top 6 Questions to Ask
How to Pick a Technology Vendor: Top 6 Questions to AskTriCorps Technologies
 
Networking Library Services: A Glimpse at the Future--Moving Library Manageme...
Networking Library Services: A Glimpse at the Future--Moving Library Manageme...Networking Library Services: A Glimpse at the Future--Moving Library Manageme...
Networking Library Services: A Glimpse at the Future--Moving Library Manageme...OCLC Research
 
Using the CAP theorem as a way to pick Cloud Service providers
Using the CAP theorem as a way to pick Cloud Service providersUsing the CAP theorem as a way to pick Cloud Service providers
Using the CAP theorem as a way to pick Cloud Service providersN. S. Amarnath
 
Open Audit
Open AuditOpen Audit
Open Auditncspa
 
A Survey on SSO Authentication protocols: Security and Performance
A Survey on SSO Authentication protocols: Security and PerformanceA Survey on SSO Authentication protocols: Security and Performance
A Survey on SSO Authentication protocols: Security and PerformanceAmin Saqi
 
Agility at Scale: Platform Versus Product Concerns - Agile Pune 2014
Agility at Scale: Platform Versus Product Concerns - Agile Pune 2014Agility at Scale: Platform Versus Product Concerns - Agile Pune 2014
Agility at Scale: Platform Versus Product Concerns - Agile Pune 2014Howard Deiner
 
Concern of Web Application Security
Concern of Web Application SecurityConcern of Web Application Security
Concern of Web Application SecurityMahmud Ahsan
 
IdeaMart: Case Study in Service Innovation Success, Shafraz Rahim, Dialog Axiata
IdeaMart: Case Study in Service Innovation Success, Shafraz Rahim, Dialog AxiataIdeaMart: Case Study in Service Innovation Success, Shafraz Rahim, Dialog Axiata
IdeaMart: Case Study in Service Innovation Success, Shafraz Rahim, Dialog AxiataAlan Quayle
 
VISION-STRATEGY-PRODUCT (VSP) Yacht: An Agile Plan to Rapidly Achieve Problem...
VISION-STRATEGY-PRODUCT (VSP) Yacht: An Agile Plan to Rapidly Achieve Problem...VISION-STRATEGY-PRODUCT (VSP) Yacht: An Agile Plan to Rapidly Achieve Problem...
VISION-STRATEGY-PRODUCT (VSP) Yacht: An Agile Plan to Rapidly Achieve Problem...Rod King, Ph.D.
 
The Art of Scalability - Managing growth
The Art of Scalability - Managing growthThe Art of Scalability - Managing growth
The Art of Scalability - Managing growthLorenzo Alberton
 
Distributed Systems: scalability and high availability
Distributed Systems: scalability and high availabilityDistributed Systems: scalability and high availability
Distributed Systems: scalability and high availabilityRenato Lucindo
 
Successful Scalability Principles - Part 1
Successful Scalability Principles - Part 1Successful Scalability Principles - Part 1
Successful Scalability Principles - Part 1Ronald Bradford
 
OpenStack NFV Edge computing for IOT microservices
OpenStack NFV Edge computing for IOT microservicesOpenStack NFV Edge computing for IOT microservices
OpenStack NFV Edge computing for IOT microservicesopenstackindia
 

Andere mochten auch (20)

Get metherightjob r_paas 2012
Get metherightjob r_paas 2012Get metherightjob r_paas 2012
Get metherightjob r_paas 2012
 
On Analyzing and Specifying Concerns for Data as a Service
On Analyzing and Specifying Concerns for Data as a ServiceOn Analyzing and Specifying Concerns for Data as a Service
On Analyzing and Specifying Concerns for Data as a Service
 
TECHNICAL CASE STUDY: STORK
TECHNICAL CASE STUDY: STORKTECHNICAL CASE STUDY: STORK
TECHNICAL CASE STUDY: STORK
 
How to Pick a Technology Vendor: Top 6 Questions to Ask
How to Pick a Technology Vendor: Top 6 Questions to AskHow to Pick a Technology Vendor: Top 6 Questions to Ask
How to Pick a Technology Vendor: Top 6 Questions to Ask
 
Networking Library Services: A Glimpse at the Future--Moving Library Manageme...
Networking Library Services: A Glimpse at the Future--Moving Library Manageme...Networking Library Services: A Glimpse at the Future--Moving Library Manageme...
Networking Library Services: A Glimpse at the Future--Moving Library Manageme...
 
Using the CAP theorem as a way to pick Cloud Service providers
Using the CAP theorem as a way to pick Cloud Service providersUsing the CAP theorem as a way to pick Cloud Service providers
Using the CAP theorem as a way to pick Cloud Service providers
 
GEOSS AIP History with Authentication and SSO
GEOSS AIP History with Authentication and SSOGEOSS AIP History with Authentication and SSO
GEOSS AIP History with Authentication and SSO
 
Open Audit
Open AuditOpen Audit
Open Audit
 
A Survey on SSO Authentication protocols: Security and Performance
A Survey on SSO Authentication protocols: Security and PerformanceA Survey on SSO Authentication protocols: Security and Performance
A Survey on SSO Authentication protocols: Security and Performance
 
Agility at Scale: Platform Versus Product Concerns - Agile Pune 2014
Agility at Scale: Platform Versus Product Concerns - Agile Pune 2014Agility at Scale: Platform Versus Product Concerns - Agile Pune 2014
Agility at Scale: Platform Versus Product Concerns - Agile Pune 2014
 
Concern of Web Application Security
Concern of Web Application SecurityConcern of Web Application Security
Concern of Web Application Security
 
IdeaMart: Case Study in Service Innovation Success, Shafraz Rahim, Dialog Axiata
IdeaMart: Case Study in Service Innovation Success, Shafraz Rahim, Dialog AxiataIdeaMart: Case Study in Service Innovation Success, Shafraz Rahim, Dialog Axiata
IdeaMart: Case Study in Service Innovation Success, Shafraz Rahim, Dialog Axiata
 
Azure Web Scalability
Azure Web ScalabilityAzure Web Scalability
Azure Web Scalability
 
VISION-STRATEGY-PRODUCT (VSP) Yacht: An Agile Plan to Rapidly Achieve Problem...
VISION-STRATEGY-PRODUCT (VSP) Yacht: An Agile Plan to Rapidly Achieve Problem...VISION-STRATEGY-PRODUCT (VSP) Yacht: An Agile Plan to Rapidly Achieve Problem...
VISION-STRATEGY-PRODUCT (VSP) Yacht: An Agile Plan to Rapidly Achieve Problem...
 
Api Design
Api DesignApi Design
Api Design
 
The Art of Scalability - Managing growth
The Art of Scalability - Managing growthThe Art of Scalability - Managing growth
The Art of Scalability - Managing growth
 
Distributed Systems: scalability and high availability
Distributed Systems: scalability and high availabilityDistributed Systems: scalability and high availability
Distributed Systems: scalability and high availability
 
Successful Scalability Principles - Part 1
Successful Scalability Principles - Part 1Successful Scalability Principles - Part 1
Successful Scalability Principles - Part 1
 
OpenStack NFV Edge computing for IOT microservices
OpenStack NFV Edge computing for IOT microservicesOpenStack NFV Edge computing for IOT microservices
OpenStack NFV Edge computing for IOT microservices
 
REST & RESTful Web Services
REST & RESTful Web ServicesREST & RESTful Web Services
REST & RESTful Web Services
 

Ähnlich wie MyCassandra: A Cloud Storage Supporting both Read Heavy and Write Heavy Workloads (SYSTOR 2012)

Webinar Sept 22: Gluster Partners with Redapt to Deliver Scale-Out NAS Storage
Webinar Sept 22: Gluster Partners with Redapt to Deliver Scale-Out NAS StorageWebinar Sept 22: Gluster Partners with Redapt to Deliver Scale-Out NAS Storage
Webinar Sept 22: Gluster Partners with Redapt to Deliver Scale-Out NAS StorageGlusterFS
 
Intro to GlusterFS Webinar - August 2011
Intro to GlusterFS Webinar - August 2011Intro to GlusterFS Webinar - August 2011
Intro to GlusterFS Webinar - August 2011GlusterFS
 
Strata + Hadoop World 2012: HDFS: Now and Future
Strata + Hadoop World 2012: HDFS: Now and FutureStrata + Hadoop World 2012: HDFS: Now and Future
Strata + Hadoop World 2012: HDFS: Now and FutureCloudera, Inc.
 
Gluster Webinar: Introduction to GlusterFS
Gluster Webinar: Introduction to GlusterFSGluster Webinar: Introduction to GlusterFS
Gluster Webinar: Introduction to GlusterFSGlusterFS
 
Introduction to GlusterFS Webinar - September 2011
Introduction to GlusterFS Webinar - September 2011Introduction to GlusterFS Webinar - September 2011
Introduction to GlusterFS Webinar - September 2011GlusterFS
 
Future of cloud storage
Future of cloud storageFuture of cloud storage
Future of cloud storageGlusterFS
 
Gluster open stack dev summit 042011
Gluster open stack dev summit 042011Gluster open stack dev summit 042011
Gluster open stack dev summit 042011Open Stack
 
Data processing at the speed of 100 Gbps@Apache Crail (Incubating)
Data processing at the speed of 100 Gbps@Apache Crail (Incubating)Data processing at the speed of 100 Gbps@Apache Crail (Incubating)
Data processing at the speed of 100 Gbps@Apache Crail (Incubating)DataWorks Summit
 
PAN 9 Datasheet
PAN 9 DatasheetPAN 9 Datasheet
PAN 9 DatasheetPanasas
 
OpenStack Cinder, Implementation Today and New Trends for Tomorrow
OpenStack Cinder, Implementation Today and New Trends for TomorrowOpenStack Cinder, Implementation Today and New Trends for Tomorrow
OpenStack Cinder, Implementation Today and New Trends for TomorrowEd Balduf
 
인메모리 클러스터링 아키텍처
인메모리 클러스터링 아키텍처인메모리 클러스터링 아키텍처
인메모리 클러스터링 아키텍처Jaehong Cheon
 
An Overview of Apache Cassandra
An Overview of Apache CassandraAn Overview of Apache Cassandra
An Overview of Apache CassandraDataStax
 
Openstack HA
Openstack HAOpenstack HA
Openstack HAYong Luo
 
End of RAID as we know it with Ceph Replication
End of RAID as we know it with Ceph ReplicationEnd of RAID as we know it with Ceph Replication
End of RAID as we know it with Ceph ReplicationCeph Community
 
PhegData X - High Performance EBS
PhegData X - High Performance EBSPhegData X - High Performance EBS
PhegData X - High Performance EBSHanson Dong
 
Under The Hood Of A Shard-Per-Core Database Architecture
Under The Hood Of A Shard-Per-Core Database ArchitectureUnder The Hood Of A Shard-Per-Core Database Architecture
Under The Hood Of A Shard-Per-Core Database ArchitectureScyllaDB
 

Ähnlich wie MyCassandra: A Cloud Storage Supporting both Read Heavy and Write Heavy Workloads (SYSTOR 2012) (20)

Webinar Sept 22: Gluster Partners with Redapt to Deliver Scale-Out NAS Storage
Webinar Sept 22: Gluster Partners with Redapt to Deliver Scale-Out NAS StorageWebinar Sept 22: Gluster Partners with Redapt to Deliver Scale-Out NAS Storage
Webinar Sept 22: Gluster Partners with Redapt to Deliver Scale-Out NAS Storage
 
Intro to GlusterFS Webinar - August 2011
Intro to GlusterFS Webinar - August 2011Intro to GlusterFS Webinar - August 2011
Intro to GlusterFS Webinar - August 2011
 
Strata + Hadoop World 2012: HDFS: Now and Future
Strata + Hadoop World 2012: HDFS: Now and FutureStrata + Hadoop World 2012: HDFS: Now and Future
Strata + Hadoop World 2012: HDFS: Now and Future
 
Gluster Webinar: Introduction to GlusterFS
Gluster Webinar: Introduction to GlusterFSGluster Webinar: Introduction to GlusterFS
Gluster Webinar: Introduction to GlusterFS
 
Introduction to GlusterFS Webinar - September 2011
Introduction to GlusterFS Webinar - September 2011Introduction to GlusterFS Webinar - September 2011
Introduction to GlusterFS Webinar - September 2011
 
Future of cloud storage
Future of cloud storageFuture of cloud storage
Future of cloud storage
 
In-memory database
In-memory databaseIn-memory database
In-memory database
 
Gluster open stack dev summit 042011
Gluster open stack dev summit 042011Gluster open stack dev summit 042011
Gluster open stack dev summit 042011
 
Data processing at the speed of 100 Gbps@Apache Crail (Incubating)
Data processing at the speed of 100 Gbps@Apache Crail (Incubating)Data processing at the speed of 100 Gbps@Apache Crail (Incubating)
Data processing at the speed of 100 Gbps@Apache Crail (Incubating)
 
PAN 9 Datasheet
PAN 9 DatasheetPAN 9 Datasheet
PAN 9 Datasheet
 
OpenStack Cinder, Implementation Today and New Trends for Tomorrow
OpenStack Cinder, Implementation Today and New Trends for TomorrowOpenStack Cinder, Implementation Today and New Trends for Tomorrow
OpenStack Cinder, Implementation Today and New Trends for Tomorrow
 
Openstorage Openstack
Openstorage OpenstackOpenstorage Openstack
Openstorage Openstack
 
인메모리 클러스터링 아키텍처
인메모리 클러스터링 아키텍처인메모리 클러스터링 아키텍처
인메모리 클러스터링 아키텍처
 
CLFS 2010
CLFS 2010CLFS 2010
CLFS 2010
 
An Overview of Apache Cassandra
An Overview of Apache CassandraAn Overview of Apache Cassandra
An Overview of Apache Cassandra
 
Openstack HA
Openstack HAOpenstack HA
Openstack HA
 
End of RAID as we know it with Ceph Replication
End of RAID as we know it with Ceph ReplicationEnd of RAID as we know it with Ceph Replication
End of RAID as we know it with Ceph Replication
 
PhegData X - High Performance EBS
PhegData X - High Performance EBSPhegData X - High Performance EBS
PhegData X - High Performance EBS
 
How swift is your Swift - SD.pptx
How swift is your Swift - SD.pptxHow swift is your Swift - SD.pptx
How swift is your Swift - SD.pptx
 
Under The Hood Of A Shard-Per-Core Database Architecture
Under The Hood Of A Shard-Per-Core Database ArchitectureUnder The Hood Of A Shard-Per-Core Database Architecture
Under The Hood Of A Shard-Per-Core Database Architecture
 

Mehr von Shun Nakamura

シリコンバレーに行ってきた!
シリコンバレーに行ってきた!シリコンバレーに行ってきた!
シリコンバレーに行ってきた!Shun Nakamura
 
第17回Cassandra勉強会: MyCassandra
第17回Cassandra勉強会: MyCassandra第17回Cassandra勉強会: MyCassandra
第17回Cassandra勉強会: MyCassandraShun Nakamura
 
読み出し性能と書き込み性能を両立させるクラウドストレージ (SACSIS2011-A6-1)
読み出し性能と書き込み性能を両立させるクラウドストレージ (SACSIS2011-A6-1)読み出し性能と書き込み性能を両立させるクラウドストレージ (SACSIS2011-A6-1)
読み出し性能と書き込み性能を両立させるクラウドストレージ (SACSIS2011-A6-1)Shun Nakamura
 
読み出し性能と書き込み性能を両立させるクラウドストレージ (OS-117-24)
読み出し性能と書き込み性能を両立させるクラウドストレージ (OS-117-24)読み出し性能と書き込み性能を両立させるクラウドストレージ (OS-117-24)
読み出し性能と書き込み性能を両立させるクラウドストレージ (OS-117-24)Shun Nakamura
 
読み出し性能と書き込み性能を選択可能なクラウドストレージ (DEIM2011-C3-3)
読み出し性能と書き込み性能を選択可能なクラウドストレージ (DEIM2011-C3-3)読み出し性能と書き込み性能を選択可能なクラウドストレージ (DEIM2011-C3-3)
読み出し性能と書き込み性能を選択可能なクラウドストレージ (DEIM2011-C3-3)Shun Nakamura
 

Mehr von Shun Nakamura (9)

HBase at LINE
HBase at LINEHBase at LINE
HBase at LINE
 
シリコンバレーに行ってきた!
シリコンバレーに行ってきた!シリコンバレーに行ってきた!
シリコンバレーに行ってきた!
 
第17回Cassandra勉強会: MyCassandra
第17回Cassandra勉強会: MyCassandra第17回Cassandra勉強会: MyCassandra
第17回Cassandra勉強会: MyCassandra
 
MyCassandra
MyCassandraMyCassandra
MyCassandra
 
読み出し性能と書き込み性能を両立させるクラウドストレージ (SACSIS2011-A6-1)
読み出し性能と書き込み性能を両立させるクラウドストレージ (SACSIS2011-A6-1)読み出し性能と書き込み性能を両立させるクラウドストレージ (SACSIS2011-A6-1)
読み出し性能と書き込み性能を両立させるクラウドストレージ (SACSIS2011-A6-1)
 
読み出し性能と書き込み性能を両立させるクラウドストレージ (OS-117-24)
読み出し性能と書き込み性能を両立させるクラウドストレージ (OS-117-24)読み出し性能と書き込み性能を両立させるクラウドストレージ (OS-117-24)
読み出し性能と書き込み性能を両立させるクラウドストレージ (OS-117-24)
 
読み出し性能と書き込み性能を選択可能なクラウドストレージ (DEIM2011-C3-3)
読み出し性能と書き込み性能を選択可能なクラウドストレージ (DEIM2011-C3-3)読み出し性能と書き込み性能を選択可能なクラウドストレージ (DEIM2011-C3-3)
読み出し性能と書き込み性能を選択可能なクラウドストレージ (DEIM2011-C3-3)
 
Cassandra勉強会
Cassandra勉強会Cassandra勉強会
Cassandra勉強会
 
ComSys WIP
ComSys WIPComSys WIP
ComSys WIP
 

Kürzlich hochgeladen

Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsMaria Levchenko
 
Maximizing Board Effectiveness 2024 Webinar.pptx
Maximizing Board Effectiveness 2024 Webinar.pptxMaximizing Board Effectiveness 2024 Webinar.pptx
Maximizing Board Effectiveness 2024 Webinar.pptxOnBoard
 
Salesforce Community Group Quito, Salesforce 101
Salesforce Community Group Quito, Salesforce 101Salesforce Community Group Quito, Salesforce 101
Salesforce Community Group Quito, Salesforce 101Paola De la Torre
 
Google AI Hackathon: LLM based Evaluator for RAG
Google AI Hackathon: LLM based Evaluator for RAGGoogle AI Hackathon: LLM based Evaluator for RAG
Google AI Hackathon: LLM based Evaluator for RAGSujit Pal
 
Scaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationScaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationRadu Cotescu
 
How to convert PDF to text with Nanonets
How to convert PDF to text with NanonetsHow to convert PDF to text with Nanonets
How to convert PDF to text with Nanonetsnaman860154
 
Understanding the Laravel MVC Architecture
Understanding the Laravel MVC ArchitectureUnderstanding the Laravel MVC Architecture
Understanding the Laravel MVC ArchitecturePixlogix Infotech
 
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...Neo4j
 
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdfThe Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdfEnterprise Knowledge
 
SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024Scott Keck-Warren
 
Histor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slideHistor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slidevu2urc
 
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | DelhiFULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhisoniya singh
 
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptxHampshireHUG
 
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking MenDelhi Call girls
 
My Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationMy Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationRidwan Fadjar
 
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 3652toLead Limited
 
Swan(sea) Song – personal research during my six years at Swansea ... and bey...
Swan(sea) Song – personal research during my six years at Swansea ... and bey...Swan(sea) Song – personal research during my six years at Swansea ... and bey...
Swan(sea) Song – personal research during my six years at Swansea ... and bey...Alan Dix
 
GenCyber Cyber Security Day Presentation
GenCyber Cyber Security Day PresentationGenCyber Cyber Security Day Presentation
GenCyber Cyber Security Day PresentationMichael W. Hawkins
 
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Miguel Araújo
 
Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024The Digital Insurer
 

Kürzlich hochgeladen (20)

Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed texts
 
Maximizing Board Effectiveness 2024 Webinar.pptx
Maximizing Board Effectiveness 2024 Webinar.pptxMaximizing Board Effectiveness 2024 Webinar.pptx
Maximizing Board Effectiveness 2024 Webinar.pptx
 
Salesforce Community Group Quito, Salesforce 101
Salesforce Community Group Quito, Salesforce 101Salesforce Community Group Quito, Salesforce 101
Salesforce Community Group Quito, Salesforce 101
 
Google AI Hackathon: LLM based Evaluator for RAG
Google AI Hackathon: LLM based Evaluator for RAGGoogle AI Hackathon: LLM based Evaluator for RAG
Google AI Hackathon: LLM based Evaluator for RAG
 
Scaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationScaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organization
 
How to convert PDF to text with Nanonets
How to convert PDF to text with NanonetsHow to convert PDF to text with Nanonets
How to convert PDF to text with Nanonets
 
Understanding the Laravel MVC Architecture
Understanding the Laravel MVC ArchitectureUnderstanding the Laravel MVC Architecture
Understanding the Laravel MVC Architecture
 
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...
 
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdfThe Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
The Role of Taxonomy and Ontology in Semantic Layers - Heather Hedden.pdf
 
SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024
 
Histor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slideHistor y of HAM Radio presentation slide
Histor y of HAM Radio presentation slide
 
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | DelhiFULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
 
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
04-2024-HHUG-Sales-and-Marketing-Alignment.pptx
 
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
 
My Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationMy Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 Presentation
 
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
Tech-Forward - Achieving Business Readiness For Copilot in Microsoft 365
 
Swan(sea) Song – personal research during my six years at Swansea ... and bey...
Swan(sea) Song – personal research during my six years at Swansea ... and bey...Swan(sea) Song – personal research during my six years at Swansea ... and bey...
Swan(sea) Song – personal research during my six years at Swansea ... and bey...
 
GenCyber Cyber Security Day Presentation
GenCyber Cyber Security Day PresentationGenCyber Cyber Security Day Presentation
GenCyber Cyber Security Day Presentation
 
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
 
Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024
 

MyCassandra: A Cloud Storage Supporting both Read Heavy and Write Heavy Workloads (SYSTOR 2012)

  • 1. + MyCassandra: A Cloud Storage Supporting both Read Heavy and Write Heavy Workloads Shunsuke Nakamura (Tokyo Institute of Technology, NHN Japan) Kazuyuki Shudo (Tokyo Inistitute of Technology) Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
  • 2. + Cloud Storage Distributed data store processing large amount of data   NoSQL, Key-Value Storage (KVS), Document-Oriented DB, GraphDB   Example: memcached, Google Bigtable, Amazon Dynamo, Amazon SimpleDB, Apache Cassandra, Voldemort, Ringo, Vpork, MongoDB, CouchDB, Tokyo Tyrant, Flare, ROMA, kumofs, Kai, Redis, LevelDB, Hadoop HBase, Hypertable,Yahoo! PNUTS, Scalaris, Dynomite, ThruDB, Neo4j, IBM ObjectGrid, Giraph, Oracle Coherence, and the others. (> 100 products)   Characteristics: “limited functions, massive volume, high performance”   Data access only by primary key   No luxury features such as join, global transaction   Scalable to much larger data and number of nodes
  • 3. + Design policies of cloud storages There are many trade-offs.   data model   key/value, multi-dimensional map, document or graph   performance - write vs. read   latency vs. persistence   latency – memory and disk utilization   persistence – synchronous vs. asynchronous (snapshot)   replication – synchronous vs. asynchronous   consistency between replicas – strong vs. weak   data partitioning – row vs. column   distribution – master/slave vs. decentralized Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
  • 4. + MyCassandra focuses on performance trade-off   data model   key/value vs. multi-dimensional map vs. document vs. graph   performance - write vs. read   latency vs. persistence   latency – memory and disk utilization   persistence – synchronous vs. asynchronous (snapshot)   replication – synchronous vs. asynchronous   consistency between replicas – strong vs. weak   data partitioning – row vs. column   distribution – master/slave vs. decentralized Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
  • 5. + Performance trade-off Write-optimized vs. read-optimized   A cloud storage with persistence is designed to optimize either write or read workload.   Storage engine determines which workload a cloud storage treats efficiently. Bigtable, Cassandra, MySQL, Yahoo! HBase Sherpa Indexing Log-Structured B-Trees [R.Bayer ’70] Merge Tree [P. O’Neil ‘96] Write to disk append random reads, writes Read to disk random reads + merge random read Performance write-optimized read-optimized Storage engine Bigtable clone MySQL Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
  • 6. + Performance trade-off - write-optimized vs. read-optimized -  Write latency for write-heavy workload Better read-optimized write-optimized 6 Yahoo! Cloud Serving Benchmark, SOCC ’10 - mycassandra - Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
  • 7. + Performance trade-off - write-optimized vs. read-optimized - Read latency for read-heavy workload write-optimized Better read-optimized Yahoo! Cloud Serving Benchmark, SOCC ’10 - mycassandra - Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
  • 8. + Research overview   Contribution: A technique to build a cloud storage performing well with both read and write workloads   Steps: 1.  MyCassandra: Storage engine enabled Apache Cassandra 2.  MyCassandra Cluster: Heterogeneous cluster with different storage engines  1. MyCassandra 2. MyCassandra Cluster read-optimized read and write-optimized select write-optimized Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
  • 9. + Apache Cassandra Open-sourced by in 2008 A top-level project in   Features:   Scalability up to hundreds of servers across multiple racks/datacenters   High availability without SPOF by adopting a decentralized architecture   Write-optimized dc1 dc2 Clustering across multiple racks/DCs Replication strategy based on region dc3 Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
  • 10. + Apache Cassandra A decentralized cloud storage without SPOF   Consistent Hashing (a decentralized algorithm): Assign identifiers to both nodes and data on its circular ID space. A-Z: hash value Num of replica := 3 ID space A F Z Roles of each node secondary 1 •  Proxy, serving clients Q •  Primary/secondary data nodes V N primary secondary 2 hash(key) = Q key values Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
  • 11. + Apache Cassandra Write-optimized storage engine, a Bigtable clone   O(1) fast write operation   Write an update to disk sequentially - Fast because of no random I/O to disk - Always writable because of no write-lock memory sync <k1, obj (v1+v2)> async flush write path Memtable 1.  Append an update to CommitLog for persistence Only sequential write disk 2.  Update Memtable, a map in <k1, v1>, <k1, v2> memory, for quick reading 3.  Acknowledge a client CommitLog 4.  Asynchronously flush Memtable <k1,obj1> to SSTable SSTable 1 5.  Delete flushed data from <k1,obj2> CommitLog and Memtable SSTable 2 <k1,obj3> Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6) SSTable 3
  • 12. + Apache Cassandra Write-optimized storage engine, a Bigtable clone   Slow read operation   Read data from Memtable and multiple SSTables, and merge them - Slow because of multiple random I/Os on disk memory <k1,obj> Memtable disk CommitLog merge <k1,obj1> SSTable 1 multiple random I/Os <k1,obj2> SSTable 2 <k1,obj3> SSTable 3 Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
  • 13. + Performance of original Cassandra Write performance is much higher.   YCSB results show:   Average: write is 9 x as fast as read.   99.9%ile: write is 43.5 x as fast as read. Better read Number of operations write avg. 6.16 ms read Latency (ms) 99.9 %ile write write: 2.0 ms avg. 0.69 ms read: 86.9 ms Latency (ms)
  • 14. 1. Storage Engine Support +  1.MyCassandra read-optimized select write-optimized Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
  • 15. + MyCassandra: A modular cloud storage Storage engines are supported   Storage engine feature inspired by MySQL   An independent and pluggable component   Perform disk I/O   A cloud storage can be either write-optimized or read-optimized by selecting storage engine   Keep Cassandra’s original distribution architecture and data model Decentralized Consistent Hashing Bigtable Gossip Protocol Bigtable MySQL Redis … selectable InnoDB MyISAM Memory … selectable Decentralized + Storage engine Storage engine
  • 16. MyCassandra implementation Cassandra’s original distribution arch. Storage Engine Interface introduced Implement Storage Engine Storage Engine Interface Interface Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
  • 17. Performance of each storage engine   storage engines   Bigtable: write-optimized (original Casssandra 0.7.5)   MySQL: read-optimized (MySQL 6.0 with InnoDB, JDBC API, stored procedure)   Redis: in-memory KVS (Redis 2.2.8) 6 nodes -  Crucial’s SSD -  allocate 6GB mem in 8GB x 11.79 1KB x 36 million data set x 9.87 workload Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
  • 18. 2. Heterogeneous cluster of different storage engines + 2.MyCassandra Cluster read and write-optimized Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
  • 19. •  W: write-optimized Basic idea •  R: read-optimized •  RW: in-memory   Replicate data on different storage engine nodes write query   Route a query to nodes processing it efficiently   Synchronously route to nodes processing quickly sync async   Asynchronously route to nodes processing slowly → Exploit each node’s advantage W R   Furthermore, maintain consistency between replicas as much as the original Cassandra Quorum Protocol: (write agreements) + (read agreements) > (num of replicas) = Guarantee retrieval of the latest data write read Consequence: At least one node processes both read and write queries synchronously and quickly → In-memory nodes play this role. W RW R Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
  • 20. •  W: write-optimized Cluster design •  R: read-optimized •  RW: in-memory   Combine nodes with different storage engines   write-optimized (W), read-optimized (R), in-memory (RW)   Disseminate storage engine types of each nodes   The type is attached to gossip messages   Place replicas on nodes with different storage engines   Proxy (any node requested) selects the storing nodes 1.  The primary node determined based on the queried key 2.  N -1 secondary nodes with different storage engines   Multiple nodes share a single server for load balance Proxy (any node) Cluster configuration (N=3) gossip RW W W RW RW R RW R W secondary2 primary secondary1 responsible nodes
  • 21. •  W: write-optimized Process for a write access •  R: read-optimized •  RW: in-memory •  Quorum parameters Client = 3, = = 2 1)  A proxy receives a write query •  Num. of replicas Write for a single record from a client. The proxy routes to nodes storing the record. W:RW:R = 1:1:1 Proxy … … 2)  The proxy waits ACKs. W, RW … nodes usually reply quickly. Wait for two ACKs … for write and return 3-a) If writing succeeds and the RW proxy receives ACKs, it Async write returns a success message. R W 3-b) If a data node fails to write, W RW R the proxy waits for ACKs including R nodes and returns a Nodes storing the record success message. 4) After returning, the proxy Write latency: max (W, RW) asynchronously waits ACKs from the remaining nodes. Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
  • 22. •  W: write-optimized Process for a read access •  R: read-optimized •  RW: in-memory •  Quorum parameters = 3, = = 2 Client 1)  A proxy receives a read query and routes to storing nodes. •  Num. of replicas Read for a single record W:RW:R = 1:1:1 2)  Theproxy waits for ACKs. R Proxy and RW nodes reply quickly. … … … 3-a) If returned values are … consistent, the proxy returns it. Async check Check consistentcy RW consistency 3-b) If the values are mismatched, and return result the proxy waits for consistent R W values including W nodes. W RW R 4) After returning, the proxy waits Nodes storing the record from the remaining nodes. If the proxy notices inconsistent Read latency: max (R, RW) values, it asynchronously updates them to the consistent one. Cassandra’s feature Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6) ReadRepair does it.
  • 23. + Performance Evaluation Demonstrate that a heterogeneous cluster performs well with both read- and write-heavy workloads   Targets   MyCassandra Cluster: 3 different nodes/server x 6 servers   Cassandra: 1 node/server x 6 servers   Quorum parameters = 3, = =2   Storage Engine   Bigtable (W), MySQL / InnoDB (R), Redis (RW)   Yahoo! Cloud Serving Benchmark (YCSB) [SOCC ’10] 1.  Load data (1KB record, 10 x 100bytes columns) from a YCSB client 2.  Warm up 3.  Run benchmark and measure response times from a client Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
  • 24. + YCSB workloads Workload Application Operation Record example ratio selection Write-Only Log Read: 0% Zipfian Write Write: 100% heavy Write-Heavy Session store Read: 50% Write: 50% Read-Heavy Photo Read: 95% Read Write: 5% heavy tagging Read-Only Cache Read: 100% Write: 0% Zipfian distribution: the access frequency of each datum is determined by its popularity, not by freshness. Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
  • 25. Write/Read latency (Response time) 1.5 avg. write-latency Cassandra + 0.57ms (max) MyCassandra Cluster 1 Better + 42.5% + 59.5% + 69.5% Performs well 0.5 with write:5% MySQL + Redis write:100% write:50% write:0% 0 (ms) max 90.4% lower in read-only workload 35 30 avg. read-latency - 26.5ms (max) 25 20 Better - 88.8% - 90.4% 15 - 83.3% 10 5 read:0% read:50% read:95% read:100% 0 (ms) Write-Only Write-Heavy Read-Heavy Read-Only
  • 26. Throughput 25000 Cassandra x 0.87 QPS for 40 clients MyCassandra Cluster 20000 15000 Better 10000 x 2.16 x 11.00 x 4.07 5000 0 [100:0] [50:50] [5:95] [0:100] [write:read] (query/sec) Write-Only Write-Heavy Read-Heavy Read-Only Write heavy Read heavy •  11.0 times as high as Cassandra in Read-Only workload •  Write performance is comparable with Cassandra. Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
  • 27. + Conclusion   A cloud storage supporting both write-heavy and read- heavy workloads by combining different storage engine nodes.   MyCassandra Cluster achieved better throughput than the original Cassandra on read heavy workload.   With a read-heavy workload   Read latency: 90.4 % lower at most   Throughput: 11.0 times at most Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
  • 28. + Related Work   Indexing algorithm whose goals include achieving both write and read performance   FD-Tree: Tree Indexing on Flash Disks, VLDB ’10   bLSM: A General Purpose Log Structured Merge Tree, SIGMOD ‘12   Fractal-Tree: It’s implemented in TokuDB (MySQL storage engine)   Modular data stores:   MySQL   Anvil, SOSP ’09   Cloudy, VLDB ’10   Dynamo, SOSP ’07   Fractured Mirrors:   MyCassandra, SYSTOR ‘12: read vs. write Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
  • 29. Discussion 1. the slight higher + write latency The cause is load balancing.   Cassandra   Write to any nodes in N nodes   MyCassandra Cluster   Write to the specified and nodes However this cost well worths improving for read performance. MyCassandra Cassandra Cluster Sync operation is write read write read Sync operation is equally distributed. fixed. W RW R Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
  • 30. + Discussion 2. in-memory node Q. Memory overflow A. In-memory node plays as LRU-like cache. The swapped data is recovered from the other persistent nodes by read repair. Q. Fault tolerance A. 1) Write to an alternative node, and if the node is recovered, it resolves inconsistency using values from the node. 2) Asynchronous snapshot (Redis’s feature) Q. Whole in-memory nodes A. This case limits capacities in cluster with the memory’s capacity. Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
  • 31. + オープンソース化 Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)
  • 32. + Session 6 - Storage, SYSTOR 2012 (Haifa, Israel, Jun 4-6)