Illustration Image

Cassandra.Link

The best knowledge base on Apache Cassandra®

Helping platform leaders, architects, engineers, and operators build scalable real time data platforms.

10/30/2017

Reading time:9 min

Everyday I’m scaling... Cassandra

by Instaclustr

Everyday I’m scaling... Cassandra SlideShare Explore You Successfully reported this slideshow.Everyday I’m scaling... CassandraUpcoming SlideShareLoading in …5× 0 Comments 0 Likes Statistics Notes Be the first to like this No DownloadsNo notes for slide 1. Ben BromheadCassandra… Every day I’m scaling 2. 2© DataStax, All Rights Reserved. 3. Who am I and What do I do?• Co-founder and CTO of Instaclustr -> www.instaclustr.com• Instaclustr provides Cassandra-as-a-Service in the cloud.• Currently support AWS, Azure, Heroku, Softlayer and Private DCs with more to come.• Approaching 1000 nodes under management• Yes… we are hiring! Come live in Australia!© DataStax, All Rights Reserved. 3 4. 1 Why scaling sucks in Cassandra2 It gets better3 Then it gets really awesome4© DataStax, All Rights Reserved. 5. Linear Scalability – In theory© DataStax, All Rights Reserved. 5 6. Linear Scalability – In practice© DataStax, All Rights Reserved. 6 7. What’s supposed to happen• Scaling Cassandra is just “bootstrap new nodes”• That works if your cluster is under provisioned and has 30% disk usage© DataStax, All Rights Reserved. 7 8. What actually happens• Add 1 node• Bootstrapping node fails (1 day)• WTF - Full disk on bootstrapping node? (5 minutes)• If STCS run SSTableSplit on large SSTables on original nodes (2 days)• Attach super sized network storage (EBS) and bind mount to bootstrapping node.© DataStax, All Rights Reserved. 8 9. What actually happens• Restart bootstrapping process• Disk alert 70% (2 days later)• Throttle streaming throughput to below compaction throughput• Bootstrapping finishes (5 days later)• Cluster latency spikes cause bootstrap finished but their was a million compactionsremaining• Take node offline and let compaction finish• Run repair on node (10 years)• Add next node.© DataStax, All Rights Reserved. 9 10. What actually happens© DataStax, All Rights Reserved. 10 11. Scalability in Cassandra sucks• Soo much over streaming• LCS and Bootstrap – Over stream then compact all the data!• STCS and bootstrap – Over stream all the data and run out of disk space© DataStax, All Rights Reserved. 11 12. Scalability in Cassandra sucks• No vnodes? – Can only double your cluster• Vnodes? – Can only add one node at a time• Bootstrap – Fragile and not guaranteed to be consistent© DataStax, All Rights Reserved. 12 13. Why does it suck for youYour database never meets your business requirements from a capacity perspective(bad) and if you try…• You could interrupt availability and performance (really bad)• You could loose data (really really bad)© DataStax, All Rights Reserved. 13 14. How did it get this way?It’s actually a hard problem:• Moving large amounts of data between nodes requires just as much attention to itfrom a CAP perspective as client facing stuff.• New features don’t tend to consider impact on scaling operations• Features that help ops tends to be less sexy© DataStax, All Rights Reserved. 14 15. Does it get better?© DataStax, All Rights Reserved. 15Yes! 16. Does it get better? Consistent bootstrapStrongly consistent membership and ownership – CASSANDRA-9667• Using LWT to propose and claim ownership of new token allocations in a consistentmanner• Work in progress• You can do this today by pre-assigning non-overlapping (inc replicas) vnode tokensand using cassandra.consistent.simultaneousmoves.allow=true as a JVM propertybefore bootstrapping your nodes© DataStax, All Rights Reserved. 16 17. Does it get better? Bootstrap stabilityKeep-alives for all streaming operations – CASSANDRA-11841• Currently implements a timeout, you can reduce this to be more aggressive, but largeSSTables will then never streamResummable bootstrap – CASSANDRA-8942 & CASSANDRA-8838• You can do this in 2.2+Incremental bootstrap – CASSANDRA-8494• Being worked on, hard to do with vnodes right now (try it… the error message usesthe word “thusly”), instead throttle streaming and uncap compaction to ensure thenode doesn’t get overloaded during bootstrap© DataStax, All Rights Reserved. 17 18. Can we make it even better?© DataStax, All Rights Reserved. 18Yes! 19. Can we make it even better?© DataStax, All Rights Reserved. 19• Let’s try scaling without data ownership changes• Take advantage of Cassandras normal partition and availability mechanisms• With a little help from our cloud providers! 20. Introducing token pinned scaling© DataStax, All Rights Reserved. 20• Probably needs a better name• Here is how it works 21. Introducing token pinned scaling© DataStax, All Rights Reserved. 21With the introduction of:• Partitioning SSTables by Range (CASSANDRA-6696)• Range Aware Compaction (CASSANDRA-10540)• A few extra lines of code to save/load a map of token to disks (coming soon)Cassandra will now keep data associated with specific tokens in a single data directory,this could let us treat a disk as a unit in which to scale around!But first what do these two features actually let us do? 22. Introducing token pinned scaling© DataStax, All Rights Reserved. 22Before Partitioning SSTables by Range and Range Aware Compaction:1 - 100901 - 10001401-1500Disk0 Disk1SSTables 23. Introducing token pinned scaling© DataStax, All Rights Reserved. 23After Partitioning SSTables by Range and Range Aware Compaction:1 - 100901 - 10001401-1500Disk0 Disk1SSTablesData within a token range is now kept on a specific disk 24. Introducing token pinned scaling© DataStax, All Rights Reserved. 24Your SSTables will converge to contain a single vnode range when things get big enough1 - 100901 - 10001401-1500Disk0 Disk1SSTables 25. Leveraging EBS to separate I/O from CPU© DataStax, All Rights Reserved. 25• Amazon Web Services provides a networked attached block store called EBS (ElasticBlock Store).• Isolated to each availability zone• We can attach and reattach EBS disk ad-hoc and in seconds/minutes 26. Adding it all together© DataStax, All Rights Reserved. 26• Make each EBS disk a data directory in Cassandra• Cassandra guarantees only data from a specific token range will exist on a given disk• When throughput is low attach all disks in a single AZ to a single node, specify all theranges from each disk via a comma separated list of tokens.• Up to 40 disks per instance!• When load is high, launch more instances and spread the disks across the newinstances. 27. Adding it all together© DataStax, All Rights Reserved. 27• Make each EBS disk a data directory in CassandrasdasddsdbsdesdcsdfAmazon EBS 28. Adding it all together© DataStax, All Rights Reserved. 28• Cassandra guarantees only data from a specific token range will exist on a given diskAmazon EBS 29. Adding it all together© DataStax, All Rights Reserved. 29• When throughput is low attach all disks in a single AZ to a single nodeAmazon EBS200 op/s 30. Adding it all together© DataStax, All Rights Reserved. 30• When load is high, launch more instances and spread the disks across the new instances.Amazon EBS10,000 op/s 31. How it works - Scaling© DataStax, All Rights Reserved. 31• Normally you have to provision your cluster at your maximum operations per second +30% (headroom in case your get it wrong).• Provision enough IOPS, CPU, RAM etc• Makes Cassandra an $$$ solutionProvisioned workloadActual workload 32. © DataStax, All Rights Reserved. 32 33. How it works - Scaling© DataStax, All Rights Reserved. 33• Let’s make our resources match our workloadProvisioned IOPS workloadActual workloadProvisioned CPU & RAM workload 34. How it works - Scaling© DataStax, All Rights Reserved. 34• Let’s make our resources match our workloadProvisioned IOPS workloadActual workloadProvisioned CPU & RAM workload 35. How it works - Consistency© DataStax, All Rights Reserved. 35• No range movements! You don’t need a Jepsen test to see how bad rangemovements are for consistency.• Tokens and ranges are fixed during all scaling operations• Range movements are where you see most consistency badness in Cassandra(bootstrap, node replacement, decommission) and need to rely on repair. 36. How it works - Consistency© DataStax, All Rights Reserved. 36• Treats Racks as a giant “meta-node”, network topology strategy ensures replicas areon different racks.• AWS Rack == AZ• As tokens for a node change based on the disk they have, replica topology stays thesame• You can only swap disks between instances within the same AZ• Scale one rack at a time… scale your cluster in constant time!• If you want to do this with a single rack, you will have a bad time 37. How it works - Consistency© DataStax, All Rights Reserved. 371,5,102,6,113, 22, 444,23,45102,134,167101,122,1551,2,3,4,5,6,10,11,22,23,44,45,101 … 38. How it works - Consistency© DataStax, All Rights Reserved. 381,5,102,6,113, 22, 444,23,45102,134,167101,122,1551,5,102,6,113,22,444,23,45102,134,167101,122,155 39. How it works - TODO© DataStax, All Rights Reserved. 39Some issues remain:• Hinted handoff breaks (handoff is based on endpoint rather than token)• Time for gossip to settle on any decent sized cluster• Currently just clearing out the system.local folder to allow booting• Can’t do this while repair is running… for some people this is all the time• You’ll need to run repair more often as scaling intentionally introduces outages• Breaks consistency and everything where RF > number of racks (usually thesystem_auth keyspace).• More work needed! 40. How it works – Real world© DataStax, All Rights Reserved. 40• No production tests yet • Have gone from a 3 node cluster to a 36 node cluster in around 50 minutes.• Plenty left to optimize (e.g. bake everything into an AMI to reduce startup time)• Could get this down to 10 minutes per rack depending on how responsive AWS is!• No performance overhead compared to Cassandra on EBS.• Check out the code here: https://github.com/benbromhead/Cassandra/tree/ic-token-pinning 41. How it works – Real world© DataStax, All Rights Reserved. 41• Really this is bending some new and impending changes to do funky stuff  42. Questions? 43. Questions? Recommended PowerPoint Tips WeeklyOnline Course - LinkedIn Learning Information LiteracyOnline Course - LinkedIn Learning PowerPoint 2016: Tips and TricksOnline Course - LinkedIn Learning Getting Started with Apache Cassandra and Apache Zeppelin (DuyHai DOAN, DataS...DataStax Time Series Processing with Apache SparkJosef Adersberger Frustration-Reduced Spark: DataFrames and the Spark Time-Series LibraryIlya Ganelin Analyzing Time Series Data with Apache Spark and CassandraPatrick McFadin Apache Cassandra Community HealthInstaclustr Instaclustr introduction to managing cassandraInstaclustr Instaclustr webinar 50,000 transactions per second with Apache Spark on Apach...Instaclustr About Blog Terms Privacy Copyright LinkedIn Corporation © 2017 Public clipboards featuring this slideNo public clipboards found for this slideSelect another clipboard ×Looks like you’ve clipped this slide to already.Create a clipboardYou just clipped your first slide! Clipping is a handy way to collect important slides you want to go back to later. Now customize the name of a clipboard to store your clips. Description Visibility Others can see my Clipboard

Illustration Image
Everyday I’m scaling... Cassandra

Successfully reported this slideshow.

Everyday I’m scaling... Cassandra
Ben Bromhead
Cassandra… Every day I’m scaling
2© DataStax, All Rights Reserved.
Who am I and What do I do?
• Co-founder and CTO of Instaclustr -> www.instaclustr.com
• Instaclustr provides Cassandra-as-...
1 Why scaling sucks in Cassandra
2 It gets better
3 Then it gets really awesome
4© DataStax, All Rights Reserved.
Linear Scalability – In theory
© DataStax, All Rights Reserved. 5
Linear Scalability – In practice
© DataStax, All Rights Reserved. 6
What’s supposed to happen
• Scaling Cassandra is just “bootstrap new nodes”
• That works if your cluster is under provisio...
What actually happens
• Add 1 node
• Bootstrapping node fails (1 day)
• WTF - Full disk on bootstrapping node? (5 minutes)...
What actually happens
• Restart bootstrapping process
• Disk alert 70% (2 days later)
• Throttle streaming throughput to b...
What actually happens
© DataStax, All Rights Reserved. 10
Scalability in Cassandra sucks
• Soo much over streaming
• LCS and Bootstrap – Over stream then compact all the data!
• ST...
Scalability in Cassandra sucks
• No vnodes? – Can only double your cluster
• Vnodes? – Can only add one node at a time
• B...
Why does it suck for you
Your database never meets your business requirements from a capacity perspective
(bad) and if you...
How did it get this way?
It’s actually a hard problem:
• Moving large amounts of data between nodes requires just as much ...
Does it get better?
© DataStax, All Rights Reserved. 15
Yes!
Does it get better? Consistent bootstrap
Strongly consistent membership and ownership – CASSANDRA-9667
• Using LWT to prop...
Does it get better? Bootstrap stability
Keep-alives for all streaming operations – CASSANDRA-11841
• Currently implements ...
Can we make it even better?
© DataStax, All Rights Reserved. 18
Yes!
Can we make it even better?
© DataStax, All Rights Reserved. 19
• Let’s try scaling without data ownership changes
• Take ...
Introducing token pinned scaling
© DataStax, All Rights Reserved. 20
• Probably needs a better name
• Here is how it works
Introducing token pinned scaling
© DataStax, All Rights Reserved. 21
With the introduction of:
• Partitioning SSTables by ...
Introducing token pinned scaling
© DataStax, All Rights Reserved. 22
Before Partitioning SSTables by Range and Range Aware...
Introducing token pinned scaling
© DataStax, All Rights Reserved. 23
After Partitioning SSTables by Range and Range Aware ...
Introducing token pinned scaling
© DataStax, All Rights Reserved. 24
Your SSTables will converge to contain a single vnode...
Leveraging EBS to separate I/O from CPU
© DataStax, All Rights Reserved. 25
• Amazon Web Services provides a networked att...
Adding it all together
© DataStax, All Rights Reserved. 26
• Make each EBS disk a data directory in Cassandra
• Cassandra ...
Adding it all together
© DataStax, All Rights Reserved. 27
• Make each EBS disk a data directory in Cassandra
sda
sdd
sdb
...
Adding it all together
© DataStax, All Rights Reserved. 28
• Cassandra guarantees only data from a specific token range wi...
Adding it all together
© DataStax, All Rights Reserved. 29
• When throughput is low attach all disks in a single AZ to a s...
Adding it all together
© DataStax, All Rights Reserved. 30
• When load is high, launch more instances and spread the disks...
How it works - Scaling
© DataStax, All Rights Reserved. 31
• Normally you have to provision your cluster at your maximum o...
© DataStax, All Rights Reserved. 32
How it works - Scaling
© DataStax, All Rights Reserved. 33
• Let’s make our resources match our workload
Provisioned IOPS ...
How it works - Scaling
© DataStax, All Rights Reserved. 34
• Let’s make our resources match our workload
Provisioned IOPS ...
How it works - Consistency
© DataStax, All Rights Reserved. 35
• No range movements! You don’t need a Jepsen test to see h...
How it works - Consistency
© DataStax, All Rights Reserved. 36
• Treats Racks as a giant “meta-node”, network topology str...
How it works - Consistency
© DataStax, All Rights Reserved. 37
1,5,10
2,6,11
3, 22, 44
4,23,45
102,134,167
101,122,155
1,2...
How it works - Consistency
© DataStax, All Rights Reserved. 38
1,5,10
2,6,11
3, 22, 44
4,23,45
102,134,167
101,122,155
1,5...
How it works - TODO
© DataStax, All Rights Reserved. 39
Some issues remain:
• Hinted handoff breaks (handoff is based on e...
How it works – Real world
© DataStax, All Rights Reserved. 40
• No production tests yet 
• Have gone from a 3 node cluste...
How it works – Real world
© DataStax, All Rights Reserved. 41
• Really this is bending some new and impending changes to d...
Questions?
Questions?

Upcoming SlideShare

Loading in …5

×

  • Be the first to like this

  1. 1. Ben Bromhead Cassandra… Every day I’m scaling
  2. 2. 2© DataStax, All Rights Reserved.
  3. 3. Who am I and What do I do? • Co-founder and CTO of Instaclustr -> www.instaclustr.com • Instaclustr provides Cassandra-as-a-Service in the cloud. • Currently support AWS, Azure, Heroku, Softlayer and Private DCs with more to come. • Approaching 1000 nodes under management • Yes… we are hiring! Come live in Australia! © DataStax, All Rights Reserved. 3
  4. 4. 1 Why scaling sucks in Cassandra 2 It gets better 3 Then it gets really awesome 4© DataStax, All Rights Reserved.
  5. 5. Linear Scalability – In theory © DataStax, All Rights Reserved. 5
  6. 6. Linear Scalability – In practice © DataStax, All Rights Reserved. 6
  7. 7. What’s supposed to happen • Scaling Cassandra is just “bootstrap new nodes” • That works if your cluster is under provisioned and has 30% disk usage © DataStax, All Rights Reserved. 7
  8. 8. What actually happens • Add 1 node • Bootstrapping node fails (1 day) • WTF - Full disk on bootstrapping node? (5 minutes) • If STCS run SSTableSplit on large SSTables on original nodes (2 days) • Attach super sized network storage (EBS) and bind mount to bootstrapping node. © DataStax, All Rights Reserved. 8
  9. 9. What actually happens • Restart bootstrapping process • Disk alert 70% (2 days later) • Throttle streaming throughput to below compaction throughput • Bootstrapping finishes (5 days later) • Cluster latency spikes cause bootstrap finished but their was a million compactions remaining • Take node offline and let compaction finish • Run repair on node (10 years) • Add next node. © DataStax, All Rights Reserved. 9
  10. 10. What actually happens © DataStax, All Rights Reserved. 10
  11. 11. Scalability in Cassandra sucks • Soo much over streaming • LCS and Bootstrap – Over stream then compact all the data! • STCS and bootstrap – Over stream all the data and run out of disk space © DataStax, All Rights Reserved. 11
  12. 12. Scalability in Cassandra sucks • No vnodes? – Can only double your cluster • Vnodes? – Can only add one node at a time • Bootstrap – Fragile and not guaranteed to be consistent © DataStax, All Rights Reserved. 12
  13. 13. Why does it suck for you Your database never meets your business requirements from a capacity perspective (bad) and if you try… • You could interrupt availability and performance (really bad) • You could loose data (really really bad) © DataStax, All Rights Reserved. 13
  14. 14. How did it get this way? It’s actually a hard problem: • Moving large amounts of data between nodes requires just as much attention to it from a CAP perspective as client facing stuff. • New features don’t tend to consider impact on scaling operations • Features that help ops tends to be less sexy © DataStax, All Rights Reserved. 14
  15. 15. Does it get better? © DataStax, All Rights Reserved. 15 Yes!
  16. 16. Does it get better? Consistent bootstrap Strongly consistent membership and ownership – CASSANDRA-9667 • Using LWT to propose and claim ownership of new token allocations in a consistent manner • Work in progress • You can do this today by pre-assigning non-overlapping (inc replicas) vnode tokens and using cassandra.consistent.simultaneousmoves.allow=true as a JVM property before bootstrapping your nodes © DataStax, All Rights Reserved. 16
  17. 17. Does it get better? Bootstrap stability Keep-alives for all streaming operations – CASSANDRA-11841 • Currently implements a timeout, you can reduce this to be more aggressive, but large SSTables will then never stream Resummable bootstrap – CASSANDRA-8942 & CASSANDRA-8838 • You can do this in 2.2+ Incremental bootstrap – CASSANDRA-8494 • Being worked on, hard to do with vnodes right now (try it… the error message uses the word “thusly”), instead throttle streaming and uncap compaction to ensure the node doesn’t get overloaded during bootstrap © DataStax, All Rights Reserved. 17
  18. 18. Can we make it even better? © DataStax, All Rights Reserved. 18 Yes!
  19. 19. Can we make it even better? © DataStax, All Rights Reserved. 19 • Let’s try scaling without data ownership changes • Take advantage of Cassandras normal partition and availability mechanisms • With a little help from our cloud providers!
  20. 20. Introducing token pinned scaling © DataStax, All Rights Reserved. 20 • Probably needs a better name • Here is how it works
  21. 21. Introducing token pinned scaling © DataStax, All Rights Reserved. 21 With the introduction of: • Partitioning SSTables by Range (CASSANDRA-6696) • Range Aware Compaction (CASSANDRA-10540) • A few extra lines of code to save/load a map of token to disks (coming soon) Cassandra will now keep data associated with specific tokens in a single data directory, this could let us treat a disk as a unit in which to scale around! But first what do these two features actually let us do?
  22. 22. Introducing token pinned scaling © DataStax, All Rights Reserved. 22 Before Partitioning SSTables by Range and Range Aware Compaction: 1 - 100 901 - 1000 1401-1500 Disk0 Disk1 SSTables
  23. 23. Introducing token pinned scaling © DataStax, All Rights Reserved. 23 After Partitioning SSTables by Range and Range Aware Compaction: 1 - 100 901 - 1000 1401-1500 Disk0 Disk1 SSTables Data within a token range is now kept on a specific disk
  24. 24. Introducing token pinned scaling © DataStax, All Rights Reserved. 24 Your SSTables will converge to contain a single vnode range when things get big enough 1 - 100 901 - 1000 1401-1500 Disk0 Disk1 SSTables
  25. 25. Leveraging EBS to separate I/O from CPU © DataStax, All Rights Reserved. 25 • Amazon Web Services provides a networked attached block store called EBS (Elastic Block Store). • Isolated to each availability zone • We can attach and reattach EBS disk ad-hoc and in seconds/minutes
  26. 26. Adding it all together © DataStax, All Rights Reserved. 26 • Make each EBS disk a data directory in Cassandra • Cassandra guarantees only data from a specific token range will exist on a given disk • When throughput is low attach all disks in a single AZ to a single node, specify all the ranges from each disk via a comma separated list of tokens. • Up to 40 disks per instance! • When load is high, launch more instances and spread the disks across the new instances.
  27. 27. Adding it all together © DataStax, All Rights Reserved. 27 • Make each EBS disk a data directory in Cassandra sda sdd sdb sde sdc sdf Amazon EBS
  28. 28. Adding it all together © DataStax, All Rights Reserved. 28 • Cassandra guarantees only data from a specific token range will exist on a given disk Amazon EBS
  29. 29. Adding it all together © DataStax, All Rights Reserved. 29 • When throughput is low attach all disks in a single AZ to a single node Amazon EBS 200 op/s
  30. 30. Adding it all together © DataStax, All Rights Reserved. 30 • When load is high, launch more instances and spread the disks across the new instances. Amazon EBS 10,000 op/s
  31. 31. How it works - Scaling © DataStax, All Rights Reserved. 31 • Normally you have to provision your cluster at your maximum operations per second + 30% (headroom in case your get it wrong). • Provision enough IOPS, CPU, RAM etc • Makes Cassandra an $$$ solution Provisioned workload Actual workload
  32. 32. © DataStax, All Rights Reserved. 32
  33. 33. How it works - Scaling © DataStax, All Rights Reserved. 33 • Let’s make our resources match our workload Provisioned IOPS workload Actual workload Provisioned CPU & RAM workload
  34. 34. How it works - Scaling © DataStax, All Rights Reserved. 34 • Let’s make our resources match our workload Provisioned IOPS workload Actual workload Provisioned CPU & RAM workload
  35. 35. How it works - Consistency © DataStax, All Rights Reserved. 35 • No range movements! You don’t need a Jepsen test to see how bad range movements are for consistency. • Tokens and ranges are fixed during all scaling operations • Range movements are where you see most consistency badness in Cassandra (bootstrap, node replacement, decommission) and need to rely on repair.
  36. 36. How it works - Consistency © DataStax, All Rights Reserved. 36 • Treats Racks as a giant “meta-node”, network topology strategy ensures replicas are on different racks. • AWS Rack == AZ • As tokens for a node change based on the disk they have, replica topology stays the same • You can only swap disks between instances within the same AZ • Scale one rack at a time… scale your cluster in constant time! • If you want to do this with a single rack, you will have a bad time
  37. 37. How it works - Consistency © DataStax, All Rights Reserved. 37 1,5,10 2,6,11 3, 22, 44 4,23,45 102,134,167 101,122,155 1,2,3,4,5,6,10,11,22,23,44,45,101 …
  38. 38. How it works - Consistency © DataStax, All Rights Reserved. 38 1,5,10 2,6,11 3, 22, 44 4,23,45 102,134,167 101,122,155 1,5,10 2,6,11 3,22,44 4,23,45 102,134,167 101,122,155
  39. 39. How it works - TODO © DataStax, All Rights Reserved. 39 Some issues remain: • Hinted handoff breaks (handoff is based on endpoint rather than token) • Time for gossip to settle on any decent sized cluster • Currently just clearing out the system.local folder to allow booting • Can’t do this while repair is running… for some people this is all the time • You’ll need to run repair more often as scaling intentionally introduces outages • Breaks consistency and everything where RF > number of racks (usually the system_auth keyspace). • More work needed!
  40. 40. How it works – Real world © DataStax, All Rights Reserved. 40 • No production tests yet  • Have gone from a 3 node cluster to a 36 node cluster in around 50 minutes. • Plenty left to optimize (e.g. bake everything into an AMI to reduce startup time) • Could get this down to 10 minutes per rack depending on how responsive AWS is! • No performance overhead compared to Cassandra on EBS. • Check out the code here: https://github.com/benbromhead/Cassandra/tree/ic-token- pinning
  41. 41. How it works – Real world © DataStax, All Rights Reserved. 41 • Really this is bending some new and impending changes to do funky stuff 
  42. 42. Questions?
  43. 43. Questions?

Related Articles

cassandra
devops

Cassandra - DevOps

niqdev

2/17/2023

cassandra.deployment
cassandra.lunch
cassandra

Checkout Planet Cassandra

Claim Your Free Planet Cassandra Contributor T-shirt!

Make your contribution and score a FREE Planet Cassandra Contributor T-Shirt! 
We value our incredible Cassandra community, and we want to express our gratitude by sending an exclusive Planet Cassandra Contributor T-Shirt you can wear with pride.

Join Our Newsletter!

Sign up below to receive email updates and see what's going on with our company

Explore Related Topics

AllKafkaSparkScyllaSStableKubernetesApiGithubGraphQl

Explore Further

cassandra