Uploaded image for project: 'Cassandra'
  1. Cassandra
  2. CASSANDRA-12510

Disallow decommission when number of replicas will drop below configured RF

Agile BoardAttach filesAttach ScreenshotBulk Copy AttachmentsBulk Move AttachmentsVotersWatch issueWatchersCreate sub-taskConvert to sub-taskMoveLinkCloneLabelsUpdate Comment AuthorReplace String in CommentUpdate Comment VisibilityDelete Comments



      Steps to replicate :

      • Create a 3 node cluster in DC1 and create a keyspace test_keyspace with table test_table with replication strategy NetworkTopologyStrategy , DC1=3 . Populate some data into this table.
      • Add 5 more nodes to this cluster, but in DC2. Also do not alter the keyspace to add the new DC2 to replication (this is intentional and the reason why the bug shows up). So the desc keyspace should still list NetworkTopologyStrategy with DC1=3 as RF
      • As expected, this will now be a 8 node cluster with 3 nodes in DC1 and 5 in DC2
      • Now start decommissioning the nodes in DC1. Note that the decommission runs fine on all the 3 nodes, but since the new nodes are in DC2 and the RF for keyspace is restricted to DC1, the new 5 nodes won't get any data.
      • You will now end with the 5 node cluster which has no data from the decommissioned 3 nodes and hence ending up in data loss

      I do understand that this problem could have been avoided if we perform an alter stmt and add DC2 replication before adding the 5 nodes. But the fact that decommission ran fine on the 3 nodes on DC1 without complaining that there were no nodes to stream its data seems a little discomforting.



          This comment will be Viewable by All Users Viewable by All Users


            KurtG Kurt Greaves Assign to me
            asood Atin Sood
            Kurt Greaves
            Paulo Motta
            0 Vote for this issue
            7 Start watching this issue




                Issue deployment