Switching snitches
Steps for switching snitches.
Because snitches determine how Cassandra distributes replicas, the procedure to switch snitches depends on whether or not the topology of the cluster will change:
- If data has not been inserted into the cluster, there is no change in the network topology. This means that you only need to set the snitch; no other steps are necessary.
- If data has been inserted into the cluster, it's possible that the topology has changed and you will need to perform additional steps.
- If data has been inserted into the cluster that must be kept, change the snitch without changing the topology. Then add a new datacenter with new nodes and racks as desired. Finally, remove nodes from the old datacenters and racks. Simply altering the snitch and replication to move some nodes to a new datacenter will result in data being replicated incorrectly.
A change in topology means that there is a change in the datacenters and/or racks where the nodes are placed. Topology changes may occur when the replicas are placed in different places by the new snitch. Specifically, the replication strategy places the replicas based on the information provided by the new snitch. The following examples demonstrate the differences:
No topology change
Change from: five nodes using the SimpleSnitch in a single datacenter
To: five nodes in one datacenter and 1 rack using a network snitch such as the GossipingPropertyFileSnitch
- Topology changes
-
Change from: 5 nodes using the SimpleSnitch in a single datacenter
To: 5 nodes in 2 datacenters using the PropertyFileSnitch (add a datacenter).Note: If "splitting" one datacenter into two, create a new datacenter with new nodes. Alter the keyspace replication settings for the keyspace that originally existed to reflect that two datacenters now exist. Once data is replicated to the new datacenter, remove the number of nodes from the original datacenter that have "moved" to the new datacenter. -
Change From: 5 nodes using the SimpleSnitch in a single datacenter
To: 5 nodes in 1 datacenter and 2 racks using the RackInferringSnitch (add rack information).
-
Procedure
-
Create a properties file with datacenter and rack information.
- cassandra-rackdc.properties
GossipingPropertyFileSnitch, Ec2Snitch, and Ec2MultiRegionSnitch only.
- cassandra-topology.properties
All other network snitches.
- cassandra-rackdc.properties
-
Copy the cassandra-rackdc.properties or
cassandra-topology.properties file to the
Cassandra configuration directory on all the cluster's nodes. They won't be used
until the new snitch is enabled.
The location of the cassandra-topology.properties file depends on the type of installation:
Package installations /etc/cassandra/cassandra-topology.properties Tarball installations install_location/conf/cassandra-topology.properties The location of the cassandra-rackdc.properties file depends on the type of installation:Package installations /etc/cassandra/cassandra-rackdc.properties Tarball installations install_location/conf/cassandra-rackdc.properties The location of the cassandra.yaml file depends on the type of installation:Package installations /etc/cassandra/cassandra.yaml Tarball installations install_location/resources/cassandra/conf/cassandra.yaml -
Change the snitch for each node in the cluster in the node's
cassandra.yaml file. For example:
endpoint_snitch: GossipingPropertyFileSnitch
-
If the topology has not changed, you can restart each node one at a time.
Any change in the cassandra.yaml file requires a node restart.
-
If the topology of the network has changed, but no datacenters are
added:
- Shut down all the nodes, then restart them.
- Run a sequential repair and nodetool cleanup on each node.
-
If the topology of the network has changed and a datacenter is added:
- Create a new datacenter.
- Replicate data into new datacenter. Remove nodes from old datacenter.
- Run a sequential repair and nodetool cleanup on each node.