def readrepair_test(self): cluster = self.cluster cluster.set_configuration_options(values={'hinted_handoff_enabled': False}) if DISABLE_VNODES: cluster.populate(2).start() else: tokens = cluster.balanced_tokens(2) cluster.populate(2, tokens=tokens).start() node1, node2 = cluster.nodelist() session = self.patient_cql_connection(node1) create_ks(session, 'ks', 2) create_c1c2_table(self, session, read_repair=1.0) node2.stop(wait_other_notice=True) insert_c1c2(session, n=10000, consistency=ConsistencyLevel.ONE) node2.start(wait_for_binary_proto=True, wait_other_notice=True) # query everything to cause RR for n in xrange(0, 10000): query_c1c2(session, n, ConsistencyLevel.QUORUM) node1.stop(wait_other_notice=True) # Check node2 for all the keys that should have been repaired session = self.patient_cql_connection(node2, keyspace='ks') for n in xrange(0, 10000): query_c1c2(session, n, ConsistencyLevel.ONE)
def readrepair_test(self): cluster = self.cluster cluster.set_configuration_options( values={'hinted_handoff_enabled': False}) if DISABLE_VNODES: cluster.populate(2).start() else: tokens = cluster.balanced_tokens(2) cluster.populate(2, tokens=tokens).start() node1, node2 = cluster.nodelist() session = self.patient_cql_connection(node1) self.create_ks(session, 'ks', 2) create_c1c2_table(self, session, read_repair=1.0) node2.stop(wait_other_notice=True) insert_c1c2(session, n=10000, consistency=ConsistencyLevel.ONE) node2.start(wait_other_notice=True) # query everything to cause RR for n in xrange(0, 10000): query_c1c2(session, n, ConsistencyLevel.QUORUM) node1.stop(wait_other_notice=True) # Check node2 for all the keys that should have been repaired session = self.patient_cql_connection(node2, keyspace='ks') for n in xrange(0, 10000): query_c1c2(session, n, ConsistencyLevel.ONE)
def _do_hinted_handoff(self, node1, node2, enabled, keyspace='ks'): """ Test that if we stop one node the other one will store hints only when hinted handoff is enabled """ session = self.patient_exclusive_cql_connection(node1) create_ks(session, keyspace, 2) create_c1c2_table(self, session) node2.stop(wait_other_notice=True) insert_c1c2(session, n=100, consistency=ConsistencyLevel.ONE) log_mark = node1.mark_log() node2.start(wait_other_notice=True) if enabled: node1.watch_log_for(["Finished hinted"], from_mark=log_mark, timeout=120) node1.stop(wait_other_notice=True) # Check node2 for all the keys that should have been delivered via HH if enabled or not if not enabled session = self.patient_exclusive_cql_connection(node2, keyspace=keyspace) for n in xrange(0, 100): if enabled: query_c1c2(session, n, ConsistencyLevel.ONE) else: query_c1c2(session, n, ConsistencyLevel.ONE, tolerate_missing=True, must_be_missing=True)
def test_blacklisted_directory(self): cluster = self.cluster cluster.set_datadir_count(3) cluster.populate(1) [node] = cluster.nodelist() cluster.start() session = self.patient_cql_connection(node) create_ks(session, 'ks', 1) create_c1c2_table(self, session) insert_c1c2(session, n=10000) node.flush() for k in range(0, 10000): query_c1c2(session, k) node.compact() mbean = make_mbean('db', type='BlacklistedDirectories') with JolokiaAgent(node) as jmx: jmx.execute_method(mbean, 'markUnwritable', [os.path.join(node.get_path(), 'data1')]) for k in range(0, 10000): query_c1c2(session, k) node.nodetool('relocatesstables') for k in range(0, 10000): query_c1c2(session, k)
def _do_hinted_handoff(self, node1, node2, enabled, keyspace='ks'): """ Test that if we stop one node the other one will store hints only when hinted handoff is enabled """ session = self.patient_exclusive_cql_connection(node1) create_ks(session, keyspace, 2) create_c1c2_table(self, session) node2.stop(wait_other_notice=True) insert_c1c2(session, n=100, consistency=ConsistencyLevel.ONE) log_mark = node1.mark_log() node2.start() if enabled: node1.watch_log_for(["Finished hinted"], from_mark=log_mark, timeout=120) node1.stop(wait_other_notice=True) # Check node2 for all the keys that should have been delivered via HH if enabled or not if not enabled session = self.patient_exclusive_cql_connection(node2, keyspace=keyspace) for n in range(0, 100): if enabled: query_c1c2(session, n, ConsistencyLevel.ONE) else: query_c1c2(session, n, ConsistencyLevel.ONE, tolerate_missing=True, must_be_missing=True)
def quorum_available_during_failure_test(self): CL = ConsistencyLevel.QUORUM RF = 3 debug("Creating a ring") cluster = self.cluster if DISABLE_VNODES: cluster.populate(3).start() else: tokens = cluster.balanced_tokens(3) cluster.populate(3, tokens=tokens).start() node1, node2, node3 = cluster.nodelist() debug("Set to talk to node 2") session = self.patient_cql_connection(node2) self.create_ks(session, 'ks', RF) create_c1c2_table(self, session) debug("Generating some data") insert_c1c2(session, n=100, consistency=CL) debug("Taking down node1") node1.stop(wait_other_notice=True) debug("Reading back data.") for n in xrange(100): query_c1c2(session, n, CL)
def quorum_available_during_failure_test(self): CL = ConsistencyLevel.QUORUM RF = 3 debug("Creating a ring") cluster = self.cluster if DISABLE_VNODES: cluster.populate(3).start() else: tokens = cluster.balanced_tokens(3) cluster.populate(3, tokens=tokens).start() node1, node2, node3 = cluster.nodelist() debug("Set to talk to node 2") session = self.patient_cql_connection(node2) create_ks(session, 'ks', RF) create_c1c2_table(self, session) debug("Generating some data") insert_c1c2(session, n=100, consistency=CL) debug("Taking down node1") node1.stop(wait_other_notice=True) debug("Reading back data.") for n in xrange(100): query_c1c2(session, n, CL)
def blacklisted_directory_test(self): cluster = self.cluster cluster.set_datadir_count(3) cluster.populate(1) [node] = cluster.nodelist() remove_perf_disable_shared_mem(node) cluster.start(wait_for_binary_proto=True) session = self.patient_cql_connection(node) create_ks(session, 'ks', 1) create_c1c2_table(self, session) insert_c1c2(session, n=10000) node.flush() for k in xrange(0, 10000): query_c1c2(session, k) node.compact() mbean = make_mbean('db', type='BlacklistedDirectories') with JolokiaAgent(node) as jmx: jmx.execute_method(mbean, 'markUnwritable', [os.path.join(node.get_path(), 'data0')]) for k in xrange(0, 10000): query_c1c2(session, k) node.nodetool('relocatesstables') for k in xrange(0, 10000): query_c1c2(session, k)
def test_consistent_reads_after_bootstrap(self): logger.debug("Creating a ring") cluster = self.cluster cluster.set_configuration_options( values={ 'hinted_handoff_enabled': False, 'write_request_timeout_in_ms': 60000, 'read_request_timeout_in_ms': 60000, 'dynamic_snitch_badness_threshold': 0.0 }) cluster.set_batch_commitlog(enabled=True) cluster.populate(2) node1, node2 = cluster.nodelist() cluster.start(wait_for_binary_proto=True, wait_other_notice=True) logger.debug("Set to talk to node 2") n2session = self.patient_cql_connection(node2) create_ks(n2session, 'ks', 2) create_c1c2_table(self, n2session) logger.debug("Generating some data for all nodes") insert_c1c2(n2session, keys=list(range(10, 20)), consistency=ConsistencyLevel.ALL) node1.flush() logger.debug("Taking down node1") node1.stop(wait_other_notice=True) logger.debug("Writing data to only node2") insert_c1c2(n2session, keys=list(range(30, 1000)), consistency=ConsistencyLevel.ONE) node2.flush() logger.debug("Restart node1") node1.start(wait_other_notice=True) logger.debug("Bootstraping node3") node3 = new_node(cluster) node3.start(wait_for_binary_proto=True) n3session = self.patient_cql_connection(node3) n3session.execute("USE ks") logger.debug("Checking that no data was lost") for n in range(10, 20): query_c1c2(n3session, n, ConsistencyLevel.ALL) for n in range(30, 1000): query_c1c2(n3session, n, ConsistencyLevel.ALL)
def test_parent_repair_session_cleanup(self): """ Calls incremental repair on 3 node cluster and verifies if all ParentRepairSession objects are cleaned @jira_ticket CASSANDRA-16446 """ self.cluster.populate(3).start() session = self.patient_cql_connection(self.cluster.nodelist()[0]) create_ks(session, 'ks', 2) create_c1c2_table(self, session) for node in self.cluster.nodelist(): node.repair(options=['ks']) assert_parent_repair_session_count(self.cluster.nodelist(), 0)
def test_non_local_read(self): """ This test reads from a coordinator we know has no copy of the data """ cluster = self.cluster cluster.populate(3).start() node1, node2, node3 = cluster.nodelist() session = self.patient_cql_connection(node1) create_ks(session, 'ks', 2) create_c1c2_table(self, session) # insert and get at CL.QUORUM (since RF=2, node1 won't have all key locally) insert_c1c2(session, n=1000, consistency=ConsistencyLevel.QUORUM) for n in range(0, 1000): query_c1c2(session, n, ConsistencyLevel.QUORUM)
def hintedhandoff_decom_test(self): self.cluster.populate(4).start(wait_for_binary_proto=True) [node1, node2, node3, node4] = self.cluster.nodelist() session = self.patient_cql_connection(node1) self.create_ks(session, 'ks', 2) create_c1c2_table(self, session) node4.stop(wait_other_notice=True) insert_c1c2(session, n=100, consistency=ConsistencyLevel.ONE) node1.decommission() node4.start(wait_for_binary_proto=True) node2.decommission() node3.decommission() time.sleep(5) for x in xrange(0, 100): query_c1c2(session, x, ConsistencyLevel.ONE)
def hintedhandoff_decom_test(self): self.cluster.populate(4).start(wait_for_binary_proto=True) [node1, node2, node3, node4] = self.cluster.nodelist() session = self.patient_cql_connection(node1) create_ks(session, 'ks', 2) create_c1c2_table(self, session) node4.stop(wait_other_notice=True) insert_c1c2(session, n=100, consistency=ConsistencyLevel.ONE) node1.decommission() node4.start(wait_for_binary_proto=True) node2.decommission() node3.decommission() time.sleep(5) for x in xrange(0, 100): query_c1c2(session, x, ConsistencyLevel.ONE)
def test_consistent_reads_after_move(self): logger.debug("Creating a ring") cluster = self.cluster cluster.set_configuration_options( values={ 'hinted_handoff_enabled': False, 'write_request_timeout_in_ms': 60000, 'read_request_timeout_in_ms': 60000, 'dynamic_snitch_badness_threshold': 0.0 }) cluster.set_batch_commitlog(enabled=True) cluster.populate(3, tokens=[0, 2**48, 2**62]).start() node1, node2, node3 = cluster.nodelist() logger.debug("Set to talk to node 2") n2session = self.patient_cql_connection(node2) create_ks(n2session, 'ks', 2) create_c1c2_table(self, n2session) logger.debug("Generating some data for all nodes") insert_c1c2(n2session, keys=list(range(10, 20)), consistency=ConsistencyLevel.ALL) node1.flush() logger.debug("Taking down node1") node1.stop(wait_other_notice=True) logger.debug("Writing data to node2") insert_c1c2(n2session, keys=list(range(30, 1000)), consistency=ConsistencyLevel.ONE) node2.flush() logger.debug("Restart node1") node1.start() logger.debug("Move token on node3") node3.move(2) logger.debug("Checking that no data was lost") for n in range(10, 20): query_c1c2(n2session, n, ConsistencyLevel.ALL, max_attempts=3) for n in range(30, 1000): query_c1c2(n2session, n, ConsistencyLevel.ALL, max_attempts=3)
def test_consistent_reads_after_bootstrap(self): logger.debug("Creating a ring") cluster = self.cluster cluster.set_configuration_options(values={'hinted_handoff_enabled': False, 'write_request_timeout_in_ms': 60000, 'read_request_timeout_in_ms': 60000, 'dynamic_snitch_badness_threshold': 0.0}) cluster.set_batch_commitlog(enabled=True) cluster.populate(2) node1, node2 = cluster.nodelist() cluster.start(wait_for_binary_proto=True, wait_other_notice=True) logger.debug("Set to talk to node 2") n2session = self.patient_cql_connection(node2) create_ks(n2session, 'ks', 2) create_c1c2_table(self, n2session) logger.debug("Generating some data for all nodes") insert_c1c2(n2session, keys=list(range(10, 20)), consistency=ConsistencyLevel.ALL) node1.flush() logger.debug("Taking down node1") node1.stop(wait_other_notice=True) logger.debug("Writing data to only node2") insert_c1c2(n2session, keys=list(range(30, 1000)), consistency=ConsistencyLevel.ONE) node2.flush() logger.debug("Restart node1") node1.start(wait_other_notice=True) logger.debug("Bootstraping node3") node3 = new_node(cluster) node3.start(wait_for_binary_proto=True) n3session = self.patient_cql_connection(node3) n3session.execute("USE ks") logger.debug("Checking that no data was lost") for n in range(10, 20): query_c1c2(n3session, n, ConsistencyLevel.ALL) for n in range(30, 1000): query_c1c2(n3session, n, ConsistencyLevel.ALL)
def consistent_reads_after_move_test(self): debug("Creating a ring") cluster = self.cluster cluster.set_configuration_options(values={'hinted_handoff_enabled': False, 'write_request_timeout_in_ms': 60000, 'read_request_timeout_in_ms': 60000, 'dynamic_snitch_badness_threshold': 0.0}) cluster.set_batch_commitlog(enabled=True) cluster.populate(3, tokens=[0, 2**48, 2**62]).start() node1, node2, node3 = cluster.nodelist() debug("Set to talk to node 2") n2session = self.patient_cql_connection(node2) create_ks(n2session, 'ks', 2) create_c1c2_table(self, n2session) debug("Generating some data for all nodes") insert_c1c2(n2session, keys=range(10, 20), consistency=ConsistencyLevel.ALL) node1.flush() debug("Taking down node1") node1.stop(wait_other_notice=True) debug("Writing data to node2") insert_c1c2(n2session, keys=range(30, 1000), consistency=ConsistencyLevel.ONE) node2.flush() debug("Restart node1") node1.start(wait_other_notice=True) debug("Move token on node3") node3.move(2) debug("Checking that no data was lost") for n in xrange(10, 20): query_c1c2(n2session, n, ConsistencyLevel.ALL) for n in xrange(30, 1000): query_c1c2(n2session, n, ConsistencyLevel.ALL)
def test_bloom_filter_false_ratio(self): """ Test for CASSANDRA-15834 Verifies if BloomFilterFalseRatio takes into account true negatives. Without this fix, the following scenario (many reads for non-existing rows) would yield BloomFilterFalseRatio=1.0. With the fix we assume it should be less then the default bloom_filter_fp_chance. """ cluster = self.cluster cluster.populate(1) node = cluster.nodelist()[0] cluster.start(wait_for_binary_proto=True) session = self.patient_exclusive_cql_connection(node) keyspace = 'bloom_ratio_test_ks' create_ks(session, keyspace, 1) create_c1c2_table(self, session) insert_c1c2(session, n=10) node.nodetool("flush " + keyspace) for key in range(10000): session.execute("SELECT * from cf where key = '{0}'".format(key)) bloom_filter_false_ratios = [ make_mbean('metrics', type='Table', name='RecentBloomFilterFalseRatio'), make_mbean('metrics', type='Table', keyspace=keyspace, scope='cf', name='BloomFilterFalseRatio'), make_mbean('metrics', type='Table', name='BloomFilterFalseRatio'), make_mbean('metrics', type='Table', keyspace=keyspace, scope='cf', name='RecentBloomFilterFalseRatio'), ] with JolokiaAgent(node) as jmx: for metric in bloom_filter_false_ratios: ratio = jmx.read_attribute(metric, "Value") # Bloom filter false positive ratio should not be greater than the default bloom_filter_fp_chance. assert ratio < 0.01
def test_hintedhandoff_window(self): """ Test that we only store at a maximum the hint window worth of hints. Prior to CASSANDRA-14309 we would store another window worth of hints if the down node was brought up and then taken back down immediately. We would also store another window of hints on a live node if the live node was restarted. @jira_ticket CASSANDRA-14309 """ # hint_window_persistent_enabled is set to true by default self.cluster.set_configuration_options({ 'max_hint_window_in_ms': 5000, 'hinted_handoff_enabled': True, 'max_hints_delivery_threads': 1, 'hints_flush_period_in_ms': 100, }) self.cluster.populate(2).start() node1, node2 = self.cluster.nodelist() session = self.patient_cql_connection(node1) create_ks(session, 'ks', 2) create_c1c2_table(self, session) # Stop handoff until very end and take node2 down for first round of hints node1.nodetool('pausehandoff') node2.nodetool('disablebinary') node2.nodetool('disablegossip') # First round of hints. We expect these to be replayed and the only # hints within the window insert_c1c2(session, n=(0, 100), consistency=ConsistencyLevel.ONE) # Let hint window pass time.sleep(10) # Re-enable and disable the node. Prior to CASSANDRA-14215 this should make the hint window on node1 reset. node2.nodetool('enablegossip') node2.nodetool('disablegossip') # Second round of inserts. We do not expect hints to be stored. insert_c1c2(session, n=(100, 200), consistency=ConsistencyLevel.ONE) # Restart node1. Prior to CASSANDRA-14215 this would reset node1's hint window. node1.stop() node1.start(wait_for_binary_proto=True, wait_other_notice=False) session = self.patient_exclusive_cql_connection(node1) session.execute('USE ks') # Third round of inserts. We do not expect hints to be stored. insert_c1c2(session, n=(200, 300), consistency=ConsistencyLevel.ONE) # Enable node2 and wait for hints to be replayed node2.nodetool('enablegossip') node2.nodetool('enablebinary') node1.nodetool('resumehandoff') node1.watch_log_for('Finished hinted handoff') # Stop node1 so that we only query node2 node1.stop() session = self.patient_exclusive_cql_connection(node2) session.execute('USE ks') # Ensure first dataset is present (through replayed hints) for x in range(0, 100): query_c1c2(session, x, ConsistencyLevel.ONE) # Ensure second and third datasets are not present for x in range(100, 300): query_c1c2(session, x, ConsistencyLevel.ONE, tolerate_missing=True, must_be_missing=True)