You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@cassandra.apache.org by mc...@apache.org on 2020/09/23 12:13:38 UTC

[cassandra-dtest] branch master updated: remove redundant params wait_for_binary_proto=True and wait_other_notice=True from Cluster.start calls

This is an automated email from the ASF dual-hosted git repository.

mck pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/cassandra-dtest.git


The following commit(s) were added to refs/heads/master by this push:
     new 4acae58  remove redundant params wait_for_binary_proto=True and wait_other_notice=True from Cluster.start calls
4acae58 is described below

commit 4acae58e0c0bbe665f5a1c38aa24a7b4f4ad06d1
Author: Christopher Lambert <Ch...@Datastax.com>
AuthorDate: Wed Sep 23 11:01:24 2020 +0200

    remove redundant params wait_for_binary_proto=True and wait_other_notice=True from Cluster.start calls
    
    Since https://github.com/riptano/ccm/pull/561 ccm Cluster.start() defaults to:
    wait_for_binary_proto=True
    wait_other_notice=True
    
    Since their presence could suggest that these are non-default values, we
    clean up the code by removing them.
    
    patch by Christopher Lambert; reviewed by Mick Semb Wever
---
 README.md                                |  1 -
 auditlog_test.py                         | 10 ++---
 auth_join_ring_false_test.py             |  2 +-
 auth_test.py                             | 14 +++---
 batch_test.py                            |  2 +-
 bootstrap_test.py                        | 28 ++++++------
 cdc_test.py                              |  2 +-
 cfid_test.py                             |  2 +-
 compaction_test.py                       | 30 ++++++-------
 compression_test.py                      |  8 ++--
 configuration_test.py                    |  6 +--
 consistency_test.py                      | 28 ++++++------
 consistent_bootstrap_test.py             |  2 +-
 counter_test.py                          |  2 +-
 cql_test.py                              | 16 +++----
 cql_tracing_test.py                      |  2 +-
 cqlsh_tests/test_cqlsh.py                | 73 ++++++++++++++++----------------
 cqlsh_tests/test_cqlsh_copy.py           |  2 +-
 delete_insert_test.py                    |  2 +-
 deletion_test.py                         |  2 +-
 disk_balance_test.py                     | 16 +++----
 fqltool_test.py                          |  6 +--
 hintedhandoff_test.py                    |  2 +-
 jmx_test.py                              | 10 ++---
 legacy_sstables_test.py                  |  2 +-
 materialized_views_test.py               |  6 +--
 metadata_test.py                         |  2 +-
 nodetool_test.py                         |  6 +--
 offline_tools_test.py                    | 24 +++++------
 paging_test.py                           |  2 +-
 paxos_test.py                            |  8 ++--
 pending_range_test.py                    |  2 +-
 pushed_notifications_test.py             | 12 +++---
 read_failures_test.py                    |  2 +-
 read_repair_test.py                      |  9 ++--
 rebuild_test.py                          |  6 +--
 refresh_test.py                          |  2 +-
 repair_tests/deprecated_repair_test.py   |  2 +-
 repair_tests/incremental_repair_test.py  |  4 +-
 repair_tests/repair_test.py              | 42 +++++++++---------
 replication_test.py                      | 10 ++---
 secondary_indexes_test.py                |  6 +--
 snapshot_test.py                         |  4 +-
 snitch_test.py                           |  2 +-
 sstable_generation_loading_test.py       |  4 +-
 sstablesplit_test.py                     |  4 +-
 sstableutil_test.py                      |  4 +-
 streaming_test.py                        |  2 +-
 stress_tool_test.py                      |  4 +-
 thrift_hsha_test.py                      |  2 +-
 thrift_test.py                           |  3 +-
 token_generator_test.py                  |  2 +-
 topology_test.py                         | 12 +++---
 transient_replication_ring_test.py       |  2 +-
 transient_replication_test.py            |  2 +-
 ttl_test.py                              |  6 +--
 upgrade_tests/compatibility_flag_test.py |  4 +-
 upgrade_tests/cql_tests.py               |  2 +-
 upgrade_tests/upgrade_base.py            |  2 +-
 user_types_test.py                       |  2 +-
 write_failures_test.py                   |  2 +-
 61 files changed, 236 insertions(+), 244 deletions(-)

diff --git a/README.md b/README.md
index a6be874..4e64587 100644
--- a/README.md
+++ b/README.md
@@ -101,7 +101,6 @@ JAVA7_HOME and JAVA8_HOME, respectively.
 Writing Tests
 -------------
 
-- Most of the time when you start a cluster with `cluster.start()`, you'll want to pass in `wait_for_binary_proto=True` so the call blocks until the cluster is ready to accept CQL connections. We tried setting this to `True` by default once, but the problems caused there (e.g. when it waited the full timeout time on a node that was deliberately down) were more unpleasant and more difficult to debug than the problems caused by having it `False` by default.
 - If you're using JMX via [the `tools.jmxutils` module](tools/jmxutils.py), make sure to call `remove_perf_disable_shared_mem` on the node or nodes you want to query with JMX _before starting the nodes_. `remove_perf_disable_shared_mem` disables a JVM option that's incompatible with JMX (see [this JMX ticket](https://github.com/rhuss/jolokia/issues/198)). It works by performing a string replacement in the node's Cassandra startup script, so changes will only propagate to the node at star [...]
 
 If you'd like to know what to expect during a code review, please see the included [CONTRIBUTING file](CONTRIBUTING.md).
diff --git a/auditlog_test.py b/auditlog_test.py
index 7a57604..c7a6d66 100644
--- a/auditlog_test.py
+++ b/auditlog_test.py
@@ -24,7 +24,7 @@ class TestAuditlog(Tester):
                                                                             'audit_logs_dir': log_dir,
                                                                             'roll_cycle': 'TEST_SECONDLY',
                                                                             'archive_command':'%s %%path'%(move_script)}})
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         node = cluster.nodelist()[0]
         node.stress(['write', 'n=100k', "no-warmup", "cl=ONE", "-rate", "threads=300"])
         node.nodetool("disableauditlog")
@@ -38,7 +38,7 @@ class TestAuditlog(Tester):
         moved_log_dir, move_script = self._create_script()
         cluster.set_configuration_options(values={'full_query_logging_options': {'log_dir': log_dir,
                                                                                  'archive_command': 'conf should not be used'}})
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         node = cluster.nodelist()[0]
         node.nodetool("enablefullquerylog --archive-command \"%s %%path\" --roll-cycle=TEST_SECONDLY"%move_script)
         node.stress(['write', 'n=100k', "no-warmup", "cl=ONE", "-rate", "threads=300"])
@@ -58,7 +58,7 @@ class TestAuditlog(Tester):
         cluster.set_configuration_options(values={'full_query_logging_options': {'log_dir': log_dir,
                                                                                  'roll_cycle': 'TEST_SECONDLY',
                                                                                  'archive_command':'%s %%path'%(move_script)}})
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         node = cluster.nodelist()[0]
         node.nodetool("enablefullquerylog")
         node.stress(['write', 'n=100k', "no-warmup", "cl=ONE", "-rate", "threads=300"])
@@ -87,7 +87,7 @@ class TestAuditlog(Tester):
         cluster.set_configuration_options(values={'full_query_logging_options': {'log_dir': log_dir,
                                                                                  'roll_cycle': 'TEST_SECONDLY',
                                                                                  'archive_command':'%s %%path'%(move_script)}})
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         node = cluster.nodelist()[0]
         node.nodetool("enablefullquerylog")
 
@@ -104,7 +104,7 @@ class TestAuditlog(Tester):
         cluster.set_configuration_options(values={'full_query_logging_options': {'log_dir': log_dir,
                                                                                  'roll_cycle': 'TEST_SECONDLY',
                                                                                  'archive_command':'%s %%path'%(move_script)}})
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         node = cluster.nodelist()[0]
         node.nodetool("enablefullquerylog")
 
diff --git a/auth_join_ring_false_test.py b/auth_join_ring_false_test.py
index 73e29e4..35c3fce 100644
--- a/auth_join_ring_false_test.py
+++ b/auth_join_ring_false_test.py
@@ -208,7 +208,7 @@ class TestAuth(Tester):
                   'permissions_validity_in_ms': 0,
                   'roles_validity_in_ms': roles_expiry}
         self.cluster.set_configuration_options(values=config)
-        self.cluster.populate(nodes).start(wait_for_binary_proto=True)
+        self.cluster.populate(nodes).start()
 
         self.cluster.wait_for_any_log('Created default superuser', 25)
 
diff --git a/auth_test.py b/auth_test.py
index 3c3bd26..4f12175 100644
--- a/auth_test.py
+++ b/auth_test.py
@@ -76,7 +76,7 @@ class TestAuth(Tester):
         logger.debug("Stopping cluster..")
         self.cluster.stop()
         logger.debug("Restarting cluster..")
-        self.cluster.start(wait_other_notice=True)
+        self.cluster.start()
 
         # check each node directly
         for i in range(3):
@@ -1026,7 +1026,7 @@ class TestAuth(Tester):
         if self.dtest_config.cassandra_version_from_build >= '4.0':
             config['network_authorizer'] = 'org.apache.cassandra.auth.AllowAllNetworkAuthorizer'
         self.cluster.set_configuration_options(values=config)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
 
         self.cluster.stop()
         config = {'authenticator': 'org.apache.cassandra.auth.PasswordAuthenticator',
@@ -1034,7 +1034,7 @@ class TestAuth(Tester):
         if self.dtest_config.cassandra_version_from_build >= '4.0':
             config['network_authorizer'] = 'org.apache.cassandra.auth.CassandraNetworkAuthorizer'
         self.cluster.set_configuration_options(values=config)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
 
         philip = self.get_session(user='philip', password='strongpass')
         cathy = self.get_session(user='cathy', password='12345')
@@ -1057,7 +1057,7 @@ class TestAuth(Tester):
         cluster.populate(1)
         [node] = cluster.nodelist()
         remove_perf_disable_shared_mem(node)
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
 
         with JolokiaAgent(node) as jmx:
             success = jmx.read_attribute(
@@ -1173,7 +1173,7 @@ class TestAuthRoles(Tester):
                 'roles_validity_in_ms': 0,
                 'num_tokens': 1
             })
-        fixture_dtest_setup.cluster.populate(1, debug=True).start(wait_for_binary_proto=True, jvm_args=['-XX:-PerfDisableSharedMem'])
+        fixture_dtest_setup.cluster.populate(1, debug=True).start(jvm_args=['-XX:-PerfDisableSharedMem'])
         nodes = fixture_dtest_setup.cluster.nodelist()
         fixture_dtest_setup.superuser = fixture_dtest_setup.patient_exclusive_cql_connection(nodes[0], user='cassandra', password='cassandra')
 
@@ -2695,7 +2695,7 @@ class TestAuthRoles(Tester):
             config['network_authorizer'] = 'org.apache.cassandra.auth.CassandraNetworkAuthorizer'
 
         self.cluster.set_configuration_options(values=config)
-        self.cluster.populate(nodes).start(wait_for_binary_proto=True)
+        self.cluster.populate(nodes).start()
 
         self.cluster.wait_for_any_log('Created default superuser', 25)
 
@@ -3040,7 +3040,7 @@ class TestNetworkAuth(Tester):
             'network_authorizer': 'org.apache.cassandra.auth.CassandraNetworkAuthorizer',
             'num_tokens': 1
         })
-        fixture_dtest_setup.cluster.populate([1, 1], debug=True).start(wait_for_binary_proto=True, jvm_args=['-XX:-PerfDisableSharedMem'])
+        fixture_dtest_setup.cluster.populate([1, 1], debug=True).start(jvm_args=['-XX:-PerfDisableSharedMem'])
         fixture_dtest_setup.dc1_node, fixture_dtest_setup.dc2_node = fixture_dtest_setup.cluster.nodelist()
         fixture_dtest_setup.superuser = fixture_dtest_setup.patient_exclusive_cql_connection(fixture_dtest_setup.dc1_node, user='cassandra', password='cassandra')
 
diff --git a/batch_test.py b/batch_test.py
index 586d888..f67a8d7 100644
--- a/batch_test.py
+++ b/batch_test.py
@@ -442,7 +442,7 @@ class TestBatch(Tester):
         for n in self.cluster.nodelist():
             remove_perf_disable_shared_mem(n)
 
-        self.cluster.start(wait_other_notice=True)
+        self.cluster.start()
 
         node1 = self.cluster.nodelist()[0]
         session = self.patient_cql_connection(node1, protocol_version=protocol_version)
diff --git a/bootstrap_test.py b/bootstrap_test.py
index e4cefba..540aa5c 100644
--- a/bootstrap_test.py
+++ b/bootstrap_test.py
@@ -75,7 +75,7 @@ class TestBootstrap(Tester):
             logger.debug("starting source node on version {}".format(bootstrap_from_version))
             node1.set_install_dir(version=bootstrap_from_version)
         node1.set_configuration_options(values={'initial_token': tokens[0]})
-        cluster.start(wait_other_notice=True)
+        cluster.start()
 
         session = self.patient_cql_connection(node1)
         create_ks(session, 'ks', 1)
@@ -177,7 +177,7 @@ class TestBootstrap(Tester):
         node1.byteman_port = '8100'
         node1.import_config_files()
 
-        cluster.start(wait_other_notice=True)
+        cluster.start()
 
         # Create more than one sstable larger than 1MB
         node1.stress(['write', 'n=1K', '-rate', 'threads=8', '-schema',
@@ -208,7 +208,7 @@ class TestBootstrap(Tester):
         cluster = self.cluster
         # Create a two-node cluster
         cluster.populate(2)
-        cluster.start(wait_other_notice=True)
+        cluster.start()
 
         # Bootstrapping a new node
         node3 = new_node(cluster)
@@ -348,7 +348,7 @@ class TestBootstrap(Tester):
         node1.byteman_port = '8100'
         node1.import_config_files()
 
-        cluster.start(wait_other_notice=True)
+        cluster.start()
         # kill stream to node3 in the middle of streaming to let it fail
         if cluster.version() < '4.0':
             node1.byteman_submit([self.byteman_submit_path_pre_4_0])
@@ -384,7 +384,7 @@ class TestBootstrap(Tester):
         """Test bootstrap with resetting bootstrap progress"""
         cluster = self.cluster
         cluster.set_configuration_options(values={'stream_throughput_outbound_megabits_per_sec': 1})
-        cluster.populate(2).start(wait_other_notice=True)
+        cluster.populate(2).start()
 
         node1 = cluster.nodes['node1']
         node1.stress(['write', 'n=100K', '-schema', 'replication(factor=2)'])
@@ -422,7 +422,7 @@ class TestBootstrap(Tester):
             @jira_ticket CASSANDRA-9022
         """
         cluster = self.cluster
-        cluster.populate(2).start(wait_other_notice=True)
+        cluster.populate(2).start()
         (node1, node2) = cluster.nodelist()
 
         node1.stress(['write', 'n=1K', 'no-warmup', '-schema', 'replication(factor=2)',
@@ -515,7 +515,7 @@ class TestBootstrap(Tester):
         """
         cluster = self.cluster
         cluster.populate(3)
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
 
         stress_table = 'keyspace1.standard1'
 
@@ -548,7 +548,7 @@ class TestBootstrap(Tester):
         """
         cluster = self.cluster
         cluster.populate(3)
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
 
         stress_table = 'keyspace1.standard1'
 
@@ -584,7 +584,7 @@ class TestBootstrap(Tester):
         """
         cluster = self.cluster
         cluster.populate(1)
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
 
         node1 = cluster.nodelist()[0]
         # Add a new node, bootstrap=True ensures that it is not a seed
@@ -627,7 +627,7 @@ class TestBootstrap(Tester):
         cluster = self.cluster
         cluster.populate(1)
         cluster.set_configuration_options(values={'stream_throughput_outbound_megabits_per_sec': 1})
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
 
         stress_table = 'keyspace1.standard1'
 
@@ -672,7 +672,7 @@ class TestBootstrap(Tester):
                                }]
         })
 
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
 
         node1 = cluster.nodelist()[0]
         node2 = cluster.nodelist()[1]
@@ -758,7 +758,7 @@ class TestBootstrap(Tester):
 
         cluster = self.cluster
         cluster.populate(1)
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
 
         node1, = cluster.nodelist()
 
@@ -795,7 +795,7 @@ class TestBootstrap(Tester):
         cluster = self.cluster
         cluster.set_configuration_options(values={'concurrent_compactors': 4})
         cluster.populate(1)
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
         node1, = cluster.nodelist()
         for x in range(0, 5):
             node1.stress(['write', 'n=100k', 'no-warmup', '-schema', 'compaction(strategy=SizeTieredCompactionStrategy,enabled=false)', 'replication(factor=1)', '-rate', 'threads=10'])
@@ -854,7 +854,7 @@ class TestBootstrap(Tester):
         node1.byteman_port = '8100'
         node1.import_config_files()
 
-        cluster.start(wait_other_notice=True)
+        cluster.start()
         # kill stream to node2 in the middle of streaming to let it fail
         if cluster.version() < '4.0':
             node1.byteman_submit([self.byteman_submit_path_pre_4_0])
diff --git a/cdc_test.py b/cdc_test.py
index a8efad6..53ae0b9 100644
--- a/cdc_test.py
+++ b/cdc_test.py
@@ -265,7 +265,7 @@ class TestCDC(Tester):
             configuration_overrides = {}
         self.cluster.populate(1)
         self.cluster.set_configuration_options(dict(config_defaults, **configuration_overrides))
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
         node = self.cluster.nodelist()[0]
         session = self.patient_cql_connection(node)
         create_ks(session, ks_name, rf=1)
diff --git a/cfid_test.py b/cfid_test.py
index 87625e0..3ad53e8 100644
--- a/cfid_test.py
+++ b/cfid_test.py
@@ -14,7 +14,7 @@ class TestCFID(Tester):
         """
         cluster = self.cluster
 
-        cluster.populate(1).start(wait_other_notice=True)
+        cluster.populate(1).start()
         [node1] = cluster.nodelist()
 
         session = self.patient_cql_connection(node1)
diff --git a/compaction_test.py b/compaction_test.py
index 49a2923..d84b5dc 100644
--- a/compaction_test.py
+++ b/compaction_test.py
@@ -33,7 +33,7 @@ class TestCompaction(Tester):
         Insert data, delete a partition of data and check that the requesite rows are tombstoned.
         """
         cluster = self.cluster
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         [node1] = cluster.nodelist()
 
         session = self.patient_cql_connection(node1)
@@ -70,7 +70,7 @@ class TestCompaction(Tester):
         Insert data and check data size before and after a compaction.
         """
         cluster = self.cluster
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         [node1] = cluster.nodelist()
 
         stress_write(node1)
@@ -119,7 +119,7 @@ class TestCompaction(Tester):
             min_bf_size = 100000
             max_bf_size = 150000
         cluster = self.cluster
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         [node1] = cluster.nodelist()
 
         for x in range(0, 5):
@@ -165,7 +165,7 @@ class TestCompaction(Tester):
         """
         self.skip_if_no_major_compaction(strategy)
         cluster = self.cluster
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         [node1] = cluster.nodelist()
         session = self.patient_cql_connection(node1)
         create_ks(session, 'ks', 1)
@@ -203,7 +203,7 @@ class TestCompaction(Tester):
             pytest.skip('Not implemented unless DateTieredCompactionStrategy is used')
 
         cluster = self.cluster
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         [node1] = cluster.nodelist()
         session = self.patient_cql_connection(node1)
         create_ks(session, 'ks', 1)
@@ -245,7 +245,7 @@ class TestCompaction(Tester):
         Set throughput, insert data and ensure compaction performance corresponds.
         """
         cluster = self.cluster
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         [node1] = cluster.nodelist()
 
         # disableautocompaction only disables compaction for existing tables,
@@ -303,7 +303,7 @@ class TestCompaction(Tester):
         if strategy in strategies:
             strategies.remove(strategy)
             cluster = self.cluster
-            cluster.populate(1).start(wait_for_binary_proto=True)
+            cluster.populate(1).start()
             [node1] = cluster.nodelist()
 
             for strat in strategies:
@@ -331,7 +331,7 @@ class TestCompaction(Tester):
                 node1.flush()
                 cluster.clear()
                 time.sleep(5)
-                cluster.start(wait_for_binary_proto=True)
+                cluster.start()
 
     def test_large_compaction_warning(self):
         """
@@ -340,7 +340,7 @@ class TestCompaction(Tester):
         """
         cluster = self.cluster
         cluster.set_configuration_options({'compaction_large_partition_warning_threshold_mb': 1})
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         [node] = cluster.nodelist()
 
         session = self.patient_cql_connection(node)
@@ -373,7 +373,7 @@ class TestCompaction(Tester):
         Make sure we can enable/disable compaction using nodetool
         """
         cluster = self.cluster
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         [node] = cluster.nodelist()
         session = self.patient_cql_connection(node)
         create_ks(session, 'ks', 1)
@@ -399,7 +399,7 @@ class TestCompaction(Tester):
         Make sure we can disable compaction via the schema compaction parameter 'enabled' = false
         """
         cluster = self.cluster
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         [node] = cluster.nodelist()
         session = self.patient_cql_connection(node)
         create_ks(session, 'ks', 1)
@@ -433,7 +433,7 @@ class TestCompaction(Tester):
         Make sure we can enable compaction using an alter-statement
         """
         cluster = self.cluster
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         [node] = cluster.nodelist()
         session = self.patient_cql_connection(node)
         create_ks(session, 'ks', 1)
@@ -462,7 +462,7 @@ class TestCompaction(Tester):
         Make sure compaction stays disabled after an alter statement where we have disabled using nodetool first
         """
         cluster = self.cluster
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         [node] = cluster.nodelist()
         session = self.patient_cql_connection(node)
         create_ks(session, 'ks', 1)
@@ -498,7 +498,7 @@ class TestCompaction(Tester):
         @jira_ticket CASSANDRA-11765
         """
         cluster = self.cluster
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         [node1] = cluster.nodelist()
 
         # disableautocompaction only disables compaction for existing tables,
@@ -527,7 +527,7 @@ class TestCompaction(Tester):
             pytest.skip('Not implemented unless LeveledCompactionStrategy is used')
 
         cluster = self.cluster
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         [node1] = cluster.nodelist()
 
         stress_write(node1, keycount=1)
diff --git a/compression_test.py b/compression_test.py
index 4688b72..0bb6344 100644
--- a/compression_test.py
+++ b/compression_test.py
@@ -29,7 +29,7 @@ class TestCompression(TestHelper):
         using new cql create table syntax to disable compression
         """
         cluster = self.cluster
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         [node] = cluster.nodelist()
 
         session = self.patient_cql_connection(node)
@@ -59,7 +59,7 @@ class TestCompression(TestHelper):
         using new cql create table syntax to configure compression
         """
         cluster = self.cluster
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         [node] = cluster.nodelist()
 
         session = self.patient_cql_connection(node)
@@ -124,7 +124,7 @@ class TestCompression(TestHelper):
         starting with compression enabled then disabling it
         """
         cluster = self.cluster
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         [node] = cluster.nodelist()
 
         session = self.patient_cql_connection(node)
@@ -155,7 +155,7 @@ class TestCompression(TestHelper):
         starting with compression disabled and enabling it
         """
         cluster = self.cluster
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         [node] = cluster.nodelist()
 
         session = self.patient_cql_connection(node)
diff --git a/configuration_test.py b/configuration_test.py
index 4a57986..5ae98f1 100644
--- a/configuration_test.py
+++ b/configuration_test.py
@@ -75,7 +75,7 @@ class TestConfiguration(Tester):
             # disable JVM option so we can use Jolokia
             # this has to happen after .set_configuration_options because of implementation details
             remove_perf_disable_shared_mem(node)
-            self.fixture_dtest_setup.cluster.start(wait_for_binary_proto=True)
+            self.fixture_dtest_setup.cluster.start()
             return node
 
         durable_node = new_commitlog_cluster_node()
@@ -117,7 +117,7 @@ class TestConfiguration(Tester):
         default_path = node1.data_directories()[0]
         node1.set_configuration_options({'saved_caches_directory': os.path.join(default_path, 'saved_caches')})
         remove_perf_disable_shared_mem(node1)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
 
         session = self.patient_exclusive_cql_connection(node1)
         session.execute("CREATE KEYSPACE ks WITH REPLICATION = {'class': 'SimpleStrategy', 'replication_factor': 1}")
@@ -130,7 +130,7 @@ class TestConfiguration(Tester):
             jmx.execute_method(cache_service, 'saveCaches')
 
         self.cluster.stop()
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
 
     def _check_chunk_length(self, session, value):
         result = session.cluster.metadata.keyspaces['ks'].tables['test_table'].as_cql_query()
diff --git a/consistency_test.py b/consistency_test.py
index fc63c9e..af4a84d 100644
--- a/consistency_test.py
+++ b/consistency_test.py
@@ -132,7 +132,7 @@ class TestHelper(Tester):
             cluster.set_configuration_options(values={
                 'endpoint_snitch': 'org.apache.cassandra.locator.PropertyFileSnitch'})
 
-        cluster.start(wait_for_binary_proto=True, wait_other_notice=True)
+        cluster.start()
 
         self.ksname = 'mytestks'
         session = self.patient_exclusive_cql_connection(cluster.nodelist()[0])
@@ -784,7 +784,7 @@ class TestConsistency(Tester):
         # set column_index_size_in_kb to 1 for a slightly easier reproduction sequence
         cluster.set_configuration_options(values={'column_index_size_in_kb': 1})
 
-        cluster.populate(1).start(wait_other_notice=True)
+        cluster.populate(1).start()
         node1 = cluster.nodelist()[0]
 
         session = self.patient_cql_connection(node1)
@@ -842,7 +842,7 @@ class TestConsistency(Tester):
         cluster.set_configuration_options(values={'column_index_size_in_kb': 1, 'hinted_handoff_enabled': False})
         cluster.set_batch_commitlog(enabled=True)
 
-        cluster.populate(3).start(wait_other_notice=True)
+        cluster.populate(3).start()
         node1, node2, node3 = cluster.nodelist()
 
         session = self.patient_exclusive_cql_connection(node1)
@@ -916,7 +916,7 @@ class TestConsistency(Tester):
         cluster.set_configuration_options(values={'hinted_handoff_enabled': False})
         cluster.set_batch_commitlog(enabled=True)
 
-        cluster.populate(2).start(wait_other_notice=True)
+        cluster.populate(2).start()
         node1, node2 = cluster.nodelist()
 
         session = self.patient_cql_connection(node2)
@@ -965,7 +965,7 @@ class TestConsistency(Tester):
         cluster.set_configuration_options(values={'hinted_handoff_enabled': False})
         cluster.set_batch_commitlog(enabled=True)
 
-        cluster.populate(2).start(wait_other_notice=True)
+        cluster.populate(2).start()
         node1, node2 = cluster.nodelist()
 
         session = self.patient_cql_connection(node1)
@@ -1018,7 +1018,7 @@ class TestConsistency(Tester):
         cluster.set_configuration_options(values={'hinted_handoff_enabled': False})
         cluster.set_batch_commitlog(enabled=True)
 
-        cluster.populate(2).start(wait_other_notice=True)
+        cluster.populate(2).start()
         node1, node2 = cluster.nodelist()
 
         session = self.patient_cql_connection(node1)
@@ -1081,7 +1081,7 @@ class TestConsistency(Tester):
         cluster.set_configuration_options(values={'hinted_handoff_enabled': False})
         cluster.set_batch_commitlog(enabled=True)
 
-        cluster.populate(2).start(wait_other_notice=True)
+        cluster.populate(2).start()
         node1, node2 = cluster.nodelist()
 
         session = self.patient_cql_connection(node1)
@@ -1141,7 +1141,7 @@ class TestConsistency(Tester):
         cluster.set_configuration_options(values={'hinted_handoff_enabled': False})
         cluster.set_batch_commitlog(enabled=True)
 
-        cluster.populate(2).start(wait_other_notice=True)
+        cluster.populate(2).start()
         node1, node2 = cluster.nodelist()
 
         session = self.patient_cql_connection(node1)
@@ -1176,7 +1176,7 @@ class TestConsistency(Tester):
         cluster.set_configuration_options(values={'hinted_handoff_enabled': False})
         cluster.set_batch_commitlog(enabled=True)
 
-        cluster.populate(2).start(wait_other_notice=True)
+        cluster.populate(2).start()
         node1, node2 = cluster.nodelist()
 
         session = self.patient_cql_connection(node1)
@@ -1232,7 +1232,7 @@ class TestConsistency(Tester):
         cluster.populate(2)
         node1, node2 = cluster.nodelist()
         remove_perf_disable_shared_mem(node1)  # necessary for jmx
-        cluster.start(wait_other_notice=True)
+        cluster.start()
 
         session = self.patient_cql_connection(node1)
 
@@ -1298,7 +1298,7 @@ class TestConsistency(Tester):
         cluster.set_configuration_options(values={'hinted_handoff_enabled': False})
         cluster.set_batch_commitlog(enabled=True)
 
-        cluster.populate(2).start(wait_other_notice=True)
+        cluster.populate(2).start()
         node1, node2 = cluster.nodelist()
 
         session = self.patient_cql_connection(node1)
@@ -1370,7 +1370,7 @@ class TestConsistency(Tester):
         cluster.set_configuration_options(values={'hinted_handoff_enabled': False})
         cluster.set_batch_commitlog(enabled=True)
 
-        cluster.populate(3).start(wait_other_notice=True)
+        cluster.populate(3).start()
         node1, node2, node3 = cluster.nodelist()
 
         session = self.patient_cql_connection(node1)
@@ -1432,7 +1432,7 @@ class TestConsistency(Tester):
         cluster.set_configuration_options(values={'hinted_handoff_enabled': False})
         cluster.set_batch_commitlog(enabled=True)
 
-        cluster.populate(2).start(wait_other_notice=True)
+        cluster.populate(2).start()
         node1, node2 = cluster.nodelist()
 
         session = self.patient_cql_connection(node1)
@@ -1472,7 +1472,7 @@ class TestConsistency(Tester):
         cluster.set_configuration_options(values={'hinted_handoff_enabled': False})
         cluster.set_batch_commitlog(enabled=True)
 
-        cluster.populate(3).start(wait_other_notice=True)
+        cluster.populate(3).start()
         node1, node2, node3 = cluster.nodelist()
 
         session = self.patient_cql_connection(node1)
diff --git a/consistent_bootstrap_test.py b/consistent_bootstrap_test.py
index 24626f3..073e8cd 100644
--- a/consistent_bootstrap_test.py
+++ b/consistent_bootstrap_test.py
@@ -65,7 +65,7 @@ class TestBootstrapConsistency(Tester):
 
         cluster.populate(2)
         node1, node2 = cluster.nodelist()
-        cluster.start(wait_for_binary_proto=True, wait_other_notice=True)
+        cluster.start()
 
         logger.debug("Set to talk to node 2")
         n2session = self.patient_cql_connection(node2)
diff --git a/counter_test.py b/counter_test.py
index ae87d95..2254038 100644
--- a/counter_test.py
+++ b/counter_test.py
@@ -102,7 +102,7 @@ class TestCounters(Tester):
             nodes[0].update_startup_byteman_script('./byteman/4.0/election_counter_leader_favor_node2.btm')
             nodes[2].update_startup_byteman_script('./byteman/4.0/election_counter_leader_favor_node2.btm')
 
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
         session = self.patient_cql_connection(nodes[0])
         create_ks(session, 'ks', 3)
         create_cf(session, 'cf', validation="CounterColumnType", columns={'c': 'counter'})
diff --git a/cql_test.py b/cql_test.py
index dcfc161..eced21d 100644
--- a/cql_test.py
+++ b/cql_test.py
@@ -53,7 +53,7 @@ class CQLTester(Tester):
             cluster.set_configuration_options(values=config)
 
         if not cluster.nodelist():
-            cluster.populate(nodes).start(wait_for_binary_proto=True)
+            cluster.populate(nodes).start()
         node1 = cluster.nodelist()[0]
 
         session = self.patient_cql_connection(node1, protocol_version=protocol_version, user=user, password=password)
@@ -803,8 +803,7 @@ class AbortedQueryTester(CQLTester):
         # introduced by CASSANDRA-7392 to pause by the specified amount of milliseconds every
         # CQL row iterated for non system queries, so that these queries take much longer to complete,
         # see ReadCommand.withStateTracking()
-        cluster.populate(1).start(wait_for_binary_proto=True,
-                                  jvm_args=["-Dcassandra.monitoring_report_interval_ms=10",
+        cluster.populate(1).start(jvm_args=["-Dcassandra.monitoring_report_interval_ms=10",
                                             "-Dcassandra.test.read_iteration_delay_ms=5"])
         node = cluster.nodelist()[0]
         session = self.patient_cql_connection(node)
@@ -919,8 +918,7 @@ class AbortedQueryTester(CQLTester):
                                                   'read_request_timeout_in_ms': 1000,
                                                   'range_request_timeout_in_ms': 1000})
 
-        cluster.populate(1).start(wait_for_binary_proto=True,
-                                  jvm_args=["-Dcassandra.monitoring_report_interval_ms=10",
+        cluster.populate(1).start(jvm_args=["-Dcassandra.monitoring_report_interval_ms=10",
                                             "-Dcassandra.test.read_iteration_delay_ms=5"])  # see above for explanation
         node = cluster.nodelist()[0]
         session = self.patient_cql_connection(node)
@@ -1034,8 +1032,7 @@ class TestCQLSlowQuery(CQLTester):
         # introduced by CASSANDRA-7392 to pause by the specified amount of milliseconds during each
         # iteration of non system queries, so that these queries take much longer to complete,
         # see ReadCommand.withStateTracking()
-        cluster.populate(1).start(wait_for_binary_proto=True,
-                                  jvm_args=["-Dcassandra.monitoring_report_interval_ms=10",
+        cluster.populate(1).start(jvm_args=["-Dcassandra.monitoring_report_interval_ms=10",
                                             "-Dcassandra.test.read_iteration_delay_ms=1"])
         node = cluster.nodelist()[0]
         session = self.patient_cql_connection(node)
@@ -1101,8 +1098,7 @@ class TestCQLSlowQuery(CQLTester):
         # introduced by CASSANDRA-7392 to pause by the specified amount of milliseconds during each
         # iteration of non system queries, so that these queries take much longer to complete,
         # see ReadCommand.withStateTracking()
-        cluster.populate(1).start(wait_for_binary_proto=True,
-                                  jvm_args=["-Dcassandra.monitoring_report_interval_ms=10",
+        cluster.populate(1).start(jvm_args=["-Dcassandra.monitoring_report_interval_ms=10",
                                             "-Dcassandra.test.read_iteration_delay_ms=1"])
         node = cluster.nodelist()[0]
         session = self.patient_cql_connection(node)
@@ -1397,7 +1393,7 @@ class TestLWTWithCQL(Tester):
     def fixture_post_initialize_cluster(self, fixture_dtest_setup):
         cluster = fixture_dtest_setup.cluster
         cluster.populate(3)
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
 
     def get_lwttester_session(self):
         node1 = self.cluster.nodelist()[0]
diff --git a/cql_tracing_test.py b/cql_tracing_test.py
index 869760b..da4414d 100644
--- a/cql_tracing_test.py
+++ b/cql_tracing_test.py
@@ -38,7 +38,7 @@ class TestCqlTracing(Tester):
         cluster.populate(nodes)
         node1 = cluster.nodelist()[0]
         remove_perf_disable_shared_mem(node1)  # necessary for jmx
-        cluster.start(wait_for_binary_proto=True, jvm_args=jvm_args)
+        cluster.start(jvm_args=jvm_args)
 
         session = self.patient_cql_connection(node1, protocol_version=protocol_version)
         if create_keyspace:
diff --git a/cqlsh_tests/test_cqlsh.py b/cqlsh_tests/test_cqlsh.py
index 86a46b2..39a1a2b 100644
--- a/cqlsh_tests/test_cqlsh.py
+++ b/cqlsh_tests/test_cqlsh.py
@@ -149,7 +149,7 @@ class TestCqlsh(Tester, CqlshMixin):
     def test_simple_insert(self):
 
         self.cluster.populate(1)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
 
         node1, = self.cluster.nodelist()
 
@@ -176,7 +176,7 @@ class TestCqlsh(Tester, CqlshMixin):
         """
 
         self.cluster.populate(1)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
 
         node1, = self.cluster.nodelist()
 
@@ -202,7 +202,7 @@ class TestCqlsh(Tester, CqlshMixin):
     @since('2.2')
     def test_past_and_future_dates(self):
         self.cluster.populate(1)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
 
         node1, = self.cluster.nodelist()
 
@@ -232,7 +232,7 @@ class TestCqlsh(Tester, CqlshMixin):
         @jira_ticket 10428
         """
         self.cluster.populate(1)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
 
         node1, = self.cluster.nodelist()
 
@@ -382,7 +382,7 @@ class TestCqlsh(Tester, CqlshMixin):
     def test_eat_glass(self):
 
         self.cluster.populate(1)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
 
         node1, = self.cluster.nodelist()
 
@@ -505,7 +505,7 @@ UPDATE varcharmaptable SET varcharvarintmap['Vitrum edere possum, mihi non nocet
     def test_source_glass(self):
 
         self.cluster.populate(1)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
 
         node1, = self.cluster.nodelist()
 
@@ -519,7 +519,7 @@ UPDATE varcharmaptable SET varcharvarintmap['Vitrum edere possum, mihi non nocet
         @jira_ticket CASSANDRA-11626
         """
         self.cluster.populate(1)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
 
         node1, = self.cluster.nodelist()
 
@@ -536,7 +536,7 @@ UPDATE varcharmaptable SET varcharvarintmap['Vitrum edere possum, mihi non nocet
         @jira_ticket CASSANDRA-11626
         """
         self.cluster.populate(1)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
 
         node1, = self.cluster.nodelist()
 
@@ -554,7 +554,7 @@ UPDATE varcharmaptable SET varcharvarintmap['Vitrum edere possum, mihi non nocet
         CASSANDRA-7196. Make sure the server returns empty values and CQLSH prints them properly
         """
         self.cluster.populate(1)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
 
         node1, = self.cluster.nodelist()
 
@@ -635,7 +635,7 @@ VALUES (4, blobAsInt(0x), '', blobAsBigint(0x), 0x, blobAsBoolean(0x), blobAsDec
         assert expected in output, "Output \n {0} \n doesn't contain expected\n {1}".format(output, expected)
 
     def test_tracing_from_system_traces(self):
-        self.cluster.populate(1).start(wait_for_binary_proto=True)
+        self.cluster.populate(1).start()
 
         node1, = self.cluster.nodelist()
 
@@ -767,7 +767,7 @@ VALUES (4, blobAsInt(0x), '', blobAsBigint(0x), 0x, blobAsBoolean(0x), blobAsDec
         @jira_ticket CASSANDRA-7814
         """
         self.cluster.populate(1)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
         node1, = self.cluster.nodelist()
 
         self.execute(
@@ -854,7 +854,7 @@ VALUES (4, blobAsInt(0x), '', blobAsBigint(0x), 0x, blobAsBoolean(0x), blobAsDec
 
     def test_describe_describes_non_default_compaction_parameters(self):
         self.cluster.populate(1)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
         node, = self.cluster.nodelist()
         session = self.patient_cql_connection(node)
         create_ks(session, 'ks', 1)
@@ -869,7 +869,7 @@ VALUES (4, blobAsInt(0x), '', blobAsBigint(0x), 0x, blobAsBoolean(0x), blobAsDec
     def test_describe_functions(self, fixture_dtest_setup_overrides):
         """Test DESCRIBE statements for functions and aggregate functions"""
         self.cluster.populate(1)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
 
         create_ks_statement = "CREATE KEYSPACE test WITH REPLICATION = {'class' : 'SimpleStrategy', 'replication_factor' : 1}"
         create_function_statement = """
@@ -941,7 +941,7 @@ CREATE OR REPLACE AGGREGATE test.average(int)
     @since('4.0')
     def test_default_keyspaces_exist(self):
         self.cluster.populate(1)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
         node1, = self.cluster.nodelist()
 
         # Describe keyspaces
@@ -962,7 +962,7 @@ CREATE OR REPLACE AGGREGATE test.average(int)
     def test_describe_types(self):
         """Test DESCRIBE statements for user defined datatypes"""
         self.cluster.populate(1)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
 
         create_ks_statement = "CREATE KEYSPACE test WITH REPLICATION = {'class' : 'SimpleStrategy', 'replication_factor' : 1}"
         create_name_type_statement = """
@@ -1003,7 +1003,7 @@ CREATE TYPE test.address_type (
         Test that we can describe tables whose name is a non-reserved CQL keyword
         """
         self.cluster.populate(1)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
         node, = self.cluster.nodelist()
         session = self.patient_cql_connection(node)
         create_ks(session, 'ks', 1)
@@ -1020,7 +1020,7 @@ CREATE TYPE test.address_type (
         """
         self.cluster.set_configuration_options({'enable_materialized_views': 'true'})
         self.cluster.populate(1)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
 
         self.execute(
             cql="""
@@ -1451,7 +1451,7 @@ CREATE TYPE test.address_type (
     def test_float_formatting(self):
         """ Tests for CASSANDRA-9224, check format of float and double values"""
         self.cluster.populate(1)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
 
         node1, = self.cluster.nodelist()
 
@@ -1629,7 +1629,7 @@ CREATE TYPE test.address_type (
     def test_int_values(self):
         """ Tests for CASSANDRA-9399, check tables with int, bigint, smallint and tinyint values"""
         self.cluster.populate(1)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
 
         node1, = self.cluster.nodelist()
 
@@ -1666,7 +1666,7 @@ CREATE TABLE int_checks.values (
     def test_datetime_values(self):
         """ Tests for CASSANDRA-9399, check tables with date and time values"""
         self.cluster.populate(1)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
 
         node1, = self.cluster.nodelist()
 
@@ -1712,7 +1712,7 @@ CREATE TABLE datetime_checks.values (
     @since('4.0')
     def test_datetime_values_40(self):
         self.cluster.populate(1)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
 
         node1, = self.cluster.nodelist()
 
@@ -1759,7 +1759,7 @@ CREATE TABLE datetime_checks.values (
         therefore make this test too brittle.
         """
         self.cluster.populate(1)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
 
         node1, = self.cluster.nodelist()
 
@@ -1806,7 +1806,7 @@ Tracing session:""")
         self.cluster.set_configuration_options({
             'unlogged_batch_across_partitions_warn_threshold': str(max_partitions_per_batch)})
 
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
 
         node1 = self.cluster.nodelist()[0]
 
@@ -1847,7 +1847,7 @@ Tracing session:""")
         @jira_ticket CASSANDRA-9601
         """
         self.cluster.populate(1)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
 
         node1, = self.cluster.nodelist()
 
@@ -1860,7 +1860,7 @@ Tracing session:""")
         @jira_ticket CASSANDRA-15193
         """
         self.cluster.populate(1)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
 
         node1, = self.cluster.nodelist()
         stdout, stderr = self.run_cqlsh(node1, cmds='USE system', cqlsh_options=['--tty'])
@@ -1876,7 +1876,7 @@ Tracing session:""")
         """
         self.cluster.populate(1)
         self.cluster.set_configuration_options({ 'native_transport_max_negotiable_protocol_version': str(3)})
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
 
         node1, = self.cluster.nodelist()
         stdout, stderr = self.run_cqlsh(node1, cmds='USE system', cqlsh_options=['--tty'])
@@ -1890,14 +1890,13 @@ Tracing session:""")
         if node1.get_cassandra_version() < '4.0':
             assert "ProtocolError returned from server while using explicitly set client protocol_version 4" in stderr
 
-
     def test_update_schema_with_down_node(self):
         """
         Test that issuing a DML statement after a DDL statement will work with a down node
         @jira_ticket CASSANDRA-9689
         """
         self.cluster.populate(3)
-        self.cluster.start(wait_for_binary_proto=True, wait_other_notice=True)
+        self.cluster.start()
 
         node1, node2, node3 = self.cluster.nodelist()
         node2.stop(wait_other_notice=True)
@@ -1930,7 +1929,7 @@ Tracing session:""")
         indicates the output of DESCRIBE is not a correct CREATE TABLE statement.
         """
         self.cluster.populate(1)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
         node1, = self.cluster.nodelist()
         session = self.patient_cql_connection(node1)
 
@@ -1960,7 +1959,7 @@ Tracing session:""")
         """
         self.cluster.set_configuration_options({'enable_materialized_views': 'true'})
         self.cluster.populate(1)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
         node1, = self.cluster.nodelist()
         session = self.patient_cql_connection(node1)
 
@@ -2042,7 +2041,7 @@ Tracing session:""")
             expression.
         """
         self.cluster.populate(1)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
         node1, = self.cluster.nodelist()
 
         out, err = self.run_cqlsh(node1, cmd, env_vars={'TERM': 'xterm'})
@@ -2058,7 +2057,7 @@ Tracing session:""")
         @jira_ticket CASSANDRA-10272
         """
         self.cluster.populate(1)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
         node1, = self.cluster.nodelist()
 
         stdout, stderr = self.run_cqlsh(node1, cmds="""
@@ -2076,7 +2075,7 @@ Tracing session:""")
         @jira_ticket CASSANDRA-15660
         """
         self.cluster.populate(1)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
         node1, = self.cluster.nodelist()
         session = self.patient_cql_connection(node1)
 
@@ -2096,7 +2095,7 @@ Tracing session:""")
         @jira_ticket CASSANDRA-15905
         """
         self.cluster.populate(1)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
         node1, = self.cluster.nodelist()
         session = self.patient_cql_connection(node1)
 
@@ -2125,7 +2124,7 @@ class TestCqlshSmoke(Tester, CqlshMixin):
 
     @pytest.fixture(scope='function', autouse=True)
     def fixture_cluster_setup(self, fixture_dtest_setup):
-        fixture_dtest_setup.cluster.populate(1).start(wait_for_binary_proto=True)
+        fixture_dtest_setup.cluster.populate(1).start()
         [self.node1] = fixture_dtest_setup.cluster.nodelist()
         self.session = fixture_dtest_setup.patient_cql_connection(self.node1)
 
@@ -2429,7 +2428,7 @@ class TestCqlLogin(Tester, CqlshMixin):
         cluster = fixture_dtest_setup.cluster
         config = {'authenticator': 'org.apache.cassandra.auth.PasswordAuthenticator'}
         cluster.set_configuration_options(values=config)
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         [self.node1] = cluster.nodelist()
         self.node1.watch_log_for('Created default superuser')
         self.session = fixture_dtest_setup.patient_cql_connection(self.node1, user='cassandra', password='cassandra')
@@ -2545,7 +2544,7 @@ class TestCqlshUnicode(Tester, CqlshMixin):
     @pytest.fixture(scope='function', autouse=True)
     def fixture_cluster_setup(self, fixture_dtest_setup):
         cluster = fixture_dtest_setup.cluster
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         [self.node1] = cluster.nodelist()
         self.session = fixture_dtest_setup.patient_cql_connection(self.node1)
 
diff --git a/cqlsh_tests/test_cqlsh_copy.py b/cqlsh_tests/test_cqlsh_copy.py
index e4155fc..88ecf44 100644
--- a/cqlsh_tests/test_cqlsh_copy.py
+++ b/cqlsh_tests/test_cqlsh_copy.py
@@ -101,7 +101,7 @@ class TestCqlshCopy(Tester):
 
             if configuration_options:
                 self.cluster.set_configuration_options(values=configuration_options)
-            self.cluster.populate(nodes, tokens=tokens).start(wait_for_binary_proto=True)
+            self.cluster.populate(nodes, tokens=tokens).start()
         else:
             assert self.cluster.partitioner == p, "Cannot reuse cluster: different partitioner"
             assert len(self.cluster.nodelist()) == nodes, "Cannot reuse cluster: different number of nodes"
diff --git a/delete_insert_test.py b/delete_insert_test.py
index e0c4bbc..a172c92 100644
--- a/delete_insert_test.py
+++ b/delete_insert_test.py
@@ -39,7 +39,7 @@ class TestDeleteInsert(Tester):
 
     def test_delete_insert_search(self):
         cluster = self.cluster
-        cluster.populate([2, 2]).start(wait_for_binary_proto=True)
+        cluster.populate([2, 2]).start()
         node1 = cluster.nodelist()[0]
 
         session = self.patient_cql_connection(node1)
diff --git a/deletion_test.py b/deletion_test.py
index d6930db..88ca6cf 100644
--- a/deletion_test.py
+++ b/deletion_test.py
@@ -51,7 +51,7 @@ class TestDeletion(Tester):
 
         remove_perf_disable_shared_mem(node1)
 
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
         [node1] = self.cluster.nodelist()
         session = self.patient_cql_connection(node1)
         create_ks(session, 'ks', 1)
diff --git a/disk_balance_test.py b/disk_balance_test.py
index 5d9918d..6238f03 100644
--- a/disk_balance_test.py
+++ b/disk_balance_test.py
@@ -37,7 +37,7 @@ class TestDiskBalance(Tester):
         cluster = self.cluster
         if self.dtest_config.use_vnodes:
             cluster.set_configuration_options(values={'num_tokens': 256})
-        cluster.populate(4).start(wait_for_binary_proto=True)
+        cluster.populate(4).start()
         node1 = cluster.nodes['node1']
 
         node1.stress(['write', 'n=50k', 'no-warmup', '-rate', 'threads=100', '-schema', 'replication(factor=3)',
@@ -54,7 +54,7 @@ class TestDiskBalance(Tester):
             cluster.set_configuration_options(values={'num_tokens': 256})
         # apparently we have legitimate errors in the log when bootstrapping (see bootstrap_test.py)
         self.fixture_dtest_setup.allow_log_errors = True
-        cluster.populate(4).start(wait_for_binary_proto=True)
+        cluster.populate(4).start()
         node1 = cluster.nodes['node1']
 
         node1.stress(['write', 'n=50k', 'no-warmup', '-rate', 'threads=100', '-schema', 'replication(factor=3)',
@@ -77,7 +77,7 @@ class TestDiskBalance(Tester):
             cluster.set_configuration_options(values={'num_tokens': 256})
         # apparently we have legitimate errors in the log when bootstrapping (see bootstrap_test.py)
         self.fixture_dtest_setup.allow_log_errors = True
-        cluster.populate(4).start(wait_for_binary_proto=True)
+        cluster.populate(4).start()
         node1 = cluster.nodes['node1']
 
         logger.debug("Populating")
@@ -107,7 +107,7 @@ class TestDiskBalance(Tester):
         cluster = self.cluster
         if self.dtest_config.use_vnodes:
             cluster.set_configuration_options(values={'num_tokens': 256})
-        cluster.populate(4).start(wait_for_binary_proto=True)
+        cluster.populate(4).start()
         node1 = cluster.nodes['node1']
         node4 = cluster.nodes['node4']
         node1.stress(['write', 'n=50k', 'no-warmup', '-rate', 'threads=100', '-schema', 'replication(factor=2)',
@@ -128,7 +128,7 @@ class TestDiskBalance(Tester):
         cluster.populate(1)
         [node] = cluster.nodelist()
         remove_perf_disable_shared_mem(node)
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
 
         session = self.patient_cql_connection(node)
         create_ks(session, 'ks', 1)
@@ -155,7 +155,7 @@ class TestDiskBalance(Tester):
         cluster = self.cluster
         if self.dtest_config.use_vnodes:
             cluster.set_configuration_options(values={'num_tokens': 256})
-        cluster.populate(3).start(wait_for_binary_proto=True)
+        cluster.populate(3).start()
         node1 = cluster.nodes['node1']
         node1.stress(['write', 'n=1', 'no-warmup', '-rate', 'threads=100', '-schema', 'replication(factor=1)'])
         cluster.flush()
@@ -208,7 +208,7 @@ class TestDiskBalance(Tester):
         cluster.set_configuration_options(values={'concurrent_compactors': num_disks})
 
         logger.debug("Starting node1 with {} data dirs and concurrent_compactors".format(num_disks))
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         [node1] = cluster.nodelist()
 
         session = self.patient_cql_connection(node1)
@@ -281,7 +281,7 @@ class TestDiskBalance(Tester):
         cluster.set_configuration_options(values={'concurrent_compactors': num_disks})
 
         logger.debug("Starting 3 nodes with {} data dirs and concurrent_compactors".format(num_disks))
-        cluster.populate(3).start(wait_for_binary_proto=True)
+        cluster.populate(3).start()
         node1 = cluster.nodelist()[0]
 
         num_flushes = 10
diff --git a/fqltool_test.py b/fqltool_test.py
index 2625bb4..1fa6bc0 100644
--- a/fqltool_test.py
+++ b/fqltool_test.py
@@ -23,7 +23,7 @@ class TestFQLTool(Tester):
         query log, then makes sure that the data is correct.
         @jira_ticket CASSANDRA-14690
         """
-        self.cluster.populate(2).start(wait_for_binary_proto=True)
+        self.cluster.populate(2).start()
         node1, node2 = self.cluster.nodelist()
 
         with tempfile.TemporaryDirectory() as temp_dir:
@@ -61,7 +61,7 @@ class TestFQLTool(Tester):
         sure that the results match
         @jira_ticket CASSANDRA-14690
         """
-        self.cluster.populate(1).start(wait_for_binary_proto=True)
+        self.cluster.populate(1).start()
         node1 = self.cluster.nodelist()[0]
 
         with tempfile.TemporaryDirectory() as temp_dir:
@@ -87,7 +87,7 @@ class TestFQLTool(Tester):
         then asserts that the replays of each generates a mismatch
         @jira_ticket CASSANDRA-14690
         """
-        self.cluster.populate(1).start(wait_for_binary_proto=True)
+        self.cluster.populate(1).start()
         node1 = self.cluster.nodelist()[0]
 
         with tempfile.TemporaryDirectory() as temp_dir:
diff --git a/hintedhandoff_test.py b/hintedhandoff_test.py
index 68d341e..d0c1709 100644
--- a/hintedhandoff_test.py
+++ b/hintedhandoff_test.py
@@ -181,7 +181,7 @@ class TestHintedHandoff(Tester):
 
     @pytest.mark.no_vnodes
     def test_hintedhandoff_decom(self):
-        self.cluster.populate(4).start(wait_for_binary_proto=True)
+        self.cluster.populate(4).start()
         [node1, node2, node3, node4] = self.cluster.nodelist()
         session = self.patient_cql_connection(node1)
         create_ks(session, 'ks', 2)
diff --git a/jmx_test.py b/jmx_test.py
index 760b801..dd90da6 100644
--- a/jmx_test.py
+++ b/jmx_test.py
@@ -34,7 +34,7 @@ class TestJMX(Tester):
         """
         #
         cluster = self.cluster
-        cluster.populate(3).start(wait_for_binary_proto=True)
+        cluster.populate(3).start()
         node1, node2, node3 = cluster.nodelist()
 
         node1.stress(['write', 'n=500K', 'no-warmup', '-schema', 'replication(factor=3)'])
@@ -73,7 +73,7 @@ class TestJMX(Tester):
         cluster.populate(3)
         node1, node2, node3 = cluster.nodelist()
         remove_perf_disable_shared_mem(node1)
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
 
         version = cluster.version()
         node1.stress(['write', 'n=10K', 'no-warmup', '-schema', 'replication(factor=3)'])
@@ -114,7 +114,7 @@ class TestJMX(Tester):
         cluster.populate(1)
         node = cluster.nodelist()[0]
         remove_perf_disable_shared_mem(node)
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
 
         node.run_cqlsh(cmds="""
             CREATE KEYSPACE mvtest WITH REPLICATION = { 'class' : 'SimpleStrategy', 'replication_factor': 1 };
@@ -188,7 +188,7 @@ class TestJMX(Tester):
         cluster.populate(1)
         node = cluster.nodelist()[0]
         remove_perf_disable_shared_mem(node)
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
 
         # Run a quick stress command to create the keyspace and table
         node.stress(['write', 'n=1', 'no-warmup'])
@@ -248,7 +248,7 @@ class TestJMX(Tester):
         @jira_ticket CASSANDRA-9526
         """
         cluster = self.cluster
-        cluster.populate(3).start(wait_for_binary_proto=True)
+        cluster.populate(3).start()
         node1, node2, node3 = cluster.nodelist()
 
         stdout = node1.nodetool("failuredetector").stdout
diff --git a/legacy_sstables_test.py b/legacy_sstables_test.py
index bedfbea..8dc15e9 100644
--- a/legacy_sstables_test.py
+++ b/legacy_sstables_test.py
@@ -28,7 +28,7 @@ class TestLegacySSTables(Tester):
         # start with 2.1.20 to generate a legacy sstable
         cluster.set_install_dir(version='2.1.20')
 
-        cluster.populate(1).start(wait_other_notice=True)
+        cluster.populate(1).start()
         node1 = cluster.nodelist()[0]
         session = self.patient_cql_connection(node1)
 
diff --git a/materialized_views_test.py b/materialized_views_test.py
index f1092fc..7305015 100644
--- a/materialized_views_test.py
+++ b/materialized_views_test.py
@@ -1118,7 +1118,7 @@ class TestMaterializedViews(Tester):
             node.mark_log(filename='debug.log')
 
         logger.debug("Restart the cluster")
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
         session = self.patient_cql_connection(node1)
         session.execute("USE ks")
 
@@ -1292,7 +1292,7 @@ class TestMaterializedViews(Tester):
         logger.debug("Restart the cluster")
         self.cluster.stop()
         marks = [node.mark_log() for node in nodes]
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
         session = self.patient_cql_connection(nodes[0])
 
         logger.debug("Verify that the MV has been successfully created")
@@ -2884,7 +2884,7 @@ class TestMaterializedViewsLockcontention(Tester):
         for node in self.nodes:
             remove_perf_disable_shared_mem(node)
 
-        self.cluster.start(wait_for_binary_proto=True, jvm_args=[
+        self.cluster.start(jvm_args=[
             "-Dcassandra.test.fail_mv_locks_count=64"
         ])
 
diff --git a/metadata_test.py b/metadata_test.py
index 90141e7..5842336 100644
--- a/metadata_test.py
+++ b/metadata_test.py
@@ -38,7 +38,7 @@ class TestMetadata(Tester):
         self.fixture_dtest_setup.ignore_log_patterns = ['.*Unknown keyspace/cf pair.*']
 
         cluster = self.cluster
-        cluster.populate(2).start(wait_other_notice=True)
+        cluster.populate(2).start()
         (node1, node2) = cluster.nodelist()
 
         node1.nodetool("disableautocompaction")
diff --git a/nodetool_test.py b/nodetool_test.py
index 381e917..abbf0bc 100644
--- a/nodetool_test.py
+++ b/nodetool_test.py
@@ -54,7 +54,7 @@ class TestNodetool(Tester):
                 for line in ["dc={}".format(node.data_center), "rack=rack{}".format(i % 2)]:
                     snitch_file.write(line + os.linesep)
 
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
 
         for i, node in enumerate(cluster.nodelist()):
             out, err, _ = node.nodetool('info')
@@ -120,7 +120,7 @@ class TestNodetool(Tester):
         @jira_ticket CASSANDRA-13526
         Test nodetool cleanup KS to remove old data when new replicas in current node instead of directly returning success.
         """
-        self.cluster.populate([1, 1]).start(wait_for_binary_proto=True, wait_other_notice=True)
+        self.cluster.populate([1, 1]).start()
 
         node_dc1 = self.cluster.nodelist()[0]
         node_dc2 = self.cluster.nodelist()[1]
@@ -360,7 +360,7 @@ class TestNodetool(Tester):
         @expected_result This test invokes nodetool describecluster and matches the output with the expected one
         """
         cluster = self.cluster
-        cluster.populate([1, 2, 1]).start(wait_for_binary_proto=True)
+        cluster.populate([1, 2, 1]).start()
 
         node1_dc1, node1_dc2, node2_dc2, node1_dc3 = cluster.nodelist()
 
diff --git a/offline_tools_test.py b/offline_tools_test.py
index 2b8b7a3..725cc65 100644
--- a/offline_tools_test.py
+++ b/offline_tools_test.py
@@ -34,7 +34,7 @@ class TestOfflineTools(Tester):
         @jira_ticket CASSANDRA-7614
         """
         cluster = self.cluster
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         node1 = cluster.nodelist()[0]
 
         # test by trying to run on nonexistent keyspace
@@ -47,7 +47,7 @@ class TestOfflineTools(Tester):
             assert e.exit_status == 1, "Expected sstablelevelreset to have a return code of 1 == but instead return code was {}".format(e.exit_status)
 
         # now test by generating keyspace but not flushing sstables
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
         node1.stress(['write', 'n=100', 'no-warmup', '-schema', 'replication(factor=1)',
                       '-rate', 'threads=8'])
         cluster.stop(gently=False)
@@ -58,7 +58,7 @@ class TestOfflineTools(Tester):
         assert rc == 0, str(rc)
 
         # test by writing small amount of data and flushing (all sstables should be level 0)
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
         session = self.patient_cql_connection(node1)
         session.execute("ALTER TABLE keyspace1.standard1 with compaction={'class': 'LeveledCompactionStrategy', 'sstable_size_in_mb':1};")
         node1.stress(['write', 'n=1K', 'no-warmup', '-schema', 'replication(factor=1)',
@@ -72,7 +72,7 @@ class TestOfflineTools(Tester):
         assert rc == 0, str(rc)
 
         # test by loading large amount data so we have multiple levels and checking all levels are 0 at end
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
         node1.stress(['write', 'n=50K', 'no-warmup', '-schema', 'replication(factor=1)',
                       '-rate', 'threads=8'])
         cluster.flush()
@@ -116,7 +116,7 @@ class TestOfflineTools(Tester):
         """
         cluster = self.cluster
         cluster.set_configuration_options(values={'compaction_throughput_mb_per_sec': 0})
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         node1 = cluster.nodelist()[0]
 
         # NOTE - As of now this does not return when it encounters Exception and causes test to hang, temporarily commented out
@@ -143,7 +143,7 @@ class TestOfflineTools(Tester):
             assert e.exit_status == 1, str(e.exit_status)
 
         # test by flushing (sstable should be level 0)
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
         session = self.patient_cql_connection(node1)
         logger.debug("Altering compaction strategy to LCS")
         session.execute("ALTER TABLE keyspace1.standard1 with compaction={'class': 'LeveledCompactionStrategy', 'sstable_size_in_mb':1, 'enabled':'false'};")
@@ -160,7 +160,7 @@ class TestOfflineTools(Tester):
         assert re.search("L0=1", output)
         assert rc == 0, str(rc)
 
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
         node1.nodetool('enableautocompaction keyspace1 standard1')
         # test by loading large amount data so we have multiple sstables
         # must write enough to create more than just L1 sstables
@@ -221,7 +221,7 @@ class TestOfflineTools(Tester):
         Test on potential situations: deleted sstables, corrupted sstables
         """
         cluster = self.cluster
-        cluster.populate(3).start(wait_for_binary_proto=True)
+        cluster.populate(3).start()
         node1, node2, node3 = cluster.nodelist()
 
         # test on nonexistent keyspace
@@ -298,7 +298,7 @@ class TestOfflineTools(Tester):
 
     def test_sstableexpiredblockers(self):
         cluster = self.cluster
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         [node1] = cluster.nodelist()
         session = self.patient_cql_connection(node1)
         create_ks(session, 'ks', 1)
@@ -351,7 +351,7 @@ class TestOfflineTools(Tester):
             cluster.set_install_dir(version='github:apache/cassandra-3.0')
 
         # Start up last major version, write out an sstable to upgrade, and stop node
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         [node1] = cluster.nodelist()
         # Check that node1 is actually what we expect
         logger.debug('Downgraded install dir: {}'.format(node1.get_install_dir()))
@@ -368,7 +368,7 @@ class TestOfflineTools(Tester):
         # Check that node1 is actually upgraded
         logger.debug('Upgraded to original install dir: {}'.format(node1.get_install_dir()))
         # Perform a node start/stop so system tables get internally updated, otherwise we may get "Unknown keyspace/table ks.cf"
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
         node1.flush()
         cluster.stop()
 
@@ -402,7 +402,7 @@ class TestOfflineTools(Tester):
         cluster = self.cluster
         # disable JBOD conf since the test expects exactly one SSTable to be written.
         cluster.set_datadir_count(1)
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         [node1] = cluster.nodelist()
         session = self.patient_cql_connection(node1)
         create_ks(session, 'ks', 1)
diff --git a/paging_test.py b/paging_test.py
index 6be3f7d..2fdd53b 100644
--- a/paging_test.py
+++ b/paging_test.py
@@ -30,7 +30,7 @@ class BasePagingTester(Tester):
         supports_v5 = self.supports_v5_protocol(self.cluster.version())
         protocol_version = 5 if supports_v5 else None
         cluster = self.cluster
-        cluster.populate(3).start(wait_for_binary_proto=True)
+        cluster.populate(3).start()
         node1 = cluster.nodelist()[0]
         session = self.patient_cql_connection(node1,
                                               protocol_version=protocol_version,
diff --git a/paxos_test.py b/paxos_test.py
index 736ca46..3ee29f7 100644
--- a/paxos_test.py
+++ b/paxos_test.py
@@ -53,10 +53,10 @@ class TestPaxos(Tester):
         self.cluster.nodelist()[1].stop()
         assert_unavailable(session.execute, "INSERT INTO test (k, v) VALUES (2, 2) IF NOT EXISTS")
 
-        self.cluster.nodelist()[1].start(wait_for_binary_proto=True, wait_other_notice=True)
+        self.cluster.nodelist()[1].start()
         session.execute("INSERT INTO test (k, v) VALUES (3, 3) IF NOT EXISTS")
 
-        self.cluster.nodelist()[2].start(wait_for_binary_proto=True)
+        self.cluster.nodelist()[2].start()
         session.execute("INSERT INTO test (k, v) VALUES (4, 4) IF NOT EXISTS")
 
     @pytest.mark.no_vnodes
@@ -75,10 +75,10 @@ class TestPaxos(Tester):
         self.cluster.nodelist()[1].stop()
         session.execute("INSERT INTO test (k, v) VALUES (3, 2) IF NOT EXISTS")
 
-        self.cluster.nodelist()[1].start(wait_for_binary_proto=True)
+        self.cluster.nodelist()[1].start()
         session.execute("INSERT INTO test (k, v) VALUES (5, 5) IF NOT EXISTS")
 
-        self.cluster.nodelist()[2].start(wait_for_binary_proto=True)
+        self.cluster.nodelist()[2].start()
         session.execute("INSERT INTO test (k, v) VALUES (6, 6) IF NOT EXISTS")
 
     def test_contention_multi_iterations(self):
diff --git a/pending_range_test.py b/pending_range_test.py
index 55d810b..65c5f80 100644
--- a/pending_range_test.py
+++ b/pending_range_test.py
@@ -23,7 +23,7 @@ class TestPendingRangeMovements(Tester):
             cluster.set_log_level('DEBUG')
 
         # Create 5 node cluster
-        cluster.populate(5).start(wait_for_binary_proto=True)
+        cluster.populate(5).start()
         node1, node2 = cluster.nodelist()[0:2]
 
         # Set up RF=3 keyspace
diff --git a/pushed_notifications_test.py b/pushed_notifications_test.py
index f9544a4..b67c626 100644
--- a/pushed_notifications_test.py
+++ b/pushed_notifications_test.py
@@ -93,7 +93,7 @@ class TestPushedNotifications(Tester):
         @jira_ticket CASSANDRA-8516
         Moving a token should result in MOVED_NODE notifications.
         """
-        self.cluster.populate(3).start(wait_for_binary_proto=True, wait_other_notice=True)
+        self.cluster.populate(3).start()
 
         waiters = [NotificationWaiter(self, node, ["TOPOLOGY_CHANGE"])
                    for node in list(self.cluster.nodes.values())]
@@ -136,7 +136,7 @@ class TestPushedNotifications(Tester):
 
         self.change_rpc_address_to_localhost()
 
-        cluster.start(wait_for_binary_proto=True, wait_other_notice=True)
+        cluster.start()
 
         waiters = [NotificationWaiter(self, node, ["TOPOLOGY_CHANGE"])
                    for node in list(self.cluster.nodes.values())]
@@ -173,7 +173,7 @@ class TestPushedNotifications(Tester):
         @jira_ticket CASSANDRA-7816
         Restarting a node should generate exactly one DOWN and one UP notification
         """
-        self.cluster.populate(2).start(wait_for_binary_proto=True, wait_other_notice=True)
+        self.cluster.populate(2).start()
         node1, node2 = self.cluster.nodelist()
 
         waiter = NotificationWaiter(self, node1, ["STATUS_CHANGE", "TOPOLOGY_CHANGE"])
@@ -223,7 +223,7 @@ class TestPushedNotifications(Tester):
 
         self.change_rpc_address_to_localhost()
 
-        cluster.start(wait_for_binary_proto=True, wait_other_notice=True)
+        cluster.start()
 
         # register for notification with node1
         waiter = NotificationWaiter(self, node1, ["STATUS_CHANGE", "TOPOLOGY_CHANGE"])
@@ -260,7 +260,7 @@ class TestPushedNotifications(Tester):
         Test that NEW_NODE and REMOVED_NODE are sent correctly as nodes join and leave.
         @jira_ticket CASSANDRA-11038
         """
-        self.cluster.populate(1).start(wait_for_binary_proto=True)
+        self.cluster.populate(1).start()
         node1 = self.cluster.nodelist()[0]
 
         waiter = NotificationWaiter(self, node1, ["STATUS_CHANGE", "TOPOLOGY_CHANGE"])
@@ -325,7 +325,7 @@ class TestPushedNotifications(Tester):
         will generate the correct schema change notifications.
         """
         self.cluster.set_configuration_options({'enable_materialized_views': 'true'})
-        self.cluster.populate(2).start(wait_for_binary_proto=True)
+        self.cluster.populate(2).start()
         node1, node2 = self.cluster.nodelist()
 
         session = self.patient_cql_connection(node1)
diff --git a/read_failures_test.py b/read_failures_test.py
index 82e132c..3f35575 100644
--- a/read_failures_test.py
+++ b/read_failures_test.py
@@ -37,7 +37,7 @@ class TestReadFailures(Tester):
             values={'tombstone_failure_threshold': self.tombstone_failure_threshold}
         )
         self.cluster.populate(3)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
         self.nodes = list(self.cluster.nodes.values())
 
         session = self.patient_exclusive_cql_connection(self.nodes[0], protocol_version=self.protocol_version)
diff --git a/read_repair_test.py b/read_repair_test.py
index 0d725d0..f096d62 100644
--- a/read_repair_test.py
+++ b/read_repair_test.py
@@ -98,7 +98,7 @@ class TestReadRepair(Tester):
                 snitch_file.write("rack=rack1" + os.linesep)
                 snitch_file.write("prefer_local=true" + os.linesep)
 
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
 
     @since('3.0')
     def test_alter_rf_and_run_read_repair(self):
@@ -302,7 +302,7 @@ class TestReadRepair(Tester):
 
         self.cluster.flush()
         self.cluster.stop()
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
         session1 = self.patient_exclusive_cql_connection(node1)
 
         for n in range(1, 1000):
@@ -444,7 +444,7 @@ class TestSpeculativeReadRepair(Tester):
                                                   'read_request_timeout_in_ms': 1000})
         cluster.populate(3, install_byteman=True, debug=True)
         byteman_validate(cluster.nodelist()[0], './byteman/read_repair/sorted_live_endpoints.btm', verbose=True)
-        cluster.start(wait_for_binary_proto=True, jvm_args=['-XX:-PerfDisableSharedMem'])
+        cluster.start(jvm_args=['-XX:-PerfDisableSharedMem'])
         session = fixture_dtest_setup.patient_exclusive_cql_connection(cluster.nodelist()[0], timeout=2)
 
         session.execute("CREATE KEYSPACE ks WITH replication = {'class': 'SimpleStrategy', 'replication_factor': 3}")
@@ -744,8 +744,7 @@ class TestReadRepairGuarantees(Tester):
                                                   'dynamic_snitch': False,
                                                   'write_request_timeout_in_ms': 500,
                                                   'read_request_timeout_in_ms': 500})
-        cluster.populate(3, install_byteman=True, debug=True).start(wait_for_binary_proto=True,
-                                                                    jvm_args=['-XX:-PerfDisableSharedMem'])
+        cluster.populate(3, install_byteman=True, debug=True).start(jvm_args=['-XX:-PerfDisableSharedMem'])
         session = fixture_dtest_setup.patient_exclusive_cql_connection(cluster.nodelist()[0], timeout=2)
 
         session.execute("CREATE KEYSPACE ks WITH replication = {'class': 'SimpleStrategy', 'replication_factor': 3}")
diff --git a/rebuild_test.py b/rebuild_test.py
index 7d90650..91bbfdd 100644
--- a/rebuild_test.py
+++ b/rebuild_test.py
@@ -331,7 +331,7 @@ class TestRebuild(Tester):
         node1.set_configuration_options(values={'initial_token': node1_token})
         node2.set_configuration_options(values={'initial_token': node2_token})
         node3.set_configuration_options(values={'initial_token': node3_token})
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
 
         session = self.patient_exclusive_cql_connection(node1)
         session.execute("CREATE KEYSPACE ks1 WITH replication = {'class':'SimpleStrategy', 'replication_factor':2};")
@@ -365,7 +365,7 @@ class TestRebuild(Tester):
         node1.set_configuration_options(values={'initial_token': node1_token})
         node2.set_configuration_options(values={'initial_token': node2_token})
         node3.set_configuration_options(values={'initial_token': node3_token})
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
 
         node3_address = node3.network_interfaces['binary'][0]
 
@@ -401,7 +401,7 @@ class TestRebuild(Tester):
         cluster.populate([1, 1], tokens=tokens[:2])
         node1, node2 = cluster.nodelist()
 
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
 
         # populate data in dc1, dc2
         session = self.patient_exclusive_cql_connection(node1)
diff --git a/refresh_test.py b/refresh_test.py
index 4177ec8..4c41494 100644
--- a/refresh_test.py
+++ b/refresh_test.py
@@ -14,7 +14,7 @@ class TestRefresh(Tester):
         node = self.cluster.nodelist()[0]
         node.byteman_port = '8100'
         node.import_config_files()
-        self.cluster.start(wait_other_notice=True)
+        self.cluster.start()
         session = self.patient_cql_connection(node)
         session.execute("CREATE KEYSPACE ks WITH replication = {'class':'SimpleStrategy', 'replication_factor':1}")
         session.execute("CREATE TABLE ks.a (id int primary key, d text)")
diff --git a/repair_tests/deprecated_repair_test.py b/repair_tests/deprecated_repair_test.py
index b11b08c..fbeceb3 100644
--- a/repair_tests/deprecated_repair_test.py
+++ b/repair_tests/deprecated_repair_test.py
@@ -228,7 +228,7 @@ class TestDeprecatedRepairNotifications(Tester):
         cluster = self.cluster
         cluster.populate(2)
         node1, node2 = cluster.nodelist()
-        cluster.start(wait_for_binary_proto=True, wait_other_notice=True)
+        cluster.start()
 
         # write some data that could be repaired
         logger.debug("Stressing node1...")
diff --git a/repair_tests/incremental_repair_test.py b/repair_tests/incremental_repair_test.py
index c8e7024..4920fc1 100644
--- a/repair_tests/incremental_repair_test.py
+++ b/repair_tests/incremental_repair_test.py
@@ -578,7 +578,7 @@ class TestIncRepair(Tester):
         @jira_ticket CASSANDRA-11172 - repeated full repairs should not cause infinite loop in getNextBackgroundTask
         """
         cluster = self.cluster
-        cluster.populate(2).start(wait_for_binary_proto=True)
+        cluster.populate(2).start()
         node1, node2 = cluster.nodelist()
         for x in range(0, 10):
             node1.stress(['write', 'n=100k', 'no-warmup', '-rate', 'threads=10', '-schema', 'compaction(strategy=LeveledCompactionStrategy,sstable_size_in_mb=10)', 'replication(factor=2)'])
@@ -663,7 +663,7 @@ class TestIncRepair(Tester):
         * Assert no extra, unrepaired sstables are generated
         """
         cluster = self.cluster
-        cluster.populate(4).start(wait_for_binary_proto=True)
+        cluster.populate(4).start()
         node1, node2, node3, node4 = cluster.nodelist()
 
         logger.debug("Inserting data with stress")
diff --git a/repair_tests/repair_test.py b/repair_tests/repair_test.py
index 5faaf76..7d1949a 100644
--- a/repair_tests/repair_test.py
+++ b/repair_tests/repair_test.py
@@ -95,7 +95,7 @@ class BaseRepairTest(Tester):
         cluster.set_configuration_options(values={'hinted_handoff_enabled': False})
         cluster.set_batch_commitlog(enabled=True)
         logger.debug("Starting cluster..")
-        cluster.populate(3).start(wait_for_binary_proto=True, wait_other_notice=True)
+        cluster.populate(3).start()
         node1, node2, node3 = cluster.nodelist()
 
         session = self.patient_cql_connection(node1, retry_policy=FlakyRetryPolicy(max_retries=15))
@@ -172,7 +172,7 @@ class TestRepair(BaseRepairTest):
         """
         cluster = self.cluster
         logger.debug("Starting cluster..")
-        cluster.populate([2, 2]).start(wait_for_binary_proto=True)
+        cluster.populate([2, 2]).start()
         node1_1, node2_1, node1_2, node2_2 = cluster.nodelist()
         node1_1.stress(stress_options=['write', 'n=50K', 'no-warmup', 'cl=ONE', '-schema', 'replication(factor=4)', '-rate', 'threads=50'])
         node1_1.nodetool("repair -local keyspace1 standard1")
@@ -197,7 +197,7 @@ class TestRepair(BaseRepairTest):
         self.fixture_dtest_setup.ignore_log_patterns = [r'Unknown keyspace/cf pair']
         cluster = self.cluster
         logger.debug('Starting nodes')
-        cluster.populate(2).start(wait_for_binary_proto=True)
+        cluster.populate(2).start()
         node1, _ = cluster.nodelist()
         logger.debug('Creating keyspace and tables')
         node1.stress(stress_options=['write', 'n=1', 'no-warmup',
@@ -241,7 +241,7 @@ class TestRepair(BaseRepairTest):
         """
         cluster = self.cluster
         logger.debug("Starting cluster..")
-        cluster.populate([2, 2]).start(wait_for_binary_proto=True)
+        cluster.populate([2, 2]).start()
         node1_1, node2_1, node1_2, node2_2 = cluster.nodelist()
         node1_1.stress(stress_options=['write', 'n=100K', 'no-warmup', 'cl=ONE', '-schema', 'replication(factor=4)', '-rate', 'threads=50'])
         node1_1.nodetool("repair -hosts 127.0.0.1,127.0.0.2,127.0.0.3,127.0.0.4 keyspace1 standard1")
@@ -262,7 +262,7 @@ class TestRepair(BaseRepairTest):
         """
         cluster = self.cluster
         logger.debug("Starting cluster..")
-        cluster.populate(3).start(wait_for_binary_proto=True)
+        cluster.populate(3).start()
         node1, node2, node3 = cluster.nodelist()
         node1.stress(stress_options=['write', 'n=50K', 'no-warmup', 'cl=ONE', '-schema', 'replication(factor=3)', '-rate', 'threads=50'])
         node1.nodetool("repair -st 0 -et 1000 keyspace1 standard1")
@@ -304,7 +304,7 @@ class TestRepair(BaseRepairTest):
         logger.debug("Starting cluster..")
         # disable JBOD conf since the test expects sstables to be on the same disk
         cluster.set_datadir_count(1)
-        cluster.populate(3).start(wait_for_binary_proto=True)
+        cluster.populate(3).start()
         node1, node2, node3 = cluster.nodelist()
         # we use RF to make sure to cover only a set of sub-ranges when doing -full -pr
         node1.stress(stress_options=['write', 'n=50K', 'no-warmup', 'cl=ONE', '-schema', 'replication(factor=2)', '-rate', 'threads=50'])
@@ -337,7 +337,7 @@ class TestRepair(BaseRepairTest):
         """
         cluster = self.cluster
         logger.debug("Starting cluster..")
-        cluster.populate([2, 2]).start(wait_for_binary_proto=True)
+        cluster.populate([2, 2]).start()
         node1_1, node2_1, node1_2, node2_2 = cluster.nodelist()
         node1_1.stress(stress_options=['write', 'n=50K', 'no-warmup', 'cl=ONE', '-schema', 'replication(factor=4)'])
         node1_1.nodetool("repair keyspace1 standard1")
@@ -709,7 +709,7 @@ class TestRepair(BaseRepairTest):
         cluster.set_batch_commitlog(enabled=True)
         logger.debug("Starting cluster..")
         # populate 2 nodes in dc1, and one node each in dc2 and dc3
-        cluster.populate([2, 1, 1]).start(wait_for_binary_proto=True)
+        cluster.populate([2, 1, 1]).start()
 
         node1, node2, node3, node4 = cluster.nodelist()
         session = self.patient_cql_connection(node1)
@@ -758,7 +758,7 @@ class TestRepair(BaseRepairTest):
 
         cluster = self.cluster
         logger.debug("Starting cluster..")
-        cluster.populate([3]).start(wait_for_binary_proto=True)
+        cluster.populate([3]).start()
         node1, node2, node3 = cluster.nodelist()
         node1.stress(stress_options=['write', 'n=10k', 'no-warmup', 'cl=ONE', '-schema', 'replication(factor=3)', '-rate', 'threads=50'])
 
@@ -795,7 +795,7 @@ class TestRepair(BaseRepairTest):
         cluster.set_configuration_options(values={'hinted_handoff_enabled': False})
         cluster.set_batch_commitlog(enabled=True)
         logger.debug("Starting cluster..")
-        cluster.populate(3).start(wait_for_binary_proto=True)
+        cluster.populate(3).start()
 
         node1, node2, node3 = cluster.nodelist()
 
@@ -819,7 +819,7 @@ class TestRepair(BaseRepairTest):
         cluster.set_configuration_options(values={'hinted_handoff_enabled': False})
         cluster.set_batch_commitlog(enabled=True)
         logger.debug("Starting cluster..")
-        cluster.populate(3).start(wait_for_binary_proto=True)
+        cluster.populate(3).start()
 
         node1, node2, node3 = cluster.nodelist()
 
@@ -868,7 +868,7 @@ class TestRepair(BaseRepairTest):
         cluster.set_configuration_options(values={'hinted_handoff_enabled': False})
         cluster.set_batch_commitlog(enabled=True)
         logger.debug("Starting cluster..")
-        cluster.populate(3).start(wait_for_binary_proto=True)
+        cluster.populate(3).start()
 
         node1, node2, node3 = cluster.nodelist()
 
@@ -892,7 +892,7 @@ class TestRepair(BaseRepairTest):
         cluster.set_configuration_options(values={'hinted_handoff_enabled': False})
         cluster.set_batch_commitlog(enabled=True)
         logger.debug("Starting cluster..")
-        cluster.populate(3).start(wait_for_binary_proto=True)
+        cluster.populate(3).start()
 
         node1, node2, node3 = cluster.nodelist()
 
@@ -966,7 +966,7 @@ class TestRepair(BaseRepairTest):
         cluster.set_configuration_options(values={'hinted_handoff_enabled': False})
         cluster.set_batch_commitlog(enabled=True)
         logger.debug("Starting cluster..")
-        cluster.populate(3).start(wait_for_binary_proto=True)
+        cluster.populate(3).start()
 
         node1, node2, node3 = cluster.nodelist()
 
@@ -1009,7 +1009,7 @@ class TestRepair(BaseRepairTest):
         cluster.set_configuration_options(values={'hinted_handoff_enabled': False})
         cluster.set_batch_commitlog(enabled=True)
         logger.debug("Starting cluster..")
-        cluster.populate(3).start(wait_for_binary_proto=True)
+        cluster.populate(3).start()
 
         node1, node2, node3 = cluster.nodelist()
 
@@ -1050,7 +1050,7 @@ class TestRepair(BaseRepairTest):
         """
         cluster = self.cluster
         cluster.set_configuration_options(values={'hinted_handoff_enabled': False})
-        cluster.populate(3).start(wait_for_binary_proto=True)
+        cluster.populate(3).start()
         node1, node2, node3 = cluster.nodelist()
         node2.stop(wait_other_notice=True)
         node1.stress(['write', 'n=1M', 'no-warmup', '-schema', 'replication(factor=3)', '-rate', 'threads=30'])
@@ -1077,7 +1077,7 @@ class TestRepair(BaseRepairTest):
         """
         cluster = self.cluster
         cluster.set_configuration_options(values={'hinted_handoff_enabled': False})
-        cluster.populate(2).start(wait_for_binary_proto=True)
+        cluster.populate(2).start()
         node1, node2 = cluster.nodelist()
         node2.stop(wait_other_notice=True)
         profile_path = os.path.join(os.getcwd(), 'stress_profiles/repair_wide_rows.yaml')
@@ -1095,7 +1095,7 @@ class TestRepair(BaseRepairTest):
         """
         cluster = self.cluster
         cluster.set_configuration_options(values={'hinted_handoff_enabled': False})
-        cluster.populate(3).start(wait_for_binary_proto=True)
+        cluster.populate(3).start()
         node1, node2, node3 = cluster.nodelist()
         node1.stress(['write', 'n=100k', '-schema', 'replication(factor=3)', '-rate', 'threads=30'])
 
@@ -1193,8 +1193,8 @@ class TestRepair(BaseRepairTest):
         node_to_kill.import_config_files()
 
         logger.debug("Starting cluster..")
-        cluster.start(wait_other_notice=True, jvm_args=['-Djdk.attach.allowAttachSelf=true'])
-        # cluster.start(wait_other_notice=True)
+        cluster.start(jvm_args=['-Djdk.attach.allowAttachSelf=true'])
+        # cluster.start()
 
         logger.debug("stopping node3")
         node3.stop(gently=False, wait_other_notice=True)
@@ -1276,7 +1276,7 @@ class TestRepairDataSystemTable(Tester):
         a 5-node cluster, then inserting 5000 values with RF=3.
         """
         self.cluster = fixture_dtest_setup.cluster
-        self.cluster.populate(5).start(wait_for_binary_proto=True)
+        self.cluster.populate(5).start()
         self.node1 = self.cluster.nodelist()[0]
         self.session = fixture_dtest_setup.patient_cql_connection(self.node1)
 
diff --git a/replication_test.py b/replication_test.py
index b199551..5ea07b2 100644
--- a/replication_test.py
+++ b/replication_test.py
@@ -204,7 +204,7 @@ class TestReplication(Tester):
         """
         Test the SimpleStrategy on a 3 node cluster
         """
-        self.cluster.populate(3).start(wait_for_binary_proto=True, wait_other_notice=True)
+        self.cluster.populate(3).start()
         node1 = self.cluster.nodelist()[0]
         session = self.patient_exclusive_cql_connection(node1, consistency_level=ConsistencyLevel.ALL)
         session.max_trace_wait = 120
@@ -240,7 +240,7 @@ class TestReplication(Tester):
         """
         Test the NetworkTopologyStrategy on a 2DC 3:3 node cluster
         """
-        self.cluster.populate([3, 3]).start(wait_for_binary_proto=True, wait_other_notice=True)
+        self.cluster.populate([3, 3]).start()
 
         node1 = self.cluster.nodelist()[0]
         ip_nodes = dict((node.address(), node) for node in self.cluster.nodelist())
@@ -541,7 +541,7 @@ class TestSnitchConfigurationUpdate(Tester):
                 for line in snitch_lines_before(i, node):
                     topo_file.write(line + os.linesep)
 
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
 
         session = self.patient_cql_connection(cluster.nodelist()[0])
 
@@ -697,7 +697,7 @@ class TestSnitchConfigurationUpdate(Tester):
                 for line in snitch_lines_before:
                     topo_file.write(line + os.linesep)
 
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
 
         # check racks are as specified
         self.wait_for_nodes_on_racks(cluster.nodelist(), racks)
@@ -742,7 +742,7 @@ class TestSnitchConfigurationUpdate(Tester):
             topo_file.write("dc=dc9" + os.linesep)
             topo_file.write("rack=rack1" + os.linesep)
 
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
 
         node.stop()
 
diff --git a/secondary_indexes_test.py b/secondary_indexes_test.py
index a447bf9..7b9459b 100644
--- a/secondary_indexes_test.py
+++ b/secondary_indexes_test.py
@@ -333,7 +333,7 @@ class TestSecondaryIndexes(Tester):
         """
 
         cluster = self.cluster
-        cluster.populate(1, install_byteman=True).start(wait_for_binary_proto=True)
+        cluster.populate(1, install_byteman=True).start()
         node = cluster.nodelist()[0]
 
         session = self.patient_cql_connection(node)
@@ -461,7 +461,7 @@ class TestSecondaryIndexes(Tester):
         """
 
         cluster = self.cluster
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         node = cluster.nodelist()[0]
 
         session = self.patient_cql_connection(node)
@@ -1178,7 +1178,7 @@ class TestPreJoinCallback(Tester):
         cluster.populate(1)
         node1 = cluster.nodelist()[0]
         node1.set_configuration_options(values={'initial_token': tokens[0]})
-        cluster.start(wait_other_notice=True)
+        cluster.start()
 
         # Create a table with 2i
         session = self.patient_cql_connection(node1)
diff --git a/snapshot_test.py b/snapshot_test.py
index 036648e..7328c18 100644
--- a/snapshot_test.py
+++ b/snapshot_test.py
@@ -402,7 +402,7 @@ class TestArchiveCommitlog(SnapshotTester):
             for snapshot_dir in snapshot_dirs:
                 self.restore_snapshot(snapshot_dir, node1, 'ks', 'cf', 'basic')
 
-            cluster.start(wait_for_binary_proto=True)
+            cluster.start()
 
             session = self.patient_cql_connection(node1)
             node1.nodetool('refresh ks cf')
@@ -487,7 +487,7 @@ class TestArchiveCommitlog(SnapshotTester):
                          (r'^restore_directories=.*$', 'restore_directories={tmp_commitlog}'.format(
                           tmp_commitlog=tmp_commitlog))])
 
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
 
         logger.debug("Creating initial connection")
         session = self.patient_cql_connection(node1)
diff --git a/snitch_test.py b/snitch_test.py
index 03b09c8..33bae7a 100644
--- a/snitch_test.py
+++ b/snitch_test.py
@@ -164,7 +164,7 @@ class TestDynamicEndpointSnitch(Tester):
         # Delay reads on the degraded node by 50 milliseconds
         degraded_node.start(jvm_args=['-Dcassandra.test.read_iteration_delay_ms=50',
                                       '-Dcassandra.allow_unsafe_join=true'])
-        cluster.start(wait_for_binary_proto=30, wait_other_notice=True)
+        cluster.start()
 
         des = make_mbean('db', type='DynamicEndpointSnitch')
         read_stage = make_mbean('metrics', type='ThreadPools', path='request',
diff --git a/sstable_generation_loading_test.py b/sstable_generation_loading_test.py
index 7b0a41d..cab39ec 100644
--- a/sstable_generation_loading_test.py
+++ b/sstable_generation_loading_test.py
@@ -301,7 +301,7 @@ class TestSSTableGenerationAndLoading(TestBaseSStableLoader):
         @jira_ticket CASSANDRA-343
         """
         cluster = self.cluster
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         node1 = cluster.nodelist()[0]
 
         # Makinge sure the cluster is ready to accept the subsequent
@@ -363,7 +363,7 @@ class TestSSTableGenerationAndLoading(TestBaseSStableLoader):
             session.execute("CREATE INDEX idx ON k.t(v)")
 
         cluster = self.cluster
-        cluster.populate(1, install_byteman=True).start(wait_for_binary_proto=True)
+        cluster.populate(1, install_byteman=True).start()
         node = cluster.nodelist()[0]
 
         session = self.patient_cql_connection(node)
diff --git a/sstablesplit_test.py b/sstablesplit_test.py
index 4d86195..1ee3fec 100644
--- a/sstablesplit_test.py
+++ b/sstablesplit_test.py
@@ -18,7 +18,7 @@ class TestSSTableSplit(Tester):
         after carrying out these operations.
         """
         cluster = self.cluster
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         node = cluster.nodelist()[0]
         version = cluster.version()
 
@@ -91,7 +91,7 @@ class TestSSTableSplit(Tester):
         Check that sstablesplit doesn't crash when splitting a single sstable at the time.
         """
         cluster = self.cluster
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         node = cluster.nodelist()[0]
 
         logger.debug("Run stress to insert data")
diff --git a/sstableutil_test.py b/sstableutil_test.py
index 30584c1..9ac93c8 100644
--- a/sstableutil_test.py
+++ b/sstableutil_test.py
@@ -37,7 +37,7 @@ class TestSSTableUtil(Tester):
         Check that we can list sstable files after a successful compaction (no temporary sstable files)
         """
         cluster = self.cluster
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         node = cluster.nodelist()[0]
 
         self._create_data(node, KeyspaceName, TableName, 100000)
@@ -58,7 +58,7 @@ class TestSSTableUtil(Tester):
         """
         log_file_name = 'debug.log'
         cluster = self.cluster
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         node = cluster.nodelist()[0]
 
         numrecords = 250000
diff --git a/streaming_test.py b/streaming_test.py
index 3410920..2d3494a 100644
--- a/streaming_test.py
+++ b/streaming_test.py
@@ -60,7 +60,7 @@ class TestStreaming(Tester):
         for i in range(0, len(nodes)):
             nodes[i].set_configuration_options(values={'initial_token': tokens[i]})
 
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
 
         session = self.patient_cql_connection(nodes[0])
 
diff --git a/stress_tool_test.py b/stress_tool_test.py
index 7d15c4d..7699974 100644
--- a/stress_tool_test.py
+++ b/stress_tool_test.py
@@ -43,7 +43,7 @@ class TestStressSparsenessRatio(Tester):
         A parameterized test for the `row-population-ratio` parameter to
         `cassandra-stress`.
         """
-        self.cluster.populate(1).start(wait_for_binary_proto=True)
+        self.cluster.populate(1).start()
         node = self.cluster.nodelist()[0]
         node.stress(['write', 'n=1000', 'no-warmup', '-rate', 'threads=50', '-col', 'n=FIXED(50)',
                      '-insert', 'row-population-ratio={ratio_spec}'.format(ratio_spec=ratio_spec)])
@@ -65,7 +65,7 @@ class TestStressWrite(Tester):
         @jira_ticket CASSANDRA-14890
         A simple write stress test should be done very quickly
         """
-        self.cluster.populate(1).start(wait_for_binary_proto=True)
+        self.cluster.populate(1).start()
         node = self.cluster.nodelist()[0]
         node.stress(['write', 'err<0.9', 'n>1', '-rate', 'threads=1'])
         out, err, _ = node.run_cqlsh('describe table keyspace1.standard1')
diff --git a/thrift_hsha_test.py b/thrift_hsha_test.py
index d1d84f8..3d1e364 100644
--- a/thrift_hsha_test.py
+++ b/thrift_hsha_test.py
@@ -49,7 +49,7 @@ class TestThriftHSHA(Tester):
         cluster.populate(1)
         (node1,) = cluster.nodelist()
         remove_perf_disable_shared_mem(node1)
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
 
         session = self.patient_cql_connection(node1)
         create_ks(session, 'test', 1)
diff --git a/thrift_test.py b/thrift_test.py
index 668b6cf..55913ad 100644
--- a/thrift_test.py
+++ b/thrift_test.py
@@ -84,8 +84,7 @@ class TestThrift(Tester):
 
         # CASSANDRA-14092 - prevent max ttl tests from failing
         fixture_dtest_setup.cluster.start(jvm_args=['-Dcassandra.expiration_date_overflow_policy=CAP',
-                                                    '-Dcassandra.expiration_overflow_warning_interval_minutes=0'],
-                                          wait_for_binary_proto=True)
+                                                    '-Dcassandra.expiration_overflow_warning_interval_minutes=0'])
         fixture_dtest_setup.cluster.nodelist()[0].watch_log_for("Listening for thrift clients")  # Wait for the thrift port to open
         time.sleep(0.1)
         # this is ugly, but the whole test module is written against a global client
diff --git a/token_generator_test.py b/token_generator_test.py
index 3d1afbd..04c0e65 100644
--- a/token_generator_test.py
+++ b/token_generator_test.py
@@ -84,7 +84,7 @@ class TestTokenGenerator(Tester):
             cluster._config_options.__delitem__('initial_token')
 
         assert not cluster.nodelist(), "nodelist() already initialized"
-        cluster.populate(nodes, use_vnodes=False, tokens=generated_tokens[0]).start(wait_for_binary_proto=True)
+        cluster.populate(nodes, use_vnodes=False, tokens=generated_tokens[0]).start()
         time.sleep(0.2)
 
         node = cluster.nodelist()[0]
diff --git a/topology_test.py b/topology_test.py
index bb0e809..6da9b7c 100644
--- a/topology_test.py
+++ b/topology_test.py
@@ -134,7 +134,7 @@ class TestTopology(Tester):
         """
         cluster = self.cluster
         cluster.populate(3)
-        cluster.start(wait_for_binary_proto=True, jvm_args=["-Dcassandra.size_recorder_interval=1"])
+        cluster.start(jvm_args=["-Dcassandra.size_recorder_interval=1"])
         node1, node2, node3 = cluster.nodelist()
 
         session = self.patient_cql_connection(node1)
@@ -161,7 +161,7 @@ class TestTopology(Tester):
         """
         cluster = self.cluster
         cluster.set_configuration_options(values={'stream_throughput_outbound_megabits_per_sec': 1})
-        cluster.populate(2).start(wait_other_notice=True)
+        cluster.populate(2).start()
         node1, node2 = cluster.nodelist()
 
         session = self.patient_cql_connection(node2)
@@ -206,7 +206,7 @@ class TestTopology(Tester):
                                                         r'Remote peer 127.0.0.2:7000 failed stream session']
         cluster = self.cluster
         cluster.set_configuration_options(values={'stream_throughput_outbound_megabits_per_sec': 1})
-        cluster.populate(3, install_byteman=True).start(wait_other_notice=True)
+        cluster.populate(3, install_byteman=True).start()
         node1, node2, node3 = cluster.nodelist()
 
         session = self.patient_cql_connection(node2)
@@ -381,7 +381,7 @@ class TestTopology(Tester):
         self.fixture_dtest_setup.ignore_log_patterns = list(self.fixture_dtest_setup.ignore_log_patterns) + [
             rejoin_err]
 
-        self.cluster.populate(3).start(wait_for_binary_proto=True)
+        self.cluster.populate(3).start()
         node1, node2, node3 = self.cluster.nodelist()
 
         logger.debug('decommissioning...')
@@ -420,7 +420,7 @@ class TestTopology(Tester):
         """
         cluster = self.cluster
         self.fixture_dtest_setup.ignore_log_patterns = [r'Streaming error occurred', 'Stream failed']
-        cluster.populate(3).start(wait_other_notice=True)
+        cluster.populate(3).start()
 
         node1, node2 = cluster.nodelist()[0:2]
 
@@ -457,7 +457,7 @@ class TestTopology(Tester):
         @expected_errors ToolError when # nodes will drop below configured replicas in NTS/SimpleStrategy
         """
         cluster = self.cluster
-        cluster.populate([2, 2]).start(wait_for_binary_proto=True)
+        cluster.populate([2, 2]).start()
         node1, node2, node3, node4 = self.cluster.nodelist()
         session = self.patient_cql_connection(node2)
         session.execute("ALTER KEYSPACE system_distributed WITH REPLICATION = {'class':'SimpleStrategy', 'replication_factor':'2'};")
diff --git a/transient_replication_ring_test.py b/transient_replication_ring_test.py
index a519c62..8a59fd1 100644
--- a/transient_replication_ring_test.py
+++ b/transient_replication_ring_test.py
@@ -111,7 +111,7 @@ class TestTransientReplicationRing(Tester):
         print(self.cluster.get_install_dir())
         self.cluster.populate(3, tokens=self.tokens, debug=True, install_byteman=True)
         # self.cluster.populate(3, debug=True, install_byteman=True)
-        self.cluster.start(wait_other_notice=True, wait_for_binary_proto=True, jvm_args=['-Dcassandra.enable_nodelocal_queries=true'])
+        self.cluster.start(jvm_args=['-Dcassandra.enable_nodelocal_queries=true'])
 
         # enable shared memory
         for node in self.cluster.nodelist():
diff --git a/transient_replication_test.py b/transient_replication_test.py
index e04162f..87b798d 100644
--- a/transient_replication_test.py
+++ b/transient_replication_test.py
@@ -213,7 +213,7 @@ class TransientReplicationBase(Tester):
                                                        'enable_transient_replication': True,
                                                        'dynamic_snitch': False})
         self.populate()
-        self.cluster.start(wait_other_notice=True, wait_for_binary_proto=True)
+        self.cluster.start()
 
         self.nodes = [patch_start(node) for node in self.cluster.nodelist()]
         self.set_nodes()
diff --git a/ttl_test.py b/ttl_test.py
index d89ca6a..40c2963 100644
--- a/ttl_test.py
+++ b/ttl_test.py
@@ -373,7 +373,7 @@ class TestTTL(Tester):
 
         # Restart node with expiration_date_overflow_policy
         self.cluster.stop()
-        self.cluster.start(jvm_args=['-Dcassandra.expiration_date_overflow_policy={}'.format(policy)], wait_for_binary_proto=True)
+        self.cluster.start(jvm_args=['-Dcassandra.expiration_date_overflow_policy={}'.format(policy)])
         self.session1 = self.patient_cql_connection(self.cluster.nodelist()[0])
         self.session1.execute("USE ks;")
 
@@ -569,7 +569,7 @@ class TestRecoverNegativeExpirationDate(TestHelper):
         cluster = self.cluster
         if self.cluster.version() >= '4':
             cluster.set_configuration_options(values={'corrupted_tombstone_strategy': 'disabled'})
-        cluster.populate(1).start(wait_for_binary_proto=True)
+        cluster.populate(1).start()
         [node] = cluster.nodelist()
 
         session = self.patient_cql_connection(node)
@@ -611,7 +611,7 @@ class TestRecoverNegativeExpirationDate(TestHelper):
         logger.debug("Executed offline scrub on {}", str(scrubbed_sstables))
 
         logger.debug("Starting node again")
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
         session = self.patient_cql_connection(node)
         session.execute("USE ks;")
 
diff --git a/upgrade_tests/compatibility_flag_test.py b/upgrade_tests/compatibility_flag_test.py
index 3711841..f778cea 100644
--- a/upgrade_tests/compatibility_flag_test.py
+++ b/upgrade_tests/compatibility_flag_test.py
@@ -26,7 +26,7 @@ class TestCompatibilityFlag(Tester):
         node1, node2 = cluster.nodelist()
         cluster.set_install_dir(version=from_version)
         self.fixture_dtest_setup.reinitialize_cluster_for_different_version()
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
 
         node1.drain()
         node1.watch_log_for("DRAINED")
@@ -48,7 +48,7 @@ class TestCompatibilityFlag(Tester):
         node1, node2 = cluster.nodelist()
         cluster.set_install_dir(version=from_version)
         self.fixture_dtest_setup.reinitialize_cluster_for_different_version()
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
 
         node1.drain()
         node1.watch_log_for("DRAINED")
diff --git a/upgrade_tests/cql_tests.py b/upgrade_tests/cql_tests.py
index 4bb9630..1ce95ba 100644
--- a/upgrade_tests/cql_tests.py
+++ b/upgrade_tests/cql_tests.py
@@ -5458,7 +5458,7 @@ class TestCQL(UpgradeTester):
 
         self.cluster.stop()
         time.sleep(0.5)
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
         time.sleep(0.5)
 
         for is_upgraded, cursor in self.do_upgrade(cursor):
diff --git a/upgrade_tests/upgrade_base.py b/upgrade_tests/upgrade_base.py
index e4480be..a1426eb 100644
--- a/upgrade_tests/upgrade_base.py
+++ b/upgrade_tests/upgrade_base.py
@@ -129,7 +129,7 @@ class UpgradeTester(Tester, metaclass=ABCMeta):
         if self.fixture_dtest_setup.enable_for_jolokia:
             remove_perf_disable_shared_mem(node1)
 
-        cluster.start(wait_for_binary_proto=True)
+        cluster.start()
 
         node1 = cluster.nodelist()[0]
         time.sleep(0.2)
diff --git a/user_types_test.py b/user_types_test.py
index b172a47..dcbb833 100644
--- a/user_types_test.py
+++ b/user_types_test.py
@@ -647,7 +647,7 @@ class TestUserTypes(Tester):
         # make sure we can define a table with a user type as a clustering column
         # and do a basic insert/query of data in that table.
         cluster = self.cluster
-        cluster.populate(3).start(wait_for_binary_proto=True)
+        cluster.populate(3).start()
         node1, node2, node3 = cluster.nodelist()
         session = self.patient_cql_connection(node1, consistency_level=ConsistencyLevel.LOCAL_QUORUM)
         create_ks(session, 'user_type_pkeys', 2)
diff --git a/write_failures_test.py b/write_failures_test.py
index 3ae721b..5c61bea 100644
--- a/write_failures_test.py
+++ b/write_failures_test.py
@@ -48,7 +48,7 @@ class TestWriteFailures(Tester):
         if start_rpc:
             self.cluster.set_configuration_options(values={'start_rpc': True})
 
-        self.cluster.start(wait_for_binary_proto=True)
+        self.cluster.start()
         self.nodes = list(self.cluster.nodes.values())
 
         session = self.patient_exclusive_cql_connection(self.nodes[0], protocol_version=self.protocol_version)


---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@cassandra.apache.org
For additional commands, e-mail: commits-help@cassandra.apache.org