Resolving query timeouts on restarted nodes

When restarting nodes with large indexes (hundreds of megabytes), initial queries might timeout due to the time it takes to build the token range filter queries.

Procedure

  1. To workaround timeouts:

  2. Run with a replication factor greater than 1 to ensure that replicas are always available.

  3. Configure the dse.yaml settings for enable_health_based_routing and uptime_ramp_up_period_seconds to be larger than the amount of time it takes for the first query to answer. 1 hour is usually enough.

  4. After restarting the node, issue several match all queries. For example, q=: to warm up the filters.

  5. If you’re using the Java Driver, create an ad-hoc session with only the node to warm up in the white list.

    Issuing many queries increase the chances that all token ranges are used.

Results

After the uptime ramp-up period, the node starts to be hit by distributed queries. The filters are warmed up already and timeouts should not occur.

Was this helpful?

Give Feedback

How can we improve the documentation?

© 2024 DataStax | Privacy policy | Terms of use

Apache, Apache Cassandra, Cassandra, Apache Tomcat, Tomcat, Apache Lucene, Apache Solr, Apache Hadoop, Hadoop, Apache Pulsar, Pulsar, Apache Spark, Spark, Apache TinkerPop, TinkerPop, Apache Kafka and Kafka are either registered trademarks or trademarks of the Apache Software Foundation or its subsidiaries in Canada, the United States and/or other countries. Kubernetes is the registered trademark of the Linux Foundation.

General Inquiries: +1 (650) 389-6000, info@datastax.com