Frequently Asked Questions

How do I use this driver as a drop-in replacement for cassandra-driver-core when it is a dependency of another project?

It is not uncommon for users to use libraries that depend on DataStax Java Driver for Apache Cassandra®. Such libraries include spring-data-cassandra, phantom, and quill among others.

Since the Java driver for DataStax Enterprise is a drop-in replacement, you can declare it explicitly with your dependency management tool while excluding cassandra-driver-core as a dependency for the library you are using which currently depends on it.

Ensure that the version of Java Driver for Apache Cassandra depended on by the third-party library is compatible with your target DataStax Enterprise Driver version. Use the following table to identify the compatible versions:

Cassandra driver version DSE driver version
3.1.x 1.1.x
3.2.x 1.2.x

To accomplish this with maven using spring-data-cassandra as an example, you may do the following:



Alternatively, using gradle:

dependencies {
  compile ('$springDataVersion') {
    exclude group: 'com.datastax.cassandra', module: 'cassandra-driver-core'
  compile 'com.datastax.dse:dse-java-driver-core:$dseDriverVersion'

Note that depending on how these libraries integrate with the driver, certain DSE specific features may not be directly available.

How do I implement paging?

When using native protocol version 2 or higher, the driver automatically pages large result sets under the hood. You can also save the paging state to resume iteration later. See this page for more information.

Native protocol v1 does not support paging, but you can emulate it in CQL with LIMIT and the token() function. See this conversation on the mailing list.

Can I check if a conditional statement (lightweight transaction) was successful?

When executing a conditional statement, the ResultSet will contain a single Row with a column named “applied” of type boolean. This tells whether the conditional statement was successful or not.

The driver provides a convenience method wasApplied to check this on the result set directly:

ResultSet rset = session.execute(conditionalStatement);

You may also inspect the value yourself:

ResultSet rset = session.execute(conditionalStatement);
Row row =;
row.getBool(0);       // this is equivalent row.getBool("applied")

Note that, unlike manual inspection, wasApplied does not consume the first row.

What is a parameterized statement and how can I use it?

Starting with Cassandra 2.0, normal statements (that is non-prepared statements) do not need to concatenate parameter values inside a query string. Instead you can use ? markers and provide the values separately:

session.execute( "INSERT INTO contacts (email, firstname, lastname)
      VALUES (?, ?, ?)", "", "Barney", "Barton");

See Simple statements for more information.

Does a parameterized statement escape parameters?

A parameterized statement sends the values of parameters separate from the query (similar to the way a prepared statement does) as bytes so there is no need to escape parameters.

What’s the difference between a parameterized statement and a Prepared statement?

The only similarity between a parameterized statement and a prepared statement is in the way that the parameters are sent. The difference is that a prepared statement:

  • is already known on the cluster side (it has been compiled and there is an execution plan available for it) which leads to better performance
  • sends only the statement id and its parameters (thus reducing the amount of data sent to the cluster)

See Prepared statements for more information.

Can I combine PreparedStatements and normal statements in a batch?

Yes. A batch can include both bound statements and simple statements:

PreparedStatement ps = session.prepare( "INSERT INTO contacts (email, firstname, lastname)
      VALUES (?, ?, ?)");
BatchStatement batch = new BatchStatement();
// here's a simple statement
batch.add(new SimpleStatement( "INSERT INTO contacts (email, firstname, lastname) VALUES (?, ?, ?)", ...));

Why do my ‘SELECT *’ PreparedStatement-based queries stop working after a schema change?

Both the driver and Cassandra maintain a mapping of PreparedStatement queries to their metadata. When a change is made to a table, such as a column being added or dropped, there is currently no mechanism for Cassandra to invalidate the existing metadata. Because of this, the driver is not able to properly react to these changes and will improperly read rows after a schema change is made.

See Prepared statements for more information.

Can I get the raw bytes of a text column?

If you need to access the raw bytes of a text column, call the Row.getBytesUnsafe("columnName") method.

Trying to use Row.getBytes("columnName") for the same purpose results in an exception, as the getBytes method can only be used if the column has the CQL type BLOB.

How do I increment counters with QueryBuilder?

Considering the following query:

UPDATE clickstream SET clicks = clicks + 1 WHERE userid = id;

To do this using QueryBuilder:

Statement query = QueryBuilder.update("clickstream")
                              .with(incr("clicks", 1)) // Use incr for counters
                              .where(eq("userid", id));

Is there a way to control the batch size of the results returned from a query?

Use the setFetchSize() method on your Statement object. The fetch size controls how many resulting rows are retrieved simultaneously (the goal being to avoid loading too many results in memory for queries yielding large result sets).

Keep in mind that if your code iterates the ResultSet entirely, the driver may run additional background queries to fetch the rest of the data. The fetch size only affects what is retrieved at a time, not the overall number of rows.

See Paging for more information.

What’s the difference between using setFetchSize() and LIMIT?

Basically, LIMIT controls the maximum number of results returned by the query, while the setFetchSize() method controls the amount of data transferred at a time.

For example, if you limit is 30 and your fetch size is 10, the ResultSet will contain 30 rows, but under the hood the driver will perform 3 requests that will transfer 10 rows each.

See Paging for more information.

I’m reading a BLOB column and the driver returns incorrect data.

Check your code to ensure that you read the returned ByteBuffer correctly. ByteBuffer is a very error-prone API, and we’ve had many reports where the problem turned out to be in user code.

See in the driver-examples module for some examples and explanations.

How do I use the driver in an OSGi application?

Read our OSGi-specific FAQ section to find out.

Why am I seeing messages about tombstone_warn_threshold or tombstone_fail_threshold being exceeded in my Cassandra logs?

Applications which use the object mapper or set null values in their statements may observe that many tombstones are being stored in their tables which subsequently may lead to poor query performance, failed queries, or columns being mysteriously deleted.

This is caused by INSERT/UPDATE statements containing null values for columns that a user does not intend to change. Common circumstances around this come from using the object mapper or writing your own persistence layer and attempting to reuse the same PreparedStatement for inserting data, even with partial updates.

Prior to cassandra 2.2, there was no means of reusing the same PreparedStatement for making partial updates to different columns.

For example, given the following code:

PreparedStatement prepared = session.prepare("INSERT INTO contacts (email, firstname, lastname) VALUES (?, ?, ?)");
BoundStatement bound = prepared.bind();
bound.set("email", "");
bound.set("firstname", "Barney");
// creates a tombstone!!
bound.set("lastname", null);

If one wanted to use this query to update only firstname this would not be achievable without binding the lastname parameter to null. This would have an undesired side effect of creating a tombstone for lastname and thus to the user giving the impression that lastname was deleted.

In cassandra 2.2 and later with protocol v4, bind parameters (?) can optionally be left unset (CASSANDRA-7304):

 PreparedStatement prepared = session.prepare("INSERT INTO contacts (email, firstname, lastname) VALUES (?, ?, ?)");
 BoundStatement bound = prepared.bind();
 bound.set("email", "");
 bound.set("firstname", "Barney");
 // lastname is left unset.

See Parameters and Binding for more details about unset parameters.

Another possible root cause for this is using the object mapper and leaving fields set to null. This also causes tombstones to be inserted unless setting saveNullFields option to false. See Mapper options for more details.

Why am I encountering an ‘illegal cyclic reference’ error when using the driver with Scala?

There is a known issue with the Scala compiler (SI-3809) that prevents Scala code from compiling when using DseCluster.

An example of such a compiler error:

src/main/scala/app/App.scala:18: illegal cyclic reference involving class Cluster 

To work around this issue, it is recommended to add -Ybreak-cycles to scalac‘s compiler arguments. To accomplish this using sbt, add the following to your sbt project file:

scalacOptions += "-Ybreak-cycles"

I’m using the Apache TinkerPop™ integration layer and cannot compile my application because of an unresolved dependency com.github.jeremyh:jBCrypt:jar:jbcrypt-0.4

This is a known problem. See our documentation on the Apache TinkerPop™ integration layer for an explanation and possible workarounds.

I am getting a “No such property: g for class: ScriptXXXX” error message when using executeGraph().

Remember that executing a query with the String execution method for DSE Graph will have the DSE Graph server interpret this String as a Groovy script. On the server side, to ease the use of DSE Graph, some variables are predefined before the execution of each script.

When no graph name is defined on the request, the DSE Graph server will predefine the Graph system management variable, named system. Hence if no graph name is defined on a request, only the system variable will be accessible.

On the other hand, if a graph name is defined for the request, the DSE Graph server will automatically predefine a GraphTraversalSource variable named g bound to the graph specified in the graph name, so that users can easily create and execute traversals with it. If no graph name is specified for a query and you try to use the variable g, the error message mentioned above will be thrown.

A graph name can be specified as a global option (via the DseCluster’s GraphOptions) or per-statement (via the method GraphStatement#setGraphName), here’s an example showing when the system variable is accessible, or when g is:

DseCluster dseCluster = DseCluster.builder().addContactPoint("localhost").build();
DseSession dseSession = dseCluster.connect();

// upon successful completion, the graph is created

// Now I can either set the graph name on each GraphStatement
dseSession.executeGraph(new SimpleGraphStatement("g.V()").setGraphName("demo"));

// Or go modify the global GraphOptions which are applied to all queries, unless overridden on a GraphStatement
// At this point, the "system" variable is not available anymore unless you reset the Cluster's Graph name to null

DseCluster’s builder also has a withGraphOptions() method to help specifying a global graph name at the time of building the DseCluster.