Accessing the Apache Spark™ SQL Thrift Server with the Simba JDBC driver

The Simba JDBC Driver for Spark provides a standard JDBC interface to the information stored in DataStax Enterprise with the Spark SQL Thrift Server running.

Your DSE 5.1 license includes a license to use the Simba drivers.

  1. Make sure you have a running DSE Analytics cluster with Spark enabled where one node in the cluster is running the Spark SQL Thrift Server.

  2. Contact DataStax Support to download the Simba JDBC Driver for Apache Spark.

  3. Expand the ZIP file containing the driver.

  4. In your JDBC application, configure the following:

    1. Add SparkJDBC41.jar and the rest of the JAR files included in the ZIP file in your classpath.

    2. The JDBC driver class is com.simba.spark.jdbc41.Driver and the JDBC data source is com.simba.spark.jdbc41.DataSource.

    3. Set the connection URL to jdbc:spark://<hostname>:<port>, such as jdbc:spark://node1.example.com:10000. <hostname> is the hostname of the node on which the Spark SQL Thrift Server is running. <port> is the port number on which the Spark SQL Thrift Server is listening.

For more information about using this driver, contact DataStax Support.

Was this helpful?

Give Feedback

How can we improve the documentation?

© Copyright IBM Corporation 2025 | Privacy policy | Terms of use Manage Privacy Choices

Apache, Apache Cassandra, Cassandra, Apache Tomcat, Tomcat, Apache Lucene, Apache Solr, Apache Hadoop, Hadoop, Apache Pulsar, Pulsar, Apache Spark, Spark, Apache TinkerPop, TinkerPop, Apache Kafka and Kafka are either registered trademarks or trademarks of the Apache Software Foundation or its subsidiaries in Canada, the United States and/or other countries. Kubernetes is the registered trademark of the Linux Foundation.

General Inquiries: Contact IBM