Use Apache Spark with DataStax Enterprise

DataStax Enterprise (DSE) integrates with Apache Spark™ to allow distributed analytic applications to run using database data.


Starting Apache Spark

How you start Apache Spark depends on the installation and if want to run in Spark mode or SearchAnalytics mode:

Running Apache Spark commands against a remote cluster

To run Spark commands against a remote cluster, you must export the DSE configuration from one of the remote nodes to the local client machine.

Monitoring Apache Spark with the web interface

A Spark web interface is bundled with DSE. The Spark web interface facilitates monitoring, debugging, and managing Apache Spark.

Using DSE Spark with third party tools and integrations

The dse exec command sets the environment variables required to run third-party tools that integrate with Apache Spark.

Was this helpful?

Give Feedback

How can we improve the documentation?

© 2025 DataStax | Privacy policy | Terms of use | Manage Privacy Choices

Apache, Apache Cassandra, Cassandra, Apache Tomcat, Tomcat, Apache Lucene, Apache Solr, Apache Hadoop, Hadoop, Apache Pulsar, Pulsar, Apache Spark, Spark, Apache TinkerPop, TinkerPop, Apache Kafka and Kafka are either registered trademarks or trademarks of the Apache Software Foundation or its subsidiaries in Canada, the United States and/or other countries. Kubernetes is the registered trademark of the Linux Foundation.

General Inquiries: +1 (650) 389-6000, info@datastax.com