Monitoring Spark with the web interface

The Spark web interface facilitates monitoring, debugging, and managing Spark.

A web interface, bundled with DataStax Enterprise, facilitates monitoring, debugging, and managing Spark.

Using the Spark web interface

To use the Spark web interface enter the listen IP address of any Spark node in a browser followed by port number 7080 (configured in the configuration file). Starting in DSE 5.1, all Spark nodes within an Analytics datacenter will redirect to the current Spark Master.

If upgrading to DSE 6.8.x, a change in the default behavior requires an authorization. This could prevent non-superuser DSE users from accessing the Spark web UI. You can grant the Spark web UI authorization to non-superusers. Update the spark.ui.filters setting in spark-daemon-defaults.conf located in the Spark configuration directory. Change the setting to com.datastax.bdp.auth.SparkUIAuthFilter or com.datastax.bdp.auth.SparkUIAuth*z*Filter.

If the Spark Master is not available, the UI will keep polling for the Spark Master every 10 seconds until the Master is available.

The Spark web interface can be secured using SSL. SSL encryption of the web interface is enabled by default when client encryption is enabled.

If authentication is enabled, and plain authentication is available, you will be prompted for authentication credentials when accessing the web UI. We recommend using SSL with authentication.

Kerberos authentication is not supported in the Spark web UI. If authentication is enabled and either LDAP or Internal authentication is not available, the Spark web UI will not be accessible. If this occurs, disable authentication for the Spark web UI only by removing the spark.ui.filters setting in spark-daemon-defaults.conf located in the Spark configuration directory.

DSE SSL encryption and authentication only apply to the Spark Master and Worker UIs, not the Spark Driver UI. To use encryption and authentication with the Driver UI, refer to the Spark security documentation.

spark master one app

The UI includes information on the number of cores and amount of memory available to Spark in total and in each work pool, and similar information for each Spark worker. The applications list the associated work pool.

See the Spark documentation for information on using the Spark web UI.

Authorization in the Spark web UI

When authorization is enabled and an authenticated user accesses the web UI, what they can see and do is controlled by their permissions. This allows administrators to control who has permission to view specific application logs, view the executors for the application, kill the application, and list all applications. Viewing and modifying applications can be configured per datacenter, work pool, or application.

See Using authorization with Spark for details on granting permissions.

Displaying fully qualified domain names in the web UI

To display fully qualified domain names (FQDNs) in the Spark web UI, set the SPARK_PUBLIC_DNS variable in on each Analytics node.

Set SPARK_PUBLIC_DNS to the FQDN of the node if you have SSL enabled for the web UI.

Redirecting to the fully qualified domain name of the master

Set the SPARK_LOCAL_IP or SPARK_LOCAL_HOSTNAME in the file on each node to the fully qualified domain name (FQDN) of the node to force any redirects to the web UI using the FQDN of the Spark master. This is useful when enabling SSL in the web UI.

export SPARK_LOCAL_HOSTNAME=<FQDN of the node>

Filtering properties in the Spark Driver UI

The Spark Driver UI has an Environment tab that lists the Spark configuration and system properties used by Spark. This can include sensitive information like passwords and security tokens. DSE Spark filters these properties and mask their values with sequences of asterisks. The spark.redaction.regex filter is configured as a regular expression that by default includes all properties that contain the string "secret", "token", or "password" as well as all system properties. To modify the filter, edit the spark.redaction.regex property in spark-defaults.conf in the Spark configuration directory.

Was this helpful?

Give Feedback

How can we improve the documentation?

© 2024 DataStax | Privacy policy | Terms of use

Apache, Apache Cassandra, Cassandra, Apache Tomcat, Tomcat, Apache Lucene, Apache Solr, Apache Hadoop, Hadoop, Apache Pulsar, Pulsar, Apache Spark, Spark, Apache TinkerPop, TinkerPop, Apache Kafka and Kafka are either registered trademarks or trademarks of the Apache Software Foundation or its subsidiaries in Canada, the United States and/or other countries. Kubernetes is the registered trademark of the Linux Foundation.

General Inquiries: +1 (650) 389-6000,