Connect to DataStax Enterprise using the Apache Spark shell on an external Spark cluster
Use the generated byos.properties configuration file and the byos-<version>.jar from a DataStax Enterprise (DSE) node to connect to the DSE cluster from the Spark shell on an external Spark cluster.
Prerequisites
You must generate the byos.properties on a node in your DSE cluster.
Procedure
-
Copy the
byos.propertiesfile you previously generated from the DSE node to the local Spark node.scp user@dsenode1.example.com:~/byos.properties .If you are using Kerberos authentication, specify the --generate-token and --token-renewer <username> options when generating
byos.properties, as described in dse client-tool configuration byos-export. -
Copy the
byos-<version>.jarfile from theclientsdirectory from a node in your DSE cluster to the local Spark node.The
byos-<version>.jarfile location depends on the type of installation.scp user@dsenode1.example.com:/usr/share/dse/clients/dse-byos_2.11-6.0.2.jar byos-6.0.jar -
Merge external Spark properties into
byos.properties.cat ${SPARK_HOME}/conf/spark-defaults.conf >> byos.properties -
If you are using Kerberos authentication, set up a CRON job or other task scheduler to periodically call dse client-tool cassandra renew-token <token> where
<token>is the encoded token string inbyos.properties. -
Start the Spark shell using the
byos.propertiesandbyos-<version>.jarfile.spark-shell --jars byos-6.0.jar --properties-file byos.properties