Connecting to DataStax Enterprise using the Spark shell on an external Spark cluster
Use the generated byos.properties
configuration file and the byos-<version>.jar
from a DataStax Enterprise node to connect to the DataStax Enterprise cluster from the Spark shell on an external Spark cluster.
Prerequisites
You must generate the byos.properties
on a node in your DataStax Enterprise cluster.
Procedure
-
Copy the
byos.properties
file you previously generated from the DataStax Enterprise node to the local Spark node.scp user@dsenode1.example.com:~/byos.properties .
If you are using Kerberos authentication, specify the --generate-token and --token-renewer <username> options when generating
byos.properties
, as described in dse client-tool configuration byos-export. -
Copy the
byos-<version>.jar
file from theclients
directory from a node in your DataStax Enterprise cluster to the local Spark node.The
byos-<version>.jar
file location depends on the type of installation.scp user@dsenode1.example.com:/usr/share/dse/clients/dse-byos_2.11-6.0.2.jar byos-6.0.jar
-
Merge external Spark properties into
byos.properties
.cat ${SPARK_HOME}/conf/spark-defaults.conf >> byos.properties
-
If you are using Kerberos authentication, set up a CRON job or other task scheduler to periodically call dse client-tool cassandra renew-token <token> where
<token>
is the encoded token string inbyos.properties
. -
Start the Spark shell using the
byos.properties
andbyos-<version>.jar
file.spark-shell --jars byos-6.0.jar --properties-file byos.properties