dse commands

The dse commands provide additional controls for starting and using DataStax Enterprise.

The dse commands provide additional controls for starting and using DataStax Enterprise.

Synopsis 

$ dse  subcommand [command_arguments]
Legend
Syntax conventions Description
Italics Variable value. Replace with a user-defined value.
[ ] Optional. Square brackets ( [ ] ) surround optional command arguments. Do not type the square brackets.
( ) Group. Parentheses ( ( ) ) identify a group to choose from. Do not type the parentheses.
| Or. A vertical bar ( | ) separates alternative elements. Type any one of the elements. Do not type the vertical bar.
[ -- ] Separate the command line options from the command arguments with two hyphens ( -- ). This syntax is useful when arguments might be mistaken for command line options.

You can provide authentication credentials in several ways, see Credentials for authentication.

This table describes the authentication command arguments that can be used with all subcommands.
Command arguments Description
-f Path to configuration file that stores credentials. If not specified, then use ~/.dserc if it exists.
-u Role to authenticate against the configured Cassandra authentication schema.
-p Password to authenticate against the configured Cassandra authentication schema.
-a User name to authenticate with secure JMX.
-b Password to authenticate with secure JMX.
-f Path to configuration file that stores credentials. The credentials in this configuration file override the ~/.dserc credentials.
-v Send the DataStax Enterprise version number to standard output. Does not require authentication.

dse subcommands 

Use the dse subcommands and none or more command arguments.
Note: When multiple flags are used, list them separately on the command line. For example, ensure there is a space between -k and -s in dse cassandra -k -s.
Subcommand Command arguments Description
advrep advrep command options Options for configuring and using DSE Advanced Replication.
beeline   Start the Beeline shell.
cassandra   Start up a real-time Cassandra node in the background. See Starting DataStax Enterprise.
cassandra -c Enable the Cassandra File System (CFS) but not the integrated DSE Job Trackers and Task Trackers. Use to start nodes for running an external Hadoop system.
cassandra -f Start up a real-time Cassandra node in the foreground. Can be used with -k, -t, or -s options.
cassandra -g Start up a node in graph mode. See Starting DataStax Enterprise. Can be used with -s and -k options.
cassandra -k Start up an analytics node in Spark mode in the background. See Starting Spark.
cassandra -k -t Start up an analytics node in Spark and DSE Hadoop mode. See Starting Spark.
cassandra -s Start up a DSE Search node in the background. See Starting DataStax Enterprise.
cassandra -t Start up an analytics node in DSE Hadoop mode in the background. See Starting DataStax Enterprise.

By default, the Hadoop 2.7.1 client libraries are used, except for Hive and Pig, which use Hadoop 1.0.4 libraries with built-in Hadoop trackers.

cassandra-stop -p pid Stop the DataStax Enterprise process number pid. See Stopping a node.
cassandra -s -Ddse.solr.data.dir=path Use path to store DSE Search data. See Moving solr.data
cassandra -Dstart_option All -D options in Cassandra start commands are supported.
client-tool subcommand See dse client-tool.
dse-nodeID command Run standard DataStax Enterprise commands for nodes on a DSE Multi-Instance host machine. See DSE Multi-Instance commands
esri-import ESRI import tool options The DataStax Enterprise custom ESRI import tool supports the Enclosed JSON format. See Spatial analytics support (deprecated).
fs Run DSEFS shell. See dsefs command line tool.
hadoop version Sends the version of the Hadoop component to standard output.
hadoop fs options Invoke the Hadoop FileSystem shell. See the Hadoop tutorial.
hadoop fs -help Send Apache Hadoop fs command descriptions to standard output. See the Hadoop tutorial.
hive   Start a Hive client. By default, the Hadoop 2.7.1 client libraries are used, except for Hive and Pig, which use Hadoop 1.0.4 libraries with built-in Hadoop trackers.
hive --service name Start a Hive server by connecting through the JDBC driver. By default, the Hadoop 2.7.1 client libraries are used, except for Hive and Pig, which use Hadoop 1.0.4 libraries with built-in Hadoop trackers.
hive-schema   Create a hive schema representing the Cassandra table when Using Hive with BYOH.
hive-metastore-migrate Hive-metastore-migrate tool options Map custom external tables to the new release format after upgrading. See dse hive-metastore-migrate -to dest_release_num.
mahout mahout_command options Run Mahout commands.
mahout hadoop hadoop_command options Add Mahout classes to classpath and execute the hadoop command. See Mahout commands.
pig   Start Pig. By default, the Hadoop 2.7.1 client libraries are used, except for Hive and Pig, which use Hadoop 1.0.4 libraries with built-in Hadoop trackers.
pyspark   PySpark.
spark   Accessing Cassandra from the Spark shell.
spark-history-server start Start Spark history server.
spark-history-server stop Stop Spark history server.
spark-sql-thriftserver start Start Spark SQL Thrift server.
spark-sql-thriftserver stop Stop Spark SQL Thrift server.
spark-jobserver start submit options Launch applications on a cluster and use Spark Jobserver.
spark-jobserver stop Stop the Spark Jobserver.
spark-sql   Spark SQL command line
spark-submit options Launch applications on a cluster and use Spark cluster managers. See dse spark-submit.
spark-beeline   Use the Beeline client with Spark SQL Thrift Server.
sqoop -help Send Apache Sqoop command line help to standard output. See the Sqoop reference and the Sqoop demo.

Hadoop, hive, mahout, and pig commands must be issued from an analytics node. The hadoop fs options, which DSE Hadoop supports with one exception (-moveToLocal), are described in the HDFS File System Shell Guide on the Apache Hadoop web site. DSE Hadoop does not support the -moveToLocal option; use the -copyToLocal option instead.

The default location of the dse tool depends on the type of installation:
Package installations /usr/bin/dse
Installer-Services installations /usr/bin/dse
Installer-No Services and Tarball installations install_location/bin/dse

DSE Multi-Instance commands

To run standard DataStax Enterprise commands for nodes on a DSE Multi-Instance host machine, specify the node name using this syntax:
  • sudo dse dse-nodeId tool [command_arguments]
    sudo dsetool dse-nodeId command [command_arguments]
  • For example:

    To run the dsetool ring command on a node named dse-node1 in a cluster on a DSE Multi-Instance host machine:

    sudo dse dse-node1 dsetool ring

    To run the dsetool ring command without DSE Multi-Instance:

    sudo dsetool ring
For details on the dse add-node and dse list-node commands, see DSE Multi-Instance commands.

Cassandra start parameters

List of Cassandra start parameters for DSE.

Cassandra start parameters can be run from the command line or specified in the cassandra-env.sh file.

The location of the cassandra.yaml file depends on the type of installation:
Installer-Services /etc/dse/cassandra/cassandra.yaml
Package installations /etc/dse/cassandra/cassandra.yaml
Installer-No Services install_location/resources/cassandra/conf/cassandra.yaml
Tarball installations install_location/resources/cassandra/conf/cassandra.yaml
The location of the cassandra-env.sh file depends on the type of installation:
Package installations /etc/dse/cassandra/cassandra-env.sh
Tarball installations install_location/resources/cassandra/conf/cassandra-env.sh

Usage 

Specify the start option on the command line:
dse cassandra option
For example:
bin/dse cassandra -Dcassandra.prepared_statements_cache_size_in_bytes=345678912
You can set start options and set options to the JVM instead of setting them in the environment. Add an entry for each option to the cassandra-env.sh file:
JVM_OPTS="$JVM_OPTS -Doption"
For example:
JVM_OPTS="$JVM_OPTS -Dcassandra.prepared_statements_cache_size_in_bytes=345678912"

Start parameters 

All of the Cassandra start parameters are supported for starting a DSE node and for use in the cassandra-env.sh for DSE. The following start parameter is supported only for DSE:

Option Description
dse.solr.data.dir=path Use path to store DSE Search data. See Managing the location of Solr data.