Install DataStax Enterprise 5.1 using the binary tarball

These instructions apply to DataStax Enterprise (DSE) 5.1 on any supported Linux-based platforms that use a binary tarball.

  • When installed from the binary tarball, DataStax Enterprise runs as a standalone process.

  • This procedure only installs DataStax Enterprise 5.1. It does not install the following developer related tools: OpsCenter, DataStax Agent, DataStax Studio, or the DSE Graph Loader.

Prerequisites

The following requirements must be met before you install DSE 5.1 using the binary tarball:

Download and deploy Datastax Enterprise (DSE) 5.1

By downloading this DataStax product, you agree to the terms of the End User License Agreement (EULA).

  1. Verify that you have installed a supported version of Java:

    • Terminal window command

    • OpenJDK sample result

    • Oracle Java sample result

    java -version
    openjdk version "1.8.0_171"
    OpenJDK Runtime Environment (build 1.8.0_171-8u171-b11-0ubuntu0.16.04.1-b11)
    OpenJDK 64-Bit Server VM (build 25.171-b11, mixed mode)
    java version "1.8.0_181"
    Java(TM) SE Runtime Environment (build 1.8.0_181-b13)
    Java HotSpot(TM) 64-Bit Server VM (build 25.181-b13, mixed mode)

    If the output shows that OpenJDK 8 or Oracle Java 8 is not installed, see Installing the JDK. Datastax recommends the latest build of a Technology Compatibility Kit (TCK) Certified OpenJDK version 8.

  2. Download and extract the tarball files manually, or use curl:

    These commands download version 5.1. For other versions, see the Release notes.

    With the manual option, the shell history retains your password. To avoid this security issue, use curl with its --netrc or --netrc-file option.

    • Manual download

    • curl download

    1. Download the tarball from Download DataStax Enterprise.

    2. Extract the binary tarball files into the directory of your choice:

      tar -xzvf dse-version_number-bin.tar.gz

      For example:

      tar -xzvf dse-5.1-bin.tar.gz

      During the manual method, the shell history retains your password. To avoid this security issue, use curl with its --netrc or --netrc-file option.

    1. Use curl to download and extract the files into the 5.1 directory:

    curl -L https://downloads.datastax.com/enterprise/dse-5.1-bin.tar.gz | tar xz

    + For example:

    +

    curl -L https://downloads.datastax.com/enterprise/dse-version_number-bin.tar.gz | tar xz
  3. Use either the default data and logging directory locations, or define your own directory locations:

    • Default directory locations

    • Define your own data and logging directory locations

    • Define your own Spark directories

    To use the default data and logging directory locations, create and change ownership for for`/var/lib/cassandra` and /var/log/cassandra: For example:

    sudo mkdir -p /var/lib/cassandra; sudo chown -R  $USER:$GROUP /var/lib/cassandra &&
    sudo mkdir -p /var/log/cassandra; sudo chown -R  $USER:$GROUP /var/log/cassandra &&
    sudo mkdir -p /var/lib/dsefs; sudo chown -R  $USER:$GROUP /var/lib/dsefs
    1. In the <installation_location>, make the directories for the data and logging directories. For example:

      mkdir <installation_location>/dse-data &&
        cd dse-data &&
        mkdir data &&
        mkdir commitlog &&
        mkdir saved_caches &&
        mkdir hints &&
        mkdir cdc_raw
    2. Change to the directory that contains the cassandra.yaml file:

      cd installation_location/resources/cassandra/conf
    3. Edit the following lines in the cassandra.yaml file to match the custom locations:

      data_file_directories:
        - full_path_to_installation_location/dse-data/data
        commitlog_directory: full_path_to_installation_location/dse-data/commitlog
        saved_caches_directory: full_path_to_installation_location/dse-data/saved_caches
        hints_directory: full_path_to_installation_location/dse-data/hints
        cdc_raw_directory: full_path_to_installation_location/cdc/raw
    1. Make the directories for the Spark lib and log directories.

    2. Edit the spark-env.sh file to match the locations of your Spark lib and log directories, as described in Configuring Spark nodes.

      The default location of the spark-env.sh file depends on the type of installation:

      Installation type Location

      Package installations + Installer-Services installations

      /etc/dse/spark/spark-env.sh

      Tarball installations + Installer-No Services installations

      <installation_location>/resources/spark/conf/spark-env.sh

    3. Change to the directory that contains the dsefs_options file.

  4. You can use either the default data and logging directory locations or define your own directory locations:

    • Default directory locations

    • Define your own directory and logging locations

    To use the default data and logging directory locations, create and change ownership for the following:

    • /var/lib/cassandra

    • /var/log/cassandra

      sudo mkdir -p /var/lib/cassandra; sudo chown -R $USER:$GROUP /var/lib/cassandra &&
      sudo mkdir -p /var/log/cassandra; sudo chown -R $USER:$GROUP /var/log/cassandra
    1. In the <installation_location> where you extracted DSE, make the directories for data and logging directories. For example:

      mkdir dse-data; chown -R $USER:$GROUP dse-data &&
      cd dse-data &&
      mkdir commitlog; chown -R $USER:$GROUP commitlog &&
      mkdir saved_caches; chown -R $USER:$GROUP saved_caches &&
      mkdir hints; chown -R $USER:$GROUP hints &&
      mkdir cdc_raw; chown -R $USER:$GROUP cdc_raw
    2. Change to the directory that contains the cassandra.yaml file:

      cd installation_location/resources/cassandra/conf
    3. Update the following lines in the cassandra.yaml file to match the custom locations:

      data_file_directories:
        - full_path_to_installation_location/dse-data
        commitlog_directory: full_path_to_installation_location/dse-data/commitlog
        saved_caches_directory: full_path_to_installation_location/dse-data/saved_caches
        hints_directory: full_path_to_installation_location/dse-data/hints
        cdc_raw_directory: full_path_to_installation_location/cdc_raw
  5. Optional: To use DSE analytics, choose either the default Spark data and logging directory locations or define your own locations::

    • Default directory locations

    • Define your own Spark directory locations

    To use the default Spark directory locations, create and change ownership for the following:

    • /var/lib/dsefs

    • /var/lib/spark

    • /var/log/spark

      sudo mkdir -p /var/lib/dsefs; sudo chown -R $USER:$GROUP /var/lib/dsefs &&
      sudo mkdir -p /var/lib/spark; sudo chown -R $USER:$GROUP /var/lib/spark &&
      sudo mkdir -p /var/log/spark; sudo chown -R $USER:$GROUP /var/log/spark &&
      sudo mkdir -p /var/lib/spark/rdd; sudo chown -R $USER:$GROUP /var/lib/spark/rdd  &&
      sudo mkdir -p /var/log/spark/master; sudo chown -R $USER:$GROUP /var/log/spark/master  &&
      sudo mkdir -p /var/log/spark/alwayson_sql; sudo chown -R $USER:$GROUP /var/log/spark/alwayson_sql  &&
      sudo mkdir -p /var/lib/spark/worker; sudo chown -R $USER:$GROUP /var/lib/spark/worker
    1. In the <installation_location> directory where you extracted DSE, make the data and logging directories. For example:

      mkdir dsefs; chown -R $USER:$GROUP dsefs &&
      mkdir spark; chown -R $USER:$GROUP spark &&
      cd spark &&
      mkdir log; chown -R $USER:$GROUP log &&
      mkdir rdd; chown -R $USER:$GROUP rdd &&
      mkdir worker; chown -R $USER:$GROUP worker &&
      cd log &&
      mkdir worker; chown -R $USER:$GROUP worker &&
      mkdir master; chown -R $USER:$GROUP master &&
      mkdir alwayson_sql; chown -R $USER:$GROUP alwayson_sql
    2. Change to the directory that contains the spark-env.sh file:

      cd installation_location/resources/spark/conf
    3. Uncomment and update the following lines in the spark-env.sh file:

      export SPARK_WORKER_DIR="full_path_to_installation_location/spark/worker"
      export SPARK_EXECUTOR_DIRS="full_path_to_installation_location/spark/rdd"
      export SPARK_WORKER_LOG_DIR="full_path_to_installation_location/spark/log/worker"
      export SPARK_MASTER_LOG_DIR="full_path_to_installation_location/spark/log/master"
      export ALWAYSON_SQL_LOG_DIR="full_path_to_installation_location/spark/log/alwayson_sql"
    4. Change to the directory that contains the dsefs_options file:

      cd installation_location/resources/dse/conf
    5. Edit dse.yaml to uncomment and update the DSEFS directory:

      work_dir: full_path_to_installation_location/dsefs
  6. Apply additional configurations to your DSE installation:

    • For production, be sure to change the default cassandra user; otherwise, it is a security risk. See Creating superuser accounts.

    • DSE provides several types of workloads; the default is transactional. See startup options for service or standalone installations.

    • Next steps provides links to related tasks and information.

  7. Optional for single-node cluster installations only:

    1. From the directory where you installed DSE 5.1, start DSE:

      bin/dse cassandra
    2. From the directory where you installed DSE 5.1, verify that DSE is running:

      bin/nodetool status
      Results using vnodes::
      Datacenter: Cassandra
      =====================
      Status=Up/Down
      |/ State=Normal/Leaving/Joining/Moving
      --  Address    Load       Tokens  Owns    Host ID                               Rack
      UN  127.0.0.1  82.43 KB   128     ?       40725dc8-7843-43ae-9c98-7c532b1f517e  rack1
      Results not using vnodes::
      Datacenter: Analytics
      =====================
      Status=Up/Down
      |/ State=Normal/Leaving/Joining/Moving
      --  Address         Load       Owns    Host ID                               Token                 Rack
      UN  172.16.222.136  103.24 KB  ?       3c1d0657-0990-4f78-a3c0-3e0c37fc3a06  1647352612226902707   rack1

Next steps

Was this helpful?

Give Feedback

How can we improve the documentation?

© 2025 DataStax | Privacy policy | Terms of use | Manage Privacy Choices

Apache, Apache Cassandra, Cassandra, Apache Tomcat, Tomcat, Apache Lucene, Apache Solr, Apache Hadoop, Hadoop, Apache Pulsar, Pulsar, Apache Spark, Spark, Apache TinkerPop, TinkerPop, Apache Kafka and Kafka are either registered trademarks or trademarks of the Apache Software Foundation or its subsidiaries in Canada, the United States and/or other countries. Kubernetes is the registered trademark of the Linux Foundation.

General Inquiries: +1 (650) 389-6000, info@datastax.com