Install DataStax Enterprise 5.1 using the binary tarball
These instructions apply to DataStax Enterprise (DSE) 5.1 on any supported Linux-based platforms that use a binary tarball.
-
When installed from the binary tarball, DataStax Enterprise runs as a standalone process.
-
This procedure only installs DataStax Enterprise 5.1. It does not install the following developer related tools: OpsCenter, DataStax Agent, DataStax Studio, or the DSE Graph Loader.
Prerequisites
The following requirements must be met before you install DSE 5.1 using the binary tarball:
-
Configure your operating system to use the latest version of Java 8:
-
Recommended. Access the latest build of a TCK (Technology Compatibility Kit) Certified OpenJDK version 8. For example, OpenJDK 8 (1.8.0_151 minimum). DataStax’s recommendation changed due to the end of public updates for Oracle JRE/JDK 8. See Oracle Java SE Support Roadmap.
-
Supported. Oracle Java SE 8 (JRE or JDK) (1.8.0_151 minimum)
-
-
RedHat-compatible distributions require EPEL (Extra Packages for Enterprise Linux).
-
Python 2.7.x or 3.6+. Both are supported for
cqlsh
. For older RHEL distributions, see Installing Python 2.7 on older RHEL-based package installations.
Download and deploy Datastax Enterprise (DSE) 5.1
By downloading this DataStax product, you agree to the terms of the End User License Agreement (EULA). |
-
Verify that you have installed a supported version of Java:
-
Terminal window command
-
OpenJDK sample result
-
Oracle Java sample result
java -version
openjdk version "1.8.0_171" OpenJDK Runtime Environment (build 1.8.0_171-8u171-b11-0ubuntu0.16.04.1-b11) OpenJDK 64-Bit Server VM (build 25.171-b11, mixed mode)
java version "1.8.0_181" Java(TM) SE Runtime Environment (build 1.8.0_181-b13) Java HotSpot(TM) 64-Bit Server VM (build 25.181-b13, mixed mode)
If the output shows that OpenJDK 8 or Oracle Java 8 is not installed, see Installing the JDK. Datastax recommends the latest build of a Technology Compatibility Kit (TCK) Certified OpenJDK version 8.
-
-
Download and extract the tarball files manually, or use
curl
:These commands download version 5.1. For other versions, see the Release notes.
With the manual option, the shell history retains your password. To avoid this security issue, use
curl
with its --netrc or --netrc-file option.-
Manual download
-
curl download
-
Download the tarball from Download DataStax Enterprise.
-
Extract the binary tarball files into the directory of your choice:
tar -xzvf dse-version_number-bin.tar.gz
For example:
tar -xzvf dse-5.1-bin.tar.gz
During the manual method, the shell history retains your password. To avoid this security issue, use
curl
with its --netrc or --netrc-file option.
-
Use curl to download and extract the files into the 5.1 directory:
curl -L https://downloads.datastax.com/enterprise/dse-5.1-bin.tar.gz | tar xz
+ For example:
+
curl -L https://downloads.datastax.com/enterprise/dse-version_number-bin.tar.gz | tar xz
-
-
Use either the default data and logging directory locations, or define your own directory locations:
-
Default directory locations
-
Define your own data and logging directory locations
-
Define your own Spark directories
To use the default data and logging directory locations, create and change ownership for for`/var/lib/cassandra` and
/var/log/cassandra
: For example:sudo mkdir -p /var/lib/cassandra; sudo chown -R $USER:$GROUP /var/lib/cassandra && sudo mkdir -p /var/log/cassandra; sudo chown -R $USER:$GROUP /var/log/cassandra && sudo mkdir -p /var/lib/dsefs; sudo chown -R $USER:$GROUP /var/lib/dsefs
-
In the <installation_location>, make the directories for the data and logging directories. For example:
mkdir <installation_location>/dse-data && cd dse-data && mkdir data && mkdir commitlog && mkdir saved_caches && mkdir hints && mkdir cdc_raw
-
Change to the directory that contains the
cassandra.yaml
file:cd installation_location/resources/cassandra/conf
-
Edit the following lines in the
cassandra.yaml
file to match the custom locations:data_file_directories: - full_path_to_installation_location/dse-data/data commitlog_directory: full_path_to_installation_location/dse-data/commitlog saved_caches_directory: full_path_to_installation_location/dse-data/saved_caches hints_directory: full_path_to_installation_location/dse-data/hints cdc_raw_directory: full_path_to_installation_location/cdc/raw
-
Make the directories for the Spark
lib
andlog
directories. -
Edit the
spark-env.sh
file to match the locations of your Sparklib
andlog
directories, as described in Configuring Spark nodes.The default location of the
spark-env.sh
file depends on the type of installation:Installation type Location Package installations + Installer-Services installations
/etc/dse/spark/spark-env.sh
Tarball installations + Installer-No Services installations
<installation_location>/resources/spark/conf/spark-env.sh
-
Change to the directory that contains the
dsefs_options
file.
-
-
You can use either the default data and logging directory locations or define your own directory locations:
-
Default directory locations
-
Define your own directory and logging locations
To use the default data and logging directory locations, create and change ownership for the following:
-
/var/lib/cassandra
-
/var/log/cassandra
sudo mkdir -p /var/lib/cassandra; sudo chown -R $USER:$GROUP /var/lib/cassandra && sudo mkdir -p /var/log/cassandra; sudo chown -R $USER:$GROUP /var/log/cassandra
-
In the <installation_location> where you extracted DSE, make the directories for data and logging directories. For example:
mkdir dse-data; chown -R $USER:$GROUP dse-data && cd dse-data && mkdir commitlog; chown -R $USER:$GROUP commitlog && mkdir saved_caches; chown -R $USER:$GROUP saved_caches && mkdir hints; chown -R $USER:$GROUP hints && mkdir cdc_raw; chown -R $USER:$GROUP cdc_raw
-
Change to the directory that contains the
cassandra.yaml
file:cd installation_location/resources/cassandra/conf
-
Update the following lines in the
cassandra.yaml
file to match the custom locations:data_file_directories: - full_path_to_installation_location/dse-data commitlog_directory: full_path_to_installation_location/dse-data/commitlog saved_caches_directory: full_path_to_installation_location/dse-data/saved_caches hints_directory: full_path_to_installation_location/dse-data/hints cdc_raw_directory: full_path_to_installation_location/cdc_raw
-
-
Optional: To use DSE analytics, choose either the default Spark data and logging directory locations or define your own locations::
-
Default directory locations
-
Define your own Spark directory locations
To use the default Spark directory locations, create and change ownership for the following:
-
/var/lib/dsefs
-
/var/lib/spark
-
/var/log/spark
sudo mkdir -p /var/lib/dsefs; sudo chown -R $USER:$GROUP /var/lib/dsefs && sudo mkdir -p /var/lib/spark; sudo chown -R $USER:$GROUP /var/lib/spark && sudo mkdir -p /var/log/spark; sudo chown -R $USER:$GROUP /var/log/spark && sudo mkdir -p /var/lib/spark/rdd; sudo chown -R $USER:$GROUP /var/lib/spark/rdd && sudo mkdir -p /var/log/spark/master; sudo chown -R $USER:$GROUP /var/log/spark/master && sudo mkdir -p /var/log/spark/alwayson_sql; sudo chown -R $USER:$GROUP /var/log/spark/alwayson_sql && sudo mkdir -p /var/lib/spark/worker; sudo chown -R $USER:$GROUP /var/lib/spark/worker
-
In the <installation_location> directory where you extracted DSE, make the data and logging directories. For example:
mkdir dsefs; chown -R $USER:$GROUP dsefs && mkdir spark; chown -R $USER:$GROUP spark && cd spark && mkdir log; chown -R $USER:$GROUP log && mkdir rdd; chown -R $USER:$GROUP rdd && mkdir worker; chown -R $USER:$GROUP worker && cd log && mkdir worker; chown -R $USER:$GROUP worker && mkdir master; chown -R $USER:$GROUP master && mkdir alwayson_sql; chown -R $USER:$GROUP alwayson_sql
-
Change to the directory that contains the
spark-env.sh
file:cd installation_location/resources/spark/conf
-
Uncomment and update the following lines in the
spark-env.sh
file:export SPARK_WORKER_DIR="full_path_to_installation_location/spark/worker" export SPARK_EXECUTOR_DIRS="full_path_to_installation_location/spark/rdd" export SPARK_WORKER_LOG_DIR="full_path_to_installation_location/spark/log/worker" export SPARK_MASTER_LOG_DIR="full_path_to_installation_location/spark/log/master" export ALWAYSON_SQL_LOG_DIR="full_path_to_installation_location/spark/log/alwayson_sql"
-
Change to the directory that contains the
dsefs_options
file:cd installation_location/resources/dse/conf
-
Edit
dse.yaml
to uncomment and update the DSEFS directory:work_dir: full_path_to_installation_location/dsefs
-
-
Apply additional configurations to your DSE installation:
-
For production, be sure to change the default
cassandra
user; otherwise, it is a security risk. See Creating superuser accounts. -
DSE provides several types of workloads; the default is transactional. See startup options for service or standalone installations.
-
Next steps provides links to related tasks and information.
-
-
Optional for single-node cluster installations only:
-
From the directory where you installed DSE 5.1, start DSE:
bin/dse cassandra
For other start options, see Starting DataStax Enterprise as a standalone process.
-
From the directory where you installed DSE 5.1, verify that DSE is running:
bin/nodetool status
Results using vnodes::
Datacenter: Cassandra ===================== Status=Up/Down |/ State=Normal/Leaving/Joining/Moving -- Address Load Tokens Owns Host ID Rack UN 127.0.0.1 82.43 KB 128 ? 40725dc8-7843-43ae-9c98-7c532b1f517e rack1
Results not using vnodes::
Datacenter: Analytics ===================== Status=Up/Down |/ State=Normal/Leaving/Joining/Moving -- Address Load Owns Host ID Token Rack UN 172.16.222.136 103.24 KB ? 3c1d0657-0990-4f78-a3c0-3e0c37fc3a06 1647352612226902707 rack1
-
Next steps
-
You must change or delete the
cassandra
user created on installation. See Creating superuser accounts. -
Configure startup options: service | standalone.
-
If performing an upgrade, go to the next step in the Upgrade Guide.
-
Configuring DataStax Enterprise - Settings for DSE Advanced Security, In-Memory, DSE Advanced Replication, DSE Multi-Instance, DSE Tiered Storage, and more.
-
Configuration and log file locations - Services and package installations.
-
Configuration and log file locations - No Services and tarball installations.
-
Changing logging locations after installation.
-
Planning and testing DSE and Apache Cassandra™ cluster deployments.
-
Configuring the heap dump directory to avoid server crashes.
-
DataStax Studio documentation.