Document version
This product has reached end of life.
About DataStax Enterprise
DataStax Enterprise is a big data platform built on Apache Cassandra that manages real-time, analytics, and enterprise search data. DataStax Enterprise leverages Cassandra, Apache Hadoop, and Apache Solr to shift your focus from the data infrastructure to using your data strategically.
Using the in-memory option
DataStax Enterprise includes the in-memory option for storing data to and accessing data from memory exclusively.
See the DataStax Upgrade Guide.
Configure data compression on a per-table basis to optimize performance of read-dominated tasks.
DataStax Enterprise installation methods include GUI or text mode, unattended command line or properties file, YUM and APT repository, and binary tarball.
Installing on RHEL-based systems
Install DataStax Enterprise and OpsCenter using Yum repositories on RHEL-based systems.
Installing on Debian-based systems
Install DataStax Enterprise and OpsCenter using APT repositories on Debian-based systems.
Installing the binary tarball
Install DataStax Enterprise on any Linux-based platform, including 32-bit platforms.
Installing on SUSE
DataStax provides a binary tarball distribution for installing DataStax Enterprise on SUSE Linux.
On cloud providers
Install on Amazon EC2 or HP cloud.
Installing prior releases
Steps for installing the same version as other nodes in your cluster.
Managing security in DataStax Enterprise including authentication, encryption, auditing, permissions, and configuration.
Security management
DataStax Enterprise includes advanced data protection for enterprise-grade databases including internal authentication, object permissions, encryption, Kerberos authentication, and data auditing.
Authenticating with Kerberos
DataStax Enterprise authentication with Kerberos protocol uses tickets to prove identity for nodes that communicate over non-secure networks.
Client-to-node encryption
Client-to-node encryption protects data in flight from client machines to a database cluster.
Node-to-node encryption
Node-to-node encryption protects data that is transferred between nodes in a cluster using SSL (Secure Sockets Layer).
Server certificates
Generate SSL certificates for client-to-node encryptions or node-to-node encryption.
Installing cqlsh security
Install packages to use cqlsh with a Kerberized cluster.
Transparent data encryption
Transparent data encryption (TDE) protects at rest data. TDE requires a secure local file system to be effective.
Data auditing
Auditing is implemented as a log4j-based integration.
Internal authentication
Internal authentication is based on Cassandra-controlled login accounts and passwords.
Managing object permissions
Use GRANT/REVOKE to grant or revoke permissions to access Cassandra data.
Configuring keyspace replication
The system_auth and dse_security keyspaces store security authentication and authorization information.
Configuring firewall ports
Opening the required ports to allow communication between the nodes.
DSE Analytics with Hadoop
DSE Hadoop topics.
Getting started
The Hadoop component in DataStax Enterprise enables analytics to be run across DataStax Enterprise's distributed, shared-nothing architecture. Instead of using the Hadoop Distributed File System (HDFS), DataStax Enterprise uses Cassandra File System (CFS) keyspaces for the underlying storage layer.
Using the job tracker node
DataStax Enterprise schedules a series of tasks on the analytics nodes for each MapReduce job that is submitted to the job tracker.
About the Cassandra File System
A Hive or Pig analytics job requires a Hadoop file system to function. For use with DSE Hadoop, DataStax Enterprise provides a replacement for the Hadoop Distributed File System (HDFS) called the Cassandra File System (CFS).
Using the cfs-archive to store huge files
The Cassandra File System (CFS) consists of two layers: cfs and cfs-archive. Using cfs-archive is recommended for long-term storage of huge files.
Using Hive
DataStax Enterprise includes a Cassandra-enabled Hive MapReduce client.
ODBC driver for Hive
The DataStax ODBC Driver for Hive provides Windows users access to the information that is stored in DSE Hadoop.
Using Mahout
DataStax Enterprise integrates Apache Mahout, a Hadoop component that offers machine learning libraries.
Using Pig
DataStax Enterprise includes a Cassandra File System (CFS) enabled Apache Pig Client to provide a high-level programming environment for MapReduce coding.
Migrating data using Sqoop topics.
DSE Search with Solr
DSE Search topics.
Getting Started with Solr
DataStax Enterprise supports Open Source Solr (OSS) tools and APIs, simplifying migration from Solr to DataStax Enterprise.
Supported and unsupported features
Supported and unsupported DSE Search and Solr features.
Defining key Solr terms
Solr terms include several names for an index of documents and configuration on a single node.
Installing Solr nodes
Installing and starting Solr nodes.
Solr tutorial
Steps for setting up Cassandra and Solr for the tutorial.
Configuring Solr
Configure Solr Type mapping.
Creating an index for searching
Requirements and steps for creating a Solr index.
Using DSE Search/Solr
A brief description and illustration of DSE Search.
Querying Solr data
A brief description about query Solr data.
Capacity planning
Use a discovery process to develop a plan to ensure sufficient memory resources.
Mixing workloads
About using real-time (Cassandra), Hadoop, or search (Solr) nodes in the same cluster.
Common operations
Topics for using DSE Search.
Tuning DSE Search performance
Topics for performance tuning and solving performance degradation, high memory consumption, or other problem swith DataStax Enterprise Search nodes.
DSE vs. Open source
A comparison of DSE Search and Open Source Solr.
Request processing and data transformation
Use the custom update request processor (URP) to extend the Solr URP. Use the field input/output transformer API as an option to the input/output transformer support in OS Solr.
Deployment topics.
Production deployment planning
Production deployment planning requires knowledge of the initial volume of data to store and an estimate of the typical application workload.
Configuring replication
Choose a data partitioner and replica placement strategy.
Single data center deployment
A deployment scenario with a mixed workload cluster has only one data center for each type of workload.
Multiple data center deployment
A deployment scenario with a mixed workload cluster has more than one data center for each type of node.
Single-token architecture deployment
Use single-token architecture deployment when you are not using virtual nodes (vnodes).
Calculating tokens
Tokens assign a range of data to a particular node within a data center.
Expanding an AMI cluster
To expand your EC2 implementations, use OpsCenter to provision a new cluster, add a new cluster, or add nodes to a cluster.
DataStax Enterprise tools
Tools include dse commands, dsetool, dfs-stress tool, pre-flight check, yaml_diff, and the Cassandra bulk loader.
The dse commands
Table of dse commands for using DataStax Enterprise
The dsetool
Use the dsetool utility for Cassandra File System (CFS) and Hadoop-related tasks, such as managing the job tracker, checking the CFS, and listing node subranges of data in a keyspace.
Configuring the disk health checker
Ways to enable, disable, and use the disk health checker>
Pre-flight check and yaml_diff tools
The pre-flight check tool is available for packaged installations. This collection of tests can be run on a node to detect and fix a configuration. The yaml_diff tool filters differences between cassandra.yaml files.
Using the Cassandra bulk loader in a secure environment
Using sstableloader tool with Kerberos/SSL.
Moving data to/from other databases
DataStax offers several solutions for migrating from other databases.
Reference topics.
Installing glibc on Oracle Linux
To install DSE on Oracle Enterprise Linux 6.x and later, install the 32-bit versions of the glibc libraries.
Tarball file locations
Locations when DataStax Enterprise was installed from a tarball.
Package file locations
Locations when DataStax Enterprise was installed from a package.
Configuration (dse.yaml)
The configuration file for Kerberos authentication, purging of expired data from the Solr indexes, and setting Solr inter-node communication.
Starting and stopping DSE
Starting and stopping DataStax Enterprise as a service or stand-alone process.
Troubleshooting examples are useful to discover and resolve problems with DSE. Also check the Cassandra troubleshooting documentation.
Cassandra Log4j appender
DataStax Enterprise allows you to stream your web and application log information into a database cluster via Apache log4j.
Release notes
Release notes for DataStax Enterprise 4.0.x.
Using the docs
Describes navigation icons and provides search tips and links to other resources.