Mission Control installation requirements

Before installing Mission Control, ensure your environment meets all necessary requirements. This guide details the hardware specifications, storage needs, networking configurations, and other prerequisites for a successful deployment.

Mission Control deployment topologies

Mission Control serves as a platform and interface for deploying and managing DataStax Enterprise (DSE), Hyper-Converged Database (HCD), and Apache Cassandra® database clusters across logical and physical regions.

Here are some examples of deployment topologies:

  • Deploy one instance of Mission Control for the entire organization.

  • Dedicate multiple targeted installations to specific groups.

  • Dedicate one instance per DEV, TEST, and PROD environments.

Organization-level Mission Control deployment

At the top-level, Mission Control consists of a centralized control plane and optional regionally deployed data planes.

Control plane

Provides the core interface for Mission Control, including the top-level API service, user interface, and observability endpoints. The control plane manages coordination across each data plane’s region boundaries.

Data plane

Each deployed data plane includes DSE, HCD, or Cassandra nodes and provides a self-contained environment for local resource management.

Mission Control topology planes

The control plane can function as a data plane since it contains all the necessary components. As a result, you don’t need a separate data plane. For a single region deployment, DataStax recommends a single control plane.

Resource capacity requirements

Installing Mission Control requires different instance types and varying resource capacities. Follow these guidelines as the minimum requirement. You might need to increase CPU capacity, memory, and disk space beyond the recommended minimums.

Platform instances

Mission Control observability includes metrics and logging from all components within the platform. These metrics and logs are centralized, aggregated, then shipped to an external object store. For more information, see [plan-storage]. Managed database instances generate most Mission Control metrics and logs. The recommended values depend on the number of clusters deployed and their individual schemas because each table holds a number of tracked metrics. The following recommendations assume minimal deployment in 1 region with a single managed cluster and associated logical database datacenter.

  • 2 Instances

  • 32 CPU cores

  • 64 GB RAM

  • 1 TB storage

To schedule platform services components, you must assign the Kubernetes label mission-control.datastax.com/role=platform. Platform services include the platform UI, API, observability stack, and operators.

Database instances

Database instances host all scheduled HCD, DSE, and Cassandra nodes.

This sizing guidance assumes each node maps to a single Kubernetes worker. Alternatively, you can co-locate multiple instances on a single worker.

Resource contention often occurs when you tune this deployment.

Adjust the values according to your expected database node resource requirements.

  • 3 Instances

  • 16 physical CPU cores-32 hyper-threaded cores

  • 64 GB RAM minimal

  • 1.5 TB storage

DataStax benchmark testing shows that resource requirements and performance are similar between bare-metal VMs and Kubernetes install types. For production, DataStax recommends running Cassandra with the following specifications:

  • 8 to 16 cores

  • 32GB to 128GB RAM

CPU and RAM setting concepts

The precise amount of RAM required depends on the workload and is usually tuned based on observation of Garbage Collection (GC) metrics. For RAM, you should set the memory requirements to double the heap size. By default, the JVM allows processes to use as much off-heap memory as the configured on heap memory. With a heap size of 8GB, the JVM may use up to 16GB of RAM, depending on off-heap memory usage. Follow this rule to simplify memory allocation: Set the heap size to a reasonable value that matches the available memory on the worker nodes. For example, given 32GB RAM per worker node and extra processes such as metrics or log collectors and Kubernetes system components that need to run, you should set your heap size to 8GB, and memory requests to 16GB. If GC pressure is too high, you can raise those values to 12 and 24GB, respectively. During the installation process, allow the monitoring stack to run on the DSE, HCD, or Cassandra nodes to enable scheduling Mimir and all on the same worker nodes. This considers all infrastructure instances available for scheduling, which may lead to unexpected utilization of hardware resources. In production, you should separate these into different node pools. You can set the CPU requests to a minimum-for example, 1 core, to allow scheduling all components concurrently. Processes continue to use as many cores as possible. Note that in production, despite scheduling the monitoring stack and the DSE, HCD, or Cassandra pods on separate node pools, you still need to leave some RAM available for system pods that allow Kubernetes operations such as networking. For example, if you have 48GB RAM per worker, leave 1GB to 2GB for those extra pods by setting your resource requests to around 46GB.

See further considerations for database instance sizing in the HCD, DSE, and Cassandra planning guides.

Storage requirements

In addition to the local storage requirements for database and observability instances, Mission Control requires one of the following bucket types for object storage:

  • Amazon Web Services Simple Storage Service (AWS S3) or S3 compatible-for example MinIO, OpenStack Swift

  • Google Cloud Storage (GCS)

  • Azure Blob Storage

Mission Control stores observability data and database backups in object storage for long-term retention.

For EKS installations, define a default storage class before you install Mission Control.

Load balancing requirements

The Mission Control UI is accessed through a specific port on any of the cluster worker nodes. DataStax recommends placing this service behind a load balancer that directs traffic from all worker nodes within the cluster to port 30880.

This ensures that the UI remains available should an instance go down.

Mission Control does not support load balancing in front of database instances.

Runtime installation requirements

Runtime installations come with an embedded Kubernetes environment that require additional prerequisites. The operating system is installed and the hosts comply with Replicated’s embedded cluster requirements. For more information, see Install the Mission Control server runtime using the embedded Kubernetes cluster.

Kubernetes management instances

Kubernetes Management nodes and load balancers are required for the runtime installation.

Kubernetes Management nodes

In addition to the number and types of instances determined during [plan-capacity], you must account for the overhead of Kubernetes Management nodes. These nodes contain metadata and services required to deploy workloads across all worker nodes. Mission Control requires three Management node instances, with each node set to:

  • 4 AMD64 (x86_64) CPUs or equivalent

  • 8 GB of RAM

  • 256 GB of disk space

Load balancers

To provide high-availability of services, Mission Control runtime-based installations require an additional load balancer. Place it in front of the Management Kubernetes nodes for core Kubernetes API traffic running on port 6443.

Plan your networking

Ensure that all database pods can route to each other. This is a critical requirement for proper operation and data consistency.

The requirement applies to:

  • All database pods within the same region or availability zone

  • All database pods across different availability zones within the same region

  • All database pods across different regions for multi-region deployments

  • All database pods across different racks in the same datacenter for multi-region deployments

To achieve this:

  • Configure proper network policies and firewall rules to allow worker to worker and pod-to-pod communication.

  • Set up routing tables, container network interfaces, or peering networks.

  • Implement network overlay solutions like Submariner or Cilium

  • Verify that the network infrastructure supports the required pod-to-pod connectivity.

Failure to establish proper pod-to-pod routing results in:

  • Connectivity issues between database pods

  • Cluster instability

  • Data consistency issues

  • Failed replication

Open TCP and UDP ports 1024-65535 between all hosts within a given region, a given control plane, or a given data plane. This allows for the seamless shifting of workloads between instances due to changing environments.

Port Listing
Instance types Server-runtime installation Kubernetes installation

Management

  • tcp:6443 - Runtime API services

  • tcp:30880 - UI service

 — 

Platform

  • tcp:30880 - UI service

  • tcp:30600 - Vector Aggregator

  • tcp:30880 - UI service

  • tcp:30600 - Vector Aggregator

Database

  • tcp:30880 - UI service

  • tcp:9042 - Cassandra Native Protocol

  • tcp:7000 - Internode communications

  • tcp:7001 - Encrypted internode communications

  • tcp:8080 - Management API

  • tcp:30880 - UI service

  • tcp:9042 - Cassandra Native Protocol

  • tcp:7000 - Internode communications

  • tcp:7001 - Encrypted internode communications

  • tcp:8080 - Management API

CIDR planning for runtime installations

When you use the runtime installer for Mission Control, plan your Kubernetes cluster CIDR ranges carefully. The CIDR range determines the IP address space available for pods and services in your cluster.

If you need to change the CIDR range after installation, you must stop the cluster, which causes downtime. Plan your CIDR ranges carefully before installation to avoid changes later.

When planning CIDR ranges, consider the following:

  • Plan for sufficient IP addresses to support your expected workload.

  • Account for future growth in your CIDR range size.

  • Verify that your chosen CIDR range doesn’t conflict with existing network ranges.

For single-region deployments:

  • Choose a CIDR range that supports all your expected pods and services.

  • Account for your planned number of database nodes and their potential growth.

For multi-region deployments, also consider:

  • Choose non-overlapping CIDR ranges for each cluster.

  • Plan separate CIDR ranges for:

    • Your control plane cluster.

    • Each data plane cluster.

    • Any future clusters you plan to add.

Next steps

Proceed to the instance configuration step based on your installation type.

Was this helpful?

Give Feedback

How can we improve the documentation?

© 2025 DataStax | Privacy policy | Terms of use | Manage Privacy Choices

Apache, Apache Cassandra, Cassandra, Apache Tomcat, Tomcat, Apache Lucene, Apache Solr, Apache Hadoop, Hadoop, Apache Pulsar, Pulsar, Apache Spark, Spark, Apache TinkerPop, TinkerPop, Apache Kafka and Kafka are either registered trademarks or trademarks of the Apache Software Foundation or its subsidiaries in Canada, the United States and/or other countries. Kubernetes is the registered trademark of the Linux Foundation.

General Inquiries: +1 (650) 389-6000, info@datastax.com