Migrate data to Hyper-Converged Database (HCD)

Hyper-Converged Database (HCD) uses several solutions for migrating data from other databases:

  • Use DataStax Bulk Loader (dsbulk) to load and unload CSV or JSON data in and out of the HCD database.

  • The CQL COPY TO command mirrors what the PostgreSQL RDBMS uses for file/export import.

    You can use COPY in the CQL shell to read CSV data to HCD and write CSV data from HCD to a file system. Typically, an RDBMS has unload utilities for writing table data to a file system.

  • The sstableloader provides the ability to bulk load external data into a cluster.

The DataStax Apache Kafka™ Connector synchronizes records from a Kafka topic with rows in one or more HCD database tables.

ETL tools

If you need more sophistication applied to a data movement situation than just extract-load, you can use any number of extract-transform-load (ETL) solutions that support HCD. These tools provide transformation routines for manipulating source data and then loading the data into a HCD target. The tools offer features such as visual, point-and-click interfaces, scheduling engines, and more.

Many ETL vendors who support HCD supply community editions of their products that are free and able to solve many different use cases. Enterprise editions are also available.

You can download ETL tools that work with HCD from Talend, Informatica, and Streamsets.

Was this helpful?

Give Feedback

How can we improve the documentation?

© 2025 DataStax | Privacy policy | Terms of use | Manage Privacy Choices

Apache, Apache Cassandra, Cassandra, Apache Tomcat, Tomcat, Apache Lucene, Apache Solr, Apache Hadoop, Hadoop, Apache Pulsar, Pulsar, Apache Spark, Spark, Apache TinkerPop, TinkerPop, Apache Kafka and Kafka are either registered trademarks or trademarks of the Apache Software Foundation or its subsidiaries in Canada, the United States and/or other countries. Kubernetes is the registered trademark of the Linux Foundation.

General Inquiries: +1 (650) 389-6000, info@datastax.com