DataStax Apache Kafka Connector

Deploy the DataStax Apache Kafka Connector to stream records from Apache Kafka® topics to your Astra Managed Cluster databases.

The Kafka Connector distribution package includes a sample JSON properties file (dse-sink-distributed.json.sample). Use the sample file as a reference when configuring your deployment. The dse-sink-distributed.json.sample file is located in the conf directory of the Kafka Connector installation.

  1. Download and install the Kafka Connector.

  2. Configure the distributed worker configuration file connect-distributed.properties for your use case. For an example, see the DataStax Kafka examples repository.

    Specify the converter for the key.converter and value.converter properties that matches the form of your Kafka data. For more information, see Configuring converters in the Confluent documentation.

  3. From your Kafka installation directory, start the distributed worker:

    bin/connect-distributed.sh config/connect-distributed.properties

    The worker startup process outputs a large number of informational messages. When the process is complete, the final output is similar to the following: [2019-10-13 19:49:25,385] INFO Finished starting connectors and tasks (org.apache.kafka.connect.runtime.distributed.DistributedHerder:852)

  4. Configure the JSON configuration file (such as dse-sink.json) to use the Secure Connect Bundle (SCB):

    { "name": "dse-sink",     "config":
      { "connector.class": "com.datastax.kafkaconnector.DseSinkConnector",
        "cloud.secureConnectBundle": "/path/to/scb.zip",
        "auth.username": "clientId",
        "auth.password": "clientSecret" ...
      }
    }

    Define the following configuration settings:

    • name: Unique name for the connector. Default: dse-sink.

    • connector.class: DataStax connector Java class provided in the kafka-connect-dse-N.N.N.jar. Default: com.datastax.kafkaconnector.DseSinkConnector.

    • cloud.secureConnectBundle: The full path to the SCB for your Managed Cluster database (secure-connect-DATABASE_NAME.zip).

      If this option is specified, you must also include the auth.username and auth.password for the database user.

    • auth.username: The database username or the literal string token.

    • auth.password: The database password or an Astra application token (AstraCS:…​).

      The user or application token must have a minimum of modify privileges on tables receiving data from the Kafka Connector.

  5. Register the connector configuration with the distributed worker:

    curl -X POST -H "Content-Type: application/json" -d @dse-sink.json "http://IP/:PORT/connectors"

    ip and port are the IP address and port number of the Kafka worker. Use the same port as the rest.port parameter set in connect-distributed.properties. The default port is 8083.

Was this helpful?

Give Feedback

How can we improve the documentation?

© Copyright IBM Corporation 2026 | Privacy policy | Terms of use Manage Privacy Choices

Apache, Apache Cassandra, Cassandra, Apache Tomcat, Tomcat, Apache Lucene, Apache Solr, Apache Hadoop, Hadoop, Apache Pulsar, Pulsar, Apache Spark, Spark, Apache TinkerPop, TinkerPop, Apache Kafka and Kafka are either registered trademarks or trademarks of the Apache Software Foundation or its subsidiaries in Canada, the United States and/or other countries. Kubernetes is the registered trademark of the Linux Foundation.

General Inquiries: Contact IBM