This guide explains how to stream data from a Kafka cluster into the MotherDuck database using native jdbc (duckdb) driver. Since this is a direct Kafka sink connector for MotherDuck, it will ingest the data into your database without S3 as an intermediary.
To set up the Connector, you will need to gather connection details and configure your Motherduck instance. Log in to your Motherduck Account and then follow the steps below.
You will need the database name to configure the connector. You can find it on the left side panel of the Motherduck UI, under the Attached databases section.
Name - A unique and memorable name for this Connector
Ingestion Mode (default: upsert) - See Inserts/Upserts for information about ingestion modes
Delete Mode (default: none) - Delete records in destination if deleted at source
Tasks - If Pipelines for this Destination have lag that’s continuously growing over time and not reducing, increase the number of Tasks, otherwise, leave as default
Schema Evolution (default: true) - If enabled, the connector will automatically adapt to changes in the source schema by adding new columns in the target table
Motherduck Token - The access token you generated earlier
Database (Case sensitive) - The name of the database
Schema (Case sensitive - default: streamkap) - The schema within the database to write data to