Skip to main content

Standard and upsert connectors for Apache Kafka®

In addition to integration with Apache Kafka® through a standard connector, Aiven for Apache Flink® also supports the use of upsert connectors, which allows you to create changelog-type data streams.

When you integrate a standard Apache Kafka® with your Aiven for Apache Flink® service, you can use the service to read data from one Kafka topic and write the processed data to another Kafka topic. Each message in the source topic that is processed and written to the sink topic is considered unique and is handled as an individual entry.

Another integration approach is upsert, or INSERT/UPDATE, which is a method of updating or deleting data based on the message key. When used as a source, Flink interprets a message that has an existing key as an update and replaces the value for that key. If the key does not exist, it is inserted as new data, and if the message data is null, it is interpreted as a DELETE operation for that key.

When used as a sink, upsert provides a method to create Kafka compacted topics (see the Log Compaction section in the Apache Kafka® documentation for details), containing only the latest value for a specific key and pushing a tombstone message on deletion.

Choose the right connector type

For most common purposes, such as filtering data or performing calculations on streamed data, a standard Kafka connector is sufficient. You can also perform data aggregation with standard connectors, but depending on the type of aggregation, it may require more careful planning and more complex SQL compared to using upsert connectors.

To use multiple data streams as a source, but have the results for matching identifiers combined into single events within the same target topic, choose upsert connectors. While it varies to some extent depending on your specific use case, this means that using upsert connectors often makes it easier and more efficient to implement data aggregation.

In addition, choose upsert connectors to provide the output as a compacted topic and read only the latest value for each message key.

For more details on the required information for each connector type, see Settings for Apache Kafka® connectors.