Kafka
Overview
The Airbyte Kafka destination allows you to sync data to Kafka. Each stream is written to the corresponding Kafka topic.
Prerequisites
- For Airbyte Open Source users using the Postgres source connector, upgrade your Airbyte platform to version
v0.40.0-alpha
or newer and upgrade your Kafka connector to version0.1.10
or newer
Sync overview
Output schema
Each stream will be output into a Kafka topic.
Currently, this connector only writes data with JSON format. More formats (e.g. Apache Avro) will be supported in the future.
Each record will contain in its key the uuid assigned by Airbyte, and in the value these 3 fields:
_airbyte_ab_id
: a uuid assigned by Airbyte to each event that is processed._airbyte_emitted_at
: a timestamp representing when the event was pulled from the data source._airbyte_data
: a json blob representing with the event data._airbyte_stream
: the name of each record's stream.
Features
Feature | Supported?(Yes/No) | Notes |
---|---|---|
Full Refresh Sync | No | |
Incremental - Append Sync | Yes | |
Incremental - Deduped History | No | As this connector does not support dbt, we don't support this sync mode on this destination. |
Namespaces | Yes |
Getting started
Requirements
To use the Kafka destination, you'll need:
- A Kafka cluster 1.0 or above.
Setup guide
Network Access
Make sure your Kafka brokers can be accessed by Airbyte.
Permissions
Airbyte should be allowed to write messages into topics, and these topics should be created before writing into Kafka or, at least, enable the configuration in the brokers auto.create.topics.enable
(which is not recommended for production environments).
Note that if you choose to use dynamic topic names, you will probably need to enable auto.create.topics.enable
to avoid your connection failing if there was an update to the source connector's schema. Otherwise a hardcoded topic name may be best.
Target topics
You can determine the topics to which messages are written via the topic_pattern
configuration parameter. Messages can be written to either a hardcoded, pre-defined topic, or dynamically written to different topics based on the namespace or stream they came from.
To write all messages to a single hardcoded topic, enter its name in the topic_pattern
field e.g: setting topic_pattern
to my-topic-name
will write all messages from all streams and namespaces to that topic.
To define the output topics dynamically, you can leverage the {namespace}
and {stream}
pattern variables, which cause messages to be written to different topics based on the values present when producing the records. For example, setting the topic_pattern
parameter to airbyte_syncs/{namespace}/{stream}
means that messages from namespace n1
and stream s1
will get written to the topic airbyte_syncs/n1/s1
, and messages from s2
to airbyte_syncs/n1/s2
etc.
If you define output topic dynamically, you might want to enable auto.create.topics.enable
to avoid your connection failing if there was an update to the source connector's schema. Otherwise, you'll need to manually create topics in Kafka as they are added/updated in the source, which is the recommended option for production environments.
NOTICE: a naming convention transformation will be applied to the target topic name using the StandardNameTransformer
so that some special characters will be replaced.
Setup the Kafka destination in Airbyte
You should now have all the requirements needed to configure Kafka as a destination in the UI. You can configure the following parameters on the Kafka destination (though many of these are optional or have default values):
- Bootstrap servers
- Topic pattern
- Test topic
- Sync producer
- Security protocol
- SASL JAAS config
- SASL mechanism
- Client ID
- ACKs
- Enable idempotence
- Compression type
- Batch size
- Linger ms
- Max in flight requests per connection
- Client DNS lookup
- Buffer memory
- Max request size
- Retries
- Socket connection setup timeout
- Socket connection setup max timeout
- Max block ms
- Request timeout
- Delivery timeout
- Send buffer bytes
- Receive buffer bytes
More info about this can be found in the Kafka producer configs documentation site.
NOTE: Some configurations for SSL are not available yet.
Changelog
Version | Date | Pull Request | Subject |
---|---|---|---|
0.1.10 | 2022-08-04 | 15287 | Update Kafka destination to use outputRecordCollector to properly store state |
0.1.9 | 2022-06-17 | 13864 | Updated stacktrace format for any trace message errors |
0.1.7 | 2022-04-19 | 12134 | Add PLAIN Auth |
0.1.6 | 2022-02-15 | 10186 | Add SCRAM-SHA-512 Auth |
0.1.5 | 2022-02-14 | 10256 | Add -XX:+ExitOnOutOfMemoryError JVM option |
0.1.4 | 2022-01-31 | 9905 | Fix SASL config read issue |
0.1.3 | 2021-12-30 | 8809 | Update connector fields title/description |
0.1.2 | 2021-09-14 | 6040 | Change spec.json and config parser |
0.1.1 | 2021-07-30 | 5125 | Enable additionalPropertities in spec.json |
0.1.0 | 2021-07-21 | 3746 | Initial Release |