The kcm-kafka-connect-adls-sink is a Kafka Connect Sink designed to help you easily send data from Kafka to Azure Data Lake Storage Gen2 (ADLS Gen2). It supports SAS authentication, handles Avro data formats, compresses data using GZIP, and organizes your data into partitions. This application simplifies data transfer and enhances your data storage solutions.
To run the kcm-kafka-connect-adls-sink, ensure your system meets these requirements:
To get started, visit the Releases page to download the application.
After downloading, you need to configure the application to connect to your Kafka and Azure accounts.
Create a Configuration File:
You need a configuration file that specifies how the application connects to Kafka and ADLS Gen2. Create a text file named kafka-connect-adls-sink.properties.
name=kcm-kafka-connect-adls-sink
connector.class=com.nitinpensia.kafka.connect.adls.sink.AdlsSinkConnector
tasks.max=1
topics=your-kafka-topic
adls.url=https://youraccount.dfs.core.windows.net/
adls.sas.token=your-sas-token
Replace your-kafka-topic, youraccount, and your-sas-token with your actual Kafka topic name, Azure Data Lake account, and SAS token.
If you encounter issues, check the following:
Common errors may include authentication failures, network issues, or incorrect file paths.
For any support needs, feel free to raise issues on our GitHub Issues page. We aim to address queries promptly.
Contributions are welcome! If you have enhancements or bug fixes, please create a pull request or open an issue for discussion.
Your feedback is valuable. Please share your thoughts or suggestions directly on the GitHub repository.
This project is licensed under the MIT License. You are free to use, modify, and distribute it in accordance with the license.
Visit the Releases page to start utilizing the kcm-kafka-connect-adls-sink today!