Azure Kafka Batch through Eventhubs with Azure Schema Registry Source
Provided by: "Apache Software Foundation"
Support Level for this Kamelet is: "Preview"
Receive data from Kafka topics in batch on Azure Eventhubs combined with Azure Schema Registry and commit them manually through KafkaManualCommit or auto commit.
Configuration Options
The following table summarizes the configuration options available for the kafka-batch-azure-schema-registry-source
Kamelet:
Property | Name | Description | Type | Default | Example |
---|---|---|---|---|---|
Azure Schema Registry URL | Required The Apicurio Schema Registry URL. | string | |||
Bootstrap Servers | Required Comma separated list of Kafka Broker URLs. | string | |||
Password | Required Password to authenticate to kafka. | string | |||
Topic Names | Required Comma separated list of Kafka topic names. | string | |||
Allow Manual Commit | Whether to allow doing manual commits. | boolean | false | ||
Auto Commit Enable | If true, periodically commit to ZooKeeper the offset of messages already fetched by the consumer. | boolean | true | ||
Auto Offset Reset | What to do when there is no initial offset. There are 3 enums and the value can be one of latest, earliest, none. | string | latest | ||
Batch Dimension | The maximum number of records returned in a single call to poll(). | int | 500 | ||
Consumer Group | A string that uniquely identifies the group of consumers to which this source belongs. | string | my-group-id | ||
Automatically Deserialize Headers | When enabled the Kamelet source will deserialize all message headers to String representation. | boolean | true | ||
Max Poll Interval | The maximum delay between invocations of poll() when using consumer group management. | int | |||
Poll On Error Behavior | What to do if kafka threw an exception while polling for new messages. There are 5 enums and the value can be one of DISCARD, ERROR_HANDLER, RECONNECT, RETRY, STOP. | string | ERROR_HANDLER | ||
Poll Timeout Interval | The timeout used when polling the KafkaConsumer. | int | 5000 | ||
SASL Mechanism | The Simple Authentication and Security Layer (SASL) Mechanism used. | string | PLAIN | ||
Security Protocol | Protocol used to communicate with brokers. SASL_PLAINTEXT, PLAINTEXT, SASL_SSL and SSL are supported. | string | SASL_SSL | ||
Specific Avro Value Type | The Specific Type Avro will have to deal with. | string | com.example.Order | ||
Value Deserializer | Deserializer class for value that implements the Deserializer interface. | string | com.microsoft.azure.schemaregistry.kafka.avro.KafkaAvroDeserializer |
Dependencies
At runtime, the kafka-batch-azure-schema-registry-source
Kamelet relies upon the presence of the following dependencies:
-
mvn:org.apache.camel.kamelets:camel-kamelets-utils:4.4.4-SNAPSHOT
-
camel:kafka
-
camel:core
-
camel:kamelet
-
camel:azure-schema-registry
-
mvn:com.microsoft.azure:azure-schemaregistry-kafka-avro:1.1.1
-
mvn:com.azure:azure-data-schemaregistry-apacheavro:1.1.13
-
mvn:com.azure:azure-identity:1.11.1
Camel JBang usage
Prerequisites
-
You’ve installed JBang.
-
You have executed the following command:
jbang app install camel@apache/camel
Supposing you have a file named route.yaml with this content:
- route:
from:
uri: "kamelet:timer-source"
parameters:
period: 10000
message: 'test'
steps:
- to:
uri: "kamelet:log-sink"
You can now run it directly through the following command
camel run route.yaml
Camel K Environment Usage
This section describes how you can use the kafka-batch-azure-schema-registry-source
.
Knative source
You can use the kafka-batch-azure-schema-registry-source
Kamelet as a Knative source by binding it to a Knative object.
apiVersion: camel.apache.org/v1
kind: Pipe
metadata:
name: kafka-batch-azure-schema-registry-source-pipe
spec:
source:
ref:
kind: Kamelet
apiVersion: camel.apache.org/v1
name: kafka-batch-azure-schema-registry-source
properties:
azureRegistryUrl: The Azure Schema Registry URL
bootstrapServers: The Bootstrap Servers
password: The Password
topic: The Topic Names
sink:
ref:
kind: Channel
apiVersion: messaging.knative.dev/v1
name: mychannel
Prerequisite
You have Camel K installed on the cluster.
Procedure for using the cluster CLI
-
Save the
kafka-batch-azure-schema-registry-source-pipe.yaml
file to your local drive, and then edit it as needed for your configuration. -
Run the source by using the following command:
kubectl apply -f kafka-batch-azure-schema-registry-source-pipe.yaml
Procedure for using the Kamel CLI
Configure and run the source by using the following command:
kamel bind channel:mychannel -p "source.azureRegistryUrl=The Azure Schema Registry URL" -p "source.bootstrapServers=The Bootstrap Servers" -p "source.password=The Password" -p "source.topic=The Topic Names" kafka-batch-azure-schema-registry-source
This command creates the Kamelet Pipe in the current namespace on the cluster.
Kafka source
You can use the kafka-batch-azure-schema-registry-source
Kamelet as a Kafka source by binding it to a Kafka topic.
apiVersion: camel.apache.org/v1
kind: Pipe
metadata:
name: kafka-batch-azure-schema-registry-source-pipe
spec:
source:
ref:
kind: Kamelet
apiVersion: camel.apache.org/v1
name: kafka-batch-azure-schema-registry-source
properties:
azureRegistryUrl: The Azure Schema Registry URL
bootstrapServers: The Bootstrap Servers
password: The Password
topic: The Topic Names
sink:
ref:
kind: KafkaTopic
apiVersion: kafka.strimzi.io/v1beta1
name: my-topic
Prerequisites
-
You’ve installed Strimzi.
-
You’ve created a topic named
my-topic
in the current namespace. -
You have Camel K installed on the cluster.
Procedure for using the cluster CLI
-
Save the
kafka-batch-azure-schema-registry-source-pipe.yaml
file to your local drive, and then edit it as needed for your configuration. -
Run the source by using the following command:
kubectl apply -f kafka-batch-azure-schema-registry-source-pipe.yaml
Procedure for using the Kamel CLI
Configure and run the source by using the following command:
kamel bind kafka.strimzi.io/v1beta1:KafkaTopic:my-topic -p "source.azureRegistryUrl=The Azure Schema Registry URL" -p "source.bootstrapServers=The Bootstrap Servers" -p "source.password=The Password" -p "source.topic=The Topic Names" kafka-batch-azure-schema-registry-source
This command creates the Kamelet Pipe in the current namespace on the cluster.