此内容没有您所选择的语言版本。
Chapter 10. AWS 2 Simple Queue Service FIFO sink
Send message to an AWS SQS FIFO Queue
10.1. Configuration Options 复制链接链接已复制到粘贴板!
The following table summarizes the configuration options available for the aws-sqs-fifo-sink
Kamelet:
Property | Name | Description | Type | Default | Example |
---|---|---|---|---|---|
accessKey * | Access Key | The access key obtained from AWS | string | ||
queueNameOrArn * | Queue Name | The SQS Queue name or ARN | string | ||
region * | AWS Region | The AWS region to connect to | string |
| |
secretKey * | Secret Key | The secret key obtained from AWS | string | ||
autoCreateQueue | Autocreate Queue | Setting the autocreation of the SQS queue. | boolean |
| |
contentBasedDeduplication | Content-Based Deduplication | Use content-based deduplication (should be enabled in the SQS FIFO queue first) | boolean |
|
Fields marked with an asterisk (*) are mandatory.
10.2. Dependencies 复制链接链接已复制到粘贴板!
At runtime, the aws-sqs-fifo-sink
Kamelet relies upon the presence of the following dependencies:
- camel:aws2-sqs
- camel:core
- camel:kamelet
10.3. Usage 复制链接链接已复制到粘贴板!
This section describes how you can use the aws-sqs-fifo-sink
.
10.3.1. Knative Sink 复制链接链接已复制到粘贴板!
You can use the aws-sqs-fifo-sink
Kamelet as a Knative sink by binding it to a Knative object.
aws-sqs-fifo-sink-binding.yaml
10.3.1.1. Prerequisite 复制链接链接已复制到粘贴板!
Make sure you have "Red Hat Integration - Camel K" installed into the OpenShift cluster you’re connected to.
10.3.1.2. Procedure for using the cluster CLI 复制链接链接已复制到粘贴板!
-
Save the
aws-sqs-fifo-sink-binding.yaml
file to your local drive, and then edit it as needed for your configuration. Run the sink by using the following command:
oc apply -f aws-sqs-fifo-sink-binding.yaml
oc apply -f aws-sqs-fifo-sink-binding.yaml
Copy to Clipboard Copied! Toggle word wrap Toggle overflow
10.3.1.3. Procedure for using the Kamel CLI 复制链接链接已复制到粘贴板!
Configure and run the sink by using the following command:
kamel bind channel:mychannel aws-sqs-fifo-sink -p "sink.accessKey=The Access Key" -p "sink.queueNameOrArn=The Queue Name" -p "sink.region=eu-west-1" -p "sink.secretKey=The Secret Key"
kamel bind channel:mychannel aws-sqs-fifo-sink -p "sink.accessKey=The Access Key" -p "sink.queueNameOrArn=The Queue Name" -p "sink.region=eu-west-1" -p "sink.secretKey=The Secret Key"
This command creates the KameletBinding in the current namespace on the cluster.
10.3.2. Kafka Sink 复制链接链接已复制到粘贴板!
You can use the aws-sqs-fifo-sink
Kamelet as a Kafka sink by binding it to a Kafka topic.
aws-sqs-fifo-sink-binding.yaml
10.3.2.1. Prerequisites 复制链接链接已复制到粘贴板!
Ensure that you’ve installed the AMQ Streams operator in your OpenShift cluster and created a topic named my-topic
in the current namespace. Make also sure you have "Red Hat Integration - Camel K" installed into the OpenShift cluster you’re connected to.
10.3.2.2. Procedure for using the cluster CLI 复制链接链接已复制到粘贴板!
-
Save the
aws-sqs-fifo-sink-binding.yaml
file to your local drive, and then edit it as needed for your configuration. Run the sink by using the following command:
oc apply -f aws-sqs-fifo-sink-binding.yaml
oc apply -f aws-sqs-fifo-sink-binding.yaml
Copy to Clipboard Copied! Toggle word wrap Toggle overflow
10.3.2.3. Procedure for using the Kamel CLI 复制链接链接已复制到粘贴板!
Configure and run the sink by using the following command:
kamel bind kafka.strimzi.io/v1beta1:KafkaTopic:my-topic aws-sqs-fifo-sink -p "sink.accessKey=The Access Key" -p "sink.queueNameOrArn=The Queue Name" -p "sink.region=eu-west-1" -p "sink.secretKey=The Secret Key"
kamel bind kafka.strimzi.io/v1beta1:KafkaTopic:my-topic aws-sqs-fifo-sink -p "sink.accessKey=The Access Key" -p "sink.queueNameOrArn=The Queue Name" -p "sink.region=eu-west-1" -p "sink.secretKey=The Secret Key"
This command creates the KameletBinding in the current namespace on the cluster.