此内容没有您所选择的语言版本。

Chapter 41. MongoDB Source


Consume documents from MongoDB.

If the persistentTailTracking option will be enabled, the consumer will keep track of the last consumed message and on the next restart, the consumption will restart from that message. In case of persistentTailTracking enabled, the tailTrackIncreasingField must be provided (by default it is optional).

If the persistentTailTracking option won’t be enabled, the consumer will consume the whole collection and wait in idle for new documents to consume.

41.1. Configuration Options

The following table summarizes the configuration options available for the mongodb-source Kamelet:

PropertyNameDescriptionTypeDefaultExample

collection *

MongoDB Collection

Sets the name of the MongoDB collection to bind to this endpoint.

string

  

database *

MongoDB Database

Sets the name of the MongoDB database to target.

string

  

hosts *

MongoDB Hosts

Comma separated list of MongoDB Host Addresses in host:port format.

string

  

password *

MongoDB Password

User password for accessing MongoDB.

string

  

username *

MongoDB Username

Username for accessing MongoDB. The username must be present in the MongoDB’s authentication database (authenticationDatabase). By default, the MongoDB authenticationDatabase is 'admin'.

string

  

persistentTailTracking

MongoDB Persistent Tail Tracking

Enable persistent tail tracking, which is a mechanism to keep track of the last consumed message across system restarts. The next time the system is up, the endpoint will recover the cursor from the point where it last stopped slurping records.

boolean

false

 

tailTrackIncreasingField

MongoDB Tail Track Increasing Field

Correlation field in the incoming record which is of increasing nature and will be used to position the tailing cursor every time it is generated.

string

  
Note

Fields marked with an asterisk (*) are mandatory.

41.2. Dependencies

At runtime, the mongodb-source Kamelet relies upon the presence of the following dependencies:

  • camel:kamelet
  • camel:mongodb
  • camel:jackson

41.3. Usage

This section describes how you can use the mongodb-source.

41.3.1. Knative Source

You can use the mongodb-source Kamelet as a Knative source by binding it to a Knative object.

mongodb-source-binding.yaml

apiVersion: camel.apache.org/v1alpha1
kind: KameletBinding
metadata:
  name: mongodb-source-binding
spec:
  source:
    ref:
      kind: Kamelet
      apiVersion: camel.apache.org/v1alpha1
      name: mongodb-source
    properties:
      collection: "The MongoDB Collection"
      database: "The MongoDB Database"
      hosts: "The MongoDB Hosts"
      password: "The MongoDB Password"
      username: "The MongoDB Username"
  sink:
    ref:
      kind: Channel
      apiVersion: messaging.knative.dev/v1
      name: mychannel

41.3.1.1. Prerequisite

Make sure you have "Red Hat Integration - Camel K" installed into the OpenShift cluster you’re connected to.

41.3.1.2. Procedure for using the cluster CLI

  1. Save the mongodb-source-binding.yaml file to your local drive, and then edit it as needed for your configuration.
  2. Run the source by using the following command:

    oc apply -f mongodb-source-binding.yaml

41.3.1.3. Procedure for using the Kamel CLI

Configure and run the source by using the following command:

kamel bind mongodb-source -p "source.collection=The MongoDB Collection" -p "source.database=The MongoDB Database" -p "source.hosts=The MongoDB Hosts" -p "source.password=The MongoDB Password" -p "source.username=The MongoDB Username" channel:mychannel

This command creates the KameletBinding in the current namespace on the cluster.

41.3.2. Kafka Source

You can use the mongodb-source Kamelet as a Kafka source by binding it to a Kafka topic.

mongodb-source-binding.yaml

apiVersion: camel.apache.org/v1alpha1
kind: KameletBinding
metadata:
  name: mongodb-source-binding
spec:
  source:
    ref:
      kind: Kamelet
      apiVersion: camel.apache.org/v1alpha1
      name: mongodb-source
    properties:
      collection: "The MongoDB Collection"
      database: "The MongoDB Database"
      hosts: "The MongoDB Hosts"
      password: "The MongoDB Password"
      username: "The MongoDB Username"
  sink:
    ref:
      kind: KafkaTopic
      apiVersion: kafka.strimzi.io/v1beta1
      name: my-topic

41.3.2.1. Prerequisites

Ensure that you’ve installed the AMQ Streams operator in your OpenShift cluster and created a topic named my-topic in the current namespace. Make also sure you have "Red Hat Integration - Camel K" installed into the OpenShift cluster you’re connected to.

41.3.2.2. Procedure for using the cluster CLI

  1. Save the mongodb-source-binding.yaml file to your local drive, and then edit it as needed for your configuration.
  2. Run the source by using the following command:

    oc apply -f mongodb-source-binding.yaml

41.3.2.3. Procedure for using the Kamel CLI

Configure and run the source by using the following command:

kamel bind mongodb-source -p "source.collection=The MongoDB Collection" -p "source.database=The MongoDB Database" -p "source.hosts=The MongoDB Hosts" -p "source.password=The MongoDB Password" -p "source.username=The MongoDB Username" kafka.strimzi.io/v1beta1:KafkaTopic:my-topic

This command creates the KameletBinding in the current namespace on the cluster.

41.4. Kamelet source file

https://github.com/openshift-integration/kamelet-catalog/blob/kamelet-catalog-1.6/mongodb-source.kamelet.yaml

Red Hat logoGithubRedditYoutubeTwitter

学习

尝试、购买和销售

社区

关于红帽文档

通过我们的产品和服务,以及可以信赖的内容,帮助红帽用户创新并实现他们的目标。

让开源更具包容性

红帽致力于替换我们的代码、文档和 Web 属性中存在问题的语言。欲了解更多详情,请参阅红帽博客.

關於紅帽

我们提供强化的解决方案,使企业能够更轻松地跨平台和环境(从核心数据中心到网络边缘)工作。

© 2024 Red Hat, Inc.