连接器消息处理属性
Overview
使用此页面上的设置可配置 MongoDB Kafka connector 的消息处理行为,包括:
消息批量大小
速率限制
并行任务数
有关按类别整理的接收器Connector配置设置列表,请参阅有关接收器连接器配置属性的指南。
设置
名称 | 说明 | ||
---|---|---|---|
max.batch.size | Type: int Description: Maximum number of sink records to batch together for processing. Consider the batch that contains the following records:
When set to 0 , the connector performs a single bulk write for
the entire batch.When set to 1 , the connector performs one bulk write for each
record in the batch, for a total of five bulk writes as shown in
the following example:
Default: 0 Accepted Values: An integer | ||
bulk.write.ordered | Type: boolean Description: Whether the connector writes a batch of records
as an ordered or unordered bulk write operation.
When set to true , the default value, the
connector writes a batch of records as an ordered bulk write
operation.To learn more about bulk write operations, see
Bulk Write Operations. Default: true Accepted Values: true or false | ||
rate.limiting.every.n | Type: int Description: Number of batches of records the sink connector processes in
order to trigger the rate limiting timeout. A value of 0 means no
rate limiting. Default: 0 Accepted Values: An integer | ||
rate.limiting.timeout | Type: int Description: How long (in milliseconds) to wait before the sink connector
should resume processing after reaching the rate limiting
threshold. Default: 0 Accepted Values: An integer | ||
tasks.max | Type: int Description: The maximum number of tasks to create for this connector. The
connector may create fewer than the maximum tasks specified if it
cannot handle the level of parallelism you specify. IMPORTANT: If you specify a value greater than 1 ,
the connector enables parallel processing of the tasks.
If your topic has multiple partition logs, which enables
the connector to read from the topic in parallel,
the tasks may process the messages out of order.Default: 1 Accepted Values: An integer |