diff --git a/docs/content/flink/cdc-ingestion/kafka-cdc.md b/docs/content/flink/cdc-ingestion/kafka-cdc.md index 0fb4ff65a393..9caf53119370 100644 --- a/docs/content/flink/cdc-ingestion/kafka-cdc.md +++ b/docs/content/flink/cdc-ingestion/kafka-cdc.md @@ -179,7 +179,6 @@ For some append data (such as log data), it can be treated as special CDC data w --kafka_conf value.format=json \ --catalog_conf metastore=hive \ --catalog_conf uri=thrift://hive-metastore:9083 \ - --table_conf bucket=4 \ --table_conf sink.parallelism=4 ``` diff --git a/docs/content/flink/cdc-ingestion/pulsar-cdc.md b/docs/content/flink/cdc-ingestion/pulsar-cdc.md index 9d482e6d79c4..76573c229098 100644 --- a/docs/content/flink/cdc-ingestion/pulsar-cdc.md +++ b/docs/content/flink/cdc-ingestion/pulsar-cdc.md @@ -178,7 +178,6 @@ For some append data (such as log data), it can be treated as special CDC data w --kafka_conf value.format=json \ --catalog_conf metastore=hive \ --catalog_conf uri=thrift://hive-metastore:9083 \ - --table_conf bucket=4 \ --table_conf sink.parallelism=4 ``` diff --git a/paimon-flink/paimon-flink-cdc/src/test/java/org/apache/paimon/flink/action/cdc/kafka/KafkaJsonSyncTableActionITCase.java b/paimon-flink/paimon-flink-cdc/src/test/java/org/apache/paimon/flink/action/cdc/kafka/KafkaJsonSyncTableActionITCase.java index e8a6d350e720..c8601bb74954 100644 --- a/paimon-flink/paimon-flink-cdc/src/test/java/org/apache/paimon/flink/action/cdc/kafka/KafkaJsonSyncTableActionITCase.java +++ b/paimon-flink/paimon-flink-cdc/src/test/java/org/apache/paimon/flink/action/cdc/kafka/KafkaJsonSyncTableActionITCase.java @@ -18,7 +18,6 @@ package org.apache.paimon.flink.action.cdc.kafka; -import org.apache.paimon.CoreOptions; import org.apache.paimon.types.DataType; import org.apache.paimon.types.DataTypes; import org.apache.paimon.types.RowType; @@ -42,7 +41,6 @@ public class KafkaJsonSyncTableActionITCase extends KafkaSyncTableActionITCase { public void testSchemaEvolution() throws Exception { String topic = "schema-evolution"; Map tableOptions = new HashMap<>(); - tableOptions.put(CoreOptions.BUCKET.key(), "-1"); createTestTopic(topic, 1, 1); writeRecordsToKafka(topic, "kafka/json/table/schemaevolution/json-data-1.txt"); @@ -98,7 +96,6 @@ public void testSchemaEvolution() throws Exception { public void testComputedColumn() throws Exception { String topic = "computed_column"; Map tableOptions = new HashMap<>(); - tableOptions.put(CoreOptions.BUCKET.key(), "-1"); createTestTopic(topic, 1, 1); writeRecordsToKafka(topic, "kafka/json/table/computedcolumn/json-data-1.txt");