Skip to content

Commit

Permalink
fix comment
Browse files Browse the repository at this point in the history
  • Loading branch information
MOBIN-F committed May 27, 2024
1 parent 9dce213 commit dadd2db
Show file tree
Hide file tree
Showing 3 changed files with 0 additions and 5 deletions.
1 change: 0 additions & 1 deletion docs/content/flink/cdc-ingestion/kafka-cdc.md
Original file line number Diff line number Diff line change
Expand Up @@ -179,7 +179,6 @@ For some append data (such as log data), it can be treated as special CDC data w
--kafka_conf value.format=json \
--catalog_conf metastore=hive \
--catalog_conf uri=thrift://hive-metastore:9083 \
--table_conf bucket=4 \
--table_conf sink.parallelism=4
```
Expand Down
1 change: 0 additions & 1 deletion docs/content/flink/cdc-ingestion/pulsar-cdc.md
Original file line number Diff line number Diff line change
Expand Up @@ -178,7 +178,6 @@ For some append data (such as log data), it can be treated as special CDC data w
--kafka_conf value.format=json \
--catalog_conf metastore=hive \
--catalog_conf uri=thrift://hive-metastore:9083 \
--table_conf bucket=4 \
--table_conf sink.parallelism=4
```
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -18,7 +18,6 @@

package org.apache.paimon.flink.action.cdc.kafka;

import org.apache.paimon.CoreOptions;
import org.apache.paimon.types.DataType;
import org.apache.paimon.types.DataTypes;
import org.apache.paimon.types.RowType;
Expand All @@ -42,7 +41,6 @@ public class KafkaJsonSyncTableActionITCase extends KafkaSyncTableActionITCase {
public void testSchemaEvolution() throws Exception {
String topic = "schema-evolution";
Map<String, String> tableOptions = new HashMap<>();
tableOptions.put(CoreOptions.BUCKET.key(), "-1");

createTestTopic(topic, 1, 1);
writeRecordsToKafka(topic, "kafka/json/table/schemaevolution/json-data-1.txt");
Expand Down Expand Up @@ -98,7 +96,6 @@ public void testSchemaEvolution() throws Exception {
public void testComputedColumn() throws Exception {
String topic = "computed_column";
Map<String, String> tableOptions = new HashMap<>();
tableOptions.put(CoreOptions.BUCKET.key(), "-1");

createTestTopic(topic, 1, 1);
writeRecordsToKafka(topic, "kafka/json/table/computedcolumn/json-data-1.txt");
Expand Down

0 comments on commit dadd2db

Please sign in to comment.