|
25 | 25 | import org.apache.hudi.common.model.EventTimeAvroPayload; |
26 | 26 | import org.apache.hudi.common.model.HoodieCleaningPolicy; |
27 | 27 | import org.apache.hudi.common.model.HoodieTableType; |
| 28 | +import org.apache.hudi.common.model.WriteOperationType; |
28 | 29 | import org.apache.hudi.config.HoodieIndexConfig; |
29 | 30 | import org.apache.hudi.config.HoodieWriteConfig; |
30 | 31 | import org.apache.hudi.exception.HoodieException; |
31 | 32 | import org.apache.hudi.hive.MultiPartKeysValueExtractor; |
| 33 | +import org.apache.hudi.hive.ddl.HiveSyncMode; |
32 | 34 | import org.apache.hudi.index.HoodieIndex; |
33 | 35 | import org.apache.hudi.keygen.constant.KeyGeneratorOptions; |
34 | 36 | import org.apache.hudi.keygen.constant.KeyGeneratorType; |
| 37 | +import org.apache.hudi.table.action.cluster.ClusteringPlanPartitionFilterMode; |
35 | 38 |
|
36 | 39 | import org.apache.flink.configuration.ConfigOption; |
37 | 40 | import org.apache.flink.configuration.ConfigOptions; |
@@ -294,7 +297,7 @@ private FlinkOptions() { |
294 | 297 | public static final ConfigOption<String> OPERATION = ConfigOptions |
295 | 298 | .key("write.operation") |
296 | 299 | .stringType() |
297 | | - .defaultValue("upsert") |
| 300 | + .defaultValue(WriteOperationType.UPSERT.value()) |
298 | 301 | .withDescription("The write operation, that this write should do"); |
299 | 302 |
|
300 | 303 | /** |
@@ -656,7 +659,7 @@ private FlinkOptions() { |
656 | 659 | public static final ConfigOption<String> CLUSTERING_PLAN_PARTITION_FILTER_MODE_NAME = ConfigOptions |
657 | 660 | .key("clustering.plan.partition.filter.mode") |
658 | 661 | .stringType() |
659 | | - .defaultValue("NONE") |
| 662 | + .defaultValue(ClusteringPlanPartitionFilterMode.NONE.name()) |
660 | 663 | .withDescription("Partition filter mode used in the creation of clustering plan. Available values are - " |
661 | 664 | + "NONE: do not filter table partition and thus the clustering plan will include all partitions that have clustering candidate." |
662 | 665 | + "RECENT_DAYS: keep a continuous range of partitions, worked together with configs '" + DAYBASED_LOOKBACK_PARTITIONS.key() + "' and '" |
@@ -725,8 +728,8 @@ private FlinkOptions() { |
725 | 728 | public static final ConfigOption<String> HIVE_SYNC_MODE = ConfigOptions |
726 | 729 | .key("hive_sync.mode") |
727 | 730 | .stringType() |
728 | | - .defaultValue("hms") |
729 | | - .withDescription("Mode to choose for Hive ops. Valid values are hms, jdbc and hiveql, default 'jdbc'"); |
| 731 | + .defaultValue(HiveSyncMode.HMS.name()) |
| 732 | + .withDescription("Mode to choose for Hive ops. Valid values are hms, jdbc and hiveql, default 'hms'"); |
730 | 733 |
|
731 | 734 | public static final ConfigOption<String> HIVE_SYNC_USERNAME = ConfigOptions |
732 | 735 | .key("hive_sync.username") |
|
0 commit comments