You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hudi.apache.org by yi...@apache.org on 2022/10/28 17:30:00 UTC
[hudi] branch master updated: [HUDI-5035] Remove usage of deprecated HoodieTimer constructor (#6952)
This is an automated email from the ASF dual-hosted git repository.
yihua pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/hudi.git
The following commit(s) were added to refs/heads/master by this push:
new 6be2057376 [HUDI-5035] Remove usage of deprecated HoodieTimer constructor (#6952)
6be2057376 is described below
commit 6be2057376fb10e79ccb690757cb172a2ad48889
Author: slfan1989 <55...@users.noreply.github.com>
AuthorDate: Sat Oct 29 01:29:49 2022 +0800
[HUDI-5035] Remove usage of deprecated HoodieTimer constructor (#6952)
Co-authored-by: slfan1989 <louj1988@@>
Co-authored-by: Y Ethan Guo <et...@gmail.com>
---
.../org/apache/hudi/cli/commands/MetadataCommand.java | 15 ++++++++-------
.../java/org/apache/hudi/index/HoodieIndexUtils.java | 7 ++++---
.../java/org/apache/hudi/io/HoodieKeyLookupHandle.java | 2 +-
.../main/java/org/apache/hudi/io/HoodieWriteHandle.java | 4 ++--
.../hudi/metadata/HoodieBackedTableMetadataWriter.java | 2 +-
.../hudi/table/action/clean/CleanActionExecutor.java | 8 +++-----
.../hudi/table/action/index/RunIndexActionExecutor.java | 3 +--
.../table/action/restore/BaseRestoreActionExecutor.java | 3 +--
.../action/rollback/BaseRollbackActionExecutor.java | 2 +-
.../rollback/CopyOnWriteRollbackActionExecutor.java | 3 +--
.../rollback/MergeOnReadRollbackActionExecutor.java | 3 +--
.../org/apache/hudi/table/marker/DirectWriteMarkers.java | 2 +-
.../table/marker/TimelineServerBasedWriteMarkers.java | 2 +-
.../hudi/io/storage/row/HoodieRowDataCreateHandle.java | 3 +--
.../hudi/client/validator/SparkPreCommitValidator.java | 2 +-
.../commit/SparkDeletePartitionCommitActionExecutor.java | 16 +++++++++-------
.../hudi/client/functional/TestHoodieBackedMetadata.java | 2 +-
.../apache/hudi/testutils/HoodieClientTestHarness.java | 2 +-
.../apache/hudi/common/fs/HoodieWrapperFileSystem.java | 4 ++--
.../common/table/view/AbstractTableFileSystemView.java | 5 ++---
.../apache/hudi/common/util/collection/RocksDBDAO.java | 3 +--
.../java/org/apache/hudi/metadata/BaseTableMetadata.java | 16 ++++++++--------
.../apache/hudi/metadata/HoodieBackedTableMetadata.java | 13 +++++--------
.../main/scala/org/apache/hudi/IncrementalRelation.scala | 9 ++++-----
.../procedures/CreateMetadataTableProcedure.scala | 2 +-
.../command/procedures/InitMetadataTableProcedure.scala | 2 +-
.../hudi/command/procedures/RunCompactionProcedure.scala | 4 +---
.../procedures/ShowMetadataTableFilesProcedure.scala | 2 +-
.../ShowMetadataTablePartitionsProcedure.scala | 2 +-
.../procedures/ValidateMetadataTableFilesProcedure.scala | 2 +-
.../org/apache/hudi/hive/ddl/HiveQueryDDLExecutor.java | 2 +-
.../org/apache/hudi/timeline/service/RequestHandler.java | 2 +-
.../handlers/marker/BatchedMarkerCreationRunnable.java | 2 +-
.../service/handlers/marker/MarkerCreationFuture.java | 2 +-
.../timeline/service/handlers/marker/MarkerDirState.java | 2 +-
35 files changed, 72 insertions(+), 83 deletions(-)
diff --git a/hudi-cli/src/main/java/org/apache/hudi/cli/commands/MetadataCommand.java b/hudi-cli/src/main/java/org/apache/hudi/cli/commands/MetadataCommand.java
index 65b01bb254..77b79d8d55 100644
--- a/hudi-cli/src/main/java/org/apache/hudi/cli/commands/MetadataCommand.java
+++ b/hudi-cli/src/main/java/org/apache/hudi/cli/commands/MetadataCommand.java
@@ -18,8 +18,6 @@
package org.apache.hudi.cli.commands;
-import org.apache.hadoop.fs.FileStatus;
-import org.apache.hadoop.fs.Path;
import org.apache.hudi.cli.HoodieCLI;
import org.apache.hudi.cli.HoodiePrintHelper;
import org.apache.hudi.cli.TableHeader;
@@ -35,6 +33,9 @@ import org.apache.hudi.config.HoodieWriteConfig;
import org.apache.hudi.metadata.HoodieBackedTableMetadata;
import org.apache.hudi.metadata.HoodieTableMetadata;
import org.apache.hudi.metadata.SparkHoodieBackedTableMetadataWriter;
+
+import org.apache.hadoop.fs.FileStatus;
+import org.apache.hadoop.fs.Path;
import org.apache.logging.log4j.LogManager;
import org.apache.logging.log4j.Logger;
import org.apache.spark.api.java.JavaSparkContext;
@@ -122,7 +123,7 @@ public class MetadataCommand {
HoodieCLI.fs.mkdirs(metadataPath);
}
- HoodieTimer timer = new HoodieTimer().startTimer();
+ HoodieTimer timer = HoodieTimer.start();
HoodieWriteConfig writeConfig = getWriteConfig();
initJavaSparkContext(Option.of(master));
SparkHoodieBackedTableMetadataWriter.create(HoodieCLI.conf, writeConfig, new HoodieSparkEngineContext(jsc));
@@ -158,7 +159,7 @@ public class MetadataCommand {
throw new RuntimeException("Metadata directory (" + metadataPath.toString() + ") does not exist.");
}
- HoodieTimer timer = new HoodieTimer().startTimer();
+ HoodieTimer timer = HoodieTimer.start();
if (!readOnly) {
HoodieWriteConfig writeConfig = getWriteConfig();
initJavaSparkContext(Option.of(master));
@@ -206,7 +207,7 @@ public class MetadataCommand {
return "[ERROR] Metadata Table not enabled/initialized\n\n";
}
- HoodieTimer timer = new HoodieTimer().startTimer();
+ HoodieTimer timer = HoodieTimer.start();
List<String> partitions = metadata.getAllPartitionPaths();
LOG.debug("Took " + timer.endTimer() + " ms");
@@ -239,7 +240,7 @@ public class MetadataCommand {
partitionPath = new Path(HoodieCLI.basePath, partition);
}
- HoodieTimer timer = new HoodieTimer().startTimer();
+ HoodieTimer timer = HoodieTimer.start();
FileStatus[] statuses = metaReader.getAllFilesInPartition(partitionPath);
LOG.debug("Took " + timer.endTimer() + " ms");
@@ -271,7 +272,7 @@ public class MetadataCommand {
HoodieBackedTableMetadata fsMetaReader = new HoodieBackedTableMetadata(
new HoodieLocalEngineContext(HoodieCLI.conf), fsConfig, HoodieCLI.basePath, "/tmp");
- HoodieTimer timer = new HoodieTimer().startTimer();
+ HoodieTimer timer = HoodieTimer.start();
List<String> metadataPartitions = metadataReader.getAllPartitionPaths();
LOG.debug("Listing partitions Took " + timer.endTimer() + " ms");
List<String> fsPartitions = fsMetaReader.getAllPartitionPaths();
diff --git a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/index/HoodieIndexUtils.java b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/index/HoodieIndexUtils.java
index 61be856d36..d6872276ac 100644
--- a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/index/HoodieIndexUtils.java
+++ b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/index/HoodieIndexUtils.java
@@ -18,8 +18,6 @@
package org.apache.hudi.index;
-import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.fs.Path;
import org.apache.hudi.common.engine.HoodieEngineContext;
import org.apache.hudi.common.fs.FSUtils;
import org.apache.hudi.common.model.FileSlice;
@@ -35,6 +33,9 @@ import org.apache.hudi.exception.HoodieIndexException;
import org.apache.hudi.io.storage.HoodieFileReader;
import org.apache.hudi.io.storage.HoodieFileReaderFactory;
import org.apache.hudi.table.HoodieTable;
+
+import org.apache.hadoop.conf.Configuration;
+import org.apache.hadoop.fs.Path;
import org.apache.log4j.LogManager;
import org.apache.log4j.Logger;
@@ -152,7 +153,7 @@ public class HoodieIndexUtils {
try {
// Load all rowKeys from the file, to double-confirm
if (!candidateRecordKeys.isEmpty()) {
- HoodieTimer timer = new HoodieTimer().startTimer();
+ HoodieTimer timer = HoodieTimer.start();
HoodieFileReader fileReader = HoodieFileReaderFactory.getFileReader(configuration, filePath);
Set<String> fileRowKeys = fileReader.filterRowKeys(new TreeSet<>(candidateRecordKeys));
foundRecordKeys.addAll(fileRowKeys);
diff --git a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/io/HoodieKeyLookupHandle.java b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/io/HoodieKeyLookupHandle.java
index a38ae7f1f1..df629b83dc 100644
--- a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/io/HoodieKeyLookupHandle.java
+++ b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/io/HoodieKeyLookupHandle.java
@@ -60,7 +60,7 @@ public class HoodieKeyLookupHandle<T extends HoodieRecordPayload, I, K, O> exten
private BloomFilter getBloomFilter() {
BloomFilter bloomFilter = null;
- HoodieTimer timer = new HoodieTimer().startTimer();
+ HoodieTimer timer = HoodieTimer.start();
try {
if (config.getBloomIndexUseMetadata()
&& hoodieTable.getMetaClient().getTableConfig().getMetadataPartitions()
diff --git a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/io/HoodieWriteHandle.java b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/io/HoodieWriteHandle.java
index abf5c0face..807f14ca28 100644
--- a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/io/HoodieWriteHandle.java
+++ b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/io/HoodieWriteHandle.java
@@ -49,8 +49,8 @@ import org.apache.log4j.Logger;
import java.io.IOException;
import java.util.Collections;
-import java.util.List;
import java.util.HashMap;
+import java.util.List;
import static org.apache.hudi.common.util.StringUtils.isNullOrEmpty;
@@ -124,7 +124,7 @@ public abstract class HoodieWriteHandle<T extends HoodieRecordPayload, I, K, O>
this.tableSchemaWithMetaFields = HoodieAvroUtils.addMetadataFields(tableSchema, config.allowOperationMetadataField());
this.writeSchema = overriddenSchema.orElseGet(() -> getWriteSchema(config));
this.writeSchemaWithMetaFields = HoodieAvroUtils.addMetadataFields(writeSchema, config.allowOperationMetadataField());
- this.timer = new HoodieTimer().startTimer();
+ this.timer = HoodieTimer.start();
this.writeStatus = (WriteStatus) ReflectionUtils.loadClass(config.getWriteStatusClassName(),
!hoodieTable.getIndex().isImplicitWithStorage(), config.getWriteStatusFailureFraction());
this.taskContextSupplier = taskContextSupplier;
diff --git a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/metadata/HoodieBackedTableMetadataWriter.java b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/metadata/HoodieBackedTableMetadataWriter.java
index 7da0a4a57a..15759a570f 100644
--- a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/metadata/HoodieBackedTableMetadataWriter.java
+++ b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/metadata/HoodieBackedTableMetadataWriter.java
@@ -382,7 +382,7 @@ public abstract class HoodieBackedTableMetadataWriter implements HoodieTableMeta
protected <T extends SpecificRecordBase> void initializeIfNeeded(HoodieTableMetaClient dataMetaClient,
Option<T> actionMetadata,
Option<String> inflightInstantTimestamp) throws IOException {
- HoodieTimer timer = new HoodieTimer().startTimer();
+ HoodieTimer timer = HoodieTimer.start();
boolean exists = metadataTableExists(dataMetaClient, actionMetadata);
diff --git a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/action/clean/CleanActionExecutor.java b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/action/clean/CleanActionExecutor.java
index 56b01ec77b..750e687947 100644
--- a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/action/clean/CleanActionExecutor.java
+++ b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/action/clean/CleanActionExecutor.java
@@ -18,9 +18,6 @@
package org.apache.hudi.table.action.clean;
-import org.apache.hadoop.fs.FileSystem;
-import org.apache.hadoop.fs.Path;
-
import org.apache.hudi.avro.model.HoodieActionInstant;
import org.apache.hudi.avro.model.HoodieCleanMetadata;
import org.apache.hudi.avro.model.HoodieCleanerPlan;
@@ -43,6 +40,8 @@ import org.apache.hudi.internal.schema.io.FileBasedInternalSchemaStorageManager;
import org.apache.hudi.table.HoodieTable;
import org.apache.hudi.table.action.BaseActionExecutor;
+import org.apache.hadoop.fs.FileSystem;
+import org.apache.hadoop.fs.Path;
import org.apache.log4j.LogManager;
import org.apache.log4j.Logger;
@@ -198,8 +197,7 @@ public class CleanActionExecutor<T extends HoodieRecordPayload, I, K, O> extends
HoodieInstant inflightInstant = null;
try {
- final HoodieTimer timer = new HoodieTimer();
- timer.startTimer();
+ final HoodieTimer timer = HoodieTimer.start();
if (cleanInstant.isRequested()) {
inflightInstant = table.getActiveTimeline().transitionCleanRequestedToInflight(cleanInstant,
TimelineMetadataUtils.serializeCleanerPlan(cleanerPlan));
diff --git a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/action/index/RunIndexActionExecutor.java b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/action/index/RunIndexActionExecutor.java
index 96d46928e7..43c7ed459d 100644
--- a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/action/index/RunIndexActionExecutor.java
+++ b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/action/index/RunIndexActionExecutor.java
@@ -104,8 +104,7 @@ public class RunIndexActionExecutor<T extends HoodieRecordPayload, I, K, O> exte
@Override
public Option<HoodieIndexCommitMetadata> execute() {
- HoodieTimer indexTimer = new HoodieTimer();
- indexTimer.startTimer();
+ HoodieTimer indexTimer = HoodieTimer.start();
HoodieInstant indexInstant = validateAndGetIndexInstant();
// read HoodieIndexPlan
diff --git a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/action/restore/BaseRestoreActionExecutor.java b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/action/restore/BaseRestoreActionExecutor.java
index 62ecbe2a31..8a577021d2 100644
--- a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/action/restore/BaseRestoreActionExecutor.java
+++ b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/action/restore/BaseRestoreActionExecutor.java
@@ -68,8 +68,7 @@ public abstract class BaseRestoreActionExecutor<T extends HoodieRecordPayload, I
@Override
public HoodieRestoreMetadata execute() {
- HoodieTimer restoreTimer = new HoodieTimer();
- restoreTimer.startTimer();
+ HoodieTimer restoreTimer = HoodieTimer.start();
Option<HoodieInstant> restoreInstant = table.getRestoreTimeline()
.filterInflightsAndRequested()
diff --git a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/action/rollback/BaseRollbackActionExecutor.java b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/action/rollback/BaseRollbackActionExecutor.java
index 4add51886f..ef2c790516 100644
--- a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/action/rollback/BaseRollbackActionExecutor.java
+++ b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/action/rollback/BaseRollbackActionExecutor.java
@@ -104,7 +104,7 @@ public abstract class BaseRollbackActionExecutor<T extends HoodieRecordPayload,
? table.getActiveTimeline().transitionRollbackRequestedToInflight(rollbackInstant)
: rollbackInstant;
- HoodieTimer rollbackTimer = new HoodieTimer().startTimer();
+ HoodieTimer rollbackTimer = HoodieTimer.start();
List<HoodieRollbackStat> stats = doRollbackAndGetStats(rollbackPlan);
HoodieRollbackMetadata rollbackMetadata = TimelineMetadataUtils.convertRollbackMetadata(
instantTime,
diff --git a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/action/rollback/CopyOnWriteRollbackActionExecutor.java b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/action/rollback/CopyOnWriteRollbackActionExecutor.java
index e766dbdc81..64b3f483e6 100644
--- a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/action/rollback/CopyOnWriteRollbackActionExecutor.java
+++ b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/action/rollback/CopyOnWriteRollbackActionExecutor.java
@@ -61,8 +61,7 @@ public class CopyOnWriteRollbackActionExecutor<T extends HoodieRecordPayload, I,
@Override
protected List<HoodieRollbackStat> executeRollback(HoodieRollbackPlan hoodieRollbackPlan) {
- HoodieTimer rollbackTimer = new HoodieTimer();
- rollbackTimer.startTimer();
+ HoodieTimer rollbackTimer = HoodieTimer.start();
List<HoodieRollbackStat> stats = new ArrayList<>();
HoodieActiveTimeline activeTimeline = table.getActiveTimeline();
diff --git a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/action/rollback/MergeOnReadRollbackActionExecutor.java b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/action/rollback/MergeOnReadRollbackActionExecutor.java
index 46d4d84ebf..097897bd15 100644
--- a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/action/rollback/MergeOnReadRollbackActionExecutor.java
+++ b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/action/rollback/MergeOnReadRollbackActionExecutor.java
@@ -61,8 +61,7 @@ public class MergeOnReadRollbackActionExecutor<T extends HoodieRecordPayload, I,
@Override
protected List<HoodieRollbackStat> executeRollback(HoodieRollbackPlan hoodieRollbackPlan) {
- HoodieTimer rollbackTimer = new HoodieTimer();
- rollbackTimer.startTimer();
+ HoodieTimer rollbackTimer = HoodieTimer.start();
LOG.info("Rolling back instant " + instantToRollback);
diff --git a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/marker/DirectWriteMarkers.java b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/marker/DirectWriteMarkers.java
index e813382079..f1a7cde432 100644
--- a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/marker/DirectWriteMarkers.java
+++ b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/marker/DirectWriteMarkers.java
@@ -156,7 +156,7 @@ public class DirectWriteMarkers extends WriteMarkers {
}
private Option<Path> create(Path markerPath, boolean checkIfExists) {
- HoodieTimer timer = new HoodieTimer().startTimer();
+ HoodieTimer timer = HoodieTimer.start();
Path dirPath = markerPath.getParent();
try {
if (!fs.exists(dirPath)) {
diff --git a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/marker/TimelineServerBasedWriteMarkers.java b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/marker/TimelineServerBasedWriteMarkers.java
index 4879e0bc60..2de9c9fdb8 100644
--- a/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/marker/TimelineServerBasedWriteMarkers.java
+++ b/hudi-client/hudi-client-common/src/main/java/org/apache/hudi/table/marker/TimelineServerBasedWriteMarkers.java
@@ -129,7 +129,7 @@ public class TimelineServerBasedWriteMarkers extends WriteMarkers {
@Override
protected Option<Path> create(String partitionPath, String dataFileName, IOType type, boolean checkIfExists) {
- HoodieTimer timer = new HoodieTimer().startTimer();
+ HoodieTimer timer = HoodieTimer.start();
String markerFileName = getMarkerFileName(dataFileName, type);
Map<String, String> paramsMap = new HashMap<>();
diff --git a/hudi-client/hudi-flink-client/src/main/java/org/apache/hudi/io/storage/row/HoodieRowDataCreateHandle.java b/hudi-client/hudi-flink-client/src/main/java/org/apache/hudi/io/storage/row/HoodieRowDataCreateHandle.java
index ec059b23cd..231dee2c7d 100644
--- a/hudi-client/hudi-flink-client/src/main/java/org/apache/hudi/io/storage/row/HoodieRowDataCreateHandle.java
+++ b/hudi-client/hudi-flink-client/src/main/java/org/apache/hudi/io/storage/row/HoodieRowDataCreateHandle.java
@@ -82,8 +82,7 @@ public class HoodieRowDataCreateHandle implements Serializable {
this.taskEpochId = taskEpochId;
this.fileId = fileId;
this.preserveHoodieMetadata = preserveHoodieMetadata;
- this.currTimer = new HoodieTimer();
- this.currTimer.startTimer();
+ this.currTimer = HoodieTimer.start();
this.fs = table.getMetaClient().getFs();
this.path = makeNewPath(partitionPath);
this.writeStatus = new HoodieInternalWriteStatus(!table.getIndex().isImplicitWithStorage(),
diff --git a/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/client/validator/SparkPreCommitValidator.java b/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/client/validator/SparkPreCommitValidator.java
index f08d11b571..c1f0f1a83b 100644
--- a/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/client/validator/SparkPreCommitValidator.java
+++ b/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/client/validator/SparkPreCommitValidator.java
@@ -70,7 +70,7 @@ public abstract class SparkPreCommitValidator<T extends HoodieRecordPayload, I,
* Throw HoodieValidationException if any unexpected data is written (Example: data files are not readable for some reason).
*/
public void validate(String instantTime, HoodieWriteMetadata<O> writeResult, Dataset<Row> before, Dataset<Row> after) throws HoodieValidationException {
- HoodieTimer timer = new HoodieTimer().startTimer();
+ HoodieTimer timer = HoodieTimer.start();
try {
validateRecordsBeforeAndAfter(before, after, getPartitionsModified(writeResult));
} finally {
diff --git a/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/table/action/commit/SparkDeletePartitionCommitActionExecutor.java b/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/table/action/commit/SparkDeletePartitionCommitActionExecutor.java
index 149aef03e2..7cc0efb34e 100644
--- a/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/table/action/commit/SparkDeletePartitionCommitActionExecutor.java
+++ b/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/table/action/commit/SparkDeletePartitionCommitActionExecutor.java
@@ -18,12 +18,6 @@
package org.apache.hudi.table.action.commit;
-import java.time.Duration;
-import java.util.Collections;
-import java.util.HashMap;
-import java.util.List;
-import java.util.Map;
-import org.apache.hadoop.fs.Path;
import org.apache.hudi.avro.model.HoodieRequestedReplaceMetadata;
import org.apache.hudi.client.WriteStatus;
import org.apache.hudi.common.data.HoodieData;
@@ -42,6 +36,14 @@ import org.apache.hudi.table.WorkloadProfile;
import org.apache.hudi.table.WorkloadStat;
import org.apache.hudi.table.action.HoodieWriteMetadata;
+import org.apache.hadoop.fs.Path;
+
+import java.time.Duration;
+import java.util.Collections;
+import java.util.HashMap;
+import java.util.List;
+import java.util.Map;
+
import static org.apache.hudi.common.table.timeline.HoodieInstant.State.REQUESTED;
import static org.apache.hudi.common.table.timeline.HoodieTimeline.REPLACE_COMMIT_ACTION;
@@ -59,7 +61,7 @@ public class SparkDeletePartitionCommitActionExecutor<T extends HoodieRecordPayl
@Override
public HoodieWriteMetadata<HoodieData<WriteStatus>> execute() {
try {
- HoodieTimer timer = new HoodieTimer().startTimer();
+ HoodieTimer timer = HoodieTimer.start();
context.setJobStatus(this.getClass().getSimpleName(), "Gather all file ids from all deleting partitions.");
Map<String, List<String>> partitionToReplaceFileIds =
HoodieJavaPairRDD.getJavaPairRDD(context.parallelize(partitions).distinct()
diff --git a/hudi-client/hudi-spark-client/src/test/java/org/apache/hudi/client/functional/TestHoodieBackedMetadata.java b/hudi-client/hudi-spark-client/src/test/java/org/apache/hudi/client/functional/TestHoodieBackedMetadata.java
index 8ea6c2adf8..a82904a8f4 100644
--- a/hudi-client/hudi-spark-client/src/test/java/org/apache/hudi/client/functional/TestHoodieBackedMetadata.java
+++ b/hudi-client/hudi-spark-client/src/test/java/org/apache/hudi/client/functional/TestHoodieBackedMetadata.java
@@ -2400,7 +2400,7 @@ public class TestHoodieBackedMetadata extends TestHoodieMetadataBase {
return;
}
- HoodieTimer timer = new HoodieTimer().startTimer();
+ HoodieTimer timer = HoodieTimer.start();
HoodieSparkEngineContext engineContext = new HoodieSparkEngineContext(jsc);
// Partitions should match
diff --git a/hudi-client/hudi-spark-client/src/test/java/org/apache/hudi/testutils/HoodieClientTestHarness.java b/hudi-client/hudi-spark-client/src/test/java/org/apache/hudi/testutils/HoodieClientTestHarness.java
index a0c093be16..a41c62cdce 100644
--- a/hudi-client/hudi-spark-client/src/test/java/org/apache/hudi/testutils/HoodieClientTestHarness.java
+++ b/hudi-client/hudi-spark-client/src/test/java/org/apache/hudi/testutils/HoodieClientTestHarness.java
@@ -578,7 +578,7 @@ public abstract class HoodieClientTestHarness extends HoodieCommonTestHarness im
}
assertEquals(inflightCommits, testTable.inflightCommits());
- HoodieTimer timer = new HoodieTimer().startTimer();
+ HoodieTimer timer = HoodieTimer.start();
HoodieSparkEngineContext engineContext = new HoodieSparkEngineContext(jsc);
// Partitions should match
diff --git a/hudi-common/src/main/java/org/apache/hudi/common/fs/HoodieWrapperFileSystem.java b/hudi-common/src/main/java/org/apache/hudi/common/fs/HoodieWrapperFileSystem.java
index 2979696be7..fba24097fb 100644
--- a/hudi-common/src/main/java/org/apache/hudi/common/fs/HoodieWrapperFileSystem.java
+++ b/hudi-common/src/main/java/org/apache/hudi/common/fs/HoodieWrapperFileSystem.java
@@ -24,6 +24,7 @@ import org.apache.hudi.common.util.HoodieTimer;
import org.apache.hudi.common.util.Option;
import org.apache.hudi.exception.HoodieException;
import org.apache.hudi.exception.HoodieIOException;
+import org.apache.hudi.hadoop.CachingPath;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.BlockLocation;
@@ -49,7 +50,6 @@ import org.apache.hadoop.fs.permission.FsPermission;
import org.apache.hadoop.security.Credentials;
import org.apache.hadoop.security.token.Token;
import org.apache.hadoop.util.Progressable;
-import org.apache.hudi.hadoop.CachingPath;
import java.io.IOException;
import java.net.URI;
@@ -102,7 +102,7 @@ public class HoodieWrapperFileSystem extends FileSystem {
}
protected static <R> R executeFuncWithTimeMetrics(String metricName, Path p, CheckedFunction<R> func) throws IOException {
- HoodieTimer timer = new HoodieTimer().startTimer();
+ HoodieTimer timer = HoodieTimer.start();
R res = func.get();
Registry registry = getMetricRegistryForPath(p);
diff --git a/hudi-common/src/main/java/org/apache/hudi/common/table/view/AbstractTableFileSystemView.java b/hudi-common/src/main/java/org/apache/hudi/common/table/view/AbstractTableFileSystemView.java
index a614523ba0..625277a375 100644
--- a/hudi-common/src/main/java/org/apache/hudi/common/table/view/AbstractTableFileSystemView.java
+++ b/hudi-common/src/main/java/org/apache/hudi/common/table/view/AbstractTableFileSystemView.java
@@ -132,7 +132,7 @@ public abstract class AbstractTableFileSystemView implements SyncableFileSystemV
* Adds the provided statuses into the file system view, and also caches it inside this object.
*/
public List<HoodieFileGroup> addFilesToView(FileStatus[] statuses) {
- HoodieTimer timer = new HoodieTimer().startTimer();
+ HoodieTimer timer = HoodieTimer.start();
List<HoodieFileGroup> fileGroups = buildFileGroups(statuses, visibleCommitsAndCompactionTimeline, true);
long fgBuildTimeTakenMs = timer.endTimer();
timer.startTimer();
@@ -216,8 +216,7 @@ public abstract class AbstractTableFileSystemView implements SyncableFileSystemV
* Get replaced instant for each file group by looking at all commit instants.
*/
private void resetFileGroupsReplaced(HoodieTimeline timeline) {
- HoodieTimer hoodieTimer = new HoodieTimer();
- hoodieTimer.startTimer();
+ HoodieTimer hoodieTimer = HoodieTimer.start();
// for each REPLACE instant, get map of (partitionPath -> deleteFileGroup)
HoodieTimeline replacedTimeline = timeline.getCompletedReplaceTimeline();
Stream<Map.Entry<HoodieFileGroupId, HoodieInstant>> resultStream = replacedTimeline.getInstants().flatMap(instant -> {
diff --git a/hudi-common/src/main/java/org/apache/hudi/common/util/collection/RocksDBDAO.java b/hudi-common/src/main/java/org/apache/hudi/common/util/collection/RocksDBDAO.java
index fe40d98594..639b5ed415 100644
--- a/hudi-common/src/main/java/org/apache/hudi/common/util/collection/RocksDBDAO.java
+++ b/hudi-common/src/main/java/org/apache/hudi/common/util/collection/RocksDBDAO.java
@@ -334,8 +334,7 @@ public class RocksDBDAO {
*/
public <T extends Serializable> Stream<Pair<String, T>> prefixSearch(String columnFamilyName, String prefix) {
ValidationUtils.checkArgument(!closed);
- final HoodieTimer timer = new HoodieTimer();
- timer.startTimer();
+ final HoodieTimer timer = HoodieTimer.start();
long timeTakenMicro = 0;
List<Pair<String, T>> results = new LinkedList<>();
try (final RocksIterator it = getRocksDB().newIterator(managedHandlesMap.get(columnFamilyName))) {
diff --git a/hudi-common/src/main/java/org/apache/hudi/metadata/BaseTableMetadata.java b/hudi-common/src/main/java/org/apache/hudi/metadata/BaseTableMetadata.java
index 37a209b0a8..e9474214d4 100644
--- a/hudi-common/src/main/java/org/apache/hudi/metadata/BaseTableMetadata.java
+++ b/hudi-common/src/main/java/org/apache/hudi/metadata/BaseTableMetadata.java
@@ -19,7 +19,6 @@
package org.apache.hudi.metadata;
-import org.apache.hadoop.fs.FileSystem;
import org.apache.hudi.avro.model.HoodieMetadataBloomFilter;
import org.apache.hudi.avro.model.HoodieMetadataColumnStats;
import org.apache.hudi.common.bloom.BloomFilter;
@@ -42,11 +41,12 @@ import org.apache.hudi.common.util.hash.FileIndexID;
import org.apache.hudi.common.util.hash.PartitionIndexID;
import org.apache.hudi.exception.HoodieIOException;
import org.apache.hudi.exception.HoodieMetadataException;
+import org.apache.hudi.hadoop.CachingPath;
+import org.apache.hudi.hadoop.SerializablePath;
import org.apache.hadoop.fs.FileStatus;
+import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
-import org.apache.hudi.hadoop.CachingPath;
-import org.apache.hudi.hadoop.SerializablePath;
import org.apache.log4j.LogManager;
import org.apache.log4j.Logger;
@@ -198,7 +198,7 @@ public abstract class BaseTableMetadata implements HoodieTableMetadata {
return Collections.emptyMap();
}
- HoodieTimer timer = new HoodieTimer().startTimer();
+ HoodieTimer timer = HoodieTimer.start();
Set<String> partitionIDFileIDSortedStrings = new TreeSet<>();
Map<String, Pair<String, String>> fileToKeyMap = new HashMap<>();
partitionNameFileNameList.forEach(partitionNameFileNamePair -> {
@@ -258,7 +258,7 @@ public abstract class BaseTableMetadata implements HoodieTableMetadata {
}
List<String> columnStatKeys = new ArrayList<>(sortedKeys);
- HoodieTimer timer = new HoodieTimer().startTimer();
+ HoodieTimer timer = HoodieTimer.start();
List<Pair<String, Option<HoodieRecord<HoodieMetadataPayload>>>> hoodieRecordList =
getRecordsByKeys(columnStatKeys, MetadataPartitionType.COLUMN_STATS.getPartitionPath());
metrics.ifPresent(m -> m.updateMetrics(HoodieMetadataMetrics.LOOKUP_COLUMN_STATS_METADATA_STR, timer.endTimer()));
@@ -287,7 +287,7 @@ public abstract class BaseTableMetadata implements HoodieTableMetadata {
* Returns a list of all partitions.
*/
protected List<String> fetchAllPartitionPaths() {
- HoodieTimer timer = new HoodieTimer().startTimer();
+ HoodieTimer timer = HoodieTimer.start();
Option<HoodieRecord<HoodieMetadataPayload>> recordOpt = getRecordByKey(RECORDKEY_PARTITION_LIST,
MetadataPartitionType.FILES.getPartitionPath());
metrics.ifPresent(m -> m.updateMetrics(HoodieMetadataMetrics.LOOKUP_PARTITIONS_STR, timer.endTimer()));
@@ -319,7 +319,7 @@ public abstract class BaseTableMetadata implements HoodieTableMetadata {
String relativePartitionPath = FSUtils.getRelativePartitionPath(dataBasePath.get(), partitionPath);
String recordKey = relativePartitionPath.isEmpty() ? NON_PARTITIONED_NAME : relativePartitionPath;
- HoodieTimer timer = new HoodieTimer().startTimer();
+ HoodieTimer timer = HoodieTimer.start();
Option<HoodieRecord<HoodieMetadataPayload>> recordOpt = getRecordByKey(recordKey,
MetadataPartitionType.FILES.getPartitionPath());
metrics.ifPresent(m -> m.updateMetrics(HoodieMetadataMetrics.LOOKUP_FILES_STR, timer.endTimer()));
@@ -349,7 +349,7 @@ public abstract class BaseTableMetadata implements HoodieTableMetadata {
}, Function.identity())
);
- HoodieTimer timer = new HoodieTimer().startTimer();
+ HoodieTimer timer = HoodieTimer.start();
List<Pair<String, Option<HoodieRecord<HoodieMetadataPayload>>>> partitionIdRecordPairs =
getRecordsByKeys(new ArrayList<>(partitionIdToPathMap.keySet()), MetadataPartitionType.FILES.getPartitionPath());
metrics.ifPresent(m -> m.updateMetrics(HoodieMetadataMetrics.LOOKUP_FILES_STR, timer.endTimer()));
diff --git a/hudi-common/src/main/java/org/apache/hudi/metadata/HoodieBackedTableMetadata.java b/hudi-common/src/main/java/org/apache/hudi/metadata/HoodieBackedTableMetadata.java
index bcaf2bcab3..001c049337 100644
--- a/hudi-common/src/main/java/org/apache/hudi/metadata/HoodieBackedTableMetadata.java
+++ b/hudi-common/src/main/java/org/apache/hudi/metadata/HoodieBackedTableMetadata.java
@@ -244,8 +244,7 @@ public class HoodieBackedTableMetadata extends BaseTableMetadata {
List<String> keys,
boolean fullKey,
List<Long> timings) {
- HoodieTimer timer = new HoodieTimer().startTimer();
- timer.startTimer();
+ HoodieTimer timer = HoodieTimer.start();
if (logRecordScanner == null) {
timings.add(timer.endTimer());
@@ -285,8 +284,7 @@ public class HoodieBackedTableMetadata extends BaseTableMetadata {
Map<String, Option<HoodieRecord<HoodieMetadataPayload>>> logRecords,
List<Long> timings,
String partitionName) throws IOException {
- HoodieTimer timer = new HoodieTimer().startTimer();
- timer.startTimer();
+ HoodieTimer timer = HoodieTimer.start();
if (baseFileReader == null) {
// No base file at all
@@ -304,8 +302,7 @@ public class HoodieBackedTableMetadata extends BaseTableMetadata {
}
}
- HoodieTimer readTimer = new HoodieTimer();
- readTimer.startTimer();
+ HoodieTimer readTimer = HoodieTimer.start();
Map<String, HoodieRecord<HoodieMetadataPayload>> records =
fetchBaseFileRecordsByKeys(baseFileReader, keys, fullKeys, partitionName);
@@ -408,7 +405,7 @@ public class HoodieBackedTableMetadata extends BaseTableMetadata {
private Pair<HoodieFileReader, HoodieMetadataMergedLogRecordReader> openReaders(String partitionName, FileSlice slice) {
try {
- HoodieTimer timer = new HoodieTimer().startTimer();
+ HoodieTimer timer = HoodieTimer.start();
// Open base file reader
Pair<HoodieFileReader, Long> baseFileReaderOpenTimePair = getBaseFileReader(slice, timer);
HoodieFileReader baseFileReader = baseFileReaderOpenTimePair.getKey();
@@ -472,7 +469,7 @@ public class HoodieBackedTableMetadata extends BaseTableMetadata {
public Pair<HoodieMetadataMergedLogRecordReader, Long> getLogRecordScanner(List<HoodieLogFile> logFiles,
String partitionName,
Option<Boolean> allowFullScanOverride) {
- HoodieTimer timer = new HoodieTimer().startTimer();
+ HoodieTimer timer = HoodieTimer.start();
List<String> sortedLogFilePaths = logFiles.stream()
.sorted(HoodieLogFile.getLogFileComparator())
.map(o -> o.getPath().toString())
diff --git a/hudi-spark-datasource/hudi-spark-common/src/main/scala/org/apache/hudi/IncrementalRelation.scala b/hudi-spark-datasource/hudi-spark-common/src/main/scala/org/apache/hudi/IncrementalRelation.scala
index db48f224f2..e5497d030a 100644
--- a/hudi-spark-datasource/hudi-spark-common/src/main/scala/org/apache/hudi/IncrementalRelation.scala
+++ b/hudi-spark-datasource/hudi-spark-common/src/main/scala/org/apache/hudi/IncrementalRelation.scala
@@ -18,15 +18,13 @@
package org.apache.hudi
import org.apache.avro.Schema
-import org.apache.hudi.common.model.{HoodieCommitMetadata, HoodieFileFormat, HoodieRecord, HoodieReplaceCommitMetadata}
-import org.apache.hudi.common.table.{HoodieTableMetaClient, TableSchemaResolver}
-
-import java.util.stream.Collectors
import org.apache.hadoop.fs.{GlobPattern, Path}
import org.apache.hudi.client.common.HoodieSparkEngineContext
import org.apache.hudi.client.utils.SparkInternalSchemaConverter
import org.apache.hudi.common.fs.FSUtils
+import org.apache.hudi.common.model.{HoodieCommitMetadata, HoodieFileFormat, HoodieRecord, HoodieReplaceCommitMetadata}
import org.apache.hudi.common.table.timeline.{HoodieInstant, HoodieTimeline}
+import org.apache.hudi.common.table.{HoodieTableMetaClient, TableSchemaResolver}
import org.apache.hudi.common.util.{HoodieTimer, InternalSchemaCache}
import org.apache.hudi.config.HoodieWriteConfig
import org.apache.hudi.exception.HoodieException
@@ -41,6 +39,7 @@ import org.apache.spark.sql.sources.{BaseRelation, TableScan}
import org.apache.spark.sql.types.StructType
import org.apache.spark.sql.{DataFrame, Row, SQLContext}
+import java.util.stream.Collectors
import scala.collection.JavaConversions._
import scala.collection.mutable
@@ -219,7 +218,7 @@ class IncrementalRelation(val sqlContext: SQLContext,
if (fallbackToFullTableScan) {
val fs = basePath.getFileSystem(sqlContext.sparkContext.hadoopConfiguration);
- val timer = new HoodieTimer().startTimer();
+ val timer = HoodieTimer.start
val allFilesToCheck = filteredMetaBootstrapFullPaths ++ filteredRegularFullPaths
val firstNotFoundPath = allFilesToCheck.find(path => !fs.exists(new Path(path)))
diff --git a/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/CreateMetadataTableProcedure.scala b/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/CreateMetadataTableProcedure.scala
index 3a16d8319a..bbed979f5c 100644
--- a/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/CreateMetadataTableProcedure.scala
+++ b/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/CreateMetadataTableProcedure.scala
@@ -61,7 +61,7 @@ class CreateMetadataTableProcedure extends BaseProcedure with ProcedureBuilder w
// Metadata directory does not exist yet
metaClient.getFs.mkdirs(metadataPath)
}
- val timer = new HoodieTimer().startTimer
+ val timer = HoodieTimer.start
val writeConfig = getWriteConfig(basePath)
SparkHoodieBackedTableMetadataWriter.create(metaClient.getHadoopConf, writeConfig, new HoodieSparkEngineContext(jsc))
Seq(Row("Created Metadata Table in " + metadataPath + " (duration=" + timer.endTimer / 1000.0 + "secs)"))
diff --git a/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/InitMetadataTableProcedure.scala b/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/InitMetadataTableProcedure.scala
index 73d1128a98..3b875e77ff 100644
--- a/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/InitMetadataTableProcedure.scala
+++ b/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/InitMetadataTableProcedure.scala
@@ -61,7 +61,7 @@ class InitMetadataTableProcedure extends BaseProcedure with ProcedureBuilder wit
throw new RuntimeException("Metadata directory (" + metadataPath.toString + ") does not exist.")
}
- val timer = new HoodieTimer().startTimer
+ val timer = HoodieTimer.start
if (!readOnly) {
val writeConfig = getWriteConfig(basePath)
SparkHoodieBackedTableMetadataWriter.create(metaClient.getHadoopConf, writeConfig, new HoodieSparkEngineContext(jsc))
diff --git a/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/RunCompactionProcedure.scala b/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/RunCompactionProcedure.scala
index 3e5a7e29e4..bd2e863b1e 100644
--- a/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/RunCompactionProcedure.scala
+++ b/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/RunCompactionProcedure.scala
@@ -28,7 +28,6 @@ import org.apache.spark.sql.Row
import org.apache.spark.sql.types._
import java.util.function.Supplier
-
import scala.collection.JavaConversions._
import scala.collection.JavaConverters._
@@ -107,8 +106,7 @@ class RunCompactionProcedure extends BaseProcedure with ProcedureBuilder with Sp
logInfo(s"No need to compaction on $basePath")
} else {
logInfo(s"Run compaction at instants: [${willCompactionInstants.mkString(",")}] on $basePath")
- val timer = new HoodieTimer
- timer.startTimer()
+ val timer = HoodieTimer.start
willCompactionInstants.foreach { compactionInstant =>
val writeResponse = client.compact(compactionInstant)
handleResponse(writeResponse.getCommitMetadata.get())
diff --git a/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/ShowMetadataTableFilesProcedure.scala b/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/ShowMetadataTableFilesProcedure.scala
index b30203dc06..edd0439a2b 100644
--- a/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/ShowMetadataTableFilesProcedure.scala
+++ b/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/ShowMetadataTableFilesProcedure.scala
@@ -65,7 +65,7 @@ class ShowMetadataTableFilesProcedure() extends BaseProcedure with ProcedureBuil
partitionPath = new Path(basePath, partition)
}
- val timer = new HoodieTimer().startTimer
+ val timer = HoodieTimer.start
val statuses = metaReader.getAllFilesInPartition(partitionPath)
logDebug("Took " + timer.endTimer + " ms")
diff --git a/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/ShowMetadataTablePartitionsProcedure.scala b/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/ShowMetadataTablePartitionsProcedure.scala
index f2eaa7ad83..f9a676abc9 100644
--- a/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/ShowMetadataTablePartitionsProcedure.scala
+++ b/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/ShowMetadataTablePartitionsProcedure.scala
@@ -57,7 +57,7 @@ class ShowMetadataTablePartitionsProcedure() extends BaseProcedure with Procedur
throw new HoodieException(s"Metadata Table not enabled/initialized.")
}
- val timer = new HoodieTimer().startTimer
+ val timer = HoodieTimer.start
val partitions = metadata.getAllPartitionPaths
Collections.sort(partitions)
logDebug("Took " + timer.endTimer + " ms")
diff --git a/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/ValidateMetadataTableFilesProcedure.scala b/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/ValidateMetadataTableFilesProcedure.scala
index 81540d9684..6d7457772b 100644
--- a/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/ValidateMetadataTableFilesProcedure.scala
+++ b/hudi-spark-datasource/hudi-spark/src/main/scala/org/apache/spark/sql/hudi/command/procedures/ValidateMetadataTableFilesProcedure.scala
@@ -73,7 +73,7 @@ class ValidateMetadataTableFilesProcedure() extends BaseProcedure with Procedure
val fsMetaReader = new HoodieBackedTableMetadata(new HoodieLocalEngineContext(metaClient.getHadoopConf),
fsConfig, basePath, "/tmp")
- val timer = new HoodieTimer().startTimer
+ val timer = HoodieTimer.start
val metadataPartitions = metadataReader.getAllPartitionPaths
logDebug("Listing partitions Took " + timer.endTimer + " ms")
val fsPartitions = fsMetaReader.getAllPartitionPaths
diff --git a/hudi-sync/hudi-hive-sync/src/main/java/org/apache/hudi/hive/ddl/HiveQueryDDLExecutor.java b/hudi-sync/hudi-hive-sync/src/main/java/org/apache/hudi/hive/ddl/HiveQueryDDLExecutor.java
index 90efd2701c..93ae3cfbf7 100644
--- a/hudi-sync/hudi-hive-sync/src/main/java/org/apache/hudi/hive/ddl/HiveQueryDDLExecutor.java
+++ b/hudi-sync/hudi-hive-sync/src/main/java/org/apache/hudi/hive/ddl/HiveQueryDDLExecutor.java
@@ -91,7 +91,7 @@ public class HiveQueryDDLExecutor extends QueryBasedDDLExecutor {
try {
for (String sql : sqls) {
if (hiveDriver != null) {
- HoodieTimer timer = new HoodieTimer().startTimer();
+ HoodieTimer timer = HoodieTimer.start();
responses.add(hiveDriver.run(sql));
LOG.info(String.format("Time taken to execute [%s]: %s ms", sql, timer.endTimer()));
}
diff --git a/hudi-timeline-service/src/main/java/org/apache/hudi/timeline/service/RequestHandler.java b/hudi-timeline-service/src/main/java/org/apache/hudi/timeline/service/RequestHandler.java
index b53c2534bc..5595c2b848 100644
--- a/hudi-timeline-service/src/main/java/org/apache/hudi/timeline/service/RequestHandler.java
+++ b/hudi-timeline-service/src/main/java/org/apache/hudi/timeline/service/RequestHandler.java
@@ -114,7 +114,7 @@ public class RequestHandler {
public static String jsonifyResult(
Context ctx, Object obj, Registry metricsRegistry, ObjectMapper objectMapper, Logger logger)
throws JsonProcessingException {
- HoodieTimer timer = new HoodieTimer().startTimer();
+ HoodieTimer timer = HoodieTimer.start();
boolean prettyPrint = ctx.queryParam("pretty") != null;
String result =
prettyPrint ? objectMapper.writerWithDefaultPrettyPrinter().writeValueAsString(obj)
diff --git a/hudi-timeline-service/src/main/java/org/apache/hudi/timeline/service/handlers/marker/BatchedMarkerCreationRunnable.java b/hudi-timeline-service/src/main/java/org/apache/hudi/timeline/service/handlers/marker/BatchedMarkerCreationRunnable.java
index 50b9913f5c..2416a5590c 100644
--- a/hudi-timeline-service/src/main/java/org/apache/hudi/timeline/service/handlers/marker/BatchedMarkerCreationRunnable.java
+++ b/hudi-timeline-service/src/main/java/org/apache/hudi/timeline/service/handlers/marker/BatchedMarkerCreationRunnable.java
@@ -40,7 +40,7 @@ public class BatchedMarkerCreationRunnable implements Runnable {
@Override
public void run() {
LOG.debug("Start processing create marker requests");
- HoodieTimer timer = new HoodieTimer().startTimer();
+ HoodieTimer timer = HoodieTimer.start();
for (BatchedMarkerCreationContext requestContext : requestContextList) {
requestContext.getMarkerDirState().processMarkerCreationRequests(
diff --git a/hudi-timeline-service/src/main/java/org/apache/hudi/timeline/service/handlers/marker/MarkerCreationFuture.java b/hudi-timeline-service/src/main/java/org/apache/hudi/timeline/service/handlers/marker/MarkerCreationFuture.java
index d965e56a01..e1deb28527 100644
--- a/hudi-timeline-service/src/main/java/org/apache/hudi/timeline/service/handlers/marker/MarkerCreationFuture.java
+++ b/hudi-timeline-service/src/main/java/org/apache/hudi/timeline/service/handlers/marker/MarkerCreationFuture.java
@@ -39,7 +39,7 @@ public class MarkerCreationFuture extends CompletableFuture<String> {
public MarkerCreationFuture(Context context, String markerDirPath, String markerName) {
super();
- this.timer = new HoodieTimer().startTimer();
+ this.timer = HoodieTimer.start();
this.context = context;
this.markerDirPath = markerDirPath;
this.markerName = markerName;
diff --git a/hudi-timeline-service/src/main/java/org/apache/hudi/timeline/service/handlers/marker/MarkerDirState.java b/hudi-timeline-service/src/main/java/org/apache/hudi/timeline/service/handlers/marker/MarkerDirState.java
index 67e850bb7d..f367ec870e 100644
--- a/hudi-timeline-service/src/main/java/org/apache/hudi/timeline/service/handlers/marker/MarkerDirState.java
+++ b/hudi-timeline-service/src/main/java/org/apache/hudi/timeline/service/handlers/marker/MarkerDirState.java
@@ -313,7 +313,7 @@ public class MarkerDirState implements Serializable {
*/
private void flushMarkersToFile(int markerFileIndex) {
LOG.debug("Write to " + markerDirPath + "/" + MARKERS_FILENAME_PREFIX + markerFileIndex);
- HoodieTimer timer = new HoodieTimer().startTimer();
+ HoodieTimer timer = HoodieTimer.start();
Path markersFilePath = new Path(markerDirPath, MARKERS_FILENAME_PREFIX + markerFileIndex);
FSDataOutputStream fsDataOutputStream = null;
BufferedWriter bufferedWriter = null;