You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@iceberg.apache.org by ao...@apache.org on 2023/02/17 23:44:45 UTC
[iceberg] branch master updated: Spark 3.3: Delete orphan files JMH benchmark (#5615)
This is an automated email from the ASF dual-hosted git repository.
aokolnychyi pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/iceberg.git
The following commit(s) were added to refs/heads/master by this push:
new f7470b84b3 Spark 3.3: Delete orphan files JMH benchmark (#5615)
f7470b84b3 is described below
commit f7470b84b334b31ae9e4ac6ac61117cb6be31586
Author: Karuppayya <ka...@gmail.com>
AuthorDate: Fri Feb 17 15:44:39 2023 -0800
Spark 3.3: Delete orphan files JMH benchmark (#5615)
---
.../spark/action/DeleteOrphanFilesBenchmark.java | 179 +++++++++++++++++++++
1 file changed, 179 insertions(+)
diff --git a/spark/v3.3/spark/src/jmh/java/org/apache/iceberg/spark/action/DeleteOrphanFilesBenchmark.java b/spark/v3.3/spark/src/jmh/java/org/apache/iceberg/spark/action/DeleteOrphanFilesBenchmark.java
new file mode 100644
index 0000000000..5a7df72837
--- /dev/null
+++ b/spark/v3.3/spark/src/jmh/java/org/apache/iceberg/spark/action/DeleteOrphanFilesBenchmark.java
@@ -0,0 +1,179 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements. See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership. The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing,
+ * software distributed under the License is distributed on an
+ * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+ * KIND, either express or implied. See the License for the
+ * specific language governing permissions and limitations
+ * under the License.
+ */
+package org.apache.iceberg.spark.action;
+
+import static org.apache.spark.sql.functions.lit;
+
+import java.sql.Timestamp;
+import java.util.List;
+import java.util.UUID;
+import java.util.concurrent.TimeUnit;
+import org.apache.iceberg.AppendFiles;
+import org.apache.iceberg.DataFile;
+import org.apache.iceberg.DataFiles;
+import org.apache.iceberg.PartitionSpec;
+import org.apache.iceberg.Table;
+import org.apache.iceberg.actions.DeleteOrphanFiles;
+import org.apache.iceberg.relocated.com.google.common.collect.Lists;
+import org.apache.iceberg.relocated.com.google.common.io.Files;
+import org.apache.iceberg.spark.Spark3Util;
+import org.apache.iceberg.spark.SparkSessionCatalog;
+import org.apache.iceberg.spark.actions.SparkActions;
+import org.apache.spark.sql.Dataset;
+import org.apache.spark.sql.Encoders;
+import org.apache.spark.sql.Row;
+import org.apache.spark.sql.SparkSession;
+import org.openjdk.jmh.annotations.Benchmark;
+import org.openjdk.jmh.annotations.BenchmarkMode;
+import org.openjdk.jmh.annotations.Fork;
+import org.openjdk.jmh.annotations.Measurement;
+import org.openjdk.jmh.annotations.Mode;
+import org.openjdk.jmh.annotations.Scope;
+import org.openjdk.jmh.annotations.Setup;
+import org.openjdk.jmh.annotations.State;
+import org.openjdk.jmh.annotations.TearDown;
+import org.openjdk.jmh.annotations.Threads;
+import org.openjdk.jmh.annotations.Timeout;
+import org.openjdk.jmh.annotations.Warmup;
+import org.openjdk.jmh.infra.Blackhole;
+
+/**
+ * A benchmark that evaluates the performance of remove orphan files action in Spark.
+ *
+ * <p>To run this benchmark for spark-3.3: <code>
+ * ./gradlew -DsparkVersions=3.3 :iceberg-spark:iceberg-spark-3.3_2.12:jmh
+ * -PjmhIncludeRegex=DeleteOrphanFilesBenchmark
+ * -PjmhOutputPath=benchmark/delete-orphan-files-benchmark-results.txt
+ * </code>
+ */
+@Fork(1)
+@State(Scope.Benchmark)
+@Warmup(iterations = 3)
+@Measurement(iterations = 5)
+@BenchmarkMode(Mode.SingleShotTime)
+@Timeout(time = 1000, timeUnit = TimeUnit.HOURS)
+public class DeleteOrphanFilesBenchmark {
+
+ private static final String TABLE_NAME = "delete_orphan_perf";
+ private static final int NUM_SNAPSHOTS = 1000;
+ private static final int NUM_FILES = 1000;
+
+ private SparkSession spark;
+ private final List<String> validAndOrphanPaths = Lists.newArrayList();
+ private Table table;
+
+ @Setup
+ public void setupBench() {
+ setupSpark();
+ initTable();
+ appendData();
+ addOrphans();
+ }
+
+ @TearDown
+ public void teardownBench() {
+ tearDownSpark();
+ }
+
+ @Benchmark
+ @Threads(1)
+ public void testDeleteOrphanFiles(Blackhole blackhole) {
+ Dataset<Row> validAndOrphanPathsDF =
+ spark
+ .createDataset(validAndOrphanPaths, Encoders.STRING())
+ .withColumnRenamed("value", "file_path")
+ .withColumn("last_modified", lit(new Timestamp(10000)));
+
+ DeleteOrphanFiles.Result results =
+ SparkActions.get(spark)
+ .deleteOrphanFiles(table())
+ .compareToFileList(validAndOrphanPathsDF)
+ .execute();
+ blackhole.consume(results);
+ }
+
+ private void initTable() {
+ spark.sql(
+ String.format(
+ "CREATE TABLE %s(id INT, name STRING)"
+ + " USING ICEBERG"
+ + " TBLPROPERTIES ( 'format-version' = '2')",
+ TABLE_NAME));
+ }
+
+ private void appendData() {
+ String location = table().location();
+ PartitionSpec partitionSpec = table().spec();
+
+ for (int i = 0; i < NUM_SNAPSHOTS; i++) {
+ AppendFiles appendFiles = table().newFastAppend();
+ for (int j = 0; j < NUM_FILES; j++) {
+ String path = String.format("%s/path/to/data-%d-%d.parquet", location, i, j);
+ validAndOrphanPaths.add(path);
+ DataFile dataFile =
+ DataFiles.builder(partitionSpec)
+ .withPath(path)
+ .withFileSizeInBytes(10)
+ .withRecordCount(1)
+ .build();
+ appendFiles.appendFile(dataFile);
+ }
+ appendFiles.commit();
+ }
+ }
+
+ private void addOrphans() {
+ String location = table.location();
+ // Generate 10% orphan files
+ int orphanFileCount = (NUM_FILES * NUM_SNAPSHOTS) / 10;
+ for (int i = 0; i < orphanFileCount; i++) {
+ validAndOrphanPaths.add(
+ String.format("%s/path/to/data-%s.parquet", location, UUID.randomUUID()));
+ }
+ }
+
+ private Table table() {
+ if (table == null) {
+ try {
+ table = Spark3Util.loadIcebergTable(spark, TABLE_NAME);
+ } catch (Exception e) {
+ throw new RuntimeException(e);
+ }
+ }
+ return table;
+ }
+
+ private String catalogWarehouse() {
+ return Files.createTempDir().getAbsolutePath() + "/" + UUID.randomUUID() + "/";
+ }
+
+ private void setupSpark() {
+ SparkSession.Builder builder =
+ SparkSession.builder()
+ .config("spark.sql.catalog.spark_catalog", SparkSessionCatalog.class.getName())
+ .config("spark.sql.catalog.spark_catalog.type", "hadoop")
+ .config("spark.sql.catalog.spark_catalog.warehouse", catalogWarehouse())
+ .master("local");
+ spark = builder.getOrCreate();
+ }
+
+ private void tearDownSpark() {
+ spark.stop();
+ }
+}