You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by do...@apache.org on 2022/07/12 18:01:48 UTC
[spark] branch master updated: [SPARK-39694][TESTS] Use `${projectName}/Test/runMain` to run benchmarks
This is an automated email from the ASF dual-hosted git repository.
dongjoon pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git
The following commit(s) were added to refs/heads/master by this push:
new 7ef883f0b13 [SPARK-39694][TESTS] Use `${projectName}/Test/runMain` to run benchmarks
7ef883f0b13 is described below
commit 7ef883f0b1362068929a6bc79cc69aa9d1e431d9
Author: yangjie01 <ya...@baidu.com>
AuthorDate: Tue Jul 12 11:01:22 2022 -0700
[SPARK-39694][TESTS] Use `${projectName}/Test/runMain` to run benchmarks
### What changes were proposed in this pull request?
The following warnings are displayed when
`build/sbt "core/test:runMain org.apache.spark.MapStatusesSerDeserBenchmark"`
is executed:
```
[warn] sbt 0.13 shell syntax is deprecated; use slash syntax instead: core / Test / runMain
```
It seems that the recommended command in the scala doc is deprecated, to this pr do the following changes:
- Update all `build/sbt "${projectName}/test:runMain <this class>"` to `build/sbt "avro/Test/runMain <this class>"` in scala doc
- Change to use `build/sbt "sql/Test/runMain org.apache.spark.sql.GenTPCDSData ...` to generate TPC-DS table data in `benchmark.yml` and `build_and_test.yml`
### Why are the changes needed?
Use the recommended sbt syntax.
### Does this PR introduce _any_ user-facing change?
No
### How was this patch tested?
- Pass Github Actions
- Manual test:
1. `Generate an input dataset for TPCDSQueryBenchmark with SF=1` run successful with this pr
2. Manually executing some commands like
```
build/sbt "core /Test/runMain org.apache.spark.MapStatusesSerDeserBenchmark"
build/sbt "avro/Test/runMain org.apache.spark.sql.execution.benchmark.AvroReadBenchmark"
build/sbt "mllib/Test/runMain org.apache.spark.mllib.linalg.UDTSerializationBenchmark"
```
can run successfully, and there are no warning messages above.
Closes #37102 from LuciferYang/SPARK-39694.
Authored-by: yangjie01 <ya...@baidu.com>
Signed-off-by: Dongjoon Hyun <do...@apache.org>
---
.github/workflows/benchmark.yml | 2 +-
.github/workflows/build_and_test.yml | 2 +-
.../apache/spark/sql/execution/benchmark/AvroReadBenchmark.scala | 4 ++--
.../apache/spark/sql/execution/benchmark/AvroWriteBenchmark.scala | 4 ++--
.../test/scala/org/apache/spark/MapStatusesConvertBenchmark.scala | 4 ++--
.../scala/org/apache/spark/MapStatusesSerDeserBenchmark.scala | 4 ++--
core/src/test/scala/org/apache/spark/io/ZStandardBenchmark.scala | 4 ++--
.../test/scala/org/apache/spark/rdd/CoalescedRDDBenchmark.scala | 4 ++--
.../test/scala/org/apache/spark/serializer/KryoBenchmark.scala | 4 ++--
.../org/apache/spark/serializer/KryoSerializerBenchmark.scala | 4 ++--
.../scala/org/apache/spark/util/PropertiesCloneBenchmark.scala | 4 ++--
.../org/apache/spark/util/random/XORShiftRandomBenchmark.scala | 4 ++--
.../src/test/scala/org/apache/spark/ml/linalg/BLASBenchmark.scala | 4 ++--
.../org/apache/spark/mllib/linalg/UDTSerializationBenchmark.scala | 4 ++--
.../src/test/scala/org/apache/spark/sql/HashBenchmark.scala | 4 ++--
.../test/scala/org/apache/spark/sql/HashByteArrayBenchmark.scala | 4 ++--
.../scala/org/apache/spark/sql/UnsafeProjectionBenchmark.scala | 4 ++--
.../spark/sql/catalyst/util/GenericArrayDataBenchmark.scala | 4 ++--
.../apache/spark/sql/connector/catalog/EnumTypeSetBenchmark.scala | 4 ++--
.../src/test/scala/org/apache/spark/sql/DatasetBenchmark.scala | 4 ++--
sql/core/src/test/scala/org/apache/spark/sql/GenTPCDSData.scala | 2 +-
.../test/scala/org/apache/spark/sql/UpdateFieldsBenchmark.scala | 4 ++--
.../spark/sql/connector/functions/V2FunctionBenchmark.scala | 4 ++--
.../apache/spark/sql/execution/SubExprEliminationBenchmark.scala | 4 ++--
.../apache/spark/sql/execution/benchmark/AggregateBenchmark.scala | 4 ++--
.../spark/sql/execution/benchmark/AnsiIntervalSortBenchmark.scala | 4 ++--
.../apache/spark/sql/execution/benchmark/Base64Benchmark.scala | 4 ++--
.../spark/sql/execution/benchmark/BloomFilterBenchmark.scala | 4 ++--
.../sql/execution/benchmark/BuiltInDataSourceWriteBenchmark.scala | 8 ++++----
.../apache/spark/sql/execution/benchmark/ByteArrayBenchmark.scala | 4 ++--
.../spark/sql/execution/benchmark/CharVarcharBenchmark.scala | 4 ++--
.../sql/execution/benchmark/ConstantColumnVectorBenchmark.scala | 4 ++--
.../spark/sql/execution/benchmark/DataSourceReadBenchmark.scala | 4 ++--
.../apache/spark/sql/execution/benchmark/DateTimeBenchmark.scala | 4 ++--
.../spark/sql/execution/benchmark/DateTimeRebaseBenchmark.scala | 4 ++--
.../apache/spark/sql/execution/benchmark/ExtractBenchmark.scala | 4 ++--
.../spark/sql/execution/benchmark/FilterPushdownBenchmark.scala | 4 ++--
.../spark/sql/execution/benchmark/GenerateExecBenchmark.scala | 4 ++--
.../sql/execution/benchmark/HashedRelationMetricsBenchmark.scala | 4 ++--
.../spark/sql/execution/benchmark/InExpressionBenchmark.scala | 4 ++--
.../benchmark/InsertTableWithDynamicPartitionsBenchmark.scala | 4 ++--
.../apache/spark/sql/execution/benchmark/IntervalBenchmark.scala | 4 ++--
.../org/apache/spark/sql/execution/benchmark/JoinBenchmark.scala | 4 ++--
.../spark/sql/execution/benchmark/MakeDateTimeBenchmark.scala | 4 ++--
.../org/apache/spark/sql/execution/benchmark/MiscBenchmark.scala | 4 ++--
.../sql/execution/benchmark/OrcNestedSchemaPruningBenchmark.scala | 4 ++--
.../execution/benchmark/OrcV2NestedSchemaPruningBenchmark.scala | 4 ++--
.../benchmark/ParquetNestedPredicatePushDownBenchmark.scala | 4 ++--
.../execution/benchmark/ParquetNestedSchemaPruningBenchmark.scala | 4 ++--
.../spark/sql/execution/benchmark/PrimitiveArrayBenchmark.scala | 4 ++--
.../org/apache/spark/sql/execution/benchmark/RangeBenchmark.scala | 4 ++--
.../org/apache/spark/sql/execution/benchmark/SortBenchmark.scala | 4 ++--
.../execution/benchmark/StateStoreBasicOperationsBenchmark.scala | 4 ++--
.../spark/sql/execution/benchmark/TPCDSQueryBenchmark.scala | 4 ++--
.../org/apache/spark/sql/execution/benchmark/UDFBenchmark.scala | 4 ++--
.../spark/sql/execution/benchmark/UnsafeArrayDataBenchmark.scala | 4 ++--
.../spark/sql/execution/benchmark/WideSchemaBenchmark.scala | 4 ++--
.../apache/spark/sql/execution/benchmark/WideTableBenchmark.scala | 4 ++--
.../spark/sql/execution/columnar/InMemoryColumnarBenchmark.scala | 4 ++--
.../columnar/compression/CompressionSchemeBenchmark.scala | 4 ++--
.../apache/spark/sql/execution/datasources/csv/CSVBenchmark.scala | 4 ++--
.../spark/sql/execution/datasources/json/JsonBenchmark.scala | 4 ++--
.../spark/sql/execution/ui/MetricsAggregationBenchmark.scala | 4 ++--
.../spark/sql/execution/vectorized/ColumnarBatchBenchmark.scala | 4 ++--
.../sql/execution/benchmark/InsertIntoHiveTableBenchmark.scala | 4 ++--
.../execution/benchmark/ObjectHashAggregateExecBenchmark.scala | 4 ++--
.../scala/org/apache/spark/sql/hive/orc/OrcReadBenchmark.scala | 4 ++--
67 files changed, 133 insertions(+), 133 deletions(-)
diff --git a/.github/workflows/benchmark.yml b/.github/workflows/benchmark.yml
index 4a5fd661c78..23fa0ccb218 100644
--- a/.github/workflows/benchmark.yml
+++ b/.github/workflows/benchmark.yml
@@ -110,7 +110,7 @@ jobs:
java-version: ${{ github.event.inputs.jdk }}
- name: Generate TPC-DS (SF=1) table data
if: steps.cache-tpcds-sf-1.outputs.cache-hit != 'true'
- run: build/sbt "sql/test:runMain org.apache.spark.sql.GenTPCDSData --dsdgenDir `pwd`/tpcds-kit/tools --location `pwd`/tpcds-sf-1 --scaleFactor 1 --numPartitions 1 --overwrite"
+ run: build/sbt "sql/Test/runMain org.apache.spark.sql.GenTPCDSData --dsdgenDir `pwd`/tpcds-kit/tools --location `pwd`/tpcds-sf-1 --scaleFactor 1 --numPartitions 1 --overwrite"
benchmark:
name: "Run benchmarks: ${{ github.event.inputs.class }} (JDK ${{ github.event.inputs.jdk }}, Scala ${{ github.event.inputs.scala }}, ${{ matrix.split }} out of ${{ github.event.inputs.num-splits }} splits)"
diff --git a/.github/workflows/build_and_test.yml b/.github/workflows/build_and_test.yml
index 43103dc31c1..133f152df50 100644
--- a/.github/workflows/build_and_test.yml
+++ b/.github/workflows/build_and_test.yml
@@ -767,7 +767,7 @@ jobs:
run: cd tpcds-kit/tools && make OS=LINUX
- name: Generate TPC-DS (SF=1) table data
if: steps.cache-tpcds-sf-1.outputs.cache-hit != 'true'
- run: build/sbt "sql/test:runMain org.apache.spark.sql.GenTPCDSData --dsdgenDir `pwd`/tpcds-kit/tools --location `pwd`/tpcds-sf-1 --scaleFactor 1 --numPartitions 1 --overwrite"
+ run: build/sbt "sql/Test/runMain org.apache.spark.sql.GenTPCDSData --dsdgenDir `pwd`/tpcds-kit/tools --location `pwd`/tpcds-sf-1 --scaleFactor 1 --numPartitions 1 --overwrite"
- name: Run TPC-DS queries (Sort merge join)
run: |
SPARK_TPCDS_DATA=`pwd`/tpcds-sf-1 build/sbt "sql/testOnly org.apache.spark.sql.TPCDSQueryTestSuite"
diff --git a/connector/avro/src/test/scala/org/apache/spark/sql/execution/benchmark/AvroReadBenchmark.scala b/connector/avro/src/test/scala/org/apache/spark/sql/execution/benchmark/AvroReadBenchmark.scala
index 7368543642b..aa0d713bbfb 100644
--- a/connector/avro/src/test/scala/org/apache/spark/sql/execution/benchmark/AvroReadBenchmark.scala
+++ b/connector/avro/src/test/scala/org/apache/spark/sql/execution/benchmark/AvroReadBenchmark.scala
@@ -33,8 +33,8 @@ import org.apache.spark.sql.types._
* To run this benchmark:
* 1. without sbt: bin/spark-submit --class <this class>
* --jars <catalyst test jar>,<core test jar>,<sql test jar>,<spark-avro jar> <avro test jar>
- * 2. build/sbt "avro/test:runMain <this class>"
- * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "avro/test:runMain <this class>"
+ * 2. build/sbt "avro/Test/runMain <this class>"
+ * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "avro/Test/runMain <this class>"
* Results will be written to "benchmarks/AvroReadBenchmark-results.txt".
* }}}
*/
diff --git a/connector/avro/src/test/scala/org/apache/spark/sql/execution/benchmark/AvroWriteBenchmark.scala b/connector/avro/src/test/scala/org/apache/spark/sql/execution/benchmark/AvroWriteBenchmark.scala
index 7f9febb5b14..585ef8f361b 100644
--- a/connector/avro/src/test/scala/org/apache/spark/sql/execution/benchmark/AvroWriteBenchmark.scala
+++ b/connector/avro/src/test/scala/org/apache/spark/sql/execution/benchmark/AvroWriteBenchmark.scala
@@ -30,8 +30,8 @@ import org.apache.spark.storage.StorageLevel
* --jars <spark core test jar>,<spark catalyst test jar>,
* <spark sql test jar>,<spark avro jar>
* <spark avro test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
- * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "avro/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
+ * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "avro/Test/runMain <this class>"
* Results will be written to "benchmarks/AvroWriteBenchmark-results.txt".
* }}}
*/
diff --git a/core/src/test/scala/org/apache/spark/MapStatusesConvertBenchmark.scala b/core/src/test/scala/org/apache/spark/MapStatusesConvertBenchmark.scala
index 7f25c86497f..df100283d52 100644
--- a/core/src/test/scala/org/apache/spark/MapStatusesConvertBenchmark.scala
+++ b/core/src/test/scala/org/apache/spark/MapStatusesConvertBenchmark.scala
@@ -29,9 +29,9 @@ import org.apache.spark.storage.BlockManagerId
* {{{
* 1. without sbt:
* bin/spark-submit --class <this class> --jars <spark core test jar>
- * 2. build/sbt "core/test:runMain <this class>"
+ * 2. build/sbt "core/Test/runMain <this class>"
* 3. generate result:
- * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "core/test:runMain <this class>"
+ * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "core/Test/runMain <this class>"
* Results will be written to "benchmarks/MapStatusesConvertBenchmark-results.txt".
* }}}
* */
diff --git a/core/src/test/scala/org/apache/spark/MapStatusesSerDeserBenchmark.scala b/core/src/test/scala/org/apache/spark/MapStatusesSerDeserBenchmark.scala
index bb627bb181d..797b650799e 100644
--- a/core/src/test/scala/org/apache/spark/MapStatusesSerDeserBenchmark.scala
+++ b/core/src/test/scala/org/apache/spark/MapStatusesSerDeserBenchmark.scala
@@ -27,8 +27,8 @@ import org.apache.spark.storage.BlockManagerId
* {{{
* To run this benchmark:
* 1. without sbt: bin/spark-submit --class <this class> <spark core test jar>
- * 2. build/sbt "core/test:runMain <this class>"
- * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "core/test:runMain <this class>"
+ * 2. build/sbt "core/Test/runMain <this class>"
+ * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "core/Test/runMain <this class>"
* Results will be written to "benchmarks/MapStatusesSerDeserBenchmark-results.txt".
* }}}
*/
diff --git a/core/src/test/scala/org/apache/spark/io/ZStandardBenchmark.scala b/core/src/test/scala/org/apache/spark/io/ZStandardBenchmark.scala
index 62a3d48d2ec..e23416177ae 100644
--- a/core/src/test/scala/org/apache/spark/io/ZStandardBenchmark.scala
+++ b/core/src/test/scala/org/apache/spark/io/ZStandardBenchmark.scala
@@ -29,8 +29,8 @@ import org.apache.spark.internal.config.{IO_COMPRESSION_ZSTD_BUFFERPOOL_ENABLED,
* {{{
* To run this benchmark:
* 1. without sbt: bin/spark-submit --class <this class> <spark core test jar>
- * 2. build/sbt "core/test:runMain <this class>"
- * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "core/test:runMain <this class>"
+ * 2. build/sbt "core/Test/runMain <this class>"
+ * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "core/Test/runMain <this class>"
* Results will be written to "benchmarks/ZStandardBenchmark-results.txt".
* }}}
*/
diff --git a/core/src/test/scala/org/apache/spark/rdd/CoalescedRDDBenchmark.scala b/core/src/test/scala/org/apache/spark/rdd/CoalescedRDDBenchmark.scala
index b622e0b1d6e..ccd96b14d63 100644
--- a/core/src/test/scala/org/apache/spark/rdd/CoalescedRDDBenchmark.scala
+++ b/core/src/test/scala/org/apache/spark/rdd/CoalescedRDDBenchmark.scala
@@ -30,9 +30,9 @@ import org.apache.spark.benchmark.{Benchmark, BenchmarkBase}
* {{{
* 1. without sbt:
* bin/spark-submit --class <this class> <spark core test jar>
- * 2. build/sbt "core/test:runMain <this class>"
+ * 2. build/sbt "core/Test/runMain <this class>"
* 3. generate result:
- * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "core/test:runMain <this class>"
+ * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "core/Test/runMain <this class>"
* Results will be written to "benchmarks/CoalescedRDD-results.txt".
* }}}
* */
diff --git a/core/src/test/scala/org/apache/spark/serializer/KryoBenchmark.scala b/core/src/test/scala/org/apache/spark/serializer/KryoBenchmark.scala
index 1c17d7b1392..61665a2219c 100644
--- a/core/src/test/scala/org/apache/spark/serializer/KryoBenchmark.scala
+++ b/core/src/test/scala/org/apache/spark/serializer/KryoBenchmark.scala
@@ -32,9 +32,9 @@ import org.apache.spark.serializer.KryoTest._
* {{{
* 1. without sbt:
* bin/spark-submit --class <this class> <spark core test jar>
- * 2. build/sbt "core/test:runMain <this class>"
+ * 2. build/sbt "core/Test/runMain <this class>"
* 3. generate result:
- * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "core/test:runMain <this class>"
+ * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "core/Test/runMain <this class>"
* Results will be written to "benchmarks/KryoBenchmark-results.txt".
* }}}
*/
diff --git a/core/src/test/scala/org/apache/spark/serializer/KryoSerializerBenchmark.scala b/core/src/test/scala/org/apache/spark/serializer/KryoSerializerBenchmark.scala
index 28e0e79a6fd..e1e4c218e9c 100644
--- a/core/src/test/scala/org/apache/spark/serializer/KryoSerializerBenchmark.scala
+++ b/core/src/test/scala/org/apache/spark/serializer/KryoSerializerBenchmark.scala
@@ -38,9 +38,9 @@ import org.apache.spark.util.ThreadUtils
* {{{
* 1. without sbt:
* bin/spark-submit --class <this class> <spark core test jar>
- * 2. build/sbt "core/test:runMain <this class>"
+ * 2. build/sbt "core/Test/runMain <this class>"
* 3. generate result:
- * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "core/test:runMain <this class>"
+ * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "core/Test/runMain <this class>"
* Results will be written to "benchmarks/KryoSerializerBenchmark-results.txt".
* }}}
*/
diff --git a/core/src/test/scala/org/apache/spark/util/PropertiesCloneBenchmark.scala b/core/src/test/scala/org/apache/spark/util/PropertiesCloneBenchmark.scala
index ff4a4941b6b..35b1a2046c4 100644
--- a/core/src/test/scala/org/apache/spark/util/PropertiesCloneBenchmark.scala
+++ b/core/src/test/scala/org/apache/spark/util/PropertiesCloneBenchmark.scala
@@ -31,9 +31,9 @@ import org.apache.spark.benchmark.{Benchmark, BenchmarkBase}
* {{{
* 1. without sbt:
* bin/spark-submit --class <this class> <spark core test jar>
- * 2. build/sbt "core/test:runMain <this class>"
+ * 2. build/sbt "core/Test/runMain <this class>"
* 3. generate result:
- * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "core/test:runMain <this class>"
+ * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "core/Test/runMain <this class>"
* Results will be written to "benchmarks/PropertiesCloneBenchmark-results.txt".
* }}}
*/
diff --git a/core/src/test/scala/org/apache/spark/util/random/XORShiftRandomBenchmark.scala b/core/src/test/scala/org/apache/spark/util/random/XORShiftRandomBenchmark.scala
index 7fd63a2c351..3c588bb59ef 100644
--- a/core/src/test/scala/org/apache/spark/util/random/XORShiftRandomBenchmark.scala
+++ b/core/src/test/scala/org/apache/spark/util/random/XORShiftRandomBenchmark.scala
@@ -28,9 +28,9 @@ import org.apache.spark.util.Utils.times
* {{{
* 1. without sbt:
* bin/spark-submit --class <this class> <spark core test jar>
- * 2. build/sbt "core/test:runMain <this class>"
+ * 2. build/sbt "core/Test/runMain <this class>"
* 3. generate result:
- * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "core/test:runMain <this class>"
+ * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "core/Test/runMain <this class>"
* Results will be written to "benchmarks/XORShiftRandomBenchmark-results.txt".
* }}}
*/
diff --git a/mllib-local/src/test/scala/org/apache/spark/ml/linalg/BLASBenchmark.scala b/mllib-local/src/test/scala/org/apache/spark/ml/linalg/BLASBenchmark.scala
index abe6c56be65..245f1faa369 100644
--- a/mllib-local/src/test/scala/org/apache/spark/ml/linalg/BLASBenchmark.scala
+++ b/mllib-local/src/test/scala/org/apache/spark/ml/linalg/BLASBenchmark.scala
@@ -28,8 +28,8 @@ import org.apache.spark.benchmark.{Benchmark, BenchmarkBase}
* To run this benchmark:
* {{{
* 1. without sbt: bin/spark-submit --class <this class> <spark mllib test jar>
- * 2. build/sbt "mllib-local/test:runMain <this class>"
- * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "mllib/test:runMain <this class>"
+ * 2. build/sbt "mllib-local/Test/runMain <this class>"
+ * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "mllib/Test/runMain <this class>"
* Results will be written to "benchmarks/BLASBenchmark-results.txt".
* }}}
*/
diff --git a/mllib/src/test/scala/org/apache/spark/mllib/linalg/UDTSerializationBenchmark.scala b/mllib/src/test/scala/org/apache/spark/mllib/linalg/UDTSerializationBenchmark.scala
index e748e3288b6..37bbe76b9b5 100644
--- a/mllib/src/test/scala/org/apache/spark/mllib/linalg/UDTSerializationBenchmark.scala
+++ b/mllib/src/test/scala/org/apache/spark/mllib/linalg/UDTSerializationBenchmark.scala
@@ -27,8 +27,8 @@ import org.apache.spark.sql.catalyst.encoders.ExpressionEncoder
* 1. without sbt:
* bin/spark-submit --class <this class>
* --jars <spark core test jar> <spark mllib test jar>
- * 2. build/sbt "mllib/test:runMain <this class>"
- * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "mllib/test:runMain <this class>"
+ * 2. build/sbt "mllib/Test/runMain <this class>"
+ * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "mllib/Test/runMain <this class>"
* Results will be written to "benchmarks/UDTSerializationBenchmark-results.txt".
* }}}
*/
diff --git a/sql/catalyst/src/test/scala/org/apache/spark/sql/HashBenchmark.scala b/sql/catalyst/src/test/scala/org/apache/spark/sql/HashBenchmark.scala
index 3f0121bcf4a..8e96faace52 100644
--- a/sql/catalyst/src/test/scala/org/apache/spark/sql/HashBenchmark.scala
+++ b/sql/catalyst/src/test/scala/org/apache/spark/sql/HashBenchmark.scala
@@ -30,9 +30,9 @@ import org.apache.spark.sql.types._
* {{{
* 1. without sbt:
* bin/spark-submit --class <this class> --jars <spark core test jar> <spark catalyst test jar>
- * 2. build/sbt "catalyst/test:runMain <this class>"
+ * 2. build/sbt "catalyst/Test/runMain <this class>"
* 3. generate result:
- * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "catalyst/test:runMain <this class>"
+ * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "catalyst/Test/runMain <this class>"
* Results will be written to "benchmarks/HashBenchmark-results.txt".
* }}}
*/
diff --git a/sql/catalyst/src/test/scala/org/apache/spark/sql/HashByteArrayBenchmark.scala b/sql/catalyst/src/test/scala/org/apache/spark/sql/HashByteArrayBenchmark.scala
index dbfa7bb18aa..1baac88bf2d 100644
--- a/sql/catalyst/src/test/scala/org/apache/spark/sql/HashByteArrayBenchmark.scala
+++ b/sql/catalyst/src/test/scala/org/apache/spark/sql/HashByteArrayBenchmark.scala
@@ -30,9 +30,9 @@ import org.apache.spark.unsafe.hash.Murmur3_x86_32
* {{{
* 1. without sbt:
* bin/spark-submit --class <this class> --jars <spark core test jar> <spark catalyst test jar>
- * 2. build/sbt "catalyst/test:runMain <this class>"
+ * 2. build/sbt "catalyst/Test/runMain <this class>"
* 3. generate result:
- * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "catalyst/test:runMain <this class>"
+ * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "catalyst/Test/runMain <this class>"
* Results will be written to "benchmarks/HashByteArrayBenchmark-results.txt".
* }}}
*/
diff --git a/sql/catalyst/src/test/scala/org/apache/spark/sql/UnsafeProjectionBenchmark.scala b/sql/catalyst/src/test/scala/org/apache/spark/sql/UnsafeProjectionBenchmark.scala
index 352afaa2740..07179a20cd0 100644
--- a/sql/catalyst/src/test/scala/org/apache/spark/sql/UnsafeProjectionBenchmark.scala
+++ b/sql/catalyst/src/test/scala/org/apache/spark/sql/UnsafeProjectionBenchmark.scala
@@ -29,8 +29,8 @@ import org.apache.spark.sql.types._
* To run this benchmark:
* 1. without sbt:
* bin/spark-submit --class <this class> --jars <spark core test jar> <spark catalyst test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
- * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
+ * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/UnsafeProjectionBenchmark-results.txt".
* }}}
*/
diff --git a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/util/GenericArrayDataBenchmark.scala b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/util/GenericArrayDataBenchmark.scala
index 3ad045f29c0..a2800b3faae 100644
--- a/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/util/GenericArrayDataBenchmark.scala
+++ b/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/util/GenericArrayDataBenchmark.scala
@@ -25,9 +25,9 @@ import org.apache.spark.benchmark.{Benchmark, BenchmarkBase}
* {{{
* 1. without sbt:
* bin/spark-submit --class <this class> --jars <spark core test jar> <spark catalyst test jar>
- * 2. build/sbt "catalyst/test:runMain <this class>"
+ * 2. build/sbt "catalyst/Test/runMain <this class>"
* 3. generate result:
- * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "catalyst/test:runMain <this class>"
+ * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "catalyst/Test/runMain <this class>"
* Results will be written to "benchmarks/GenericArrayDataBenchmark-results.txt".
* }}}
*/
diff --git a/sql/catalyst/src/test/scala/org/apache/spark/sql/connector/catalog/EnumTypeSetBenchmark.scala b/sql/catalyst/src/test/scala/org/apache/spark/sql/connector/catalog/EnumTypeSetBenchmark.scala
index a918bae4a84..d8bc3ed28dd 100644
--- a/sql/catalyst/src/test/scala/org/apache/spark/sql/connector/catalog/EnumTypeSetBenchmark.scala
+++ b/sql/catalyst/src/test/scala/org/apache/spark/sql/connector/catalog/EnumTypeSetBenchmark.scala
@@ -30,9 +30,9 @@ import org.apache.spark.sql.connector.catalog.TableCapability._
* {{{
* 1. without sbt:
* bin/spark-submit --class <this class> --jars <spark core test jar> <spark catalyst test jar>
- * 2. build/sbt "catalyst/test:runMain <this class>"
+ * 2. build/sbt "catalyst/Test/runMain <this class>"
* 3. generate result:
- * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "catalyst/test:runMain <this class>"
+ * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "catalyst/Test/runMain <this class>"
* Results will be written to "benchmarks/EnumTypeSetBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/DatasetBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/DatasetBenchmark.scala
index 955d0f9193e..076332f68ac 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/DatasetBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/DatasetBenchmark.scala
@@ -31,8 +31,8 @@ import org.apache.spark.sql.types.StringType
* 1. without sbt:
* bin/spark-submit --class <this class>
* --jars <spark core test jar>,<spark catalyst test jar> <spark sql test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
- * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
+ * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/DatasetBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/GenTPCDSData.scala b/sql/core/src/test/scala/org/apache/spark/sql/GenTPCDSData.scala
index 104ea6c9cfe..1ab8b1196d5 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/GenTPCDSData.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/GenTPCDSData.scala
@@ -410,7 +410,7 @@ class GenTPCDSDataConfig(args: Array[String]) {
*
* To run this:
* {{{
- * build/sbt "sql/test:runMain <this class> --dsdgenDir <path> --location <path> --scaleFactor 1"
+ * build/sbt "sql/Test/runMain <this class> --dsdgenDir <path> --location <path> --scaleFactor 1"
* }}}
*/
object GenTPCDSData {
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/UpdateFieldsBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/UpdateFieldsBenchmark.scala
index 13221c9c79c..bfbf97f553b 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/UpdateFieldsBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/UpdateFieldsBenchmark.scala
@@ -31,9 +31,9 @@ import org.apache.spark.sql.types.{IntegerType, StructField, StructType}
* bin/spark-submit --class <this class>
* --jars <spark core test jar>,<spark catalyst test jar> <spark sql test jar>
* 2. with sbt:
- * build/sbt "sql/test:runMain <this class>"
+ * build/sbt "sql/Test/runMain <this class>"
* 3. generate result:
- * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/UpdateFieldsBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/connector/functions/V2FunctionBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/connector/functions/V2FunctionBenchmark.scala
index 16e86a7597e..38f016c2b63 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/connector/functions/V2FunctionBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/connector/functions/V2FunctionBenchmark.scala
@@ -40,8 +40,8 @@ import org.apache.spark.sql.types.{AbstractDataType, DataType, LongType, Numeric
* 1. without sbt:
* bin/spark-submit --class <this class>
* --jars <spark core test jar>,<spark catalyst test jar> <sql core test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
- * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
+ * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/V2FunctionBenchmark-results.txt".
* }}}
* '''NOTE''': to update the result of this benchmark, please use Github benchmark action:
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/SubExprEliminationBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/SubExprEliminationBenchmark.scala
index a72106837fe..e2ff7dc1c9a 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/SubExprEliminationBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/SubExprEliminationBenchmark.scala
@@ -31,9 +31,9 @@ import org.apache.spark.sql.internal.SQLConf
* 1. without sbt:
* bin/spark-submit --class <this class> --jars <spark core test jar>,
* <spark catalyst test jar> <spark sql test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
* 3. generate result:
- * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/SubExprEliminationBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/AggregateBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/AggregateBenchmark.scala
index b8d7b774d83..ae4281cd639 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/AggregateBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/AggregateBenchmark.scala
@@ -38,8 +38,8 @@ import org.apache.spark.unsafe.map.BytesToBytesMap
* {{{
* 1. without sbt: bin/spark-submit --class <this class>
* --jars <spark core test jar>,<spark catalyst test jar> <spark sql test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
- * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
+ * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/AggregateBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/AnsiIntervalSortBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/AnsiIntervalSortBenchmark.scala
index 0537527b855..fbab85621c0 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/AnsiIntervalSortBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/AnsiIntervalSortBenchmark.scala
@@ -26,9 +26,9 @@ import org.apache.spark.sql.internal.SQLConf
* {{{
* 1. without sbt:
* bin/spark-submit --class <this class> --jars <spark core test jar> <sql core test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
* 3. generate result:
- * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/IntervalBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/Base64Benchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/Base64Benchmark.scala
index eb0b896574a..3ad6baea84f 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/Base64Benchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/Base64Benchmark.scala
@@ -25,9 +25,9 @@ import org.apache.spark.benchmark.Benchmark
* {{{
* 1. without sbt:
* bin/spark-submit --class <this class> --jars <spark core test jar> <sql core test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
* 3. generate result:
- * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/Base64Benchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/BloomFilterBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/BloomFilterBenchmark.scala
index 5f193109ca2..1cb987458de 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/BloomFilterBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/BloomFilterBenchmark.scala
@@ -30,8 +30,8 @@ import org.apache.spark.benchmark.Benchmark
* {{{
* 1. without sbt: bin/spark-submit --class <this class>
* --jars <spark core test jar>,<spark catalyst test jar> <spark sql test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
- * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
+ * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/BloomFilterBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/BuiltInDataSourceWriteBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/BuiltInDataSourceWriteBenchmark.scala
index 45d50b5e11a..4752787c501 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/BuiltInDataSourceWriteBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/BuiltInDataSourceWriteBenchmark.scala
@@ -28,17 +28,17 @@ import org.apache.spark.sql.internal.SQLConf
* By default it measures 4 data source format: Parquet, ORC, JSON, CSV.
* 1. without sbt: bin/spark-submit --class <this class>
* --jars <spark core test jar>,<spark catalyst test jar> <spark sql test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
- * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
+ * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/BuiltInDataSourceWriteBenchmark-results.txt".
*
* To measure specified formats, run it with arguments.
* 1. without sbt:
* bin/spark-submit --class <this class> --jars <spark core test jar>,
* <spark catalyst test jar> <spark sql test jar> format1 [format2] [...]
- * 2. build/sbt "sql/test:runMain <this class> format1 [format2] [...]"
+ * 2. build/sbt "sql/Test/runMain <this class> format1 [format2] [...]"
* 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt
- * "sql/test:runMain <this class> format1 [format2] [...]"
+ * "sql/Test/runMain <this class> format1 [format2] [...]"
* Results will be written to "benchmarks/BuiltInDataSourceWriteBenchmark-results.txt".
* }}}
*
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/ByteArrayBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/ByteArrayBenchmark.scala
index 99016842d8c..f6bd881a82a 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/ByteArrayBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/ByteArrayBenchmark.scala
@@ -29,8 +29,8 @@ import org.apache.spark.unsafe.types.{ByteArray, UTF8String}
* To run this benchmark:
* 1. without sbt:
* bin/spark-submit --class <this class> --jars <spark core test jar> <sql core test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
- * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
+ * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/<this class>-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/CharVarcharBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/CharVarcharBenchmark.scala
index 112fd98d701..05148f5494e 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/CharVarcharBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/CharVarcharBenchmark.scala
@@ -27,9 +27,9 @@ import org.apache.spark.benchmark.Benchmark
* 1. without sbt:
* bin/spark-submit --class <this class>
* --jars <spark core test jar>,<spark catalyst test jar> <sql core test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
* 3. generate result:
- * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/CharVarcharBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/ConstantColumnVectorBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/ConstantColumnVectorBenchmark.scala
index 8046a4b6cc5..078954f1a60 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/ConstantColumnVectorBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/ConstantColumnVectorBenchmark.scala
@@ -34,8 +34,8 @@ import org.apache.spark.unsafe.UTF8StringBuilder
* {{{
* 1. without sbt: bin/spark-submit --class <this class>
* --jars <spark core test jar>,<spark catalyst test jar> <spark sql test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
- * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
+ * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/ConstantColumnVectorBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/DataSourceReadBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/DataSourceReadBenchmark.scala
index b35aa73e146..d439ac43243 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/DataSourceReadBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/DataSourceReadBenchmark.scala
@@ -40,8 +40,8 @@ import org.apache.spark.sql.vectorized.ColumnVector
* {{{
* 1. without sbt: bin/spark-submit --class <this class>
* --jars <spark core test jar>,<spark catalyst test jar> <spark sql test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
- * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
+ * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/DataSourceReadBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/DateTimeBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/DateTimeBenchmark.scala
index 918f665238f..6359f1b5f4f 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/DateTimeBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/DateTimeBenchmark.scala
@@ -34,9 +34,9 @@ import org.apache.spark.sql.internal.SQLConf
* 1. without sbt:
* bin/spark-submit --class <this class>
* --jars <spark core test jar>,<spark catalyst test jar> <sql core test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
* 3. generate result:
- * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/DateTimeBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/DateTimeRebaseBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/DateTimeRebaseBenchmark.scala
index 090ef34e1cf..e7f1f15ae32 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/DateTimeRebaseBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/DateTimeRebaseBenchmark.scala
@@ -39,9 +39,9 @@ object DateTime extends Enumeration {
* 1. without sbt:
* bin/spark-submit --class <this class>
* --jars <spark core test jar>,<spark catalyst test jar> <sql core test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
* 3. generate result:
- * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/DateTimeRebaseBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/ExtractBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/ExtractBenchmark.scala
index 6af20e8696a..2fffa265cb7 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/ExtractBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/ExtractBenchmark.scala
@@ -29,9 +29,9 @@ import org.apache.spark.sql.internal.SQLConf
* 1. without sbt:
* bin/spark-submit --class <this class>
* --jars <spark core test jar>,<spark catalyst test jar> <sql core test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
* 3. generate result:
- * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/ExtractBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/FilterPushdownBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/FilterPushdownBenchmark.scala
index dd2852eea78..b572444ac4e 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/FilterPushdownBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/FilterPushdownBenchmark.scala
@@ -35,8 +35,8 @@ import org.apache.spark.sql.types.{ByteType, Decimal, DecimalType}
* {{{
* 1. without sbt: bin/spark-submit --class <this class>
* --jars <spark core test jar>,<spark catalyst test jar> <spark sql test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
- * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
+ * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/FilterPushdownBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/GenerateExecBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/GenerateExecBenchmark.scala
index 97bf95090f1..333e0155d03 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/GenerateExecBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/GenerateExecBenchmark.scala
@@ -26,8 +26,8 @@ import org.apache.spark.sql.functions.explode
* 1. without sbt:
* bin/spark-submit --class <this class>
* --jars <spark core test jar>,<spark catalyst test jar> <sql core test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
- * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
+ * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/GenerateExecBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/HashedRelationMetricsBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/HashedRelationMetricsBenchmark.scala
index f03c22aecfc..857a86ab1c6 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/HashedRelationMetricsBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/HashedRelationMetricsBenchmark.scala
@@ -32,8 +32,8 @@ import org.apache.spark.sql.types.LongType
* {{{
* 1. without sbt: bin/spark-submit --class <this class>
* --jars <spark core test jar>,<spark catalyst test jar> <spark sql test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
- * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
+ * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/HashedRelationMetricsBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/InExpressionBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/InExpressionBenchmark.scala
index f5a3483bdf3..faaea51c002 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/InExpressionBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/InExpressionBenchmark.scala
@@ -34,8 +34,8 @@ import org.apache.spark.sql.types._
* 1. without sbt:
* bin/spark-submit --class <this class>
* --jars <spark core test jar>,<spark catalyst test jar> <spark sql test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
- * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
+ * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/InExpressionBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/InsertTableWithDynamicPartitionsBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/InsertTableWithDynamicPartitionsBenchmark.scala
index 32bc2001dcd..399ac2a9f05 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/InsertTableWithDynamicPartitionsBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/InsertTableWithDynamicPartitionsBenchmark.scala
@@ -26,8 +26,8 @@ import org.apache.spark.benchmark.Benchmark
* 1. without sbt:
* bin/spark-submit --class <this class>
* --jars <spark core test jar>,<spark catalyst test jar> < spark sql test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
- * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
+ * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to
* "benchmarks/InsertTableWithDynamicPartitionsBenchmark-results.txt".
* }}}
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/IntervalBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/IntervalBenchmark.scala
index a9696e6718d..84bffef64db 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/IntervalBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/IntervalBenchmark.scala
@@ -30,9 +30,9 @@ import org.apache.spark.sql.internal.SQLConf
* {{{
* 1. without sbt:
* bin/spark-submit --class <this class> --jars <spark core test jar> <sql core test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
* 3. generate result:
- * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/IntervalBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/JoinBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/JoinBenchmark.scala
index 787fdc7b59d..9bca5970143 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/JoinBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/JoinBenchmark.scala
@@ -29,9 +29,9 @@ import org.apache.spark.sql.types.IntegerType
* 1. without sbt:
* bin/spark-submit --class <this class>
* --jars <spark core test jar>,<spark catalyst test jar> <spark sql test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
* 3. generate result:
- * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/JoinBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/MakeDateTimeBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/MakeDateTimeBenchmark.scala
index 6ab4fcc9c00..429e8f1741f 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/MakeDateTimeBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/MakeDateTimeBenchmark.scala
@@ -27,9 +27,9 @@ import org.apache.spark.sql.internal.SQLConf
* 1. without sbt:
* bin/spark-submit --class <this class>
* --jars <spark core test jar>,<spark catalyst test jar> <sql core test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
* 3. generate result:
- * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/MakeDateTimeBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/MiscBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/MiscBenchmark.scala
index 18605468adb..ad7850ef092 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/MiscBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/MiscBenchmark.scala
@@ -27,8 +27,8 @@ import org.apache.spark.sql.internal.SQLConf
* 1. without sbt:
* bin/spark-submit --class <this class>
* --jars <spark core test jar>,<spark catalyst test jar> <sql core test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
- * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
+ * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/MiscBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/OrcNestedSchemaPruningBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/OrcNestedSchemaPruningBenchmark.scala
index a59da45f1fe..0d541f62641 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/OrcNestedSchemaPruningBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/OrcNestedSchemaPruningBenchmark.scala
@@ -26,9 +26,9 @@ import org.apache.spark.sql.internal.SQLConf
* 1. without sbt:
* bin/spark-submit --class <this class>
* --jars <spark core test jar>,<spark catalyst test jar> <spark sql test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
* 3. generate result:
- * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/OrcNestedSchemaPruningBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/OrcV2NestedSchemaPruningBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/OrcV2NestedSchemaPruningBenchmark.scala
index d0289df380d..3716f982684 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/OrcV2NestedSchemaPruningBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/OrcV2NestedSchemaPruningBenchmark.scala
@@ -24,9 +24,9 @@ package org.apache.spark.sql.execution.benchmark
* 1. without sbt:
* bin/spark-submit --class <this class>
* --jars <spark core test jar>,<spark catalyst test jar> <sql core test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
* 3. generate result:
- * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/OrcV2NestedSchemaPruningBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/ParquetNestedPredicatePushDownBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/ParquetNestedPredicatePushDownBenchmark.scala
index 5604f6856b0..218677e8e40 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/ParquetNestedPredicatePushDownBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/ParquetNestedPredicatePushDownBenchmark.scala
@@ -28,9 +28,9 @@ import org.apache.spark.sql.internal.SQLConf
* 1. without sbt:
* bin/spark-submit --class <this class>
* --jars <spark core test jar>,<spark catalyst test jar> <sql core test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
* 3. generate result:
- * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/ParquetNestedPredicatePushDownBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/ParquetNestedSchemaPruningBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/ParquetNestedSchemaPruningBenchmark.scala
index f8f0ab10b24..2c956d84868 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/ParquetNestedSchemaPruningBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/ParquetNestedSchemaPruningBenchmark.scala
@@ -24,9 +24,9 @@ package org.apache.spark.sql.execution.benchmark
* 1. without sbt:
* bin/spark-submit --class <this class>
* --jars <spark core test jar>,<spark catalyst test jar> <sql core test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
* 3. generate result:
- * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/ParquetNestedSchemaPruningBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/PrimitiveArrayBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/PrimitiveArrayBenchmark.scala
index b09549be167..c9676159a9c 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/PrimitiveArrayBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/PrimitiveArrayBenchmark.scala
@@ -26,8 +26,8 @@ import org.apache.spark.sql.SparkSession
* 1. without sbt:
* bin/spark-submit --class <this class>
* --jars <spark core test jar>,<spark catalyst test jar> < spark sql test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
- * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
+ * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/PrimitiveArrayBenchmark-results.txt".
*/
object PrimitiveArrayBenchmark extends SqlBasedBenchmark {
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/RangeBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/RangeBenchmark.scala
index e7c1790b2e3..3bfe5e9833e 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/RangeBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/RangeBenchmark.scala
@@ -26,8 +26,8 @@ import org.apache.spark.benchmark.Benchmark
* 1. without sbt:
* bin/spark-submit --class <this class>
* --jars <spark core test jar>,<spark catalyst test jar> <spark sql test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
- * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
+ * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/RangeBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/SortBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/SortBenchmark.scala
index 8b8710d96b5..bc4dc9a88d8 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/SortBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/SortBenchmark.scala
@@ -32,8 +32,8 @@ import org.apache.spark.util.random.XORShiftRandom
* To run this benchmark:
* 1. without sbt:
* bin/spark-submit --class <this class> --jars <spark core test jar> <sql core test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
- * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
+ * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/<this class>-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/StateStoreBasicOperationsBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/StateStoreBasicOperationsBenchmark.scala
index a98c8d8a23a..41715563e63 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/StateStoreBasicOperationsBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/StateStoreBasicOperationsBenchmark.scala
@@ -35,9 +35,9 @@ import org.apache.spark.util.Utils
* 1. without sbt:
* bin/spark-submit --class <this class>
* --jars <spark core test jar>,<spark catalyst test jar> <sql core test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
* 3. generate result:
- * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/StateStoreBasicOperationsBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/TPCDSQueryBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/TPCDSQueryBenchmark.scala
index b4f1802cedb..be0b89bc6bc 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/TPCDSQueryBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/TPCDSQueryBenchmark.scala
@@ -39,9 +39,9 @@ import org.apache.spark.sql.types.StructType
* 1. without sbt:
* bin/spark-submit --jars <spark core test jar>,<spark catalyst test jar>
* --class <this class> <spark sql test jar> --data-location <location>
- * 2. build/sbt "sql/test:runMain <this class> --data-location <TPCDS data location>"
+ * 2. build/sbt "sql/Test/runMain <this class> --data-location <TPCDS data location>"
* 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt
- * "sql/test:runMain <this class> --data-location <location>"
+ * "sql/Test/runMain <this class> --data-location <location>"
* Results will be written to "benchmarks/TPCDSQueryBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/UDFBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/UDFBenchmark.scala
index 34e51efc1f7..6ea65d863a9 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/UDFBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/UDFBenchmark.scala
@@ -30,9 +30,9 @@ import org.apache.spark.sql.types.{IntegerType, StringType}
* 1. without sbt:
* bin/spark-submit --class <this class>
* --jars <spark core test jar>,<spark catalyst test jar> <sql core test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
* 3. generate result:
- * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/UDFBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/UnsafeArrayDataBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/UnsafeArrayDataBenchmark.scala
index 67eb20c42a9..9b095647c7f 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/UnsafeArrayDataBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/UnsafeArrayDataBenchmark.scala
@@ -29,8 +29,8 @@ import org.apache.spark.sql.catalyst.expressions.UnsafeArrayData
* {{{
* 1. without sbt:
* bin/spark-submit --class <this class> --jars <spark core test jar> <spark sql test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
- * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
+ * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/UnsafeArrayDataBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/WideSchemaBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/WideSchemaBenchmark.scala
index c209a6da3a3..58e92297d3b 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/WideSchemaBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/WideSchemaBenchmark.scala
@@ -30,8 +30,8 @@ import org.apache.spark.util.Utils
* 1. without sbt:
* bin/spark-submit --class <this class>
* --jars <spark core test jar>,<spark catalyst test jar> <spark sql test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
- * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
+ * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/WideSchemaBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/WideTableBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/WideTableBenchmark.scala
index ba79c12c461..f4f1380b947 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/WideTableBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/benchmark/WideTableBenchmark.scala
@@ -26,8 +26,8 @@ import org.apache.spark.sql.internal.SQLConf
* To run this benchmark:
* 1. without sbt: bin/spark-submit --class <this class>
* --jars <spark core test jar>,<spark catalyst test jar> <spark sql test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
- * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
+ * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/WideTableBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/columnar/InMemoryColumnarBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/columnar/InMemoryColumnarBenchmark.scala
index b975451e135..d8f154bfb1e 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/columnar/InMemoryColumnarBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/columnar/InMemoryColumnarBenchmark.scala
@@ -27,8 +27,8 @@ import org.apache.spark.sql.execution.benchmark.SqlBasedBenchmark
* 1. without sbt:
* bin/spark-submit --class <this class>
* --jars <spark core test jar> <spark sql test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
- * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
+ * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/InMemoryColumnarBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/columnar/compression/CompressionSchemeBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/columnar/compression/CompressionSchemeBenchmark.scala
index cad7f7a1121..ddf5ba2f20a 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/columnar/compression/CompressionSchemeBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/columnar/compression/CompressionSchemeBenchmark.scala
@@ -36,8 +36,8 @@ import org.apache.spark.util.Utils._
* 1. without sbt:
* bin/spark-submit --class <this class>
* --jars <spark core test jar>,<spark catalyst test jar> <spark sql test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
- * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
+ * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/CompressionSchemeBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/csv/CSVBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/csv/CSVBenchmark.scala
index 53d287b32f8..7e2a2688e65 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/csv/CSVBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/csv/CSVBenchmark.scala
@@ -33,9 +33,9 @@ import org.apache.spark.sql.types._
* 1. without sbt:
* bin/spark-submit --class <this class> --jars <spark core test jar>,
* <spark catalyst test jar> <spark sql test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
* 3. generate result:
- * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/CSVBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/json/JsonBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/json/JsonBenchmark.scala
index 19317035691..a7794848434 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/json/JsonBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/json/JsonBenchmark.scala
@@ -33,9 +33,9 @@ import org.apache.spark.sql.types._
* 1. without sbt:
* bin/spark-submit --class <this class> --jars <spark core test jar>,
* <spark catalyst test jar> <spark sql test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
* 3. generate result:
- * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/JSONBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/ui/MetricsAggregationBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/ui/MetricsAggregationBenchmark.scala
index 5d9bb8a0a4c..d994126fe63 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/ui/MetricsAggregationBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/ui/MetricsAggregationBenchmark.scala
@@ -41,8 +41,8 @@ import org.apache.spark.util.kvstore.InMemoryStore
* To run this benchmark:
* 1. without sbt:
* bin/spark-submit --class <this class> --jars <core test jar> <spark sql test jar>
- * 2. build/sbt "core/test:runMain <this class>"
- * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "core/test:runMain <this class>"
+ * 2. build/sbt "core/Test/runMain <this class>"
+ * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "core/Test/runMain <this class>"
* Results will be written to "benchmarks/MetricsAggregationBenchmark-results.txt".
* }}}
*/
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/vectorized/ColumnarBatchBenchmark.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/vectorized/ColumnarBatchBenchmark.scala
index f9ae611691a..bffd006cbbf 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/vectorized/ColumnarBatchBenchmark.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/vectorized/ColumnarBatchBenchmark.scala
@@ -35,8 +35,8 @@ import org.apache.spark.util.collection.BitSet
* 1. without sbt:
* bin/spark-submit --class <this class>
* --jars <spark core test jar> <spark sql test jar>
- * 2. build/sbt "sql/test:runMain <this class>"
- * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/test:runMain <this class>"
+ * 2. build/sbt "sql/Test/runMain <this class>"
+ * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "sql/Test/runMain <this class>"
* Results will be written to "benchmarks/ColumnarBatchBenchmark-results.txt".
* }}}
*/
diff --git a/sql/hive/src/test/scala/org/apache/spark/sql/execution/benchmark/InsertIntoHiveTableBenchmark.scala b/sql/hive/src/test/scala/org/apache/spark/sql/execution/benchmark/InsertIntoHiveTableBenchmark.scala
index 1df91a1174b..76345985698 100644
--- a/sql/hive/src/test/scala/org/apache/spark/sql/execution/benchmark/InsertIntoHiveTableBenchmark.scala
+++ b/sql/hive/src/test/scala/org/apache/spark/sql/execution/benchmark/InsertIntoHiveTableBenchmark.scala
@@ -28,9 +28,9 @@ import org.apache.spark.sql.hive.test.TestHive
* 1. without sbt: bin/spark-submit --class <this class>
* --jars <spark catalyst test jar>,<spark core test jar>,<spark sql test jar>
* <spark hive test jar>
- * 2. build/sbt "hive/test:runMain <this class>"
+ * 2. build/sbt "hive/Test/runMain <this class>"
* 3. generate result:
- * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "hive/test:runMain <this class>"
+ * SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "hive/Test/runMain <this class>"
* Results will be written to "benchmarks/InsertIntoHiveTableBenchmark-hive2.3-results.txt".
* }}}
*/
diff --git a/sql/hive/src/test/scala/org/apache/spark/sql/execution/benchmark/ObjectHashAggregateExecBenchmark.scala b/sql/hive/src/test/scala/org/apache/spark/sql/execution/benchmark/ObjectHashAggregateExecBenchmark.scala
index 46a60efc5b5..5d0a5ce0957 100644
--- a/sql/hive/src/test/scala/org/apache/spark/sql/execution/benchmark/ObjectHashAggregateExecBenchmark.scala
+++ b/sql/hive/src/test/scala/org/apache/spark/sql/execution/benchmark/ObjectHashAggregateExecBenchmark.scala
@@ -37,8 +37,8 @@ import org.apache.spark.sql.types.LongType
* 1. without sbt: bin/spark-submit --class <this class>
* --jars <spark catalyst test jar>,<spark core test jar>,<spark sql test jar>
* <spark hive test jar>
- * 2. build/sbt "hive/test:runMain <this class>"
- * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "hive/test:runMain <this class>"
+ * 2. build/sbt "hive/Test/runMain <this class>"
+ * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "hive/Test/runMain <this class>"
* Results will be written to "benchmarks/ObjectHashAggregateExecBenchmark-results.txt".
* }}}
*/
diff --git a/sql/hive/src/test/scala/org/apache/spark/sql/hive/orc/OrcReadBenchmark.scala b/sql/hive/src/test/scala/org/apache/spark/sql/hive/orc/OrcReadBenchmark.scala
index 61a93606841..9ee9ebc2282 100644
--- a/sql/hive/src/test/scala/org/apache/spark/sql/hive/orc/OrcReadBenchmark.scala
+++ b/sql/hive/src/test/scala/org/apache/spark/sql/hive/orc/OrcReadBenchmark.scala
@@ -34,8 +34,8 @@ import org.apache.spark.sql.types._
* To run this benchmark:
* 1. without sbt: bin/spark-submit --class <this class>
* --jars <catalyst test jar>,<core test jar>,<spark sql test jar> <spark-hive test jar>
- * 2. build/sbt "hive/test:runMain <this class>"
- * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "hive/test:runMain <this class>"
+ * 2. build/sbt "hive/Test/runMain <this class>"
+ * 3. generate result: SPARK_GENERATE_BENCHMARK_FILES=1 build/sbt "hive/Test/runMain <this class>"
* Results will be written to "benchmarks/OrcReadBenchmark-results.txt".
* }}}
*
---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@spark.apache.org
For additional commands, e-mail: commits-help@spark.apache.org