You are viewing a plain text version of this content. The canonical link for it is here.
Posted to reviews@spark.apache.org by GitBox <gi...@apache.org> on 2022/03/21 18:28:44 UTC

[GitHub] [spark] revans2 commented on pull request #35925: [SPARK-38604][SQL] Keep ceil and floor with only a single argument the same as before

revans2 commented on pull request #35925:
URL: https://github.com/apache/spark/pull/35925#issuecomment-1074266533


   The test failure appears to be unrelated 
   
   ```
   2022-03-21T15:44:04.1286684Z [info] - SPARK-37753: Inhibit broadcast in left outer join when there are many empty partitions on outer/left side *** FAILED *** (297 milliseconds)
   2022-03-21T15:44:04.1292629Z [info]   ArrayBuffer(BroadcastHashJoin [key#113281], [a#113291], LeftOuter, BuildRight, false
   2022-03-21T15:44:04.1319286Z [info]   :- AQEShuffleRead local
   2022-03-21T15:44:04.1322776Z [info]   :  +- ShuffleQueryStage 0
   2022-03-21T15:44:04.1326338Z [info]   :     +- Exchange hashpartitioning(key#113281, 5), ENSURE_REQUIREMENTS, [id=#257658]
   2022-03-21T15:44:04.1329839Z [info]   :        +- *(1) Filter (isnotnull(value#113282) AND (value#113282 = 1))
   2022-03-21T15:44:04.1335598Z [info]   :           +- *(1) SerializeFromObject [knownnotnull(assertnotnull(input[0, org.apache.spark.sql.test.SQLTestData$TestData, true])).key AS key#113281, staticinvoke(class org.apache.spark.unsafe.types
   .UTF8String, StringType, fromString, knownnotnull(assertnotnull(input[0, org.apache.spark.sql.test.SQLTestData$TestData, true])).value, true, false, true) AS value#113282]
   2022-03-21T15:44:04.1340122Z [info]   :              +- Scan[obj#113280]
   2022-03-21T15:44:04.1344085Z [info]   +- BroadcastQueryStage 2
   2022-03-21T15:44:04.1349457Z [info]      +- BroadcastExchange HashedRelationBroadcastMode(List(cast(input[0, int, false] as bigint)),false), [id=#257698]
   2022-03-21T15:44:04.1353255Z [info]         +- AQEShuffleRead local
   2022-03-21T15:44:04.1356728Z [info]            +- ShuffleQueryStage 1
   2022-03-21T15:44:04.1360576Z [info]               +- Exchange hashpartitioning(a#113291, 5), ENSURE_REQUIREMENTS, [id=#257677]
   2022-03-21T15:44:04.1365128Z [info]                  +- *(2) SerializeFromObject [knownnotnull(assertnotnull(input[0, org.apache.spark.sql.test.SQLTestData$TestData2, true])).a AS a#113291, knownnotnull(assertnotnull(input[0, org.apache.s
   park.sql.test.SQLTestData$TestData2, true])).b AS b#113292]
   2022-03-21T15:44:04.1369134Z [info]                     +- Scan[obj#113290]
   2022-03-21T15:44:04.1372542Z [info]   ) was not empty (AdaptiveQueryExecSuite.scala:717)
   2022-03-21T15:44:04.1376177Z [info]   org.scalatest.exceptions.TestFailedException:
   2022-03-21T15:44:04.1382168Z [info]   at org.scalatest.Assertions.newAssertionFailedException(Assertions.scala:472)
   2022-03-21T15:44:04.1386962Z [info]   at org.scalatest.Assertions.newAssertionFailedException$(Assertions.scala:471)
   2022-03-21T15:44:04.1391699Z [info]   at org.scalatest.Assertions$.newAssertionFailedException(Assertions.scala:1231)
   2022-03-21T15:44:04.1397524Z [info]   at org.scalatest.Assertions$AssertionsHelper.macroAssert(Assertions.scala:1295)
   2022-03-21T15:44:04.1402345Z [info]   at org.apache.spark.sql.execution.adaptive.AdaptiveQueryExecSuite.$anonfun$new$59(AdaptiveQueryExecSuite.scala:717)
   2022-03-21T15:44:04.1407153Z [info]   at org.apache.spark.sql.catalyst.plans.SQLHelper.withSQLConf(SQLHelper.scala:54)
   2022-03-21T15:44:04.1411886Z [info]   at org.apache.spark.sql.catalyst.plans.SQLHelper.withSQLConf$(SQLHelper.scala:38)
   ```
   
   The query has no ceil or floor in it.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org

For queries about this service, please contact Infrastructure at:
users@infra.apache.org



---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org