You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@spark.apache.org by Maxim Gekk <ma...@databricks.com.INVALID> on 2022/05/16 12:43:37 UTC

[VOTE] Release Spark 3.3.0 (RC2)

Please vote on releasing the following candidate as
Apache Spark version 3.3.0.

The vote is open until 11:59pm Pacific time May 19th and passes if a
majority +1 PMC votes are cast, with a minimum of 3 +1 votes.

[ ] +1 Release this package as Apache Spark 3.3.0
[ ] -1 Do not release this package because ...

To learn more about Apache Spark, please see http://spark.apache.org/

The tag to be voted on is v3.3.0-rc2 (commit
c8c657b922ac8fd8dcf9553113e11a80079db059):
https://github.com/apache/spark/tree/v3.3.0-rc2

The release files, including signatures, digests, etc. can be found at:
https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-bin/

Signatures used for Spark RCs can be found in this file:
https://dist.apache.org/repos/dist/dev/spark/KEYS

The staging repository for this release can be found at:
https://repository.apache.org/content/repositories/orgapachespark-1403

The documentation corresponding to this release can be found at:
https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-docs/

The list of bug fixes going into 3.3.0 can be found at the following URL:
https://issues.apache.org/jira/projects/SPARK/versions/12350369

This release is using the release script of the tag v3.3.0-rc2.


FAQ

=========================
How can I help test this release?
=========================
If you are a Spark user, you can help us test this release by taking
an existing Spark workload and running on this release candidate, then
reporting any regressions.

If you're working in PySpark you can set up a virtual env and install
the current RC and see if anything important breaks, in the Java/Scala
you can add the staging repository to your projects resolvers and test
with the RC (make sure to clean up the artifact cache before/after so
you don't end up building with a out of date RC going forward).

===========================================
What should happen to JIRA tickets still targeting 3.3.0?
===========================================
The current list of open tickets targeted at 3.3.0 can be found at:
https://issues.apache.org/jira/projects/SPARK and search for "Target
Version/s" = 3.3.0

Committers should look at those and triage. Extremely important bug
fixes, documentation, and API tweaks that impact compatibility should
be worked on immediately. Everything else please retarget to an
appropriate release.

==================
But my bug isn't fixed?
==================
In order to make timely releases, we will typically not hold the
release unless the bug in question is a regression from the previous
release. That being said, if there is something which is a regression
that has not been correctly targeted please ping me or a committer to
help target the issue.

Maxim Gekk

Software Engineer

Databricks, Inc.

Re: [VOTE] Release Spark 3.3.0 (RC2)

Posted by Kent Yao <ya...@apache.org>.
Thanks for the quick fix, Gengliang.

BR,
Kent

Gengliang Wang <lt...@gmail.com> 于2022年5月19日周四 18:25写道:
>
> Hi Kent and Wenchen,
>
> Thanks for reporting. I just created https://github.com/apache/spark/pull/36609 to fix the issue.
>
> Gengliang
>
> On Thu, May 19, 2022 at 5:40 PM Wenchen Fan <cl...@gmail.com> wrote:
>>
>> I think it should have been fixed  by https://github.com/apache/spark/commit/0fdb6757946e2a0991256a3b73c0c09d6e764eed . Maybe the fix is not completed...
>>
>> On Thu, May 19, 2022 at 2:16 PM Kent Yao <ya...@apache.org> wrote:
>>>
>>> Thanks, Maxim.
>>>
>>> Leave my -1 for this release candidate.
>>>
>>> Unfortunately, I don't know which PR fixed this.
>>> Does anyone happen to know?
>>>
>>> BR,
>>> Kent Yao
>>>
>>> Maxim Gekk <ma...@databricks.com> 于2022年5月19日周四 13:42写道:
>>> >
>>> > Hi Kent,
>>> >
>>> > > Shall we backport the fix from the master to 3.3 too?
>>> >
>>> > Yes, we shall.
>>> >
>>> > Maxim Gekk
>>> >
>>> > Software Engineer
>>> >
>>> > Databricks, Inc.
>>> >
>>> >
>>> >
>>> > On Thu, May 19, 2022 at 6:44 AM Kent Yao <ya...@apache.org> wrote:
>>> >>
>>> >> Hi,
>>> >>
>>> >> I verified the simple case below with the binary release, and it looks
>>> >> like a bug to me.
>>> >>
>>> >> bin/spark-sql -e "select date '2018-11-17' > 1"
>>> >>
>>> >> Error in query: Invalid call to toAttribute on unresolved object;
>>> >> 'Project [unresolvedalias((2018-11-17 > 1), None)]
>>> >> +- OneRowRelation
>>> >>
>>> >> Both 3.2 releases and the master branch work fine with correct errors
>>> >> -  'due to data type mismatch'.
>>> >>
>>> >> Shall we backport the fix from the master to 3.3 too?
>>> >>
>>> >> Bests
>>> >>
>>> >> Kent Yao
>>> >>
>>> >>
>>> >> Yuming Wang <wg...@gmail.com> 于2022年5月18日周三 19:04写道:
>>> >> >
>>> >> > -1. There is a regression: https://github.com/apache/spark/pull/36595
>>> >> >
>>> >> > On Wed, May 18, 2022 at 4:11 PM Martin Grigorov <mg...@apache.org> wrote:
>>> >> >>
>>> >> >> Hi,
>>> >> >>
>>> >> >> [X] +1 Release this package as Apache Spark 3.3.0
>>> >> >>
>>> >> >> Tested:
>>> >> >> - make local distribution from sources (with ./dev/make-distribution.sh --tgz --name with-volcano -Pkubernetes,volcano,hadoop-3)
>>> >> >> - create a Docker image (with JDK 11)
>>> >> >> - run Pi example on
>>> >> >> -- local
>>> >> >> -- Kubernetes with default scheduler
>>> >> >> -- Kubernetes with Volcano scheduler
>>> >> >>
>>> >> >> On both x86_64 and aarch64 !
>>> >> >>
>>> >> >> Regards,
>>> >> >> Martin
>>> >> >>
>>> >> >>
>>> >> >> On Mon, May 16, 2022 at 3:44 PM Maxim Gekk <ma...@databricks.com.invalid> wrote:
>>> >> >>>
>>> >> >>> Please vote on releasing the following candidate as Apache Spark version 3.3.0.
>>> >> >>>
>>> >> >>> The vote is open until 11:59pm Pacific time May 19th and passes if a majority +1 PMC votes are cast, with a minimum of 3 +1 votes.
>>> >> >>>
>>> >> >>> [ ] +1 Release this package as Apache Spark 3.3.0
>>> >> >>> [ ] -1 Do not release this package because ...
>>> >> >>>
>>> >> >>> To learn more about Apache Spark, please see http://spark.apache.org/
>>> >> >>>
>>> >> >>> The tag to be voted on is v3.3.0-rc2 (commit c8c657b922ac8fd8dcf9553113e11a80079db059):
>>> >> >>> https://github.com/apache/spark/tree/v3.3.0-rc2
>>> >> >>>
>>> >> >>> The release files, including signatures, digests, etc. can be found at:
>>> >> >>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-bin/
>>> >> >>>
>>> >> >>> Signatures used for Spark RCs can be found in this file:
>>> >> >>> https://dist.apache.org/repos/dist/dev/spark/KEYS
>>> >> >>>
>>> >> >>> The staging repository for this release can be found at:
>>> >> >>> https://repository.apache.org/content/repositories/orgapachespark-1403
>>> >> >>>
>>> >> >>> The documentation corresponding to this release can be found at:
>>> >> >>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-docs/
>>> >> >>>
>>> >> >>> The list of bug fixes going into 3.3.0 can be found at the following URL:
>>> >> >>> https://issues.apache.org/jira/projects/SPARK/versions/12350369
>>> >> >>>
>>> >> >>> This release is using the release script of the tag v3.3.0-rc2.
>>> >> >>>
>>> >> >>>
>>> >> >>> FAQ
>>> >> >>>
>>> >> >>> =========================
>>> >> >>> How can I help test this release?
>>> >> >>> =========================
>>> >> >>> If you are a Spark user, you can help us test this release by taking
>>> >> >>> an existing Spark workload and running on this release candidate, then
>>> >> >>> reporting any regressions.
>>> >> >>>
>>> >> >>> If you're working in PySpark you can set up a virtual env and install
>>> >> >>> the current RC and see if anything important breaks, in the Java/Scala
>>> >> >>> you can add the staging repository to your projects resolvers and test
>>> >> >>> with the RC (make sure to clean up the artifact cache before/after so
>>> >> >>> you don't end up building with a out of date RC going forward).
>>> >> >>>
>>> >> >>> ===========================================
>>> >> >>> What should happen to JIRA tickets still targeting 3.3.0?
>>> >> >>> ===========================================
>>> >> >>> The current list of open tickets targeted at 3.3.0 can be found at:
>>> >> >>> https://issues.apache.org/jira/projects/SPARK and search for "Target Version/s" = 3.3.0
>>> >> >>>
>>> >> >>> Committers should look at those and triage. Extremely important bug
>>> >> >>> fixes, documentation, and API tweaks that impact compatibility should
>>> >> >>> be worked on immediately. Everything else please retarget to an
>>> >> >>> appropriate release.
>>> >> >>>
>>> >> >>> ==================
>>> >> >>> But my bug isn't fixed?
>>> >> >>> ==================
>>> >> >>> In order to make timely releases, we will typically not hold the
>>> >> >>> release unless the bug in question is a regression from the previous
>>> >> >>> release. That being said, if there is something which is a regression
>>> >> >>> that has not been correctly targeted please ping me or a committer to
>>> >> >>> help target the issue.
>>> >> >>>
>>> >> >>> Maxim Gekk
>>> >> >>>
>>> >> >>> Software Engineer
>>> >> >>>
>>> >> >>> Databricks, Inc.
>>> >>
>>> >> ---------------------------------------------------------------------
>>> >> To unsubscribe e-mail: dev-unsubscribe@spark.apache.org
>>> >>
>>>
>>> ---------------------------------------------------------------------
>>> To unsubscribe e-mail: dev-unsubscribe@spark.apache.org
>>>

---------------------------------------------------------------------
To unsubscribe e-mail: dev-unsubscribe@spark.apache.org


Re: [VOTE] Release Spark 3.3.0 (RC2)

Posted by Gengliang Wang <lt...@gmail.com>.
Hi Kent and Wenchen,

Thanks for reporting. I just created
https://github.com/apache/spark/pull/36609 to fix the issue.

Gengliang

On Thu, May 19, 2022 at 5:40 PM Wenchen Fan <cl...@gmail.com> wrote:

> I think it should have been fixed  by
> https://github.com/apache/spark/commit/0fdb6757946e2a0991256a3b73c0c09d6e764eed
> . Maybe the fix is not completed...
>
> On Thu, May 19, 2022 at 2:16 PM Kent Yao <ya...@apache.org> wrote:
>
>> Thanks, Maxim.
>>
>> Leave my -1 for this release candidate.
>>
>> Unfortunately, I don't know which PR fixed this.
>> Does anyone happen to know?
>>
>> BR,
>> Kent Yao
>>
>> Maxim Gekk <ma...@databricks.com> 于2022年5月19日周四 13:42写道:
>> >
>> > Hi Kent,
>> >
>> > > Shall we backport the fix from the master to 3.3 too?
>> >
>> > Yes, we shall.
>> >
>> > Maxim Gekk
>> >
>> > Software Engineer
>> >
>> > Databricks, Inc.
>> >
>> >
>> >
>> > On Thu, May 19, 2022 at 6:44 AM Kent Yao <ya...@apache.org> wrote:
>> >>
>> >> Hi,
>> >>
>> >> I verified the simple case below with the binary release, and it looks
>> >> like a bug to me.
>> >>
>> >> bin/spark-sql -e "select date '2018-11-17' > 1"
>> >>
>> >> Error in query: Invalid call to toAttribute on unresolved object;
>> >> 'Project [unresolvedalias((2018-11-17 > 1), None)]
>> >> +- OneRowRelation
>> >>
>> >> Both 3.2 releases and the master branch work fine with correct errors
>> >> -  'due to data type mismatch'.
>> >>
>> >> Shall we backport the fix from the master to 3.3 too?
>> >>
>> >> Bests
>> >>
>> >> Kent Yao
>> >>
>> >>
>> >> Yuming Wang <wg...@gmail.com> 于2022年5月18日周三 19:04写道:
>> >> >
>> >> > -1. There is a regression:
>> https://github.com/apache/spark/pull/36595
>> >> >
>> >> > On Wed, May 18, 2022 at 4:11 PM Martin Grigorov <
>> mgrigorov@apache.org> wrote:
>> >> >>
>> >> >> Hi,
>> >> >>
>> >> >> [X] +1 Release this package as Apache Spark 3.3.0
>> >> >>
>> >> >> Tested:
>> >> >> - make local distribution from sources (with
>> ./dev/make-distribution.sh --tgz --name with-volcano
>> -Pkubernetes,volcano,hadoop-3)
>> >> >> - create a Docker image (with JDK 11)
>> >> >> - run Pi example on
>> >> >> -- local
>> >> >> -- Kubernetes with default scheduler
>> >> >> -- Kubernetes with Volcano scheduler
>> >> >>
>> >> >> On both x86_64 and aarch64 !
>> >> >>
>> >> >> Regards,
>> >> >> Martin
>> >> >>
>> >> >>
>> >> >> On Mon, May 16, 2022 at 3:44 PM Maxim Gekk <
>> maxim.gekk@databricks.com.invalid> wrote:
>> >> >>>
>> >> >>> Please vote on releasing the following candidate as Apache Spark
>> version 3.3.0.
>> >> >>>
>> >> >>> The vote is open until 11:59pm Pacific time May 19th and passes if
>> a majority +1 PMC votes are cast, with a minimum of 3 +1 votes.
>> >> >>>
>> >> >>> [ ] +1 Release this package as Apache Spark 3.3.0
>> >> >>> [ ] -1 Do not release this package because ...
>> >> >>>
>> >> >>> To learn more about Apache Spark, please see
>> http://spark.apache.org/
>> >> >>>
>> >> >>> The tag to be voted on is v3.3.0-rc2 (commit
>> c8c657b922ac8fd8dcf9553113e11a80079db059):
>> >> >>> https://github.com/apache/spark/tree/v3.3.0-rc2
>> >> >>>
>> >> >>> The release files, including signatures, digests, etc. can be
>> found at:
>> >> >>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-bin/
>> >> >>>
>> >> >>> Signatures used for Spark RCs can be found in this file:
>> >> >>> https://dist.apache.org/repos/dist/dev/spark/KEYS
>> >> >>>
>> >> >>> The staging repository for this release can be found at:
>> >> >>>
>> https://repository.apache.org/content/repositories/orgapachespark-1403
>> >> >>>
>> >> >>> The documentation corresponding to this release can be found at:
>> >> >>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-docs/
>> >> >>>
>> >> >>> The list of bug fixes going into 3.3.0 can be found at the
>> following URL:
>> >> >>> https://issues.apache.org/jira/projects/SPARK/versions/12350369
>> >> >>>
>> >> >>> This release is using the release script of the tag v3.3.0-rc2.
>> >> >>>
>> >> >>>
>> >> >>> FAQ
>> >> >>>
>> >> >>> =========================
>> >> >>> How can I help test this release?
>> >> >>> =========================
>> >> >>> If you are a Spark user, you can help us test this release by
>> taking
>> >> >>> an existing Spark workload and running on this release candidate,
>> then
>> >> >>> reporting any regressions.
>> >> >>>
>> >> >>> If you're working in PySpark you can set up a virtual env and
>> install
>> >> >>> the current RC and see if anything important breaks, in the
>> Java/Scala
>> >> >>> you can add the staging repository to your projects resolvers and
>> test
>> >> >>> with the RC (make sure to clean up the artifact cache before/after
>> so
>> >> >>> you don't end up building with a out of date RC going forward).
>> >> >>>
>> >> >>> ===========================================
>> >> >>> What should happen to JIRA tickets still targeting 3.3.0?
>> >> >>> ===========================================
>> >> >>> The current list of open tickets targeted at 3.3.0 can be found at:
>> >> >>> https://issues.apache.org/jira/projects/SPARK and search for
>> "Target Version/s" = 3.3.0
>> >> >>>
>> >> >>> Committers should look at those and triage. Extremely important bug
>> >> >>> fixes, documentation, and API tweaks that impact compatibility
>> should
>> >> >>> be worked on immediately. Everything else please retarget to an
>> >> >>> appropriate release.
>> >> >>>
>> >> >>> ==================
>> >> >>> But my bug isn't fixed?
>> >> >>> ==================
>> >> >>> In order to make timely releases, we will typically not hold the
>> >> >>> release unless the bug in question is a regression from the
>> previous
>> >> >>> release. That being said, if there is something which is a
>> regression
>> >> >>> that has not been correctly targeted please ping me or a committer
>> to
>> >> >>> help target the issue.
>> >> >>>
>> >> >>> Maxim Gekk
>> >> >>>
>> >> >>> Software Engineer
>> >> >>>
>> >> >>> Databricks, Inc.
>> >>
>> >> ---------------------------------------------------------------------
>> >> To unsubscribe e-mail: dev-unsubscribe@spark.apache.org
>> >>
>>
>> ---------------------------------------------------------------------
>> To unsubscribe e-mail: dev-unsubscribe@spark.apache.org
>>
>>

Re: [VOTE] Release Spark 3.3.0 (RC2)

Posted by Wenchen Fan <cl...@gmail.com>.
I think it should have been fixed  by
https://github.com/apache/spark/commit/0fdb6757946e2a0991256a3b73c0c09d6e764eed
. Maybe the fix is not completed...

On Thu, May 19, 2022 at 2:16 PM Kent Yao <ya...@apache.org> wrote:

> Thanks, Maxim.
>
> Leave my -1 for this release candidate.
>
> Unfortunately, I don't know which PR fixed this.
> Does anyone happen to know?
>
> BR,
> Kent Yao
>
> Maxim Gekk <ma...@databricks.com> 于2022年5月19日周四 13:42写道:
> >
> > Hi Kent,
> >
> > > Shall we backport the fix from the master to 3.3 too?
> >
> > Yes, we shall.
> >
> > Maxim Gekk
> >
> > Software Engineer
> >
> > Databricks, Inc.
> >
> >
> >
> > On Thu, May 19, 2022 at 6:44 AM Kent Yao <ya...@apache.org> wrote:
> >>
> >> Hi,
> >>
> >> I verified the simple case below with the binary release, and it looks
> >> like a bug to me.
> >>
> >> bin/spark-sql -e "select date '2018-11-17' > 1"
> >>
> >> Error in query: Invalid call to toAttribute on unresolved object;
> >> 'Project [unresolvedalias((2018-11-17 > 1), None)]
> >> +- OneRowRelation
> >>
> >> Both 3.2 releases and the master branch work fine with correct errors
> >> -  'due to data type mismatch'.
> >>
> >> Shall we backport the fix from the master to 3.3 too?
> >>
> >> Bests
> >>
> >> Kent Yao
> >>
> >>
> >> Yuming Wang <wg...@gmail.com> 于2022年5月18日周三 19:04写道:
> >> >
> >> > -1. There is a regression: https://github.com/apache/spark/pull/36595
> >> >
> >> > On Wed, May 18, 2022 at 4:11 PM Martin Grigorov <mg...@apache.org>
> wrote:
> >> >>
> >> >> Hi,
> >> >>
> >> >> [X] +1 Release this package as Apache Spark 3.3.0
> >> >>
> >> >> Tested:
> >> >> - make local distribution from sources (with
> ./dev/make-distribution.sh --tgz --name with-volcano
> -Pkubernetes,volcano,hadoop-3)
> >> >> - create a Docker image (with JDK 11)
> >> >> - run Pi example on
> >> >> -- local
> >> >> -- Kubernetes with default scheduler
> >> >> -- Kubernetes with Volcano scheduler
> >> >>
> >> >> On both x86_64 and aarch64 !
> >> >>
> >> >> Regards,
> >> >> Martin
> >> >>
> >> >>
> >> >> On Mon, May 16, 2022 at 3:44 PM Maxim Gekk <
> maxim.gekk@databricks.com.invalid> wrote:
> >> >>>
> >> >>> Please vote on releasing the following candidate as Apache Spark
> version 3.3.0.
> >> >>>
> >> >>> The vote is open until 11:59pm Pacific time May 19th and passes if
> a majority +1 PMC votes are cast, with a minimum of 3 +1 votes.
> >> >>>
> >> >>> [ ] +1 Release this package as Apache Spark 3.3.0
> >> >>> [ ] -1 Do not release this package because ...
> >> >>>
> >> >>> To learn more about Apache Spark, please see
> http://spark.apache.org/
> >> >>>
> >> >>> The tag to be voted on is v3.3.0-rc2 (commit
> c8c657b922ac8fd8dcf9553113e11a80079db059):
> >> >>> https://github.com/apache/spark/tree/v3.3.0-rc2
> >> >>>
> >> >>> The release files, including signatures, digests, etc. can be found
> at:
> >> >>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-bin/
> >> >>>
> >> >>> Signatures used for Spark RCs can be found in this file:
> >> >>> https://dist.apache.org/repos/dist/dev/spark/KEYS
> >> >>>
> >> >>> The staging repository for this release can be found at:
> >> >>>
> https://repository.apache.org/content/repositories/orgapachespark-1403
> >> >>>
> >> >>> The documentation corresponding to this release can be found at:
> >> >>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-docs/
> >> >>>
> >> >>> The list of bug fixes going into 3.3.0 can be found at the
> following URL:
> >> >>> https://issues.apache.org/jira/projects/SPARK/versions/12350369
> >> >>>
> >> >>> This release is using the release script of the tag v3.3.0-rc2.
> >> >>>
> >> >>>
> >> >>> FAQ
> >> >>>
> >> >>> =========================
> >> >>> How can I help test this release?
> >> >>> =========================
> >> >>> If you are a Spark user, you can help us test this release by taking
> >> >>> an existing Spark workload and running on this release candidate,
> then
> >> >>> reporting any regressions.
> >> >>>
> >> >>> If you're working in PySpark you can set up a virtual env and
> install
> >> >>> the current RC and see if anything important breaks, in the
> Java/Scala
> >> >>> you can add the staging repository to your projects resolvers and
> test
> >> >>> with the RC (make sure to clean up the artifact cache before/after
> so
> >> >>> you don't end up building with a out of date RC going forward).
> >> >>>
> >> >>> ===========================================
> >> >>> What should happen to JIRA tickets still targeting 3.3.0?
> >> >>> ===========================================
> >> >>> The current list of open tickets targeted at 3.3.0 can be found at:
> >> >>> https://issues.apache.org/jira/projects/SPARK and search for
> "Target Version/s" = 3.3.0
> >> >>>
> >> >>> Committers should look at those and triage. Extremely important bug
> >> >>> fixes, documentation, and API tweaks that impact compatibility
> should
> >> >>> be worked on immediately. Everything else please retarget to an
> >> >>> appropriate release.
> >> >>>
> >> >>> ==================
> >> >>> But my bug isn't fixed?
> >> >>> ==================
> >> >>> In order to make timely releases, we will typically not hold the
> >> >>> release unless the bug in question is a regression from the previous
> >> >>> release. That being said, if there is something which is a
> regression
> >> >>> that has not been correctly targeted please ping me or a committer
> to
> >> >>> help target the issue.
> >> >>>
> >> >>> Maxim Gekk
> >> >>>
> >> >>> Software Engineer
> >> >>>
> >> >>> Databricks, Inc.
> >>
> >> ---------------------------------------------------------------------
> >> To unsubscribe e-mail: dev-unsubscribe@spark.apache.org
> >>
>
> ---------------------------------------------------------------------
> To unsubscribe e-mail: dev-unsubscribe@spark.apache.org
>
>

Re: [VOTE] Release Spark 3.3.0 (RC2)

Posted by Kent Yao <ya...@apache.org>.
Thanks, Maxim.

Leave my -1 for this release candidate.

Unfortunately, I don't know which PR fixed this.
Does anyone happen to know?

BR,
Kent Yao

Maxim Gekk <ma...@databricks.com> 于2022年5月19日周四 13:42写道:
>
> Hi Kent,
>
> > Shall we backport the fix from the master to 3.3 too?
>
> Yes, we shall.
>
> Maxim Gekk
>
> Software Engineer
>
> Databricks, Inc.
>
>
>
> On Thu, May 19, 2022 at 6:44 AM Kent Yao <ya...@apache.org> wrote:
>>
>> Hi,
>>
>> I verified the simple case below with the binary release, and it looks
>> like a bug to me.
>>
>> bin/spark-sql -e "select date '2018-11-17' > 1"
>>
>> Error in query: Invalid call to toAttribute on unresolved object;
>> 'Project [unresolvedalias((2018-11-17 > 1), None)]
>> +- OneRowRelation
>>
>> Both 3.2 releases and the master branch work fine with correct errors
>> -  'due to data type mismatch'.
>>
>> Shall we backport the fix from the master to 3.3 too?
>>
>> Bests
>>
>> Kent Yao
>>
>>
>> Yuming Wang <wg...@gmail.com> 于2022年5月18日周三 19:04写道:
>> >
>> > -1. There is a regression: https://github.com/apache/spark/pull/36595
>> >
>> > On Wed, May 18, 2022 at 4:11 PM Martin Grigorov <mg...@apache.org> wrote:
>> >>
>> >> Hi,
>> >>
>> >> [X] +1 Release this package as Apache Spark 3.3.0
>> >>
>> >> Tested:
>> >> - make local distribution from sources (with ./dev/make-distribution.sh --tgz --name with-volcano -Pkubernetes,volcano,hadoop-3)
>> >> - create a Docker image (with JDK 11)
>> >> - run Pi example on
>> >> -- local
>> >> -- Kubernetes with default scheduler
>> >> -- Kubernetes with Volcano scheduler
>> >>
>> >> On both x86_64 and aarch64 !
>> >>
>> >> Regards,
>> >> Martin
>> >>
>> >>
>> >> On Mon, May 16, 2022 at 3:44 PM Maxim Gekk <ma...@databricks.com.invalid> wrote:
>> >>>
>> >>> Please vote on releasing the following candidate as Apache Spark version 3.3.0.
>> >>>
>> >>> The vote is open until 11:59pm Pacific time May 19th and passes if a majority +1 PMC votes are cast, with a minimum of 3 +1 votes.
>> >>>
>> >>> [ ] +1 Release this package as Apache Spark 3.3.0
>> >>> [ ] -1 Do not release this package because ...
>> >>>
>> >>> To learn more about Apache Spark, please see http://spark.apache.org/
>> >>>
>> >>> The tag to be voted on is v3.3.0-rc2 (commit c8c657b922ac8fd8dcf9553113e11a80079db059):
>> >>> https://github.com/apache/spark/tree/v3.3.0-rc2
>> >>>
>> >>> The release files, including signatures, digests, etc. can be found at:
>> >>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-bin/
>> >>>
>> >>> Signatures used for Spark RCs can be found in this file:
>> >>> https://dist.apache.org/repos/dist/dev/spark/KEYS
>> >>>
>> >>> The staging repository for this release can be found at:
>> >>> https://repository.apache.org/content/repositories/orgapachespark-1403
>> >>>
>> >>> The documentation corresponding to this release can be found at:
>> >>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-docs/
>> >>>
>> >>> The list of bug fixes going into 3.3.0 can be found at the following URL:
>> >>> https://issues.apache.org/jira/projects/SPARK/versions/12350369
>> >>>
>> >>> This release is using the release script of the tag v3.3.0-rc2.
>> >>>
>> >>>
>> >>> FAQ
>> >>>
>> >>> =========================
>> >>> How can I help test this release?
>> >>> =========================
>> >>> If you are a Spark user, you can help us test this release by taking
>> >>> an existing Spark workload and running on this release candidate, then
>> >>> reporting any regressions.
>> >>>
>> >>> If you're working in PySpark you can set up a virtual env and install
>> >>> the current RC and see if anything important breaks, in the Java/Scala
>> >>> you can add the staging repository to your projects resolvers and test
>> >>> with the RC (make sure to clean up the artifact cache before/after so
>> >>> you don't end up building with a out of date RC going forward).
>> >>>
>> >>> ===========================================
>> >>> What should happen to JIRA tickets still targeting 3.3.0?
>> >>> ===========================================
>> >>> The current list of open tickets targeted at 3.3.0 can be found at:
>> >>> https://issues.apache.org/jira/projects/SPARK and search for "Target Version/s" = 3.3.0
>> >>>
>> >>> Committers should look at those and triage. Extremely important bug
>> >>> fixes, documentation, and API tweaks that impact compatibility should
>> >>> be worked on immediately. Everything else please retarget to an
>> >>> appropriate release.
>> >>>
>> >>> ==================
>> >>> But my bug isn't fixed?
>> >>> ==================
>> >>> In order to make timely releases, we will typically not hold the
>> >>> release unless the bug in question is a regression from the previous
>> >>> release. That being said, if there is something which is a regression
>> >>> that has not been correctly targeted please ping me or a committer to
>> >>> help target the issue.
>> >>>
>> >>> Maxim Gekk
>> >>>
>> >>> Software Engineer
>> >>>
>> >>> Databricks, Inc.
>>
>> ---------------------------------------------------------------------
>> To unsubscribe e-mail: dev-unsubscribe@spark.apache.org
>>

---------------------------------------------------------------------
To unsubscribe e-mail: dev-unsubscribe@spark.apache.org


Re: [VOTE] Release Spark 3.3.0 (RC2)

Posted by Maxim Gekk <ma...@databricks.com.INVALID>.
Hi Kent,

> Shall we backport the fix from the master to 3.3 too?

Yes, we shall.

Maxim Gekk

Software Engineer

Databricks, Inc.


On Thu, May 19, 2022 at 6:44 AM Kent Yao <ya...@apache.org> wrote:

> Hi,
>
> I verified the simple case below with the binary release, and it looks
> like a bug to me.
>
> bin/spark-sql -e "select date '2018-11-17' > 1"
>
> Error in query: Invalid call to toAttribute on unresolved object;
> 'Project [unresolvedalias((2018-11-17 > 1), None)]
> +- OneRowRelation
>
> Both 3.2 releases and the master branch work fine with correct errors
> -  'due to data type mismatch'.
>
> Shall we backport the fix from the master to 3.3 too?
>
> Bests
>
> Kent Yao
>
>
> Yuming Wang <wg...@gmail.com> 于2022年5月18日周三 19:04写道:
> >
> > -1. There is a regression: https://github.com/apache/spark/pull/36595
> >
> > On Wed, May 18, 2022 at 4:11 PM Martin Grigorov <mg...@apache.org>
> wrote:
> >>
> >> Hi,
> >>
> >> [X] +1 Release this package as Apache Spark 3.3.0
> >>
> >> Tested:
> >> - make local distribution from sources (with ./dev/make-distribution.sh
> --tgz --name with-volcano -Pkubernetes,volcano,hadoop-3)
> >> - create a Docker image (with JDK 11)
> >> - run Pi example on
> >> -- local
> >> -- Kubernetes with default scheduler
> >> -- Kubernetes with Volcano scheduler
> >>
> >> On both x86_64 and aarch64 !
> >>
> >> Regards,
> >> Martin
> >>
> >>
> >> On Mon, May 16, 2022 at 3:44 PM Maxim Gekk <ma...@databricks.com.invalid>
> wrote:
> >>>
> >>> Please vote on releasing the following candidate as Apache Spark
> version 3.3.0.
> >>>
> >>> The vote is open until 11:59pm Pacific time May 19th and passes if a
> majority +1 PMC votes are cast, with a minimum of 3 +1 votes.
> >>>
> >>> [ ] +1 Release this package as Apache Spark 3.3.0
> >>> [ ] -1 Do not release this package because ...
> >>>
> >>> To learn more about Apache Spark, please see http://spark.apache.org/
> >>>
> >>> The tag to be voted on is v3.3.0-rc2 (commit
> c8c657b922ac8fd8dcf9553113e11a80079db059):
> >>> https://github.com/apache/spark/tree/v3.3.0-rc2
> >>>
> >>> The release files, including signatures, digests, etc. can be found at:
> >>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-bin/
> >>>
> >>> Signatures used for Spark RCs can be found in this file:
> >>> https://dist.apache.org/repos/dist/dev/spark/KEYS
> >>>
> >>> The staging repository for this release can be found at:
> >>> https://repository.apache.org/content/repositories/orgapachespark-1403
> >>>
> >>> The documentation corresponding to this release can be found at:
> >>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-docs/
> >>>
> >>> The list of bug fixes going into 3.3.0 can be found at the following
> URL:
> >>> https://issues.apache.org/jira/projects/SPARK/versions/12350369
> >>>
> >>> This release is using the release script of the tag v3.3.0-rc2.
> >>>
> >>>
> >>> FAQ
> >>>
> >>> =========================
> >>> How can I help test this release?
> >>> =========================
> >>> If you are a Spark user, you can help us test this release by taking
> >>> an existing Spark workload and running on this release candidate, then
> >>> reporting any regressions.
> >>>
> >>> If you're working in PySpark you can set up a virtual env and install
> >>> the current RC and see if anything important breaks, in the Java/Scala
> >>> you can add the staging repository to your projects resolvers and test
> >>> with the RC (make sure to clean up the artifact cache before/after so
> >>> you don't end up building with a out of date RC going forward).
> >>>
> >>> ===========================================
> >>> What should happen to JIRA tickets still targeting 3.3.0?
> >>> ===========================================
> >>> The current list of open tickets targeted at 3.3.0 can be found at:
> >>> https://issues.apache.org/jira/projects/SPARK and search for "Target
> Version/s" = 3.3.0
> >>>
> >>> Committers should look at those and triage. Extremely important bug
> >>> fixes, documentation, and API tweaks that impact compatibility should
> >>> be worked on immediately. Everything else please retarget to an
> >>> appropriate release.
> >>>
> >>> ==================
> >>> But my bug isn't fixed?
> >>> ==================
> >>> In order to make timely releases, we will typically not hold the
> >>> release unless the bug in question is a regression from the previous
> >>> release. That being said, if there is something which is a regression
> >>> that has not been correctly targeted please ping me or a committer to
> >>> help target the issue.
> >>>
> >>> Maxim Gekk
> >>>
> >>> Software Engineer
> >>>
> >>> Databricks, Inc.
>
> ---------------------------------------------------------------------
> To unsubscribe e-mail: dev-unsubscribe@spark.apache.org
>
>

Re: [VOTE] Release Spark 3.3.0 (RC2)

Posted by Kent Yao <ya...@apache.org>.
Hi,

I verified the simple case below with the binary release, and it looks
like a bug to me.

bin/spark-sql -e "select date '2018-11-17' > 1"

Error in query: Invalid call to toAttribute on unresolved object;
'Project [unresolvedalias((2018-11-17 > 1), None)]
+- OneRowRelation

Both 3.2 releases and the master branch work fine with correct errors
-  'due to data type mismatch'.

Shall we backport the fix from the master to 3.3 too?

Bests

Kent Yao


Yuming Wang <wg...@gmail.com> 于2022年5月18日周三 19:04写道:
>
> -1. There is a regression: https://github.com/apache/spark/pull/36595
>
> On Wed, May 18, 2022 at 4:11 PM Martin Grigorov <mg...@apache.org> wrote:
>>
>> Hi,
>>
>> [X] +1 Release this package as Apache Spark 3.3.0
>>
>> Tested:
>> - make local distribution from sources (with ./dev/make-distribution.sh --tgz --name with-volcano -Pkubernetes,volcano,hadoop-3)
>> - create a Docker image (with JDK 11)
>> - run Pi example on
>> -- local
>> -- Kubernetes with default scheduler
>> -- Kubernetes with Volcano scheduler
>>
>> On both x86_64 and aarch64 !
>>
>> Regards,
>> Martin
>>
>>
>> On Mon, May 16, 2022 at 3:44 PM Maxim Gekk <ma...@databricks.com.invalid> wrote:
>>>
>>> Please vote on releasing the following candidate as Apache Spark version 3.3.0.
>>>
>>> The vote is open until 11:59pm Pacific time May 19th and passes if a majority +1 PMC votes are cast, with a minimum of 3 +1 votes.
>>>
>>> [ ] +1 Release this package as Apache Spark 3.3.0
>>> [ ] -1 Do not release this package because ...
>>>
>>> To learn more about Apache Spark, please see http://spark.apache.org/
>>>
>>> The tag to be voted on is v3.3.0-rc2 (commit c8c657b922ac8fd8dcf9553113e11a80079db059):
>>> https://github.com/apache/spark/tree/v3.3.0-rc2
>>>
>>> The release files, including signatures, digests, etc. can be found at:
>>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-bin/
>>>
>>> Signatures used for Spark RCs can be found in this file:
>>> https://dist.apache.org/repos/dist/dev/spark/KEYS
>>>
>>> The staging repository for this release can be found at:
>>> https://repository.apache.org/content/repositories/orgapachespark-1403
>>>
>>> The documentation corresponding to this release can be found at:
>>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-docs/
>>>
>>> The list of bug fixes going into 3.3.0 can be found at the following URL:
>>> https://issues.apache.org/jira/projects/SPARK/versions/12350369
>>>
>>> This release is using the release script of the tag v3.3.0-rc2.
>>>
>>>
>>> FAQ
>>>
>>> =========================
>>> How can I help test this release?
>>> =========================
>>> If you are a Spark user, you can help us test this release by taking
>>> an existing Spark workload and running on this release candidate, then
>>> reporting any regressions.
>>>
>>> If you're working in PySpark you can set up a virtual env and install
>>> the current RC and see if anything important breaks, in the Java/Scala
>>> you can add the staging repository to your projects resolvers and test
>>> with the RC (make sure to clean up the artifact cache before/after so
>>> you don't end up building with a out of date RC going forward).
>>>
>>> ===========================================
>>> What should happen to JIRA tickets still targeting 3.3.0?
>>> ===========================================
>>> The current list of open tickets targeted at 3.3.0 can be found at:
>>> https://issues.apache.org/jira/projects/SPARK and search for "Target Version/s" = 3.3.0
>>>
>>> Committers should look at those and triage. Extremely important bug
>>> fixes, documentation, and API tweaks that impact compatibility should
>>> be worked on immediately. Everything else please retarget to an
>>> appropriate release.
>>>
>>> ==================
>>> But my bug isn't fixed?
>>> ==================
>>> In order to make timely releases, we will typically not hold the
>>> release unless the bug in question is a regression from the previous
>>> release. That being said, if there is something which is a regression
>>> that has not been correctly targeted please ping me or a committer to
>>> help target the issue.
>>>
>>> Maxim Gekk
>>>
>>> Software Engineer
>>>
>>> Databricks, Inc.

---------------------------------------------------------------------
To unsubscribe e-mail: dev-unsubscribe@spark.apache.org


Re: [VOTE] Release Spark 3.3.0 (RC2)

Posted by Yuming Wang <wg...@gmail.com>.
-1. There is a regression: https://github.com/apache/spark/pull/36595

On Wed, May 18, 2022 at 4:11 PM Martin Grigorov <mg...@apache.org>
wrote:

> Hi,
>
> [X] +1 Release this package as Apache Spark 3.3.0
>
> Tested:
> - make local distribution from sources (with ./dev/make-distribution.sh
> --tgz --name with-volcano -Pkubernetes,volcano,hadoop-3)
> - create a Docker image (with JDK 11)
> - run Pi example on
> -- local
> -- Kubernetes with default scheduler
> -- Kubernetes with Volcano scheduler
>
> On both x86_64 and aarch64 !
>
> Regards,
> Martin
>
>
> On Mon, May 16, 2022 at 3:44 PM Maxim Gekk
> <ma...@databricks.com.invalid> wrote:
>
>> Please vote on releasing the following candidate as
>> Apache Spark version 3.3.0.
>>
>> The vote is open until 11:59pm Pacific time May 19th and passes if a
>> majority +1 PMC votes are cast, with a minimum of 3 +1 votes.
>>
>> [ ] +1 Release this package as Apache Spark 3.3.0
>> [ ] -1 Do not release this package because ...
>>
>> To learn more about Apache Spark, please see http://spark.apache.org/
>>
>> The tag to be voted on is v3.3.0-rc2 (commit
>> c8c657b922ac8fd8dcf9553113e11a80079db059):
>> https://github.com/apache/spark/tree/v3.3.0-rc2
>>
>> The release files, including signatures, digests, etc. can be found at:
>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-bin/
>>
>> Signatures used for Spark RCs can be found in this file:
>> https://dist.apache.org/repos/dist/dev/spark/KEYS
>>
>> The staging repository for this release can be found at:
>> https://repository.apache.org/content/repositories/orgapachespark-1403
>>
>> The documentation corresponding to this release can be found at:
>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-docs/
>>
>> The list of bug fixes going into 3.3.0 can be found at the following URL:
>> https://issues.apache.org/jira/projects/SPARK/versions/12350369
>>
>> This release is using the release script of the tag v3.3.0-rc2.
>>
>>
>> FAQ
>>
>> =========================
>> How can I help test this release?
>> =========================
>> If you are a Spark user, you can help us test this release by taking
>> an existing Spark workload and running on this release candidate, then
>> reporting any regressions.
>>
>> If you're working in PySpark you can set up a virtual env and install
>> the current RC and see if anything important breaks, in the Java/Scala
>> you can add the staging repository to your projects resolvers and test
>> with the RC (make sure to clean up the artifact cache before/after so
>> you don't end up building with a out of date RC going forward).
>>
>> ===========================================
>> What should happen to JIRA tickets still targeting 3.3.0?
>> ===========================================
>> The current list of open tickets targeted at 3.3.0 can be found at:
>> https://issues.apache.org/jira/projects/SPARK and search for "Target
>> Version/s" = 3.3.0
>>
>> Committers should look at those and triage. Extremely important bug
>> fixes, documentation, and API tweaks that impact compatibility should
>> be worked on immediately. Everything else please retarget to an
>> appropriate release.
>>
>> ==================
>> But my bug isn't fixed?
>> ==================
>> In order to make timely releases, we will typically not hold the
>> release unless the bug in question is a regression from the previous
>> release. That being said, if there is something which is a regression
>> that has not been correctly targeted please ping me or a committer to
>> help target the issue.
>>
>> Maxim Gekk
>>
>> Software Engineer
>>
>> Databricks, Inc.
>>
>

Re: [VOTE] Release Spark 3.3.0 (RC2)

Posted by Martin Grigorov <mg...@apache.org>.
Hi,

[X] +1 Release this package as Apache Spark 3.3.0

Tested:
- make local distribution from sources (with ./dev/make-distribution.sh
--tgz --name with-volcano -Pkubernetes,volcano,hadoop-3)
- create a Docker image (with JDK 11)
- run Pi example on
-- local
-- Kubernetes with default scheduler
-- Kubernetes with Volcano scheduler

On both x86_64 and aarch64 !

Regards,
Martin


On Mon, May 16, 2022 at 3:44 PM Maxim Gekk
<ma...@databricks.com.invalid> wrote:

> Please vote on releasing the following candidate as
> Apache Spark version 3.3.0.
>
> The vote is open until 11:59pm Pacific time May 19th and passes if a
> majority +1 PMC votes are cast, with a minimum of 3 +1 votes.
>
> [ ] +1 Release this package as Apache Spark 3.3.0
> [ ] -1 Do not release this package because ...
>
> To learn more about Apache Spark, please see http://spark.apache.org/
>
> The tag to be voted on is v3.3.0-rc2 (commit
> c8c657b922ac8fd8dcf9553113e11a80079db059):
> https://github.com/apache/spark/tree/v3.3.0-rc2
>
> The release files, including signatures, digests, etc. can be found at:
> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-bin/
>
> Signatures used for Spark RCs can be found in this file:
> https://dist.apache.org/repos/dist/dev/spark/KEYS
>
> The staging repository for this release can be found at:
> https://repository.apache.org/content/repositories/orgapachespark-1403
>
> The documentation corresponding to this release can be found at:
> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-docs/
>
> The list of bug fixes going into 3.3.0 can be found at the following URL:
> https://issues.apache.org/jira/projects/SPARK/versions/12350369
>
> This release is using the release script of the tag v3.3.0-rc2.
>
>
> FAQ
>
> =========================
> How can I help test this release?
> =========================
> If you are a Spark user, you can help us test this release by taking
> an existing Spark workload and running on this release candidate, then
> reporting any regressions.
>
> If you're working in PySpark you can set up a virtual env and install
> the current RC and see if anything important breaks, in the Java/Scala
> you can add the staging repository to your projects resolvers and test
> with the RC (make sure to clean up the artifact cache before/after so
> you don't end up building with a out of date RC going forward).
>
> ===========================================
> What should happen to JIRA tickets still targeting 3.3.0?
> ===========================================
> The current list of open tickets targeted at 3.3.0 can be found at:
> https://issues.apache.org/jira/projects/SPARK and search for "Target
> Version/s" = 3.3.0
>
> Committers should look at those and triage. Extremely important bug
> fixes, documentation, and API tweaks that impact compatibility should
> be worked on immediately. Everything else please retarget to an
> appropriate release.
>
> ==================
> But my bug isn't fixed?
> ==================
> In order to make timely releases, we will typically not hold the
> release unless the bug in question is a regression from the previous
> release. That being said, if there is something which is a regression
> that has not been correctly targeted please ping me or a committer to
> help target the issue.
>
> Maxim Gekk
>
> Software Engineer
>
> Databricks, Inc.
>

Re: Re: Re: [VOTE] Release Spark 3.3.0 (RC2)

Posted by Hyukjin Kwon <gu...@gmail.com>.
And seems like it won't break it because adding a new method won't break
binary compatibility.

On Tue, 17 May 2022 at 19:59, Hyukjin Kwon <gu...@gmail.com> wrote:

> I think most users won't be affected since aggregate pushdown is disabled
> by default.
>
> On Tue, 17 May 2022 at 19:53, beliefer <be...@163.com> wrote:
>
>> If we not contains https://github.com/apache/spark/pull/36556, we will
>> break change when we merge it into 3.3.1
>>
>> At 2022-05-17 18:26:12, "Hyukjin Kwon" <gu...@gmail.com> wrote:
>>
>> We need add https://github.com/apache/spark/pull/36556 to RC2.
>>
>> We will likely have to change the version being added if RC2 passes.
>> Since this is a new API/improvement, I would prefer to not block the
>> release by that.
>>
>> On Tue, 17 May 2022 at 19:19, beliefer <be...@163.com> wrote:
>>
>>> We need add https://github.com/apache/spark/pull/36556 to RC2.
>>>
>>>
>>> 在 2022-05-17 17:37:13,"Hyukjin Kwon" <gu...@gmail.com> 写道:
>>>
>>> That seems like a test-only issue. I made a quick followup at
>>> https://github.com/apache/spark/pull/36576.
>>>
>>> On Tue, 17 May 2022 at 03:56, Sean Owen <sr...@gmail.com> wrote:
>>>
>>>> I'm still seeing failures related to the function registry, like:
>>>>
>>>> ExpressionsSchemaSuite:
>>>> - Check schemas for expression examples *** FAILED ***
>>>>   396 did not equal 398 Expected 396 blocks in result file but got 398.
>>>> Try regenerating the result files. (ExpressionsSchemaSuite.scala:161)
>>>>
>>>> - SPARK-14415: All functions should have own descriptions *** FAILED ***
>>>>   "Function: bloom_filter_aggClass:
>>>> org.apache.spark.sql.catalyst.expressions.aggregate.BloomFilterAggregateUsage:
>>>> N/A." contained "N/A." Failed for [function_desc: string] (N/A. existed in
>>>> the result) (QueryTest.scala:54)
>>>>
>>>> There seems to be consistently a difference of 2 in the list of
>>>> expected functions and actual. I haven't looked closely, don't know this
>>>> code. I'm on Ubuntu 22.04. Anyone else seeing something like this?
>>>> Wondering if it's something weird to do with case sensitivity, hidden files
>>>> lurking somewhere, etc.
>>>>
>>>> I suspect it's not a 'real' error as the Linux-based testers work fine,
>>>> but I also can't think of why this is failing.
>>>>
>>>>
>>>>
>>>> On Mon, May 16, 2022 at 7:44 AM Maxim Gekk
>>>> <ma...@databricks.com.invalid> wrote:
>>>>
>>>>> Please vote on releasing the following candidate as
>>>>> Apache Spark version 3.3.0.
>>>>>
>>>>> The vote is open until 11:59pm Pacific time May 19th and passes if a
>>>>> majority +1 PMC votes are cast, with a minimum of 3 +1 votes.
>>>>>
>>>>> [ ] +1 Release this package as Apache Spark 3.3.0
>>>>> [ ] -1 Do not release this package because ...
>>>>>
>>>>> To learn more about Apache Spark, please see http://spark.apache.org/
>>>>>
>>>>> The tag to be voted on is v3.3.0-rc2 (commit
>>>>> c8c657b922ac8fd8dcf9553113e11a80079db059):
>>>>> https://github.com/apache/spark/tree/v3.3.0-rc2
>>>>>
>>>>> The release files, including signatures, digests, etc. can be found at:
>>>>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-bin/
>>>>>
>>>>> Signatures used for Spark RCs can be found in this file:
>>>>> https://dist.apache.org/repos/dist/dev/spark/KEYS
>>>>>
>>>>> The staging repository for this release can be found at:
>>>>> https://repository.apache.org/content/repositories/orgapachespark-1403
>>>>>
>>>>> The documentation corresponding to this release can be found at:
>>>>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-docs/
>>>>>
>>>>> The list of bug fixes going into 3.3.0 can be found at the following
>>>>> URL:
>>>>> https://issues.apache.org/jira/projects/SPARK/versions/12350369
>>>>>
>>>>> This release is using the release script of the tag v3.3.0-rc2.
>>>>>
>>>>>
>>>>> FAQ
>>>>>
>>>>> =========================
>>>>> How can I help test this release?
>>>>> =========================
>>>>> If you are a Spark user, you can help us test this release by taking
>>>>> an existing Spark workload and running on this release candidate, then
>>>>> reporting any regressions.
>>>>>
>>>>> If you're working in PySpark you can set up a virtual env and install
>>>>> the current RC and see if anything important breaks, in the Java/Scala
>>>>> you can add the staging repository to your projects resolvers and test
>>>>> with the RC (make sure to clean up the artifact cache before/after so
>>>>> you don't end up building with a out of date RC going forward).
>>>>>
>>>>> ===========================================
>>>>> What should happen to JIRA tickets still targeting 3.3.0?
>>>>> ===========================================
>>>>> The current list of open tickets targeted at 3.3.0 can be found at:
>>>>> https://issues.apache.org/jira/projects/SPARK and search for "Target
>>>>> Version/s" = 3.3.0
>>>>>
>>>>> Committers should look at those and triage. Extremely important bug
>>>>> fixes, documentation, and API tweaks that impact compatibility should
>>>>> be worked on immediately. Everything else please retarget to an
>>>>> appropriate release.
>>>>>
>>>>> ==================
>>>>> But my bug isn't fixed?
>>>>> ==================
>>>>> In order to make timely releases, we will typically not hold the
>>>>> release unless the bug in question is a regression from the previous
>>>>> release. That being said, if there is something which is a regression
>>>>> that has not been correctly targeted please ping me or a committer to
>>>>> help target the issue.
>>>>>
>>>>> Maxim Gekk
>>>>>
>>>>> Software Engineer
>>>>>
>>>>> Databricks, Inc.
>>>>>
>>>>
>>>
>>>
>>>
>>
>>
>>
>>
>

Re: Re: Re: Re: [VOTE] Release Spark 3.3.0 (RC2)

Posted by Hyukjin Kwon <gu...@gmail.com>.
There might be other blockers. Lets wait and see.

On Tue, May 17, 2022 at 8:59 PM beliefer <be...@163.com> wrote:

> OK. let it into 3.3.1
>
>
> 在 2022-05-17 18:59:13,"Hyukjin Kwon" <gu...@gmail.com> 写道:
>
> I think most users won't be affected since aggregate pushdown is disabled
> by default.
>
> On Tue, 17 May 2022 at 19:53, beliefer <be...@163.com> wrote:
>
>> If we not contains https://github.com/apache/spark/pull/36556, we will
>> break change when we merge it into 3.3.1
>>
>> At 2022-05-17 18:26:12, "Hyukjin Kwon" <gu...@gmail.com> wrote:
>>
>> We need add https://github.com/apache/spark/pull/36556 to RC2.
>>
>> We will likely have to change the version being added if RC2 passes.
>> Since this is a new API/improvement, I would prefer to not block the
>> release by that.
>>
>> On Tue, 17 May 2022 at 19:19, beliefer <be...@163.com> wrote:
>>
>>> We need add https://github.com/apache/spark/pull/36556 to RC2.
>>>
>>>
>>> 在 2022-05-17 17:37:13,"Hyukjin Kwon" <gu...@gmail.com> 写道:
>>>
>>> That seems like a test-only issue. I made a quick followup at
>>> https://github.com/apache/spark/pull/36576.
>>>
>>> On Tue, 17 May 2022 at 03:56, Sean Owen <sr...@gmail.com> wrote:
>>>
>>>> I'm still seeing failures related to the function registry, like:
>>>>
>>>> ExpressionsSchemaSuite:
>>>> - Check schemas for expression examples *** FAILED ***
>>>>   396 did not equal 398 Expected 396 blocks in result file but got 398.
>>>> Try regenerating the result files. (ExpressionsSchemaSuite.scala:161)
>>>>
>>>> - SPARK-14415: All functions should have own descriptions *** FAILED ***
>>>>   "Function: bloom_filter_aggClass:
>>>> org.apache.spark.sql.catalyst.expressions.aggregate.BloomFilterAggregateUsage:
>>>> N/A." contained "N/A." Failed for [function_desc: string] (N/A. existed in
>>>> the result) (QueryTest.scala:54)
>>>>
>>>> There seems to be consistently a difference of 2 in the list of
>>>> expected functions and actual. I haven't looked closely, don't know this
>>>> code. I'm on Ubuntu 22.04. Anyone else seeing something like this?
>>>> Wondering if it's something weird to do with case sensitivity, hidden files
>>>> lurking somewhere, etc.
>>>>
>>>> I suspect it's not a 'real' error as the Linux-based testers work fine,
>>>> but I also can't think of why this is failing.
>>>>
>>>>
>>>>
>>>> On Mon, May 16, 2022 at 7:44 AM Maxim Gekk
>>>> <ma...@databricks.com.invalid> wrote:
>>>>
>>>>> Please vote on releasing the following candidate as
>>>>> Apache Spark version 3.3.0.
>>>>>
>>>>> The vote is open until 11:59pm Pacific time May 19th and passes if a
>>>>> majority +1 PMC votes are cast, with a minimum of 3 +1 votes.
>>>>>
>>>>> [ ] +1 Release this package as Apache Spark 3.3.0
>>>>> [ ] -1 Do not release this package because ...
>>>>>
>>>>> To learn more about Apache Spark, please see http://spark.apache.org/
>>>>>
>>>>> The tag to be voted on is v3.3.0-rc2 (commit
>>>>> c8c657b922ac8fd8dcf9553113e11a80079db059):
>>>>> https://github.com/apache/spark/tree/v3.3.0-rc2
>>>>>
>>>>> The release files, including signatures, digests, etc. can be found at:
>>>>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-bin/
>>>>>
>>>>> Signatures used for Spark RCs can be found in this file:
>>>>> https://dist.apache.org/repos/dist/dev/spark/KEYS
>>>>>
>>>>> The staging repository for this release can be found at:
>>>>> https://repository.apache.org/content/repositories/orgapachespark-1403
>>>>>
>>>>> The documentation corresponding to this release can be found at:
>>>>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-docs/
>>>>>
>>>>> The list of bug fixes going into 3.3.0 can be found at the following
>>>>> URL:
>>>>> https://issues.apache.org/jira/projects/SPARK/versions/12350369
>>>>>
>>>>> This release is using the release script of the tag v3.3.0-rc2.
>>>>>
>>>>>
>>>>> FAQ
>>>>>
>>>>> =========================
>>>>> How can I help test this release?
>>>>> =========================
>>>>> If you are a Spark user, you can help us test this release by taking
>>>>> an existing Spark workload and running on this release candidate, then
>>>>> reporting any regressions.
>>>>>
>>>>> If you're working in PySpark you can set up a virtual env and install
>>>>> the current RC and see if anything important breaks, in the Java/Scala
>>>>> you can add the staging repository to your projects resolvers and test
>>>>> with the RC (make sure to clean up the artifact cache before/after so
>>>>> you don't end up building with a out of date RC going forward).
>>>>>
>>>>> ===========================================
>>>>> What should happen to JIRA tickets still targeting 3.3.0?
>>>>> ===========================================
>>>>> The current list of open tickets targeted at 3.3.0 can be found at:
>>>>> https://issues.apache.org/jira/projects/SPARK and search for "Target
>>>>> Version/s" = 3.3.0
>>>>>
>>>>> Committers should look at those and triage. Extremely important bug
>>>>> fixes, documentation, and API tweaks that impact compatibility should
>>>>> be worked on immediately. Everything else please retarget to an
>>>>> appropriate release.
>>>>>
>>>>> ==================
>>>>> But my bug isn't fixed?
>>>>> ==================
>>>>> In order to make timely releases, we will typically not hold the
>>>>> release unless the bug in question is a regression from the previous
>>>>> release. That being said, if there is something which is a regression
>>>>> that has not been correctly targeted please ping me or a committer to
>>>>> help target the issue.
>>>>>
>>>>> Maxim Gekk
>>>>>
>>>>> Software Engineer
>>>>>
>>>>> Databricks, Inc.
>>>>>
>>>>
>>>
>>>
>>>
>>
>>
>>
>>
>
>
>
>

Re:Re: Re: Re: [VOTE] Release Spark 3.3.0 (RC2)

Posted by beliefer <be...@163.com>.
OK. let it into 3.3.1




在 2022-05-17 18:59:13,"Hyukjin Kwon" <gu...@gmail.com> 写道:

I think most users won't be affected since aggregate pushdown is disabled by default.


On Tue, 17 May 2022 at 19:53, beliefer <be...@163.com> wrote:


If we not contains https://github.com/apache/spark/pull/36556, we will break change when we merge it into 3.3.1

At 2022-05-17 18:26:12, "Hyukjin Kwon" <gu...@gmail.com> wrote:

We need add https://github.com/apache/spark/pull/36556 to RC2.

We will likely have to change the version being added if RC2 passes.
Since this is a new API/improvement, I would prefer to not block the release by that.



On Tue, 17 May 2022 at 19:19, beliefer <be...@163.com> wrote:

We need add https://github.com/apache/spark/pull/36556 to RC2.




在 2022-05-17 17:37:13,"Hyukjin Kwon" <gu...@gmail.com> 写道:

That seems like a test-only issue. I made a quick followup at https://github.com/apache/spark/pull/36576.


On Tue, 17 May 2022 at 03:56, Sean Owen <sr...@gmail.com> wrote:

I'm still seeing failures related to the function registry, like:


ExpressionsSchemaSuite:
- Check schemas for expression examples *** FAILED ***
  396 did not equal 398 Expected 396 blocks in result file but got 398. Try regenerating the result files. (ExpressionsSchemaSuite.scala:161)

- SPARK-14415: All functions should have own descriptions *** FAILED ***
  "Function: bloom_filter_aggClass: org.apache.spark.sql.catalyst.expressions.aggregate.BloomFilterAggregateUsage: N/A." contained "N/A." Failed for [function_desc: string] (N/A. existed in the result) (QueryTest.scala:54)



There seems to be consistently a difference of 2 in the list of expected functions and actual. I haven't looked closely, don't know this code. I'm on Ubuntu 22.04. Anyone else seeing something like this? Wondering if it's something weird to do with case sensitivity, hidden files lurking somewhere, etc.


I suspect it's not a 'real' error as the Linux-based testers work fine, but I also can't think of why this is failing.






On Mon, May 16, 2022 at 7:44 AM Maxim Gekk <ma...@databricks.com.invalid> wrote:

Please vote on releasing the following candidate as Apache Spark version 3.3.0.



The vote is open until 11:59pm Pacific time May 19th and passes if a majority +1 PMC votes are cast, with a minimum of 3 +1 votes.



[ ] +1 Release this package as Apache Spark 3.3.0

[ ] -1 Do not release this package because ...



To learn more about Apache Spark, please see http://spark.apache.org/



The tag to be voted on is v3.3.0-rc2 (commit c8c657b922ac8fd8dcf9553113e11a80079db059):

https://github.com/apache/spark/tree/v3.3.0-rc2



The release files, including signatures, digests, etc. can be found at:

https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-bin/



Signatures used for Spark RCs can be found in this file:

https://dist.apache.org/repos/dist/dev/spark/KEYS



The staging repository for this release can be found at:

https://repository.apache.org/content/repositories/orgapachespark-1403



The documentation corresponding to this release can be found at:

https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-docs/



The list of bug fixes going into 3.3.0 can be found at the following URL:

https://issues.apache.org/jira/projects/SPARK/versions/12350369


This release is using the release script of the tag v3.3.0-rc2.





FAQ



=========================

How can I help test this release?

=========================

If you are a Spark user, you can help us test this release by taking

an existing Spark workload and running on this release candidate, then

reporting any regressions.



If you're working in PySpark you can set up a virtual env and install

the current RC and see if anything important breaks, in the Java/Scala

you can add the staging repository to your projects resolvers and test

with the RC (make sure to clean up the artifact cache before/after so

you don't end up building with a out of date RC going forward).



===========================================

What should happen to JIRA tickets still targeting 3.3.0?

===========================================

The current list of open tickets targeted at 3.3.0 can be found at:

https://issues.apache.org/jira/projects/SPARK and search for "Target Version/s" = 3.3.0



Committers should look at those and triage. Extremely important bug

fixes, documentation, and API tweaks that impact compatibility should

be worked on immediately. Everything else please retarget to an

appropriate release.



==================

But my bug isn't fixed?

==================

In order to make timely releases, we will typically not hold the

release unless the bug in question is a regression from the previous

release. That being said, if there is something which is a regression

that has not been correctly targeted please ping me or a committer to

help target the issue.


Maxim Gekk


Software Engineer

Databricks, Inc.





 





 

Re: Re: Re: [VOTE] Release Spark 3.3.0 (RC2)

Posted by Hyukjin Kwon <gu...@gmail.com>.
I think most users won't be affected since aggregate pushdown is disabled
by default.

On Tue, 17 May 2022 at 19:53, beliefer <be...@163.com> wrote:

> If we not contains https://github.com/apache/spark/pull/36556, we will
> break change when we merge it into 3.3.1
>
> At 2022-05-17 18:26:12, "Hyukjin Kwon" <gu...@gmail.com> wrote:
>
> We need add https://github.com/apache/spark/pull/36556 to RC2.
>
> We will likely have to change the version being added if RC2 passes.
> Since this is a new API/improvement, I would prefer to not block the
> release by that.
>
> On Tue, 17 May 2022 at 19:19, beliefer <be...@163.com> wrote:
>
>> We need add https://github.com/apache/spark/pull/36556 to RC2.
>>
>>
>> 在 2022-05-17 17:37:13,"Hyukjin Kwon" <gu...@gmail.com> 写道:
>>
>> That seems like a test-only issue. I made a quick followup at
>> https://github.com/apache/spark/pull/36576.
>>
>> On Tue, 17 May 2022 at 03:56, Sean Owen <sr...@gmail.com> wrote:
>>
>>> I'm still seeing failures related to the function registry, like:
>>>
>>> ExpressionsSchemaSuite:
>>> - Check schemas for expression examples *** FAILED ***
>>>   396 did not equal 398 Expected 396 blocks in result file but got 398.
>>> Try regenerating the result files. (ExpressionsSchemaSuite.scala:161)
>>>
>>> - SPARK-14415: All functions should have own descriptions *** FAILED ***
>>>   "Function: bloom_filter_aggClass:
>>> org.apache.spark.sql.catalyst.expressions.aggregate.BloomFilterAggregateUsage:
>>> N/A." contained "N/A." Failed for [function_desc: string] (N/A. existed in
>>> the result) (QueryTest.scala:54)
>>>
>>> There seems to be consistently a difference of 2 in the list of expected
>>> functions and actual. I haven't looked closely, don't know this code. I'm
>>> on Ubuntu 22.04. Anyone else seeing something like this? Wondering if it's
>>> something weird to do with case sensitivity, hidden files lurking
>>> somewhere, etc.
>>>
>>> I suspect it's not a 'real' error as the Linux-based testers work fine,
>>> but I also can't think of why this is failing.
>>>
>>>
>>>
>>> On Mon, May 16, 2022 at 7:44 AM Maxim Gekk
>>> <ma...@databricks.com.invalid> wrote:
>>>
>>>> Please vote on releasing the following candidate as
>>>> Apache Spark version 3.3.0.
>>>>
>>>> The vote is open until 11:59pm Pacific time May 19th and passes if a
>>>> majority +1 PMC votes are cast, with a minimum of 3 +1 votes.
>>>>
>>>> [ ] +1 Release this package as Apache Spark 3.3.0
>>>> [ ] -1 Do not release this package because ...
>>>>
>>>> To learn more about Apache Spark, please see http://spark.apache.org/
>>>>
>>>> The tag to be voted on is v3.3.0-rc2 (commit
>>>> c8c657b922ac8fd8dcf9553113e11a80079db059):
>>>> https://github.com/apache/spark/tree/v3.3.0-rc2
>>>>
>>>> The release files, including signatures, digests, etc. can be found at:
>>>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-bin/
>>>>
>>>> Signatures used for Spark RCs can be found in this file:
>>>> https://dist.apache.org/repos/dist/dev/spark/KEYS
>>>>
>>>> The staging repository for this release can be found at:
>>>> https://repository.apache.org/content/repositories/orgapachespark-1403
>>>>
>>>> The documentation corresponding to this release can be found at:
>>>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-docs/
>>>>
>>>> The list of bug fixes going into 3.3.0 can be found at the following
>>>> URL:
>>>> https://issues.apache.org/jira/projects/SPARK/versions/12350369
>>>>
>>>> This release is using the release script of the tag v3.3.0-rc2.
>>>>
>>>>
>>>> FAQ
>>>>
>>>> =========================
>>>> How can I help test this release?
>>>> =========================
>>>> If you are a Spark user, you can help us test this release by taking
>>>> an existing Spark workload and running on this release candidate, then
>>>> reporting any regressions.
>>>>
>>>> If you're working in PySpark you can set up a virtual env and install
>>>> the current RC and see if anything important breaks, in the Java/Scala
>>>> you can add the staging repository to your projects resolvers and test
>>>> with the RC (make sure to clean up the artifact cache before/after so
>>>> you don't end up building with a out of date RC going forward).
>>>>
>>>> ===========================================
>>>> What should happen to JIRA tickets still targeting 3.3.0?
>>>> ===========================================
>>>> The current list of open tickets targeted at 3.3.0 can be found at:
>>>> https://issues.apache.org/jira/projects/SPARK and search for "Target
>>>> Version/s" = 3.3.0
>>>>
>>>> Committers should look at those and triage. Extremely important bug
>>>> fixes, documentation, and API tweaks that impact compatibility should
>>>> be worked on immediately. Everything else please retarget to an
>>>> appropriate release.
>>>>
>>>> ==================
>>>> But my bug isn't fixed?
>>>> ==================
>>>> In order to make timely releases, we will typically not hold the
>>>> release unless the bug in question is a regression from the previous
>>>> release. That being said, if there is something which is a regression
>>>> that has not been correctly targeted please ping me or a committer to
>>>> help target the issue.
>>>>
>>>> Maxim Gekk
>>>>
>>>> Software Engineer
>>>>
>>>> Databricks, Inc.
>>>>
>>>
>>
>>
>>
>
>
>
>

Re:Re: Re: [VOTE] Release Spark 3.3.0 (RC2)

Posted by beliefer <be...@163.com>.
If we not contains https://github.com/apache/spark/pull/36556, we will break change when we merge it into 3.3.1

At 2022-05-17 18:26:12, "Hyukjin Kwon" <gu...@gmail.com> wrote:

We need add https://github.com/apache/spark/pull/36556 to RC2.

We will likely have to change the version being added if RC2 passes.
Since this is a new API/improvement, I would prefer to not block the release by that.



On Tue, 17 May 2022 at 19:19, beliefer <be...@163.com> wrote:

We need add https://github.com/apache/spark/pull/36556 to RC2.




在 2022-05-17 17:37:13,"Hyukjin Kwon" <gu...@gmail.com> 写道:

That seems like a test-only issue. I made a quick followup at https://github.com/apache/spark/pull/36576.


On Tue, 17 May 2022 at 03:56, Sean Owen <sr...@gmail.com> wrote:

I'm still seeing failures related to the function registry, like:


ExpressionsSchemaSuite:
- Check schemas for expression examples *** FAILED ***
  396 did not equal 398 Expected 396 blocks in result file but got 398. Try regenerating the result files. (ExpressionsSchemaSuite.scala:161)

- SPARK-14415: All functions should have own descriptions *** FAILED ***
  "Function: bloom_filter_aggClass: org.apache.spark.sql.catalyst.expressions.aggregate.BloomFilterAggregateUsage: N/A." contained "N/A." Failed for [function_desc: string] (N/A. existed in the result) (QueryTest.scala:54)



There seems to be consistently a difference of 2 in the list of expected functions and actual. I haven't looked closely, don't know this code. I'm on Ubuntu 22.04. Anyone else seeing something like this? Wondering if it's something weird to do with case sensitivity, hidden files lurking somewhere, etc.


I suspect it's not a 'real' error as the Linux-based testers work fine, but I also can't think of why this is failing.






On Mon, May 16, 2022 at 7:44 AM Maxim Gekk <ma...@databricks.com.invalid> wrote:

Please vote on releasing the following candidate as Apache Spark version 3.3.0.



The vote is open until 11:59pm Pacific time May 19th and passes if a majority +1 PMC votes are cast, with a minimum of 3 +1 votes.



[ ] +1 Release this package as Apache Spark 3.3.0

[ ] -1 Do not release this package because ...



To learn more about Apache Spark, please see http://spark.apache.org/



The tag to be voted on is v3.3.0-rc2 (commit c8c657b922ac8fd8dcf9553113e11a80079db059):

https://github.com/apache/spark/tree/v3.3.0-rc2



The release files, including signatures, digests, etc. can be found at:

https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-bin/



Signatures used for Spark RCs can be found in this file:

https://dist.apache.org/repos/dist/dev/spark/KEYS



The staging repository for this release can be found at:

https://repository.apache.org/content/repositories/orgapachespark-1403



The documentation corresponding to this release can be found at:

https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-docs/



The list of bug fixes going into 3.3.0 can be found at the following URL:

https://issues.apache.org/jira/projects/SPARK/versions/12350369


This release is using the release script of the tag v3.3.0-rc2.





FAQ



=========================

How can I help test this release?

=========================

If you are a Spark user, you can help us test this release by taking

an existing Spark workload and running on this release candidate, then

reporting any regressions.



If you're working in PySpark you can set up a virtual env and install

the current RC and see if anything important breaks, in the Java/Scala

you can add the staging repository to your projects resolvers and test

with the RC (make sure to clean up the artifact cache before/after so

you don't end up building with a out of date RC going forward).



===========================================

What should happen to JIRA tickets still targeting 3.3.0?

===========================================

The current list of open tickets targeted at 3.3.0 can be found at:

https://issues.apache.org/jira/projects/SPARK and search for "Target Version/s" = 3.3.0



Committers should look at those and triage. Extremely important bug

fixes, documentation, and API tweaks that impact compatibility should

be worked on immediately. Everything else please retarget to an

appropriate release.



==================

But my bug isn't fixed?

==================

In order to make timely releases, we will typically not hold the

release unless the bug in question is a regression from the previous

release. That being said, if there is something which is a regression

that has not been correctly targeted please ping me or a committer to

help target the issue.


Maxim Gekk


Software Engineer

Databricks, Inc.





 

Re: Re: [VOTE] Release Spark 3.3.0 (RC2)

Posted by Hyukjin Kwon <gu...@gmail.com>.
We need add https://github.com/apache/spark/pull/36556 to RC2.

We will likely have to change the version being added if RC2 passes.
Since this is a new API/improvement, I would prefer to not block the
release by that.

On Tue, 17 May 2022 at 19:19, beliefer <be...@163.com> wrote:

> We need add https://github.com/apache/spark/pull/36556 to RC2.
>
>
> 在 2022-05-17 17:37:13,"Hyukjin Kwon" <gu...@gmail.com> 写道:
>
> That seems like a test-only issue. I made a quick followup at
> https://github.com/apache/spark/pull/36576.
>
> On Tue, 17 May 2022 at 03:56, Sean Owen <sr...@gmail.com> wrote:
>
>> I'm still seeing failures related to the function registry, like:
>>
>> ExpressionsSchemaSuite:
>> - Check schemas for expression examples *** FAILED ***
>>   396 did not equal 398 Expected 396 blocks in result file but got 398.
>> Try regenerating the result files. (ExpressionsSchemaSuite.scala:161)
>>
>> - SPARK-14415: All functions should have own descriptions *** FAILED ***
>>   "Function: bloom_filter_aggClass:
>> org.apache.spark.sql.catalyst.expressions.aggregate.BloomFilterAggregateUsage:
>> N/A." contained "N/A." Failed for [function_desc: string] (N/A. existed in
>> the result) (QueryTest.scala:54)
>>
>> There seems to be consistently a difference of 2 in the list of expected
>> functions and actual. I haven't looked closely, don't know this code. I'm
>> on Ubuntu 22.04. Anyone else seeing something like this? Wondering if it's
>> something weird to do with case sensitivity, hidden files lurking
>> somewhere, etc.
>>
>> I suspect it's not a 'real' error as the Linux-based testers work fine,
>> but I also can't think of why this is failing.
>>
>>
>>
>> On Mon, May 16, 2022 at 7:44 AM Maxim Gekk
>> <ma...@databricks.com.invalid> wrote:
>>
>>> Please vote on releasing the following candidate as
>>> Apache Spark version 3.3.0.
>>>
>>> The vote is open until 11:59pm Pacific time May 19th and passes if a
>>> majority +1 PMC votes are cast, with a minimum of 3 +1 votes.
>>>
>>> [ ] +1 Release this package as Apache Spark 3.3.0
>>> [ ] -1 Do not release this package because ...
>>>
>>> To learn more about Apache Spark, please see http://spark.apache.org/
>>>
>>> The tag to be voted on is v3.3.0-rc2 (commit
>>> c8c657b922ac8fd8dcf9553113e11a80079db059):
>>> https://github.com/apache/spark/tree/v3.3.0-rc2
>>>
>>> The release files, including signatures, digests, etc. can be found at:
>>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-bin/
>>>
>>> Signatures used for Spark RCs can be found in this file:
>>> https://dist.apache.org/repos/dist/dev/spark/KEYS
>>>
>>> The staging repository for this release can be found at:
>>> https://repository.apache.org/content/repositories/orgapachespark-1403
>>>
>>> The documentation corresponding to this release can be found at:
>>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-docs/
>>>
>>> The list of bug fixes going into 3.3.0 can be found at the following URL:
>>> https://issues.apache.org/jira/projects/SPARK/versions/12350369
>>>
>>> This release is using the release script of the tag v3.3.0-rc2.
>>>
>>>
>>> FAQ
>>>
>>> =========================
>>> How can I help test this release?
>>> =========================
>>> If you are a Spark user, you can help us test this release by taking
>>> an existing Spark workload and running on this release candidate, then
>>> reporting any regressions.
>>>
>>> If you're working in PySpark you can set up a virtual env and install
>>> the current RC and see if anything important breaks, in the Java/Scala
>>> you can add the staging repository to your projects resolvers and test
>>> with the RC (make sure to clean up the artifact cache before/after so
>>> you don't end up building with a out of date RC going forward).
>>>
>>> ===========================================
>>> What should happen to JIRA tickets still targeting 3.3.0?
>>> ===========================================
>>> The current list of open tickets targeted at 3.3.0 can be found at:
>>> https://issues.apache.org/jira/projects/SPARK and search for "Target
>>> Version/s" = 3.3.0
>>>
>>> Committers should look at those and triage. Extremely important bug
>>> fixes, documentation, and API tweaks that impact compatibility should
>>> be worked on immediately. Everything else please retarget to an
>>> appropriate release.
>>>
>>> ==================
>>> But my bug isn't fixed?
>>> ==================
>>> In order to make timely releases, we will typically not hold the
>>> release unless the bug in question is a regression from the previous
>>> release. That being said, if there is something which is a regression
>>> that has not been correctly targeted please ping me or a committer to
>>> help target the issue.
>>>
>>> Maxim Gekk
>>>
>>> Software Engineer
>>>
>>> Databricks, Inc.
>>>
>>
>
>
>

Re:Re: [VOTE] Release Spark 3.3.0 (RC2)

Posted by beliefer <be...@163.com>.
We need add https://github.com/apache/spark/pull/36556 to RC2.




在 2022-05-17 17:37:13,"Hyukjin Kwon" <gu...@gmail.com> 写道:

That seems like a test-only issue. I made a quick followup at https://github.com/apache/spark/pull/36576.


On Tue, 17 May 2022 at 03:56, Sean Owen <sr...@gmail.com> wrote:

I'm still seeing failures related to the function registry, like:


ExpressionsSchemaSuite:
- Check schemas for expression examples *** FAILED ***
  396 did not equal 398 Expected 396 blocks in result file but got 398. Try regenerating the result files. (ExpressionsSchemaSuite.scala:161)

- SPARK-14415: All functions should have own descriptions *** FAILED ***
  "Function: bloom_filter_aggClass: org.apache.spark.sql.catalyst.expressions.aggregate.BloomFilterAggregateUsage: N/A." contained "N/A." Failed for [function_desc: string] (N/A. existed in the result) (QueryTest.scala:54)



There seems to be consistently a difference of 2 in the list of expected functions and actual. I haven't looked closely, don't know this code. I'm on Ubuntu 22.04. Anyone else seeing something like this? Wondering if it's something weird to do with case sensitivity, hidden files lurking somewhere, etc.


I suspect it's not a 'real' error as the Linux-based testers work fine, but I also can't think of why this is failing.






On Mon, May 16, 2022 at 7:44 AM Maxim Gekk <ma...@databricks.com.invalid> wrote:

Please vote on releasing the following candidate as Apache Spark version 3.3.0.



The vote is open until 11:59pm Pacific time May 19th and passes if a majority +1 PMC votes are cast, with a minimum of 3 +1 votes.



[ ] +1 Release this package as Apache Spark 3.3.0

[ ] -1 Do not release this package because ...



To learn more about Apache Spark, please see http://spark.apache.org/



The tag to be voted on is v3.3.0-rc2 (commit c8c657b922ac8fd8dcf9553113e11a80079db059):

https://github.com/apache/spark/tree/v3.3.0-rc2



The release files, including signatures, digests, etc. can be found at:

https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-bin/



Signatures used for Spark RCs can be found in this file:

https://dist.apache.org/repos/dist/dev/spark/KEYS



The staging repository for this release can be found at:

https://repository.apache.org/content/repositories/orgapachespark-1403



The documentation corresponding to this release can be found at:

https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-docs/



The list of bug fixes going into 3.3.0 can be found at the following URL:

https://issues.apache.org/jira/projects/SPARK/versions/12350369


This release is using the release script of the tag v3.3.0-rc2.





FAQ



=========================

How can I help test this release?

=========================

If you are a Spark user, you can help us test this release by taking

an existing Spark workload and running on this release candidate, then

reporting any regressions.



If you're working in PySpark you can set up a virtual env and install

the current RC and see if anything important breaks, in the Java/Scala

you can add the staging repository to your projects resolvers and test

with the RC (make sure to clean up the artifact cache before/after so

you don't end up building with a out of date RC going forward).



===========================================

What should happen to JIRA tickets still targeting 3.3.0?

===========================================

The current list of open tickets targeted at 3.3.0 can be found at:

https://issues.apache.org/jira/projects/SPARK and search for "Target Version/s" = 3.3.0



Committers should look at those and triage. Extremely important bug

fixes, documentation, and API tweaks that impact compatibility should

be worked on immediately. Everything else please retarget to an

appropriate release.



==================

But my bug isn't fixed?

==================

In order to make timely releases, we will typically not hold the

release unless the bug in question is a regression from the previous

release. That being said, if there is something which is a regression

that has not been correctly targeted please ping me or a committer to

help target the issue.


Maxim Gekk


Software Engineer

Databricks, Inc.

Re: [VOTE] Release Spark 3.3.0 (RC2)

Posted by Hyukjin Kwon <gu...@gmail.com>.
That seems like a test-only issue. I made a quick followup at
https://github.com/apache/spark/pull/36576.

On Tue, 17 May 2022 at 03:56, Sean Owen <sr...@gmail.com> wrote:

> I'm still seeing failures related to the function registry, like:
>
> ExpressionsSchemaSuite:
> - Check schemas for expression examples *** FAILED ***
>   396 did not equal 398 Expected 396 blocks in result file but got 398.
> Try regenerating the result files. (ExpressionsSchemaSuite.scala:161)
>
> - SPARK-14415: All functions should have own descriptions *** FAILED ***
>   "Function: bloom_filter_aggClass:
> org.apache.spark.sql.catalyst.expressions.aggregate.BloomFilterAggregateUsage:
> N/A." contained "N/A." Failed for [function_desc: string] (N/A. existed in
> the result) (QueryTest.scala:54)
>
> There seems to be consistently a difference of 2 in the list of expected
> functions and actual. I haven't looked closely, don't know this code. I'm
> on Ubuntu 22.04. Anyone else seeing something like this? Wondering if it's
> something weird to do with case sensitivity, hidden files lurking
> somewhere, etc.
>
> I suspect it's not a 'real' error as the Linux-based testers work fine,
> but I also can't think of why this is failing.
>
>
>
> On Mon, May 16, 2022 at 7:44 AM Maxim Gekk
> <ma...@databricks.com.invalid> wrote:
>
>> Please vote on releasing the following candidate as
>> Apache Spark version 3.3.0.
>>
>> The vote is open until 11:59pm Pacific time May 19th and passes if a
>> majority +1 PMC votes are cast, with a minimum of 3 +1 votes.
>>
>> [ ] +1 Release this package as Apache Spark 3.3.0
>> [ ] -1 Do not release this package because ...
>>
>> To learn more about Apache Spark, please see http://spark.apache.org/
>>
>> The tag to be voted on is v3.3.0-rc2 (commit
>> c8c657b922ac8fd8dcf9553113e11a80079db059):
>> https://github.com/apache/spark/tree/v3.3.0-rc2
>>
>> The release files, including signatures, digests, etc. can be found at:
>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-bin/
>>
>> Signatures used for Spark RCs can be found in this file:
>> https://dist.apache.org/repos/dist/dev/spark/KEYS
>>
>> The staging repository for this release can be found at:
>> https://repository.apache.org/content/repositories/orgapachespark-1403
>>
>> The documentation corresponding to this release can be found at:
>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-docs/
>>
>> The list of bug fixes going into 3.3.0 can be found at the following URL:
>> https://issues.apache.org/jira/projects/SPARK/versions/12350369
>>
>> This release is using the release script of the tag v3.3.0-rc2.
>>
>>
>> FAQ
>>
>> =========================
>> How can I help test this release?
>> =========================
>> If you are a Spark user, you can help us test this release by taking
>> an existing Spark workload and running on this release candidate, then
>> reporting any regressions.
>>
>> If you're working in PySpark you can set up a virtual env and install
>> the current RC and see if anything important breaks, in the Java/Scala
>> you can add the staging repository to your projects resolvers and test
>> with the RC (make sure to clean up the artifact cache before/after so
>> you don't end up building with a out of date RC going forward).
>>
>> ===========================================
>> What should happen to JIRA tickets still targeting 3.3.0?
>> ===========================================
>> The current list of open tickets targeted at 3.3.0 can be found at:
>> https://issues.apache.org/jira/projects/SPARK and search for "Target
>> Version/s" = 3.3.0
>>
>> Committers should look at those and triage. Extremely important bug
>> fixes, documentation, and API tweaks that impact compatibility should
>> be worked on immediately. Everything else please retarget to an
>> appropriate release.
>>
>> ==================
>> But my bug isn't fixed?
>> ==================
>> In order to make timely releases, we will typically not hold the
>> release unless the bug in question is a regression from the previous
>> release. That being said, if there is something which is a regression
>> that has not been correctly targeted please ping me or a committer to
>> help target the issue.
>>
>> Maxim Gekk
>>
>> Software Engineer
>>
>> Databricks, Inc.
>>
>

Re: [VOTE] Release Spark 3.3.0 (RC2)

Posted by Sean Owen <sr...@gmail.com>.
I'm still seeing failures related to the function registry, like:

ExpressionsSchemaSuite:
- Check schemas for expression examples *** FAILED ***
  396 did not equal 398 Expected 396 blocks in result file but got 398. Try
regenerating the result files. (ExpressionsSchemaSuite.scala:161)

- SPARK-14415: All functions should have own descriptions *** FAILED ***
  "Function: bloom_filter_aggClass:
org.apache.spark.sql.catalyst.expressions.aggregate.BloomFilterAggregateUsage:
N/A." contained "N/A." Failed for [function_desc: string] (N/A. existed in
the result) (QueryTest.scala:54)

There seems to be consistently a difference of 2 in the list of expected
functions and actual. I haven't looked closely, don't know this code. I'm
on Ubuntu 22.04. Anyone else seeing something like this? Wondering if it's
something weird to do with case sensitivity, hidden files lurking
somewhere, etc.

I suspect it's not a 'real' error as the Linux-based testers work fine, but
I also can't think of why this is failing.



On Mon, May 16, 2022 at 7:44 AM Maxim Gekk
<ma...@databricks.com.invalid> wrote:

> Please vote on releasing the following candidate as
> Apache Spark version 3.3.0.
>
> The vote is open until 11:59pm Pacific time May 19th and passes if a
> majority +1 PMC votes are cast, with a minimum of 3 +1 votes.
>
> [ ] +1 Release this package as Apache Spark 3.3.0
> [ ] -1 Do not release this package because ...
>
> To learn more about Apache Spark, please see http://spark.apache.org/
>
> The tag to be voted on is v3.3.0-rc2 (commit
> c8c657b922ac8fd8dcf9553113e11a80079db059):
> https://github.com/apache/spark/tree/v3.3.0-rc2
>
> The release files, including signatures, digests, etc. can be found at:
> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-bin/
>
> Signatures used for Spark RCs can be found in this file:
> https://dist.apache.org/repos/dist/dev/spark/KEYS
>
> The staging repository for this release can be found at:
> https://repository.apache.org/content/repositories/orgapachespark-1403
>
> The documentation corresponding to this release can be found at:
> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-docs/
>
> The list of bug fixes going into 3.3.0 can be found at the following URL:
> https://issues.apache.org/jira/projects/SPARK/versions/12350369
>
> This release is using the release script of the tag v3.3.0-rc2.
>
>
> FAQ
>
> =========================
> How can I help test this release?
> =========================
> If you are a Spark user, you can help us test this release by taking
> an existing Spark workload and running on this release candidate, then
> reporting any regressions.
>
> If you're working in PySpark you can set up a virtual env and install
> the current RC and see if anything important breaks, in the Java/Scala
> you can add the staging repository to your projects resolvers and test
> with the RC (make sure to clean up the artifact cache before/after so
> you don't end up building with a out of date RC going forward).
>
> ===========================================
> What should happen to JIRA tickets still targeting 3.3.0?
> ===========================================
> The current list of open tickets targeted at 3.3.0 can be found at:
> https://issues.apache.org/jira/projects/SPARK and search for "Target
> Version/s" = 3.3.0
>
> Committers should look at those and triage. Extremely important bug
> fixes, documentation, and API tweaks that impact compatibility should
> be worked on immediately. Everything else please retarget to an
> appropriate release.
>
> ==================
> But my bug isn't fixed?
> ==================
> In order to make timely releases, we will typically not hold the
> release unless the bug in question is a regression from the previous
> release. That being said, if there is something which is a regression
> that has not been correctly targeted please ping me or a committer to
> help target the issue.
>
> Maxim Gekk
>
> Software Engineer
>
> Databricks, Inc.
>

Re: [VOTE] Release Spark 3.3.0 (RC2)

Posted by Dongjoon Hyun <do...@gmail.com>.
Thank you, Maxim!

Dongjoon.

On Thu, May 19, 2022 at 11:49 PM Maxim Gekk
<ma...@databricks.com.invalid> wrote:

> Hi All,
>
> The voting for Spark 3.3.0 RC2 has failed since there aren't enough +1 and
> due to reported bugs. I will prepare RC3 at the beginning of next week.
>
> All known issues have been resolved in 3.3 already (at least the issues
> reported in the thread). Please, test the current branch-3.3 and report any
> issues.
>
> Maxim Gekk
>
> Software Engineer
>
> Databricks, Inc.
>
>
> On Thu, May 19, 2022 at 1:59 PM Emil Ejbyfeldt
> <ee...@liveintent.com.invalid> wrote:
>
>> Hi,
>>
>> When testing out Spark 3.3.0 on our production spark workload it was
>> noticed that https://issues.apache.org/jira/browse/SPARK-38681 is
>> actually a regression from 3.2 (I did not know this a the time of
>> creating the ticket) seem like the bug was introduced in
>> https://github.com/apache/spark/pull/33205
>>
>> I already have a PR here that fixes the issue:
>> https://github.com/apache/spark/pull/36004
>>
>> Since this is a breaking regression for us I think it might be for other
>> people as well.
>>
>> Best,
>> Emil
>>
>> On 16/05/2022 14:43, Maxim Gekk wrote:
>> > Please vote on releasing the following candidate as
>> > Apache Spark version 3.3.0.
>> >
>> > The vote is open until 11:59pm Pacific time May 19th and passes if a
>> > majority +1 PMC votes are cast, with a minimum of 3 +1 votes.
>> >
>> > [ ] +1 Release this package as Apache Spark 3.3.0
>> > [ ] -1 Do not release this package because ...
>> >
>> > To learn more about Apache Spark, please see http://spark.apache.org/
>> > <http://spark.apache.org/>
>> >
>> > The tag to be voted on is v3.3.0-rc2 (commit
>> > c8c657b922ac8fd8dcf9553113e11a80079db059):
>> > https://github.com/apache/spark/tree/v3.3.0-rc2
>> > <https://github.com/apache/spark/tree/v3.3.0-rc2>
>> >
>> > The release files, including signatures, digests, etc. can be found at:
>> > https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-bin/
>> > <https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-bin/>
>> >
>> > Signatures used for Spark RCs can be found in this file:
>> > https://dist.apache.org/repos/dist/dev/spark/KEYS
>> > <https://dist.apache.org/repos/dist/dev/spark/KEYS>
>> >
>> > The staging repository for this release can be found at:
>> > https://repository.apache.org/content/repositories/orgapachespark-1403
>> > <https://repository.apache.org/content/repositories/orgapachespark-1403
>> >
>> >
>> > The documentation corresponding to this release can be found at:
>> > https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-docs/
>> > <https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-docs/>
>> >
>> > The list of bug fixes going into 3.3.0 can be found at the following
>> URL:
>> > https://issues.apache.org/jira/projects/SPARK/versions/12350369
>> > <https://issues.apache.org/jira/projects/SPARK/versions/12350369>
>> >
>> > This release is using the release script of the tag v3.3.0-rc2.
>> >
>> >
>> > FAQ
>> >
>> > =========================
>> > How can I help test this release?
>> > =========================
>> > If you are a Spark user, you can help us test this release by taking
>> > an existing Spark workload and running on this release candidate, then
>> > reporting any regressions.
>> >
>> > If you're working in PySpark you can set up a virtual env and install
>> > the current RC and see if anything important breaks, in the Java/Scala
>> > you can add the staging repository to your projects resolvers and test
>> > with the RC (make sure to clean up the artifact cache before/after so
>> > you don't end up building with a out of date RC going forward).
>> >
>> > ===========================================
>> > What should happen to JIRA tickets still targeting 3.3.0?
>> > ===========================================
>> > The current list of open tickets targeted at 3.3.0 can be found at:
>> > https://issues.apache.org/jira/projects/SPARK
>> > <https://issues.apache.org/jira/projects/SPARK> and search for "Target
>> > Version/s" = 3.3.0
>> >
>> > Committers should look at those and triage. Extremely important bug
>> > fixes, documentation, and API tweaks that impact compatibility should
>> > be worked on immediately. Everything else please retarget to an
>> > appropriate release.
>> >
>> > ==================
>> > But my bug isn't fixed?
>> > ==================
>> > In order to make timely releases, we will typically not hold the
>> > release unless the bug in question is a regression from the previous
>> > release. That being said, if there is something which is a regression
>> > that has not been correctly targeted please ping me or a committer to
>> > help target the issue.
>> >
>> > Maxim Gekk
>> >
>> > Software Engineer
>> >
>> > Databricks, Inc.
>> >
>>
>> ---------------------------------------------------------------------
>> To unsubscribe e-mail: dev-unsubscribe@spark.apache.org
>>
>>

Re: [VOTE] Release Spark 3.3.0 (RC2)

Posted by Maxim Gekk <ma...@databricks.com.INVALID>.
Hi All,

The voting for Spark 3.3.0 RC2 has failed since there aren't enough +1 and
due to reported bugs. I will prepare RC3 at the beginning of next week.

All known issues have been resolved in 3.3 already (at least the issues
reported in the thread). Please, test the current branch-3.3 and report any
issues.

Maxim Gekk

Software Engineer

Databricks, Inc.


On Thu, May 19, 2022 at 1:59 PM Emil Ejbyfeldt
<ee...@liveintent.com.invalid> wrote:

> Hi,
>
> When testing out Spark 3.3.0 on our production spark workload it was
> noticed that https://issues.apache.org/jira/browse/SPARK-38681 is
> actually a regression from 3.2 (I did not know this a the time of
> creating the ticket) seem like the bug was introduced in
> https://github.com/apache/spark/pull/33205
>
> I already have a PR here that fixes the issue:
> https://github.com/apache/spark/pull/36004
>
> Since this is a breaking regression for us I think it might be for other
> people as well.
>
> Best,
> Emil
>
> On 16/05/2022 14:43, Maxim Gekk wrote:
> > Please vote on releasing the following candidate as
> > Apache Spark version 3.3.0.
> >
> > The vote is open until 11:59pm Pacific time May 19th and passes if a
> > majority +1 PMC votes are cast, with a minimum of 3 +1 votes.
> >
> > [ ] +1 Release this package as Apache Spark 3.3.0
> > [ ] -1 Do not release this package because ...
> >
> > To learn more about Apache Spark, please see http://spark.apache.org/
> > <http://spark.apache.org/>
> >
> > The tag to be voted on is v3.3.0-rc2 (commit
> > c8c657b922ac8fd8dcf9553113e11a80079db059):
> > https://github.com/apache/spark/tree/v3.3.0-rc2
> > <https://github.com/apache/spark/tree/v3.3.0-rc2>
> >
> > The release files, including signatures, digests, etc. can be found at:
> > https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-bin/
> > <https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-bin/>
> >
> > Signatures used for Spark RCs can be found in this file:
> > https://dist.apache.org/repos/dist/dev/spark/KEYS
> > <https://dist.apache.org/repos/dist/dev/spark/KEYS>
> >
> > The staging repository for this release can be found at:
> > https://repository.apache.org/content/repositories/orgapachespark-1403
> > <https://repository.apache.org/content/repositories/orgapachespark-1403>
> >
> > The documentation corresponding to this release can be found at:
> > https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-docs/
> > <https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-docs/>
> >
> > The list of bug fixes going into 3.3.0 can be found at the following URL:
> > https://issues.apache.org/jira/projects/SPARK/versions/12350369
> > <https://issues.apache.org/jira/projects/SPARK/versions/12350369>
> >
> > This release is using the release script of the tag v3.3.0-rc2.
> >
> >
> > FAQ
> >
> > =========================
> > How can I help test this release?
> > =========================
> > If you are a Spark user, you can help us test this release by taking
> > an existing Spark workload and running on this release candidate, then
> > reporting any regressions.
> >
> > If you're working in PySpark you can set up a virtual env and install
> > the current RC and see if anything important breaks, in the Java/Scala
> > you can add the staging repository to your projects resolvers and test
> > with the RC (make sure to clean up the artifact cache before/after so
> > you don't end up building with a out of date RC going forward).
> >
> > ===========================================
> > What should happen to JIRA tickets still targeting 3.3.0?
> > ===========================================
> > The current list of open tickets targeted at 3.3.0 can be found at:
> > https://issues.apache.org/jira/projects/SPARK
> > <https://issues.apache.org/jira/projects/SPARK> and search for "Target
> > Version/s" = 3.3.0
> >
> > Committers should look at those and triage. Extremely important bug
> > fixes, documentation, and API tweaks that impact compatibility should
> > be worked on immediately. Everything else please retarget to an
> > appropriate release.
> >
> > ==================
> > But my bug isn't fixed?
> > ==================
> > In order to make timely releases, we will typically not hold the
> > release unless the bug in question is a regression from the previous
> > release. That being said, if there is something which is a regression
> > that has not been correctly targeted please ping me or a committer to
> > help target the issue.
> >
> > Maxim Gekk
> >
> > Software Engineer
> >
> > Databricks, Inc.
> >
>
> ---------------------------------------------------------------------
> To unsubscribe e-mail: dev-unsubscribe@spark.apache.org
>
>

Re: [VOTE] Release Spark 3.3.0 (RC2)

Posted by Emil Ejbyfeldt <ee...@liveintent.com.INVALID>.
Hi,

When testing out Spark 3.3.0 on our production spark workload it was 
noticed that https://issues.apache.org/jira/browse/SPARK-38681 is 
actually a regression from 3.2 (I did not know this a the time of 
creating the ticket) seem like the bug was introduced in 
https://github.com/apache/spark/pull/33205

I already have a PR here that fixes the issue:
https://github.com/apache/spark/pull/36004

Since this is a breaking regression for us I think it might be for other 
people as well.

Best,
Emil

On 16/05/2022 14:43, Maxim Gekk wrote:
> Please vote on releasing the following candidate as 
> Apache Spark version 3.3.0.
> 
> The vote is open until 11:59pm Pacific time May 19th and passes if a 
> majority +1 PMC votes are cast, with a minimum of 3 +1 votes.
> 
> [ ] +1 Release this package as Apache Spark 3.3.0
> [ ] -1 Do not release this package because ...
> 
> To learn more about Apache Spark, please see http://spark.apache.org/ 
> <http://spark.apache.org/>
> 
> The tag to be voted on is v3.3.0-rc2 (commit 
> c8c657b922ac8fd8dcf9553113e11a80079db059):
> https://github.com/apache/spark/tree/v3.3.0-rc2 
> <https://github.com/apache/spark/tree/v3.3.0-rc2>
> 
> The release files, including signatures, digests, etc. can be found at:
> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-bin/ 
> <https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-bin/>
> 
> Signatures used for Spark RCs can be found in this file:
> https://dist.apache.org/repos/dist/dev/spark/KEYS 
> <https://dist.apache.org/repos/dist/dev/spark/KEYS>
> 
> The staging repository for this release can be found at:
> https://repository.apache.org/content/repositories/orgapachespark-1403 
> <https://repository.apache.org/content/repositories/orgapachespark-1403>
> 
> The documentation corresponding to this release can be found at:
> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-docs/ 
> <https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc2-docs/>
> 
> The list of bug fixes going into 3.3.0 can be found at the following URL:
> https://issues.apache.org/jira/projects/SPARK/versions/12350369 
> <https://issues.apache.org/jira/projects/SPARK/versions/12350369>
> 
> This release is using the release script of the tag v3.3.0-rc2.
> 
> 
> FAQ
> 
> =========================
> How can I help test this release?
> =========================
> If you are a Spark user, you can help us test this release by taking
> an existing Spark workload and running on this release candidate, then
> reporting any regressions.
> 
> If you're working in PySpark you can set up a virtual env and install
> the current RC and see if anything important breaks, in the Java/Scala
> you can add the staging repository to your projects resolvers and test
> with the RC (make sure to clean up the artifact cache before/after so
> you don't end up building with a out of date RC going forward).
> 
> ===========================================
> What should happen to JIRA tickets still targeting 3.3.0?
> ===========================================
> The current list of open tickets targeted at 3.3.0 can be found at:
> https://issues.apache.org/jira/projects/SPARK 
> <https://issues.apache.org/jira/projects/SPARK> and search for "Target 
> Version/s" = 3.3.0
> 
> Committers should look at those and triage. Extremely important bug
> fixes, documentation, and API tweaks that impact compatibility should
> be worked on immediately. Everything else please retarget to an
> appropriate release.
> 
> ==================
> But my bug isn't fixed?
> ==================
> In order to make timely releases, we will typically not hold the
> release unless the bug in question is a regression from the previous
> release. That being said, if there is something which is a regression
> that has not been correctly targeted please ping me or a committer to
> help target the issue.
> 
> Maxim Gekk
> 
> Software Engineer
> 
> Databricks, Inc.
> 

---------------------------------------------------------------------
To unsubscribe e-mail: dev-unsubscribe@spark.apache.org