You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Stefaan Lippens (Jira)" <ji...@apache.org> on 2023/01/11 15:54:00 UTC

[jira] [Created] (SPARK-41989) PYARROW_IGNORE_TIMEZONE warning can break application logging setup

Stefaan Lippens created SPARK-41989:
---------------------------------------

             Summary: PYARROW_IGNORE_TIMEZONE warning can break application logging setup
                 Key: SPARK-41989
                 URL: https://issues.apache.org/jira/browse/SPARK-41989
             Project: Spark
          Issue Type: Bug
          Components: PySpark
    Affects Versions: 3.2.3
         Environment: python 3.9 env with pyspark installed
            Reporter: Stefaan Lippens


in {code}python/pyspark/pandas/__init__.py{code}  there is currently a warning when {{PYARROW_IGNORE_TIMEZONE}} env var is not set (https://github.com/apache/spark/blob/187c4a9c66758e973633c5c309b551b1d9094e6e/python/pyspark/pandas/__init__.py#L44-L59):

{code:python}
    import logging

    logging.warning(
        "'PYARROW_IGNORE_TIMEZONE' environment variable was not set. It is required to "...
{code}

The {{logging.warning()}} call  will silently do a {{logging.basicConfig()}} call (at least in python 3.9, which I tried). Something like {{logging.getLogger(...).warning()}} would not do this FYI.


This has a a very hard to figure out side-effect that importing `pyspark.pandas`  might break your logging setup (if PYARROW_IGNORE_TIMEZONE is not set).

For example:

{code:python}
import logging
import pyspark.pandas

logging.basicConfig(level=logging.DEBUG)

logger = logging.getLogger("test")
logger.warning("I warn you")
logger.debug("I debug you")
{code}

Will only produce the warning, not the debug line.
By removing the {{import pyspark.pandas}}, the debug like is produced



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org