You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@airflow.apache.org by GitBox <gi...@apache.org> on 2022/08/16 12:58:08 UTC

[GitHub] [airflow] potiuk commented on issue #25721: Status of testing Providers that were prepared on August 15, 2022

potiuk commented on issue #25721:
URL: https://github.com/apache/airflow/issues/25721#issuecomment-1216603447

   Tested: 
   
   #25299:
   
   ```
   >>> from airflow.providers.apache.hive.hooks.hive import HiveServer2Hook
   >>> hook = HiveServer2Hook()
   >>> hook.get_pandas_df(hql='xxx')
   Traceback (most recent call last):
     File "<stdin>", line 1, in <module>
   TypeError: get_pandas_df() missing 1 required positional argument: 'sql'
   ```
   #25293
   
   ```
   root@a49eb20ef97a:/opt/airflow# airflow providers list
   package_name                        | description                                                                 | version
   ====================================+=============================================================================+========
   apache-airflow-providers-common-sql | Common SQL Provider https://en.wikipedia.org/wiki/SQL                       | 1.0.0  
   apache-airflow-providers-ftp        | File Transfer Protocol (FTP) https://tools.ietf.org/html/rfc114             | 3.0.0  
   apache-airflow-providers-http       | Hypertext Transfer Protocol (HTTP) https://www.w3.org/Protocols/            | 3.0.0  
   apache-airflow-providers-imap       | Internet Message Access Protocol (IMAP) https://tools.ietf.org/html/rfc3501 | 3.0.0  
   apache-airflow-providers-postgres   | PostgreSQL https://www.postgresql.org/                                      | 5.2.0  
   apache-airflow-providers-sqlite     | SQLite https://www.sqlite.org/                                              | 3.0.0  
                                                                                                                              
   root@a49eb20ef97a:/opt/airflow# python
   Python 3.7.13 (default, Aug  2 2022, 12:15:43) 
   [GCC 10.2.1 20210110] on linux
   Type "help", "copyright", "credits" or "license" for more information.
   >>> from airflow.sensors.sql import SqlSensor
   >>> sensor = SqlSensor(conn_id='postgres_default', sql='SELECT * FROM DUAL', task_id='id')
   >>> sensor._get_hook()
   [2022-08-16 12:44:15,582] {base.py:68} INFO - Using connection ID 'postgres_default' for task execution.
   Traceback (most recent call last):
     File "<stdin>", line 1, in <module>
     File "/usr/local/lib/python3.7/site-packages/airflow/sensors/sql.py", line 84, in _get_hook
       f'The connection type is not supported by {self.__class__.__name__}. '
   airflow.exceptions.AirflowException: The connection type is not supported by SqlSensor. The associated hook should be a subclass of `DbApiHook`. Got PostgresHook
   >>> quit()
   root@a49eb20ef97a:/opt/airflow# pip install apache-airflow-providers-common-sql==1.1.0rc4
   Collecting apache-airflow-providers-common-sql==1.1.0rc4
     Downloading apache_airflow_providers_common_sql-1.1.0rc4-py3-none-any.whl (27 kB)
   Requirement already satisfied: sqlparse>=0.4.2 in /usr/local/lib/python3.7/site-packages (from apache-airflow-providers-common-sql==1.1.0rc4) (0.4.2)
   Installing collected packages: apache-airflow-providers-common-sql
     Attempting uninstall: apache-airflow-providers-common-sql
       Found existing installation: apache-airflow-providers-common-sql 1.0.0
       Uninstalling apache-airflow-providers-common-sql-1.0.0:
         Successfully uninstalled apache-airflow-providers-common-sql-1.0.0
   Successfully installed apache-airflow-providers-common-sql-1.1.0rc4
   WARNING: Running pip as the 'root' user can result in broken permissions and conflicting behaviour with the system package manager. It is recommended to use a virtual environment instead: https://pip.pypa.io/warnings/venv
   root@a49eb20ef97a:/opt/airflow# python
   Python 3.7.13 (default, Aug  2 2022, 12:15:43) 
   [GCC 10.2.1 20210110] on linux
   Type "help", "copyright", "credits" or "license" for more information.
   >>> from airflow.sensors.sql import SqlSensor
   >>> sensor = SqlSensor(conn_id='postgres_default', sql='SELECT * FROM DUAL', task_id='id')
   >>> sensor._get_hook()
   [2022-08-16 12:45:24,882] {base.py:68} INFO - Using connection ID 'postgres_default' for task execution.
   <airflow.providers.postgres.hooks.postgres.PostgresHook object at 0x7f3ddfa16750>
   >>> 
   ```
   #25350 
   ```
   [GCC 10.2.1 20210110] on linux
   Type "help", "copyright", "credits" or "license" for more information.
   >>> from airflow.providers.common.sql.operators.sql import (
       SQLCheckOperator,
       SQLIntervalCheckOperator,
       SQLValueCheckOperator,
   )
   >>> 
   ```
   
   #25713 
   
   common-sql: 1.1.0rc4:
   ```
   >>> from airflow.providers.databricks.hooks.databricks_sql import DatabricksSqlHook
   >>> hook = DatabricksSqlHook()
   >>> hook.run(sql='SELECT * FROM TABLE')
   [2022-08-16 12:51:21,954] {base.py:68} INFO - Using connection ID 'databricks_default' for task execution.
   Traceback (most recent call last):
     File "<stdin>", line 1, in <module>
     File "/usr/local/lib/python3.7/site-packages/airflow/providers/databricks/hooks/databricks_sql.py", line 179, in run
       with closing(self.get_conn()) as conn:
     File "/usr/local/lib/python3.7/site-packages/airflow/providers/databricks/hooks/databricks_sql.py", line 105, in get_conn
       "http_path should be provided either explicitly, "
   airflow.exceptions.AirflowException: http_path should be provided either explicitly, or in extra parameter of Databricks connection, or sql_endpoint_name should be specified
   >>> import logging
   >>> hook.log.setLevel(logging.DEBUG)
   >>> hook.run(sql='SELECT * FROM TABLE')
   [2022-08-16 12:53:00,963] {databricks_sql.py:172} DEBUG - Executing following statements against Databricks DB: ['SELECT * FROM TABLE']
   ```
   
   common-sql: 1.1.0rc3:
   
   ```
   root@a49eb20ef97a:/opt/airflow# pip install apache-airflow-providers-common-sql==1.1.0rc3
   Collecting apache-airflow-providers-common-sql==1.1.0rc3
     Downloading apache_airflow_providers_common_sql-1.1.0rc3-py3-none-any.whl (27 kB)
   Requirement already satisfied: sqlparse>=0.4.2 in /usr/local/lib/python3.7/site-packages (from apache-airflow-providers-common-sql==1.1.0rc3) (0.4.2)
   Installing collected packages: apache-airflow-providers-common-sql
     Attempting uninstall: apache-airflow-providers-common-sql
       Found existing installation: apache-airflow-providers-common-sql 1.0.0
       Uninstalling apache-airflow-providers-common-sql-1.0.0:
         Successfully uninstalled apache-airflow-providers-common-sql-1.0.0
   Successfully installed apache-airflow-providers-common-sql-1.1.0rc3
   WARNING: Running pip as the 'root' user can result in broken permissions and conflicting behaviour with the system package manager. It is recommended to use a virtual environment instead: https://pip.pypa.io/warnings/venv
   root@a49eb20ef97a:/opt/airflow# python
   Python 3.7.13 (default, Aug  2 2022, 12:15:43) 
   [GCC 10.2.1 20210110] on linux
   Type "help", "copyright", "credits" or "license" for more information.
   >>> from airflow.providers.databricks.hooks.databricks_sql import DatabricksSqlHook
   >>> hook = DatabricksSqlHook()
   >>> import logging
   >>> hook.log.setLevel(logging.DEBUG)
   >>> hook.run(sql='SELECT * FROM TABLE')
   Traceback (most recent call last):
     File "<stdin>", line 1, in <module>
     File "/usr/local/lib/python3.7/site-packages/airflow/providers/databricks/hooks/databricks_sql.py", line 174, in run
       raise ValueError("List of SQL statements is empty")
   ValueError: List of SQL statements is empty
   ```
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: commits-unsubscribe@airflow.apache.org

For queries about this service, please contact Infrastructure at:
users@infra.apache.org