You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Shivu Sondur (Jira)" <ji...@apache.org> on 2019/08/26 08:31:00 UTC
[jira] [Commented] (SPARK-28873) [UDF]show functions behaves
different in hive and spark
[ https://issues.apache.org/jira/browse/SPARK-28873?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16915593#comment-16915593 ]
Shivu Sondur commented on SPARK-28873:
--------------------------------------
i will check this issue
> [UDF]show functions behaves different in hive and spark
> -------------------------------------------------------
>
> Key: SPARK-28873
> URL: https://issues.apache.org/jira/browse/SPARK-28873
> Project: Spark
> Issue Type: Bug
> Components: SQL
> Affects Versions: 2.4.3
> Reporter: ABHISHEK KUMAR GUPTA
> Priority: Major
>
> Description:
> Launch spark-beeline
> show functions;
> Result will list all the system level functions and permanent UDF created inside deafult db.
> jdbc:hive2://10.18.18.214:23040/default> create function func.mul_651 AS 'com.huawei.bigdata.hive.example.udf.multiply' using jar 'hdfs://hacluster/user/Multiply.jar';
> create function default.mul_651 AS 'com.huawei.bigdata.hive.example.udf.multiply' using jar 'hdfs://hacluster/user/Multiply.jar';
> show functions;
> Will list give total count of functions created inside default DB only.
> In Hive
> jdbc:hive2://10.18.98.147:21066/> create function func.add_5 AS 'com.huawei.bigdata.hive.example.udf.AddDoublesUDF' using jar 'hdfs://hacluster/user/AddDoublesUDF.jar';
> jdbc:hive2://10.18.98.147:21066/> create function default.add_5 AS 'com.huawei.bigdata.hive.example.udf.AddDoublesUDF' using jar 'hdfs://hacluster/user/AddDoublesUDF.jar';
> Show functions; will list all the UDF functions created inside the Database func as well as default.
> Expected: show functions; should list all the functions of user created as well as systems functions.
--
This message was sent by Atlassian Jira
(v8.3.2#803003)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org