You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Franck Thang (Jira)" <ji...@apache.org> on 2022/01/17 22:48:00 UTC
[jira] [Comment Edited] (SPARK-37857) Any depth search not working in get_json_object ($..foo)
[ https://issues.apache.org/jira/browse/SPARK-37857?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17477433#comment-17477433 ]
Franck Thang edited comment on SPARK-37857 at 1/17/22, 10:47 PM:
-----------------------------------------------------------------
As [~venk] is not answering, I can start taking a look if you want [~hyukjin.kwon]
Post-update:
It seems the first implementation of Spark get_json_object does replicate Hive behaviour (https://github.com/apache/spark/pull/7901)
In hive documentation ([https://cwiki.apache.org/confluence/display/hive/languagemanual+udf)] it is stated that recursive decent (i.e ".." notation) is not supported
I guess that's probably why it's not working
What do you think ?
was (Author: stelyus):
As [~venk] is not answering, I can start taking a look if you want [~hyukjin.kwon]
> Any depth search not working in get_json_object ($..foo)
> --------------------------------------------------------
>
> Key: SPARK-37857
> URL: https://issues.apache.org/jira/browse/SPARK-37857
> Project: Spark
> Issue Type: Bug
> Components: SQL
> Affects Versions: 3.2.0
> Reporter: Venk
> Priority: Minor
>
> The following example should return a value _abc_ but instead returns null
> {code:java}
> spark.sql("""select get_json_object('{"k":{"value":"abc"}}', '$..value') as j""").show(){code}
> I checked the [spark test suite for Json expression parser | [spark/JsonExpressionsSuite.scala at f12793de205a46c75529ac20649e2e61edc0b144 · apache/spark · GitHub|https://github.com/apache/spark/blob/f12793de205a46c75529ac20649e2e61edc0b144/sql/catalyst/src/test/scala/org/apache/spark/sql/catalyst/expressions/JsonExpressionsSuite.scala]] and the example there seems to be for a negative test case (non-existent key), so I'm not sure if this is expected.
--
This message was sent by Atlassian Jira
(v8.20.1#820001)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org