You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@systemml.apache.org by GitBox <gi...@apache.org> on 2019/03/21 20:49:21 UTC

[GitHub] [systemml] niketanpansare edited a comment on issue #857: [SYSTEMML-2523] Update SystemML to Support Spark 2.3.0

niketanpansare edited a comment on issue #857: [SYSTEMML-2523] Update SystemML to Support Spark 2.3.0
URL: https://github.com/apache/systemml/pull/857#issuecomment-475394794
 
 
   @romeokienzler You are getting the error because the setup contains two SystemML (possibly conflicting dependencies) jars. There are two possible solutions to your problem:
   1. *Recommended:* Remove the older incubating jar and do not include the corresponding 1.2.0 or 1.3.0-snapshot jars (i.e. no need for `ln -s` trick).
   2. Use the python package compiled by this PR.
   
   Since there is something weird happening here, I am including the logs. I apologize it advance for the long trace. Please ignore the below logs if you agree to the above statements.
   
   Setup 1. With only incubating jar (FAILS !!)
   ```
   $ ~/spark-2.3.0-bin-hadoop2.7/bin/pyspark --driver-memory 20g --master local[*] --driver-class-path systemml-0.14.0-incubating.jar
   Python 3.6.3 (default, Mar 20 2018, 13:50:41) 
   [GCC 4.8.5 20150623 (Red Hat 4.8.5-16)] on linux
   Type "help", "copyright", "credits" or "license" for more information.
   2019-03-21 13:07:11 WARN  NativeCodeLoader:62 - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
   Setting default log level to "WARN".
   To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel).
   Welcome to
         ____              __
        / __/__  ___ _____/ /__
       _\ \/ _ \/ _ `/ __/  '_/
      /__ / .__/\_,_/_/ /_/\_\   version 2.3.0
         /_/
   
   Using Python version 3.6.3 (default, Mar 20 2018 13:50:41)
   SparkSession available as 'spark'.
   >>> from systemml import MLContext
   >>> ml = MLContext(spark)
   2019-03-21 13:07:20 WARN  ObjectStore:568 - Failed to get database global_temp, returning NoSuchObjectException
   
   Welcome to Apache SystemML!
   
   >>> ml.version()
   '0.14.0-incubating'
   >>> df=spark.read.parquet('shake.parquet')
   >>> df.show()
   +-----+---------+-----+-----+-----+
   |CLASS| SENSORID|    X|    Y|    Z|
   +-----+---------+-----+-----+-----+
   |    2| qqqqqqqq| 0.12| 0.12| 0.12|
   |    2|aUniqueID| 0.03| 0.03| 0.03|
   |    2| qqqqqqqq|-3.84|-3.84|-3.84|
   |    2| 12345678| -0.1| -0.1| -0.1|
   |    2| 12345678|-0.15|-0.15|-0.15|
   |    2| 12345678| 0.47| 0.47| 0.47|
   |    2| 12345678|-0.06|-0.06|-0.06|
   |    2| 12345678|-0.09|-0.09|-0.09|
   |    2| 12345678| 0.21| 0.21| 0.21|
   |    2| 12345678|-0.08|-0.08|-0.08|
   |    2| 12345678| 0.44| 0.44| 0.44|
   |    2|    gholi| 0.76| 0.76| 0.76|
   |    2|    gholi| 1.62| 1.62| 1.62|
   |    2|    gholi| 5.81| 5.81| 5.81|
   |    2| bcbcbcbc| 0.58| 0.58| 0.58|
   |    2| bcbcbcbc|-8.24|-8.24|-8.24|
   |    2| bcbcbcbc|-0.45|-0.45|-0.45|
   |    2| bcbcbcbc| 1.03| 1.03| 1.03|
   |    2|aUniqueID|-0.05|-0.05|-0.05|
   |    2| qqqqqqqq|-0.44|-0.44|-0.44|
   +-----+---------+-----+-----+-----+
   only showing top 20 rows
   
   >>> df.createOrReplaceTempView("df")
   ANTLR Tool version 4.7 used for code generation does not match the current runtime version 4.5.3ANTLR Runtime version 4.7 used for parser compilation does not match the current runtime version 4.5.3Traceback (most recent call last):
     File "<stdin>", line 1, in <module>
     File "/home/npansar/spark-2.3.0-bin-hadoop2.7/python/pyspark/sql/dataframe.py", line 176, in createOrReplaceTempView
       self._jdf.createOrReplaceTempView(name)
     File "/home/npansar/spark-2.3.0-bin-hadoop2.7/python/lib/py4j-0.10.6-src.zip/py4j/java_gateway.py", line 1160, in __call__
     File "/home/npansar/spark-2.3.0-bin-hadoop2.7/python/pyspark/sql/utils.py", line 63, in deco
       return f(*a, **kw)
     File "/home/npansar/spark-2.3.0-bin-hadoop2.7/python/lib/py4j-0.10.6-src.zip/py4j/protocol.py", line 320, in get_return_value
   py4j.protocol.Py4JJavaError: An error occurred while calling o52.createOrReplaceTempView.
   : java.lang.ExceptionInInitializerError
   	at org.apache.spark.sql.catalyst.parser.AbstractSqlParser.parse(ParseDriver.scala:84)
   	at org.apache.spark.sql.execution.SparkSqlParser.parse(SparkSqlParser.scala:48)
   	at org.apache.spark.sql.catalyst.parser.AbstractSqlParser.parseTableIdentifier(ParseDriver.scala:49)
   	at org.apache.spark.sql.Dataset.createTempViewCommand(Dataset.scala:3079)
   	at org.apache.spark.sql.Dataset.createOrReplaceTempView(Dataset.scala:3034)
   	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
   	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   	at java.lang.reflect.Method.invoke(Method.java:498)
   	at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
   	at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
   	at py4j.Gateway.invoke(Gateway.java:282)
   	at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
   	at py4j.commands.CallCommand.execute(CallCommand.java:79)
   	at py4j.GatewayConnection.run(GatewayConnection.java:214)
   	at java.lang.Thread.run(Thread.java:748)
   Caused by: java.lang.UnsupportedOperationException: java.io.InvalidClassException: org.antlr.v4.runtime.atn.ATN; Could not deserialize ATN with UUID 59627784-3be5-417a-b9eb-8131a7286089 (expected aadb8d7e-aeef-4415-ad2b-8204d6cf042e or a legacy UUID).
   	at org.antlr.v4.runtime.atn.ATNDeserializer.deserialize(ATNDeserializer.java:153)
   	at org.apache.spark.sql.catalyst.parser.SqlBaseLexer.<clinit>(SqlBaseLexer.java:1153)
   	... 16 more
   Caused by: java.io.InvalidClassException: org.antlr.v4.runtime.atn.ATN; Could not deserialize ATN with UUID 59627784-3be5-417a-b9eb-8131a7286089 (expected aadb8d7e-aeef-4415-ad2b-8204d6cf042e or a legacy UUID).
   	... 18 more
   
   >>>
   ```
   
   Setup 2: Put the older incubating jar before the current SystemML 1.2.0 jars (FAILS !!)
   ```
   $ ~/spark-2.3.0-bin-hadoop2.7/bin/pyspark --driver-memory 20g --master local[*] --driver-class-path systemml-0.14.0-incubating.jar:systemml-1.2.0-extra.jar:systemml-1.2.0.jar
   Python 3.6.3 (default, Mar 20 2018, 13:50:41) 
   [GCC 4.8.5 20150623 (Red Hat 4.8.5-16)] on linux
   Type "help", "copyright", "credits" or "license" for more information.
   2019-03-21 13:12:11 WARN  NativeCodeLoader:62 - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
   Setting default log level to "WARN".
   To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel).
   Welcome to
         ____              __
        / __/__  ___ _____/ /__
       _\ \/ _ \/ _ `/ __/  '_/
      /__ / .__/\_,_/_/ /_/\_\   version 2.3.0
         /_/
   
   Using Python version 3.6.3 (default, Mar 20 2018 13:50:41)
   SparkSession available as 'spark'.
   >>> from systemml import MLContext
   >>> ml = MLContext(spark)
   2019-03-21 13:12:21 WARN  ObjectStore:568 - Failed to get database global_temp, returning NoSuchObjectException
   
   Welcome to Apache SystemML!
   
   >>> ml.version()
   '0.14.0-incubating'
   >>> df=spark.read.parquet('shake.parquet')
   >>> df.show()
   +-----+---------+-----+-----+-----+
   |CLASS| SENSORID|    X|    Y|    Z|
   +-----+---------+-----+-----+-----+
   |    2| qqqqqqqq| 0.12| 0.12| 0.12|
   |    2|aUniqueID| 0.03| 0.03| 0.03|
   |    2| qqqqqqqq|-3.84|-3.84|-3.84|
   |    2| 12345678| -0.1| -0.1| -0.1|
   |    2| 12345678|-0.15|-0.15|-0.15|
   |    2| 12345678| 0.47| 0.47| 0.47|
   |    2| 12345678|-0.06|-0.06|-0.06|
   |    2| 12345678|-0.09|-0.09|-0.09|
   |    2| 12345678| 0.21| 0.21| 0.21|
   |    2| 12345678|-0.08|-0.08|-0.08|
   |    2| 12345678| 0.44| 0.44| 0.44|
   |    2|    gholi| 0.76| 0.76| 0.76|
   |    2|    gholi| 1.62| 1.62| 1.62|
   |    2|    gholi| 5.81| 5.81| 5.81|
   |    2| bcbcbcbc| 0.58| 0.58| 0.58|
   |    2| bcbcbcbc|-8.24|-8.24|-8.24|
   |    2| bcbcbcbc|-0.45|-0.45|-0.45|
   |    2| bcbcbcbc| 1.03| 1.03| 1.03|
   |    2|aUniqueID|-0.05|-0.05|-0.05|
   |    2| qqqqqqqq|-0.44|-0.44|-0.44|
   +-----+---------+-----+-----+-----+
   only showing top 20 rows
   
   >>> df.createOrReplaceTempView("df")
   ANTLR Tool version 4.7 used for code generation does not match the current runtime version 4.5.3ANTLR Runtime version 4.7 used for parser compilation does not match the current runtime version 4.5.3Traceback (most recent call last):
     File "<stdin>", line 1, in <module>
     File "/home/npansar/spark-2.3.0-bin-hadoop2.7/python/pyspark/sql/dataframe.py", line 176, in createOrReplaceTempView
       self._jdf.createOrReplaceTempView(name)
     File "/home/npansar/spark-2.3.0-bin-hadoop2.7/python/lib/py4j-0.10.6-src.zip/py4j/java_gateway.py", line 1160, in __call__
     File "/home/npansar/spark-2.3.0-bin-hadoop2.7/python/pyspark/sql/utils.py", line 63, in deco
       return f(*a, **kw)
     File "/home/npansar/spark-2.3.0-bin-hadoop2.7/python/lib/py4j-0.10.6-src.zip/py4j/protocol.py", line 320, in get_return_value
   py4j.protocol.Py4JJavaError: An error occurred while calling o52.createOrReplaceTempView.
   : java.lang.ExceptionInInitializerError
   	at org.apache.spark.sql.catalyst.parser.AbstractSqlParser.parse(ParseDriver.scala:84)
   	at org.apache.spark.sql.execution.SparkSqlParser.parse(SparkSqlParser.scala:48)
   	at org.apache.spark.sql.catalyst.parser.AbstractSqlParser.parseTableIdentifier(ParseDriver.scala:49)
   	at org.apache.spark.sql.Dataset.createTempViewCommand(Dataset.scala:3079)
   	at org.apache.spark.sql.Dataset.createOrReplaceTempView(Dataset.scala:3034)
   	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
   	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   	at java.lang.reflect.Method.invoke(Method.java:498)
   	at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
   	at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
   	at py4j.Gateway.invoke(Gateway.java:282)
   	at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
   	at py4j.commands.CallCommand.execute(CallCommand.java:79)
   	at py4j.GatewayConnection.run(GatewayConnection.java:214)
   	at java.lang.Thread.run(Thread.java:748)
   Caused by: java.lang.UnsupportedOperationException: java.io.InvalidClassException: org.antlr.v4.runtime.atn.ATN; Could not deserialize ATN with UUID 59627784-3be5-417a-b9eb-8131a7286089 (expected aadb8d7e-aeef-4415-ad2b-8204d6cf042e or a legacy UUID).
   	at org.antlr.v4.runtime.atn.ATNDeserializer.deserialize(ATNDeserializer.java:153)
   	at org.apache.spark.sql.catalyst.parser.SqlBaseLexer.<clinit>(SqlBaseLexer.java:1153)
   	... 16 more
   Caused by: java.io.InvalidClassException: org.antlr.v4.runtime.atn.ATN; Could not deserialize ATN with UUID 59627784-3be5-417a-b9eb-8131a7286089 (expected aadb8d7e-aeef-4415-ad2b-8204d6cf042e or a legacy UUID).
   	... 18 more
   
   >>>
   ```
   
   Setup 3: Put the the current SystemML 1.2.0 jars before the older incubating jar (FAILS !!)
   ```
   $ ~/spark-2.3.0-bin-hadoop2.7/bin/pyspark --driver-memory 20g --master local[*] --driver-class-path systemml-1.2.0-extra.jar:systemml-1.2.0.jar:systemml-0.14.0-incubating.jar
   Python 3.6.3 (default, Mar 20 2018, 13:50:41) 
   [GCC 4.8.5 20150623 (Red Hat 4.8.5-16)] on linux
   Type "help", "copyright", "credits" or "license" for more information.
   2019-03-21 13:14:49 WARN  NativeCodeLoader:62 - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
   Setting default log level to "WARN".
   To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel).
   Welcome to
         ____              __
        / __/__  ___ _____/ /__
       _\ \/ _ \/ _ `/ __/  '_/
      /__ / .__/\_,_/_/ /_/\_\   version 2.3.0
         /_/
   
   Using Python version 3.6.3 (default, Mar 20 2018 13:50:41)
   SparkSession available as 'spark'.
   >>> from systemml import MLContext
   >>> ml = MLContext(spark)
   2019-03-21 13:15:11 WARN  ObjectStore:568 - Failed to get database global_temp, returning NoSuchObjectException
   
   Welcome to Apache SystemML!
   Version 1.2.0
   >>> ml.version()
   '1.2.0'
   >>> df=spark.read.parquet('shake.parquet')
   >>> df.show()
   +-----+---------+-----+-----+-----+
   |CLASS| SENSORID|    X|    Y|    Z|
   +-----+---------+-----+-----+-----+
   |    2| qqqqqqqq| 0.12| 0.12| 0.12|
   |    2|aUniqueID| 0.03| 0.03| 0.03|
   |    2| qqqqqqqq|-3.84|-3.84|-3.84|
   |    2| 12345678| -0.1| -0.1| -0.1|
   |    2| 12345678|-0.15|-0.15|-0.15|
   |    2| 12345678| 0.47| 0.47| 0.47|
   |    2| 12345678|-0.06|-0.06|-0.06|
   |    2| 12345678|-0.09|-0.09|-0.09|
   |    2| 12345678| 0.21| 0.21| 0.21|
   |    2| 12345678|-0.08|-0.08|-0.08|
   |    2| 12345678| 0.44| 0.44| 0.44|
   |    2|    gholi| 0.76| 0.76| 0.76|
   |    2|    gholi| 1.62| 1.62| 1.62|
   |    2|    gholi| 5.81| 5.81| 5.81|
   |    2| bcbcbcbc| 0.58| 0.58| 0.58|
   |    2| bcbcbcbc|-8.24|-8.24|-8.24|
   |    2| bcbcbcbc|-0.45|-0.45|-0.45|
   |    2| bcbcbcbc| 1.03| 1.03| 1.03|
   |    2|aUniqueID|-0.05|-0.05|-0.05|
   |    2| qqqqqqqq|-0.44|-0.44|-0.44|
   +-----+---------+-----+-----+-----+
   only showing top 20 rows
   
   >>> df.createOrReplaceTempView("df")
   ANTLR Tool version 4.7 used for code generation does not match the current runtime version 4.5.3ANTLR Runtime version 4.7 used for parser compilation does not match the current runtime version 4.5.3Traceback (most recent call last):
     File "<stdin>", line 1, in <module>
     File "/home/npansar/spark-2.3.0-bin-hadoop2.7/python/pyspark/sql/dataframe.py", line 176, in createOrReplaceTempView
       self._jdf.createOrReplaceTempView(name)
     File "/home/npansar/spark-2.3.0-bin-hadoop2.7/python/lib/py4j-0.10.6-src.zip/py4j/java_gateway.py", line 1160, in __call__
     File "/home/npansar/spark-2.3.0-bin-hadoop2.7/python/pyspark/sql/utils.py", line 63, in deco
       return f(*a, **kw)
     File "/home/npansar/spark-2.3.0-bin-hadoop2.7/python/lib/py4j-0.10.6-src.zip/py4j/protocol.py", line 320, in get_return_value
   py4j.protocol.Py4JJavaError: An error occurred while calling o52.createOrReplaceTempView.
   : java.lang.ExceptionInInitializerError
   	at org.apache.spark.sql.catalyst.parser.AbstractSqlParser.parse(ParseDriver.scala:84)
   	at org.apache.spark.sql.execution.SparkSqlParser.parse(SparkSqlParser.scala:48)
   	at org.apache.spark.sql.catalyst.parser.AbstractSqlParser.parseTableIdentifier(ParseDriver.scala:49)
   	at org.apache.spark.sql.Dataset.createTempViewCommand(Dataset.scala:3079)
   	at org.apache.spark.sql.Dataset.createOrReplaceTempView(Dataset.scala:3034)
   	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
   	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   	at java.lang.reflect.Method.invoke(Method.java:498)
   	at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
   	at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
   	at py4j.Gateway.invoke(Gateway.java:282)
   	at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
   	at py4j.commands.CallCommand.execute(CallCommand.java:79)
   	at py4j.GatewayConnection.run(GatewayConnection.java:214)
   	at java.lang.Thread.run(Thread.java:748)
   Caused by: java.lang.UnsupportedOperationException: java.io.InvalidClassException: org.antlr.v4.runtime.atn.ATN; Could not deserialize ATN with UUID 59627784-3be5-417a-b9eb-8131a7286089 (expected aadb8d7e-aeef-4415-ad2b-8204d6cf042e or a legacy UUID).
   	at org.antlr.v4.runtime.atn.ATNDeserializer.deserialize(ATNDeserializer.java:153)
   	at org.apache.spark.sql.catalyst.parser.SqlBaseLexer.<clinit>(SqlBaseLexer.java:1153)
   	... 16 more
   Caused by: java.io.InvalidClassException: org.antlr.v4.runtime.atn.ATN; Could not deserialize ATN with UUID 59627784-3be5-417a-b9eb-8131a7286089 (expected aadb8d7e-aeef-4415-ad2b-8204d6cf042e or a legacy UUID).
   	... 18 more
   
   >>>
   ```
   
   Setup 4: Put the jar from the PR before the older incubating jar (SUCCEEDS !!)
   ```
   $ ~/spark-2.3.0-bin-hadoop2.7/bin/pyspark --driver-memory 20g --master local[*] --driver-class-path systemml-1.3.0-SNAPSHOT-extra-pr.jar:systemml-1.3.0-SNAPSHOT-pr.jar:systemml-0.14.0-incubating.jar
   Python 3.6.3 (default, Mar 20 2018, 13:50:41) 
   [GCC 4.8.5 20150623 (Red Hat 4.8.5-16)] on linux
   Type "help", "copyright", "credits" or "license" for more information.
   2019-03-21 13:19:59 WARN  NativeCodeLoader:62 - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
   Setting default log level to "WARN".
   To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel).
   Welcome to
         ____              __
        / __/__  ___ _____/ /__
       _\ \/ _ \/ _ `/ __/  '_/
      /__ / .__/\_,_/_/ /_/\_\   version 2.3.0
         /_/
   
   Using Python version 3.6.3 (default, Mar 20 2018 13:50:41)
   SparkSession available as 'spark'.
   >>> from systemml import MLContext
   >>> ml = MLContext(spark)
   2019-03-21 13:20:22 WARN  ObjectStore:568 - Failed to get database global_temp, returning NoSuchObjectException
   
   Welcome to Apache SystemML!
   Version 1.3.0-SNAPSHOT
   >>> ml.version()
   '1.3.0-SNAPSHOT'
   >>> df=spark.read.parquet('shake.parquet')
   >>> df.show()
   +-----+---------+-----+-----+-----+
   |CLASS| SENSORID|    X|    Y|    Z|
   +-----+---------+-----+-----+-----+
   |    2| qqqqqqqq| 0.12| 0.12| 0.12|
   |    2|aUniqueID| 0.03| 0.03| 0.03|
   |    2| qqqqqqqq|-3.84|-3.84|-3.84|
   |    2| 12345678| -0.1| -0.1| -0.1|
   |    2| 12345678|-0.15|-0.15|-0.15|
   |    2| 12345678| 0.47| 0.47| 0.47|
   |    2| 12345678|-0.06|-0.06|-0.06|
   |    2| 12345678|-0.09|-0.09|-0.09|
   |    2| 12345678| 0.21| 0.21| 0.21|
   |    2| 12345678|-0.08|-0.08|-0.08|
   |    2| 12345678| 0.44| 0.44| 0.44|
   |    2|    gholi| 0.76| 0.76| 0.76|
   |    2|    gholi| 1.62| 1.62| 1.62|
   |    2|    gholi| 5.81| 5.81| 5.81|
   |    2| bcbcbcbc| 0.58| 0.58| 0.58|
   |    2| bcbcbcbc|-8.24|-8.24|-8.24|
   |    2| bcbcbcbc|-0.45|-0.45|-0.45|
   |    2| bcbcbcbc| 1.03| 1.03| 1.03|
   |    2|aUniqueID|-0.05|-0.05|-0.05|
   |    2| qqqqqqqq|-0.44|-0.44|-0.44|
   +-----+---------+-----+-----+-----+
   only showing top 20 rows
   
   >>> df.createOrReplaceTempView("df")
   >>>
   ```
   
   Setup 5: No jar provided (SUCCEEDS !!)
   
   ```
   $ ~/spark-2.3.0-bin-hadoop2.7/bin/pyspark --driver-memory 20g --master local[*]
   Python 3.6.3 (default, Mar 20 2018, 13:50:41) 
   [GCC 4.8.5 20150623 (Red Hat 4.8.5-16)] on linux
   Type "help", "copyright", "credits" or "license" for more information.
   2019-03-21 13:23:26 WARN  NativeCodeLoader:62 - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
   Setting default log level to "WARN".
   To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel).
   Welcome to
         ____              __
        / __/__  ___ _____/ /__
       _\ \/ _ \/ _ `/ __/  '_/
      /__ / .__/\_,_/_/ /_/\_\   version 2.3.0
         /_/
   
   Using Python version 3.6.3 (default, Mar 20 2018 13:50:41)
   SparkSession available as 'spark'.
   >>> from systemml import MLContext
   >>> ml = MLContext(spark)
   2019-03-21 13:23:46 WARN  ObjectStore:568 - Failed to get database global_temp, returning NoSuchObjectException
   
   Welcome to Apache SystemML!
   Version 1.2.0
   >>> ml.version()
   '1.2.0'
   >>> df=spark.read.parquet('shake.parquet')
   >>> df.show()
   +-----+---------+-----+-----+-----+
   |CLASS| SENSORID|    X|    Y|    Z|
   +-----+---------+-----+-----+-----+
   |    2| qqqqqqqq| 0.12| 0.12| 0.12|
   |    2|aUniqueID| 0.03| 0.03| 0.03|
   |    2| qqqqqqqq|-3.84|-3.84|-3.84|
   |    2| 12345678| -0.1| -0.1| -0.1|
   |    2| 12345678|-0.15|-0.15|-0.15|
   |    2| 12345678| 0.47| 0.47| 0.47|
   |    2| 12345678|-0.06|-0.06|-0.06|
   |    2| 12345678|-0.09|-0.09|-0.09|
   |    2| 12345678| 0.21| 0.21| 0.21|
   |    2| 12345678|-0.08|-0.08|-0.08|
   |    2| 12345678| 0.44| 0.44| 0.44|
   |    2|    gholi| 0.76| 0.76| 0.76|
   |    2|    gholi| 1.62| 1.62| 1.62|
   |    2|    gholi| 5.81| 5.81| 5.81|
   |    2| bcbcbcbc| 0.58| 0.58| 0.58|
   |    2| bcbcbcbc|-8.24|-8.24|-8.24|
   |    2| bcbcbcbc|-0.45|-0.45|-0.45|
   |    2| bcbcbcbc| 1.03| 1.03| 1.03|
   |    2|aUniqueID|-0.05|-0.05|-0.05|
   |    2| qqqqqqqq|-0.44|-0.44|-0.44|
   +-----+---------+-----+-----+-----+
   only showing top 20 rows
   
   >>> df.createOrReplaceTempView("df")
   >>>
   ```
   
   Setup 6: Provide just `1.2.0` jars (FAILS !!)
   ```
   $ ~/spark-2.3.0-bin-hadoop2.7/bin/pyspark --driver-memory 20g --master local[*] --driver-class-path systemml-1.2.0
   systemml-1.2.0-extra.jar  systemml-1.2.0.jar        
   [npansar@dml3 debug_classpath]$ ~/spark-2.3.0-bin-hadoop2.7/bin/pyspark --driver-memory 20g --master local[*] --driver-class-path systemml-1.2.0.jar:systemml-1.2.0-extra.jar
   Python 3.6.3 (default, Mar 20 2018, 13:50:41) 
   [GCC 4.8.5 20150623 (Red Hat 4.8.5-16)] on linux
   Type "help", "copyright", "credits" or "license" for more information.
   2019-03-21 13:32:09 WARN  NativeCodeLoader:62 - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
   Setting default log level to "WARN".
   To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel).
   Welcome to
         ____              __
        / __/__  ___ _____/ /__
       _\ \/ _ \/ _ `/ __/  '_/
      /__ / .__/\_,_/_/ /_/\_\   version 2.3.0
         /_/
   
   Using Python version 3.6.3 (default, Mar 20 2018 13:50:41)
   SparkSession available as 'spark'.
   >>> from systemml import MLContext
   >>> ml = MLContext(spark)
   2019-03-21 13:32:25 WARN  ObjectStore:568 - Failed to get database global_temp, returning NoSuchObjectException
   
   Welcome to Apache SystemML!
   Version 1.2.0
   >>> ml.version()
   '1.2.0'
   >>> df=spark.read.parquet('shake.parquet')
   >>> df.show()
   +-----+---------+-----+-----+-----+
   |CLASS| SENSORID|    X|    Y|    Z|
   +-----+---------+-----+-----+-----+
   |    2| qqqqqqqq| 0.12| 0.12| 0.12|
   |    2|aUniqueID| 0.03| 0.03| 0.03|
   |    2| qqqqqqqq|-3.84|-3.84|-3.84|
   |    2| 12345678| -0.1| -0.1| -0.1|
   |    2| 12345678|-0.15|-0.15|-0.15|
   |    2| 12345678| 0.47| 0.47| 0.47|
   |    2| 12345678|-0.06|-0.06|-0.06|
   |    2| 12345678|-0.09|-0.09|-0.09|
   |    2| 12345678| 0.21| 0.21| 0.21|
   |    2| 12345678|-0.08|-0.08|-0.08|
   |    2| 12345678| 0.44| 0.44| 0.44|
   |    2|    gholi| 0.76| 0.76| 0.76|
   |    2|    gholi| 1.62| 1.62| 1.62|
   |    2|    gholi| 5.81| 5.81| 5.81|
   |    2| bcbcbcbc| 0.58| 0.58| 0.58|
   |    2| bcbcbcbc|-8.24|-8.24|-8.24|
   |    2| bcbcbcbc|-0.45|-0.45|-0.45|
   |    2| bcbcbcbc| 1.03| 1.03| 1.03|
   |    2|aUniqueID|-0.05|-0.05|-0.05|
   |    2| qqqqqqqq|-0.44|-0.44|-0.44|
   +-----+---------+-----+-----+-----+
   only showing top 20 rows
   
   >>> df.createOrReplaceTempView("df")
   ANTLR Tool version 4.7 used for code generation does not match the current runtime version 4.5.3ANTLR Runtime version 4.7 used for parser compilation does not match the current runtime version 4.5.3Traceback (most recent call last):
     File "<stdin>", line 1, in <module>
     File "/home/npansar/spark-2.3.0-bin-hadoop2.7/python/pyspark/sql/dataframe.py", line 176, in createOrReplaceTempView
       self._jdf.createOrReplaceTempView(name)
     File "/home/npansar/spark-2.3.0-bin-hadoop2.7/python/lib/py4j-0.10.6-src.zip/py4j/java_gateway.py", line 1160, in __call__
     File "/home/npansar/spark-2.3.0-bin-hadoop2.7/python/pyspark/sql/utils.py", line 63, in deco
       return f(*a, **kw)
     File "/home/npansar/spark-2.3.0-bin-hadoop2.7/python/lib/py4j-0.10.6-src.zip/py4j/protocol.py", line 320, in get_return_value
   py4j.protocol.Py4JJavaError: An error occurred while calling o52.createOrReplaceTempView.
   : java.lang.ExceptionInInitializerError
   	at org.apache.spark.sql.catalyst.parser.AbstractSqlParser.parse(ParseDriver.scala:84)
   	at org.apache.spark.sql.execution.SparkSqlParser.parse(SparkSqlParser.scala:48)
   	at org.apache.spark.sql.catalyst.parser.AbstractSqlParser.parseTableIdentifier(ParseDriver.scala:49)
   	at org.apache.spark.sql.Dataset.createTempViewCommand(Dataset.scala:3079)
   	at org.apache.spark.sql.Dataset.createOrReplaceTempView(Dataset.scala:3034)
   	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
   	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   	at java.lang.reflect.Method.invoke(Method.java:498)
   	at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
   	at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
   	at py4j.Gateway.invoke(Gateway.java:282)
   	at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
   	at py4j.commands.CallCommand.execute(CallCommand.java:79)
   	at py4j.GatewayConnection.run(GatewayConnection.java:214)
   	at java.lang.Thread.run(Thread.java:748)
   Caused by: java.lang.UnsupportedOperationException: java.io.InvalidClassException: org.antlr.v4.runtime.atn.ATN; Could not deserialize ATN with UUID 59627784-3be5-417a-b9eb-8131a7286089 (expected aadb8d7e-aeef-4415-ad2b-8204d6cf042e or a legacy UUID).
   	at org.antlr.v4.runtime.atn.ATNDeserializer.deserialize(ATNDeserializer.java:153)
   	at org.apache.spark.sql.catalyst.parser.SqlBaseLexer.<clinit>(SqlBaseLexer.java:1153)
   	... 16 more
   Caused by: java.io.InvalidClassException: org.antlr.v4.runtime.atn.ATN; Could not deserialize ATN with UUID 59627784-3be5-417a-b9eb-8131a7286089 (expected aadb8d7e-aeef-4415-ad2b-8204d6cf042e or a legacy UUID).
   	... 18 more
   
   >>> 
   ```

----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
users@infra.apache.org


With regards,
Apache Git Services