You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Apache Spark (JIRA)" <ji...@apache.org> on 2015/06/23 18:43:01 UTC
[jira] [Assigned] (SPARK-8558) Script /dev/run-tests fails when
_JAVA_OPTIONS env var set
[ https://issues.apache.org/jira/browse/SPARK-8558?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Apache Spark reassigned SPARK-8558:
-----------------------------------
Assignee: (was: Apache Spark)
> Script /dev/run-tests fails when _JAVA_OPTIONS env var set
> ----------------------------------------------------------
>
> Key: SPARK-8558
> URL: https://issues.apache.org/jira/browse/SPARK-8558
> Project: Spark
> Issue Type: Bug
> Components: Build, Tests
> Affects Versions: 1.4.0
> Environment: Centos 6
> Reporter: Oleksiy Dyagilev
> Priority: Minor
>
> Script /dev/run-tests.py fails when _JAVA_OPTIONS env. var set.
> Steps to reproduce in linux:
> 1. export _JAVA_OPTIONS="-Xmx2048M
> 2. ./dev/run-tests
> [pivot@fe2s spark]$ ./dev/run-tests
> Traceback (most recent call last):
> File "./dev/run-tests.py", line 793, in <module>
> main()
> File "./dev/run-tests.py", line 722, in main
> java_version = determine_java_version(java_exe)
> File "./dev/run-tests.py", line 484, in determine_java_version
> version, update = version_str.split('_') # eg ['1.8.0', '25']
> ValueError: need more than 1 value to unpack
> The problem is in 'determine_java_version' function in run-tests.py.
> It runs 'java' and extracts version from output. However when _JAVA_OPTIONS set the output of 'java' command is different and it breaks parser. See the first line
> [pivot@fe2s spark]$ java -version
> Picked up _JAVA_OPTIONS: -Xmx2048M
> java version "1.8.0_31"
> Java(TM) SE Runtime Environment (build 1.8.0_31-b13)
> Java HotSpot(TM) 64-Bit Server VM (build 25.31-b07, mixed mode)
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org