You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Dat Tran (JIRA)" <ji...@apache.org> on 2016/01/11 15:54:39 UTC
[jira] [Created] (SPARK-12753) Import error during unit test while
calling a function from reduceByKey()
Dat Tran created SPARK-12753:
--------------------------------
Summary: Import error during unit test while calling a function from reduceByKey()
Key: SPARK-12753
URL: https://issues.apache.org/jira/browse/SPARK-12753
Project: Spark
Issue Type: Question
Components: PySpark
Affects Versions: 1.6.0
Environment: El Capitan, Single cluster Hadoop, Python 3, Spark 1.6, Anaconda
Reporter: Dat Tran
Priority: Trivial
The current directory structure for my test script is as follows:
project/
script/
__init__.py
map.py
test/
__init.py__
test_map.py
I have attached map.py and test_map.py file with this issue.
When I run the nosetest in the test directory, the test fails. I get no module named "script" found error.
However when I modify the map_add function to replace the call to add within reduceByKey in map.py like this:
def map_add(df):
result = df.map(lambda x: (x.key, x.value)).reduceByKey(lambda x,y: x+y)
return result
The test passes.
Also, when I run the original test_map.py from the project directory, the test passes.
I am not able to figure out why the test doesn't detect the script module when it is within the test directory.
I have also attached the log error file. Any help will be much appreciated.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org