You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Sean Owen (JIRA)" <ji...@apache.org> on 2016/01/21 13:56:39 UTC
[jira] [Resolved] (SPARK-12954) pyspark API 1.3.0 how we can
patitionning by columns
[ https://issues.apache.org/jira/browse/SPARK-12954?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Sean Owen resolved SPARK-12954.
-------------------------------
Resolution: Invalid
Target Version/s: (was: 1.3.0)
[~Malouke] a lot is wrong with this. Please don't open a JIRA until you read https://cwiki.apache.org/confluence/display/SPARK/Contributing+to+Spark
Don't set Blocker or Target Version. Ask questions on user@spark.apache.org
> pyspark API 1.3.0 how we can patitionning by columns
> -------------------------------------------------------
>
> Key: SPARK-12954
> URL: https://issues.apache.org/jira/browse/SPARK-12954
> Project: Spark
> Issue Type: Bug
> Components: PySpark
> Affects Versions: 1.3.0
> Environment: spark 1.3.0
> cloudera manger
> linux platfrome
> pyspark
> Reporter: malouke
> Priority: Blocker
> Labels: documentation, features, performance, test
>
> hi,
> before posting this question i try lot of things , but i dont found solution.
> i have 9 table and i join thems with two ways:
> -1 first test with df.join(df2, df.id == df.id2,'left_outer')
> -2 sqlcontext.sql("select * from t1 left join t2 on id_t1=id_t2")
> after that i want partition by date the result of join :
> -in pyspark 1.5.2 i try partitionBy if table it's not comming from result of at most two tables evry thiings ok. but when i join more than three tables i dont have result after severals hours .
> - in pyspark 1.3.0 i dont found in api one function let me partition by dat columns
> Q: some one can help me to resolve this probleme
> thank you in advance
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org