You are viewing a plain text version of this content. The canonical link for it is here.
Posted to reviews@spark.apache.org by orenmazor <gi...@git.apache.org> on 2014/12/08 23:14:51 UTC
[GitHub] spark pull request: Cdh5
GitHub user orenmazor opened a pull request:
https://github.com/apache/spark/pull/3639
Cdh5
https://github.com/Shopify/dataops/issues/2
You can merge this pull request into a Git repository by running:
$ git pull https://github.com/Shopify/spark cdh5
Alternatively you can review and apply these changes as the patch at:
https://github.com/apache/spark/pull/3639.patch
To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:
This closes #3639
----
commit 422de4cc2a823e16b86fd22095e35d1ebe842a12
Author: Harry Brundage <ha...@gmail.com>
Date: 2014-01-15T01:29:43Z
Add compile script for packserv
commit 4ffa04cc6cc7bb8086a422a94d4f2e4105a69786
Author: Harry Brundage <ha...@gmail.com>
Date: 2014-01-15T02:18:29Z
Don't compile streaming when assembling cause it doesn't build against CDH4.4.0
commit b7bf08171e8eb796d86408ce5712175d781e0f8d
Author: Harry Brundage <ha...@gmail.com>
Date: 2014-01-15T02:22:10Z
Make script compile executable
commit 65033e665c75f4e82b56c8113c99308f8b419704
Author: Harry Brundage <ha...@gmail.com>
Date: 2014-01-15T02:22:57Z
Make script compile bash
commit 9e6fc96f461864f4ffdd6c8aefaa53b6fd8c4ae0
Author: Mark Cooper <mc...@quantcast.com>
Date: 2013-11-20T22:26:42Z
Add a environment variable that allows for configuring a different path to Spark binaries when running Spark from a different location locally
commit fdb0ce298048832f75b24b464fdf59fb791f869f
Author: Harry Brundage <ha...@gmail.com>
Date: 2014-01-15T19:24:05Z
Add fixed conf file with proper master and remote spark home
commit a837356d7d84641ab504522e74cedc4b5d865aa3
Author: Harry Brundage <ha...@gmail.com>
Date: 2014-01-21T01:50:22Z
Copy in hadoop core-site.xml so local clones know where to find hdfs.
commit 4d0f3682e0931c21ba6e5b01fc42ee33a44453e1
Author: Harry Brundage <ha...@gmail.com>
Date: 2014-01-21T02:12:36Z
Update the given spark env to actually work, and only if a custom master isn't provided.
commit 01cf4c51f2c3c3089ee91dd64d6cab32dd17aa70
Author: Harry Brundage <ha...@gmail.com>
Date: 2014-01-21T04:22:11Z
Allow controlling the number of cores pyspark uses using the `-c` option, like spark-shell.
- Turns out there isn't actually a way right now to control the number of cores an interactive pyspark session uses, which is annoying if more than one person is trying to work on a cluster interactively at once.
- Use the python 2.7 stdlib argparse library to pull out the -c option
- This requires changing the bin/pyspark shell script to pass all arguments to the python script instead of allowing the python interpreter program to parse any of them.
commit 91ddfb4c43a88a4cf0082e445e2e82bcde069969
Author: Harry Brundage <ha...@gmail.com>
Date: 2014-01-21T04:22:34Z
Merge branch 'pyspark_cores'
commit 0b44511492131b60f744527eee467fd147e4f4c0
Author: Harry Brundage <ha...@gmail.com>
Date: 2014-01-21T14:51:05Z
Revert "Merge branch 'pyspark_cores'"
This reverts commit 91ddfb4c43a88a4cf0082e445e2e82bcde069969, reversing
changes made to 4d0f3682e0931c21ba6e5b01fc42ee33a44453e1.
commit b4c5ff7e7d6d550743e3aa97710fa514744b0c6e
Author: Harry Brundage <ha...@gmail.com>
Date: 2014-01-21T15:22:09Z
Auto setup python and warn if the vpn isn't connected
commit 4c2c45eaf14197b79cf5949bb370a74c52a38ff0
Author: Harry Brundage <ha...@gmail.com>
Date: 2014-01-21T15:34:05Z
Add an applescript to the spark conf file that autoconnects the VPN if it can't find the interface the VPN should create
commit 712b8856e4b14f88d34da569505c59884d8e8155
Author: Harry Brundage <ha...@gmail.com>
Date: 2014-01-21T19:24:21Z
Check to see if Viscosity is a thing before trying to tell it to connect in spark env setup
commit 986a60c0b9a880ee0fd7242e53458efbebfab73e
Author: Harry Brundage <ha...@gmail.com>
Date: 2014-01-21T21:21:18Z
Merge pull request #1 from Shopify/autoconnect_vpn
Autoconnect VPN
commit b944fc6ff5f5866b93e937f7d7629370c24944f0
Author: Dana Klassen <kl...@gmail.com>
Date: 2014-01-23T01:36:53Z
change configuration to be set through environment variable
commit 5eace91604360da5b446a96582f141b09ab109c1
Author: Erik Selin <er...@jadedpixel.com>
Date: 2014-01-23T04:14:17Z
apply pr 494 and 496
commit 42dc1708daec21a3ba302f61f473afa57fb5c12c
Author: Dana Klassen <kl...@gmail.com>
Date: 2014-01-23T12:15:27Z
Merge pull request #2 from Shopify/config_hdfs
Config hdfs
commit 2b6c170b50f58ccdbe1e2faaf4ff3439bdf9e01e
Author: Erik Selin <ty...@gmail.com>
Date: 2014-01-23T15:28:32Z
Merge pull request #3 from Shopify/apply_494_and_496
apply pr 494 and 496
commit 25c5a0d90c5926133b32e43f5e6a8d1a58c0685c
Author: Patrick Wendell <pw...@gmail.com>
Date: 2014-01-15T18:58:02Z
Workers should use working directory as spark home if it's not specified
commit c0cd09186b5cef2f15f3dfb4bd9dadea6392385c
Author: Erik Selin <ty...@gmail.com>
Date: 2014-01-24T00:37:45Z
Merge pull request #5 from Shopify/apply_commit_00a3f7e
Workers should use working directory as spark home if it's not specified
commit df6098e1254a5d7128574d2c2fc31dae005c8fd7
Author: Harry Brundage <ha...@gmail.com>
Date: 2014-01-24T00:14:06Z
Fix broken quoting in spark-env.sh and echo to stderr to not break stdout communication with pyspark
commit 1b71ba315487b3bb6214631fece93cbd8ef27fb3
Author: Erik Selin <er...@jadedpixel.com>
Date: 2014-01-26T19:47:37Z
Add custom writeUTF method to PythonRDD and make it replace DataOutputStream.writeUTF so that we can use items with greater size in pyspark. Also modify pyspark MUTF8Deserializer to expect the size of the data blocks to be represented by an integer instead of an unsigned short.
commit af58a246f3babb5848c248eb133e88336c15d3ec
Author: Erik Selin <er...@jadedpixel.com>
Date: 2014-01-26T22:21:39Z
remove alignment
commit 39330f3a41dd6ccdd8a8211ea36e33b5600e27ec
Author: Erik Selin <er...@jadedpixel.com>
Date: 2014-01-26T22:38:27Z
remove using MUTF to send data to pyspark and use UTF from getBytes instead.
commit 80c976e3c5befcc02f5581a15058afd825fe16e2
Author: Bouke van der Bijl <bo...@gmail.com>
Date: 2014-01-27T16:11:23Z
Add test for writing large strings
commit 9de42052c8c622d11d20e82bdf18f312fc1e0d0e
Author: Harry Brundage <ha...@gmail.com>
Date: 2014-01-27T21:34:26Z
Merge pull request #6 from Shopify/conf_fixes_for_standalone_pyspark
Fix broken quoting in spark-env.sh and echo to stderr to not break stdout communication with pyspark
commit 57d86ee81ac02a9c8357ab56c16c0577491bc4c4
Author: Harry Brundage <ha...@gmail.com>
Date: 2014-02-05T02:57:38Z
Remove erroneous local SPARK_HOME environment variable setting
commit 47e44f87963e5ba0c9fa51f14db55d1a06f6fa38
Author: Harry Brundage <ha...@gmail.com>
Date: 2014-02-05T03:24:15Z
Don't let spark spend time figuring out what IP to bind the web server to when executing in local mode
commit 812a7880938a415e7aa89d9e82e13dc827b7d119
Author: Harry Brundage <ha...@gmail.com>
Date: 2014-02-05T03:25:09Z
Get rid of kerberos auth testing. No more "Unable to load realm info from SCDynamicStore"! Woo!
See http://stackoverflow.com/questions/7134723/hadoop-on-osx-unable-to-load-realm-info-from-scdynamicstore
----
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---
---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org
[GitHub] spark pull request: Cdh5
Posted by orenmazor <gi...@git.apache.org>.
Github user orenmazor closed the pull request at:
https://github.com/apache/spark/pull/3639
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastructure@apache.org or file a JIRA ticket
with INFRA.
---
---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscribe@spark.apache.org
For additional commands, e-mail: reviews-help@spark.apache.org