GitHub user tgravescs opened a pull request:

    https://github.com/apache/spark/pull/128

    [SPARK-1198] Allow pipes tasks to run in different sub-directories

    This works as is on Linux/Mac/etc but doesn't cover working on Windows.  In 
here I use ln -sf for symlinks. Putting this up for comments on that. Do we 
want to create perhaps some classes for doing shell commands - Linux vs 
Windows.  Is there some other way we want to do this?   I assume we are still 
supporting jdk1.6?
    
    Also should I update the Java API for pipes to allow this parameter?

You can merge this pull request into a Git repository by running:

    $ git pull https://github.com/tgravescs/spark SPARK1198

Alternatively you can review and apply these changes as the patch at:

    https://github.com/apache/spark/pull/128.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

    This closes #128
    
----
commit 1ab49ca90b7cae82efa26e018d9d285c948bf25c
Author: Thomas Graves <tgra...@apache.org>
Date:   2014-03-12T14:11:46Z

    Add support for running pipe tasks is separate directories

commit 6b783bdb5e09b7c96cbb76111876fbb6c9ca9a6f
Author: Thomas Graves <tgra...@apache.org>
Date:   2014-03-12T14:47:13Z

    style fixes

----


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---

Reply via email to