You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@mahout.apache.org by "Hudson (Jira)" <ji...@apache.org> on 2019/11/12 00:06:00 UTC

[jira] [Commented] (MAHOUT-2074) Dockerfile(s)

    [ https://issues.apache.org/jira/browse/MAHOUT-2074?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16971918#comment-16971918 ] 

Hudson commented on MAHOUT-2074:
--------------------------------

SUCCESS: Integrated in Jenkins build Mahout-Quality #3601 (See [https://builds.apache.org/job/Mahout-Quality/3601/])
[MAHOUT-2074] - Dockerfiles!! (apalumbo: rev ff2c23da1ca6b4d5cfde64d64ea9eb2105a72f9a)
* (edit) Dockerfile
* (edit) entrypoint.sh
[MAHOUT-2074] - entrypioint work (apalumbo: rev 3073fa7c0ff2e6d947a41deccb37cc0bfd60b2ed)
* (edit) entrypoint.sh


> Dockerfile(s) 
> --------------
>
>                 Key: MAHOUT-2074
>                 URL: https://issues.apache.org/jira/browse/MAHOUT-2074
>             Project: Mahout
>          Issue Type: New Feature
>    Affects Versions: 14.1
>            Reporter: Andrew Palumbo
>            Assignee: Andrew Palumbo
>            Priority: Critical
>             Fix For: 14.1
>
>
> Have a [WIP] Dockerfile for which (assuming a binary release,) pulls the appropriate version of Spark and places both Spark and Mahout in {{/opt/spark}} and {{/opt/mahout respectively.}}
> Would like to add full build mahout build capabilities (this should not be difficult) in a second file.
> these files currently use an ENTRYPOINT["entrypoint.sh"] command and some environment variables (none uncommon to Spark or Mahout aside from a $MAHOUT_CLASSPATH {{env}} variable).  
> the {{entrypiont.sh}} essentially. cheeks to see if the command is form a worker or a driver, and runs as such.  Currently I'm just dumping the entire {{$MAHOUT_HOME/lib/*.jar}} into the {{$MAHOUT_CLASSPATH}} and adding it to the {{SPARK_CLASSPATH}}.
> If the {{entrypoint.sh}} file detects a driver. it will launch {{spark-submit}}.  IIRC, which I so not think that I do, spark submit can handle any driver [~pferrel] Does this sound correct.  Otherwise we just add the mahout class to be passed to spark-submit class as a command parameter. 
> Though this may be better to migrate in 14.2 or 15.0 to an entire new build-chain. E.g. CMake.   (I would suggest) given our large amount of native code (hopefully soon to be added :)) 
> Though its nearly finished may want to punt this {{Dockerfile}} for 14.1, or mark it Experimental, is likely a better option.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)