You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@mahout.apache.org by sr...@apache.org on 2010/04/02 11:12:19 UTC
svn commit: r930192 - in
/lucene/mahout/trunk/core/src/main/java/org/apache/mahout:
cf/taste/hadoop/item/RecommenderJob.java
cf/taste/hadoop/pseudo/RecommenderJob.java common/AbstractJob.java
Author: srowen
Date: Fri Apr 2 09:12:19 2010
New Revision: 930192
URL: http://svn.apache.org/viewvc?rev=930192&view=rev
Log:
MAHOUT-350: doc updates, -D args must come first
Modified:
lucene/mahout/trunk/core/src/main/java/org/apache/mahout/cf/taste/hadoop/item/RecommenderJob.java
lucene/mahout/trunk/core/src/main/java/org/apache/mahout/cf/taste/hadoop/pseudo/RecommenderJob.java
lucene/mahout/trunk/core/src/main/java/org/apache/mahout/common/AbstractJob.java
Modified: lucene/mahout/trunk/core/src/main/java/org/apache/mahout/cf/taste/hadoop/item/RecommenderJob.java
URL: http://svn.apache.org/viewvc/lucene/mahout/trunk/core/src/main/java/org/apache/mahout/cf/taste/hadoop/item/RecommenderJob.java?rev=930192&r1=930191&r2=930192&view=diff
==============================================================================
--- lucene/mahout/trunk/core/src/main/java/org/apache/mahout/cf/taste/hadoop/item/RecommenderJob.java (original)
+++ lucene/mahout/trunk/core/src/main/java/org/apache/mahout/cf/taste/hadoop/item/RecommenderJob.java Fri Apr 2 09:12:19 2010
@@ -47,14 +47,17 @@ import org.apache.mahout.math.VectorWrit
* <p>Command line arguments specific to this class are:</p>
*
* <ol>
- * <li>--numRecommendations (integer): Number of recommendations to compute per user (optional; default 10)</li>
* <li>-Dmapred.input.dir=(path): Directory containing a text file containing user IDs for which recommendations should be
* computed, one per line</li>
* <li>-Dmapred.output.dir=(path): output path where recommender output should go</li>
* <li>--usersFile (path): file containing user IDs to recommend for (optional)</li>
+ * <li>--numRecommendations (integer): Number of recommendations to compute per user (optional; default 10)</li>
* </ol>
*
* <p>General command line options are documented in {@link AbstractJob}.</p>
+ *
+ * <p>Note that because of how Hadoop parses arguments, all "-D" arguments must appear before all other
+ * arguments.</p>
*/
public final class RecommenderJob extends AbstractJob {
Modified: lucene/mahout/trunk/core/src/main/java/org/apache/mahout/cf/taste/hadoop/pseudo/RecommenderJob.java
URL: http://svn.apache.org/viewvc/lucene/mahout/trunk/core/src/main/java/org/apache/mahout/cf/taste/hadoop/pseudo/RecommenderJob.java?rev=930192&r1=930191&r2=930192&view=diff
==============================================================================
--- lucene/mahout/trunk/core/src/main/java/org/apache/mahout/cf/taste/hadoop/pseudo/RecommenderJob.java (original)
+++ lucene/mahout/trunk/core/src/main/java/org/apache/mahout/cf/taste/hadoop/pseudo/RecommenderJob.java Fri Apr 2 09:12:19 2010
@@ -46,17 +46,20 @@ import org.apache.mahout.cf.taste.recomm
* <p>Command line arguments specific to this class are:</p>
*
* <ol>
+ * <li>-Dmapred.input.dir=(path): Location of a data model file containing preference data, suitable for use with
+ * {@link org.apache.mahout.cf.taste.impl.model.file.FileDataModel}</li>
+ * <li>-Dmapred.output.dir=(path): output path where recommender output should go</li>
* <li>--recommenderClassName (string): Fully-qualified class name of {@link Recommender} to use to make recommendations.
* Note that it must have a constructor which takes a {@link org.apache.mahout.cf.taste.model.DataModel}
* argument.</li>
* <li>--numRecommendations (integer): Number of recommendations to compute per user</li>
- * <li>-Dmapred.input.dir=(path): Location of a data model file containing preference data, suitable for use with
- * {@link org.apache.mahout.cf.taste.impl.model.file.FileDataModel}</li>
- * <li>-Dmapred.output.dir=(path): output path where recommender output should go</li>
* <li>--usersFile (path): file containing user IDs to recommend for (optional)</li>
* </ol>
*
* <p>General command line options are documented in {@link AbstractJob}.</p>
+ *
+ * <p>Note that because of how Hadoop parses arguments, all "-D" arguments must appear before all other
+ * arguments.</p>
*
* <p>
* For example, to get started trying this out, set up Hadoop in a pseudo-distributed manner:
@@ -70,7 +73,7 @@ import org.apache.mahout.cf.taste.recomm
* HDFS like so:
* </p>
*
- * {@code hadoop fs -put input.csv input/input.csv hadoop fs -put users.txt input/users.txt * }
+ * {@code hadoop fs -put input.csv input/input.csv; hadoop fs -put users.txt input/users.txt * }
*
* <p>
* Build Mahout code with <code>mvn package</code> in the core/ directory. Locate
@@ -89,9 +92,13 @@ import org.apache.mahout.cf.taste.recomm
* And launch:
* </p>
*
- * {@code hadoop jar recommender.jar org.apache.mahout.cf.taste.hadoop.pseudo.RecommenderJob \
- * --recommenderClassName your.project.Recommender \ --numRecommendations 10 -Dmapred.input.dir=input/users.csv \
- * -Dmapred.output.dir=output * }
+ * {@code hadoop jar recommender.jar \
+ * org.apache.mahout.cf.taste.hadoop.pseudo.RecommenderJob \
+ * -Dmapred.input.dir=input/users.csv \
+ * -Dmapred.output.dir=output \
+ * --recommenderClassName your.project.Recommender \
+ * --numRecommendations 10 *
+ * }
*/
public final class RecommenderJob extends AbstractJob {
Modified: lucene/mahout/trunk/core/src/main/java/org/apache/mahout/common/AbstractJob.java
URL: http://svn.apache.org/viewvc/lucene/mahout/trunk/core/src/main/java/org/apache/mahout/common/AbstractJob.java?rev=930192&r1=930191&r2=930192&view=diff
==============================================================================
--- lucene/mahout/trunk/core/src/main/java/org/apache/mahout/common/AbstractJob.java (original)
+++ lucene/mahout/trunk/core/src/main/java/org/apache/mahout/common/AbstractJob.java Fri Apr 2 09:12:19 2010
@@ -67,6 +67,9 @@ import org.slf4j.LoggerFactory;
* <li>-Dmapred.input.dir=(path): input file, or directory containing input files (required)</li>
* <li>-Dmapred.output.dir=(path): path to write output files (required)</li>
* </ul>
+ *
+ * <p>Note that because of how Hadoop parses arguments, all "-D" arguments must appear before all other
+ * arguments.</p>
*/
public abstract class AbstractJob extends Configured implements Tool {