You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@lucene.apache.org by ma...@apache.org on 2014/03/19 18:39:39 UTC
svn commit: r1579307 - in /lucene/dev/branches/lucene_solr_4_7: ./ solr/
solr/CHANGES.txt solr/contrib/
solr/contrib/map-reduce/src/java/org/apache/solr/hadoop/MapReduceIndexerTool.java
Author: markrmiller
Date: Wed Mar 19 17:39:39 2014
New Revision: 1579307
URL: http://svn.apache.org/r1579307
Log:
SOLR-5782: The full MapReduceIndexer help text does not display when using --help.
Modified:
lucene/dev/branches/lucene_solr_4_7/ (props changed)
lucene/dev/branches/lucene_solr_4_7/solr/ (props changed)
lucene/dev/branches/lucene_solr_4_7/solr/CHANGES.txt (contents, props changed)
lucene/dev/branches/lucene_solr_4_7/solr/contrib/ (props changed)
lucene/dev/branches/lucene_solr_4_7/solr/contrib/map-reduce/src/java/org/apache/solr/hadoop/MapReduceIndexerTool.java
Modified: lucene/dev/branches/lucene_solr_4_7/solr/CHANGES.txt
URL: http://svn.apache.org/viewvc/lucene/dev/branches/lucene_solr_4_7/solr/CHANGES.txt?rev=1579307&r1=1579306&r2=1579307&view=diff
==============================================================================
--- lucene/dev/branches/lucene_solr_4_7/solr/CHANGES.txt (original)
+++ lucene/dev/branches/lucene_solr_4_7/solr/CHANGES.txt Wed Mar 19 17:39:39 2014
@@ -56,6 +56,9 @@ Bug Fixes
* SOLR-5866: UpdateShardHandler needs to use the system default scheme registry to
properly handle https via javax.net.ssl.* properties. (Steve Davids via shalin)
+
+* SOLR-5782: The full MapReduceIndexer help text does not display when using
+ --help. (Mark Miller, Wolfgang Hoschek)
================== 4.7.0 ==================
Modified: lucene/dev/branches/lucene_solr_4_7/solr/contrib/map-reduce/src/java/org/apache/solr/hadoop/MapReduceIndexerTool.java
URL: http://svn.apache.org/viewvc/lucene/dev/branches/lucene_solr_4_7/solr/contrib/map-reduce/src/java/org/apache/solr/hadoop/MapReduceIndexerTool.java?rev=1579307&r1=1579306&r2=1579307&view=diff
==============================================================================
--- lucene/dev/branches/lucene_solr_4_7/solr/contrib/map-reduce/src/java/org/apache/solr/hadoop/MapReduceIndexerTool.java (original)
+++ lucene/dev/branches/lucene_solr_4_7/solr/contrib/map-reduce/src/java/org/apache/solr/hadoop/MapReduceIndexerTool.java Wed Mar 19 17:39:39 2014
@@ -28,7 +28,6 @@ import java.io.InputStream;
import java.io.InputStreamReader;
import java.io.OutputStreamWriter;
import java.io.PrintWriter;
-import java.io.UnsupportedEncodingException;
import java.io.Writer;
import java.net.URISyntaxException;
import java.net.URL;
@@ -81,8 +80,8 @@ import org.apache.solr.hadoop.morphline.
import org.apache.solr.hadoop.morphline.MorphlineMapper;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
-
import org.kitesdk.morphline.base.Fields;
+
import com.google.common.base.Charsets;
import com.google.common.base.Preconditions;
import com.google.common.io.ByteStreams;
@@ -183,11 +182,7 @@ public class MapReduceIndexerTool extend
.action(new HelpArgumentAction() {
@Override
public void run(ArgumentParser parser, Argument arg, Map<String, Object> attrs, String flag, Object value) throws ArgumentParserException {
- try {
- parser.printHelp(new PrintWriter(new OutputStreamWriter(System.out, "UTF-8")));
- } catch (UnsupportedEncodingException e) {
- throw new RuntimeException("Won't Happen for UTF-8");
- }
+ parser.printHelp();
System.out.println();
System.out.print(ToolRunnerHelpFormatter.getGenericCommandUsage());
//ToolRunner.printGenericCommandUsage(System.out);
@@ -207,31 +202,6 @@ public class MapReduceIndexerTool extend
" --shards 1 \\\n" +
" hdfs:///user/$USER/test-documents/sample-statuses-20120906-141433.avro\n" +
"\n" +
- "# (Re)index all files that match all of the following conditions:\n" +
- "# 1) File is contained in dir tree hdfs:///user/$USER/solrloadtest/twitter/tweets\n" +
- "# 2) file name matches the glob pattern 'sample-statuses*.gz'\n" +
- "# 3) file was last modified less than 100000 minutes ago\n" +
- "# 4) file size is between 1 MB and 1 GB\n" +
- "# Also include extra library jar file containing JSON tweet Java parser:\n" +
- "hadoop jar target/solr-map-reduce-*.jar " + "com.cloudera.cdk.morphline.hadoop.find.HdfsFindTool" + " \\\n" +
- " -find hdfs:///user/$USER/solrloadtest/twitter/tweets \\\n" +
- " -type f \\\n" +
- " -name 'sample-statuses*.gz' \\\n" +
- " -mmin -1000000 \\\n" +
- " -size -100000000c \\\n" +
- " -size +1000000c \\\n" +
- "| sudo -u hdfs hadoop \\\n" +
- " --config /etc/hadoop/conf.cloudera.mapreduce1 \\\n" +
- " jar target/solr-map-reduce-*.jar \\\n" +
- " -D 'mapred.child.java.opts=-Xmx500m' \\\n" +
-// " -D 'mapreduce.child.java.opts=-Xmx500m' \\\n" +
- " --log4j src/test/resources/log4j.properties \\\n" +
- " --morphline-file ../search-core/src/test/resources/test-morphlines/tutorialReadJsonTestTweets.conf \\\n" +
- " --solr-home-dir src/test/resources/solr/minimr \\\n" +
- " --output-dir hdfs://c2202.mycompany.com/user/$USER/test \\\n" +
- " --shards 100 \\\n" +
- " --input-list -\n" +
- "\n" +
"# Go live by merging resulting index shards into a live Solr cluster\n" +
"# (explicitly specify Solr URLs - for a SolrCloud cluster see next example):\n" +
"sudo -u hdfs hadoop \\\n" +