You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@accumulo.apache.org by mm...@apache.org on 2016/12/13 19:27:00 UTC

accumulo-wikisearch git commit: Updates to INSTALL.md with 1.8.0 upgrade

Repository: accumulo-wikisearch
Updated Branches:
  refs/heads/master 5ab605107 -> 074fa7729


Updates to INSTALL.md with 1.8.0 upgrade


Project: http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/repo
Commit: http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/commit/074fa772
Tree: http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/tree/074fa772
Diff: http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/diff/074fa772

Branch: refs/heads/master
Commit: 074fa7729ee837932dc1a665c98419e9a33fe622
Parents: 5ab6051
Author: Mike Miller <mm...@apache.org>
Authored: Tue Dec 13 14:25:52 2016 -0500
Committer: Mike Miller <mm...@apache.org>
Committed: Tue Dec 13 14:25:52 2016 -0500

----------------------------------------------------------------------
 INSTALL.md | 5 +++--
 1 file changed, 3 insertions(+), 2 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/blob/074fa772/INSTALL.md
----------------------------------------------------------------------
diff --git a/INSTALL.md b/INSTALL.md
index fff2bc0..9a85105 100644
--- a/INSTALL.md
+++ b/INSTALL.md
@@ -27,7 +27,8 @@ Instructions for installing and running the Accumulo Wikisearch example.
 	 You will want to grab the files with the link name of pages-articles.xml.bz2. Though not strictly
 	 required, the ingest will go more quickly if the files are decompressed:
 
-        $ bunzip2 < enwiki-*-pages-articles.xml.bz2 | hadoop fs -put - /wikipedia/enwiki-pages-articles.xml
+        $ bunzip2 enwiki-*-pages-articles.xml.bz2
+        $ hadoop fs -put enwiki-*-pages-articles.xml /wikipedia/enwiki-pages-articles.xml
 
 ### Instructions
 	
@@ -39,7 +40,7 @@ Instructions for installing and running the Accumulo Wikisearch example.
         $ cp wikipedia.xml.example wikipedia.xml
         $ vim wikipedia.xml
  
-1. Copy `ingest/lib/wikisearch-*.jar` and `ingest/lib/protobuf*.jar` to `$ACCUMULO_HOME/lib/ext`
+1. Copy `ingest/lib/wikisearch-*.jar` to `$ACCUMULO_HOME/lib/ext`
 1. Run `ingest/bin/ingest.sh` (or `ingest_parallel.sh` if running parallel version) with one
    argument (the name of the directory in HDFS where the wikipedia XML files reside) and this will
    kick off a MapReduce job to ingest the data into Accumulo.