You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@accumulo.apache.org by mm...@apache.org on 2016/12/12 19:21:55 UTC
accumulo-wikisearch git commit: ACCUMULO-2446: many updates for
Accumulo 1.8.0
Repository: accumulo-wikisearch
Updated Branches:
refs/heads/1.8 [created] dd23f1071
ACCUMULO-2446: many updates for Accumulo 1.8.0
Project: http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/repo
Commit: http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/commit/dd23f107
Tree: http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/tree/dd23f107
Diff: http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/diff/dd23f107
Branch: refs/heads/1.8
Commit: dd23f1071068625c6958001a1c0bc7ad39e30227
Parents: 9c30660
Author: Mike Miller <mm...@apache.org>
Authored: Thu Dec 1 15:18:30 2016 -0500
Committer: Mike Miller <mm...@apache.org>
Committed: Fri Dec 9 15:22:46 2016 -0500
----------------------------------------------------------------------
.gitignore | 4 +
ingest/bin/ingest.sh | 2 +-
ingest/conf/wikipedia.xml.uno | 43 ++++++++
ingest/pom.xml | 57 ++---------
.../wikisearch/ingest/WikipediaIngester.java | 13 ++-
.../ingest/WikipediaPartitionedIngester.java | 12 ++-
.../output/BufferingRFileRecordWriter.java | 47 ++++-----
.../output/SortingRFileOutputFormat.java | 9 +-
pom.xml | 102 ++++++-------------
query-war/pom.xml | 2 +-
query/pom.xml | 95 ++---------------
.../iterator/DefaultIteratorEnvironment.java | 22 ++++
.../examples/wikisearch/query/Query.java | 5 +-
13 files changed, 159 insertions(+), 254 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/blob/dd23f107/.gitignore
----------------------------------------------------------------------
diff --git a/.gitignore b/.gitignore
index 698a832..8a09ac0 100644
--- a/.gitignore
+++ b/.gitignore
@@ -2,3 +2,7 @@
.idea
**/*.iml
**/lib
+.project
+.settings/
+.classpath
+
http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/blob/dd23f107/ingest/bin/ingest.sh
----------------------------------------------------------------------
diff --git a/ingest/bin/ingest.sh b/ingest/bin/ingest.sh
index 73d582d..f9f3495 100755
--- a/ingest/bin/ingest.sh
+++ b/ingest/bin/ingest.sh
@@ -38,7 +38,7 @@ LIBJARS=`echo $CLASSPATH | sed 's/^://' | sed 's/:/,/g'`
#
# Map/Reduce job
#
-JAR=$SCRIPT_DIR/../lib/wikisearch-ingest-1.5.0.jar
+JAR=$SCRIPT_DIR/../lib/wikisearch-ingest-*.jar
CONF=$SCRIPT_DIR/../conf/wikipedia.xml
HDFS_DATA_DIR=$1
export HADOOP_CLASSPATH=$CLASSPATH
http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/blob/dd23f107/ingest/conf/wikipedia.xml.uno
----------------------------------------------------------------------
diff --git a/ingest/conf/wikipedia.xml.uno b/ingest/conf/wikipedia.xml.uno
new file mode 100644
index 0000000..093551c
--- /dev/null
+++ b/ingest/conf/wikipedia.xml.uno
@@ -0,0 +1,43 @@
+<?xml version="1.0"?>
+<!--
+ Licensed to the Apache Software Foundation (ASF) under one or more
+ contributor license agreements. See the NOTICE file distributed with
+ this work for additional information regarding copyright ownership.
+ The ASF licenses this file to You under the Apache License, Version 2.0
+ (the "License"); you may not use this file except in compliance with
+ the License. You may obtain a copy of the License at
+
+ http://www.apache.org/licenses/LICENSE-2.0
+
+ Unless required by applicable law or agreed to in writing, software
+ distributed under the License is distributed on an "AS IS" BASIS,
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ See the License for the specific language governing permissions and
+ limitations under the License.
+-->
+<configuration>
+ <property>
+ <name>wikipedia.accumulo.zookeepers</name>
+ <value>localhost:2181</value>
+ </property>
+ <property>
+ <name>wikipedia.accumulo.instance_name</name>
+ <value>uno</value>
+ </property>
+ <property>
+ <name>wikipedia.accumulo.user</name>
+ <value>root</value>
+ </property>
+ <property>
+ <name>wikipedia.accumulo.password</name>
+ <value>secret</value>
+ </property>
+ <property>
+ <name>wikipedia.accumulo.table</name>
+ <value>wikipedia</value>
+ </property>
+ <property>
+ <name>wikipedia.ingest.partitions</name>
+ <value>1</value>
+ </property>
+</configuration>
http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/blob/dd23f107/ingest/pom.xml
----------------------------------------------------------------------
diff --git a/ingest/pom.xml b/ingest/pom.xml
index 1d17e27..426cff2 100644
--- a/ingest/pom.xml
+++ b/ingest/pom.xml
@@ -20,7 +20,7 @@
<parent>
<groupId>org.apache.accumulo</groupId>
<artifactId>accumulo-wikisearch</artifactId>
- <version>1.5.0</version>
+ <version>1.8.0</version>
</parent>
<artifactId>wikisearch-ingest</artifactId>
<name>wikisearch-ingest</name>
@@ -48,6 +48,12 @@
<dependency>
<groupId>org.apache.accumulo</groupId>
<artifactId>accumulo-core</artifactId>
+ <exclusions>
+ <exclusion>
+ <groupId>commons-digester</groupId>
+ <artifactId>commons-digester</artifactId>
+ </exclusion>
+ </exclusions>
</dependency>
<dependency>
<groupId>org.apache.lucene</groupId>
@@ -101,53 +107,4 @@
</plugin>
</plugins>
</build>
- <profiles>
- <!-- profile for building against Hadoop 1.0.x
- Activate by not specifying hadoop.profile -->
- <profile>
- <id>hadoop-1.0</id>
- <activation>
- <property>
- <name>!hadoop.profile</name>
- </property>
- </activation>
- <dependencies>
- <dependency>
- <groupId>org.apache.hadoop</groupId>
- <artifactId>hadoop-core</artifactId>
- <version>${version.hadoop}</version>
- <exclusions>
- <exclusion>
- <groupId>commons-logging</groupId>
- <artifactId>commons-logging</artifactId>
- </exclusion>
- </exclusions>
- </dependency>
- </dependencies>
- </profile>
- <!-- profile for building against Hadoop 2.0.x
- Activate using: mvn -Dhadoop.profile=2.0 -->
- <profile>
- <id>hadoop-2.0</id>
- <activation>
- <property>
- <name>hadoop.profile</name>
- <value>2.0</value>
- </property>
- </activation>
- <dependencies>
- <dependency>
- <groupId>org.apache.hadoop</groupId>
- <artifactId>hadoop-client</artifactId>
- <version>${version.hadoop}</version>
- <exclusions>
- <exclusion>
- <groupId>org.codehaus.jackson</groupId>
- <artifactId>jackson-mapper-asl</artifactId>
- </exclusion>
- </exclusions>
- </dependency>
- </dependencies>
- </profile>
- </profiles>
</project>
http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/blob/dd23f107/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/ingest/WikipediaIngester.java
----------------------------------------------------------------------
diff --git a/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/ingest/WikipediaIngester.java b/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/ingest/WikipediaIngester.java
index d4fa1c6..1a495ed 100644
--- a/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/ingest/WikipediaIngester.java
+++ b/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/ingest/WikipediaIngester.java
@@ -29,13 +29,16 @@ import java.util.regex.Pattern;
import org.apache.accumulo.core.client.AccumuloException;
import org.apache.accumulo.core.client.AccumuloSecurityException;
+import org.apache.accumulo.core.client.ClientConfiguration;
import org.apache.accumulo.core.client.Connector;
import org.apache.accumulo.core.client.IteratorSetting;
import org.apache.accumulo.core.client.IteratorSetting.Column;
import org.apache.accumulo.core.client.TableExistsException;
import org.apache.accumulo.core.client.TableNotFoundException;
+import org.apache.accumulo.core.client.ClientConfiguration.ClientProperty;
import org.apache.accumulo.core.client.admin.TableOperations;
import org.apache.accumulo.core.client.mapreduce.AccumuloOutputFormat;
+import org.apache.accumulo.core.client.security.tokens.PasswordToken;
import org.apache.accumulo.core.data.Mutation;
import org.apache.accumulo.core.iterators.IteratorUtil.IteratorScope;
import org.apache.accumulo.core.iterators.user.SummingCombiner;
@@ -135,9 +138,9 @@ public class WikipediaIngester extends Configured implements Tool {
conf.set("mapred.map.tasks.speculative.execution", "false");
String tablename = WikipediaConfiguration.getTableName(conf);
-
- String zookeepers = WikipediaConfiguration.getZookeepers(conf);
- String instanceName = WikipediaConfiguration.getInstanceName(conf);
+ ClientConfiguration clientConfig = new ClientConfiguration();
+ clientConfig.setProperty(ClientProperty.INSTANCE_NAME, WikipediaConfiguration.getInstanceName(conf));
+ clientConfig.setProperty(ClientProperty.INSTANCE_ZK_HOST, WikipediaConfiguration.getZookeepers(conf));
String user = WikipediaConfiguration.getUser(conf);
byte[] password = WikipediaConfiguration.getPassword(conf);
@@ -168,8 +171,8 @@ public class WikipediaIngester extends Configured implements Tool {
job.setMapOutputKeyClass(Text.class);
job.setMapOutputValueClass(Mutation.class);
job.setOutputFormatClass(AccumuloOutputFormat.class);
- AccumuloOutputFormat.setOutputInfo(job.getConfiguration(), user, password, true, tablename);
- AccumuloOutputFormat.setZooKeeperInstance(job.getConfiguration(), instanceName, zookeepers);
+ AccumuloOutputFormat.setConnectorInfo(job, user, new PasswordToken(password));
+ AccumuloOutputFormat.setZooKeeperInstance(job, clientConfig);
return job.waitForCompletion(true) ? 0 : 1;
}
http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/blob/dd23f107/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/ingest/WikipediaPartitionedIngester.java
----------------------------------------------------------------------
diff --git a/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/ingest/WikipediaPartitionedIngester.java b/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/ingest/WikipediaPartitionedIngester.java
index 59035dc..841f169 100644
--- a/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/ingest/WikipediaPartitionedIngester.java
+++ b/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/ingest/WikipediaPartitionedIngester.java
@@ -29,13 +29,16 @@ import java.util.regex.Pattern;
import org.apache.accumulo.core.client.AccumuloException;
import org.apache.accumulo.core.client.AccumuloSecurityException;
+import org.apache.accumulo.core.client.ClientConfiguration;
import org.apache.accumulo.core.client.Connector;
import org.apache.accumulo.core.client.IteratorSetting;
import org.apache.accumulo.core.client.IteratorSetting.Column;
import org.apache.accumulo.core.client.TableExistsException;
import org.apache.accumulo.core.client.TableNotFoundException;
+import org.apache.accumulo.core.client.ClientConfiguration.ClientProperty;
import org.apache.accumulo.core.client.admin.TableOperations;
import org.apache.accumulo.core.client.mapreduce.AccumuloOutputFormat;
+import org.apache.accumulo.core.client.security.tokens.PasswordToken;
import org.apache.accumulo.core.data.Mutation;
import org.apache.accumulo.core.iterators.IteratorUtil.IteratorScope;
import org.apache.accumulo.core.iterators.user.SummingCombiner;
@@ -236,12 +239,13 @@ public class WikipediaPartitionedIngester extends Configured implements Tool {
SortingRFileOutputFormat.setPathName(ingestConf, WikipediaConfiguration.bulkIngestDir(ingestConf));
} else {
ingestJob.setOutputFormatClass(AccumuloOutputFormat.class);
- String zookeepers = WikipediaConfiguration.getZookeepers(ingestConf);
- String instanceName = WikipediaConfiguration.getInstanceName(ingestConf);
+ ClientConfiguration clientConfig = new ClientConfiguration();
+ clientConfig.setProperty(ClientProperty.INSTANCE_NAME, WikipediaConfiguration.getInstanceName(ingestConf));
+ clientConfig.setProperty(ClientProperty.INSTANCE_ZK_HOST, WikipediaConfiguration.getZookeepers(ingestConf));
String user = WikipediaConfiguration.getUser(ingestConf);
byte[] password = WikipediaConfiguration.getPassword(ingestConf);
- AccumuloOutputFormat.setOutputInfo(ingestJob.getConfiguration(), user, password, true, tablename);
- AccumuloOutputFormat.setZooKeeperInstance(ingestJob.getConfiguration(), instanceName, zookeepers);
+ AccumuloOutputFormat.setConnectorInfo(ingestJob, user, new PasswordToken(password));
+ AccumuloOutputFormat.setZooKeeperInstance(ingestJob, clientConfig);
}
return ingestJob.waitForCompletion(true) ? 0 : 1;
http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/blob/dd23f107/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/output/BufferingRFileRecordWriter.java
----------------------------------------------------------------------
diff --git a/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/output/BufferingRFileRecordWriter.java b/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/output/BufferingRFileRecordWriter.java
index 9b663de..aa6e357 100644
--- a/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/output/BufferingRFileRecordWriter.java
+++ b/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/output/BufferingRFileRecordWriter.java
@@ -22,13 +22,18 @@ import java.util.Map;
import java.util.Map.Entry;
import java.util.TreeMap;
+import org.apache.accumulo.core.client.AccumuloException;
+import org.apache.accumulo.core.client.AccumuloSecurityException;
+import org.apache.accumulo.core.client.BatchWriter;
+import org.apache.accumulo.core.client.BatchWriterConfig;
+import org.apache.accumulo.core.client.Connector;
+import org.apache.accumulo.core.client.TableNotFoundException;
import org.apache.accumulo.core.conf.AccumuloConfiguration;
import org.apache.accumulo.core.data.ColumnUpdate;
import org.apache.accumulo.core.data.Key;
import org.apache.accumulo.core.data.Mutation;
import org.apache.accumulo.core.data.Value;
-import org.apache.accumulo.core.file.FileSKVWriter;
-import org.apache.accumulo.core.file.rfile.RFileOperations;
+import org.apache.accumulo.examples.wikisearch.ingest.WikipediaConfiguration;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.io.Text;
@@ -37,12 +42,7 @@ import org.apache.hadoop.mapreduce.TaskAttemptContext;
final class BufferingRFileRecordWriter extends RecordWriter<Text,Mutation> {
private final long maxSize;
- private final AccumuloConfiguration acuconf;
private final Configuration conf;
- private final String filenamePrefix;
- private final String taskID;
- private final FileSystem fs;
- private int fileCount = 0;
private long size;
private Map<Text,TreeMap<Key,Value>> buffers = new HashMap<Text,TreeMap<Key,Value>>();
@@ -79,31 +79,32 @@ final class BufferingRFileRecordWriter extends RecordWriter<Text,Mutation> {
if (buffer.size() == 0)
return;
- String file = filenamePrefix + "/" + tablename + "/" + taskID + "_" + (fileCount++) + ".rf";
+ Connector conn;
+ try {
+ conn = WikipediaConfiguration.getConnector(conf);
+ BatchWriterConfig bwconfig = new BatchWriterConfig();
+ BatchWriter writer = conn.createBatchWriter(tablename.toString(), bwconfig);
+ for (Entry<Key,Value> e : buffer.entrySet()) {
+ Key k = e.getKey();
+ Mutation m = new Mutation();
+ m.put(k.getColumnFamily(), k.getColumnQualifier(), e.getValue());
+ writer.addMutation(m);
+ }
+ writer.close();
+ } catch (AccumuloException | AccumuloSecurityException | TableNotFoundException e1) {
+ // TODO Auto-generated catch block
+ e1.printStackTrace();
+ }
// TODO get the table configuration for the given table?
- FileSKVWriter writer = RFileOperations.getInstance().openWriter(file, fs, conf, acuconf);
-
- // forget locality groups for now, just write everything to the default
- writer.startDefaultLocalityGroup();
-
- for (Entry<Key,Value> e : buffer.entrySet()) {
- writer.append(e.getKey(), e.getValue());
- }
-
- writer.close();
size -= bufferSize;
buffer.clear();
bufferSizes.put(tablename, 0l);
}
- BufferingRFileRecordWriter(long maxSize, AccumuloConfiguration acuconf, Configuration conf, String filenamePrefix, String taskID, FileSystem fs) {
+ BufferingRFileRecordWriter(long maxSize, Configuration conf) {
this.maxSize = maxSize;
- this.acuconf = acuconf;
this.conf = conf;
- this.filenamePrefix = filenamePrefix;
- this.taskID = taskID;
- this.fs = fs;
}
@Override
http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/blob/dd23f107/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/output/SortingRFileOutputFormat.java
----------------------------------------------------------------------
diff --git a/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/output/SortingRFileOutputFormat.java b/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/output/SortingRFileOutputFormat.java
index 1fa8fdc..c54db06 100644
--- a/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/output/SortingRFileOutputFormat.java
+++ b/ingest/src/main/java/org/apache/accumulo/examples/wikisearch/output/SortingRFileOutputFormat.java
@@ -105,17 +105,10 @@ public class SortingRFileOutputFormat extends OutputFormat<Text,Mutation> {
// grab the configuration
final Configuration conf = attempt.getConfiguration();
- // create a filename
- final String filenamePrefix = getPathName(conf);
- final String taskID = attempt.getTaskAttemptID().toString();
// grab the max size
final long maxSize = getMaxBufferSize(conf);
- // grab the FileSystem
- final FileSystem fs = FileSystem.get(conf);
- // create a default AccumuloConfiguration
- final AccumuloConfiguration acuconf = AccumuloConfiguration.getDefaultConfiguration();
- return new BufferingRFileRecordWriter(maxSize, acuconf, conf, filenamePrefix, taskID, fs);
+ return new BufferingRFileRecordWriter(maxSize, conf);
}
}
http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/blob/dd23f107/pom.xml
----------------------------------------------------------------------
diff --git a/pom.xml b/pom.xml
index 2e3cb3f..87fd3b2 100644
--- a/pom.xml
+++ b/pom.xml
@@ -21,12 +21,12 @@
<parent>
<groupId>org.apache</groupId>
<artifactId>apache</artifactId>
- <version>14</version>
+ <version>18</version>
</parent>
<groupId>org.apache.accumulo</groupId>
<artifactId>accumulo-wikisearch</artifactId>
- <version>1.5.0</version>
+ <version>1.8.0</version>
<packaging>pom</packaging>
<name>accumulo-wikisearch</name>
@@ -44,7 +44,7 @@
<version.commons-lang>2.4</version.commons-lang>
<version.ejb-spec-api>1.0.1.Final</version.ejb-spec-api>
<version.guava>11.0.2</version.guava>
- <version.hadoop>1.0.4</version.hadoop>
+ <version.hadoop>2.6.4</version.hadoop>
<version.jaxrs>2.1.0.GA</version.jaxrs>
<version.junit>4.11</version.junit>
<version.kryo>1.04</version.kryo>
@@ -55,7 +55,7 @@
<version.lucene-wikipedia>3.0.2</version.lucene-wikipedia>
<version.minlog>1.2</version.minlog>
<version.protobuf>2.3.0</version.protobuf>
- <version.thrift>0.9.0</version.thrift>
+ <version.thrift>0.9.1</version.thrift>
<version.zookeeper>3.3.1</version.zookeeper>
</properties>
<dependencyManagement>
@@ -76,6 +76,16 @@
<version>${version.collections}</version>
</dependency>
<dependency>
+ <groupId>commons-logging</groupId>
+ <artifactId>commons-logging</artifactId>
+ <version>1.1.1</version>
+ </dependency>
+ <dependency>
+ <groupId>org.apache.commons</groupId>
+ <artifactId>commons-math3</artifactId>
+ <version>3.6.1</version>
+ </dependency>
+ <dependency>
<groupId>junit</groupId>
<artifactId>junit</artifactId>
<version>${version.junit}</version>
@@ -206,6 +216,21 @@
</exclusion>
</exclusions>
</dependency>
+ <dependency>
+ <groupId>org.slf4j</groupId>
+ <artifactId>slf4j-api</artifactId>
+ <version>1.7.21</version>
+ </dependency>
+ <dependency>
+ <groupId>org.codehaus.jackson</groupId>
+ <artifactId>jackson-core-asl</artifactId>
+ <version>1.9.13</version>
+ </dependency>
+ <dependency>
+ <groupId>org.codehaus.jackson</groupId>
+ <artifactId>jackson-mapper-asl</artifactId>
+ <version>1.9.13</version>
+ </dependency>
</dependencies>
</dependencyManagement>
<repositories>
@@ -351,73 +376,4 @@
</plugin>
</plugins>
</build>
- <profiles>
- <profile>
- <!-- profile for building against Hadoop 1.0.x
- Activate by not specifying hadoop.profile -->
- <id>hadoop-1.0</id>
- <activation>
- <property>
- <name>!hadoop.profile</name>
- </property>
- </activation>
- <build>
- <plugins>
- <plugin>
- <groupId>org.codehaus.mojo</groupId>
- <artifactId>build-helper-maven-plugin</artifactId>
- <version>1.8</version>
- <executions>
- <execution>
- <id>add-test-source</id>
- <goals>
- <goal>add-test-source</goal>
- </goals>
- <phase>generate-test-sources</phase>
- <configuration>
- <sources>
- <source>src/test/hadoop1</source>
- </sources>
- </configuration>
- </execution>
- </executions>
- </plugin>
- </plugins>
- </build>
- </profile>
- <profile>
- <!-- profile for building against Hadoop 2.0.x
- Activate using: mvn -Dhadoop.profile=2.0 -->
- <id>hadoop-2.0</id>
- <activation>
- <property>
- <name>hadoop.profile</name>
- <value>2.0</value>
- </property>
- </activation>
- <build>
- <plugins>
- <plugin>
- <groupId>org.codehaus.mojo</groupId>
- <artifactId>build-helper-maven-plugin</artifactId>
- <version>1.8</version>
- <executions>
- <execution>
- <id>add-test-source</id>
- <goals>
- <goal>add-test-source</goal>
- </goals>
- <phase>generate-test-sources</phase>
- <configuration>
- <sources>
- <source>src/test/hadoop2</source>
- </sources>
- </configuration>
- </execution>
- </executions>
- </plugin>
- </plugins>
- </build>
- </profile>
- </profiles>
</project>
http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/blob/dd23f107/query-war/pom.xml
----------------------------------------------------------------------
diff --git a/query-war/pom.xml b/query-war/pom.xml
index 3ee0bc2..45a1f1b 100644
--- a/query-war/pom.xml
+++ b/query-war/pom.xml
@@ -20,7 +20,7 @@
<parent>
<groupId>org.apache.accumulo</groupId>
<artifactId>accumulo-wikisearch</artifactId>
- <version>1.5.0</version>
+ <version>1.8.0</version>
</parent>
<artifactId>wikisearch-query-war</artifactId>
http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/blob/dd23f107/query/pom.xml
----------------------------------------------------------------------
diff --git a/query/pom.xml b/query/pom.xml
index e779945..eeb2337 100644
--- a/query/pom.xml
+++ b/query/pom.xml
@@ -20,7 +20,7 @@
<parent>
<groupId>org.apache.accumulo</groupId>
<artifactId>accumulo-wikisearch</artifactId>
- <version>1.5.0</version>
+ <version>1.8.0</version>
</parent>
<artifactId>wikisearch-query</artifactId>
<packaging>ejb</packaging>
@@ -51,6 +51,11 @@
<artifactId>commons-collections</artifactId>
</dependency>
<dependency>
+ <groupId>commons-configuration</groupId>
+ <artifactId>commons-configuration</artifactId>
+ <version>1.6</version>
+ </dependency>
+ <dependency>
<groupId>commons-lang</groupId>
<artifactId>commons-lang</artifactId>
</dependency>
@@ -83,21 +88,9 @@
<scope>runtime</scope>
</dependency>
<dependency>
- <groupId>commons-configuration</groupId>
- <artifactId>commons-configuration</artifactId>
- <version>1.6</version>
- <exclusions>
- <exclusion>
- <groupId>commons-logging</groupId>
- <artifactId>commons-logging</artifactId>
- </exclusion>
- </exclusions>
- <scope>runtime</scope>
- </dependency>
- <dependency>
<groupId>commons-io</groupId>
<artifactId>commons-io</artifactId>
- <version>2.1</version>
+ <version>2.4</version>
<scope>runtime</scope>
</dependency>
<dependency>
@@ -142,7 +135,7 @@
<outputDirectory>lib</outputDirectory>
<!-- just grab the non-provided runtime dependencies -->
<!-- XXX we include guava at the same version as hadoop 2 provides so that we have it on hadoop 1 -->
- <includeArtifactIds>commons-io,commons-configuration,commons-lang,commons-codec,protobuf-java,libthrift,zookeeper,hadoop-core,commons-jexl,guava,kryo,asm,minlog,reflectasm,wikisearch-ingest,accumulo-core,accumulo-fate,accumulo-trace</includeArtifactIds>
+ <includeArtifactIds>commons-io,commons-configuration,commons-lang,commons-codec,protobuf-java,libthrift,zookeeper,hadoop-client,commons-jexl,guava,kryo,asm,minlog,reflectasm,wikisearch-ingest,accumulo-core,accumulo-fate,accumulo-trace</includeArtifactIds>
<excludeTransitive>true</excludeTransitive>
</configuration>
</execution>
@@ -178,76 +171,4 @@
</plugin>
</plugins>
</build>
- <profiles>
- <!-- profile for building against Hadoop 1.0.x
- Activate by not specifying hadoop.profile -->
- <profile>
- <id>hadoop-1.0</id>
- <activation>
- <property>
- <name>!hadoop.profile</name>
- </property>
- </activation>
- <dependencies>
- <dependency>
- <groupId>org.apache.hadoop</groupId>
- <artifactId>hadoop-core</artifactId>
- <version>${version.hadoop}</version>
- <exclusions>
- <exclusion>
- <groupId>commons-logging</groupId>
- <artifactId>commons-logging</artifactId>
- </exclusion>
- </exclusions>
- </dependency>
- </dependencies>
- </profile>
- <!-- profile for building against Hadoop 2.0.x
- Activate using: mvn -Dhadoop.profile=2.0 -->
- <profile>
- <id>hadoop-2.0</id>
- <activation>
- <property>
- <name>hadoop.profile</name>
- <value>2.0</value>
- </property>
- </activation>
- <dependencies>
- <dependency>
- <groupId>org.apache.hadoop</groupId>
- <artifactId>hadoop-client</artifactId>
- <version>${version.hadoop}</version>
- <exclusions>
- <exclusion>
- <groupId>org.codehaus.jackson</groupId>
- <artifactId>jackson-mapper-asl</artifactId>
- </exclusion>
- </exclusions>
- </dependency>
- </dependencies>
- <build>
- <plugins>
- <plugin>
- <groupId>org.apache.maven.plugins</groupId>
- <artifactId>maven-dependency-plugin</artifactId>
- <executions>
- <execution>
- <id>copy-dependencies</id>
- <goals>
- <goal>copy-dependencies</goal>
- </goals>
- <phase>prepare-package</phase>
- <configuration>
- <outputDirectory>lib</outputDirectory>
- <!-- just grab the non-provided runtime dependencies -->
- <includeArtifactIds>commons-io,commons-configuration,commons-lang,commons-codec,protobuf-java,libthrift,zookeeper,hadoop-client,hadoop-common,hadoop-hdfs,commons-jexl,guava,kryo,asm,minlog,reflectasm,wikisearch-ingest,accumulo-core,accumulo-fate,accumulo-trace</includeArtifactIds>
- <excludeTransitive>false</excludeTransitive>
- </configuration>
- </execution>
- </executions>
- </plugin>
- </plugins>
- </build>
- </profile>
- </profiles>
</project>
http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/blob/dd23f107/query/src/main/java/org/apache/accumulo/examples/wikisearch/iterator/DefaultIteratorEnvironment.java
----------------------------------------------------------------------
diff --git a/query/src/main/java/org/apache/accumulo/examples/wikisearch/iterator/DefaultIteratorEnvironment.java b/query/src/main/java/org/apache/accumulo/examples/wikisearch/iterator/DefaultIteratorEnvironment.java
index 6b58d08..f05304a 100644
--- a/query/src/main/java/org/apache/accumulo/examples/wikisearch/iterator/DefaultIteratorEnvironment.java
+++ b/query/src/main/java/org/apache/accumulo/examples/wikisearch/iterator/DefaultIteratorEnvironment.java
@@ -18,6 +18,7 @@ package org.apache.accumulo.examples.wikisearch.iterator;
import java.io.IOException;
+import org.apache.accumulo.core.client.sample.SamplerConfiguration;
import org.apache.accumulo.core.conf.AccumuloConfiguration;
import org.apache.accumulo.core.data.Key;
import org.apache.accumulo.core.data.Value;
@@ -25,6 +26,7 @@ import org.apache.accumulo.core.iterators.IteratorEnvironment;
import org.apache.accumulo.core.iterators.IteratorUtil.IteratorScope;
import org.apache.accumulo.core.iterators.SortedKeyValueIterator;
import org.apache.accumulo.core.iterators.system.MapFileIterator;
+import org.apache.accumulo.core.security.Authorizations;
import org.apache.accumulo.core.util.CachedConfiguration;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FileSystem;
@@ -70,5 +72,25 @@ public class DefaultIteratorEnvironment implements IteratorEnvironment {
public void registerSideChannel(SortedKeyValueIterator<Key,Value> iter) {
throw new UnsupportedOperationException();
}
+
+@Override
+public Authorizations getAuthorizations() {
+ throw new UnsupportedOperationException();
+}
+
+@Override
+public IteratorEnvironment cloneWithSamplingEnabled() {
+ throw new UnsupportedOperationException();
+}
+
+@Override
+public boolean isSamplingEnabled() {
+ throw new UnsupportedOperationException();
+}
+
+@Override
+public SamplerConfiguration getSamplerConfiguration() {
+ throw new UnsupportedOperationException();
+}
}
http://git-wip-us.apache.org/repos/asf/accumulo-wikisearch/blob/dd23f107/query/src/main/java/org/apache/accumulo/examples/wikisearch/query/Query.java
----------------------------------------------------------------------
diff --git a/query/src/main/java/org/apache/accumulo/examples/wikisearch/query/Query.java b/query/src/main/java/org/apache/accumulo/examples/wikisearch/query/Query.java
index 024a865..62aa4e2 100644
--- a/query/src/main/java/org/apache/accumulo/examples/wikisearch/query/Query.java
+++ b/query/src/main/java/org/apache/accumulo/examples/wikisearch/query/Query.java
@@ -43,6 +43,7 @@ import javax.xml.transform.stream.StreamSource;
import org.apache.accumulo.core.client.Connector;
import org.apache.accumulo.core.client.Instance;
import org.apache.accumulo.core.client.ZooKeeperInstance;
+import org.apache.accumulo.core.client.security.tokens.PasswordToken;
import org.apache.accumulo.examples.wikisearch.logic.ContentLogic;
import org.apache.accumulo.examples.wikisearch.logic.QueryLogic;
import org.apache.accumulo.examples.wikisearch.sample.Results;
@@ -179,7 +180,7 @@ public class Query implements IQuery {
Instance instance = new ZooKeeperInstance(this.instanceName, this.zooKeepers);
try {
log.info("Connecting to [instanceName = " + this.instanceName + ", zookeepers = " + this.zooKeepers + ", username = " + this.username + "].");
- connector = instance.getConnector(this.username, this.password.getBytes());
+ connector = instance.getConnector(this.username, new PasswordToken(this.password.getBytes()));
} catch (Exception e) {
throw new EJBException("Error getting connector from instance", e);
}
@@ -212,7 +213,7 @@ public class Query implements IQuery {
Instance instance = new ZooKeeperInstance(this.instanceName, this.zooKeepers);
try {
log.info("Connecting to [instanceName = " + this.instanceName + ", zookeepers = " + this.zooKeepers + ", username = " + this.username + "].");
- connector = instance.getConnector(this.username, this.password.getBytes());
+ connector = instance.getConnector(this.username, new PasswordToken(this.password.getBytes()));
} catch (Exception e) {
throw new EJBException("Error getting connector from instance", e);
}