You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@accumulo.apache.org by el...@apache.org on 2014/03/26 01:35:18 UTC
[04/11] git commit: ACCUMULO-2532 Relax access on cleanupIndexOp and
add tests that show failure case.
ACCUMULO-2532 Relax access on cleanupIndexOp and add tests that show failure case.
Project: http://git-wip-us.apache.org/repos/asf/accumulo/repo
Commit: http://git-wip-us.apache.org/repos/asf/accumulo/commit/6420dd3d
Tree: http://git-wip-us.apache.org/repos/asf/accumulo/tree/6420dd3d
Diff: http://git-wip-us.apache.org/repos/asf/accumulo/diff/6420dd3d
Branch: refs/heads/master
Commit: 6420dd3d49c97ee5c9a5e173914ca5374a26c8e5
Parents: ce38eea
Author: Josh Elser <el...@apache.org>
Authored: Tue Mar 25 15:27:43 2014 -0700
Committer: Josh Elser <el...@apache.org>
Committed: Tue Mar 25 16:08:33 2014 -0700
----------------------------------------------------------------------
.../apache/accumulo/server/util/FileUtil.java | 2 +-
.../accumulo/server/util/FileUtilTest.java | 162 +++++++++++++++++++
2 files changed, 163 insertions(+), 1 deletion(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6420dd3d/server/base/src/main/java/org/apache/accumulo/server/util/FileUtil.java
----------------------------------------------------------------------
diff --git a/server/base/src/main/java/org/apache/accumulo/server/util/FileUtil.java b/server/base/src/main/java/org/apache/accumulo/server/util/FileUtil.java
index b1e39f0..fa186aa 100644
--- a/server/base/src/main/java/org/apache/accumulo/server/util/FileUtil.java
+++ b/server/base/src/main/java/org/apache/accumulo/server/util/FileUtil.java
@@ -360,7 +360,7 @@ public class FileUtil {
}
}
- private static void cleanupIndexOp(AccumuloConfiguration acuConf, Path tmpDir, VolumeManager fs, ArrayList<FileSKVIterator> readers) throws IOException {
+ protected static void cleanupIndexOp(AccumuloConfiguration acuConf, Path tmpDir, VolumeManager fs, ArrayList<FileSKVIterator> readers) throws IOException {
// close all of the index sequence files
for (FileSKVIterator r : readers) {
try {
http://git-wip-us.apache.org/repos/asf/accumulo/blob/6420dd3d/server/base/src/test/java/org/apache/accumulo/server/util/FileUtilTest.java
----------------------------------------------------------------------
diff --git a/server/base/src/test/java/org/apache/accumulo/server/util/FileUtilTest.java b/server/base/src/test/java/org/apache/accumulo/server/util/FileUtilTest.java
new file mode 100644
index 0000000..546b4ee
--- /dev/null
+++ b/server/base/src/test/java/org/apache/accumulo/server/util/FileUtilTest.java
@@ -0,0 +1,162 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements. See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package org.apache.accumulo.server.util;
+
+import java.io.File;
+import java.io.IOException;
+import java.util.ArrayList;
+import java.util.HashMap;
+import java.util.Map;
+
+import org.apache.accumulo.core.conf.AccumuloConfiguration;
+import org.apache.accumulo.core.conf.DefaultConfiguration;
+import org.apache.accumulo.core.conf.Property;
+import org.apache.accumulo.core.file.FileSKVIterator;
+import org.apache.accumulo.server.fs.VolumeManager;
+import org.apache.accumulo.server.fs.VolumeManagerImpl;
+import org.apache.commons.io.FileUtils;
+import org.apache.hadoop.fs.Path;
+import org.junit.Assert;
+import org.junit.Test;
+
+import com.google.common.io.Files;
+
+/**
+ *
+ */
+public class FileUtilTest {
+
+ @SuppressWarnings("deprecation")
+ @Test
+ public void testCleanupIndexOpWithDfsDir() throws IOException {
+ File dfsDir = Files.createTempDir();
+
+ try {
+ // And a "unique" tmp directory for each volume
+ File tmp1 = new File(dfsDir, "tmp");
+ tmp1.mkdirs();
+ Path tmpPath1 = new Path(tmp1.toURI());
+
+ HashMap<Property,String> testProps = new HashMap<Property,String>();
+ testProps.put(Property.INSTANCE_DFS_DIR, dfsDir.getAbsolutePath());
+
+ AccumuloConfiguration testConf = new FileUtilTestConfiguration(testProps);
+ VolumeManager fs = VolumeManagerImpl.getLocal(dfsDir.getAbsolutePath());
+
+ FileUtil.cleanupIndexOp(testConf, tmpPath1, fs, new ArrayList<FileSKVIterator>());
+
+ Assert.assertFalse("Expected " + tmp1 + " to be cleaned up but it wasn't", tmp1.exists());
+ } finally {
+ FileUtils.deleteQuietly(dfsDir);
+ }
+ }
+
+ @Test
+ public void testCleanupIndexOpWithCommonParentVolume() throws IOException {
+ File accumuloDir = Files.createTempDir();
+
+ try {
+ File volumeDir = new File(accumuloDir, "volumes");
+ volumeDir.mkdirs();
+
+ // Make some directories to simulate multiple volumes
+ File v1 = new File(volumeDir, "v1"), v2 = new File(volumeDir, "v2");
+ v1.mkdirs();
+ v2.mkdirs();
+
+ // And a "unique" tmp directory for each volume
+ File tmp1 = new File(v1, "tmp"), tmp2 = new File(v2, "tmp");
+ tmp1.mkdirs();
+ tmp2.mkdirs();
+ Path tmpPath1 = new Path(tmp1.toURI()), tmpPath2 = new Path(tmp2.toURI());
+
+ HashMap<Property,String> testProps = new HashMap<Property,String>();
+ testProps.put(Property.INSTANCE_VOLUMES, v1.toURI().toString() + "," + v2.toURI().toString());
+
+ AccumuloConfiguration testConf = new FileUtilTestConfiguration(testProps);
+ VolumeManager fs = VolumeManagerImpl.getLocal(accumuloDir.getAbsolutePath());
+
+ FileUtil.cleanupIndexOp(testConf, tmpPath1, fs, new ArrayList<FileSKVIterator>());
+
+ Assert.assertFalse("Expected " + tmp1 + " to be cleaned up but it wasn't", tmp1.exists());
+
+ FileUtil.cleanupIndexOp(testConf, tmpPath2, fs, new ArrayList<FileSKVIterator>());
+
+ Assert.assertFalse("Expected " + tmp2 + " to be cleaned up but it wasn't", tmp2.exists());
+ } finally {
+ FileUtils.deleteQuietly(accumuloDir);
+ }
+ }
+
+ @Test
+ public void testCleanupIndexOpWithoutCommonParentVolume() throws IOException {
+ File accumuloDir = Files.createTempDir();
+
+ try {
+ // Make some directories to simulate multiple volumes
+ File v1 = new File(accumuloDir, "v1"), v2 = new File(accumuloDir, "v2");
+ v1.mkdirs();
+ v2.mkdirs();
+
+ // And a "unique" tmp directory for each volume
+ File tmp1 = new File(v1, "tmp"), tmp2 = new File(v2, "tmp");
+ tmp1.mkdirs();
+ tmp2.mkdirs();
+ Path tmpPath1 = new Path(tmp1.toURI()), tmpPath2 = new Path(tmp2.toURI());
+
+ HashMap<Property,String> testProps = new HashMap<Property,String>();
+ testProps.put(Property.INSTANCE_VOLUMES, v1.toURI().toString() + "," + v2.toURI().toString());
+
+ AccumuloConfiguration testConf = new FileUtilTestConfiguration(testProps);
+ VolumeManager fs = VolumeManagerImpl.getLocal(accumuloDir.getAbsolutePath());
+
+ FileUtil.cleanupIndexOp(testConf, tmpPath1, fs, new ArrayList<FileSKVIterator>());
+
+ Assert.assertFalse("Expected " + tmp1 + " to be cleaned up but it wasn't", tmp1.exists());
+
+ FileUtil.cleanupIndexOp(testConf, tmpPath2, fs, new ArrayList<FileSKVIterator>());
+
+ Assert.assertFalse("Expected " + tmp2 + " to be cleaned up but it wasn't", tmp2.exists());
+ } finally {
+ FileUtils.deleteQuietly(accumuloDir);
+ }
+ }
+
+ private static class FileUtilTestConfiguration extends AccumuloConfiguration {
+ private DefaultConfiguration defaultConf = new DefaultConfiguration();
+ private Map<Property,String> properties;
+
+ public FileUtilTestConfiguration(Map<Property,String> properties) {
+ this.properties = properties;
+ }
+
+ @Override
+ public String get(Property property) {
+ String value = properties.get(property);
+ if (null != value) {
+ return value;
+ }
+ return defaultConf.get(property);
+ }
+
+ @Override
+ public void getProperties(Map<String,String> props, PropertyFilter filter) {
+ throw new UnsupportedOperationException();
+ }
+
+ }
+}