You are viewing a plain text version of this content. The canonical link for it is here.
Posted to oak-commits@jackrabbit.apache.org by al...@apache.org on 2014/12/16 16:00:52 UTC
svn commit: r1645966 -
/jackrabbit/oak/trunk/oak-core/src/test/java/org/apache/jackrabbit/oak/plugins/segment/CompactionAndCleanupTest.java
Author: alexparvulescu
Date: Tue Dec 16 15:00:52 2014
New Revision: 1645966
URL: http://svn.apache.org/r1645966
Log:
OAK-2192 Concurrent commit during compaction results in mixed segments
- added #close call as test was failing in Win, more debug logs
Modified:
jackrabbit/oak/trunk/oak-core/src/test/java/org/apache/jackrabbit/oak/plugins/segment/CompactionAndCleanupTest.java
Modified: jackrabbit/oak/trunk/oak-core/src/test/java/org/apache/jackrabbit/oak/plugins/segment/CompactionAndCleanupTest.java
URL: http://svn.apache.org/viewvc/jackrabbit/oak/trunk/oak-core/src/test/java/org/apache/jackrabbit/oak/plugins/segment/CompactionAndCleanupTest.java?rev=1645966&r1=1645965&r2=1645966&view=diff
==============================================================================
--- jackrabbit/oak/trunk/oak-core/src/test/java/org/apache/jackrabbit/oak/plugins/segment/CompactionAndCleanupTest.java (original)
+++ jackrabbit/oak/trunk/oak-core/src/test/java/org/apache/jackrabbit/oak/plugins/segment/CompactionAndCleanupTest.java Tue Dec 16 15:00:52 2014
@@ -20,6 +20,7 @@
package org.apache.jackrabbit.oak.plugins.segment;
import static com.google.common.collect.Lists.newArrayList;
+import static org.apache.commons.io.FileUtils.byteCountToDisplaySize;
import static org.apache.commons.io.FileUtils.deleteDirectory;
import static org.apache.jackrabbit.oak.plugins.segment.compaction.CompactionStrategy.CleanupType.CLEAN_NONE;
import static org.apache.jackrabbit.oak.plugins.segment.compaction.CompactionStrategy.CleanupType.CLEAN_OLD;
@@ -55,11 +56,16 @@ import org.apache.jackrabbit.oak.spi.sta
import org.junit.After;
import org.junit.Before;
import org.junit.Test;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
import com.google.common.io.ByteStreams;
public class CompactionAndCleanupTest {
+ private static final Logger log = LoggerFactory
+ .getLogger(CompactionAndCleanupTest.class);
+
private File directory;
@Before
@@ -80,13 +86,14 @@ public class CompactionAndCleanupTest {
FileStore fileStore = new FileStore(directory, 1);
final SegmentNodeStore nodeStore = new SegmentNodeStore(fileStore);
- CompactionStrategy custom = new CompactionStrategy(false,
- false, CLEAN_OLD, TimeUnit.HOURS.toMillis(1), (byte) 0) {
- @Override
- public boolean compacted(@Nonnull Callable<Boolean> setHead) throws Exception {
- return nodeStore.locked(setHead);
- }
- };
+ CompactionStrategy custom = new CompactionStrategy(false, false,
+ CLEAN_OLD, TimeUnit.HOURS.toMillis(1), (byte) 0) {
+ @Override
+ public boolean compacted(@Nonnull Callable<Boolean> setHead)
+ throws Exception {
+ return nodeStore.locked(setHead);
+ }
+ };
fileStore.setCompactionStrategy(custom);
// 1a. Create a bunch of data
@@ -102,89 +109,85 @@ public class CompactionAndCleanupTest {
// ----
final long dataSize = fileStore.size();
- // System.out.printf("File store dataSize %s%n",
- // byteCountToDisplaySize(dataSize));
+ log.debug("File store dataSize {}", byteCountToDisplaySize(dataSize));
- // 1. Create a property with 5 MB blob
- NodeBuilder builder = nodeStore.getRoot().builder();
- builder.setProperty("a1", createBlob(nodeStore, blobSize));
- builder.setProperty("b", "foo");
- nodeStore.merge(builder, EmptyHook.INSTANCE, CommitInfo.EMPTY);
-
- // System.out.printf("File store pre removal %s expecting %s %n",
- // byteCountToDisplaySize(fileStore.size()),
- // byteCountToDisplaySize(blobSize + dataSize));
- assertEquals(mb(blobSize + dataSize), mb(fileStore.size()));
-
- // 2. Now remove the property
- builder = nodeStore.getRoot().builder();
- builder.removeProperty("a1");
- nodeStore.merge(builder, EmptyHook.INSTANCE, CommitInfo.EMPTY);
-
- // Size remains same, no cleanup happened yet
- // System.out.printf("File store pre compaction %s expecting %s%n",
- // byteCountToDisplaySize(fileStore.size()),
- // byteCountToDisplaySize(blobSize + dataSize));
- assertEquals(mb(blobSize + dataSize), mb(fileStore.size()));
-
- // 3. Compact
- assertTrue(fileStore.maybeCompact(false));
-
- // Size doesn't shrink: ran compaction with a '1 Hour' cleanup
- // strategy
- // System.out.printf("File store post compaction %s expecting %s%n",
- // byteCountToDisplaySize(fileStore.size()),
- // byteCountToDisplaySize(blobSize + dataSize));
- assertSize("post compaction", fileStore.size(), blobSize + dataSize,
- blobSize + 2 * dataSize);
-
- // 4. Add some more property to flush the current TarWriter
- builder = nodeStore.getRoot().builder();
- builder.setProperty("a2", createBlob(nodeStore, blobSize));
- nodeStore.merge(builder, EmptyHook.INSTANCE, CommitInfo.EMPTY);
-
- // Size is double
- // System.out.printf("File store pre cleanup %s expecting %s%n",
- // byteCountToDisplaySize(fileStore.size()),
- // byteCountToDisplaySize(2 * blobSize + dataSize));
- assertSize("post compaction", fileStore.size(),
- 2 * blobSize + dataSize, 2 * blobSize + 2 * dataSize);
-
- // 5. Cleanup, expecting store size:
- // no data content =>
- // fileStore.size() == blobSize
- // some data content =>
- // fileStore.size() in [blobSize + dataSize, blobSize + 2xdataSize]
- assertTrue(fileStore.maybeCompact(false));
- fileStore.cleanup();
- // System.out.printf(
- // "File store post cleanup %s expecting between [%s,%s]%n",
- // byteCountToDisplaySize(fileStore.size()),
- // byteCountToDisplaySize(blobSize + dataSize),
- // byteCountToDisplaySize(blobSize + 2 * dataSize));
- assertSize("post cleanup", fileStore.size(),
- blobSize + dataSize, blobSize + 2 * dataSize);
-
- // refresh the ts ref, to simulate a long wait time
- custom.setOlderThan(0);
- TimeUnit.MILLISECONDS.sleep(5);
-
- boolean needsCompaction = true;
- for (int i = 0; i < 3 && needsCompaction; i++) {
- needsCompaction = fileStore.maybeCompact(false);
+ try {
+ // 1. Create a property with 5 MB blob
+ NodeBuilder builder = nodeStore.getRoot().builder();
+ builder.setProperty("a1", createBlob(nodeStore, blobSize));
+ builder.setProperty("b", "foo");
+ nodeStore.merge(builder, EmptyHook.INSTANCE, CommitInfo.EMPTY);
+
+ log.debug("File store pre removal {}, expecting {}",
+ byteCountToDisplaySize(fileStore.size()),
+ byteCountToDisplaySize(blobSize + dataSize));
+ assertEquals(mb(blobSize + dataSize), mb(fileStore.size()));
+
+ // 2. Now remove the property
+ builder = nodeStore.getRoot().builder();
+ builder.removeProperty("a1");
+ nodeStore.merge(builder, EmptyHook.INSTANCE, CommitInfo.EMPTY);
+
+ // Size remains same, no cleanup happened yet
+ log.debug("File store pre compaction {}, expecting {}",
+ byteCountToDisplaySize(fileStore.size()),
+ byteCountToDisplaySize(blobSize + dataSize));
+ assertEquals(mb(blobSize + dataSize), mb(fileStore.size()));
+
+ // 3. Compact
+ assertTrue(fileStore.maybeCompact(false));
+
+ // Size doesn't shrink: ran compaction with a '1 Hour' cleanup
+ // strategy
+ assertSize("post compaction", fileStore.size(),
+ blobSize + dataSize, blobSize + 2 * dataSize);
+
+ // 4. Add some more property to flush the current TarWriter
+ builder = nodeStore.getRoot().builder();
+ builder.setProperty("a2", createBlob(nodeStore, blobSize));
+ nodeStore.merge(builder, EmptyHook.INSTANCE, CommitInfo.EMPTY);
+
+ // Size is double
+ assertSize("pre cleanup", fileStore.size(), 2 * blobSize
+ + dataSize, 2 * blobSize + 2 * dataSize);
+
+ // 5. Cleanup, expecting store size:
+ // no data content =>
+ // fileStore.size() == blobSize
+ // some data content =>
+ // fileStore.size() in [blobSize + dataSize, blobSize + 2xdataSize]
+ assertTrue(fileStore.maybeCompact(false));
fileStore.cleanup();
- }
+ assertSize("post cleanup", fileStore.size(), blobSize + dataSize,
+ blobSize + 2 * dataSize);
- // gain is finally 0%
- assertFalse(fileStore.maybeCompact(false));
+ // refresh the ts ref, to simulate a long wait time
+ custom.setOlderThan(0);
+ TimeUnit.MILLISECONDS.sleep(5);
+
+ boolean needsCompaction = true;
+ for (int i = 0; i < 3 && needsCompaction; i++) {
+ needsCompaction = fileStore.maybeCompact(false);
+ fileStore.cleanup();
+ }
+
+ // gain is finally 0%
+ assertFalse(fileStore.maybeCompact(false));
- // no data loss happened
- byte[] blob = ByteStreams.toByteArray(nodeStore.getRoot()
- .getProperty("a2").getValue(Type.BINARY).getNewStream());
- assertEquals(blobSize, blob.length);
+ // no data loss happened
+ byte[] blob = ByteStreams.toByteArray(nodeStore.getRoot()
+ .getProperty("a2").getValue(Type.BINARY).getNewStream());
+ assertEquals(blobSize, blob.length);
+ } finally {
+ fileStore.close();
+ }
}
- private static void assertSize(String log, long size, long lower, long upper) {
+ private static void assertSize(String info, long size, long lower,
+ long upper) {
+ log.debug("File Store {} size {}, expected in interval [{},{}]", info,
+ byteCountToDisplaySize(size), byteCountToDisplaySize(lower),
+ byteCountToDisplaySize(upper));
assertTrue("File Store " + log + " size expected in interval ["
+ mb(lower) + "," + mb(upper) + "] but was: " + mb(size),
mb(size) >= mb(lower) && mb(size) <= mb(upper));