You are viewing a plain text version of this content. The canonical link for it is here.
Posted to hdfs-commits@hadoop.apache.org by sz...@apache.org on 2009/07/31 20:40:37 UTC
svn commit: r799688 - in /hadoop/hdfs/trunk: CHANGES.txt
src/java/org/apache/hadoop/hdfs/DFSClient.java
src/java/org/apache/hadoop/hdfs/server/datanode/BlockReceiver.java
Author: szetszwo
Date: Fri Jul 31 18:40:37 2009
New Revision: 799688
URL: http://svn.apache.org/viewvc?rev=799688&view=rev
Log:
HDFS-496. Use PureJavaCrc32 in HDFS. Contributed by Todd Lipcon
Modified:
hadoop/hdfs/trunk/CHANGES.txt
hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/DFSClient.java
hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/server/datanode/BlockReceiver.java
Modified: hadoop/hdfs/trunk/CHANGES.txt
URL: http://svn.apache.org/viewvc/hadoop/hdfs/trunk/CHANGES.txt?rev=799688&r1=799687&r2=799688&view=diff
==============================================================================
--- hadoop/hdfs/trunk/CHANGES.txt (original)
+++ hadoop/hdfs/trunk/CHANGES.txt Fri Jul 31 18:40:37 2009
@@ -63,6 +63,8 @@
HDFS-514. Change DFSClient.namenode from public to private. (Bill Zeller
via szetszwo)
+ HDFS-496. Use PureJavaCrc32 in HDFS. (Todd Lipcon via szetszwo)
+
BUG FIXES
HDFS-76. Better error message to users when commands fail because of
lack of quota. Allow quota to be set even if the limit is lower than
Modified: hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/DFSClient.java
URL: http://svn.apache.org/viewvc/hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/DFSClient.java?rev=799688&r1=799687&r2=799688&view=diff
==============================================================================
--- hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/DFSClient.java (original)
+++ hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/DFSClient.java Fri Jul 31 18:40:37 2009
@@ -50,7 +50,6 @@
import java.util.concurrent.ConcurrentHashMap;
import java.util.concurrent.TimeUnit;
import java.util.concurrent.atomic.AtomicBoolean;
-import java.util.zip.CRC32;
import javax.net.SocketFactory;
import javax.security.auth.login.LoginException;
@@ -110,6 +109,7 @@
import org.apache.hadoop.util.Daemon;
import org.apache.hadoop.util.DataChecksum;
import org.apache.hadoop.util.Progressable;
+import org.apache.hadoop.util.PureJavaCrc32;
import org.apache.hadoop.util.StringUtils;
/********************************************************
@@ -2983,7 +2983,7 @@
private DFSOutputStream(String src, long blockSize, Progressable progress,
int bytesPerChecksum) throws IOException {
- super(new CRC32(), bytesPerChecksum, 4);
+ super(new PureJavaCrc32(), bytesPerChecksum, 4);
this.src = src;
this.blockSize = blockSize;
this.progress = progress;
Modified: hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/server/datanode/BlockReceiver.java
URL: http://svn.apache.org/viewvc/hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/server/datanode/BlockReceiver.java?rev=799688&r1=799687&r2=799688&view=diff
==============================================================================
--- hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/server/datanode/BlockReceiver.java (original)
+++ hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/server/datanode/BlockReceiver.java Fri Jul 31 18:40:37 2009
@@ -29,7 +29,6 @@
import java.io.OutputStream;
import java.nio.ByteBuffer;
import java.util.LinkedList;
-import java.util.zip.CRC32;
import java.util.zip.Checksum;
import org.apache.commons.logging.Log;
@@ -44,6 +43,7 @@
import org.apache.hadoop.io.IOUtils;
import org.apache.hadoop.util.Daemon;
import org.apache.hadoop.util.DataChecksum;
+import org.apache.hadoop.util.PureJavaCrc32;
import org.apache.hadoop.util.StringUtils;
/** A class that receives a block and writes to its own disk, meanwhile
@@ -676,7 +676,7 @@
}
// compute crc of partial chunk from data read in the block file.
- partialCrc = new CRC32();
+ partialCrc = new PureJavaCrc32();
partialCrc.update(buf, 0, sizePartialChunk);
LOG.info("Read in partial CRC chunk from disk for block " + block);