You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by ha...@apache.org on 2014/06/07 17:49:02 UTC
svn commit: r1601136 -
/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/WriteBuffers.java
Author: hashutosh
Date: Sat Jun 7 15:49:02 2014
New Revision: 1601136
URL: http://svn.apache.org/r1601136
Log:
HIVE-7191 : optimized map join hash table has a bug when it reaches 2Gb (Sergey Shelukhin via Ashutosh Chauhan)
Modified:
hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/WriteBuffers.java
Modified: hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/WriteBuffers.java
URL: http://svn.apache.org/viewvc/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/WriteBuffers.java?rev=1601136&r1=1601135&r2=1601136&view=diff
==============================================================================
--- hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/WriteBuffers.java (original)
+++ hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/WriteBuffers.java Sat Jun 7 15:49:02 2014
@@ -199,7 +199,7 @@ public final class WriteBuffers implemen
private void nextBufferToWrite() {
if (currentWriteBufferIndex == (writeBuffers.size() - 1)) {
- if ((1 + writeBuffers.size()) * wbSize > maxSize) {
+ if ((1 + writeBuffers.size()) * ((long)wbSize) > maxSize) {
// We could verify precisely at write time, but just do approximate at allocation time.
throw new RuntimeException("Too much memory used by write buffers");
}
@@ -283,11 +283,11 @@ public final class WriteBuffers implemen
}
public long getWritePoint() {
- return (currentWriteBufferIndex * wbSize) + currentWriteOffset;
+ return (currentWriteBufferIndex * (long)wbSize) + currentWriteOffset;
}
public long getReadPoint() {
- return (currentReadBufferIndex * wbSize) + currentReadOffset;
+ return (currentReadBufferIndex * (long)wbSize) + currentReadOffset;
}
public void writeVLong(long value) {