You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hive.apache.org by ha...@apache.org on 2014/06/07 17:49:02 UTC

svn commit: r1601136 - /hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/WriteBuffers.java

Author: hashutosh
Date: Sat Jun  7 15:49:02 2014
New Revision: 1601136

URL: http://svn.apache.org/r1601136
Log:
HIVE-7191 : optimized map join hash table has a bug when it reaches 2Gb (Sergey Shelukhin via Ashutosh Chauhan)

Modified:
    hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/WriteBuffers.java

Modified: hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/WriteBuffers.java
URL: http://svn.apache.org/viewvc/hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/WriteBuffers.java?rev=1601136&r1=1601135&r2=1601136&view=diff
==============================================================================
--- hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/WriteBuffers.java (original)
+++ hive/trunk/serde/src/java/org/apache/hadoop/hive/serde2/WriteBuffers.java Sat Jun  7 15:49:02 2014
@@ -199,7 +199,7 @@ public final class WriteBuffers implemen
 
   private void nextBufferToWrite() {
     if (currentWriteBufferIndex == (writeBuffers.size() - 1)) {
-      if ((1 + writeBuffers.size()) * wbSize > maxSize) {
+      if ((1 + writeBuffers.size()) * ((long)wbSize) > maxSize) {
         // We could verify precisely at write time, but just do approximate at allocation time.
         throw new RuntimeException("Too much memory used by write buffers");
       }
@@ -283,11 +283,11 @@ public final class WriteBuffers implemen
   }
 
   public long getWritePoint() {
-    return (currentWriteBufferIndex * wbSize) + currentWriteOffset;
+    return (currentWriteBufferIndex * (long)wbSize) + currentWriteOffset;
   }
 
   public long getReadPoint() {
-    return (currentReadBufferIndex * wbSize) + currentReadOffset;
+    return (currentReadBufferIndex * (long)wbSize) + currentReadOffset;
   }
 
   public void writeVLong(long value) {