You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-commits@hadoop.apache.org by cd...@apache.org on 2010/02/19 08:48:15 UTC
svn commit: r911729 - in /hadoop/common/branches/branch-0.20: CHANGES.txt
src/docs/releasenotes.html
Author: cdouglas
Date: Fri Feb 19 07:48:15 2010
New Revision: 911729
URL: http://svn.apache.org/viewvc?rev=911729&view=rev
Log:
Prepare for 0.20.2-rc4
Modified:
hadoop/common/branches/branch-0.20/CHANGES.txt
hadoop/common/branches/branch-0.20/src/docs/releasenotes.html
Modified: hadoop/common/branches/branch-0.20/CHANGES.txt
URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-0.20/CHANGES.txt?rev=911729&r1=911728&r2=911729&view=diff
==============================================================================
--- hadoop/common/branches/branch-0.20/CHANGES.txt (original)
+++ hadoop/common/branches/branch-0.20/CHANGES.txt Fri Feb 19 07:48:15 2010
@@ -2,7 +2,7 @@
Release 0.20.3 - Unreleased
-Release 0.20.2 - 2010-2-18
+Release 0.20.2 - 2010-2-19
NEW FEATURES
Modified: hadoop/common/branches/branch-0.20/src/docs/releasenotes.html
URL: http://svn.apache.org/viewvc/hadoop/common/branches/branch-0.20/src/docs/releasenotes.html?rev=911729&r1=911728&r2=911729&view=diff
==============================================================================
--- hadoop/common/branches/branch-0.20/src/docs/releasenotes.html (original)
+++ hadoop/common/branches/branch-0.20/src/docs/releasenotes.html Fri Feb 19 07:48:15 2010
@@ -17,18 +17,20 @@
<a name="changes"></a>
<h2>Changes Since Hadoop 0.20.1</h2>
-<h3>Common</h3>
+<h3>Common</h3>
<h4> Bug
</h4>
<ul>
<li>[<a href='https://issues.apache.org/jira/browse/HADOOP-4802'>HADOOP-4802</a>] - RPC Server send buffer retains size of largest response ever sent
</li>
+<li>[<a href='https://issues.apache.org/jira/browse/HADOOP-5611'>HADOOP-5611</a>] - C++ libraries do not build on Debian Lenny
+</li>
+<li>[<a href='https://issues.apache.org/jira/browse/HADOOP-5612'>HADOOP-5612</a>] - Some c++ scripts are not chmodded before ant execution
+</li>
<li>[<a href='https://issues.apache.org/jira/browse/HADOOP-5623'>HADOOP-5623</a>] - Streaming: process provided status messages are overwritten every 10 seoncds
</li>
<li>[<a href='https://issues.apache.org/jira/browse/HADOOP-5759'>HADOOP-5759</a>] - IllegalArgumentException when CombineFileInputFormat is used as job InputFormat
</li>
-<li>[<a href='https://issues.apache.org/jira/browse/HADOOP-5997'>HADOOP-5997</a>] - Many test jobs write to HDFS under /
-</li>
<li>[<a href='https://issues.apache.org/jira/browse/HADOOP-6097'>HADOOP-6097</a>] - Multiple bugs w/ Hadoop archives
</li>
<li>[<a href='https://issues.apache.org/jira/browse/HADOOP-6231'>HADOOP-6231</a>] - Allow caching of filesystem instances to be disabled on a per-instance basis
@@ -39,39 +41,22 @@
</li>
<li>[<a href='https://issues.apache.org/jira/browse/HADOOP-6386'>HADOOP-6386</a>] - NameNode's HttpServer can't instantiate InetSocketAddress: IllegalArgumentException is thrown
</li>
-<li>[<a href='https://issues.apache.org/jira/browse/HADOOP-6417'>HADOOP-6417</a>] - Alternative Java Distributions in the Hadoop Documention
-</li>
<li>[<a href='https://issues.apache.org/jira/browse/HADOOP-6428'>HADOOP-6428</a>] - HttpServer sleeps with negative values
</li>
-<li>[<a href='https://issues.apache.org/jira/browse/HADOOP-6453'>HADOOP-6453</a>] - Hadoop wrapper script shouldn't ignore an existing JAVA_LIBRARY_PATH
-</li>
<li>[<a href='https://issues.apache.org/jira/browse/HADOOP-6460'>HADOOP-6460</a>] - Namenode runs of out of memory due to memory leak in ipc Server
</li>
<li>[<a href='https://issues.apache.org/jira/browse/HADOOP-6498'>HADOOP-6498</a>] - IPC client bug may cause rpc call hang
</li>
-<li>[<a href='https://issues.apache.org/jira/browse/HADOOP-6502'>HADOOP-6502</a>] - DistributedFileSystem#listStatus is very slow when listing a directory with a size of 1300
-</li>
<li>[<a href='https://issues.apache.org/jira/browse/HADOOP-6506'>HADOOP-6506</a>] - Failing tests prevent the rest of test targets from execution.
</li>
<li>[<a href='https://issues.apache.org/jira/browse/HADOOP-6524'>HADOOP-6524</a>] - Contrib tests are failing Clover'ed build
</li>
-</ul>
-
-<h4> Improvement
-</h4>
-<ul>
-<li>[<a href='https://issues.apache.org/jira/browse/HADOOP-3659'>HADOOP-3659</a>] - Patch to allow hadoop native to compile on Mac OS X
-</li>
-<li>[<a href='https://issues.apache.org/jira/browse/HADOOP-6304'>HADOOP-6304</a>] - Use java.io.File.set{Readable|Writable|Executable} where possible in RawLocalFileSystem
-</li>
-<li>[<a href='https://issues.apache.org/jira/browse/HADOOP-6376'>HADOOP-6376</a>] - slaves file to have a header specifying the format of conf/slaves file
+<li>[<a href='https://issues.apache.org/jira/browse/HADOOP-6575'>HADOOP-6575</a>] - Tests do not run on 0.20 branch
</li>
-<li>[<a href='https://issues.apache.org/jira/browse/HADOOP-6475'>HADOOP-6475</a>] - Improvements to the hadoop-config script
-</li>
-<li>[<a href='https://issues.apache.org/jira/browse/HADOOP-6542'>HADOOP-6542</a>] - Add a -Dno-docs option to build.xml
+<li>[<a href='https://issues.apache.org/jira/browse/HADOOP-6576'>HADOOP-6576</a>] - TestStreamingStatus is failing on 0.20 branch
</li>
</ul>
-
+
<h4> Task
</h4>
<ul>
@@ -79,6 +64,7 @@
</li>
</ul>
+
<h3>HDFS</h3>
<h4> Bug
</h4>
@@ -89,26 +75,16 @@
</li>
<li>[<a href='https://issues.apache.org/jira/browse/HDFS-187'>HDFS-187</a>] - TestStartup fails if hdfs is running in the same machine
</li>
-<li>[<a href='https://issues.apache.org/jira/browse/HDFS-442'>HDFS-442</a>] - dfsthroughput in test.jar throws NPE
-</li>
<li>[<a href='https://issues.apache.org/jira/browse/HDFS-495'>HDFS-495</a>] - Hadoop FSNamesystem startFileInternal() getLease() has bug
</li>
<li>[<a href='https://issues.apache.org/jira/browse/HDFS-579'>HDFS-579</a>] - HADOOP-3792 update of DfsTask incomplete
</li>
<li>[<a href='https://issues.apache.org/jira/browse/HDFS-596'>HDFS-596</a>] - Memory leak in libhdfs: hdfsFreeFileInfo() in libhdfs does not free memory for mOwner and mGroup
</li>
-<li>[<a href='https://issues.apache.org/jira/browse/HDFS-645'>HDFS-645</a>] - Namenode does not leave safe mode even if all blocks are available
-</li>
-<li>[<a href='https://issues.apache.org/jira/browse/HDFS-667'>HDFS-667</a>] - test-contrib target fails on hdfsproxy tests
-</li>
<li>[<a href='https://issues.apache.org/jira/browse/HDFS-677'>HDFS-677</a>] - Rename failure due to quota results in deletion of src directory
</li>
-<li>[<a href='https://issues.apache.org/jira/browse/HDFS-686'>HDFS-686</a>] - NullPointerException is thrown while merging edit log and image
-</li>
<li>[<a href='https://issues.apache.org/jira/browse/HDFS-723'>HDFS-723</a>] - Deadlock in DFSClient#DFSOutputStream
</li>
-<li>[<a href='https://issues.apache.org/jira/browse/HDFS-727'>HDFS-727</a>] - bug setting block size hdfsOpenFile
-</li>
<li>[<a href='https://issues.apache.org/jira/browse/HDFS-732'>HDFS-732</a>] - HDFS files are ending up truncated
</li>
<li>[<a href='https://issues.apache.org/jira/browse/HDFS-734'>HDFS-734</a>] - TestDatanodeBlockScanner times out in branch 0.20
@@ -123,29 +99,12 @@
</li>
<li>[<a href='https://issues.apache.org/jira/browse/HDFS-795'>HDFS-795</a>] - DFS Write pipeline does not detect defective datanode correctly in some cases (HADOOP-3339)
</li>
-<li>[<a href='https://issues.apache.org/jira/browse/HDFS-846'>HDFS-846</a>] - SetSpaceQuota of value 9223372036854775807 does not apply quota.
-</li>
<li>[<a href='https://issues.apache.org/jira/browse/HDFS-872'>HDFS-872</a>] - DFSClient 0.20.1 is incompatible with HDFS 0.20.2
</li>
-<li>[<a href='https://issues.apache.org/jira/browse/HDFS-886'>HDFS-886</a>] - TestHDFSTrash fails on Windows
-</li>
-<li>[<a href='https://issues.apache.org/jira/browse/HDFS-920'>HDFS-920</a>] - Incorrect metrics reporting of transcations metrics.
-</li>
<li>[<a href='https://issues.apache.org/jira/browse/HDFS-927'>HDFS-927</a>] - DFSInputStream retries too many times for new block locations
</li>
-<li>[<a href='https://issues.apache.org/jira/browse/HDFS-937'>HDFS-937</a>] - Port HDFS-101 to branch 0.20
-</li>
-<li>[<a href='https://issues.apache.org/jira/browse/HDFS-961'>HDFS-961</a>] - dfs_readdir incorrectly parses paths
-</li>
-</ul>
-
-<h4> Improvement
-</h4>
-<ul>
-<li>[<a href='https://issues.apache.org/jira/browse/HDFS-959'>HDFS-959</a>] - Performance improvements to DFSClient and DataNode for faster DFS write at replication factor of 1
-</li>
</ul>
-
+
<h4> Test
</h4>
<ul>
@@ -157,70 +116,46 @@
</li>
</ul>
-<h3>MapReduce</h3>
+<h3>MapReduce</h3>
<h4> Bug
</h4>
<ul>
<li>[<a href='https://issues.apache.org/jira/browse/MAPREDUCE-112'>MAPREDUCE-112</a>] - Reduce Input Records and Reduce Output Records counters are not being set when using the new Mapreduce reducer API
</li>
-<li>[<a href='https://issues.apache.org/jira/browse/MAPREDUCE-118'>MAPREDUCE-118</a>] - Job.getJobID() will always return null
-</li>
<li>[<a href='https://issues.apache.org/jira/browse/MAPREDUCE-433'>MAPREDUCE-433</a>] - TestReduceFetch failed.
</li>
-<li>[<a href='https://issues.apache.org/jira/browse/MAPREDUCE-806'>MAPREDUCE-806</a>] - WordCount example does not compile given the current instructions
-</li>
<li>[<a href='https://issues.apache.org/jira/browse/MAPREDUCE-826'>MAPREDUCE-826</a>] - harchive doesn't use ToolRunner / harchive returns 0 even if the job fails with exception
</li>
<li>[<a href='https://issues.apache.org/jira/browse/MAPREDUCE-979'>MAPREDUCE-979</a>] - JobConf.getMemoryFor{Map|Reduce}Task doesn't fallback to newer config knobs when mapred.taskmaxvmem is set to DISABLED_MEMORY_LIMIT of -1
</li>
<li>[<a href='https://issues.apache.org/jira/browse/MAPREDUCE-1010'>MAPREDUCE-1010</a>] - Adding tests for changes in archives.
</li>
-<li>[<a href='https://issues.apache.org/jira/browse/MAPREDUCE-1057'>MAPREDUCE-1057</a>] - java tasks are not honouring the value of mapred.userlog.limit.kb
-</li>
<li>[<a href='https://issues.apache.org/jira/browse/MAPREDUCE-1068'>MAPREDUCE-1068</a>] - In hadoop-0.20.0 streaming job do not throw proper verbose error message if file is not present
</li>
<li>[<a href='https://issues.apache.org/jira/browse/MAPREDUCE-1070'>MAPREDUCE-1070</a>] - Deadlock in FairSchedulerServlet
</li>
-<li>[<a href='https://issues.apache.org/jira/browse/MAPREDUCE-1088'>MAPREDUCE-1088</a>] - JobHistory files should have narrower 0600 perms
-</li>
<li>[<a href='https://issues.apache.org/jira/browse/MAPREDUCE-1112'>MAPREDUCE-1112</a>] - Fix CombineFileInputFormat for hadoop 0.20
</li>
<li>[<a href='https://issues.apache.org/jira/browse/MAPREDUCE-1147'>MAPREDUCE-1147</a>] - Map output records counter missing for map-only jobs in new API
</li>
-<li>[<a href='https://issues.apache.org/jira/browse/MAPREDUCE-1157'>MAPREDUCE-1157</a>] - JT UI shows incorrect scheduling info for failed/killed retired jobs
-</li>
<li>[<a href='https://issues.apache.org/jira/browse/MAPREDUCE-1163'>MAPREDUCE-1163</a>] - hdfsJniHelper.h: Yahoo! specific paths are encoded
</li>
<li>[<a href='https://issues.apache.org/jira/browse/MAPREDUCE-1182'>MAPREDUCE-1182</a>] - Reducers fail with OutOfMemoryError while copying Map outputs
</li>
-<li>[<a href='https://issues.apache.org/jira/browse/MAPREDUCE-1262'>MAPREDUCE-1262</a>] - Eclipse Plugin does not build for Hadoop 0.20.1
-</li>
-<li>[<a href='https://issues.apache.org/jira/browse/MAPREDUCE-1264'>MAPREDUCE-1264</a>] - Error Recovery failed, task will continue but run forever as new data only comes in very very slowly
-</li>
-<li>[<a href='https://issues.apache.org/jira/browse/MAPREDUCE-1321'>MAPREDUCE-1321</a>] - Spurios logs with org.apache.hadoop.util.DiskChecker$DiskErrorException in TaskTracker
+<li>[<a href='https://issues.apache.org/jira/browse/MAPREDUCE-1251'>MAPREDUCE-1251</a>] - c++ utils doesn't compile
</li>
<li>[<a href='https://issues.apache.org/jira/browse/MAPREDUCE-1328'>MAPREDUCE-1328</a>] - contrib/index - modify build / ivy files as appropriate
</li>
-<li>[<a href='https://issues.apache.org/jira/browse/MAPREDUCE-1346'>MAPREDUCE-1346</a>] - TestStreamingExitStatus / TestStreamingKeyValue - correct text fixtures in place
-</li>
-<li>[<a href='https://issues.apache.org/jira/browse/MAPREDUCE-1381'>MAPREDUCE-1381</a>] - Incorrect values being displayed for blacklisted_maps and blacklisted_reduces
-</li>
-<li>[<a href='https://issues.apache.org/jira/browse/MAPREDUCE-1389'>MAPREDUCE-1389</a>] - TestDFSIO creates TestDFSIO_results.log file directly under hadoop.home
-</li>
-<li>[<a href='https://issues.apache.org/jira/browse/MAPREDUCE-1397'>MAPREDUCE-1397</a>] - NullPointerException observed during task failures
-</li>
</ul>
-
+
<h4> Improvement
</h4>
<ul>
-<li>[<a href='https://issues.apache.org/jira/browse/MAPREDUCE-1315'>MAPREDUCE-1315</a>] - taskdetails.jsp and jobfailures.jsp should have consistent convention for machine names in case of lost task tracker
-</li>
-<li>[<a href='https://issues.apache.org/jira/browse/MAPREDUCE-1361'>MAPREDUCE-1361</a>] - In the pools with minimum slots, new job will always receive slots even if the minimum slots limit has been fulfilled
+<li>[<a href='https://issues.apache.org/jira/browse/MAPREDUCE-623'>MAPREDUCE-623</a>] - Resolve javac warnings in mapred
</li>
</ul>
-
+
<h4> New Feature
</h4>
<ul>
@@ -229,14 +164,8 @@
<li>[<a href='https://issues.apache.org/jira/browse/MAPREDUCE-1170'>MAPREDUCE-1170</a>] - MultipleInputs doesn't work with new API in 0.20 branch
</li>
</ul>
-
-<h4> Test
-</h4>
-<ul>
-<li>[<a href='https://issues.apache.org/jira/browse/MAPREDUCE-1336'>MAPREDUCE-1336</a>] - TestStreamingExitStatus - Fix deprecated use of StreamJob submission API
-</li>
-</ul>
-
+
+
<h2>Changes Since Hadoop 0.20.0</h2>
<h3>Common</h3>