You are viewing a plain text version of this content. The canonical link for it is here.
Posted to common-issues@hadoop.apache.org by "Allen Wittenauer (JIRA)" <ji...@apache.org> on 2015/04/13 14:58:13 UTC

[jira] [Updated] (HADOOP-9324) Out of date API document

     [ https://issues.apache.org/jira/browse/HADOOP-9324?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Allen Wittenauer updated HADOOP-9324:
-------------------------------------
    Component/s: documentation

> Out of date API document
> ------------------------
>
>                 Key: HADOOP-9324
>                 URL: https://issues.apache.org/jira/browse/HADOOP-9324
>             Project: Hadoop Common
>          Issue Type: Bug
>          Components: documentation
>    Affects Versions: 2.0.3-alpha
>            Reporter: Hao Zhong
>
> The documentation is out of date. Some code references are broken:
> 1. http://hadoop.apache.org/docs/current/api/org/apache/hadoop/fs/FSDataInputStream.html
> "All Implemented Interfaces:
>     Closeable, DataInput, *org.apache.hadoop.fs.ByteBufferReadable*, *org.apache.hadoop.fs.HasFileDescriptor*, PositionedReadable, Seekable "
> 2.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapreduce/Cluster.html
> renewDelegationToken(*org.apache.hadoop.security.token.Token<org.apache.hadoop.mapreduce.security.token.delegation.DelegationTokenIdentifier>* token)
>           Deprecated. Use Token.renew(*org.apache.hadoop.conf.Configuration*) instead
> 3.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapred/JobConf.html
> "Use MRAsyncDiskService.moveAndDeleteAllVolumes instead. "
> I cannot find the MRAsyncDiskService class in the documentation of 2.0.3. 
> 4.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapred/join/CompositeRecordReader.html
>  "protected *org.apache.hadoop.mapred.join.CompositeRecordReader.JoinCollector* 	jc"
> Please globally search JoinCollector. It is deleted, but mentioned many times in the current documentation.
> 5.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapred/OutputCommitter.html
> "abortJob(JobContext context, *org.apache.hadoop.mapreduce.JobStatus.State runState*)"  
> http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapreduce/Job.html
> "public *org.apache.hadoop.mapreduce.JobStatus.State* getJobState()"
> 4.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapred/SequenceFileOutputFormat.html
> " static *org.apache.hadoop.io.SequenceFile.CompressionType* getOutputCompressionType"
> " static *org.apache.hadoop.io.SequenceFile.Reader[]* 	getReaders"
> 5.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapred/TaskCompletionEvent.html
> "Returns enum Status.SUCESS or Status.FAILURE."->Status.SUCCEEDED? 
> 6.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapreduce/Job.html
> " static *org.apache.hadoop.mapreduce.Job.TaskStatusFilter* 	getTaskOutputFilter"
> "  org.apache.hadoop.mapreduce.TaskReport[] 	getTaskReports(TaskType type) "
> 7.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapreduce/Reducer.html
> "cleanup(*org.apache.hadoop.mapreduce.Reducer.Context* context) "
> 8.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapred/SequenceFileOutputFormat.html
>  "static *org.apache.hadoop.io.SequenceFile.CompressionType* 	getOutputCompressionType(JobConf conf)
>           Get the *SequenceFile.CompressionType* for the output SequenceFile."
> " static *org.apache.hadoop.io.SequenceFile.Reader[]* 	getReaders" 
> 9.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapreduce/lib/partition/InputSampler.html
> "writePartitionFile(Job job, *org.apache.hadoop.mapreduce.lib.partition.InputSampler.Sampler<K,V>* sampler) "
> 10.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapreduce/lib/partition/TotalOrderPartitioner.html
> contain JobContextImpl.getNumReduceTasks() - 1 keys. 
> The JobContextImpl class is already deleted.
> 11. http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapreduce/OutputCommitter.html
> "Note that this is invoked for jobs with final runstate as JobStatus.State.FAILED or JobStatus.State.KILLED."->JobStatus.FAILED JobStatus.KILLED?
> 12.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapred/TaskAttemptContext.html
> "All Superinterfaces:
>     JobContext, *org.apache.hadoop.mapreduce.MRJobConfig*, Progressable, TaskAttemptContext "
> 13.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/metrics/file/FileContext.html
> "All Implemented Interfaces:
>     *org.apache.hadoop.metrics.MetricsContext*"
> 14.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/metrics/spi/AbstractMetricsContext.html   
> "*org.apache.hadoop.metrics.MetricsRecord* 	createRecord(String recordName)"
> 15. http://hadoop.apache.org/docs/current/api/org/apache/hadoop/net/DNSToSwitchMapping.html
> "If a name cannot be resolved to a rack, the implementation should return NetworkTopology.DEFAULT_RACK."
> NetworkTopology is deleted.
> 16.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/metrics2/package-summary.html
> "myprefix.sink.file.class=org.hadoop.metrics2.sink.FileSink" ->
> org.apache.hadoop.metrics2.sink.FileSink?
> "org.apache.hadoop.metrics2.impl" -> The package is not found.
> 17.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/ha/HAServiceTarget.html
> " abstract  *org.apache.hadoop.ha.NodeFencer* 	getFencer() "
> 18.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapreduce/MarkableIterator.html
> "MarkableIterator is a wrapper iterator class that implements the MarkableIteratorInterface. "
> MarkableIteratorInterface is deleted.
> 19.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/metrics/spi/NoEmitMetricsContext.html
>  "A MetricsContext that does not emit data, but, unlike NullContextWithUpdate"
> NullContextWithUpdate is deleted.
> 20.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/net/ConnectTimeoutException.html
> "Thrown by NetUtils.connect(java.net.Socket, java.net.SocketAddress, int) "
> The NetUtils class is deleted.
> Please revise the documentation.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)