Details
-
Bug
-
Status: Resolved
-
Major
-
Resolution: Duplicate
-
2.0.3-alpha
-
None
-
None
Description
The documentation is out of date. Some code references are broken:
1. http://hadoop.apache.org/docs/current/api/org/apache/hadoop/fs/FSDataInputStream.html
"All Implemented Interfaces:
Closeable, DataInput, org.apache.hadoop.fs.ByteBufferReadable, org.apache.hadoop.fs.HasFileDescriptor, PositionedReadable, Seekable "
2.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapreduce/Cluster.html
renewDelegationToken(org.apache.hadoop.security.token.Token<org.apache.hadoop.mapreduce.security.token.delegation.DelegationTokenIdentifier> token)
Deprecated. Use Token.renew(org.apache.hadoop.conf.Configuration) instead
3.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapred/JobConf.html
"Use MRAsyncDiskService.moveAndDeleteAllVolumes instead. "
I cannot find the MRAsyncDiskService class in the documentation of 2.0.3.
4.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapred/join/CompositeRecordReader.html
"protected org.apache.hadoop.mapred.join.CompositeRecordReader.JoinCollector jc"
Please globally search JoinCollector. It is deleted, but mentioned many times in the current documentation.
5.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapred/OutputCommitter.html
"abortJob(JobContext context, org.apache.hadoop.mapreduce.JobStatus.State runState)"
http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapreduce/Job.html
"public org.apache.hadoop.mapreduce.JobStatus.State getJobState()"
4.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapred/SequenceFileOutputFormat.html
" static org.apache.hadoop.io.SequenceFile.CompressionType getOutputCompressionType"
" static org.apache.hadoop.io.SequenceFile.Reader[] getReaders"
5.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapred/TaskCompletionEvent.html
"Returns enum Status.SUCESS or Status.FAILURE."->Status.SUCCEEDED?
6.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapreduce/Job.html
" static org.apache.hadoop.mapreduce.Job.TaskStatusFilter getTaskOutputFilter"
" org.apache.hadoop.mapreduce.TaskReport[] getTaskReports(TaskType type) "
7.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapreduce/Reducer.html
"cleanup(org.apache.hadoop.mapreduce.Reducer.Context context) "
8.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapred/SequenceFileOutputFormat.html
"static org.apache.hadoop.io.SequenceFile.CompressionType getOutputCompressionType(JobConf conf)
Get the SequenceFile.CompressionType for the output SequenceFile."
" static org.apache.hadoop.io.SequenceFile.Reader[] getReaders"
9.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapreduce/lib/partition/InputSampler.html
"writePartitionFile(Job job, org.apache.hadoop.mapreduce.lib.partition.InputSampler.Sampler<K,V> sampler) "
10.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapreduce/lib/partition/TotalOrderPartitioner.html
contain JobContextImpl.getNumReduceTasks() - 1 keys.
The JobContextImpl class is already deleted.
11. http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapreduce/OutputCommitter.html
"Note that this is invoked for jobs with final runstate as JobStatus.State.FAILED or JobStatus.State.KILLED."->JobStatus.FAILED JobStatus.KILLED?
12.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapred/TaskAttemptContext.html
"All Superinterfaces:
JobContext, org.apache.hadoop.mapreduce.MRJobConfig, Progressable, TaskAttemptContext "
13.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/metrics/file/FileContext.html
"All Implemented Interfaces:
org.apache.hadoop.metrics.MetricsContext"
14.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/metrics/spi/AbstractMetricsContext.html
"org.apache.hadoop.metrics.MetricsRecord createRecord(String recordName)"
15. http://hadoop.apache.org/docs/current/api/org/apache/hadoop/net/DNSToSwitchMapping.html
"If a name cannot be resolved to a rack, the implementation should return NetworkTopology.DEFAULT_RACK."
NetworkTopology is deleted.
16.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/metrics2/package-summary.html
"myprefix.sink.file.class=org.hadoop.metrics2.sink.FileSink" ->
org.apache.hadoop.metrics2.sink.FileSink?
"org.apache.hadoop.metrics2.impl" -> The package is not found.
17.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/ha/HAServiceTarget.html
" abstract org.apache.hadoop.ha.NodeFencer getFencer() "
18.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapreduce/MarkableIterator.html
"MarkableIterator is a wrapper iterator class that implements the MarkableIteratorInterface. "
MarkableIteratorInterface is deleted.
19.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/metrics/spi/NoEmitMetricsContext.html
"A MetricsContext that does not emit data, but, unlike NullContextWithUpdate"
NullContextWithUpdate is deleted.
20.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/net/ConnectTimeoutException.html
"Thrown by NetUtils.connect(java.net.Socket, java.net.SocketAddress, int) "
The NetUtils class is deleted.
Please revise the documentation.
Attachments
Issue Links
- duplicates
-
HADOOP-15021 Excluding private and limitiedprivate from javadoc causes broken links
- Open