Uploaded image for project: 'Hadoop Common'
  1. Hadoop Common
  2. HADOOP-9324

Out of date API document

    XMLWordPrintableJSON

Details

    • Bug
    • Status: Resolved
    • Major
    • Resolution: Duplicate
    • 2.0.3-alpha
    • None
    • documentation
    • None

    Description

      The documentation is out of date. Some code references are broken:
      1. http://hadoop.apache.org/docs/current/api/org/apache/hadoop/fs/FSDataInputStream.html
      "All Implemented Interfaces:
      Closeable, DataInput, org.apache.hadoop.fs.ByteBufferReadable, org.apache.hadoop.fs.HasFileDescriptor, PositionedReadable, Seekable "

      2.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapreduce/Cluster.html
      renewDelegationToken(org.apache.hadoop.security.token.Token<org.apache.hadoop.mapreduce.security.token.delegation.DelegationTokenIdentifier> token)
      Deprecated. Use Token.renew(org.apache.hadoop.conf.Configuration) instead

      3.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapred/JobConf.html
      "Use MRAsyncDiskService.moveAndDeleteAllVolumes instead. "
      I cannot find the MRAsyncDiskService class in the documentation of 2.0.3.

      4.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapred/join/CompositeRecordReader.html
      "protected org.apache.hadoop.mapred.join.CompositeRecordReader.JoinCollector jc"
      Please globally search JoinCollector. It is deleted, but mentioned many times in the current documentation.

      5.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapred/OutputCommitter.html
      "abortJob(JobContext context, org.apache.hadoop.mapreduce.JobStatus.State runState)"
      http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapreduce/Job.html
      "public org.apache.hadoop.mapreduce.JobStatus.State getJobState()"

      4.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapred/SequenceFileOutputFormat.html
      " static org.apache.hadoop.io.SequenceFile.CompressionType getOutputCompressionType"
      " static org.apache.hadoop.io.SequenceFile.Reader[] getReaders"

      5.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapred/TaskCompletionEvent.html
      "Returns enum Status.SUCESS or Status.FAILURE."->Status.SUCCEEDED?

      6.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapreduce/Job.html
      " static org.apache.hadoop.mapreduce.Job.TaskStatusFilter getTaskOutputFilter"
      " org.apache.hadoop.mapreduce.TaskReport[] getTaskReports(TaskType type) "

      7.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapreduce/Reducer.html
      "cleanup(org.apache.hadoop.mapreduce.Reducer.Context context) "

      8.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapred/SequenceFileOutputFormat.html
      "static org.apache.hadoop.io.SequenceFile.CompressionType getOutputCompressionType(JobConf conf)
      Get the SequenceFile.CompressionType for the output SequenceFile."
      " static org.apache.hadoop.io.SequenceFile.Reader[] getReaders"

      9.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapreduce/lib/partition/InputSampler.html
      "writePartitionFile(Job job, org.apache.hadoop.mapreduce.lib.partition.InputSampler.Sampler<K,V> sampler) "

      10.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapreduce/lib/partition/TotalOrderPartitioner.html
      contain JobContextImpl.getNumReduceTasks() - 1 keys.
      The JobContextImpl class is already deleted.

      11. http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapreduce/OutputCommitter.html
      "Note that this is invoked for jobs with final runstate as JobStatus.State.FAILED or JobStatus.State.KILLED."->JobStatus.FAILED JobStatus.KILLED?

      12.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapred/TaskAttemptContext.html
      "All Superinterfaces:
      JobContext, org.apache.hadoop.mapreduce.MRJobConfig, Progressable, TaskAttemptContext "

      13.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/metrics/file/FileContext.html
      "All Implemented Interfaces:
      org.apache.hadoop.metrics.MetricsContext"

      14.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/metrics/spi/AbstractMetricsContext.html
      "org.apache.hadoop.metrics.MetricsRecord createRecord(String recordName)"

      15. http://hadoop.apache.org/docs/current/api/org/apache/hadoop/net/DNSToSwitchMapping.html
      "If a name cannot be resolved to a rack, the implementation should return NetworkTopology.DEFAULT_RACK."
      NetworkTopology is deleted.

      16.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/metrics2/package-summary.html
      "myprefix.sink.file.class=org.hadoop.metrics2.sink.FileSink" ->
      org.apache.hadoop.metrics2.sink.FileSink?
      "org.apache.hadoop.metrics2.impl" -> The package is not found.

      17.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/ha/HAServiceTarget.html
      " abstract org.apache.hadoop.ha.NodeFencer getFencer() "

      18.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapreduce/MarkableIterator.html
      "MarkableIterator is a wrapper iterator class that implements the MarkableIteratorInterface. "
      MarkableIteratorInterface is deleted.

      19.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/metrics/spi/NoEmitMetricsContext.html
      "A MetricsContext that does not emit data, but, unlike NullContextWithUpdate"
      NullContextWithUpdate is deleted.

      20.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/net/ConnectTimeoutException.html
      "Thrown by NetUtils.connect(java.net.Socket, java.net.SocketAddress, int) "
      The NetUtils class is deleted.

      Please revise the documentation.

      Attachments

        Issue Links

          Activity

            People

              Unassigned Unassigned
              drzhonghao Hao Zhong
              Votes:
              0 Vote for this issue
              Watchers:
              3 Start watching this issue

              Dates

                Created:
                Updated:
                Resolved: