Details
-
Bug
-
Status: Closed
-
Major
-
Resolution: Invalid
-
0.8.1, 0.8.2, 0.9.0
-
None
-
None
-
Hadoop-0.6.2
Description
org.apache.nutch.indexer.DeleteDuplicates fails to build against Hadoop-0.6.2
Specifically, the interface org.apache.hadoop.mapred.RecordReader has two new methods:
WritableComparable createKey();
Writable createValue();
Which are not implemented by the anonymous inner class declared starting on line 155 of DeleteDuplicates.java 438670:
149: public RecordReader getRecordReader(final FileSystem fs,
150: final FileSplit split,
151: final JobConf job,
152: Reporter reporter) throws IOException {
153: final UTF8 index = new UTF8(split.getPath().toString());
154: reporter.setStatus(index.toString());
155: return new RecordReader() {
Attachments
Attachments
Issue Links
- is related to
-
NUTCH-371 DeleteDuplicates should remove documents with duplicate URLs
- Closed