Index: lucene/contrib/highlighter/src/test/org/apache/lucene/search/highlight/HighlighterTest.java =================================================================== --- lucene/contrib/highlighter/src/test/org/apache/lucene/search/highlight/HighlighterTest.java (revision 1297029) +++ lucene/contrib/highlighter/src/test/org/apache/lucene/search/highlight/HighlighterTest.java (working copy) @@ -50,6 +50,7 @@ import org.apache.lucene.search.spans.*; import org.apache.lucene.store.Directory; import org.apache.lucene.util.BytesRef; +import org.apache.lucene.util.LuceneTestCase; import org.apache.lucene.util.automaton.BasicAutomata; import org.apache.lucene.util.automaton.CharacterRunAutomaton; import org.apache.lucene.util.automaton.RegExp; @@ -1969,16 +1970,16 @@ String result = highlighter.getBestFragments(tokenStream, text, maxNumFragmentsRequired, fragmentSeparator); - if (HighlighterTest.VERBOSE) System.out.println("\t" + result); + if (LuceneTestCase.VERBOSE) System.out.println("\t" + result); } } abstract void run() throws Exception; void start() throws Exception { - if (HighlighterTest.VERBOSE) System.out.println("Run QueryScorer"); + if (LuceneTestCase.VERBOSE) System.out.println("Run QueryScorer"); run(); - if (HighlighterTest.VERBOSE) System.out.println("Run QueryTermScorer"); + if (LuceneTestCase.VERBOSE) System.out.println("Run QueryTermScorer"); mode = QUERY_TERM; run(); } Index: lucene/contrib/sandbox/src/java/org/apache/lucene/sandbox/queries/DuplicateFilter.java =================================================================== --- lucene/contrib/sandbox/src/java/org/apache/lucene/sandbox/queries/DuplicateFilter.java (revision 1297029) +++ lucene/contrib/sandbox/src/java/org/apache/lucene/sandbox/queries/DuplicateFilter.java (working copy) @@ -18,6 +18,7 @@ import org.apache.lucene.index.*; import org.apache.lucene.search.DocIdSet; +import org.apache.lucene.search.DocIdSetIterator; import org.apache.lucene.search.Filter; import org.apache.lucene.util.Bits; import org.apache.lucene.util.BytesRef; @@ -94,7 +95,7 @@ } else { docs = termsEnum.docs(acceptDocs, docs, false); int doc = docs.nextDoc(); - if (doc != DocsEnum.NO_MORE_DOCS) { + if (doc != DocIdSetIterator.NO_MORE_DOCS) { if (keepMode == KeepMode.KM_USE_FIRST_OCCURRENCE) { bits.set(doc); } else { @@ -102,7 +103,7 @@ while (true) { lastDoc = doc; doc = docs.nextDoc(); - if (doc == DocsEnum.NO_MORE_DOCS) { + if (doc == DocIdSetIterator.NO_MORE_DOCS) { break; } } @@ -134,7 +135,7 @@ // unset potential duplicates docs = termsEnum.docs(acceptDocs, docs, false); int doc = docs.nextDoc(); - if (doc != DocsEnum.NO_MORE_DOCS) { + if (doc != DocIdSetIterator.NO_MORE_DOCS) { if (keepMode == KeepMode.KM_USE_FIRST_OCCURRENCE) { doc = docs.nextDoc(); } @@ -145,7 +146,7 @@ lastDoc = doc; bits.clear(lastDoc); doc = docs.nextDoc(); - if (doc == DocsEnum.NO_MORE_DOCS) { + if (doc == DocIdSetIterator.NO_MORE_DOCS) { break; } } Index: lucene/contrib/sandbox/src/test/org/apache/lucene/sandbox/queries/DuplicateFilterTest.java =================================================================== --- lucene/contrib/sandbox/src/test/org/apache/lucene/sandbox/queries/DuplicateFilterTest.java (revision 1297029) +++ lucene/contrib/sandbox/src/test/org/apache/lucene/sandbox/queries/DuplicateFilterTest.java (working copy) @@ -25,6 +25,7 @@ import org.apache.lucene.document.StringField; import org.apache.lucene.document.TextField; import org.apache.lucene.index.*; +import org.apache.lucene.search.DocIdSetIterator; import org.apache.lucene.search.IndexSearcher; import org.apache.lucene.search.ScoreDoc; import org.apache.lucene.search.TermQuery; @@ -142,7 +143,7 @@ false); int lastDoc = 0; - while (td.nextDoc() != DocsEnum.NO_MORE_DOCS) { + while (td.nextDoc() != DocIdSetIterator.NO_MORE_DOCS) { lastDoc = td.docID(); } assertEquals("Duplicate urls should return last doc", lastDoc, hit.doc); Index: lucene/core/src/java/org/apache/lucene/codecs/BlockTreeTermsReader.java =================================================================== --- lucene/core/src/java/org/apache/lucene/codecs/BlockTreeTermsReader.java (revision 1297029) +++ lucene/core/src/java/org/apache/lucene/codecs/BlockTreeTermsReader.java (working copy) @@ -488,7 +488,7 @@ private Frame[] stack; - @SuppressWarnings("unchecked") private FST.Arc[] arcs = new FST.Arc[5]; + @SuppressWarnings({"rawtypes","unchecked"}) private FST.Arc[] arcs = new FST.Arc[5]; private final RunAutomaton runAutomaton; private final CompiledAutomaton compiledAutomaton; @@ -821,7 +821,8 @@ private FST.Arc getArc(int ord) { if (ord >= arcs.length) { - @SuppressWarnings("unchecked") final FST.Arc[] next = new FST.Arc[ArrayUtil.oversize(1+ord, RamUsageEstimator.NUM_BYTES_OBJECT_REF)]; + @SuppressWarnings({"rawtypes","unchecked"}) final FST.Arc[] next = + new FST.Arc[ArrayUtil.oversize(1+ord, RamUsageEstimator.NUM_BYTES_OBJECT_REF)]; System.arraycopy(arcs, 0, next, 0, arcs.length); for(int arcOrd=arcs.length;arcOrd(); @@ -1198,7 +1199,8 @@ final BytesRef term = new BytesRef(); private final FST.BytesReader fstReader; - @SuppressWarnings("unchecked") private FST.Arc[] arcs = new FST.Arc[1]; + @SuppressWarnings({"rawtypes","unchecked"}) private FST.Arc[] arcs = + new FST.Arc[1]; public SegmentTermsEnum() throws IOException { //if (DEBUG) System.out.println("BTTR.init seg=" + segment); @@ -1354,7 +1356,8 @@ private FST.Arc getArc(int ord) { if (ord >= arcs.length) { - @SuppressWarnings("unchecked") final FST.Arc[] next = new FST.Arc[ArrayUtil.oversize(1+ord, RamUsageEstimator.NUM_BYTES_OBJECT_REF)]; + @SuppressWarnings({"rawtypes","unchecked"}) final FST.Arc[] next = + new FST.Arc[ArrayUtil.oversize(1+ord, RamUsageEstimator.NUM_BYTES_OBJECT_REF)]; System.arraycopy(arcs, 0, next, 0, arcs.length); for(int arcOrd=arcs.length;arcOrd(); Index: lucene/core/src/java/org/apache/lucene/codecs/BlockTreeTermsWriter.java =================================================================== --- lucene/core/src/java/org/apache/lucene/codecs/BlockTreeTermsWriter.java (revision 1297029) +++ lucene/core/src/java/org/apache/lucene/codecs/BlockTreeTermsWriter.java (working copy) @@ -639,7 +639,6 @@ } // for debugging - @SuppressWarnings("unused") private String toString(BytesRef b) { try { return b.utf8ToString() + " " + b; Index: lucene/core/src/java/org/apache/lucene/codecs/TermVectorsWriter.java =================================================================== --- lucene/core/src/java/org/apache/lucene/codecs/TermVectorsWriter.java (revision 1297029) +++ lucene/core/src/java/org/apache/lucene/codecs/TermVectorsWriter.java (working copy) @@ -30,6 +30,7 @@ import org.apache.lucene.index.MergeState; import org.apache.lucene.index.Terms; import org.apache.lucene.index.TermsEnum; +import org.apache.lucene.search.DocIdSetIterator; import org.apache.lucene.store.DataInput; import org.apache.lucene.util.Bits; import org.apache.lucene.util.BytesRef; @@ -236,7 +237,7 @@ if (docsAndPositionsEnum != null) { final int docID = docsAndPositionsEnum.nextDoc(); - assert docID != DocsEnum.NO_MORE_DOCS; + assert docID != DocIdSetIterator.NO_MORE_DOCS; assert docsAndPositionsEnum.freq() == freq; for(int posUpto=0; posUpto 5*postings[0].docFreq; chunkStates[i] = new ChunkState(postings[i].postings, -postings[i].position, useAdvance); - if (i > 0 && postings[i].postings.nextDoc() == DocsEnum.NO_MORE_DOCS) { + if (i > 0 && postings[i].postings.nextDoc() == DocIdSetIterator.NO_MORE_DOCS) { noDocs = true; return; } @@ -89,7 +89,7 @@ // first (rarest) term final int doc = chunkStates[0].posEnum.nextDoc(); - if (doc == DocsEnum.NO_MORE_DOCS) { + if (doc == DocIdSetIterator.NO_MORE_DOCS) { docID = doc; return doc; } @@ -140,8 +140,8 @@ // first term int doc = chunkStates[0].posEnum.advance(target); - if (doc == DocsEnum.NO_MORE_DOCS) { - docID = DocsEnum.NO_MORE_DOCS; + if (doc == DocIdSetIterator.NO_MORE_DOCS) { + docID = DocIdSetIterator.NO_MORE_DOCS; return doc; } @@ -171,7 +171,7 @@ } doc = chunkStates[0].posEnum.nextDoc(); - if (doc == DocsEnum.NO_MORE_DOCS) { + if (doc == DocIdSetIterator.NO_MORE_DOCS) { docID = doc; return doc; } Index: lucene/core/src/java/org/apache/lucene/search/FieldCacheImpl.java =================================================================== --- lucene/core/src/java/org/apache/lucene/search/FieldCacheImpl.java (revision 1297029) +++ lucene/core/src/java/org/apache/lucene/search/FieldCacheImpl.java (working copy) @@ -367,7 +367,7 @@ docs = termsEnum.docs(null, docs, false); while (true) { final int docID = docs.nextDoc(); - if (docID == DocsEnum.NO_MORE_DOCS) { + if (docID == DocIdSetIterator.NO_MORE_DOCS) { break; } retArray[docID] = termval; @@ -440,7 +440,7 @@ docs = termsEnum.docs(null, docs, false); while (true) { final int docID = docs.nextDoc(); - if (docID == DocsEnum.NO_MORE_DOCS) { + if (docID == DocIdSetIterator.NO_MORE_DOCS) { break; } retArray[docID] = termval; @@ -544,7 +544,7 @@ docs = termsEnum.docs(null, docs, false); while (true) { final int docID = docs.nextDoc(); - if (docID == DocsEnum.NO_MORE_DOCS) { + if (docID == DocIdSetIterator.NO_MORE_DOCS) { break; } retArray[docID] = termval; @@ -612,7 +612,7 @@ // TODO: use bulk API while (true) { final int docID = docs.nextDoc(); - if (docID == DocsEnum.NO_MORE_DOCS) { + if (docID == DocIdSetIterator.NO_MORE_DOCS) { break; } res.set(docID); @@ -694,7 +694,7 @@ docs = termsEnum.docs(null, docs, false); while (true) { final int docID = docs.nextDoc(); - if (docID == DocsEnum.NO_MORE_DOCS) { + if (docID == DocIdSetIterator.NO_MORE_DOCS) { break; } retArray[docID] = termval; @@ -782,7 +782,7 @@ docs = termsEnum.docs(null, docs, false); while (true) { final int docID = docs.nextDoc(); - if (docID == DocsEnum.NO_MORE_DOCS) { + if (docID == DocIdSetIterator.NO_MORE_DOCS) { break; } retArray[docID] = termval; @@ -871,7 +871,7 @@ docs = termsEnum.docs(null, docs, false); while (true) { final int docID = docs.nextDoc(); - if (docID == DocsEnum.NO_MORE_DOCS) { + if (docID == DocIdSetIterator.NO_MORE_DOCS) { break; } retArray[docID] = termval; @@ -1172,7 +1172,7 @@ docs = termsEnum.docs(null, docs, false); while (true) { final int docID = docs.nextDoc(); - if (docID == DocsEnum.NO_MORE_DOCS) { + if (docID == DocIdSetIterator.NO_MORE_DOCS) { break; } docToTermOrd.set(docID, termOrd); @@ -1293,7 +1293,7 @@ docs = termsEnum.docs(null, docs, false); while (true) { final int docID = docs.nextDoc(); - if (docID == DocsEnum.NO_MORE_DOCS) { + if (docID == DocIdSetIterator.NO_MORE_DOCS) { break; } docToOffset.set(docID, pointer); Index: lucene/core/src/java/org/apache/lucene/search/FieldCacheRangeFilter.java =================================================================== --- lucene/core/src/java/org/apache/lucene/search/FieldCacheRangeFilter.java (revision 1297029) +++ lucene/core/src/java/org/apache/lucene/search/FieldCacheRangeFilter.java (working copy) @@ -459,6 +459,7 @@ } @Override + @SuppressWarnings({"unchecked","rawtypes"}) public final boolean equals(Object o) { if (this == o) return true; if (!(o instanceof FieldCacheRangeFilter)) return false; Index: lucene/core/src/java/org/apache/lucene/search/FieldComparator.java =================================================================== --- lucene/core/src/java/org/apache/lucene/search/FieldComparator.java (revision 1297029) +++ lucene/core/src/java/org/apache/lucene/search/FieldComparator.java (working copy) @@ -150,7 +150,7 @@ * comparator across segments * @throws IOException */ - public abstract FieldComparator setNextReader(AtomicReaderContext context) throws IOException; + public abstract FieldComparator setNextReader(AtomicReaderContext context) throws IOException; /** Sets the Scorer to use in case a document's score is * needed. @@ -201,7 +201,7 @@ } @Override - public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { + public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { if (missingValue != null) { docsWithField = FieldCache.DEFAULT.getDocsWithField(context.reader(), field); // optimization to remove unneeded checks on the bit interface: @@ -258,7 +258,7 @@ } @Override - public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { + public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { // NOTE: must do this before calling super otherwise // we compute the docsWithField Bits twice! currentReaderValues = FieldCache.DEFAULT.getBytes(context.reader(), field, parser, missingValue != null); @@ -335,7 +335,7 @@ } @Override - public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { + public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { // NOTE: must do this before calling super otherwise // we compute the docsWithField Bits twice! currentReaderValues = FieldCache.DEFAULT.getDoubles(context.reader(), field, parser, missingValue != null); @@ -396,7 +396,7 @@ } @Override - public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { + public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { final DocValues docValues = context.reader().docValues(field); if (docValues != null) { currentReaderValues = docValues.getSource(); @@ -478,7 +478,7 @@ } @Override - public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { + public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { // NOTE: must do this before calling super otherwise // we compute the docsWithField Bits twice! currentReaderValues = FieldCache.DEFAULT.getFloats(context.reader(), field, parser, missingValue != null); @@ -540,7 +540,7 @@ } @Override - public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { + public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { // NOTE: must do this before calling super otherwise // we compute the docsWithField Bits twice! currentReaderValues = FieldCache.DEFAULT.getShorts(context.reader(), field, parser, missingValue != null); @@ -624,7 +624,7 @@ } @Override - public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { + public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { // NOTE: must do this before calling super otherwise // we compute the docsWithField Bits twice! currentReaderValues = FieldCache.DEFAULT.getInts(context.reader(), field, parser, missingValue != null); @@ -689,7 +689,7 @@ } @Override - public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { + public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { DocValues docValues = context.reader().docValues(field); if (docValues != null) { currentReaderValues = docValues.getSource(); @@ -772,7 +772,7 @@ } @Override - public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { + public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { // NOTE: must do this before calling super otherwise // we compute the docsWithField Bits twice! currentReaderValues = FieldCache.DEFAULT.getLongs(context.reader(), field, parser, missingValue != null); @@ -824,7 +824,7 @@ } @Override - public FieldComparator setNextReader(AtomicReaderContext context) { + public FieldComparator setNextReader(AtomicReaderContext context) { return this; } @@ -887,7 +887,7 @@ } @Override - public FieldComparator setNextReader(AtomicReaderContext context) { + public FieldComparator setNextReader(AtomicReaderContext context) { // TODO: can we "map" our docIDs to the current // reader? saves having to then subtract on every // compare call @@ -1007,7 +1007,7 @@ abstract class PerSegmentComparator extends FieldComparator { @Override - public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { + public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { return TermOrdValComparator.this.setNextReader(context); } @@ -1226,11 +1226,11 @@ } @Override - public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { + public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { final int docBase = context.docBase; termsIndex = FieldCache.DEFAULT.getTermsIndex(context.reader(), field); final PackedInts.Reader docToOrd = termsIndex.getDocToOrd(); - FieldComparator perSegComp = null; + FieldComparator perSegComp = null; if (docToOrd.hasArray()) { final Object arr = docToOrd.getArray(); if (arr instanceof byte[]) { @@ -1397,7 +1397,7 @@ abstract class PerSegmentComparator extends FieldComparator { @Override - public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { + public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { return TermOrdValDocValuesComparator.this.setNextReader(context); } @@ -1625,7 +1625,7 @@ } @Override - public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { + public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { final int docBase = context.docBase; final DocValues dv = context.reader().docValues(field); @@ -1646,7 +1646,7 @@ comp = termsIndex.getComparator(); - FieldComparator perSegComp = null; + FieldComparator perSegComp = null; if (termsIndex.hasPackedDocToOrd()) { final PackedInts.Reader docToOrd = termsIndex.getDocToOrd(); if (docToOrd.hasArray()) { @@ -1774,7 +1774,7 @@ } @Override - public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { + public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { docTerms = FieldCache.DEFAULT.getTerms(context.reader(), field); return this; } @@ -1843,7 +1843,7 @@ } @Override - public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { + public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { final DocValues dv = context.reader().docValues(field); if (dv != null) { docTerms = dv.getSource(); Index: lucene/core/src/java/org/apache/lucene/search/FieldComparatorSource.java =================================================================== --- lucene/core/src/java/org/apache/lucene/search/FieldComparatorSource.java (revision 1297029) +++ lucene/core/src/java/org/apache/lucene/search/FieldComparatorSource.java (working copy) @@ -36,6 +36,6 @@ * @throws IOException * If an error occurs reading the index. */ - public abstract FieldComparator newComparator(String fieldname, int numHits, int sortPos, boolean reversed) + public abstract FieldComparator newComparator(String fieldname, int numHits, int sortPos, boolean reversed) throws IOException; } Index: lucene/core/src/java/org/apache/lucene/search/FieldValueHitQueue.java =================================================================== --- lucene/core/src/java/org/apache/lucene/search/FieldValueHitQueue.java (revision 1297029) +++ lucene/core/src/java/org/apache/lucene/search/FieldValueHitQueue.java (working copy) @@ -129,6 +129,7 @@ } // prevent instantiation and extension. + @SuppressWarnings({"rawtypes","unchecked"}) private FieldValueHitQueue(SortField[] fields, int size) { super(size); // When we get here, fields.length is guaranteed to be > 0, therefore no @@ -169,7 +170,7 @@ } } - public FieldComparator[] getComparators() { + public FieldComparator[] getComparators() { return comparators; } @@ -177,15 +178,15 @@ return reverseMul; } - public void setComparator(int pos, FieldComparator comparator) { + public void setComparator(int pos, FieldComparator comparator) { if (pos==0) firstComparator = comparator; comparators[pos] = comparator; } /** Stores the sort criteria being used. */ protected final SortField[] fields; - protected final FieldComparator[] comparators; // use setComparator to change this array - protected FieldComparator firstComparator; // this must always be equal to comparators[0] + protected final FieldComparator[] comparators; // use setComparator to change this array + protected FieldComparator firstComparator; // this must always be equal to comparators[0] protected final int[] reverseMul; @Override Index: lucene/core/src/java/org/apache/lucene/search/MultiPhraseQuery.java =================================================================== --- lucene/core/src/java/org/apache/lucene/search/MultiPhraseQuery.java (revision 1297029) +++ lucene/core/src/java/org/apache/lucene/search/MultiPhraseQuery.java (working copy) @@ -408,7 +408,7 @@ Iterator i = docsEnums.iterator(); while (i.hasNext()) { DocsAndPositionsEnum postings = i.next(); - if (postings.nextDoc() != DocsAndPositionsEnum.NO_MORE_DOCS) { + if (postings.nextDoc() != DocIdSetIterator.NO_MORE_DOCS) { add(postings); } } Index: lucene/core/src/java/org/apache/lucene/search/MultiTermQueryWrapperFilter.java =================================================================== --- lucene/core/src/java/org/apache/lucene/search/MultiTermQueryWrapperFilter.java (revision 1297029) +++ lucene/core/src/java/org/apache/lucene/search/MultiTermQueryWrapperFilter.java (working copy) @@ -60,6 +60,7 @@ } @Override + @SuppressWarnings({"unchecked","rawtypes"}) public final boolean equals(final Object o) { if (o==this) return true; if (o==null) return false; Index: lucene/core/src/java/org/apache/lucene/search/NumericRangeQuery.java =================================================================== --- lucene/core/src/java/org/apache/lucene/search/NumericRangeQuery.java (revision 1297029) +++ lucene/core/src/java/org/apache/lucene/search/NumericRangeQuery.java (working copy) @@ -352,6 +352,7 @@ } @Override + @SuppressWarnings({"unchecked","rawtypes"}) public final boolean equals(final Object o) { if (o==this) return true; if (!super.equals(o)) Index: lucene/core/src/java/org/apache/lucene/search/payloads/PayloadSpanUtil.java =================================================================== --- lucene/core/src/java/org/apache/lucene/search/payloads/PayloadSpanUtil.java (revision 1297029) +++ lucene/core/src/java/org/apache/lucene/search/payloads/PayloadSpanUtil.java (working copy) @@ -135,7 +135,8 @@ } } - @SuppressWarnings("unchecked") final List[] disjunctLists = new List[maxPosition + 1]; + @SuppressWarnings({"rawtypes","unchecked"}) final List[] disjunctLists = + new List[maxPosition + 1]; int distinctPositions = 0; for (int i = 0; i < termArrays.size(); ++i) { Index: lucene/core/src/java/org/apache/lucene/search/SortField.java =================================================================== --- lucene/core/src/java/org/apache/lucene/search/SortField.java (revision 1297029) +++ lucene/core/src/java/org/apache/lucene/search/SortField.java (working copy) @@ -376,7 +376,7 @@ * optimize themselves when they are the primary sort. * @return {@link FieldComparator} to use when sorting */ - public FieldComparator getComparator(final int numHits, final int sortPos) throws IOException { + public FieldComparator getComparator(final int numHits, final int sortPos) throws IOException { switch (type) { case SCORE: Index: lucene/core/src/java/org/apache/lucene/search/spans/SpanMultiTermQueryWrapper.java =================================================================== --- lucene/core/src/java/org/apache/lucene/search/spans/SpanMultiTermQueryWrapper.java (revision 1297029) +++ lucene/core/src/java/org/apache/lucene/search/spans/SpanMultiTermQueryWrapper.java (working copy) @@ -60,6 +60,7 @@ * Be sure to not change the rewrite method on the wrapped query afterwards! Doing so will * throw {@link UnsupportedOperationException} on rewriting this query! */ + @SuppressWarnings({"rawtypes","unchecked"}) public SpanMultiTermQueryWrapper(Q query) { this.query = query; @@ -123,6 +124,7 @@ } @Override + @SuppressWarnings({"rawtypes","unchecked"}) public boolean equals(Object obj) { if (this == obj) return true; if (obj == null) return false; Index: lucene/core/src/java/org/apache/lucene/search/spans/TermSpans.java =================================================================== --- lucene/core/src/java/org/apache/lucene/search/spans/TermSpans.java (revision 1297029) +++ lucene/core/src/java/org/apache/lucene/search/spans/TermSpans.java (working copy) @@ -56,7 +56,7 @@ return false; } doc = postings.nextDoc(); - if (doc == DocsAndPositionsEnum.NO_MORE_DOCS) { + if (doc == DocIdSetIterator.NO_MORE_DOCS) { return false; } freq = postings.freq(); @@ -70,7 +70,7 @@ @Override public boolean skipTo(int target) throws IOException { doc = postings.advance(target); - if (doc == DocsAndPositionsEnum.NO_MORE_DOCS) { + if (doc == DocIdSetIterator.NO_MORE_DOCS) { return false; } Index: lucene/core/src/java/org/apache/lucene/search/TopDocs.java =================================================================== --- lucene/core/src/java/org/apache/lucene/search/TopDocs.java (revision 1297029) +++ lucene/core/src/java/org/apache/lucene/search/TopDocs.java (working copy) @@ -116,10 +116,11 @@ } } + @SuppressWarnings({"rawtypes","unchecked"}) private static class MergeSortQueue extends PriorityQueue { // These are really FieldDoc instances: final ScoreDoc[][] shardHits; - final FieldComparator[] comparators; + final FieldComparator[] comparators; final int[] reverseMul; public MergeSortQueue(Sort sort, TopDocs[] shardHits) throws IOException { @@ -155,7 +156,7 @@ } // Returns true if first is < second - @SuppressWarnings("unchecked") + @SuppressWarnings({"unchecked","rawtypes"}) public boolean lessThan(ShardRef first, ShardRef second) { assert first != second; final FieldDoc firstFD = (FieldDoc) shardHits[first.shardIndex][first.hitIndex]; Index: lucene/core/src/java/org/apache/lucene/search/TopFieldCollector.java =================================================================== --- lucene/core/src/java/org/apache/lucene/search/TopFieldCollector.java (revision 1297029) +++ lucene/core/src/java/org/apache/lucene/search/TopFieldCollector.java (working copy) @@ -46,7 +46,7 @@ private static class OneComparatorNonScoringCollector extends TopFieldCollector { - FieldComparator comparator; + FieldComparator comparator; final int reverseMul; final FieldValueHitQueue queue; @@ -382,7 +382,7 @@ */ private static class MultiComparatorNonScoringCollector extends TopFieldCollector { - final FieldComparator[] comparators; + final FieldComparator[] comparators; final int[] reverseMul; final FieldValueHitQueue queue; public MultiComparatorNonScoringCollector(FieldValueHitQueue queue, Index: lucene/core/src/java/org/apache/lucene/util/automaton/Automaton.java =================================================================== --- lucene/core/src/java/org/apache/lucene/util/automaton/Automaton.java (revision 1297029) +++ lucene/core/src/java/org/apache/lucene/util/automaton/Automaton.java (working copy) @@ -430,7 +430,7 @@ } } // map> - @SuppressWarnings("unchecked") Set map[] = new Set[states.length]; + @SuppressWarnings({"rawtypes","unchecked"}) Set map[] = new Set[states.length]; for (int i = 0; i < map.length; i++) map[i] = new HashSet(); for (State s : states) { Index: lucene/core/src/java/org/apache/lucene/util/automaton/MinimizationOperations.java =================================================================== --- lucene/core/src/java/org/apache/lucene/util/automaton/MinimizationOperations.java (revision 1297029) +++ lucene/core/src/java/org/apache/lucene/util/automaton/MinimizationOperations.java (working copy) @@ -74,11 +74,11 @@ final int[] sigma = a.getStartPoints(); final State[] states = a.getNumberedStates(); final int sigmaLen = sigma.length, statesLen = states.length; - @SuppressWarnings("unchecked") final ArrayList[][] reverse = + @SuppressWarnings({"rawtypes","unchecked"}) final ArrayList[][] reverse = (ArrayList[][]) new ArrayList[statesLen][sigmaLen]; - @SuppressWarnings("unchecked") final HashSet[] partition = + @SuppressWarnings({"rawtypes","unchecked"}) final HashSet[] partition = (HashSet[]) new HashSet[statesLen]; - @SuppressWarnings("unchecked") final ArrayList[] splitblock = + @SuppressWarnings({"rawtypes","unchecked"}) final ArrayList[] splitblock = (ArrayList[]) new ArrayList[statesLen]; final int[] block = new int[statesLen]; final StateList[][] active = new StateList[statesLen][sigmaLen]; Index: lucene/core/src/java/org/apache/lucene/util/fst/Builder.java =================================================================== --- lucene/core/src/java/org/apache/lucene/util/fst/Builder.java (revision 1297029) +++ lucene/core/src/java/org/apache/lucene/util/fst/Builder.java (working copy) @@ -144,7 +144,8 @@ } NO_OUTPUT = outputs.getNoOutput(); - @SuppressWarnings("unchecked") final UnCompiledNode[] f = (UnCompiledNode[]) new UnCompiledNode[10]; + @SuppressWarnings({"rawtypes","unchecked"}) final UnCompiledNode[] f = + (UnCompiledNode[]) new UnCompiledNode[10]; frontier = f; for(int idx=0;idx(this, idx); @@ -239,7 +240,8 @@ if (node.inputCount < minSuffixCount2 || (minSuffixCount2 == 1 && node.inputCount == 1 && idx > 1)) { // drop all arcs for(int arcIdx=0;arcIdx target = (UnCompiledNode) node.arcs[arcIdx].target; + @SuppressWarnings({"rawtypes","unchecked"}) final UnCompiledNode target = + (UnCompiledNode) node.arcs[arcIdx].target; target.clear(); } node.numArcs = 0; @@ -356,7 +358,7 @@ final int prefixLenPlus1 = pos1+1; if (frontier.length < input.length+1) { - @SuppressWarnings("unchecked") final UnCompiledNode[] next = + @SuppressWarnings({"rawtypes","unchecked"}) final UnCompiledNode[] next = new UnCompiledNode[ArrayUtil.oversize(input.length+1, RamUsageEstimator.NUM_BYTES_OBJECT_REF)]; System.arraycopy(frontier, 0, next, 0, frontier.length); for(int idx=frontier.length;idx arc = node.arcs[arcIdx]; if (!arc.target.isCompiled()) { // not yet compiled - @SuppressWarnings("unchecked") final UnCompiledNode n = (UnCompiledNode) arc.target; + @SuppressWarnings({"rawtypes","unchecked"}) final UnCompiledNode n = (UnCompiledNode) arc.target; if (n.numArcs == 0) { //System.out.println("seg=" + segment + " FORCE final arc=" + (char) arc.label); arc.isFinal = n.isFinal = true; @@ -512,7 +514,7 @@ * LUCENE-2934 (node expansion based on conditions other than the * fanout size). */ - @SuppressWarnings("unchecked") + @SuppressWarnings({"rawtypes","unchecked"}) public UnCompiledNode(Builder owner, int depth) { this.owner = owner; arcs = (Arc[]) new Arc[1]; @@ -545,7 +547,7 @@ assert label >= 0; assert numArcs == 0 || label > arcs[numArcs-1].label: "arc[-1].label=" + arcs[numArcs-1].label + " new label=" + label + " numArcs=" + numArcs; if (numArcs == arcs.length) { - @SuppressWarnings("unchecked") final Arc[] newArcs = + @SuppressWarnings({"rawtypes","unchecked"}) final Arc[] newArcs = new Arc[ArrayUtil.oversize(numArcs+1, RamUsageEstimator.NUM_BYTES_OBJECT_REF)]; System.arraycopy(arcs, 0, newArcs, 0, arcs.length); for(int arcIdx=numArcs;arcIdx[]) new Arc[0x80]; final Arc arc = new Arc(); Index: lucene/core/src/java/org/apache/lucene/util/fst/FSTEnum.java =================================================================== --- lucene/core/src/java/org/apache/lucene/util/fst/FSTEnum.java (revision 1297029) +++ lucene/core/src/java/org/apache/lucene/util/fst/FSTEnum.java (working copy) @@ -30,9 +30,9 @@ abstract class FSTEnum { protected final FST fst; - @SuppressWarnings("unchecked") protected FST.Arc[] arcs = new FST.Arc[10]; + @SuppressWarnings({"rawtypes","unchecked"}) protected FST.Arc[] arcs = new FST.Arc[10]; // outputs are cumulative - @SuppressWarnings("unchecked") protected T[] output = (T[]) new Object[10]; + @SuppressWarnings({"rawtypes","unchecked"}) protected T[] output = (T[]) new Object[10]; protected final T NO_OUTPUT; protected final FST.Arc scratchArc = new FST.Arc(); @@ -462,13 +462,13 @@ upto++; grow(); if (arcs.length <= upto) { - @SuppressWarnings("unchecked") final FST.Arc[] newArcs = + @SuppressWarnings({"rawtypes","unchecked"}) final FST.Arc[] newArcs = new FST.Arc[ArrayUtil.oversize(1+upto, RamUsageEstimator.NUM_BYTES_OBJECT_REF)]; System.arraycopy(arcs, 0, newArcs, 0, arcs.length); arcs = newArcs; } if (output.length <= upto) { - @SuppressWarnings("unchecked") final T[] newOutput = + @SuppressWarnings({"rawtypes","unchecked"}) final T[] newOutput = (T[]) new Object[ArrayUtil.oversize(1+upto, RamUsageEstimator.NUM_BYTES_OBJECT_REF)]; System.arraycopy(output, 0, newOutput, 0, output.length); output = newOutput; Index: lucene/core/src/test/org/apache/lucene/analysis/TestCachingTokenFilter.java =================================================================== --- lucene/core/src/test/org/apache/lucene/analysis/TestCachingTokenFilter.java (revision 1297029) +++ lucene/core/src/test/org/apache/lucene/analysis/TestCachingTokenFilter.java (working copy) @@ -28,6 +28,7 @@ import org.apache.lucene.index.MultiFields; import org.apache.lucene.index.DocsAndPositionsEnum; import org.apache.lucene.index.RandomIndexWriter; +import org.apache.lucene.search.DocIdSetIterator; import org.apache.lucene.store.Directory; import org.apache.lucene.util.BytesRef; @@ -76,7 +77,7 @@ "preanalyzed", new BytesRef("term1"), false); - assertTrue(termPositions.nextDoc() != termPositions.NO_MORE_DOCS); + assertTrue(termPositions.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); assertEquals(1, termPositions.freq()); assertEquals(0, termPositions.nextPosition()); @@ -85,7 +86,7 @@ "preanalyzed", new BytesRef("term2"), false); - assertTrue(termPositions.nextDoc() != termPositions.NO_MORE_DOCS); + assertTrue(termPositions.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); assertEquals(2, termPositions.freq()); assertEquals(1, termPositions.nextPosition()); assertEquals(3, termPositions.nextPosition()); @@ -95,7 +96,7 @@ "preanalyzed", new BytesRef("term3"), false); - assertTrue(termPositions.nextDoc() != termPositions.NO_MORE_DOCS); + assertTrue(termPositions.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); assertEquals(1, termPositions.freq()); assertEquals(2, termPositions.nextPosition()); reader.close(); Index: lucene/core/src/test/org/apache/lucene/codecs/appending/TestAppendingCodec.java =================================================================== --- lucene/core/src/test/org/apache/lucene/codecs/appending/TestAppendingCodec.java (revision 1297029) +++ lucene/core/src/test/org/apache/lucene/codecs/appending/TestAppendingCodec.java (working copy) @@ -35,6 +35,7 @@ import org.apache.lucene.index.TermsEnum.SeekStatus; import org.apache.lucene.index.TermsEnum; import org.apache.lucene.index.TieredMergePolicy; +import org.apache.lucene.search.DocIdSetIterator; import org.apache.lucene.store.Directory; import org.apache.lucene.store.IOContext; import org.apache.lucene.store.IndexOutput; @@ -141,10 +142,10 @@ assertEquals(SeekStatus.FOUND, te.seekCeil(new BytesRef("dog"))); assertEquals(SeekStatus.FOUND, te.seekCeil(new BytesRef("the"))); DocsEnum de = te.docs(null, null, true); - assertTrue(de.advance(0) != DocsEnum.NO_MORE_DOCS); + assertTrue(de.advance(0) != DocIdSetIterator.NO_MORE_DOCS); assertEquals(2, de.freq()); - assertTrue(de.advance(1) != DocsEnum.NO_MORE_DOCS); - assertTrue(de.advance(2) == DocsEnum.NO_MORE_DOCS); + assertTrue(de.advance(1) != DocIdSetIterator.NO_MORE_DOCS); + assertTrue(de.advance(2) == DocIdSetIterator.NO_MORE_DOCS); reader.close(); } Index: lucene/core/src/test/org/apache/lucene/index/TestCodecs.java =================================================================== --- lucene/core/src/test/org/apache/lucene/index/TestCodecs.java (revision 1297029) +++ lucene/core/src/test/org/apache/lucene/index/TestCodecs.java (working copy) @@ -81,7 +81,7 @@ NUM_TEST_ITER = atLeast(20); } - class FieldData implements Comparable { + class FieldData implements Comparable { final FieldInfo fieldInfo; final TermData[] terms; final boolean omitTF; @@ -102,8 +102,8 @@ Arrays.sort(terms); } - public int compareTo(final Object other) { - return fieldInfo.name.compareTo(((FieldData) other).fieldInfo.name); + public int compareTo(final FieldData other) { + return fieldInfo.name.compareTo(other.fieldInfo.name); } public void write(final FieldsConsumer consumer) throws Throwable { @@ -133,7 +133,7 @@ } } - class TermData implements Comparable { + class TermData implements Comparable { String text2; final BytesRef text; int[] docs; @@ -147,8 +147,8 @@ this.positions = positions; } - public int compareTo(final Object o) { - return text.compareTo(((TermData) o).text); + public int compareTo(final TermData o) { + return text.compareTo(o.text); } public long write(final TermsConsumer termsConsumer) throws Throwable { @@ -281,7 +281,7 @@ for(int iter=0;iter<2;iter++) { docsEnum = _TestUtil.docs(random, termsEnum, null, docsEnum, false); assertEquals(terms[i].docs[0], docsEnum.nextDoc()); - assertEquals(DocsEnum.NO_MORE_DOCS, docsEnum.nextDoc()); + assertEquals(DocIdSetIterator.NO_MORE_DOCS, docsEnum.nextDoc()); } } assertNull(termsEnum.next()); @@ -439,7 +439,7 @@ assertEquals(positions[i].pos, pos); if (positions[i].payload != null) { assertTrue(posEnum.hasPayload()); - if (TestCodecs.random.nextInt(3) < 2) { + if (LuceneTestCase.random.nextInt(3) < 2) { // Verify the payload bytes final BytesRef otherPayload = posEnum.getPayload(); assertTrue("expected=" + positions[i].payload.toString() + " got=" + otherPayload.toString(), positions[i].payload.equals(otherPayload)); @@ -453,7 +453,7 @@ public void _run() throws Throwable { for(int iter=0;iter= 1) { - final int inc = 1+TestCodecs.random.nextInt(left-1); + if (LuceneTestCase.random.nextInt(3) == 1 && left >= 1) { + final int inc = 1+LuceneTestCase.random.nextInt(left-1); upto2 += inc; - if (TestCodecs.random.nextInt(2) == 1) { + if (LuceneTestCase.random.nextInt(2) == 1) { doc = docs.advance(term.docs[upto2]); assertEquals(term.docs[upto2], doc); } else { @@ -597,7 +597,7 @@ assertEquals(term.docs[upto2], doc); if (!field.omitTF) { assertEquals(term.positions[upto2].length, postings.freq()); - if (TestCodecs.random.nextInt(2) == 1) { + if (LuceneTestCase.random.nextInt(2) == 1) { this.verifyPositions(term.positions[upto2], postings); } } Index: lucene/core/src/test/org/apache/lucene/index/TestDirectoryReader.java =================================================================== --- lucene/core/src/test/org/apache/lucene/index/TestDirectoryReader.java (revision 1297029) +++ lucene/core/src/test/org/apache/lucene/index/TestDirectoryReader.java (working copy) @@ -110,7 +110,7 @@ // This should blow up if we forget to check that the TermEnum is from the same // reader as the TermDocs. - while (td.nextDoc() != td.NO_MORE_DOCS) ret += td.docID(); + while (td.nextDoc() != DocIdSetIterator.NO_MORE_DOCS) ret += td.docID(); // really a dummy assert to ensure that we got some docs and to ensure that // nothing is eliminated by hotspot Index: lucene/core/src/test/org/apache/lucene/index/TestDocsAndPositions.java =================================================================== --- lucene/core/src/test/org/apache/lucene/index/TestDocsAndPositions.java (revision 1297029) +++ lucene/core/src/test/org/apache/lucene/index/TestDocsAndPositions.java (working copy) @@ -84,7 +84,7 @@ assertEquals(msg, 20, docsAndPosEnum.nextPosition()); assertEquals(msg, 4, docsAndPosEnum.freq()); assertEquals(msg, 30, docsAndPosEnum.nextPosition()); - } while (docsAndPosEnum.nextDoc() != DocsAndPositionsEnum.NO_MORE_DOCS); + } while (docsAndPosEnum.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); } } reader.close(); @@ -156,7 +156,7 @@ // now run through the scorer and check if all positions are there... do { int docID = docsAndPosEnum.docID(); - if (docID == DocsAndPositionsEnum.NO_MORE_DOCS) { + if (docID == DocIdSetIterator.NO_MORE_DOCS) { break; } Integer[] pos = positionsInDoc[atomicReaderContext.docBase + docID]; @@ -177,7 +177,7 @@ .advance(docID + 1 + random.nextInt((maxDoc - docID))); } - } while (docsAndPosEnum.nextDoc() != DocsAndPositionsEnum.NO_MORE_DOCS); + } while (docsAndPosEnum.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); } } @@ -234,7 +234,7 @@ int next = findNext(freqInDoc, context.docBase+j+1, context.docBase + maxDoc) - context.docBase; int advancedTo = docsEnum.advance(next); if (next >= maxDoc) { - assertEquals(DocsEnum.NO_MORE_DOCS, advancedTo); + assertEquals(DocIdSetIterator.NO_MORE_DOCS, advancedTo); } else { assertTrue("advanced to: " +advancedTo + " but should be <= " + next, next >= advancedTo); } @@ -243,7 +243,7 @@ } } } - assertEquals("docBase: " + context.docBase + " maxDoc: " + maxDoc + " " + docsEnum.getClass(), DocsEnum.NO_MORE_DOCS, docsEnum.docID()); + assertEquals("docBase: " + context.docBase + " maxDoc: " + maxDoc + " " + docsEnum.getClass(), DocIdSetIterator.NO_MORE_DOCS, docsEnum.docID()); } } Index: lucene/core/src/test/org/apache/lucene/index/TestDocumentWriter.java =================================================================== --- lucene/core/src/test/org/apache/lucene/index/TestDocumentWriter.java (revision 1297029) +++ lucene/core/src/test/org/apache/lucene/index/TestDocumentWriter.java (working copy) @@ -30,6 +30,7 @@ import org.apache.lucene.document.StringField; import org.apache.lucene.document.TextField; import org.apache.lucene.index.FieldInfo.IndexOptions; +import org.apache.lucene.search.DocIdSetIterator; import org.apache.lucene.store.Directory; import org.apache.lucene.util.AttributeSource; import org.apache.lucene.util.BytesRef; @@ -129,7 +130,7 @@ DocsAndPositionsEnum termPositions = MultiFields.getTermPositionsEnum(reader, MultiFields.getLiveDocs(reader), "repeated", new BytesRef("repeated"), false); - assertTrue(termPositions.nextDoc() != termPositions.NO_MORE_DOCS); + assertTrue(termPositions.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); int freq = termPositions.freq(); assertEquals(2, freq); assertEquals(0, termPositions.nextPosition()); @@ -200,7 +201,7 @@ SegmentReader reader = new SegmentReader(info, DirectoryReader.DEFAULT_TERMS_INDEX_DIVISOR, newIOContext(random)); DocsAndPositionsEnum termPositions = MultiFields.getTermPositionsEnum(reader, reader.getLiveDocs(), "f1", new BytesRef("a"), false); - assertTrue(termPositions.nextDoc() != termPositions.NO_MORE_DOCS); + assertTrue(termPositions.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); int freq = termPositions.freq(); assertEquals(3, freq); assertEquals(0, termPositions.nextPosition()); @@ -244,18 +245,18 @@ SegmentReader reader = new SegmentReader(info, DirectoryReader.DEFAULT_TERMS_INDEX_DIVISOR, newIOContext(random)); DocsAndPositionsEnum termPositions = reader.termPositionsEnum(reader.getLiveDocs(), "preanalyzed", new BytesRef("term1"), false); - assertTrue(termPositions.nextDoc() != termPositions.NO_MORE_DOCS); + assertTrue(termPositions.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); assertEquals(1, termPositions.freq()); assertEquals(0, termPositions.nextPosition()); termPositions = reader.termPositionsEnum(reader.getLiveDocs(), "preanalyzed", new BytesRef("term2"), false); - assertTrue(termPositions.nextDoc() != termPositions.NO_MORE_DOCS); + assertTrue(termPositions.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); assertEquals(2, termPositions.freq()); assertEquals(1, termPositions.nextPosition()); assertEquals(3, termPositions.nextPosition()); termPositions = reader.termPositionsEnum(reader.getLiveDocs(), "preanalyzed", new BytesRef("term3"), false); - assertTrue(termPositions.nextDoc() != termPositions.NO_MORE_DOCS); + assertTrue(termPositions.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); assertEquals(1, termPositions.freq()); assertEquals(2, termPositions.nextPosition()); reader.close(); Index: lucene/core/src/test/org/apache/lucene/index/TestDocValuesIndexing.java =================================================================== --- lucene/core/src/test/org/apache/lucene/index/TestDocValuesIndexing.java (revision 1297029) +++ lucene/core/src/test/org/apache/lucene/index/TestDocValuesIndexing.java (working copy) @@ -816,7 +816,7 @@ assertEquals(1, docFreq); DocsEnum termDocsEnum = reader.termDocsEnum(null, term.field, term.bytes, false); int nextDoc = termDocsEnum.nextDoc(); - assertEquals(DocsEnum.NO_MORE_DOCS, termDocsEnum.nextDoc()); + assertEquals(DocIdSetIterator.NO_MORE_DOCS, termDocsEnum.nextDoc()); return nextDoc; } Index: lucene/core/src/test/org/apache/lucene/index/TestFilterAtomicReader.java =================================================================== --- lucene/core/src/test/org/apache/lucene/index/TestFilterAtomicReader.java (revision 1297029) +++ lucene/core/src/test/org/apache/lucene/index/TestFilterAtomicReader.java (working copy) @@ -24,6 +24,7 @@ import org.apache.lucene.analysis.MockAnalyzer; import org.apache.lucene.document.Document; import org.apache.lucene.document.TextField; +import org.apache.lucene.search.DocIdSetIterator; import org.apache.lucene.store.Directory; import org.apache.lucene.store.MockDirectoryWrapper; import org.apache.lucene.util.Bits; @@ -165,7 +166,7 @@ DocsAndPositionsEnum positions = terms.docsAndPositions(MultiFields.getLiveDocs(reader), null, false); - while (positions.nextDoc() != DocsEnum.NO_MORE_DOCS) { + while (positions.nextDoc() != DocIdSetIterator.NO_MORE_DOCS) { assertTrue((positions.docID() % 2) == 1); } Index: lucene/core/src/test/org/apache/lucene/index/TestIndexableField.java =================================================================== --- lucene/core/src/test/org/apache/lucene/index/TestIndexableField.java (revision 1297029) +++ lucene/core/src/test/org/apache/lucene/index/TestIndexableField.java (working copy) @@ -28,6 +28,7 @@ import org.apache.lucene.document.StringField; import org.apache.lucene.search.BooleanClause; import org.apache.lucene.search.BooleanQuery; +import org.apache.lucene.search.DocIdSetIterator; import org.apache.lucene.search.IndexSearcher; import org.apache.lucene.search.TermQuery; import org.apache.lucene.search.TopDocs; @@ -264,14 +265,14 @@ assertEquals(new BytesRef(""+counter), termsEnum.next()); assertEquals(1, termsEnum.totalTermFreq()); DocsAndPositionsEnum dpEnum = termsEnum.docsAndPositions(null, null, false); - assertTrue(dpEnum.nextDoc() != DocsEnum.NO_MORE_DOCS); + assertTrue(dpEnum.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); assertEquals(1, dpEnum.freq()); assertEquals(1, dpEnum.nextPosition()); assertEquals(new BytesRef("text"), termsEnum.next()); assertEquals(1, termsEnum.totalTermFreq()); dpEnum = termsEnum.docsAndPositions(null, dpEnum, false); - assertTrue(dpEnum.nextDoc() != DocsEnum.NO_MORE_DOCS); + assertTrue(dpEnum.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); assertEquals(1, dpEnum.freq()); assertEquals(0, dpEnum.nextPosition()); Index: lucene/core/src/test/org/apache/lucene/index/TestIndexWriter.java =================================================================== --- lucene/core/src/test/org/apache/lucene/index/TestIndexWriter.java (revision 1297029) +++ lucene/core/src/test/org/apache/lucene/index/TestIndexWriter.java (working copy) @@ -969,14 +969,14 @@ assertNotNull(termsEnum.next()); DocsAndPositionsEnum dpEnum = termsEnum.docsAndPositions(null, null, false); assertNotNull(dpEnum); - assertTrue(dpEnum.nextDoc() != DocsEnum.NO_MORE_DOCS); + assertTrue(dpEnum.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); assertEquals(1, dpEnum.freq()); assertEquals(100, dpEnum.nextPosition()); assertNotNull(termsEnum.next()); dpEnum = termsEnum.docsAndPositions(null, dpEnum, false); assertNotNull(dpEnum); - assertTrue(dpEnum.nextDoc() != DocsEnum.NO_MORE_DOCS); + assertTrue(dpEnum.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); assertEquals(1, dpEnum.freq()); assertEquals(101, dpEnum.nextPosition()); assertNull(termsEnum.next()); Index: lucene/core/src/test/org/apache/lucene/index/TestIndexWriterReader.java =================================================================== --- lucene/core/src/test/org/apache/lucene/index/TestIndexWriterReader.java (revision 1297029) +++ lucene/core/src/test/org/apache/lucene/index/TestIndexWriterReader.java (working copy) @@ -31,6 +31,7 @@ import org.apache.lucene.document.Field; import org.apache.lucene.document.StringField; import org.apache.lucene.document.TextField; +import org.apache.lucene.search.DocIdSetIterator; import org.apache.lucene.search.IndexSearcher; import org.apache.lucene.search.Query; import org.apache.lucene.search.TermQuery; @@ -57,7 +58,7 @@ false); if (td != null) { - while (td.nextDoc() != DocsEnum.NO_MORE_DOCS) { + while (td.nextDoc() != DocIdSetIterator.NO_MORE_DOCS) { td.docID(); count++; } Index: lucene/core/src/test/org/apache/lucene/index/TestIndexWriterWithThreads.java =================================================================== --- lucene/core/src/test/org/apache/lucene/index/TestIndexWriterWithThreads.java (revision 1297029) +++ lucene/core/src/test/org/apache/lucene/index/TestIndexWriterWithThreads.java (working copy) @@ -25,6 +25,7 @@ import org.apache.lucene.document.Field; import org.apache.lucene.document.FieldType; import org.apache.lucene.document.TextField; +import org.apache.lucene.search.DocIdSetIterator; import org.apache.lucene.store.AlreadyClosedException; import org.apache.lucene.store.Directory; import org.apache.lucene.store.MockDirectoryWrapper; @@ -217,7 +218,7 @@ null, false); int count = 0; - while(tdocs.nextDoc() != DocsEnum.NO_MORE_DOCS) { + while(tdocs.nextDoc() != DocIdSetIterator.NO_MORE_DOCS) { count++; } assertTrue(count > 0); Index: lucene/core/src/test/org/apache/lucene/index/TestLongPostings.java =================================================================== --- lucene/core/src/test/org/apache/lucene/index/TestLongPostings.java (revision 1297029) +++ lucene/core/src/test/org/apache/lucene/index/TestLongPostings.java (working copy) @@ -29,6 +29,7 @@ import org.apache.lucene.document.FieldType; import org.apache.lucene.index.FieldInfo.IndexOptions; import org.apache.lucene.document.TextField; +import org.apache.lucene.search.DocIdSetIterator; import org.apache.lucene.store.Directory; import org.apache.lucene.util.BytesRef; import org.apache.lucene.util.LuceneTestCase; @@ -176,7 +177,7 @@ final DocsAndPositionsEnum postings = MultiFields.getTermPositionsEnum(r, null, "field", new BytesRef(term), false); int docID = -1; - while(docID < DocsEnum.NO_MORE_DOCS) { + while(docID < DocIdSetIterator.NO_MORE_DOCS) { final int what = random.nextInt(3); if (what == 0) { if (VERBOSE) { @@ -199,7 +200,7 @@ System.out.println(" got docID=" + docID); } assertEquals(expected, docID); - if (docID == DocsEnum.NO_MORE_DOCS) { + if (docID == DocIdSetIterator.NO_MORE_DOCS) { break; } @@ -241,7 +242,7 @@ System.out.println(" got docID=" + docID); } assertEquals(expected, docID); - if (docID == DocsEnum.NO_MORE_DOCS) { + if (docID == DocIdSetIterator.NO_MORE_DOCS) { break; } @@ -380,7 +381,7 @@ assert docs != null; int docID = -1; - while(docID < DocsEnum.NO_MORE_DOCS) { + while(docID < DocIdSetIterator.NO_MORE_DOCS) { final int what = random.nextInt(3); if (what == 0) { if (VERBOSE) { @@ -403,7 +404,7 @@ System.out.println(" got docID=" + docID); } assertEquals(expected, docID); - if (docID == DocsEnum.NO_MORE_DOCS) { + if (docID == DocIdSetIterator.NO_MORE_DOCS) { break; } @@ -439,7 +440,7 @@ System.out.println(" got docID=" + docID); } assertEquals(expected, docID); - if (docID == DocsEnum.NO_MORE_DOCS) { + if (docID == DocIdSetIterator.NO_MORE_DOCS) { break; } Index: lucene/core/src/test/org/apache/lucene/index/TestMultiFields.java =================================================================== --- lucene/core/src/test/org/apache/lucene/index/TestMultiFields.java (revision 1297029) +++ lucene/core/src/test/org/apache/lucene/index/TestMultiFields.java (working copy) @@ -17,6 +17,7 @@ * limitations under the License. */ +import org.apache.lucene.search.DocIdSetIterator; import org.apache.lucene.store.*; import org.apache.lucene.util.*; import org.apache.lucene.document.*; @@ -128,7 +129,7 @@ assertEquals(docID, docsEnum.nextDoc()); } } - assertEquals(docsEnum.NO_MORE_DOCS, docsEnum.nextDoc()); + assertEquals(DocIdSetIterator.NO_MORE_DOCS, docsEnum.nextDoc()); } reader.close(); Index: lucene/core/src/test/org/apache/lucene/index/TestParallelTermEnum.java =================================================================== --- lucene/core/src/test/org/apache/lucene/index/TestParallelTermEnum.java (revision 1297029) +++ lucene/core/src/test/org/apache/lucene/index/TestParallelTermEnum.java (working copy) @@ -22,6 +22,7 @@ import org.apache.lucene.analysis.MockAnalyzer; import org.apache.lucene.document.Document; import org.apache.lucene.document.TextField; +import org.apache.lucene.search.DocIdSetIterator; import org.apache.lucene.store.Directory; import org.apache.lucene.util.Bits; import org.apache.lucene.util.BytesRef; @@ -79,9 +80,9 @@ assertNotNull(b); assertEquals(t, b.utf8ToString()); DocsEnum td = _TestUtil.docs(random, te, liveDocs, null, false); - assertTrue(td.nextDoc() != DocsEnum.NO_MORE_DOCS); + assertTrue(td.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); assertEquals(0, td.docID()); - assertEquals(td.nextDoc(), DocsEnum.NO_MORE_DOCS); + assertEquals(td.nextDoc(), DocIdSetIterator.NO_MORE_DOCS); } assertNull(te.next()); } Index: lucene/core/src/test/org/apache/lucene/index/TestPerSegmentDeletes.java =================================================================== --- lucene/core/src/test/org/apache/lucene/index/TestPerSegmentDeletes.java (revision 1297029) +++ lucene/core/src/test/org/apache/lucene/index/TestPerSegmentDeletes.java (working copy) @@ -24,6 +24,7 @@ import java.util.Random; import org.apache.lucene.analysis.MockAnalyzer; +import org.apache.lucene.search.DocIdSetIterator; import org.apache.lucene.store.Directory; import org.apache.lucene.store.MockDirectoryWrapper; import org.apache.lucene.store.RAMDirectory; @@ -233,7 +234,7 @@ public static int[] toArray(DocsEnum docsEnum) throws IOException { List docs = new ArrayList(); - while (docsEnum.nextDoc() != DocsEnum.NO_MORE_DOCS) { + while (docsEnum.nextDoc() != DocIdSetIterator.NO_MORE_DOCS) { int docID = docsEnum.docID(); docs.add(docID); } Index: lucene/core/src/test/org/apache/lucene/index/TestPostingsOffsets.java =================================================================== --- lucene/core/src/test/org/apache/lucene/index/TestPostingsOffsets.java (revision 1297029) +++ lucene/core/src/test/org/apache/lucene/index/TestPostingsOffsets.java (working copy) @@ -88,7 +88,7 @@ assertEquals(2, dp.nextPosition()); assertEquals(9, dp.startOffset()); assertEquals(17, dp.endOffset()); - assertEquals(DocsEnum.NO_MORE_DOCS, dp.nextDoc()); + assertEquals(DocIdSetIterator.NO_MORE_DOCS, dp.nextDoc()); dp = MultiFields.getTermPositionsEnum(r, null, "content", new BytesRef("b"), true); assertNotNull(dp); @@ -97,7 +97,7 @@ assertEquals(1, dp.nextPosition()); assertEquals(8, dp.startOffset()); assertEquals(9, dp.endOffset()); - assertEquals(DocsEnum.NO_MORE_DOCS, dp.nextDoc()); + assertEquals(DocIdSetIterator.NO_MORE_DOCS, dp.nextDoc()); dp = MultiFields.getTermPositionsEnum(r, null, "content", new BytesRef("c"), true); assertNotNull(dp); @@ -106,7 +106,7 @@ assertEquals(3, dp.nextPosition()); assertEquals(19, dp.startOffset()); assertEquals(50, dp.endOffset()); - assertEquals(DocsEnum.NO_MORE_DOCS, dp.nextDoc()); + assertEquals(DocIdSetIterator.NO_MORE_DOCS, dp.nextDoc()); r.close(); dir.close(); @@ -156,7 +156,7 @@ for (String term : terms) { DocsAndPositionsEnum dp = MultiFields.getTermPositionsEnum(reader, null, "numbers", new BytesRef(term), true); int doc; - while((doc = dp.nextDoc()) != DocsEnum.NO_MORE_DOCS) { + while((doc = dp.nextDoc()) != DocIdSetIterator.NO_MORE_DOCS) { String storedNumbers = reader.document(doc).get("numbers"); int freq = dp.freq(); for (int i = 0; i < freq; i++) { @@ -304,7 +304,7 @@ assertNotNull(docs); int doc; //System.out.println(" doc/freq"); - while((doc = docs.nextDoc()) != DocsEnum.NO_MORE_DOCS) { + while((doc = docs.nextDoc()) != DocIdSetIterator.NO_MORE_DOCS) { final List expected = actualTokens.get(term).get(docIDToID[doc]); //System.out.println(" doc=" + docIDToID[doc] + " docID=" + doc + " " + expected.size() + " freq"); assertNotNull(expected); @@ -314,7 +314,7 @@ docsAndPositions = termsEnum.docsAndPositions(null, docsAndPositions, false); assertNotNull(docsAndPositions); //System.out.println(" doc/freq/pos"); - while((doc = docsAndPositions.nextDoc()) != DocsEnum.NO_MORE_DOCS) { + while((doc = docsAndPositions.nextDoc()) != DocIdSetIterator.NO_MORE_DOCS) { final List expected = actualTokens.get(term).get(docIDToID[doc]); //System.out.println(" doc=" + docIDToID[doc] + " " + expected.size() + " freq"); assertNotNull(expected); @@ -329,7 +329,7 @@ docsAndPositionsAndOffsets = termsEnum.docsAndPositions(null, docsAndPositions, true); assertNotNull(docsAndPositionsAndOffsets); //System.out.println(" doc/freq/pos/offs"); - while((doc = docsAndPositions.nextDoc()) != DocsEnum.NO_MORE_DOCS) { + while((doc = docsAndPositions.nextDoc()) != DocIdSetIterator.NO_MORE_DOCS) { final List expected = actualTokens.get(term).get(docIDToID[doc]); //System.out.println(" doc=" + docIDToID[doc] + " " + expected.size() + " freq"); assertNotNull(expected); Index: lucene/core/src/test/org/apache/lucene/index/TestSegmentMerger.java =================================================================== --- lucene/core/src/test/org/apache/lucene/index/TestSegmentMerger.java (revision 1297029) +++ lucene/core/src/test/org/apache/lucene/index/TestSegmentMerger.java (working copy) @@ -23,6 +23,7 @@ import org.apache.lucene.codecs.Codec; import org.apache.lucene.document.Document; import org.apache.lucene.document.TextField; +import org.apache.lucene.search.DocIdSetIterator; import org.apache.lucene.store.Directory; import org.apache.lucene.util.BytesRef; import org.apache.lucene.util.InfoStream; @@ -105,7 +106,7 @@ null, false); assertTrue(termDocs != null); - assertTrue(termDocs.nextDoc() != DocsEnum.NO_MORE_DOCS); + assertTrue(termDocs.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); int tvCount = 0; for(FieldInfo fieldInfo : mergedReader.getFieldInfos()) { Index: lucene/core/src/test/org/apache/lucene/index/TestSegmentReader.java =================================================================== --- lucene/core/src/test/org/apache/lucene/index/TestSegmentReader.java (revision 1297029) +++ lucene/core/src/test/org/apache/lucene/index/TestSegmentReader.java (working copy) @@ -23,6 +23,7 @@ import java.util.List; import org.apache.lucene.document.Document; +import org.apache.lucene.search.DocIdSetIterator; import org.apache.lucene.store.Directory; import org.apache.lucene.store.IOContext; import org.apache.lucene.util.BytesRef; @@ -133,7 +134,7 @@ MultiFields.getLiveDocs(reader), null, false); - assertTrue(termDocs.nextDoc() != DocsEnum.NO_MORE_DOCS); + assertTrue(termDocs.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); termDocs = _TestUtil.docs(random, reader, DocHelper.NO_NORMS_KEY, @@ -142,7 +143,7 @@ null, false); - assertTrue(termDocs.nextDoc() != DocsEnum.NO_MORE_DOCS); + assertTrue(termDocs.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); DocsAndPositionsEnum positions = MultiFields.getTermPositionsEnum(reader, @@ -152,7 +153,7 @@ false); // NOTE: prior rev of this test was failing to first // call next here: - assertTrue(positions.nextDoc() != DocsEnum.NO_MORE_DOCS); + assertTrue(positions.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); assertTrue(positions.docID() == 0); assertTrue(positions.nextPosition() >= 0); } Index: lucene/core/src/test/org/apache/lucene/index/TestSegmentTermDocs.java =================================================================== --- lucene/core/src/test/org/apache/lucene/index/TestSegmentTermDocs.java (revision 1297029) +++ lucene/core/src/test/org/apache/lucene/index/TestSegmentTermDocs.java (working copy) @@ -22,6 +22,7 @@ import org.apache.lucene.analysis.MockAnalyzer; import org.apache.lucene.document.Document; import org.apache.lucene.document.TextField; +import org.apache.lucene.search.DocIdSetIterator; import org.apache.lucene.store.Directory; import org.apache.lucene.util.BytesRef; import org.apache.lucene.util.LuceneTestCase; @@ -63,7 +64,7 @@ TermsEnum terms = reader.fields().terms(DocHelper.TEXT_FIELD_2_KEY).iterator(null); terms.seekCeil(new BytesRef("field")); DocsEnum termDocs = _TestUtil.docs(random, terms, reader.getLiveDocs(), null, true); - if (termDocs.nextDoc() != DocsEnum.NO_MORE_DOCS) { + if (termDocs.nextDoc() != DocIdSetIterator.NO_MORE_DOCS) { int docId = termDocs.docID(); assertTrue(docId == 0); int freq = termDocs.freq(); @@ -142,19 +143,19 @@ // without optimization (assumption skipInterval == 16) // with next - assertTrue(tdocs.nextDoc() != DocsEnum.NO_MORE_DOCS); + assertTrue(tdocs.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); assertEquals(0, tdocs.docID()); assertEquals(4, tdocs.freq()); - assertTrue(tdocs.nextDoc() != DocsEnum.NO_MORE_DOCS); + assertTrue(tdocs.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); assertEquals(1, tdocs.docID()); assertEquals(4, tdocs.freq()); - assertTrue(tdocs.advance(0) != DocsEnum.NO_MORE_DOCS); + assertTrue(tdocs.advance(0) != DocIdSetIterator.NO_MORE_DOCS); assertEquals(2, tdocs.docID()); - assertTrue(tdocs.advance(4) != DocsEnum.NO_MORE_DOCS); + assertTrue(tdocs.advance(4) != DocIdSetIterator.NO_MORE_DOCS); assertEquals(4, tdocs.docID()); - assertTrue(tdocs.advance(9) != DocsEnum.NO_MORE_DOCS); + assertTrue(tdocs.advance(9) != DocIdSetIterator.NO_MORE_DOCS); assertEquals(9, tdocs.docID()); - assertFalse(tdocs.advance(10) != DocsEnum.NO_MORE_DOCS); + assertFalse(tdocs.advance(10) != DocIdSetIterator.NO_MORE_DOCS); // without next tdocs = _TestUtil.docs(random, reader, @@ -164,13 +165,13 @@ null, false); - assertTrue(tdocs.advance(0) != DocsEnum.NO_MORE_DOCS); + assertTrue(tdocs.advance(0) != DocIdSetIterator.NO_MORE_DOCS); assertEquals(0, tdocs.docID()); - assertTrue(tdocs.advance(4) != DocsEnum.NO_MORE_DOCS); + assertTrue(tdocs.advance(4) != DocIdSetIterator.NO_MORE_DOCS); assertEquals(4, tdocs.docID()); - assertTrue(tdocs.advance(9) != DocsEnum.NO_MORE_DOCS); + assertTrue(tdocs.advance(9) != DocIdSetIterator.NO_MORE_DOCS); assertEquals(9, tdocs.docID()); - assertFalse(tdocs.advance(10) != DocsEnum.NO_MORE_DOCS); + assertFalse(tdocs.advance(10) != DocIdSetIterator.NO_MORE_DOCS); // exactly skipInterval documents and therefore with optimization @@ -182,21 +183,21 @@ null, true); - assertTrue(tdocs.nextDoc() != DocsEnum.NO_MORE_DOCS); + assertTrue(tdocs.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); assertEquals(10, tdocs.docID()); assertEquals(4, tdocs.freq()); - assertTrue(tdocs.nextDoc() != DocsEnum.NO_MORE_DOCS); + assertTrue(tdocs.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); assertEquals(11, tdocs.docID()); assertEquals(4, tdocs.freq()); - assertTrue(tdocs.advance(5) != DocsEnum.NO_MORE_DOCS); + assertTrue(tdocs.advance(5) != DocIdSetIterator.NO_MORE_DOCS); assertEquals(12, tdocs.docID()); - assertTrue(tdocs.advance(15) != DocsEnum.NO_MORE_DOCS); + assertTrue(tdocs.advance(15) != DocIdSetIterator.NO_MORE_DOCS); assertEquals(15, tdocs.docID()); - assertTrue(tdocs.advance(24) != DocsEnum.NO_MORE_DOCS); + assertTrue(tdocs.advance(24) != DocIdSetIterator.NO_MORE_DOCS); assertEquals(24, tdocs.docID()); - assertTrue(tdocs.advance(25) != DocsEnum.NO_MORE_DOCS); + assertTrue(tdocs.advance(25) != DocIdSetIterator.NO_MORE_DOCS); assertEquals(25, tdocs.docID()); - assertFalse(tdocs.advance(26) != DocsEnum.NO_MORE_DOCS); + assertFalse(tdocs.advance(26) != DocIdSetIterator.NO_MORE_DOCS); // without next tdocs = _TestUtil.docs(random, reader, @@ -206,15 +207,15 @@ null, true); - assertTrue(tdocs.advance(5) != DocsEnum.NO_MORE_DOCS); + assertTrue(tdocs.advance(5) != DocIdSetIterator.NO_MORE_DOCS); assertEquals(10, tdocs.docID()); - assertTrue(tdocs.advance(15) != DocsEnum.NO_MORE_DOCS); + assertTrue(tdocs.advance(15) != DocIdSetIterator.NO_MORE_DOCS); assertEquals(15, tdocs.docID()); - assertTrue(tdocs.advance(24) != DocsEnum.NO_MORE_DOCS); + assertTrue(tdocs.advance(24) != DocIdSetIterator.NO_MORE_DOCS); assertEquals(24, tdocs.docID()); - assertTrue(tdocs.advance(25) != DocsEnum.NO_MORE_DOCS); + assertTrue(tdocs.advance(25) != DocIdSetIterator.NO_MORE_DOCS); assertEquals(25, tdocs.docID()); - assertFalse(tdocs.advance(26) != DocsEnum.NO_MORE_DOCS); + assertFalse(tdocs.advance(26) != DocIdSetIterator.NO_MORE_DOCS); // much more than skipInterval documents and therefore with optimization @@ -226,23 +227,23 @@ null, true); - assertTrue(tdocs.nextDoc() != DocsEnum.NO_MORE_DOCS); + assertTrue(tdocs.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); assertEquals(26, tdocs.docID()); assertEquals(4, tdocs.freq()); - assertTrue(tdocs.nextDoc() != DocsEnum.NO_MORE_DOCS); + assertTrue(tdocs.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); assertEquals(27, tdocs.docID()); assertEquals(4, tdocs.freq()); - assertTrue(tdocs.advance(5) != DocsEnum.NO_MORE_DOCS); + assertTrue(tdocs.advance(5) != DocIdSetIterator.NO_MORE_DOCS); assertEquals(28, tdocs.docID()); - assertTrue(tdocs.advance(40) != DocsEnum.NO_MORE_DOCS); + assertTrue(tdocs.advance(40) != DocIdSetIterator.NO_MORE_DOCS); assertEquals(40, tdocs.docID()); - assertTrue(tdocs.advance(57) != DocsEnum.NO_MORE_DOCS); + assertTrue(tdocs.advance(57) != DocIdSetIterator.NO_MORE_DOCS); assertEquals(57, tdocs.docID()); - assertTrue(tdocs.advance(74) != DocsEnum.NO_MORE_DOCS); + assertTrue(tdocs.advance(74) != DocIdSetIterator.NO_MORE_DOCS); assertEquals(74, tdocs.docID()); - assertTrue(tdocs.advance(75) != DocsEnum.NO_MORE_DOCS); + assertTrue(tdocs.advance(75) != DocIdSetIterator.NO_MORE_DOCS); assertEquals(75, tdocs.docID()); - assertFalse(tdocs.advance(76) != DocsEnum.NO_MORE_DOCS); + assertFalse(tdocs.advance(76) != DocIdSetIterator.NO_MORE_DOCS); //without next tdocs = _TestUtil.docs(random, reader, @@ -251,17 +252,17 @@ MultiFields.getLiveDocs(reader), null, false); - assertTrue(tdocs.advance(5) != DocsEnum.NO_MORE_DOCS); + assertTrue(tdocs.advance(5) != DocIdSetIterator.NO_MORE_DOCS); assertEquals(26, tdocs.docID()); - assertTrue(tdocs.advance(40) != DocsEnum.NO_MORE_DOCS); + assertTrue(tdocs.advance(40) != DocIdSetIterator.NO_MORE_DOCS); assertEquals(40, tdocs.docID()); - assertTrue(tdocs.advance(57) != DocsEnum.NO_MORE_DOCS); + assertTrue(tdocs.advance(57) != DocIdSetIterator.NO_MORE_DOCS); assertEquals(57, tdocs.docID()); - assertTrue(tdocs.advance(74) != DocsEnum.NO_MORE_DOCS); + assertTrue(tdocs.advance(74) != DocIdSetIterator.NO_MORE_DOCS); assertEquals(74, tdocs.docID()); - assertTrue(tdocs.advance(75) != DocsEnum.NO_MORE_DOCS); + assertTrue(tdocs.advance(75) != DocIdSetIterator.NO_MORE_DOCS); assertEquals(75, tdocs.docID()); - assertFalse(tdocs.advance(76) != DocsEnum.NO_MORE_DOCS); + assertFalse(tdocs.advance(76) != DocIdSetIterator.NO_MORE_DOCS); reader.close(); dir.close(); Index: lucene/core/src/test/org/apache/lucene/index/TestStressAdvance.java =================================================================== --- lucene/core/src/test/org/apache/lucene/index/TestStressAdvance.java (revision 1297029) +++ lucene/core/src/test/org/apache/lucene/index/TestStressAdvance.java (working copy) @@ -22,6 +22,7 @@ import java.util.Set; import org.apache.lucene.util.*; +import org.apache.lucene.search.DocIdSetIterator; import org.apache.lucene.store.*; import org.apache.lucene.document.*; @@ -117,14 +118,14 @@ } if (upto == expected.size()) { if (VERBOSE) { - System.out.println(" expect docID=" + DocsEnum.NO_MORE_DOCS + " actual=" + docID); + System.out.println(" expect docID=" + DocIdSetIterator.NO_MORE_DOCS + " actual=" + docID); } - assertEquals(DocsEnum.NO_MORE_DOCS, docID); + assertEquals(DocIdSetIterator.NO_MORE_DOCS, docID); } else { if (VERBOSE) { System.out.println(" expect docID=" + expected.get(upto) + " actual=" + docID); } - assertTrue(docID != DocsEnum.NO_MORE_DOCS); + assertTrue(docID != DocIdSetIterator.NO_MORE_DOCS); assertEquals(expected.get(upto).intValue(), docID); } } Index: lucene/core/src/test/org/apache/lucene/index/TestStressIndexing2.java =================================================================== --- lucene/core/src/test/org/apache/lucene/index/TestStressIndexing2.java (revision 1297029) +++ lucene/core/src/test/org/apache/lucene/index/TestStressIndexing2.java (working copy) @@ -34,6 +34,7 @@ import org.apache.lucene.document.FieldType; import org.apache.lucene.document.TextField; import org.apache.lucene.index.IndexWriterConfig.OpenMode; +import org.apache.lucene.search.DocIdSetIterator; import org.apache.lucene.search.TermQuery; import org.apache.lucene.store.Directory; import org.apache.lucene.util.*; @@ -336,7 +337,7 @@ DocsEnum docs = null; while(termsEnum.next() != null) { docs = _TestUtil.docs(random, termsEnum, null, docs, false); - while(docs.nextDoc() != DocsEnum.NO_MORE_DOCS) { + while(docs.nextDoc() != DocIdSetIterator.NO_MORE_DOCS) { fail("r1 is not empty but r2 is"); } } @@ -362,18 +363,18 @@ termDocs2 = null; } - if (termDocs1.nextDoc() == DocsEnum.NO_MORE_DOCS) { + if (termDocs1.nextDoc() == DocIdSetIterator.NO_MORE_DOCS) { // This doc is deleted and wasn't replaced - assertTrue(termDocs2 == null || termDocs2.nextDoc() == DocsEnum.NO_MORE_DOCS); + assertTrue(termDocs2 == null || termDocs2.nextDoc() == DocIdSetIterator.NO_MORE_DOCS); continue; } int id1 = termDocs1.docID(); - assertEquals(DocsEnum.NO_MORE_DOCS, termDocs1.nextDoc()); + assertEquals(DocIdSetIterator.NO_MORE_DOCS, termDocs1.nextDoc()); - assertTrue(termDocs2.nextDoc() != DocsEnum.NO_MORE_DOCS); + assertTrue(termDocs2.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); int id2 = termDocs2.docID(); - assertEquals(DocsEnum.NO_MORE_DOCS, termDocs2.nextDoc()); + assertEquals(DocIdSetIterator.NO_MORE_DOCS, termDocs2.nextDoc()); r2r1[id2] = id1; @@ -409,7 +410,7 @@ System.out.println(" " + term2.utf8ToString() + ": freq=" + termsEnum3.totalTermFreq()); dpEnum = termsEnum3.docsAndPositions(null, dpEnum, false); if (dpEnum != null) { - assertTrue(dpEnum.nextDoc() != DocsEnum.NO_MORE_DOCS); + assertTrue(dpEnum.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); final int freq = dpEnum.freq(); System.out.println(" doc=" + dpEnum.docID() + " freq=" + freq); for(int posUpto=0;posUpto setNextReader(AtomicReaderContext context) throws IOException { throw new UnsupportedOperationException(UNSUPPORTED_MSG); } @@ -144,7 +144,7 @@ static final class JustCompileFieldComparatorSource extends FieldComparatorSource { @Override - public FieldComparator newComparator(String fieldname, int numHits, + public FieldComparator newComparator(String fieldname, int numHits, int sortPos, boolean reversed) throws IOException { throw new UnsupportedOperationException(UNSUPPORTED_MSG); } Index: lucene/core/src/test/org/apache/lucene/search/spans/MultiSpansWrapper.java =================================================================== --- lucene/core/src/test/org/apache/lucene/search/spans/MultiSpansWrapper.java (revision 1297029) +++ lucene/core/src/test/org/apache/lucene/search/spans/MultiSpansWrapper.java (working copy) @@ -28,6 +28,7 @@ import org.apache.lucene.index.DocsEnum; import org.apache.lucene.index.IndexReaderContext; import org.apache.lucene.index.Term; +import org.apache.lucene.search.DocIdSetIterator; import org.apache.lucene.util.ReaderUtil; import org.apache.lucene.util.TermContext; @@ -121,7 +122,7 @@ @Override public int doc() { if (current == null) { - return DocsEnum.NO_MORE_DOCS; + return DocIdSetIterator.NO_MORE_DOCS; } return current.doc() + leaves[leafOrd].docBase; } @@ -129,7 +130,7 @@ @Override public int start() { if (current == null) { - return DocsEnum.NO_MORE_DOCS; + return DocIdSetIterator.NO_MORE_DOCS; } return current.start(); } @@ -137,7 +138,7 @@ @Override public int end() { if (current == null) { - return DocsEnum.NO_MORE_DOCS; + return DocIdSetIterator.NO_MORE_DOCS; } return current.end(); } Index: lucene/core/src/test/org/apache/lucene/search/TestCustomSearcherSort.java =================================================================== --- lucene/core/src/test/org/apache/lucene/search/TestCustomSearcherSort.java (revision 1297029) +++ lucene/core/src/test/org/apache/lucene/search/TestCustomSearcherSort.java (working copy) @@ -51,7 +51,7 @@ INDEX_SIZE = atLeast(2000); index = newDirectory(); RandomIndexWriter writer = new RandomIndexWriter(random, index); - RandomGen random = new RandomGen(this.random); + RandomGen random = new RandomGen(LuceneTestCase.random); for (int i = 0; i < INDEX_SIZE; ++i) { // don't decrease; if to low the // problem doesn't show up Document doc = new Document(); Index: lucene/core/src/test/org/apache/lucene/search/TestElevationComparator.java =================================================================== --- lucene/core/src/test/org/apache/lucene/search/TestElevationComparator.java (revision 1297029) +++ lucene/core/src/test/org/apache/lucene/search/TestElevationComparator.java (working copy) @@ -139,7 +139,7 @@ } @Override - public FieldComparator newComparator(final String fieldname, final int numHits, int sortPos, boolean reversed) throws IOException { + public FieldComparator newComparator(final String fieldname, final int numHits, int sortPos, boolean reversed) throws IOException { return new FieldComparator() { FieldCache.DocTermsIndex idIndex; @@ -179,7 +179,7 @@ } @Override - public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { + public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { idIndex = FieldCache.DEFAULT.getTermsIndex(context.reader(), fieldname); return this; } Index: lucene/core/src/test/org/apache/lucene/search/TestPositionIncrement.java =================================================================== --- lucene/core/src/test/org/apache/lucene/search/TestPositionIncrement.java (revision 1297029) +++ lucene/core/src/test/org/apache/lucene/search/TestPositionIncrement.java (working copy) @@ -218,7 +218,7 @@ false); int count = 0; - assertTrue(tp.nextDoc() != DocsAndPositionsEnum.NO_MORE_DOCS); + assertTrue(tp.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); // "a" occurs 4 times assertEquals(4, tp.freq()); int expected = 0; @@ -228,7 +228,7 @@ assertEquals(6, tp.nextPosition()); // only one doc has "a" - assertEquals(DocsAndPositionsEnum.NO_MORE_DOCS, tp.nextDoc()); + assertEquals(DocIdSetIterator.NO_MORE_DOCS, tp.nextDoc()); IndexSearcher is = newSearcher(readerFromWriter); Index: lucene/core/src/test/org/apache/lucene/search/TestSort.java =================================================================== --- lucene/core/src/test/org/apache/lucene/search/TestSort.java (revision 1297029) +++ lucene/core/src/test/org/apache/lucene/search/TestSort.java (working copy) @@ -699,7 +699,7 @@ }; @Override - public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { + public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { docValues = FieldCache.DEFAULT.getInts(context.reader(), "parser", testIntParser, false); return this; } @@ -712,7 +712,7 @@ static class MyFieldComparatorSource extends FieldComparatorSource { @Override - public FieldComparator newComparator(String fieldname, int numHits, int sortPos, boolean reversed) { + public FieldComparator newComparator(String fieldname, int numHits, int sortPos, boolean reversed) { return new MyFieldComparator(numHits); } } Index: lucene/core/src/test/org/apache/lucene/search/TestSubScorerFreqs.java =================================================================== --- lucene/core/src/test/org/apache/lucene/search/TestSubScorerFreqs.java (revision 1297029) +++ lucene/core/src/test/org/apache/lucene/search/TestSubScorerFreqs.java (working copy) @@ -161,7 +161,7 @@ query.add(inner, Occur.MUST); query.add(aQuery, Occur.MUST); query.add(dQuery, Occur.MUST); - Set[] occurList = new Set[] { + @SuppressWarnings({"rawtypes","unchecked"}) Set[] occurList = new Set[] { Collections.singleton(Occur.MUST.toString()), new HashSet(Arrays.asList(Occur.MUST.toString(), Occur.SHOULD.toString())) }; Index: lucene/core/src/test/org/apache/lucene/search/TestTermVectors.java =================================================================== --- lucene/core/src/test/org/apache/lucene/search/TestTermVectors.java (revision 1297029) +++ lucene/core/src/test/org/apache/lucene/search/TestTermVectors.java (working copy) @@ -135,19 +135,19 @@ TermsEnum termsEnum = terms.iterator(null); assertEquals("content", termsEnum.next().utf8ToString()); dpEnum = termsEnum.docsAndPositions(null, dpEnum, false); - assertTrue(dpEnum.nextDoc() != DocsEnum.NO_MORE_DOCS); + assertTrue(dpEnum.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); assertEquals(1, dpEnum.freq()); assertEquals(expectedPositions[0], dpEnum.nextPosition()); assertEquals("here", termsEnum.next().utf8ToString()); dpEnum = termsEnum.docsAndPositions(null, dpEnum, false); - assertTrue(dpEnum.nextDoc() != DocsEnum.NO_MORE_DOCS); + assertTrue(dpEnum.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); assertEquals(1, dpEnum.freq()); assertEquals(expectedPositions[1], dpEnum.nextPosition()); assertEquals("some", termsEnum.next().utf8ToString()); dpEnum = termsEnum.docsAndPositions(null, dpEnum, false); - assertTrue(dpEnum.nextDoc() != DocsEnum.NO_MORE_DOCS); + assertTrue(dpEnum.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); assertEquals(1, dpEnum.freq()); assertEquals(expectedPositions[2], dpEnum.nextPosition()); @@ -178,7 +178,7 @@ while(true) { dpEnum = termsEnum.docsAndPositions(null, dpEnum, shouldBeOffVector); assertNotNull(dpEnum); - assertTrue(dpEnum.nextDoc() != DocsEnum.NO_MORE_DOCS); + assertTrue(dpEnum.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); dpEnum.nextPosition(); @@ -263,7 +263,7 @@ String text = termsEnum.term().utf8ToString(); docs = _TestUtil.docs(random, termsEnum, MultiFields.getLiveDocs(knownSearcher.reader), docs, true); - while (docs.nextDoc() != DocsEnum.NO_MORE_DOCS) { + while (docs.nextDoc() != DocIdSetIterator.NO_MORE_DOCS) { int docId = docs.docID(); int freq = docs.freq(); //System.out.println("Doc Id: " + docId + " freq " + freq); @@ -428,7 +428,7 @@ assertEquals(5, termsEnum.totalTermFreq()); DocsAndPositionsEnum dpEnum = termsEnum.docsAndPositions(null, null, false); assertNotNull(dpEnum); - assertTrue(dpEnum.nextDoc() != DocsEnum.NO_MORE_DOCS); + assertTrue(dpEnum.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); assertEquals(5, dpEnum.freq()); for(int i=0;i<5;i++) { assertEquals(i, dpEnum.nextPosition()); @@ -436,7 +436,7 @@ dpEnum = termsEnum.docsAndPositions(null, dpEnum, true); assertNotNull(dpEnum); - assertTrue(dpEnum.nextDoc() != DocsEnum.NO_MORE_DOCS); + assertTrue(dpEnum.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); assertEquals(5, dpEnum.freq()); for(int i=0;i<5;i++) { dpEnum.nextPosition(); Index: lucene/core/src/test/org/apache/lucene/util/fst/TestFSTs.java =================================================================== --- lucene/core/src/test/org/apache/lucene/util/fst/TestFSTs.java (revision 1297029) +++ lucene/core/src/test/org/apache/lucene/util/fst/TestFSTs.java (working copy) @@ -57,6 +57,7 @@ import org.apache.lucene.util.LuceneTestCase.UseNoMemoryExpensiveCodec; import org.apache.lucene.util.UnicodeUtil; import org.apache.lucene.util._TestUtil; +import org.apache.lucene.util.fst.BytesRefFSTEnum.InputOutput; import org.apache.lucene.util.fst.FST.Arc; import org.apache.lucene.util.fst.FST.BytesReader; import org.apache.lucene.util.fst.PairOutputs.Pair; @@ -494,7 +495,7 @@ if (random.nextBoolean() && fst != null && !willRewrite) { TestFSTs t = new TestFSTs(); - IOContext context = t.newIOContext(random); + IOContext context = LuceneTestCase.newIOContext(random); IndexOutput out = dir.createOutput("fst.bin", context); fst.save(out); out.close(); @@ -984,7 +985,7 @@ if (VERBOSE) { System.out.println(" fstEnum.next prefix=" + inputToString(inputMode, current.input, false) + " output=" + outputs.outputToString(current.output)); } - final CountMinOutput cmo = prefixes.get(current.input); + final CountMinOutput cmo = prefixes.get(current.input); assertNotNull(cmo); assertTrue(cmo.isLeaf || cmo.isFinal); //if (cmo.isFinal && !cmo.isLeaf) { @@ -1183,7 +1184,7 @@ } final TermsEnum.SeekStatus seekResult = termsEnum.seekCeil(randomTerm); - final BytesRefFSTEnum.InputOutput fstSeekResult = fstEnum.seekCeil(randomTerm); + final InputOutput fstSeekResult = fstEnum.seekCeil(randomTerm); if (seekResult == TermsEnum.SeekStatus.END) { assertNull("got " + (fstSeekResult == null ? "null" : fstSeekResult.input.utf8ToString()) + " but expected null", fstSeekResult); @@ -1224,7 +1225,7 @@ dir.close(); } - private void assertSame(TermsEnum termsEnum, BytesRefFSTEnum fstEnum, boolean storeOrd) throws Exception { + private void assertSame(TermsEnum termsEnum, BytesRefFSTEnum fstEnum, boolean storeOrd) throws Exception { if (termsEnum.term() == null) { assertNull(fstEnum.current()); } else { @@ -1829,7 +1830,7 @@ public int verifyStateAndBelow(FST fst, Arc arc, int depth) throws IOException { - if (fst.targetHasArcs(arc)) { + if (FST.targetHasArcs(arc)) { int childCount = 0; for (arc = fst.readFirstTargetArc(arc, arc);; arc = fst.readNextArc(arc), childCount++) Index: lucene/core/src/test/org/apache/lucene/util/TestAttributeSource.java =================================================================== --- lucene/core/src/test/org/apache/lucene/util/TestAttributeSource.java (revision 1297029) +++ lucene/core/src/test/org/apache/lucene/util/TestAttributeSource.java (working copy) @@ -126,7 +126,7 @@ src.addAttribute(TypeAttribute.class) instanceof TypeAttributeImpl); } - @SuppressWarnings("unchecked") + @SuppressWarnings({"rawtypes","unchecked"}) public void testInvalidArguments() throws Exception { try { AttributeSource src = new AttributeSource(); Index: lucene/core/src/test/org/apache/lucene/util/TestVirtualMethod.java =================================================================== --- lucene/core/src/test/org/apache/lucene/util/TestVirtualMethod.java (revision 1297029) +++ lucene/core/src/test/org/apache/lucene/util/TestVirtualMethod.java (working copy) @@ -69,7 +69,7 @@ assertEquals(0, VirtualMethod.compareImplementationDistance(TestClass5.class, publicTestMethod, protectedTestMethod)); } - @SuppressWarnings("unchecked") + @SuppressWarnings({"rawtypes","unchecked"}) public void testExceptions() { try { // cast to Class to remove generics: Index: modules/analysis/common/src/test/org/apache/lucene/analysis/core/TestClassicAnalyzer.java =================================================================== --- modules/analysis/common/src/test/org/apache/lucene/analysis/core/TestClassicAnalyzer.java (revision 1297029) +++ modules/analysis/common/src/test/org/apache/lucene/analysis/core/TestClassicAnalyzer.java (working copy) @@ -12,6 +12,7 @@ import org.apache.lucene.index.IndexWriterConfig; import org.apache.lucene.index.MultiFields; import org.apache.lucene.index.Term; +import org.apache.lucene.search.DocIdSetIterator; import org.apache.lucene.store.RAMDirectory; import org.apache.lucene.util.BytesRef; import org.apache.lucene.util.Version; @@ -285,7 +286,7 @@ "content", new BytesRef("another"), false); - assertTrue(tps.nextDoc() != DocsEnum.NO_MORE_DOCS); + assertTrue(tps.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); assertEquals(1, tps.freq()); assertEquals(3, tps.nextPosition()); Index: modules/analysis/common/src/test/org/apache/lucene/analysis/core/TestKeywordAnalyzer.java =================================================================== --- modules/analysis/common/src/test/org/apache/lucene/analysis/core/TestKeywordAnalyzer.java (revision 1297029) +++ modules/analysis/common/src/test/org/apache/lucene/analysis/core/TestKeywordAnalyzer.java (working copy) @@ -31,6 +31,7 @@ import org.apache.lucene.index.IndexWriter; import org.apache.lucene.index.IndexWriterConfig; import org.apache.lucene.index.MultiFields; +import org.apache.lucene.search.DocIdSetIterator; import org.apache.lucene.search.IndexSearcher; import org.apache.lucene.store.Directory; import org.apache.lucene.store.RAMDirectory; @@ -102,7 +103,7 @@ MultiFields.getLiveDocs(reader), null, false); - assertTrue(td.nextDoc() != DocsEnum.NO_MORE_DOCS); + assertTrue(td.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); td = _TestUtil.docs(random, reader, "partnum", @@ -110,7 +111,7 @@ MultiFields.getLiveDocs(reader), null, false); - assertTrue(td.nextDoc() != DocsEnum.NO_MORE_DOCS); + assertTrue(td.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); } // LUCENE-1441 Index: modules/analysis/common/src/test/org/apache/lucene/analysis/shingle/ShingleFilterTest.java =================================================================== --- modules/analysis/common/src/test/org/apache/lucene/analysis/shingle/ShingleFilterTest.java (revision 1297029) +++ modules/analysis/common/src/test/org/apache/lucene/analysis/shingle/ShingleFilterTest.java (working copy) @@ -59,7 +59,7 @@ termAtt.copyBuffer(t.buffer(), 0, t.length()); offsetAtt.setOffset(t.startOffset(), t.endOffset()); posIncrAtt.setPositionIncrement(t.getPositionIncrement()); - typeAtt.setType(TypeAttributeImpl.DEFAULT_TYPE); + typeAtt.setType(TypeAttribute.DEFAULT_TYPE); return true; } else { return false; @@ -1018,14 +1018,14 @@ assertTokenStreamContents(filter, new String[]{"please","please divide","divide","divide this","this","this sentence","sentence"}, new int[]{0,0,7,7,14,14,19}, new int[]{6,13,13,18,18,27,27}, - new String[]{TypeAttributeImpl.DEFAULT_TYPE,"shingle",TypeAttributeImpl.DEFAULT_TYPE,"shingle",TypeAttributeImpl.DEFAULT_TYPE,"shingle",TypeAttributeImpl.DEFAULT_TYPE}, + new String[]{TypeAttribute.DEFAULT_TYPE,"shingle",TypeAttribute.DEFAULT_TYPE,"shingle",TypeAttribute.DEFAULT_TYPE,"shingle",TypeAttribute.DEFAULT_TYPE}, new int[]{1,0,1,0,1,0,1} ); wsTokenizer.reset(new StringReader("please divide this sentence")); assertTokenStreamContents(filter, new String[]{"please","please divide","divide","divide this","this","this sentence","sentence"}, new int[]{0,0,7,7,14,14,19}, new int[]{6,13,13,18,18,27,27}, - new String[]{TypeAttributeImpl.DEFAULT_TYPE,"shingle",TypeAttributeImpl.DEFAULT_TYPE,"shingle",TypeAttributeImpl.DEFAULT_TYPE,"shingle",TypeAttributeImpl.DEFAULT_TYPE}, + new String[]{TypeAttribute.DEFAULT_TYPE,"shingle",TypeAttribute.DEFAULT_TYPE,"shingle",TypeAttribute.DEFAULT_TYPE,"shingle",TypeAttribute.DEFAULT_TYPE}, new int[]{1,0,1,0,1,0,1} ); } Index: modules/analysis/common/src/test/org/apache/lucene/analysis/sinks/TestTeeSinkTokenFilter.java =================================================================== --- modules/analysis/common/src/test/org/apache/lucene/analysis/sinks/TestTeeSinkTokenFilter.java (revision 1297029) +++ modules/analysis/common/src/test/org/apache/lucene/analysis/sinks/TestTeeSinkTokenFilter.java (working copy) @@ -35,6 +35,7 @@ import org.apache.lucene.index.IndexWriter; import org.apache.lucene.index.Terms; import org.apache.lucene.index.TermsEnum; +import org.apache.lucene.search.DocIdSetIterator; import org.apache.lucene.store.Directory; import org.apache.lucene.util.AttributeSource; import org.apache.lucene.util.English; @@ -110,7 +111,7 @@ termsEnum.next(); assertEquals(2, termsEnum.totalTermFreq()); DocsAndPositionsEnum positions = termsEnum.docsAndPositions(null, null, true); - assertTrue(positions.nextDoc() != DocsEnum.NO_MORE_DOCS); + assertTrue(positions.nextDoc() != DocIdSetIterator.NO_MORE_DOCS); assertEquals(2, positions.freq()); positions.nextPosition(); assertEquals(0, positions.startOffset()); @@ -118,7 +119,7 @@ positions.nextPosition(); assertEquals(8, positions.startOffset()); assertEquals(12, positions.endOffset()); - assertEquals(DocsEnum.NO_MORE_DOCS, positions.nextDoc()); + assertEquals(DocIdSetIterator.NO_MORE_DOCS, positions.nextDoc()); r.close(); dir.close(); } Index: modules/analysis/icu/src/java/org/apache/lucene/analysis/icu/segmentation/ScriptIterator.java =================================================================== --- modules/analysis/icu/src/java/org/apache/lucene/analysis/icu/segmentation/ScriptIterator.java (revision 1297029) +++ modules/analysis/icu/src/java/org/apache/lucene/analysis/icu/segmentation/ScriptIterator.java (working copy) @@ -29,6 +29,7 @@ */ import com.ibm.icu.lang.UCharacter; +import com.ibm.icu.lang.UCharacterEnums.ECharacterCategory; import com.ibm.icu.lang.UScript; import com.ibm.icu.text.UTF16; @@ -110,7 +111,7 @@ * value — should inherit the script value of its base character. */ if (isSameScript(scriptCode, sc) - || UCharacter.getType(ch) == UCharacter.NON_SPACING_MARK) { + || UCharacter.getType(ch) == ECharacterCategory.NON_SPACING_MARK) { index += UTF16.getCharCount(ch); /* Index: modules/analysis/smartcn/src/java/org/apache/lucene/analysis/cn/smart/hhmm/AbstractDictionary.java =================================================================== --- modules/analysis/smartcn/src/java/org/apache/lucene/analysis/cn/smart/hhmm/AbstractDictionary.java (revision 1297029) +++ modules/analysis/smartcn/src/java/org/apache/lucene/analysis/cn/smart/hhmm/AbstractDictionary.java (working copy) @@ -85,7 +85,7 @@ * @return unicode String */ public String getCCByGB2312Id(int ccid) { - if (ccid < 0 || ccid > WordDictionary.GB2312_CHAR_NUM) + if (ccid < 0 || ccid > AbstractDictionary.GB2312_CHAR_NUM) return ""; int cc1 = ccid / 94 + 161; int cc2 = ccid % 94 + 161; Index: modules/benchmark/src/java/org/apache/lucene/benchmark/quality/utils/SimpleQQParser.java =================================================================== --- modules/benchmark/src/java/org/apache/lucene/benchmark/quality/utils/SimpleQQParser.java (revision 1297029) +++ modules/benchmark/src/java/org/apache/lucene/benchmark/quality/utils/SimpleQQParser.java (working copy) @@ -21,6 +21,7 @@ import org.apache.lucene.benchmark.quality.QualityQueryParser; import org.apache.lucene.queryparser.classic.ParseException; import org.apache.lucene.queryparser.classic.QueryParser; +import org.apache.lucene.queryparser.classic.QueryParserBase; import org.apache.lucene.search.BooleanClause; import org.apache.lucene.search.BooleanQuery; import org.apache.lucene.search.Query; @@ -66,7 +67,7 @@ } BooleanQuery bq = new BooleanQuery(); for (int i = 0; i < qqNames.length; i++) - bq.add(qp.parse(QueryParser.escape(qq.getValue(qqNames[i]))), BooleanClause.Occur.SHOULD); + bq.add(qp.parse(QueryParserBase.escape(qq.getValue(qqNames[i]))), BooleanClause.Occur.SHOULD); return bq; } Index: modules/benchmark/src/test/org/apache/lucene/benchmark/byTask/TestPerfTasksLogic.java =================================================================== --- modules/benchmark/src/test/org/apache/lucene/benchmark/byTask/TestPerfTasksLogic.java (revision 1297029) +++ modules/benchmark/src/test/org/apache/lucene/benchmark/byTask/TestPerfTasksLogic.java (working copy) @@ -55,6 +55,7 @@ import org.apache.lucene.index.SlowCompositeReaderWrapper; import org.apache.lucene.index.Terms; import org.apache.lucene.index.TermsEnum; +import org.apache.lucene.search.DocIdSetIterator; import org.apache.lucene.search.FieldCache.DocTermsIndex; import org.apache.lucene.search.FieldCache; import org.apache.lucene.store.Directory; @@ -497,7 +498,7 @@ DocsEnum docs = null; while(termsEnum.next() != null) { docs = _TestUtil.docs(random, termsEnum, MultiFields.getLiveDocs(reader), docs, true); - while(docs.nextDoc() != docs.NO_MORE_DOCS) { + while(docs.nextDoc() != DocIdSetIterator.NO_MORE_DOCS) { totalTokenCount2 += docs.freq(); } } Index: modules/facet/src/java/org/apache/lucene/facet/taxonomy/directory/ParentArray.java =================================================================== --- modules/facet/src/java/org/apache/lucene/facet/taxonomy/directory/ParentArray.java (revision 1297029) +++ modules/facet/src/java/org/apache/lucene/facet/taxonomy/directory/ParentArray.java (working copy) @@ -6,6 +6,7 @@ import org.apache.lucene.index.DocsAndPositionsEnum; import org.apache.lucene.index.IndexReader; import org.apache.lucene.index.MultiFields; +import org.apache.lucene.search.DocIdSetIterator; import org.apache.lucene.util.Bits; import org.apache.lucene.util.BytesRef; @@ -106,7 +107,7 @@ DocsAndPositionsEnum positions = MultiFields.getTermPositionsEnum(indexReader, liveDocs, Consts.FIELD_PAYLOADS, new BytesRef(Consts.PAYLOAD_PARENT), false); - if ((positions == null || positions.advance(first) == DocsAndPositionsEnum.NO_MORE_DOCS) && first < num) { + if ((positions == null || positions.advance(first) == DocIdSetIterator.NO_MORE_DOCS) && first < num) { throw new CorruptIndexException("Missing parent data for category " + first); } for (int i=first; i 0) { name = unknownFieldPrefix + name; sf = schema.getFieldOrNull(name); - } else if (sf == null && defaultField.length() > 0 && name.equals(Metadata.RESOURCE_NAME_KEY) == false /*let the fall through below handle this*/){ + } else if (sf == null && defaultField.length() > 0 && name.equals(TikaMetadataKeys.RESOURCE_NAME_KEY) == false /*let the fall through below handle this*/){ name = defaultField; sf = schema.getFieldOrNull(name); } @@ -201,7 +202,7 @@ // ExtractingDocumentLoader.load(). You shouldn't have to define a mapping for this // field just because you specified a resource.name parameter to the handler, should // you? - if (sf == null && unknownFieldPrefix.length()==0 && name == Metadata.RESOURCE_NAME_KEY) { + if (sf == null && unknownFieldPrefix.length()==0 && name == TikaMetadataKeys.RESOURCE_NAME_KEY) { return; } Index: solr/contrib/velocity/src/java/org/apache/solr/response/VelocityResponseWriter.java =================================================================== --- solr/contrib/velocity/src/java/org/apache/solr/response/VelocityResponseWriter.java (revision 1297029) +++ solr/contrib/velocity/src/java/org/apache/solr/response/VelocityResponseWriter.java (working copy) @@ -25,6 +25,7 @@ import org.apache.velocity.Template; import org.apache.velocity.VelocityContext; import org.apache.velocity.app.VelocityEngine; +import org.apache.velocity.runtime.RuntimeConstants; import org.apache.velocity.tools.generic.*; import java.io.*; @@ -117,14 +118,14 @@ if (template_root != null) { baseDir = new File(template_root); } - engine.setProperty(VelocityEngine.FILE_RESOURCE_LOADER_PATH, baseDir.getAbsolutePath()); + engine.setProperty(RuntimeConstants.FILE_RESOURCE_LOADER_PATH, baseDir.getAbsolutePath()); engine.setProperty("params.resource.loader.instance", new SolrParamResourceLoader(request)); SolrVelocityResourceLoader resourceLoader = new SolrVelocityResourceLoader(request.getCore().getSolrConfig().getResourceLoader()); engine.setProperty("solr.resource.loader.instance", resourceLoader); // TODO: Externalize Velocity properties - engine.setProperty(VelocityEngine.RESOURCE_LOADER, "params,file,solr"); + engine.setProperty(RuntimeConstants.RESOURCE_LOADER, "params,file,solr"); String propFile = request.getParams().get("v.properties"); try { if (propFile == null) Index: solr/core/src/java/org/apache/solr/handler/admin/LukeRequestHandler.java =================================================================== --- solr/core/src/java/org/apache/solr/handler/admin/LukeRequestHandler.java (revision 1297029) +++ solr/core/src/java/org/apache/solr/handler/admin/LukeRequestHandler.java (working copy) @@ -27,6 +27,7 @@ import org.apache.lucene.document.Field; import org.apache.lucene.index.*; import org.apache.lucene.index.FieldInfo.IndexOptions; +import org.apache.lucene.search.DocIdSetIterator; import org.apache.lucene.store.Directory; import org.apache.lucene.util.BytesRef; import org.apache.lucene.util.CharsRef; @@ -399,7 +400,7 @@ false); if (docsEnum != null) { int docId; - if ((docId = docsEnum.nextDoc()) != DocsEnum.NO_MORE_DOCS) { + if ((docId = docsEnum.nextDoc()) != DocIdSetIterator.NO_MORE_DOCS) { return reader.document(docId); } } Index: solr/core/src/java/org/apache/solr/handler/ReplicationHandler.java =================================================================== --- solr/core/src/java/org/apache/solr/handler/ReplicationHandler.java (revision 1297029) +++ solr/core/src/java/org/apache/solr/handler/ReplicationHandler.java (working copy) @@ -150,7 +150,7 @@ // the CMD_GET_FILE_LIST command. // core.getDeletionPolicy().setReserveDuration(commitPoint.getGeneration(), reserveCommitDuration); - rsp.add(CMD_INDEX_VERSION, core.getDeletionPolicy().getCommitTimestamp(commitPoint)); + rsp.add(CMD_INDEX_VERSION, IndexDeletionPolicyWrapper.getCommitTimestamp(commitPoint)); rsp.add(GENERATION, commitPoint.getGeneration()); } else { // This happens when replication is not configured to happen after startup and no commit/optimize @@ -229,7 +229,7 @@ for (IndexCommit c : commits.values()) { try { NamedList nl = new NamedList(); - nl.add("indexVersion", core.getDeletionPolicy().getCommitTimestamp(c)); + nl.add("indexVersion", IndexDeletionPolicyWrapper.getCommitTimestamp(c)); nl.add(GENERATION, c.getGeneration()); nl.add(CMD_GET_FILE_LIST, c.getFileNames()); l.add(nl); Index: solr/core/src/java/org/apache/solr/request/SimpleFacets.java =================================================================== --- solr/core/src/java/org/apache/solr/request/SimpleFacets.java (revision 1297029) +++ solr/core/src/java/org/apache/solr/request/SimpleFacets.java (working copy) @@ -821,7 +821,7 @@ } final String gap = required.getFieldParam(f,FacetParams.FACET_DATE_GAP); - final DateMathParser dmp = new DateMathParser(ft.UTC, Locale.US); + final DateMathParser dmp = new DateMathParser(DateField.UTC, Locale.US); final int minCount = params.getFieldInt(f,FacetParams.FACET_MINCOUNT, 0); Index: solr/core/src/java/org/apache/solr/schema/LatLonType.java =================================================================== --- solr/core/src/java/org/apache/solr/schema/LatLonType.java (revision 1297029) +++ solr/core/src/java/org/apache/solr/schema/LatLonType.java (working copy) @@ -349,8 +349,8 @@ public SpatialWeight(IndexSearcher searcher) throws IOException { this.searcher = searcher; - this.latContext = latSource.newContext(searcher); - this.lonContext = lonSource.newContext(searcher); + this.latContext = ValueSource.newContext(searcher); + this.lonContext = ValueSource.newContext(searcher); latSource.createWeight(latContext, searcher); lonSource.createWeight(lonContext, searcher); } Index: solr/core/src/java/org/apache/solr/search/ExtendedDismaxQParserPlugin.java =================================================================== --- solr/core/src/java/org/apache/solr/search/ExtendedDismaxQParserPlugin.java (revision 1297029) +++ solr/core/src/java/org/apache/solr/search/ExtendedDismaxQParserPlugin.java (working copy) @@ -103,25 +103,25 @@ final String minShouldMatch = DisMaxQParser.parseMinShouldMatch(req.getSchema(), solrParams); - queryFields = U.parseFieldBoosts(solrParams.getParams(DMP.QF)); + queryFields = SolrPluginUtils.parseFieldBoosts(solrParams.getParams(DisMaxParams.QF)); if (0 == queryFields.size()) { queryFields.put(req.getSchema().getDefaultSearchFieldName(), 1.0f); } // Boosted phrase of the full query string Map phraseFields = - U.parseFieldBoosts(solrParams.getParams(DMP.PF)); + SolrPluginUtils.parseFieldBoosts(solrParams.getParams(DisMaxParams.PF)); // Boosted Bi-Term Shingles from the query string Map phraseFields2 = - U.parseFieldBoosts(solrParams.getParams("pf2")); + SolrPluginUtils.parseFieldBoosts(solrParams.getParams("pf2")); // Boosted Tri-Term Shingles from the query string Map phraseFields3 = - U.parseFieldBoosts(solrParams.getParams("pf3")); + SolrPluginUtils.parseFieldBoosts(solrParams.getParams("pf3")); - float tiebreaker = solrParams.getFloat(DMP.TIE, 0.0f); + float tiebreaker = solrParams.getFloat(DisMaxParams.TIE, 0.0f); - int pslop = solrParams.getInt(DMP.PS, 0); - int qslop = solrParams.getInt(DMP.QS, 0); + int pslop = solrParams.getInt(DisMaxParams.PS, 0); + int qslop = solrParams.getInt(DisMaxParams.QS, 0); // remove stopwords from mandatory "matching" component? boolean stopwords = solrParams.getBool("stopwords", true); @@ -137,7 +137,7 @@ altUserQuery = null; if( userQuery == null || userQuery.length() < 1 ) { // If no query is specified, we may have an alternate - String altQ = solrParams.get( DMP.ALTQ ); + String altQ = solrParams.get( DisMaxParams.ALTQ ); if (altQ != null) { altQParser = subQuery(altQ, null); altUserQuery = altQParser.getQuery(); @@ -248,7 +248,7 @@ if (parsedUserQuery != null && doMinMatched) { if (parsedUserQuery instanceof BooleanQuery) { - U.setMinShouldMatch((BooleanQuery)parsedUserQuery, minShouldMatch); + SolrPluginUtils.setMinShouldMatch((BooleanQuery)parsedUserQuery, minShouldMatch); } } @@ -285,8 +285,8 @@ if (parsedUserQuery instanceof BooleanQuery) { BooleanQuery t = new BooleanQuery(); - U.flattenBooleanQuery(t, (BooleanQuery)parsedUserQuery); - U.setMinShouldMatch(t, minShouldMatch); + SolrPluginUtils.flattenBooleanQuery(t, (BooleanQuery)parsedUserQuery); + SolrPluginUtils.setMinShouldMatch(t, minShouldMatch); parsedUserQuery = t; } } @@ -326,7 +326,7 @@ /* * * Boosting Query * * */ - boostParams = solrParams.getParams(DMP.BQ); + boostParams = solrParams.getParams(DisMaxParams.BQ); //List boostQueries = U.parseQueryStrings(req, boostParams); boostQueries=null; if (boostParams!=null && boostParams.length>0) { @@ -345,7 +345,7 @@ /* * * Boosting Functions * * */ - String[] boostFuncs = solrParams.getParams(DMP.BF); + String[] boostFuncs = solrParams.getParams(DisMaxParams.BF); if (null != boostFuncs && 0 != boostFuncs.length) { for (String boostFunc : boostFuncs) { if(null == boostFunc || "".equals(boostFunc)) continue; Index: solr/core/src/java/org/apache/solr/search/function/FileFloatSource.java =================================================================== --- solr/core/src/java/org/apache/solr/search/function/FileFloatSource.java (revision 1297029) +++ solr/core/src/java/org/apache/solr/search/function/FileFloatSource.java (working copy) @@ -36,6 +36,7 @@ import org.apache.lucene.queries.function.FunctionValues; import org.apache.lucene.queries.function.ValueSource; import org.apache.lucene.queries.function.docvalues.FloatDocValues; +import org.apache.lucene.search.DocIdSetIterator; import org.apache.lucene.util.BytesRef; import org.apache.lucene.util.ReaderUtil; import org.apache.solr.core.SolrCore; @@ -278,7 +279,7 @@ docsEnum = termsEnum.docs(null, docsEnum, false); int doc; - while ((doc = docsEnum.nextDoc()) != DocsEnum.NO_MORE_DOCS) { + while ((doc = docsEnum.nextDoc()) != DocIdSetIterator.NO_MORE_DOCS) { vals[doc] = fval; } } Index: solr/core/src/test/org/apache/solr/analysis/TestWordDelimiterFilterFactory.java =================================================================== --- solr/core/src/test/org/apache/solr/analysis/TestWordDelimiterFilterFactory.java (revision 1297029) +++ solr/core/src/test/org/apache/solr/analysis/TestWordDelimiterFilterFactory.java (working copy) @@ -21,6 +21,7 @@ import java.util.HashMap; import java.util.Map; +import org.apache.lucene.analysis.BaseTokenStreamTestCase; import org.apache.lucene.analysis.MockTokenizer; import org.apache.lucene.analysis.TokenStream; import org.apache.solr.SolrTestCaseJ4; @@ -211,12 +212,12 @@ TokenStream ts = factoryDefault.create( new MockTokenizer(new StringReader(testText), MockTokenizer.WHITESPACE, false)); - BaseTokenTestCase.assertTokenStreamContents(ts, + BaseTokenStreamTestCase.assertTokenStreamContents(ts, new String[] { "I", "borrowed", "5", "400", "00", "540000", "at", "25", "interest", "rate", "interestrate" }); ts = factoryDefault.create( new MockTokenizer(new StringReader("foo\u200Dbar"), MockTokenizer.WHITESPACE, false)); - BaseTokenTestCase.assertTokenStreamContents(ts, + BaseTokenStreamTestCase.assertTokenStreamContents(ts, new String[] { "foo", "bar", "foobar" }); @@ -229,13 +230,13 @@ ts = factoryCustom.create( new MockTokenizer(new StringReader(testText), MockTokenizer.WHITESPACE, false)); - BaseTokenTestCase.assertTokenStreamContents(ts, + BaseTokenStreamTestCase.assertTokenStreamContents(ts, new String[] { "I", "borrowed", "$5,400.00", "at", "25%", "interest", "rate", "interestrate" }); /* test custom behavior with a char > 0x7F, because we had to make a larger byte[] */ ts = factoryCustom.create( new MockTokenizer(new StringReader("foo\u200Dbar"), MockTokenizer.WHITESPACE, false)); - BaseTokenTestCase.assertTokenStreamContents(ts, + BaseTokenStreamTestCase.assertTokenStreamContents(ts, new String[] { "foo\u200Dbar" }); } } Index: solr/core/src/test/org/apache/solr/cloud/ZkSolrClientTest.java =================================================================== --- solr/core/src/test/org/apache/solr/cloud/ZkSolrClientTest.java (revision 1297029) +++ solr/core/src/test/org/apache/solr/cloud/ZkSolrClientTest.java (working copy) @@ -20,6 +20,7 @@ import java.io.File; import java.util.concurrent.atomic.AtomicInteger; +import junit.framework.Assert; import junit.framework.TestCase; import org.apache.solr.common.cloud.SolrZkClient; @@ -91,7 +92,7 @@ try { zkClient.makePath("collections/collection2", false); - TestCase.fail("Server should be down here"); + Assert.fail("Server should be down here"); } catch (KeeperException.ConnectionLossException e) { } Index: solr/core/src/test/org/apache/solr/core/TestArbitraryIndexDir.java =================================================================== --- solr/core/src/test/org/apache/solr/core/TestArbitraryIndexDir.java (revision 1297029) +++ solr/core/src/test/org/apache/solr/core/TestArbitraryIndexDir.java (working copy) @@ -71,7 +71,7 @@ + System.getProperty("file.separator") + "data"); dataDir.mkdirs(); - solrConfig = h.createConfig("solrconfig.xml"); + solrConfig = TestHarness.createConfig("solrconfig.xml"); h = new TestHarness( dataDir.getAbsolutePath(), solrConfig, "schema12.xml"); Index: solr/core/src/test/org/apache/solr/handler/component/DistributedSpellCheckComponentTest.java =================================================================== --- solr/core/src/test/org/apache/solr/handler/component/DistributedSpellCheckComponentTest.java (revision 1297029) +++ solr/core/src/test/org/apache/solr/handler/component/DistributedSpellCheckComponentTest.java (working copy) @@ -17,12 +17,14 @@ * limitations under the License. */ +import junit.framework.Assert; import junit.framework.TestCase; import org.apache.solr.BaseDistributedSearchTestCase; import org.apache.solr.client.solrj.SolrServer; import org.apache.solr.client.solrj.response.QueryResponse; import org.apache.solr.common.params.ModifiableSolrParams; +import org.apache.solr.common.params.SpellingParams; import org.apache.solr.common.util.NamedList; /** @@ -85,7 +87,7 @@ NamedList sc = (NamedList) nl.get("spellcheck"); NamedList sug = (NamedList) sc.get("suggestions"); if(sug.size()==0) { - TestCase.fail("Control data did not return any suggestions."); + Assert.fail("Control data did not return any suggestions."); } } @@ -124,16 +126,16 @@ // we care only about the spellcheck results handle.put("response", SKIP); - q("q", "*:*", "spellcheck", "true", SpellCheckComponent.SPELLCHECK_BUILD, "true", "qt", "spellCheckCompRH", "shards.qt", "spellCheckCompRH"); + q("q", "*:*", "spellcheck", "true", SpellingParams.SPELLCHECK_BUILD, "true", "qt", "spellCheckCompRH", "shards.qt", "spellCheckCompRH"); query("q", "*:*", "fl", "id,lowerfilt", "spellcheck.q","toyata", "spellcheck", "true", "qt", requestHandlerName, "shards.qt", requestHandlerName); - query("q", "*:*", "fl", "id,lowerfilt", "spellcheck.q","toyata", "spellcheck", "true", "qt", requestHandlerName, "shards.qt", requestHandlerName, SpellCheckComponent.SPELLCHECK_EXTENDED_RESULTS, "true"); - query("q", "*:*", "fl", "id,lowerfilt", "spellcheck.q","bluo", "spellcheck", "true", "qt", requestHandlerName, "shards.qt", requestHandlerName, SpellCheckComponent.SPELLCHECK_EXTENDED_RESULTS, "true", SpellCheckComponent.SPELLCHECK_COUNT, "4"); - query("q", "The quick reb fox jumped over the lazy brown dogs", "fl", "id,lowerfilt", "spellcheck", "true", "qt", requestHandlerName, "shards.qt", requestHandlerName, SpellCheckComponent.SPELLCHECK_EXTENDED_RESULTS, "true", SpellCheckComponent.SPELLCHECK_COUNT, "4", SpellCheckComponent.SPELLCHECK_COLLATE, "true"); + query("q", "*:*", "fl", "id,lowerfilt", "spellcheck.q","toyata", "spellcheck", "true", "qt", requestHandlerName, "shards.qt", requestHandlerName, SpellingParams.SPELLCHECK_EXTENDED_RESULTS, "true"); + query("q", "*:*", "fl", "id,lowerfilt", "spellcheck.q","bluo", "spellcheck", "true", "qt", requestHandlerName, "shards.qt", requestHandlerName, SpellingParams.SPELLCHECK_EXTENDED_RESULTS, "true", SpellingParams.SPELLCHECK_COUNT, "4"); + query("q", "The quick reb fox jumped over the lazy brown dogs", "fl", "id,lowerfilt", "spellcheck", "true", "qt", requestHandlerName, "shards.qt", requestHandlerName, SpellingParams.SPELLCHECK_EXTENDED_RESULTS, "true", SpellingParams.SPELLCHECK_COUNT, "4", SpellingParams.SPELLCHECK_COLLATE, "true"); - query("q", "lowerfilt:(+quock +reb)", "fl", "id,lowerfilt", "spellcheck", "true", "qt", requestHandlerName, "shards.qt", requestHandlerName, SpellCheckComponent.SPELLCHECK_EXTENDED_RESULTS, "true", SpellCheckComponent.SPELLCHECK_COUNT, "10", SpellCheckComponent.SPELLCHECK_COLLATE, "true", SpellCheckComponent.SPELLCHECK_MAX_COLLATION_TRIES, "10", SpellCheckComponent.SPELLCHECK_MAX_COLLATIONS, "10", SpellCheckComponent.SPELLCHECK_COLLATE_EXTENDED_RESULTS, "true"); - query("q", "lowerfilt:(+quock +reb)", "fl", "id,lowerfilt", "spellcheck", "true", "qt", requestHandlerName, "shards.qt", requestHandlerName, SpellCheckComponent.SPELLCHECK_EXTENDED_RESULTS, "true", SpellCheckComponent.SPELLCHECK_COUNT, "10", SpellCheckComponent.SPELLCHECK_COLLATE, "true", SpellCheckComponent.SPELLCHECK_MAX_COLLATION_TRIES, "10", SpellCheckComponent.SPELLCHECK_MAX_COLLATIONS, "10", SpellCheckComponent.SPELLCHECK_COLLATE_EXTENDED_RESULTS, "false"); - query("q", "lowerfilt:(+quock +reb)", "fl", "id,lowerfilt", "spellcheck", "true", "qt", requestHandlerName, "shards.qt", requestHandlerName, SpellCheckComponent.SPELLCHECK_EXTENDED_RESULTS, "true", SpellCheckComponent.SPELLCHECK_COUNT, "10", SpellCheckComponent.SPELLCHECK_COLLATE, "true", SpellCheckComponent.SPELLCHECK_MAX_COLLATION_TRIES, "0", SpellCheckComponent.SPELLCHECK_MAX_COLLATIONS, "1", SpellCheckComponent.SPELLCHECK_COLLATE_EXTENDED_RESULTS, "false"); + query("q", "lowerfilt:(+quock +reb)", "fl", "id,lowerfilt", "spellcheck", "true", "qt", requestHandlerName, "shards.qt", requestHandlerName, SpellingParams.SPELLCHECK_EXTENDED_RESULTS, "true", SpellingParams.SPELLCHECK_COUNT, "10", SpellingParams.SPELLCHECK_COLLATE, "true", SpellingParams.SPELLCHECK_MAX_COLLATION_TRIES, "10", SpellingParams.SPELLCHECK_MAX_COLLATIONS, "10", SpellingParams.SPELLCHECK_COLLATE_EXTENDED_RESULTS, "true"); + query("q", "lowerfilt:(+quock +reb)", "fl", "id,lowerfilt", "spellcheck", "true", "qt", requestHandlerName, "shards.qt", requestHandlerName, SpellingParams.SPELLCHECK_EXTENDED_RESULTS, "true", SpellingParams.SPELLCHECK_COUNT, "10", SpellingParams.SPELLCHECK_COLLATE, "true", SpellingParams.SPELLCHECK_MAX_COLLATION_TRIES, "10", SpellingParams.SPELLCHECK_MAX_COLLATIONS, "10", SpellingParams.SPELLCHECK_COLLATE_EXTENDED_RESULTS, "false"); + query("q", "lowerfilt:(+quock +reb)", "fl", "id,lowerfilt", "spellcheck", "true", "qt", requestHandlerName, "shards.qt", requestHandlerName, SpellingParams.SPELLCHECK_EXTENDED_RESULTS, "true", SpellingParams.SPELLCHECK_COUNT, "10", SpellingParams.SPELLCHECK_COLLATE, "true", SpellingParams.SPELLCHECK_MAX_COLLATION_TRIES, "0", SpellingParams.SPELLCHECK_MAX_COLLATIONS, "1", SpellingParams.SPELLCHECK_COLLATE_EXTENDED_RESULTS, "false"); } } Index: solr/core/src/test/org/apache/solr/handler/component/SpellCheckComponentTest.java =================================================================== --- solr/core/src/test/org/apache/solr/handler/component/SpellCheckComponentTest.java (revision 1297029) +++ solr/core/src/test/org/apache/solr/handler/component/SpellCheckComponentTest.java (working copy) @@ -32,6 +32,7 @@ import org.apache.solr.request.SolrRequestHandler; import org.apache.solr.response.SolrQueryResponse; import org.apache.solr.spelling.AbstractLuceneSpellChecker; +import org.apache.solr.spelling.SolrSpellChecker; import org.junit.BeforeClass; import org.junit.Test; @@ -76,12 +77,12 @@ @Test public void testExtendedResultsCount() throws Exception { - assertJQ(req("qt",rh, SpellCheckComponent.COMPONENT_NAME, "true", SpellCheckComponent.SPELLCHECK_BUILD, "true", "q","bluo", SpellCheckComponent.SPELLCHECK_COUNT,"5", SpellCheckComponent.SPELLCHECK_EXTENDED_RESULTS,"false") + assertJQ(req("qt",rh, SpellCheckComponent.COMPONENT_NAME, "true", SpellingParams.SPELLCHECK_BUILD, "true", "q","bluo", SpellingParams.SPELLCHECK_COUNT,"5", SpellingParams.SPELLCHECK_EXTENDED_RESULTS,"false") ,"/spellcheck/suggestions/[0]=='bluo'" ,"/spellcheck/suggestions/[1]/numFound==5" ); - assertJQ(req("qt",rh, SpellCheckComponent.COMPONENT_NAME, "true", "q","bluo", SpellCheckComponent.SPELLCHECK_COUNT,"3", SpellCheckComponent.SPELLCHECK_EXTENDED_RESULTS,"true") + assertJQ(req("qt",rh, SpellCheckComponent.COMPONENT_NAME, "true", "q","bluo", SpellingParams.SPELLCHECK_COUNT,"3", SpellingParams.SPELLCHECK_EXTENDED_RESULTS,"true") ,"/spellcheck/suggestions/[1]/suggestion==[{'word':'blud','freq':1}, {'word':'blue','freq':1}, {'word':'blee','freq':1}]" ); } @@ -96,7 +97,7 @@ @Test public void testPerDictionary() throws Exception { - assertJQ(req("json.nl","map", "qt",rh, SpellCheckComponent.COMPONENT_NAME, "true", SpellCheckComponent.SPELLCHECK_BUILD, "true", "q","documemt" + assertJQ(req("json.nl","map", "qt",rh, SpellCheckComponent.COMPONENT_NAME, "true", SpellingParams.SPELLCHECK_BUILD, "true", "q","documemt" , SpellingParams.SPELLCHECK_DICT, "perDict", SpellingParams.SPELLCHECK_PREFIX + ".perDict.foo", "bar", SpellingParams.SPELLCHECK_PREFIX + ".perDict.bar", "foo") ,"/spellcheck/suggestions/bar=={'numFound':1, 'startOffset':0, 'endOffset':1, 'suggestion':['foo']}" ,"/spellcheck/suggestions/foo=={'numFound':1, 'startOffset':2, 'endOffset':3, 'suggestion':['bar']}" @@ -105,16 +106,16 @@ @Test public void testCollate() throws Exception { - assertJQ(req("json.nl","map", "qt",rh, SpellCheckComponent.COMPONENT_NAME, "true", SpellCheckComponent.SPELLCHECK_BUILD, "true", "q","documemt", SpellCheckComponent.SPELLCHECK_COLLATE, "true") + assertJQ(req("json.nl","map", "qt",rh, SpellCheckComponent.COMPONENT_NAME, "true", SpellingParams.SPELLCHECK_BUILD, "true", "q","documemt", SpellingParams.SPELLCHECK_COLLATE, "true") ,"/spellcheck/suggestions/collation=='document'" ); - assertJQ(req("json.nl","map", "qt",rh, SpellCheckComponent.COMPONENT_NAME, "true", "q","documemt lowerfilt:broen^4", SpellCheckComponent.SPELLCHECK_COLLATE, "true") + assertJQ(req("json.nl","map", "qt",rh, SpellCheckComponent.COMPONENT_NAME, "true", "q","documemt lowerfilt:broen^4", SpellingParams.SPELLCHECK_COLLATE, "true") ,"/spellcheck/suggestions/collation=='document lowerfilt:brown^4'" ); - assertJQ(req("json.nl","map", "qt",rh, SpellCheckComponent.COMPONENT_NAME, "true", "q","documemtsss broens", SpellCheckComponent.SPELLCHECK_COLLATE, "true") + assertJQ(req("json.nl","map", "qt",rh, SpellCheckComponent.COMPONENT_NAME, "true", "q","documemtsss broens", SpellingParams.SPELLCHECK_COLLATE, "true") ,"/spellcheck/suggestions/collation=='document brown'" ); - assertJQ(req("json.nl","map", "qt",rh, SpellCheckComponent.COMPONENT_NAME, "true", "q","pixma", SpellCheckComponent.SPELLCHECK_COLLATE, "true") + assertJQ(req("json.nl","map", "qt",rh, SpellCheckComponent.COMPONENT_NAME, "true", "q","pixma", SpellingParams.SPELLCHECK_COLLATE, "true") ,"/spellcheck/suggestions/collation=='pixmaa'" ); } @@ -123,10 +124,10 @@ @Test public void testCorrectSpelling() throws Exception { // Make sure correct spellings are signaled in the response - assertJQ(req("json.nl","map", "qt",rh, SpellCheckComponent.COMPONENT_NAME, "true", "q","lowerfilt:lazy lowerfilt:brown", SpellCheckComponent.SPELLCHECK_EXTENDED_RESULTS, "true") + assertJQ(req("json.nl","map", "qt",rh, SpellCheckComponent.COMPONENT_NAME, "true", "q","lowerfilt:lazy lowerfilt:brown", SpellingParams.SPELLCHECK_EXTENDED_RESULTS, "true") ,"/spellcheck/suggestions=={'correctlySpelled':true}" ); - assertJQ(req("json.nl","map", "qt",rh, SpellCheckComponent.COMPONENT_NAME, "true", "q","lakkle", SpellCheckComponent.SPELLCHECK_EXTENDED_RESULTS, "true") + assertJQ(req("json.nl","map", "qt",rh, SpellCheckComponent.COMPONENT_NAME, "true", "q","lakkle", SpellingParams.SPELLCHECK_EXTENDED_RESULTS, "true") ,"/spellcheck/suggestions/correctlySpelled==false" ); } @@ -156,7 +157,7 @@ NamedList args = new NamedList(); NamedList spellchecker = new NamedList(); - spellchecker.add(AbstractLuceneSpellChecker.DICTIONARY_NAME, "default"); + spellchecker.add(SolrSpellChecker.DICTIONARY_NAME, "default"); spellchecker.add(AbstractLuceneSpellChecker.FIELD, "lowerfilt"); spellchecker.add(AbstractLuceneSpellChecker.INDEX_DIR, "spellchecker1"); args.add("spellchecker", spellchecker); @@ -201,11 +202,11 @@ //So with a threshold of 29%, "another" is absent from the dictionary //while "document" is present. - assertJQ(req("qt",rh, SpellCheckComponent.COMPONENT_NAME, "true", "q","documenq", SpellCheckComponent.SPELLCHECK_DICT, "threshold", SpellCheckComponent.SPELLCHECK_COUNT,"5", SpellCheckComponent.SPELLCHECK_EXTENDED_RESULTS,"true") + assertJQ(req("qt",rh, SpellCheckComponent.COMPONENT_NAME, "true", "q","documenq", SpellingParams.SPELLCHECK_DICT, "threshold", SpellingParams.SPELLCHECK_COUNT,"5", SpellingParams.SPELLCHECK_EXTENDED_RESULTS,"true") ,"/spellcheck/suggestions/[1]/suggestion==[{'word':'document','freq':2}]" ); - assertJQ(req("qt",rh, SpellCheckComponent.COMPONENT_NAME, "true", "q","documenq", SpellCheckComponent.SPELLCHECK_DICT, "threshold_direct", SpellCheckComponent.SPELLCHECK_COUNT,"5", SpellCheckComponent.SPELLCHECK_EXTENDED_RESULTS,"true") + assertJQ(req("qt",rh, SpellCheckComponent.COMPONENT_NAME, "true", "q","documenq", SpellingParams.SPELLCHECK_DICT, "threshold_direct", SpellingParams.SPELLCHECK_COUNT,"5", SpellingParams.SPELLCHECK_EXTENDED_RESULTS,"true") ,"/spellcheck/suggestions/[1]/suggestion==[{'word':'document','freq':2}]" ); @@ -216,9 +217,9 @@ ModifiableSolrParams params = new ModifiableSolrParams(); params.add(SpellCheckComponent.COMPONENT_NAME, "true"); - params.add(SpellCheckComponent.SPELLCHECK_COUNT, "10"); - params.add(SpellCheckComponent.SPELLCHECK_DICT, "threshold"); - params.add(SpellCheckComponent.SPELLCHECK_EXTENDED_RESULTS,"true"); + params.add(SpellingParams.SPELLCHECK_COUNT, "10"); + params.add(SpellingParams.SPELLCHECK_DICT, "threshold"); + params.add(SpellingParams.SPELLCHECK_EXTENDED_RESULTS,"true"); params.add(CommonParams.Q, "anotheq"); SolrRequestHandler handler = core.getRequestHandler("spellCheckCompRH"); @@ -233,8 +234,8 @@ assertTrue(suggestions.get("suggestion")==null); assertTrue((Boolean) suggestions.get("correctlySpelled")==false); - params.remove(SpellCheckComponent.SPELLCHECK_DICT); - params.add(SpellCheckComponent.SPELLCHECK_DICT, "threshold_direct"); + params.remove(SpellingParams.SPELLCHECK_DICT); + params.add(SpellingParams.SPELLCHECK_DICT, "threshold_direct"); rsp = new SolrQueryResponse(); rsp.add("responseHeader", new SimpleOrderedMap()); req = new LocalSolrQueryRequest(core, params); Index: solr/core/src/test/org/apache/solr/handler/TestReplicationHandler.java =================================================================== --- solr/core/src/test/org/apache/solr/handler/TestReplicationHandler.java (revision 1297029) +++ solr/core/src/test/org/apache/solr/handler/TestReplicationHandler.java (working copy) @@ -35,6 +35,7 @@ import org.apache.lucene.search.TopDocs; import org.apache.lucene.store.Directory; import org.apache.lucene.store.SimpleFSDirectory; +import org.apache.solr.BaseDistributedSearchTestCase; import org.apache.solr.SolrTestCaseJ4; import org.apache.solr.TestDistributedSearch; import org.apache.solr.client.solrj.SolrServer; @@ -389,7 +390,7 @@ assertEquals(nDocs, slaveQueryResult.getNumFound()); //compare results - String cmp = TestDistributedSearch.compare(masterQueryResult, slaveQueryResult, 0, null); + String cmp = BaseDistributedSearchTestCase.compare(masterQueryResult, slaveQueryResult, 0, null); assertEquals(null, cmp); //start config files replication test @@ -447,7 +448,7 @@ assertEquals(nDocs, slaveQueryResult.getNumFound()); //compare results - String cmp = TestDistributedSearch.compare(masterQueryResult, slaveQueryResult, 0, null); + String cmp = BaseDistributedSearchTestCase.compare(masterQueryResult, slaveQueryResult, 0, null); assertEquals(null, cmp); // start stop polling test @@ -527,7 +528,7 @@ SolrDocumentList slaveQueryResult = (SolrDocumentList) slaveQueryRsp.get("response"); assertEquals(nDocs, slaveQueryResult.getNumFound()); //compare results - String cmp = TestDistributedSearch.compare(masterQueryResult, slaveQueryResult, 0, null); + String cmp = BaseDistributedSearchTestCase.compare(masterQueryResult, slaveQueryResult, 0, null); assertEquals(null, cmp); System.out.println("replicate slave to master"); @@ -599,7 +600,7 @@ assertEquals(nDocs, slaveQueryResult.getNumFound()); //compare results - String cmp = TestDistributedSearch.compare(masterQueryResult, slaveQueryResult, 0, null); + String cmp = BaseDistributedSearchTestCase.compare(masterQueryResult, slaveQueryResult, 0, null); assertEquals(null, cmp); // NOTE: the master only replicates after startup now! @@ -654,7 +655,7 @@ assertEquals(10, slaveQueryResult.getNumFound()); //compare results - String cmp = TestDistributedSearch.compare(masterQueryResult, slaveQueryResult, 0, null); + String cmp = BaseDistributedSearchTestCase.compare(masterQueryResult, slaveQueryResult, 0, null); assertEquals(null, cmp); Object version = getIndexVersion(masterClient).get("indexversion"); @@ -714,7 +715,7 @@ assertEquals(nDocs, slaveQueryResult.getNumFound()); //compare results - String cmp = TestDistributedSearch.compare(masterQueryResult, slaveQueryResult, 0, null); + String cmp = BaseDistributedSearchTestCase.compare(masterQueryResult, slaveQueryResult, 0, null); assertEquals(null, cmp); //start config files replication test Index: solr/core/src/test/org/apache/solr/search/TestRealTimeGet.java =================================================================== --- solr/core/src/test/org/apache/solr/search/TestRealTimeGet.java (revision 1297029) +++ solr/core/src/test/org/apache/solr/search/TestRealTimeGet.java (working copy) @@ -34,6 +34,7 @@ import org.apache.solr.update.UpdateHandler; import org.apache.solr.update.UpdateLog; import org.apache.solr.update.VersionInfo; +import org.apache.solr.util.TestHarness; import org.junit.BeforeClass; import org.junit.Test; @@ -355,7 +356,7 @@ if (rand.nextInt(100) < softCommitPercent) { verbose("softCommit start"); - assertU(h.commit("softCommit","true")); + assertU(TestHarness.commit("softCommit","true")); verbose("softCommit end"); } else { verbose("hardCommit start"); @@ -577,7 +578,7 @@ if (rand.nextInt(100) < softCommitPercent) { verbose("softCommit start"); - assertU(h.commit("softCommit","true")); + assertU(TestHarness.commit("softCommit","true")); verbose("softCommit end"); } else { verbose("hardCommit start"); @@ -815,7 +816,7 @@ if (rand.nextInt(100) < softCommitPercent) { verbose("softCommit start"); - assertU(h.commit("softCommit","true")); + assertU(TestHarness.commit("softCommit","true")); verbose("softCommit end"); } else { verbose("hardCommit start"); @@ -1085,7 +1086,7 @@ if (uLog.getState() != UpdateLog.State.ACTIVE) version = -1; if (rand.nextInt(100) < softCommitPercent) { verbose("softCommit start"); - assertU(h.commit("softCommit","true")); + assertU(TestHarness.commit("softCommit","true")); verbose("softCommit end"); } else { verbose("hardCommit start"); Index: solr/core/src/test/org/apache/solr/spelling/DirectSolrSpellCheckerTest.java =================================================================== --- solr/core/src/test/org/apache/solr/spelling/DirectSolrSpellCheckerTest.java (revision 1297029) +++ solr/core/src/test/org/apache/solr/spelling/DirectSolrSpellCheckerTest.java (working copy) @@ -23,6 +23,7 @@ import org.apache.lucene.analysis.Token; import org.apache.solr.SolrTestCaseJ4; import org.apache.solr.common.params.CommonParams; +import org.apache.solr.common.params.SpellingParams; import org.apache.solr.common.util.NamedList; import org.apache.solr.core.SolrCore; import org.apache.solr.handler.component.SpellCheckComponent; @@ -59,7 +60,7 @@ DirectSolrSpellChecker checker = new DirectSolrSpellChecker(); NamedList spellchecker = new NamedList(); spellchecker.add("classname", DirectSolrSpellChecker.class.getName()); - spellchecker.add(DirectSolrSpellChecker.FIELD, "teststop"); + spellchecker.add(SolrSpellChecker.FIELD, "teststop"); spellchecker.add(DirectSolrSpellChecker.MINQUERYLENGTH, 2); // we will try "fob" SolrCore core = h.getCore(); @@ -85,7 +86,7 @@ @Test public void testOnlyMorePopularWithExtendedResults() throws Exception { - assertQ(req("q", "teststop:fox", "qt", "spellCheckCompRH", SpellCheckComponent.COMPONENT_NAME, "true", SpellCheckComponent.SPELLCHECK_DICT, "direct", SpellCheckComponent.SPELLCHECK_EXTENDED_RESULTS, "true", SpellCheckComponent.SPELLCHECK_ONLY_MORE_POPULAR, "true"), + assertQ(req("q", "teststop:fox", "qt", "spellCheckCompRH", SpellCheckComponent.COMPONENT_NAME, "true", SpellingParams.SPELLCHECK_DICT, "direct", SpellingParams.SPELLCHECK_EXTENDED_RESULTS, "true", SpellingParams.SPELLCHECK_ONLY_MORE_POPULAR, "true"), "//lst[@name='spellcheck']/lst[@name='suggestions']/lst[@name='fox']/int[@name='origFreq']=1", "//lst[@name='spellcheck']/lst[@name='suggestions']/lst[@name='fox']/arr[@name='suggestion']/lst/str[@name='word']='foo'", "//lst[@name='spellcheck']/lst[@name='suggestions']/lst[@name='fox']/arr[@name='suggestion']/lst/int[@name='freq']=2", Index: solr/core/src/test/org/apache/solr/spelling/FileBasedSpellCheckerTest.java =================================================================== --- solr/core/src/test/org/apache/solr/spelling/FileBasedSpellCheckerTest.java (revision 1297029) +++ solr/core/src/test/org/apache/solr/spelling/FileBasedSpellCheckerTest.java (working copy) @@ -66,11 +66,11 @@ spellchecker.add(SolrSpellChecker.DICTIONARY_NAME, "external"); spellchecker.add(AbstractLuceneSpellChecker.LOCATION, "spellings.txt"); - spellchecker.add(IndexBasedSpellChecker.FIELD, "teststop"); + spellchecker.add(AbstractLuceneSpellChecker.FIELD, "teststop"); spellchecker.add(FileBasedSpellChecker.SOURCE_FILE_CHAR_ENCODING, "UTF-8"); File indexDir = new File(TEMP_DIR, "spellingIdx" + new Date().getTime()); indexDir.mkdirs(); - spellchecker.add(FileBasedSpellChecker.INDEX_DIR, indexDir.getAbsolutePath()); + spellchecker.add(AbstractLuceneSpellChecker.INDEX_DIR, indexDir.getAbsolutePath()); SolrCore core = h.getCore(); String dictName = checker.init(spellchecker, core); assertTrue(dictName + " is not equal to " + "external", dictName.equals("external") == true); @@ -102,12 +102,12 @@ spellchecker.add("classname", FileBasedSpellChecker.class.getName()); spellchecker.add(SolrSpellChecker.DICTIONARY_NAME, "external"); spellchecker.add(AbstractLuceneSpellChecker.LOCATION, "spellings.txt"); - spellchecker.add(IndexBasedSpellChecker.FIELD, "teststop"); + spellchecker.add(AbstractLuceneSpellChecker.FIELD, "teststop"); spellchecker.add(FileBasedSpellChecker.SOURCE_FILE_CHAR_ENCODING, "UTF-8"); File indexDir = new File(TEMP_DIR, "spellingIdx" + new Date().getTime()); indexDir.mkdirs(); - spellchecker.add(FileBasedSpellChecker.INDEX_DIR, indexDir.getAbsolutePath()); - spellchecker.add(FileBasedSpellChecker.FIELD_TYPE, "teststop"); + spellchecker.add(AbstractLuceneSpellChecker.INDEX_DIR, indexDir.getAbsolutePath()); + spellchecker.add(SolrSpellChecker.FIELD_TYPE, "teststop"); spellchecker.add(AbstractLuceneSpellChecker.SPELLCHECKER_ARG_NAME, spellchecker); SolrCore core = h.getCore(); String dictName = checker.init(spellchecker, core); @@ -149,8 +149,8 @@ spellchecker.add(SolrSpellChecker.DICTIONARY_NAME, "external"); spellchecker.add(AbstractLuceneSpellChecker.LOCATION, "spellings.txt"); spellchecker.add(FileBasedSpellChecker.SOURCE_FILE_CHAR_ENCODING, "UTF-8"); - spellchecker.add(IndexBasedSpellChecker.FIELD, "teststop"); - spellchecker.add(FileBasedSpellChecker.FIELD_TYPE, "teststop"); + spellchecker.add(AbstractLuceneSpellChecker.FIELD, "teststop"); + spellchecker.add(SolrSpellChecker.FIELD_TYPE, "teststop"); spellchecker.add(AbstractLuceneSpellChecker.SPELLCHECKER_ARG_NAME, spellchecker); SolrCore core = h.getCore(); Index: solr/core/src/test/org/apache/solr/spelling/IndexBasedSpellCheckerTest.java =================================================================== --- solr/core/src/test/org/apache/solr/spelling/IndexBasedSpellCheckerTest.java (revision 1297029) +++ solr/core/src/test/org/apache/solr/spelling/IndexBasedSpellCheckerTest.java (working copy) @@ -110,7 +110,7 @@ File indexDir = new File(TEMP_DIR, "spellingIdx" + new Date().getTime()); indexDir.mkdirs(); spellchecker.add(AbstractLuceneSpellChecker.INDEX_DIR, indexDir.getAbsolutePath()); - spellchecker.add(IndexBasedSpellChecker.FIELD, "title"); + spellchecker.add(AbstractLuceneSpellChecker.FIELD, "title"); spellchecker.add(AbstractLuceneSpellChecker.SPELLCHECKER_ARG_NAME, spellchecker); SolrCore core = h.getCore(); @@ -186,7 +186,7 @@ File indexDir = new File(TEMP_DIR, "spellingIdx" + new Date().getTime()); indexDir.mkdirs(); spellchecker.add(AbstractLuceneSpellChecker.INDEX_DIR, indexDir.getAbsolutePath()); - spellchecker.add(IndexBasedSpellChecker.FIELD, "title"); + spellchecker.add(AbstractLuceneSpellChecker.FIELD, "title"); spellchecker.add(AbstractLuceneSpellChecker.SPELLCHECKER_ARG_NAME, spellchecker); SolrCore core = h.getCore(); String dictName = checker.init(spellchecker, core); @@ -243,7 +243,7 @@ File indexDir = new File(TEMP_DIR, "spellingIdx" + new Date().getTime()); indexDir.mkdirs(); spellchecker.add(AbstractLuceneSpellChecker.INDEX_DIR, indexDir.getAbsolutePath()); - spellchecker.add(IndexBasedSpellChecker.FIELD, "title"); + spellchecker.add(AbstractLuceneSpellChecker.FIELD, "title"); spellchecker.add(AbstractLuceneSpellChecker.SPELLCHECKER_ARG_NAME, spellchecker); spellchecker.add(AbstractLuceneSpellChecker.STRING_DISTANCE, JaroWinklerDistance.class.getName()); SolrCore core = h.getCore(); @@ -299,7 +299,7 @@ indexDir.mkdirs(); spellchecker.add(AbstractLuceneSpellChecker.INDEX_DIR, indexDir.getAbsolutePath()); spellchecker.add(AbstractLuceneSpellChecker.LOCATION, altIndexDir.getAbsolutePath()); - spellchecker.add(IndexBasedSpellChecker.FIELD, "title"); + spellchecker.add(AbstractLuceneSpellChecker.FIELD, "title"); spellchecker.add(AbstractLuceneSpellChecker.SPELLCHECKER_ARG_NAME, spellchecker); SolrCore core = h.getCore(); String dictName = checker.init(spellchecker, core); Index: solr/core/src/test/org/apache/solr/spelling/SpellCheckCollatorTest.java =================================================================== --- solr/core/src/test/org/apache/solr/spelling/SpellCheckCollatorTest.java (revision 1297029) +++ solr/core/src/test/org/apache/solr/spelling/SpellCheckCollatorTest.java (working copy) @@ -24,6 +24,7 @@ import org.apache.solr.common.params.CommonParams; import org.apache.solr.common.params.GroupParams; import org.apache.solr.common.params.ModifiableSolrParams; +import org.apache.solr.common.params.SpellingParams; import org.apache.solr.common.util.NamedList; import org.apache.solr.common.util.SimpleOrderedMap; import org.apache.solr.core.SolrCore; @@ -59,9 +60,9 @@ ModifiableSolrParams params = new ModifiableSolrParams(); params.add(SpellCheckComponent.COMPONENT_NAME, "true"); - params.add(SpellCheckComponent.SPELLCHECK_BUILD, "true"); - params.add(SpellCheckComponent.SPELLCHECK_COUNT, "10"); - params.add(SpellCheckComponent.SPELLCHECK_COLLATE, "true"); + params.add(SpellingParams.SPELLCHECK_BUILD, "true"); + params.add(SpellingParams.SPELLCHECK_COUNT, "10"); + params.add(SpellingParams.SPELLCHECK_COLLATE, "true"); params.add(CommonParams.Q, "lowerfilt:(hypenated-wotd)"); { @@ -111,11 +112,11 @@ ModifiableSolrParams params = new ModifiableSolrParams(); params.add(SpellCheckComponent.COMPONENT_NAME, "true"); - params.add(SpellCheckComponent.SPELLCHECK_BUILD, "true"); - params.add(SpellCheckComponent.SPELLCHECK_COUNT, "10"); - params.add(SpellCheckComponent.SPELLCHECK_COLLATE, "true"); - params.add(SpellCheckComponent.SPELLCHECK_MAX_COLLATION_TRIES, "10"); - params.add(SpellCheckComponent.SPELLCHECK_MAX_COLLATIONS, "10"); + params.add(SpellingParams.SPELLCHECK_BUILD, "true"); + params.add(SpellingParams.SPELLCHECK_COUNT, "10"); + params.add(SpellingParams.SPELLCHECK_COLLATE, "true"); + params.add(SpellingParams.SPELLCHECK_MAX_COLLATION_TRIES, "10"); + params.add(SpellingParams.SPELLCHECK_MAX_COLLATIONS, "10"); params.add(CommonParams.Q, "lowerfilt:(+fauth +home +loane)"); params.add(CommonParams.FQ, "NOT(id:1)"); @@ -146,12 +147,12 @@ ModifiableSolrParams params = new ModifiableSolrParams(); params.add(SpellCheckComponent.COMPONENT_NAME, "true"); - params.add(SpellCheckComponent.SPELLCHECK_DICT, "multipleFields"); - params.add(SpellCheckComponent.SPELLCHECK_BUILD, "true"); - params.add(SpellCheckComponent.SPELLCHECK_COUNT, "10"); - params.add(SpellCheckComponent.SPELLCHECK_COLLATE, "true"); - params.add(SpellCheckComponent.SPELLCHECK_MAX_COLLATION_TRIES, "1"); - params.add(SpellCheckComponent.SPELLCHECK_MAX_COLLATIONS, "1"); + params.add(SpellingParams.SPELLCHECK_DICT, "multipleFields"); + params.add(SpellingParams.SPELLCHECK_BUILD, "true"); + params.add(SpellingParams.SPELLCHECK_COUNT, "10"); + params.add(SpellingParams.SPELLCHECK_COLLATE, "true"); + params.add(SpellingParams.SPELLCHECK_MAX_COLLATION_TRIES, "1"); + params.add(SpellingParams.SPELLCHECK_MAX_COLLATIONS, "1"); params.add(CommonParams.Q, "peac"); //SpellCheckCompRH has no "qf" defined. It will not find "peace" from "peac" despite it being in the dictionary @@ -170,7 +171,7 @@ //SpellCheckCompRH1 has "lowerfilt1" defined in the "qf" param. It will find "peace" from "peac" because //requrying field "lowerfilt1" returns the hit. - params.remove(SpellCheckComponent.SPELLCHECK_BUILD); + params.remove(SpellingParams.SPELLCHECK_BUILD); handler = core.getRequestHandler("spellCheckCompRH1"); rsp = new SolrQueryResponse(); rsp.add("responseHeader", new SimpleOrderedMap()); @@ -193,11 +194,11 @@ ModifiableSolrParams params = new ModifiableSolrParams(); params.add(CommonParams.QT, "spellCheckCompRH"); params.add(CommonParams.Q, "lowerfilt:(+fauth +home +loane)"); - params.add(SpellCheckComponent.SPELLCHECK_EXTENDED_RESULTS, "true"); + params.add(SpellingParams.SPELLCHECK_EXTENDED_RESULTS, "true"); params.add(SpellCheckComponent.COMPONENT_NAME, "true"); - params.add(SpellCheckComponent.SPELLCHECK_BUILD, "true"); - params.add(SpellCheckComponent.SPELLCHECK_COUNT, "10"); - params.add(SpellCheckComponent.SPELLCHECK_COLLATE, "true"); + params.add(SpellingParams.SPELLCHECK_BUILD, "true"); + params.add(SpellingParams.SPELLCHECK_COUNT, "10"); + params.add(SpellingParams.SPELLCHECK_COLLATE, "true"); // Testing backwards-compatible behavior. // Returns 1 collation as a single string. @@ -217,9 +218,9 @@ // Testing backwards-compatible response format but will only return a // collation that would return results. - params.remove(SpellCheckComponent.SPELLCHECK_BUILD); - params.add(SpellCheckComponent.SPELLCHECK_MAX_COLLATION_TRIES, "5"); - params.add(SpellCheckComponent.SPELLCHECK_MAX_COLLATIONS, "1"); + params.remove(SpellingParams.SPELLCHECK_BUILD); + params.add(SpellingParams.SPELLCHECK_MAX_COLLATION_TRIES, "5"); + params.add(SpellingParams.SPELLCHECK_MAX_COLLATIONS, "1"); handler = core.getRequestHandler("spellCheckCompRH"); rsp = new SolrQueryResponse(); rsp.add("responseHeader", new SimpleOrderedMap()); @@ -234,10 +235,10 @@ // Testing returning multiple collations if more than one valid // combination exists. - params.remove(SpellCheckComponent.SPELLCHECK_MAX_COLLATION_TRIES); - params.remove(SpellCheckComponent.SPELLCHECK_MAX_COLLATIONS); - params.add(SpellCheckComponent.SPELLCHECK_MAX_COLLATION_TRIES, "10"); - params.add(SpellCheckComponent.SPELLCHECK_MAX_COLLATIONS, "2"); + params.remove(SpellingParams.SPELLCHECK_MAX_COLLATION_TRIES); + params.remove(SpellingParams.SPELLCHECK_MAX_COLLATIONS); + params.add(SpellingParams.SPELLCHECK_MAX_COLLATION_TRIES, "10"); + params.add(SpellingParams.SPELLCHECK_MAX_COLLATIONS, "2"); handler = core.getRequestHandler("spellCheckCompRH"); rsp = new SolrQueryResponse(); rsp.add("responseHeader", new SimpleOrderedMap()); @@ -256,7 +257,7 @@ // Testing return multiple collations with expanded collation response // format. - params.add(SpellCheckComponent.SPELLCHECK_COLLATE_EXTENDED_RESULTS, "true"); + params.add(SpellingParams.SPELLCHECK_COLLATE_EXTENDED_RESULTS, "true"); handler = core.getRequestHandler("spellCheckCompRH"); rsp = new SolrQueryResponse(); rsp.add("responseHeader", new SimpleOrderedMap()); @@ -300,11 +301,11 @@ ModifiableSolrParams params = new ModifiableSolrParams(); params.add(SpellCheckComponent.COMPONENT_NAME, "true"); - params.add(SpellCheckComponent.SPELLCHECK_BUILD, "true"); - params.add(SpellCheckComponent.SPELLCHECK_COUNT, "10"); - params.add(SpellCheckComponent.SPELLCHECK_COLLATE, "true"); - params.add(SpellCheckComponent.SPELLCHECK_MAX_COLLATION_TRIES, "5"); - params.add(SpellCheckComponent.SPELLCHECK_MAX_COLLATIONS, "1"); + params.add(SpellingParams.SPELLCHECK_BUILD, "true"); + params.add(SpellingParams.SPELLCHECK_COUNT, "10"); + params.add(SpellingParams.SPELLCHECK_COLLATE, "true"); + params.add(SpellingParams.SPELLCHECK_MAX_COLLATION_TRIES, "5"); + params.add(SpellingParams.SPELLCHECK_MAX_COLLATIONS, "1"); params.add(CommonParams.Q, "lowerfilt:(+fauth)"); params.add(GroupParams.GROUP, "true"); params.add(GroupParams.GROUP_FIELD, "id"); Index: solr/core/src/test/org/apache/solr/util/DateMathParserTest.java =================================================================== --- solr/core/src/test/org/apache/solr/util/DateMathParserTest.java (revision 1297029) +++ solr/core/src/test/org/apache/solr/util/DateMathParserTest.java (working copy) @@ -167,7 +167,8 @@ Date trash = p.parseMath("+7YEARS"); trash = p.parseMath("/MONTH"); trash = p.parseMath("-5DAYS+20MINUTES"); - Thread.currentThread().sleep(5); + Thread.currentThread(); + Thread.sleep(5); String a = fmt.format(p.parseMath("")); assertEquals("State of DateMathParser changed", e, a); Index: solr/solrj/src/test/org/apache/solr/client/solrj/embedded/JettyWebappTest.java =================================================================== --- solr/solrj/src/test/org/apache/solr/client/solrj/embedded/JettyWebappTest.java (revision 1297029) +++ solr/solrj/src/test/org/apache/solr/client/solrj/embedded/JettyWebappTest.java (working copy) @@ -56,7 +56,7 @@ super.setUp(); System.setProperty("solr.solr.home", ExternalPaths.EXAMPLE_HOME); - File dataDir = new File(SolrTestCaseJ4.TEMP_DIR, + File dataDir = new File(LuceneTestCase.TEMP_DIR, getClass().getName() + "-" + System.currentTimeMillis()); dataDir.mkdirs(); System.setProperty("solr.data.dir", dataDir.getCanonicalPath()); Index: solr/solrj/src/test/org/apache/solr/client/solrj/TestLBHttpSolrServer.java =================================================================== --- solr/solrj/src/test/org/apache/solr/client/solrj/TestLBHttpSolrServer.java (revision 1297029) +++ solr/solrj/src/test/org/apache/solr/client/solrj/TestLBHttpSolrServer.java (working copy) @@ -248,7 +248,7 @@ } public void setUp() throws Exception { - File home = new File(SolrTestCaseJ4.TEMP_DIR, + File home = new File(LuceneTestCase.TEMP_DIR, getClass().getName() + "-" + System.currentTimeMillis()); Index: solr/test-framework/src/java/org/apache/solr/BaseDistributedSearchTestCase.java =================================================================== --- solr/test-framework/src/java/org/apache/solr/BaseDistributedSearchTestCase.java (revision 1297029) +++ solr/test-framework/src/java/org/apache/solr/BaseDistributedSearchTestCase.java (working copy) @@ -30,6 +30,7 @@ import java.util.Random; import java.util.Set; +import junit.framework.Assert; import junit.framework.TestCase; import org.apache.solr.client.solrj.SolrServer; @@ -652,7 +653,7 @@ cmp = compare(a.getResponse(), b.getResponse(), flags, handle); if (cmp != null) { log.error("Mismatched responses:\n" + a + "\n" + b); - TestCase.fail(cmp); + Assert.fail(cmp); } } Index: solr/test-framework/src/java/org/apache/solr/SolrTestCaseJ4.java =================================================================== --- solr/test-framework/src/java/org/apache/solr/SolrTestCaseJ4.java (revision 1297029) +++ solr/test-framework/src/java/org/apache/solr/SolrTestCaseJ4.java (working copy) @@ -347,7 +347,7 @@ } public static void createCore() throws Exception { - solrConfig = h.createConfig(getSolrConfigFile()); + solrConfig = TestHarness.createConfig(getSolrConfigFile()); h = new TestHarness( dataDir.getAbsolutePath(), solrConfig, getSchemaFile()); @@ -612,13 +612,13 @@ * @see TestHarness#optimize */ public static String optimize(String... args) { - return h.optimize(args); + return TestHarness.optimize(args); } /** * @see TestHarness#commit */ public static String commit(String... args) { - return h.commit(args); + return TestHarness.commit(args); } /** @@ -683,7 +683,7 @@ * @see TestHarness#deleteById */ public static String delI(String id) { - return h.deleteById(id); + return TestHarness.deleteById(id); } /** * Generates a <delete>... XML string for an query @@ -691,7 +691,7 @@ * @see TestHarness#deleteByQuery */ public static String delQ(String q) { - return h.deleteByQuery(q); + return TestHarness.deleteByQuery(q); } /** @@ -702,7 +702,7 @@ */ public static XmlDoc doc(String... fieldsAndValues) { XmlDoc d = new XmlDoc(); - d.xml = h.makeSimpleDoc(fieldsAndValues).toString(); + d.xml = TestHarness.makeSimpleDoc(fieldsAndValues).toString(); return d; } Index: solr/test-framework/src/java/org/apache/solr/util/AbstractSolrTestCase.java =================================================================== --- solr/test-framework/src/java/org/apache/solr/util/AbstractSolrTestCase.java (revision 1297029) +++ solr/test-framework/src/java/org/apache/solr/util/AbstractSolrTestCase.java (working copy) @@ -155,7 +155,7 @@ System.setProperty("solr.solr.home", getSolrHome()); if (configFile != null) { - solrConfig = h.createConfig(getSolrConfigFile()); + solrConfig = TestHarness.createConfig(getSolrConfigFile()); h = new TestHarness( dataDir.getAbsolutePath(), solrConfig, getSchemaFile()); @@ -312,13 +312,13 @@ * @see TestHarness#optimize */ public String optimize(String... args) { - return h.optimize(args); + return TestHarness.optimize(args); } /** * @see TestHarness#commit */ public String commit(String... args) { - return h.commit(args); + return TestHarness.commit(args); } /** @@ -397,7 +397,7 @@ * @see TestHarness#deleteById */ public String delI(String id, String... args) { - return h.deleteById(id, args); + return TestHarness.deleteById(id, args); } /** @@ -406,7 +406,7 @@ * @see TestHarness#deleteByQuery */ public String delQ(String q, String... args) { - return h.deleteByQuery(q, args); + return TestHarness.deleteByQuery(q, args); } /** @@ -417,7 +417,7 @@ */ public Doc doc(String... fieldsAndValues) { Doc d = new Doc(); - d.xml = h.makeSimpleDoc(fieldsAndValues).toString(); + d.xml = TestHarness.makeSimpleDoc(fieldsAndValues).toString(); return d; }