Index: solr/core/src/test/org/apache/solr/search/TestRealTimeGet.java =================================================================== --- solr/core/src/test/org/apache/solr/search/TestRealTimeGet.java (revision 1245181) +++ solr/core/src/test/org/apache/solr/search/TestRealTimeGet.java (working copy) @@ -28,8 +28,6 @@ import org.apache.noggit.ObjectBuilder; import org.apache.solr.SolrTestCaseJ4; import org.apache.solr.common.SolrException; -import org.apache.solr.common.SolrInputDocument; -import org.apache.solr.common.params.SolrParams; import org.apache.solr.request.SolrQueryRequest; import org.apache.solr.update.UpdateHandler; import org.apache.solr.update.UpdateLog; @@ -1673,7 +1671,7 @@ public int getFirstMatch(IndexReader r, Term t) throws IOException { - Fields fields = MultiFields.getFields(r); + InvertedFields fields = MultiFields.getFields(r); if (fields == null) return -1; Terms terms = fields.terms(t.field()); if (terms == null) return -1; Index: solr/core/src/java/org/apache/solr/request/SimpleFacets.java =================================================================== --- solr/core/src/java/org/apache/solr/request/SimpleFacets.java (revision 1245181) +++ solr/core/src/java/org/apache/solr/request/SimpleFacets.java (working copy) @@ -34,7 +34,6 @@ import org.apache.solr.common.util.NamedList; import org.apache.solr.common.util.SimpleOrderedMap; import org.apache.solr.common.util.StrUtils; -import org.apache.solr.core.SolrCore; import org.apache.solr.schema.*; import org.apache.solr.search.*; import org.apache.solr.util.BoundedTreeSet; @@ -627,7 +626,7 @@ startTermBytes = new BytesRef(indexedPrefix); } - Fields fields = r.fields(); + InvertedFields fields = r.fields(); Terms terms = fields==null ? null : fields.terms(field); TermsEnum termsEnum = null; SolrIndexSearcher.DocsEnumState deState = null; Index: solr/core/src/java/org/apache/solr/search/JoinQParserPlugin.java =================================================================== --- solr/core/src/java/org/apache/solr/search/JoinQParserPlugin.java (revision 1245181) +++ solr/core/src/java/org/apache/solr/search/JoinQParserPlugin.java (working copy) @@ -262,8 +262,8 @@ fastForRandomSet = new HashDocSet(sset.getDocs(), 0, sset.size()); } - Fields fromFields = fromSearcher.getAtomicReader().fields(); - Fields toFields = fromSearcher==toSearcher ? fromFields : toSearcher.getAtomicReader().fields(); + InvertedFields fromFields = fromSearcher.getAtomicReader().fields(); + InvertedFields toFields = fromSearcher==toSearcher ? fromFields : toSearcher.getAtomicReader().fields(); if (fromFields == null) return DocSet.EMPTY; Terms terms = fromFields.terms(fromField); Terms toTerms = toFields.terms(toField); Index: solr/core/src/java/org/apache/solr/search/SolrIndexSearcher.java =================================================================== --- solr/core/src/java/org/apache/solr/search/SolrIndexSearcher.java (revision 1245181) +++ solr/core/src/java/org/apache/solr/search/SolrIndexSearcher.java (working copy) @@ -38,7 +38,6 @@ import org.apache.lucene.util.Bits; import org.apache.lucene.util.BytesRef; import org.apache.lucene.util.OpenBitSet; -import org.apache.lucene.util.ReaderUtil; import org.apache.solr.common.params.ModifiableSolrParams; import org.apache.solr.common.util.NamedList; import org.apache.solr.common.util.SimpleOrderedMap; @@ -569,7 +568,7 @@ * @return the first document number containing the term */ public int getFirstMatch(Term t) throws IOException { - Fields fields = atomicReader.fields(); + InvertedFields fields = atomicReader.fields(); if (fields == null) return -1; Terms terms = fields.terms(t.field()); if (terms == null) return -1; @@ -597,7 +596,7 @@ final AtomicReaderContext leaf = leaves[i]; final AtomicReader reader = leaf.reader(); - final Fields fields = reader.fields(); + final InvertedFields fields = reader.fields(); if (fields == null) continue; final Bits liveDocs = reader.getLiveDocs(); @@ -983,7 +982,7 @@ final AtomicReaderContext leaf = leaves[i]; final AtomicReader reader = leaf.reader(); collector.setNextReader(leaf); - Fields fields = reader.fields(); + InvertedFields fields = reader.fields(); Terms terms = fields.terms(t.field()); BytesRef termBytes = t.bytes(); Index: solr/core/src/java/org/apache/solr/handler/admin/LukeRequestHandler.java =================================================================== --- solr/core/src/java/org/apache/solr/handler/admin/LukeRequestHandler.java (revision 1245181) +++ solr/core/src/java/org/apache/solr/handler/admin/LukeRequestHandler.java (working copy) @@ -574,7 +574,7 @@ final CharsRef spare = new CharsRef(); - Fields fields = MultiFields.getFields(req.getSearcher().getIndexReader()); + InvertedFields fields = MultiFields.getFields(req.getSearcher().getIndexReader()); if (fields == null) { // No indexed fields return; Index: solr/core/src/java/org/apache/solr/handler/component/QueryElevationComponent.java =================================================================== --- solr/core/src/java/org/apache/solr/handler/component/QueryElevationComponent.java (revision 1245181) +++ solr/core/src/java/org/apache/solr/handler/component/QueryElevationComponent.java (working copy) @@ -535,7 +535,7 @@ public FieldComparator setNextReader(AtomicReaderContext context) throws IOException { //convert the ids to Lucene doc ids, the ordSet and termValues needs to be the same size as the number of elevation docs we have ordSet.clear(); - Fields fields = context.reader().fields(); + InvertedFields fields = context.reader().fields(); if (fields == null) return this; Terms terms = fields.terms(fieldname); if (terms == null) return this; Index: solr/core/src/java/org/apache/solr/handler/component/TermVectorComponent.java =================================================================== --- solr/core/src/java/org/apache/solr/handler/component/TermVectorComponent.java (revision 1245181) +++ solr/core/src/java/org/apache/solr/handler/component/TermVectorComponent.java (working copy) @@ -8,14 +8,8 @@ import java.util.List; import java.util.Map; -import org.apache.lucene.index.DocsAndPositionsEnum; -import org.apache.lucene.index.FieldInfo; -import org.apache.lucene.index.Fields; -import org.apache.lucene.index.FieldsEnum; -import org.apache.lucene.index.IndexReader; -import org.apache.lucene.index.StoredFieldVisitor; -import org.apache.lucene.index.Terms; -import org.apache.lucene.index.TermsEnum; +import org.apache.lucene.index.*; +import org.apache.lucene.index.InvertedFields; import org.apache.lucene.util.BytesRef; import org.apache.solr.common.SolrException; import org.apache.solr.common.params.CommonParams; @@ -251,7 +245,7 @@ } } else { // extract all fields - final Fields vectors = reader.getTermVectors(docId); + final InvertedFields vectors = reader.getTermVectors(docId); final FieldsEnum fieldsEnum = vectors.iterator(); String field; while((field = fieldsEnum.next()) != null) { Index: solr/core/src/java/org/apache/solr/handler/component/TermsComponent.java =================================================================== --- solr/core/src/java/org/apache/solr/handler/component/TermsComponent.java (revision 1245181) +++ solr/core/src/java/org/apache/solr/handler/component/TermsComponent.java (working copy) @@ -118,7 +118,7 @@ final AtomicReader indexReader = rb.req.getSearcher().getAtomicReader(); - Fields lfields = indexReader.fields(); + InvertedFields lfields = indexReader.fields(); for (String field : fields) { NamedList fieldTerms = new NamedList(); Index: modules/benchmark/src/test/org/apache/lucene/benchmark/byTask/TestPerfTasksLogic.java =================================================================== --- modules/benchmark/src/test/org/apache/lucene/benchmark/byTask/TestPerfTasksLogic.java (revision 1245181) +++ modules/benchmark/src/test/org/apache/lucene/benchmark/byTask/TestPerfTasksLogic.java (working copy) @@ -39,22 +39,9 @@ import org.apache.lucene.benchmark.byTask.tasks.WriteLineDocTask; import org.apache.lucene.collation.CollationKeyAnalyzer; import org.apache.lucene.facet.taxonomy.TaxonomyReader; -import org.apache.lucene.index.DirectoryReader; -import org.apache.lucene.index.DocsEnum; -import org.apache.lucene.index.Fields; -import org.apache.lucene.index.FieldsEnum; -import org.apache.lucene.index.IndexReader; -import org.apache.lucene.index.IndexWriter; +import org.apache.lucene.index.*; +import org.apache.lucene.index.InvertedFields; import org.apache.lucene.index.IndexWriterConfig.OpenMode; -import org.apache.lucene.index.IndexWriterConfig; -import org.apache.lucene.index.LogDocMergePolicy; -import org.apache.lucene.index.LogMergePolicy; -import org.apache.lucene.index.MultiFields; -import org.apache.lucene.index.SegmentInfos; -import org.apache.lucene.index.SerialMergeScheduler; -import org.apache.lucene.index.SlowCompositeReaderWrapper; -import org.apache.lucene.index.Terms; -import org.apache.lucene.index.TermsEnum; import org.apache.lucene.search.FieldCache.DocTermsIndex; import org.apache.lucene.search.FieldCache; import org.apache.lucene.store.Directory; @@ -749,7 +736,7 @@ writer.close(); Directory dir = benchmark.getRunData().getDirectory(); IndexReader reader = IndexReader.open(dir); - Fields tfv = reader.getTermVectors(0); + InvertedFields tfv = reader.getTermVectors(0); assertNotNull(tfv); assertTrue(tfv.getUniqueFieldCount() > 0); reader.close(); Index: modules/queries/src/java/org/apache/lucene/queries/TermsFilter.java =================================================================== --- modules/queries/src/java/org/apache/lucene/queries/TermsFilter.java (revision 1245181) +++ modules/queries/src/java/org/apache/lucene/queries/TermsFilter.java (working copy) @@ -56,7 +56,7 @@ public DocIdSet getDocIdSet(AtomicReaderContext context, Bits acceptDocs) throws IOException { AtomicReader reader = context.reader(); FixedBitSet result = new FixedBitSet(reader.maxDoc()); - Fields fields = reader.fields(); + InvertedFields fields = reader.fields(); if (fields == null) { return result; Index: modules/queries/src/java/org/apache/lucene/queries/function/valuesource/SumTotalTermFreqValueSource.java =================================================================== --- modules/queries/src/java/org/apache/lucene/queries/function/valuesource/SumTotalTermFreqValueSource.java (revision 1245181) +++ modules/queries/src/java/org/apache/lucene/queries/function/valuesource/SumTotalTermFreqValueSource.java (working copy) @@ -18,13 +18,12 @@ package org.apache.lucene.queries.function.valuesource; import org.apache.lucene.index.AtomicReaderContext; -import org.apache.lucene.index.Fields; +import org.apache.lucene.index.InvertedFields; import org.apache.lucene.index.Terms; import org.apache.lucene.queries.function.FunctionValues; import org.apache.lucene.queries.function.ValueSource; import org.apache.lucene.queries.function.docvalues.LongDocValues; import org.apache.lucene.search.IndexSearcher; -import org.apache.lucene.util.BytesRef; import java.io.IOException; import java.util.Map; @@ -58,7 +57,7 @@ public void createWeight(Map context, IndexSearcher searcher) throws IOException { long sumTotalTermFreq = 0; for (AtomicReaderContext readerContext : searcher.getTopReaderContext().leaves()) { - Fields fields = readerContext.reader().fields(); + InvertedFields fields = readerContext.reader().fields(); if (fields == null) continue; Terms terms = fields.terms(indexedField); if (terms == null) continue; Index: modules/queries/src/java/org/apache/lucene/queries/function/valuesource/TermFreqValueSource.java =================================================================== --- modules/queries/src/java/org/apache/lucene/queries/function/valuesource/TermFreqValueSource.java (revision 1245181) +++ modules/queries/src/java/org/apache/lucene/queries/function/valuesource/TermFreqValueSource.java (working copy) @@ -38,7 +38,7 @@ @Override public FunctionValues getValues(Map context, AtomicReaderContext readerContext) throws IOException { - Fields fields = readerContext.reader().fields(); + InvertedFields fields = readerContext.reader().fields(); final Terms terms = fields.terms(field); return new IntDocValues(this) { Index: modules/queries/src/java/org/apache/lucene/queries/function/valuesource/TFValueSource.java =================================================================== --- modules/queries/src/java/org/apache/lucene/queries/function/valuesource/TFValueSource.java (revision 1245181) +++ modules/queries/src/java/org/apache/lucene/queries/function/valuesource/TFValueSource.java (working copy) @@ -40,7 +40,7 @@ @Override public FunctionValues getValues(Map context, AtomicReaderContext readerContext) throws IOException { - Fields fields = readerContext.reader().fields(); + InvertedFields fields = readerContext.reader().fields(); final Terms terms = fields.terms(field); IndexSearcher searcher = (IndexSearcher)context.get("searcher"); final TFIDFSimilarity similarity = IDFValueSource.asTFIDF(searcher.getSimilarity(), field); Index: modules/queries/src/java/org/apache/lucene/queries/mlt/MoreLikeThis.java =================================================================== --- modules/queries/src/java/org/apache/lucene/queries/mlt/MoreLikeThis.java (revision 1245181) +++ modules/queries/src/java/org/apache/lucene/queries/mlt/MoreLikeThis.java (working copy) @@ -22,7 +22,7 @@ import org.apache.lucene.analysis.TokenStream; import org.apache.lucene.analysis.tokenattributes.CharTermAttribute; import org.apache.lucene.document.Document; -import org.apache.lucene.index.Fields; +import org.apache.lucene.index.InvertedFields; import org.apache.lucene.index.IndexReader; import org.apache.lucene.index.IndexableField; import org.apache.lucene.index.MultiFields; @@ -705,7 +705,7 @@ public PriorityQueue retrieveTerms(int docNum) throws IOException { Map termFreqMap = new HashMap(); for (String fieldName : fieldNames) { - final Fields vectors = ir.getTermVectors(docNum); + final InvertedFields vectors = ir.getTermVectors(docNum); final Terms vector; if (vectors != null) { vector = vectors.terms(fieldName); Index: lucene/core/src/test/org/apache/lucene/codecs/appending/TestAppendingCodec.java =================================================================== --- lucene/core/src/test/org/apache/lucene/codecs/appending/TestAppendingCodec.java (revision 1245181) +++ lucene/core/src/test/org/apache/lucene/codecs/appending/TestAppendingCodec.java (working copy) @@ -21,12 +21,11 @@ import java.util.Random; import org.apache.lucene.analysis.MockAnalyzer; -import org.apache.lucene.codecs.appending.AppendingCodec; import org.apache.lucene.document.Document; import org.apache.lucene.document.FieldType; import org.apache.lucene.document.TextField; import org.apache.lucene.index.DocsEnum; -import org.apache.lucene.index.Fields; +import org.apache.lucene.index.InvertedFields; import org.apache.lucene.index.IndexReader; import org.apache.lucene.index.IndexWriter; import org.apache.lucene.index.IndexWriterConfig; @@ -128,7 +127,7 @@ assertEquals(2, reader.numDocs()); Document doc2 = reader.document(0); assertEquals(text, doc2.get("f")); - Fields fields = MultiFields.getFields(reader); + InvertedFields fields = MultiFields.getFields(reader); Terms terms = fields.terms("f"); assertNotNull(terms); TermsEnum te = terms.iterator(null); Index: lucene/core/src/test/org/apache/lucene/codecs/lucene3x/TestSurrogates.java =================================================================== --- lucene/core/src/test/org/apache/lucene/codecs/lucene3x/TestSurrogates.java (revision 1245181) +++ lucene/core/src/test/org/apache/lucene/codecs/lucene3x/TestSurrogates.java (working copy) @@ -342,7 +342,7 @@ //SegmentInfo si = makePreFlexSegment(r, "_0", dir, fieldInfos, codec, fieldTerms); - //FieldsProducer fields = codec.fieldsProducer(new SegmentReadState(dir, si, fieldInfos, 1024, 1)); + //InvertedFieldsProducer fields = codec.fieldsProducer(new SegmentReadState(dir, si, fieldInfos, 1024, 1)); //assertNotNull(fields); doTestStraightEnum(fieldTerms, reader, uniqueTermCount); Index: lucene/core/src/test/org/apache/lucene/search/TestTermVectors.java =================================================================== --- lucene/core/src/test/org/apache/lucene/search/TestTermVectors.java (revision 1245181) +++ lucene/core/src/test/org/apache/lucene/search/TestTermVectors.java (working copy) @@ -96,7 +96,7 @@ assertEquals(100, hits.length); for (int i = 0; i < hits.length; i++) { - Fields vectors = searcher.reader.getTermVectors(hits[i].doc); + InvertedFields vectors = searcher.reader.getTermVectors(hits[i].doc); assertNotNull(vectors); assertEquals("doc=" + hits[i].doc + " tv=" + vectors, 1, vectors.getUniqueFieldCount()); } @@ -120,7 +120,7 @@ writer.addDocument(doc); IndexReader reader = writer.getReader(); writer.close(); - Fields v = reader.getTermVectors(0); + InvertedFields v = reader.getTermVectors(0); assertEquals(4, v.getUniqueFieldCount()); String[] expectedFields = new String[]{"a", "b", "c", "x"}; int[] expectedPositions = new int[]{1, 2, 0}; @@ -164,7 +164,7 @@ DocsAndPositionsEnum dpEnum = null; for (int i = 0; i < hits.length; i++) { - Fields vectors = searcher.reader.getTermVectors(hits[i].doc); + InvertedFields vectors = searcher.reader.getTermVectors(hits[i].doc); assertNotNull(vectors); assertEquals(1, vectors.getUniqueFieldCount()); @@ -203,7 +203,7 @@ assertEquals(100, hits.length); for (int i = 0; i < hits.length; i++) { - Fields vectors = searcher.reader.getTermVectors(hits[i].doc); + InvertedFields vectors = searcher.reader.getTermVectors(hits[i].doc); assertNotNull(vectors); assertEquals(1, vectors.getUniqueFieldCount()); } @@ -369,7 +369,7 @@ assertEquals(10, hits.length); for (int i = 0; i < hits.length; i++) { - Fields vectors = searcher.reader.getTermVectors(hits[i].doc); + InvertedFields vectors = searcher.reader.getTermVectors(hits[i].doc); assertNotNull(vectors); assertEquals(1, vectors.getUniqueFieldCount()); } @@ -416,7 +416,7 @@ ScoreDoc[] hits = searcher.search(query, null, 1000).scoreDocs; assertEquals(1, hits.length); - Fields vectors = searcher.reader.getTermVectors(hits[0].doc); + InvertedFields vectors = searcher.reader.getTermVectors(hits[0].doc); assertNotNull(vectors); assertEquals(1, vectors.getUniqueFieldCount()); Terms vector = vectors.terms("field"); Index: lucene/core/src/test/org/apache/lucene/search/TestMultiThreadTermVectors.java =================================================================== --- lucene/core/src/test/org/apache/lucene/search/TestMultiThreadTermVectors.java (revision 1245181) +++ lucene/core/src/test/org/apache/lucene/search/TestMultiThreadTermVectors.java (working copy) @@ -21,7 +21,7 @@ import org.apache.lucene.analysis.MockAnalyzer; import org.apache.lucene.document.*; -import org.apache.lucene.index.Fields; +import org.apache.lucene.index.InvertedFields; import org.apache.lucene.index.FieldsEnum; import org.apache.lucene.index.IndexReader; import org.apache.lucene.index.IndexWriter; @@ -163,7 +163,7 @@ long start = 0L; for (int docId = 0; docId < numDocs; docId++) { start = System.currentTimeMillis(); - Fields vectors = reader.getTermVectors(docId); + InvertedFields vectors = reader.getTermVectors(docId); timeElapsed += System.currentTimeMillis()-start; // verify vectors result @@ -177,7 +177,7 @@ } } - private void verifyVectors(Fields vectors, int num) throws IOException { + private void verifyVectors(InvertedFields vectors, int num) throws IOException { FieldsEnum fieldsEnum = vectors.iterator(); while(fieldsEnum.next() != null) { Terms terms = fieldsEnum.terms(); Index: lucene/core/src/test/org/apache/lucene/index/TestIndexableField.java =================================================================== --- lucene/core/src/test/org/apache/lucene/index/TestIndexableField.java (revision 1245181) +++ lucene/core/src/test/org/apache/lucene/index/TestIndexableField.java (working copy) @@ -280,7 +280,7 @@ // TODO: offsets } else { - Fields vectors = r.getTermVectors(docID); + InvertedFields vectors = r.getTermVectors(docID); assertTrue(vectors == null || vectors.terms(name) == null); } Index: lucene/core/src/test/org/apache/lucene/index/TestFilterAtomicReader.java =================================================================== --- lucene/core/src/test/org/apache/lucene/index/TestFilterAtomicReader.java (revision 1245181) +++ lucene/core/src/test/org/apache/lucene/index/TestFilterAtomicReader.java (working copy) @@ -21,7 +21,6 @@ import java.io.IOException; import java.lang.reflect.Method; import java.lang.reflect.Modifier; -import java.util.HashSet; import org.apache.lucene.analysis.MockAnalyzer; import org.apache.lucene.document.Document; @@ -31,7 +30,6 @@ import org.apache.lucene.util.Bits; import org.apache.lucene.util.BytesRef; import org.apache.lucene.util.LuceneTestCase; -import org.apache.lucene.util.ReaderUtil; public class TestFilterAtomicReader extends LuceneTestCase { @@ -39,7 +37,7 @@ /** Filter that only permits terms containing 'e'.*/ private static class TestFields extends FilterFields { - TestFields(Fields in) { + TestFields(InvertedFields in) { super(in); } @Override @@ -119,7 +117,7 @@ } @Override - public Fields fields() throws IOException { + public InvertedFields fields() throws IOException { return new TestFields(super.fields()); } } Index: lucene/core/src/test/org/apache/lucene/index/TestDuelingCodecs.java =================================================================== --- lucene/core/src/test/org/apache/lucene/index/TestDuelingCodecs.java (revision 1245181) +++ lucene/core/src/test/org/apache/lucene/index/TestDuelingCodecs.java (working copy) @@ -34,7 +34,6 @@ import org.apache.lucene.util.FixedBitSet; import org.apache.lucene.util.LineFileDocs; import org.apache.lucene.util.LuceneTestCase; -import org.apache.lucene.util.ReaderUtil; import org.apache.lucene.util.automaton.AutomatonTestUtil; import org.apache.lucene.util.automaton.CompiledAutomaton; import org.apache.lucene.util.automaton.RegExp; @@ -154,10 +153,10 @@ } /** - * Fields api equivalency + * InvertedFields api equivalency */ - public void assertFields(Fields leftFields, Fields rightFields, boolean deep) throws Exception { - // Fields could be null if there are no postings, + public void assertFields(InvertedFields leftFields, InvertedFields rightFields, boolean deep) throws Exception { + // InvertedFields could be null if there are no postings, // but then it must be null for both if (leftFields == null || rightFields == null) { assertNull(info, leftFields); @@ -178,9 +177,9 @@ } /** - * checks that top-level statistics on Fields are the same + * checks that top-level statistics on InvertedFields are the same */ - public void assertFieldStatistics(Fields leftFields, Fields rightFields) throws Exception { + public void assertFieldStatistics(InvertedFields leftFields, InvertedFields rightFields) throws Exception { if (leftFields.getUniqueFieldCount() != -1 && rightFields.getUniqueFieldCount() != -1) { assertEquals(info, leftFields.getUniqueFieldCount(), rightFields.getUniqueFieldCount()); } @@ -445,9 +444,9 @@ * checks that norms are the same across all fields */ public void assertNorms(IndexReader leftReader, IndexReader rightReader) throws Exception { - Fields leftFields = MultiFields.getFields(leftReader); - Fields rightFields = MultiFields.getFields(rightReader); - // Fields could be null if there are no postings, + InvertedFields leftFields = MultiFields.getFields(leftReader); + InvertedFields rightFields = MultiFields.getFields(rightReader); + // InvertedFields could be null if there are no postings, // but then it must be null for both if (leftFields == null || rightFields == null) { assertNull(info, leftFields); @@ -509,8 +508,8 @@ public void assertTermVectors(IndexReader leftReader, IndexReader rightReader) throws Exception { assert leftReader.maxDoc() == rightReader.maxDoc(); for (int i = 0; i < leftReader.maxDoc(); i++) { - Fields leftFields = leftReader.getTermVectors(i); - Fields rightFields = rightReader.getTermVectors(i); + InvertedFields leftFields = leftReader.getTermVectors(i); + InvertedFields rightFields = rightReader.getTermVectors(i); assertFields(leftFields, rightFields, rarely()); } } Index: lucene/core/src/test/org/apache/lucene/index/TestSumDocFreq.java =================================================================== --- lucene/core/src/test/org/apache/lucene/index/TestSumDocFreq.java (revision 1245181) +++ lucene/core/src/test/org/apache/lucene/index/TestSumDocFreq.java (working copy) @@ -75,7 +75,7 @@ private void assertSumDocFreq(IndexReader ir) throws Exception { // compute sumDocFreq across all fields - Fields fields = MultiFields.getFields(ir); + InvertedFields fields = MultiFields.getFields(ir); FieldsEnum fieldEnum = fields.iterator(); String f = null; while ((f = fieldEnum.next()) != null) { Index: lucene/core/src/test/org/apache/lucene/index/TestPerSegmentDeletes.java =================================================================== --- lucene/core/src/test/org/apache/lucene/index/TestPerSegmentDeletes.java (revision 1245181) +++ lucene/core/src/test/org/apache/lucene/index/TestPerSegmentDeletes.java (working copy) @@ -221,7 +221,7 @@ public static int[] toDocsArray(Term term, Bits bits, IndexReader reader) throws IOException { - Fields fields = MultiFields.getFields(reader); + InvertedFields fields = MultiFields.getFields(reader); Terms cterms = fields.terms(term.field); TermsEnum ctermsEnum = cterms.iterator(null); if (ctermsEnum.seekExact(new BytesRef(term.text()), false)) { Index: lucene/core/src/test/org/apache/lucene/index/TestDocCount.java =================================================================== --- lucene/core/src/test/org/apache/lucene/index/TestDocCount.java (revision 1245181) +++ lucene/core/src/test/org/apache/lucene/index/TestDocCount.java (working copy) @@ -60,7 +60,7 @@ } private void verifyCount(IndexReader ir) throws Exception { - Fields fields = MultiFields.getFields(ir); + InvertedFields fields = MultiFields.getFields(ir); if (fields == null) { return; } Index: lucene/core/src/test/org/apache/lucene/index/TestStressIndexing2.java =================================================================== --- lucene/core/src/test/org/apache/lucene/index/TestStressIndexing2.java (revision 1245181) +++ lucene/core/src/test/org/apache/lucene/index/TestStressIndexing2.java (working copy) @@ -312,7 +312,7 @@ int[] r2r1 = new int[r2.maxDoc()]; // r2 id to r1 id mapping // create mapping from id2 space to id2 based on idField - final Fields f1 = MultiFields.getFields(r1); + final InvertedFields f1 = MultiFields.getFields(r1); if (f1 == null) { // make sure r2 is empty assertNull(MultiFields.getFields(r2)); @@ -329,7 +329,7 @@ final Bits liveDocs1 = MultiFields.getLiveDocs(r1); final Bits liveDocs2 = MultiFields.getLiveDocs(r2); - Fields fields = MultiFields.getFields(r2); + InvertedFields fields = MultiFields.getFields(r2); if (fields == null) { // make sure r1 is in fact empty (eg has only all // deleted docs): @@ -392,7 +392,7 @@ verifyEquals(r1.getTermVectors(id1), r2.getTermVectors(id2)); } catch (Throwable e) { System.out.println("FAILED id=" + term + " id1=" + id1 + " id2=" + id2); - Fields tv1 = r1.getTermVectors(id1); + InvertedFields tv1 = r1.getTermVectors(id1); System.out.println(" d1=" + tv1); if (tv1 != null) { FieldsEnum fieldsEnum = tv1.iterator(); @@ -426,7 +426,7 @@ } } - Fields tv2 = r2.getTermVectors(id2); + InvertedFields tv2 = r2.getTermVectors(id2); System.out.println(" d2=" + tv2); if (tv2 != null) { FieldsEnum fieldsEnum = tv2.iterator(); @@ -597,7 +597,7 @@ } } - public static void verifyEquals(Fields d1, Fields d2) throws IOException { + public static void verifyEquals(InvertedFields d1, InvertedFields d2) throws IOException { if (d1 == null) { assertTrue(d2 == null || d2.getUniqueFieldCount() == 0); return; Index: lucene/core/src/test/org/apache/lucene/index/TestCodecs.java =================================================================== --- lucene/core/src/test/org/apache/lucene/index/TestCodecs.java (revision 1245181) +++ lucene/core/src/test/org/apache/lucene/index/TestCodecs.java (working copy) @@ -23,8 +23,8 @@ import org.apache.lucene.analysis.MockAnalyzer; import org.apache.lucene.codecs.Codec; -import org.apache.lucene.codecs.FieldsConsumer; -import org.apache.lucene.codecs.FieldsProducer; +import org.apache.lucene.codecs.InvertedFieldsConsumer; +import org.apache.lucene.codecs.InvertedFieldsProducer; import org.apache.lucene.codecs.PostingsConsumer; import org.apache.lucene.codecs.TermStats; import org.apache.lucene.codecs.TermsConsumer; @@ -106,7 +106,7 @@ return fieldInfo.name.compareTo(((FieldData) other).fieldInfo.name); } - public void write(final FieldsConsumer consumer) throws Throwable { + public void write(final InvertedFieldsConsumer consumer) throws Throwable { Arrays.sort(terms); final TermsConsumer termsConsumer = consumer.addField(fieldInfo); long sumTotalTermCount = 0; @@ -260,7 +260,7 @@ Codec codec = Codec.getDefault(); final SegmentInfo si = new SegmentInfo(SEGMENT, 10000, dir, false, codec, clonedFieldInfos); - final FieldsProducer reader = codec.postingsFormat().fieldsProducer(new SegmentReadState(dir, si, fieldInfos, newIOContext(random), DirectoryReader.DEFAULT_TERMS_INDEX_DIVISOR)); + final InvertedFieldsProducer reader = codec.postingsFormat().fieldsProducer(new SegmentReadState(dir, si, fieldInfos, newIOContext(random), DirectoryReader.DEFAULT_TERMS_INDEX_DIVISOR)); final FieldsEnum fieldsEnum = reader.iterator(); assertNotNull(fieldsEnum.next()); @@ -319,7 +319,7 @@ if (VERBOSE) { System.out.println("TEST: now read postings"); } - final FieldsProducer terms = codec.postingsFormat().fieldsProducer(new SegmentReadState(dir, si, fieldInfos, newIOContext(random), DirectoryReader.DEFAULT_TERMS_INDEX_DIVISOR)); + final InvertedFieldsProducer terms = codec.postingsFormat().fieldsProducer(new SegmentReadState(dir, si, fieldInfos, newIOContext(random), DirectoryReader.DEFAULT_TERMS_INDEX_DIVISOR)); final Verify[] threads = new Verify[NUM_TEST_THREADS-1]; for(int i=0;i 0); } - Fields results = reader.getTermVectors(0); + InvertedFields results = reader.getTermVectors(0); assertTrue(results != null); assertEquals("We do not have 3 term freq vectors", 3, results.getUniqueFieldCount()); } Index: lucene/core/src/test/org/apache/lucene/document/TestDocument.java =================================================================== --- lucene/core/src/test/org/apache/lucene/document/TestDocument.java (revision 1245181) +++ lucene/core/src/test/org/apache/lucene/document/TestDocument.java (working copy) @@ -20,9 +20,8 @@ import java.io.StringReader; import org.apache.lucene.analysis.EmptyTokenizer; -import org.apache.lucene.analysis.Tokenizer; import org.apache.lucene.index.DocsAndPositionsEnum; -import org.apache.lucene.index.Fields; +import org.apache.lucene.index.InvertedFields; import org.apache.lucene.index.IndexReader; import org.apache.lucene.index.IndexableField; import org.apache.lucene.index.RandomIndexWriter; @@ -338,7 +337,7 @@ assertEquals(1, s.search(new TermQuery(new Term("tokenized_tokenstream", "xyz")), 1).totalHits); for(String field : new String[] {"tv", "tv_pos", "tv_off", "tv_pos_off"}) { - Fields tvFields = r.getTermVectors(0); + InvertedFields tvFields = r.getTermVectors(0); Terms tvs = tvFields.terms(field); assertNotNull(tvs); assertEquals(2, tvs.getUniqueTermCount()); Index: lucene/core/src/java/org/apache/lucene/codecs/lucene40/Lucene40TermVectorsReader.java =================================================================== --- lucene/core/src/java/org/apache/lucene/codecs/lucene40/Lucene40TermVectorsReader.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/codecs/lucene40/Lucene40TermVectorsReader.java (working copy) @@ -30,7 +30,7 @@ import org.apache.lucene.index.DocsEnum; import org.apache.lucene.index.FieldInfo; import org.apache.lucene.index.FieldInfos; -import org.apache.lucene.index.Fields; +import org.apache.lucene.index.InvertedFields; import org.apache.lucene.index.FieldsEnum; import org.apache.lucene.index.IndexFileNames; import org.apache.lucene.index.IndexFormatTooNewException; @@ -225,7 +225,7 @@ return size; } - private class TVFields extends Fields { + private class TVFields extends InvertedFields { private final int[] fieldNumbers; private final long[] fieldFPs; private final Map fieldNumberToIndex = new HashMap(); @@ -663,12 +663,12 @@ } @Override - public Fields get(int docID) throws IOException { + public InvertedFields get(int docID) throws IOException { if (docID < 0 || docID >= numTotalDocs) { throw new IllegalArgumentException("doID=" + docID + " is out of bounds [0.." + (numTotalDocs-1) + "]"); } if (tvx != null) { - Fields fields = new TVFields(docID); + InvertedFields fields = new TVFields(docID); if (fields.getUniqueFieldCount() == 0) { // TODO: we can improve writer here, eg write 0 into // tvx file, so we know on first read from tvx that Index: lucene/core/src/java/org/apache/lucene/codecs/lucene40/Lucene40TermVectorsWriter.java =================================================================== --- lucene/core/src/java/org/apache/lucene/codecs/lucene40/Lucene40TermVectorsWriter.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/codecs/lucene40/Lucene40TermVectorsWriter.java (working copy) @@ -23,7 +23,7 @@ import org.apache.lucene.codecs.TermVectorsReader; import org.apache.lucene.codecs.TermVectorsWriter; import org.apache.lucene.index.FieldInfo; -import org.apache.lucene.index.Fields; +import org.apache.lucene.index.InvertedFields; import org.apache.lucene.index.IndexFileNames; import org.apache.lucene.index.MergePolicy.MergeAbortedException; import org.apache.lucene.index.MergeState; @@ -309,7 +309,7 @@ // NOTE: it's very important to first assign to vectors then pass it to // termVectorsWriter.addAllDocVectors; see LUCENE-1282 - Fields vectors = reader.reader.getTermVectors(docNum); + InvertedFields vectors = reader.reader.getTermVectors(docNum); addAllDocVectors(vectors, mergeState.fieldInfos); totalNumDocs++; mergeState.checkAbort.work(300); @@ -339,7 +339,7 @@ for (int docNum = 0; docNum < maxDoc; docNum++) { // NOTE: it's very important to first assign to vectors then pass it to // termVectorsWriter.addAllDocVectors; see LUCENE-1282 - Fields vectors = reader.reader.getTermVectors(docNum); + InvertedFields vectors = reader.reader.getTermVectors(docNum); addAllDocVectors(vectors, mergeState.fieldInfos); mergeState.checkAbort.work(300); } Index: lucene/core/src/java/org/apache/lucene/codecs/lucene40/Lucene40StoredFieldsReader.java =================================================================== --- lucene/core/src/java/org/apache/lucene/codecs/lucene40/Lucene40StoredFieldsReader.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/codecs/lucene40/Lucene40StoredFieldsReader.java (working copy) @@ -118,7 +118,7 @@ /** * Closes the underlying {@link org.apache.lucene.store.IndexInput} streams. - * This means that the Fields values will not be accessible. + * This means that the InvertedFields values will not be accessible. * * @throws IOException */ Index: lucene/core/src/java/org/apache/lucene/codecs/lucene40/Lucene40PostingsFormat.java =================================================================== --- lucene/core/src/java/org/apache/lucene/codecs/lucene40/Lucene40PostingsFormat.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/codecs/lucene40/Lucene40PostingsFormat.java (working copy) @@ -20,13 +20,8 @@ import java.io.IOException; import java.util.Set; -import org.apache.lucene.codecs.BlockTreeTermsReader; -import org.apache.lucene.codecs.BlockTreeTermsWriter; -import org.apache.lucene.codecs.FieldsConsumer; -import org.apache.lucene.codecs.FieldsProducer; -import org.apache.lucene.codecs.PostingsFormat; -import org.apache.lucene.codecs.PostingsReaderBase; -import org.apache.lucene.codecs.PostingsWriterBase; +import org.apache.lucene.codecs.*; +import org.apache.lucene.codecs.InvertedFieldsConsumer; import org.apache.lucene.index.SegmentInfo; import org.apache.lucene.index.SegmentReadState; import org.apache.lucene.index.SegmentWriteState; @@ -54,7 +49,7 @@ } @Override - public FieldsConsumer fieldsConsumer(SegmentWriteState state) throws IOException { + public InvertedFieldsConsumer fieldsConsumer(SegmentWriteState state) throws IOException { PostingsWriterBase docs = new Lucene40PostingsWriter(state); // TODO: should we make the terms index more easily @@ -63,7 +58,7 @@ // Or... you must make a new Codec for this? boolean success = false; try { - FieldsConsumer ret = new BlockTreeTermsWriter(state, docs, minBlockSize, maxBlockSize); + InvertedFieldsConsumer ret = new BlockTreeTermsWriter(state, docs, minBlockSize, maxBlockSize); success = true; return ret; } finally { @@ -76,12 +71,12 @@ public final static int TERMS_CACHE_SIZE = 1024; @Override - public FieldsProducer fieldsProducer(SegmentReadState state) throws IOException { + public InvertedFieldsProducer fieldsProducer(SegmentReadState state) throws IOException { PostingsReaderBase postings = new Lucene40PostingsReader(state.dir, state.segmentInfo, state.context, state.segmentSuffix); boolean success = false; try { - FieldsProducer ret = new BlockTreeTermsReader( + InvertedFieldsProducer ret = new BlockTreeTermsReader( state.dir, state.fieldInfos, state.segmentInfo.name, Index: lucene/core/src/java/org/apache/lucene/codecs/memory/MemoryPostingsFormat.java =================================================================== --- lucene/core/src/java/org/apache/lucene/codecs/memory/MemoryPostingsFormat.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/codecs/memory/MemoryPostingsFormat.java (working copy) @@ -24,12 +24,8 @@ import java.util.SortedMap; import java.util.TreeMap; -import org.apache.lucene.codecs.FieldsConsumer; -import org.apache.lucene.codecs.FieldsProducer; -import org.apache.lucene.codecs.PostingsConsumer; -import org.apache.lucene.codecs.PostingsFormat; -import org.apache.lucene.codecs.TermStats; -import org.apache.lucene.codecs.TermsConsumer; +import org.apache.lucene.codecs.*; +import org.apache.lucene.codecs.InvertedFieldsConsumer; import org.apache.lucene.index.DocsAndPositionsEnum; import org.apache.lucene.index.DocsEnum; import org.apache.lucene.index.FieldInfo.IndexOptions; @@ -262,12 +258,12 @@ private static String EXTENSION = "ram"; @Override - public FieldsConsumer fieldsConsumer(SegmentWriteState state) throws IOException { + public InvertedFieldsConsumer fieldsConsumer(SegmentWriteState state) throws IOException { final String fileName = IndexFileNames.segmentFileName(state.segmentName, state.segmentSuffix, EXTENSION); final IndexOutput out = state.directory.createOutput(fileName, state.context); - return new FieldsConsumer() { + return new InvertedFieldsConsumer() { @Override public TermsConsumer addField(FieldInfo field) { if (field.indexOptions.compareTo(IndexOptions.DOCS_AND_FREQS_AND_POSITIONS_AND_OFFSETS) >= 0) { @@ -781,7 +777,7 @@ } @Override - public FieldsProducer fieldsProducer(SegmentReadState state) throws IOException { + public InvertedFieldsProducer fieldsProducer(SegmentReadState state) throws IOException { final String fileName = IndexFileNames.segmentFileName(state.segmentInfo.name, state.segmentSuffix, EXTENSION); final IndexInput in = state.dir.openInput(fileName, IOContext.READONCE); @@ -803,7 +799,7 @@ in.close(); } - return new FieldsProducer() { + return new InvertedFieldsProducer() { @Override public FieldsEnum iterator() { final Iterator iter = fields.values().iterator(); Index: lucene/core/src/java/org/apache/lucene/codecs/lucene3x/Lucene3xFields.java =================================================================== --- lucene/core/src/java/org/apache/lucene/codecs/lucene3x/Lucene3xFields.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/codecs/lucene3x/Lucene3xFields.java (working copy) @@ -25,7 +25,7 @@ import java.util.Map; import java.util.TreeMap; -import org.apache.lucene.codecs.FieldsProducer; +import org.apache.lucene.codecs.InvertedFieldsProducer; import org.apache.lucene.index.DocsAndPositionsEnum; import org.apache.lucene.index.DocsEnum; import org.apache.lucene.index.FieldInfo.IndexOptions; @@ -49,7 +49,7 @@ * @deprecated (4.0) */ @Deprecated -class Lucene3xFields extends FieldsProducer { +class Lucene3xFields extends InvertedFieldsProducer { private static final boolean DEBUG_SURROGATES = false; Index: lucene/core/src/java/org/apache/lucene/codecs/lucene3x/Lucene3xTermVectorsReader.java =================================================================== --- lucene/core/src/java/org/apache/lucene/codecs/lucene3x/Lucene3xTermVectorsReader.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/codecs/lucene3x/Lucene3xTermVectorsReader.java (working copy) @@ -25,19 +25,8 @@ import java.util.Set; import org.apache.lucene.codecs.TermVectorsReader; -import org.apache.lucene.index.CorruptIndexException; -import org.apache.lucene.index.DocsAndPositionsEnum; -import org.apache.lucene.index.DocsEnum; -import org.apache.lucene.index.FieldInfo; -import org.apache.lucene.index.FieldInfos; -import org.apache.lucene.index.Fields; -import org.apache.lucene.index.FieldsEnum; -import org.apache.lucene.index.IndexFileNames; -import org.apache.lucene.index.IndexFormatTooNewException; -import org.apache.lucene.index.IndexFormatTooOldException; -import org.apache.lucene.index.SegmentInfo; -import org.apache.lucene.index.Terms; -import org.apache.lucene.index.TermsEnum; +import org.apache.lucene.index.*; +import org.apache.lucene.index.InvertedFields; import org.apache.lucene.store.CompoundFileDirectory; import org.apache.lucene.store.Directory; import org.apache.lucene.store.IOContext; @@ -195,7 +184,7 @@ return size; } - private class TVFields extends Fields { + private class TVFields extends InvertedFields { private final int[] fieldNumbers; private final long[] fieldFPs; private final Map fieldNumberToIndex = new HashMap(); @@ -654,12 +643,12 @@ } @Override - public Fields get(int docID) throws IOException { + public InvertedFields get(int docID) throws IOException { if (docID < 0 || docID >= numTotalDocs) { throw new IllegalArgumentException("doID=" + docID + " is out of bounds [0.." + (numTotalDocs-1) + "]"); } if (tvx != null) { - Fields fields = new TVFields(docID); + InvertedFields fields = new TVFields(docID); if (fields.getUniqueFieldCount() == 0) { // TODO: we can improve writer here, eg write 0 into // tvx file, so we know on first read from tvx that Index: lucene/core/src/java/org/apache/lucene/codecs/lucene3x/Lucene3xPostingsFormat.java =================================================================== --- lucene/core/src/java/org/apache/lucene/codecs/lucene3x/Lucene3xPostingsFormat.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/codecs/lucene3x/Lucene3xPostingsFormat.java (working copy) @@ -20,8 +20,8 @@ import java.util.Set; import java.io.IOException; -import org.apache.lucene.codecs.FieldsConsumer; -import org.apache.lucene.codecs.FieldsProducer; +import org.apache.lucene.codecs.InvertedFieldsConsumer; +import org.apache.lucene.codecs.InvertedFieldsProducer; import org.apache.lucene.codecs.PostingsFormat; import org.apache.lucene.index.SegmentInfo; import org.apache.lucene.index.SegmentWriteState; @@ -55,12 +55,12 @@ } @Override - public FieldsConsumer fieldsConsumer(SegmentWriteState state) throws IOException { + public InvertedFieldsConsumer fieldsConsumer(SegmentWriteState state) throws IOException { throw new UnsupportedOperationException("this codec can only be used for reading"); } @Override - public FieldsProducer fieldsProducer(SegmentReadState state) throws IOException { + public InvertedFieldsProducer fieldsProducer(SegmentReadState state) throws IOException { return new Lucene3xFields(state.dir, state.fieldInfos, state.segmentInfo, state.context, state.termsIndexDivisor); } Index: lucene/core/src/java/org/apache/lucene/codecs/BlockTermsReader.java =================================================================== --- lucene/core/src/java/org/apache/lucene/codecs/BlockTermsReader.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/codecs/BlockTermsReader.java (working copy) @@ -56,7 +56,7 @@ * implementation of the terms dict index. * @lucene.experimental */ -public class BlockTermsReader extends FieldsProducer { +public class BlockTermsReader extends InvertedFieldsProducer { // Open input to the main terms dict file (_X.tis) private final IndexInput in; Index: lucene/core/src/java/org/apache/lucene/codecs/pulsing/PulsingPostingsFormat.java =================================================================== --- lucene/core/src/java/org/apache/lucene/codecs/pulsing/PulsingPostingsFormat.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/codecs/pulsing/PulsingPostingsFormat.java (working copy) @@ -20,14 +20,8 @@ import java.io.IOException; import java.util.Set; -import org.apache.lucene.codecs.BlockTreeTermsReader; -import org.apache.lucene.codecs.BlockTreeTermsWriter; -import org.apache.lucene.codecs.FieldsConsumer; -import org.apache.lucene.codecs.FieldsProducer; -import org.apache.lucene.codecs.PostingsBaseFormat; -import org.apache.lucene.codecs.PostingsFormat; -import org.apache.lucene.codecs.PostingsReaderBase; -import org.apache.lucene.codecs.PostingsWriterBase; +import org.apache.lucene.codecs.*; +import org.apache.lucene.codecs.InvertedFieldsConsumer; import org.apache.lucene.index.SegmentInfo; import org.apache.lucene.index.SegmentReadState; import org.apache.lucene.index.SegmentWriteState; @@ -66,7 +60,7 @@ } @Override - public FieldsConsumer fieldsConsumer(SegmentWriteState state) throws IOException { + public InvertedFieldsConsumer fieldsConsumer(SegmentWriteState state) throws IOException { PostingsWriterBase docsWriter = wrappedPostingsBaseFormat.postingsWriterBase(state); // Terms that have <= freqCutoff number of docs are @@ -76,7 +70,7 @@ // Terms dict boolean success = false; try { - FieldsConsumer ret = new BlockTreeTermsWriter(state, pulsingWriter, minBlockSize, maxBlockSize); + InvertedFieldsConsumer ret = new BlockTreeTermsWriter(state, pulsingWriter, minBlockSize, maxBlockSize); success = true; return ret; } finally { @@ -87,14 +81,14 @@ } @Override - public FieldsProducer fieldsProducer(SegmentReadState state) throws IOException { + public InvertedFieldsProducer fieldsProducer(SegmentReadState state) throws IOException { PostingsReaderBase docsReader = wrappedPostingsBaseFormat.postingsReaderBase(state); PostingsReaderBase pulsingReader = new PulsingPostingsReader(docsReader); boolean success = false; try { - FieldsProducer ret = new BlockTreeTermsReader( + InvertedFieldsProducer ret = new BlockTreeTermsReader( state.dir, state.fieldInfos, state.segmentInfo.name, pulsingReader, state.context, Index: lucene/core/src/java/org/apache/lucene/codecs/BlockTermsWriter.java =================================================================== --- lucene/core/src/java/org/apache/lucene/codecs/BlockTermsWriter.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/codecs/BlockTermsWriter.java (working copy) @@ -47,7 +47,7 @@ * @lucene.experimental */ -public class BlockTermsWriter extends FieldsConsumer { +public class BlockTermsWriter extends InvertedFieldsConsumer { final static String CODEC_NAME = "BLOCK_TERMS_DICT"; Index: lucene/core/src/java/org/apache/lucene/codecs/simpletext/SimpleTextFieldsReader.java =================================================================== --- lucene/core/src/java/org/apache/lucene/codecs/simpletext/SimpleTextFieldsReader.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/codecs/simpletext/SimpleTextFieldsReader.java (working copy) @@ -22,7 +22,7 @@ import java.util.HashMap; import java.util.Map; -import org.apache.lucene.codecs.FieldsProducer; +import org.apache.lucene.codecs.InvertedFieldsProducer; import org.apache.lucene.index.DocsAndPositionsEnum; import org.apache.lucene.index.DocsEnum; import org.apache.lucene.index.FieldInfo.IndexOptions; @@ -47,7 +47,7 @@ import org.apache.lucene.util.fst.PositiveIntOutputs; import org.apache.lucene.util.fst.Util; -class SimpleTextFieldsReader extends FieldsProducer { +class SimpleTextFieldsReader extends InvertedFieldsProducer { private final IndexInput in; private final FieldInfos fieldInfos; Index: lucene/core/src/java/org/apache/lucene/codecs/simpletext/SimpleTextFieldsWriter.java =================================================================== --- lucene/core/src/java/org/apache/lucene/codecs/simpletext/SimpleTextFieldsWriter.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/codecs/simpletext/SimpleTextFieldsWriter.java (working copy) @@ -18,7 +18,7 @@ */ import org.apache.lucene.util.BytesRef; -import org.apache.lucene.codecs.FieldsConsumer; +import org.apache.lucene.codecs.InvertedFieldsConsumer; import org.apache.lucene.codecs.PostingsConsumer; import org.apache.lucene.codecs.TermStats; import org.apache.lucene.codecs.TermsConsumer; @@ -30,7 +30,7 @@ import java.io.IOException; import java.util.Comparator; -class SimpleTextFieldsWriter extends FieldsConsumer { +class SimpleTextFieldsWriter extends InvertedFieldsConsumer { private final IndexOutput out; private final BytesRef scratch = new BytesRef(10); Index: lucene/core/src/java/org/apache/lucene/codecs/simpletext/SimpleTextTermVectorsReader.java =================================================================== --- lucene/core/src/java/org/apache/lucene/codecs/simpletext/SimpleTextTermVectorsReader.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/codecs/simpletext/SimpleTextTermVectorsReader.java (working copy) @@ -29,7 +29,7 @@ import org.apache.lucene.codecs.TermVectorsReader; import org.apache.lucene.index.DocsAndPositionsEnum; import org.apache.lucene.index.DocsEnum; -import org.apache.lucene.index.Fields; +import org.apache.lucene.index.InvertedFields; import org.apache.lucene.index.FieldsEnum; import org.apache.lucene.index.IndexFileNames; import org.apache.lucene.index.SegmentInfo; @@ -94,7 +94,7 @@ } @Override - public Fields get(int doc) throws IOException { + public InvertedFields get(int doc) throws IOException { // TestTV tests for this in testBadParams... but is this // really guaranteed by the API? if (doc < 0 || doc >= offsets.size()) { @@ -221,7 +221,7 @@ return scratchUTF16.toString(); } - private class SimpleTVFields extends Fields { + private class SimpleTVFields extends InvertedFields { private final SortedMap fields; SimpleTVFields(SortedMap fields) throws IOException { Index: lucene/core/src/java/org/apache/lucene/codecs/simpletext/SimpleTextPostingsFormat.java =================================================================== --- lucene/core/src/java/org/apache/lucene/codecs/simpletext/SimpleTextPostingsFormat.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/codecs/simpletext/SimpleTextPostingsFormat.java (working copy) @@ -20,8 +20,8 @@ import java.io.IOException; import java.util.Set; -import org.apache.lucene.codecs.FieldsConsumer; -import org.apache.lucene.codecs.FieldsProducer; +import org.apache.lucene.codecs.InvertedFieldsConsumer; +import org.apache.lucene.codecs.InvertedFieldsProducer; import org.apache.lucene.codecs.PostingsFormat; import org.apache.lucene.index.SegmentInfo; import org.apache.lucene.index.SegmentWriteState; @@ -43,12 +43,12 @@ } @Override - public FieldsConsumer fieldsConsumer(SegmentWriteState state) throws IOException { + public InvertedFieldsConsumer fieldsConsumer(SegmentWriteState state) throws IOException { return new SimpleTextFieldsWriter(state); } @Override - public FieldsProducer fieldsProducer(SegmentReadState state) throws IOException { + public InvertedFieldsProducer fieldsProducer(SegmentReadState state) throws IOException { return new SimpleTextFieldsReader(state); } Index: lucene/core/src/java/org/apache/lucene/codecs/PostingsFormat.java =================================================================== --- lucene/core/src/java/org/apache/lucene/codecs/PostingsFormat.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/codecs/PostingsFormat.java (working copy) @@ -49,12 +49,12 @@ } /** Writes a new segment */ - public abstract FieldsConsumer fieldsConsumer(SegmentWriteState state) throws IOException; + public abstract InvertedFieldsConsumer fieldsConsumer(SegmentWriteState state) throws IOException; /** Reads a segment. NOTE: by the time this call * returns, it must hold open any files it will need to * use; else, those files may be deleted. */ - public abstract FieldsProducer fieldsProducer(SegmentReadState state) throws IOException; + public abstract InvertedFieldsProducer fieldsProducer(SegmentReadState state) throws IOException; /** * Gathers files associated with this segment Index: lucene/core/src/java/org/apache/lucene/codecs/appending/AppendingPostingsFormat.java =================================================================== --- lucene/core/src/java/org/apache/lucene/codecs/appending/AppendingPostingsFormat.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/codecs/appending/AppendingPostingsFormat.java (working copy) @@ -20,19 +20,13 @@ import java.io.IOException; import java.util.Set; -import org.apache.lucene.codecs.BlockTreeTermsReader; -import org.apache.lucene.codecs.BlockTreeTermsWriter; -import org.apache.lucene.codecs.FieldsConsumer; -import org.apache.lucene.codecs.FieldsProducer; -import org.apache.lucene.codecs.PostingsFormat; -import org.apache.lucene.codecs.PostingsReaderBase; -import org.apache.lucene.codecs.PostingsWriterBase; +import org.apache.lucene.codecs.*; +import org.apache.lucene.codecs.InvertedFieldsConsumer; import org.apache.lucene.codecs.lucene40.Lucene40PostingsReader; import org.apache.lucene.codecs.lucene40.Lucene40PostingsWriter; import org.apache.lucene.index.SegmentInfo; import org.apache.lucene.index.SegmentReadState; import org.apache.lucene.index.SegmentWriteState; -import org.apache.lucene.store.Directory; /** * Appending postings impl @@ -45,11 +39,11 @@ } @Override - public FieldsConsumer fieldsConsumer(SegmentWriteState state) throws IOException { + public InvertedFieldsConsumer fieldsConsumer(SegmentWriteState state) throws IOException { PostingsWriterBase docsWriter = new Lucene40PostingsWriter(state); boolean success = false; try { - FieldsConsumer ret = new AppendingTermsWriter(state, docsWriter, BlockTreeTermsWriter.DEFAULT_MIN_BLOCK_SIZE, BlockTreeTermsWriter.DEFAULT_MAX_BLOCK_SIZE); + InvertedFieldsConsumer ret = new AppendingTermsWriter(state, docsWriter, BlockTreeTermsWriter.DEFAULT_MIN_BLOCK_SIZE, BlockTreeTermsWriter.DEFAULT_MAX_BLOCK_SIZE); success = true; return ret; } finally { @@ -60,12 +54,12 @@ } @Override - public FieldsProducer fieldsProducer(SegmentReadState state) throws IOException { + public InvertedFieldsProducer fieldsProducer(SegmentReadState state) throws IOException { PostingsReaderBase postings = new Lucene40PostingsReader(state.dir, state.segmentInfo, state.context, state.segmentSuffix); boolean success = false; try { - FieldsProducer ret = new AppendingTermsReader( + InvertedFieldsProducer ret = new AppendingTermsReader( state.dir, state.fieldInfos, state.segmentInfo.name, Index: lucene/core/src/java/org/apache/lucene/codecs/TermVectorsReader.java =================================================================== --- lucene/core/src/java/org/apache/lucene/codecs/TermVectorsReader.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/codecs/TermVectorsReader.java (working copy) @@ -22,7 +22,7 @@ import org.apache.lucene.analysis.tokenattributes.OffsetAttribute; // javadocs import org.apache.lucene.index.DocsAndPositionsEnum; // javadocs -import org.apache.lucene.index.Fields; +import org.apache.lucene.index.InvertedFields; /** * Codec API for reading term vectors: @@ -35,7 +35,7 @@ * term vectors were not indexed. If offsets are * available they are in an {@link OffsetAttribute} * available from the {@link DocsAndPositionsEnum}. */ - public abstract Fields get(int doc) throws IOException; + public abstract InvertedFields get(int doc) throws IOException; /** Create a clone that one caller at a time may use to * read term vectors. */ Index: lucene/core/src/java/org/apache/lucene/codecs/BlockTreeTermsReader.java =================================================================== --- lucene/core/src/java/org/apache/lucene/codecs/BlockTreeTermsReader.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/codecs/BlockTreeTermsReader.java (working copy) @@ -35,7 +35,6 @@ import org.apache.lucene.index.SegmentInfo; import org.apache.lucene.index.TermState; import org.apache.lucene.index.Terms; -import org.apache.lucene.index.TermsEnum.SeekStatus; import org.apache.lucene.index.TermsEnum; import org.apache.lucene.store.ByteArrayDataInput; import org.apache.lucene.store.Directory; @@ -86,7 +85,7 @@ * @lucene.experimental */ -public class BlockTreeTermsReader extends FieldsProducer { +public class BlockTreeTermsReader extends InvertedFieldsProducer { // Open input to the main terms dict file (_X.tib) private final IndexInput in; Index: lucene/core/src/java/org/apache/lucene/codecs/TermVectorsWriter.java =================================================================== --- lucene/core/src/java/org/apache/lucene/codecs/TermVectorsWriter.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/codecs/TermVectorsWriter.java (working copy) @@ -21,15 +21,8 @@ import java.io.IOException; import java.util.Comparator; -import org.apache.lucene.index.DocsAndPositionsEnum; -import org.apache.lucene.index.DocsEnum; -import org.apache.lucene.index.FieldInfo; -import org.apache.lucene.index.FieldInfos; -import org.apache.lucene.index.Fields; -import org.apache.lucene.index.FieldsEnum; -import org.apache.lucene.index.MergeState; -import org.apache.lucene.index.Terms; -import org.apache.lucene.index.TermsEnum; +import org.apache.lucene.index.*; +import org.apache.lucene.index.InvertedFields; import org.apache.lucene.store.DataInput; import org.apache.lucene.util.Bits; import org.apache.lucene.util.BytesRef; @@ -153,7 +146,7 @@ } // NOTE: it's very important to first assign to vectors then pass it to // termVectorsWriter.addAllDocVectors; see LUCENE-1282 - Fields vectors = reader.reader.getTermVectors(docID); + InvertedFields vectors = reader.reader.getTermVectors(docID); addAllDocVectors(vectors, mergeState.fieldInfos); docCount++; mergeState.checkAbort.work(300); @@ -166,9 +159,9 @@ /** Safe (but, slowish) default method to write every * vector field in the document. This default * implementation requires that the vectors implement - * both Fields.getUniqueFieldCount and + * both InvertedFields.getUniqueFieldCount and * Terms.getUniqueTermCount. */ - protected final void addAllDocVectors(Fields vectors, FieldInfos fieldInfos) throws IOException { + protected final void addAllDocVectors(InvertedFields vectors, FieldInfos fieldInfos) throws IOException { if (vectors == null) { startDocument(0); return; Index: lucene/core/src/java/org/apache/lucene/codecs/perfield/PerFieldPostingsFormat.java =================================================================== --- lucene/core/src/java/org/apache/lucene/codecs/perfield/PerFieldPostingsFormat.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/codecs/perfield/PerFieldPostingsFormat.java (working copy) @@ -27,8 +27,8 @@ import java.util.Set; import java.util.TreeMap; -import org.apache.lucene.codecs.FieldsConsumer; -import org.apache.lucene.codecs.FieldsProducer; +import org.apache.lucene.codecs.InvertedFieldsConsumer; +import org.apache.lucene.codecs.InvertedFieldsProducer; import org.apache.lucene.codecs.PostingsFormat; import org.apache.lucene.codecs.TermsConsumer; import org.apache.lucene.index.FieldInfo; @@ -64,17 +64,17 @@ } @Override - public FieldsConsumer fieldsConsumer(SegmentWriteState state) + public InvertedFieldsConsumer fieldsConsumer(SegmentWriteState state) throws IOException { return new FieldsWriter(state); } // NOTE: not private to avoid $accessN at runtime!! static class FieldsConsumerAndID implements Closeable { - final FieldsConsumer fieldsConsumer; + final InvertedFieldsConsumer fieldsConsumer; final String segmentSuffix; - public FieldsConsumerAndID(FieldsConsumer fieldsConsumer, String segmentSuffix) { + public FieldsConsumerAndID(InvertedFieldsConsumer fieldsConsumer, String segmentSuffix) { this.fieldsConsumer = fieldsConsumer; this.segmentSuffix = segmentSuffix; } @@ -85,7 +85,7 @@ } }; - private class FieldsWriter extends FieldsConsumer { + private class FieldsWriter extends InvertedFieldsConsumer { private final Map formats = new IdentityHashMap(); @@ -173,10 +173,10 @@ } } - private class FieldsReader extends FieldsProducer { + private class FieldsReader extends InvertedFieldsProducer { - private final Map fields = new TreeMap(); - private final Map formats = new IdentityHashMap(); + private final Map fields = new TreeMap(); + private final Map formats = new IdentityHashMap(); public FieldsReader(final SegmentReadState readState) throws IOException { @@ -235,7 +235,7 @@ @Override public Terms terms(String field) throws IOException { - FieldsProducer fieldsProducer = fields.get(field); + InvertedFieldsProducer fieldsProducer = fields.get(field); return fieldsProducer == null ? null : fieldsProducer.terms(field); } @@ -251,7 +251,7 @@ } @Override - public FieldsProducer fieldsProducer(SegmentReadState state) + public InvertedFieldsProducer fieldsProducer(SegmentReadState state) throws IOException { return new FieldsReader(state); } Index: lucene/core/src/java/org/apache/lucene/codecs/BlockTreeTermsWriter.java =================================================================== --- lucene/core/src/java/org/apache/lucene/codecs/BlockTreeTermsWriter.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/codecs/BlockTreeTermsWriter.java (working copy) @@ -83,7 +83,7 @@ * @lucene.experimental */ -public class BlockTreeTermsWriter extends FieldsConsumer { +public class BlockTreeTermsWriter extends InvertedFieldsConsumer { public final static int DEFAULT_MIN_BLOCK_SIZE = 25; public final static int DEFAULT_MAX_BLOCK_SIZE = 48; Index: lucene/core/src/java/org/apache/lucene/search/MultiTermQueryWrapperFilter.java =================================================================== --- lucene/core/src/java/org/apache/lucene/search/MultiTermQueryWrapperFilter.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/search/MultiTermQueryWrapperFilter.java (working copy) @@ -21,7 +21,7 @@ import org.apache.lucene.index.AtomicReaderContext; import org.apache.lucene.index.DocsEnum; -import org.apache.lucene.index.Fields; +import org.apache.lucene.index.InvertedFields; import org.apache.lucene.index.AtomicReader; import org.apache.lucene.index.Terms; import org.apache.lucene.index.TermsEnum; @@ -84,7 +84,7 @@ @Override public DocIdSet getDocIdSet(AtomicReaderContext context, Bits acceptDocs) throws IOException { final AtomicReader reader = context.reader(); - final Fields fields = reader.fields(); + final InvertedFields fields = reader.fields(); if (fields == null) { // reader has no fields return DocIdSet.EMPTY_DOCIDSET; Index: lucene/core/src/java/org/apache/lucene/search/spans/SpanTermQuery.java =================================================================== --- lucene/core/src/java/org/apache/lucene/search/spans/SpanTermQuery.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/search/spans/SpanTermQuery.java (working copy) @@ -17,13 +17,8 @@ * limitations under the License. */ -import org.apache.lucene.index.AtomicReaderContext; -import org.apache.lucene.index.Fields; -import org.apache.lucene.index.Term; -import org.apache.lucene.index.DocsAndPositionsEnum; -import org.apache.lucene.index.TermState; -import org.apache.lucene.index.Terms; -import org.apache.lucene.index.TermsEnum; +import org.apache.lucene.index.*; +import org.apache.lucene.index.InvertedFields; import org.apache.lucene.util.Bits; import org.apache.lucene.util.TermContext; import org.apache.lucene.util.ToStringUtils; @@ -93,7 +88,7 @@ if (termContext == null) { // this happens with span-not query, as it doesn't include the NOT side in extractTerms() // so we seek to the term now in this segment..., this sucks because its ugly mostly! - final Fields fields = context.reader().fields(); + final InvertedFields fields = context.reader().fields(); if (fields != null) { final Terms terms = fields.terms(term.field()); if (terms != null) { Index: lucene/core/src/java/org/apache/lucene/search/TermCollectingRewrite.java =================================================================== --- lucene/core/src/java/org/apache/lucene/search/TermCollectingRewrite.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/search/TermCollectingRewrite.java (working copy) @@ -20,13 +20,8 @@ import java.io.IOException; import java.util.Comparator; -import org.apache.lucene.index.AtomicReaderContext; -import org.apache.lucene.index.Fields; -import org.apache.lucene.index.IndexReader; -import org.apache.lucene.index.IndexReaderContext; -import org.apache.lucene.index.Term; -import org.apache.lucene.index.Terms; -import org.apache.lucene.index.TermsEnum; +import org.apache.lucene.index.*; +import org.apache.lucene.index.InvertedFields; import org.apache.lucene.util.AttributeSource; import org.apache.lucene.util.BytesRef; import org.apache.lucene.util.TermContext; @@ -50,7 +45,7 @@ Comparator lastTermComp = null; final AtomicReaderContext[] leaves = topReaderContext.leaves(); for (AtomicReaderContext context : leaves) { - final Fields fields = context.reader().fields(); + final InvertedFields fields = context.reader().fields(); if (fields == null) { // reader has no fields continue; Index: lucene/core/src/java/org/apache/lucene/index/DocTermOrds.java =================================================================== --- lucene/core/src/java/org/apache/lucene/index/DocTermOrds.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/index/DocTermOrds.java (working copy) @@ -200,7 +200,7 @@ } if (indexedTermsArray == null) { //System.out.println("GET normal enum"); - final Fields fields = reader.fields(); + final InvertedFields fields = reader.fields(); if (fields == null) { return null; } @@ -234,7 +234,7 @@ final int[] lastTerm = new int[maxDoc]; // last term we saw for this document final byte[][] bytes = new byte[maxDoc][]; // list of term numbers for the doc (delta encoded vInts) - final Fields fields = reader.fields(); + final InvertedFields fields = reader.fields(); if (fields == null) { // No terms return; Index: lucene/core/src/java/org/apache/lucene/index/SegmentReader.java =================================================================== --- lucene/core/src/java/org/apache/lucene/index/SegmentReader.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/index/SegmentReader.java (working copy) @@ -133,7 +133,7 @@ } @Override - public Fields fields() throws IOException { + public InvertedFields fields() throws IOException { ensureOpen(); return core.fields; } @@ -170,7 +170,7 @@ * @throws IOException */ @Override - public Fields getTermVectors(int docID) throws IOException { + public InvertedFields getTermVectors(int docID) throws IOException { TermVectorsReader termVectorsReader = getTermVectorsReader(); if (termVectorsReader == null) { return null; Index: lucene/core/src/java/org/apache/lucene/index/BufferedDeletesStream.java =================================================================== --- lucene/core/src/java/org/apache/lucene/index/BufferedDeletesStream.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/index/BufferedDeletesStream.java (working copy) @@ -355,7 +355,7 @@ // Delete by Term private synchronized long applyTermDeletes(Iterable termsIter, IndexWriter.ReadersAndLiveDocs rld, SegmentReader reader) throws IOException { long delCount = 0; - Fields fields = reader.fields(); + InvertedFields fields = reader.fields(); if (fields == null) { // This reader has no postings return 0; Index: lucene/core/src/java/org/apache/lucene/index/AtomicReader.java =================================================================== --- lucene/core/src/java/org/apache/lucene/index/AtomicReader.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/index/AtomicReader.java (working copy) @@ -19,11 +19,8 @@ import java.io.IOException; -import org.apache.lucene.search.SearcherManager; // javadocs -import org.apache.lucene.store.*; import org.apache.lucene.util.Bits; import org.apache.lucene.util.BytesRef; -import org.apache.lucene.util.ReaderUtil; // for javadocs /** {@code AtomicReader} is an abstract class, providing an interface for accessing an index. Search of an index is done entirely through this abstract interface, @@ -70,15 +67,15 @@ } /** - * Returns {@link Fields} for this reader. + * Returns {@link InvertedFields} for this reader. * This method may return null if the reader has no * postings. */ - public abstract Fields fields() throws IOException; + public abstract InvertedFields fields() throws IOException; @Override public final int docFreq(String field, BytesRef term) throws IOException { - final Fields fields = fields(); + final InvertedFields fields = fields(); if (fields == null) { return 0; } @@ -100,7 +97,7 @@ * account deleted documents that have not yet been merged * away. */ public final long totalTermFreq(String field, BytesRef term) throws IOException { - final Fields fields = fields(); + final InvertedFields fields = fields(); if (fields == null) { return 0; } @@ -118,7 +115,7 @@ /** This may return null if the field does not exist.*/ public final Terms terms(String field) throws IOException { - final Fields fields = fields(); + final InvertedFields fields = fields(); if (fields == null) { return null; } @@ -131,7 +128,7 @@ public final DocsEnum termDocsEnum(Bits liveDocs, String field, BytesRef term, boolean needsFreqs) throws IOException { assert field != null; assert term != null; - final Fields fields = fields(); + final InvertedFields fields = fields(); if (fields != null) { final Terms terms = fields.terms(field); if (terms != null) { @@ -151,7 +148,7 @@ public final DocsAndPositionsEnum termPositionsEnum(Bits liveDocs, String field, BytesRef term, boolean needsOffsets) throws IOException { assert field != null; assert term != null; - final Fields fields = fields(); + final InvertedFields fields = fields(); if (fields != null) { final Terms terms = fields.terms(field); if (terms != null) { @@ -172,7 +169,7 @@ public final DocsEnum termDocsEnum(Bits liveDocs, String field, BytesRef term, TermState state, boolean needsFreqs) throws IOException { assert state != null; assert field != null; - final Fields fields = fields(); + final InvertedFields fields = fields(); if (fields != null) { final Terms terms = fields.terms(field); if (terms != null) { @@ -193,7 +190,7 @@ public final DocsAndPositionsEnum termPositionsEnum(Bits liveDocs, String field, BytesRef term, TermState state, boolean needsOffsets) throws IOException { assert state != null; assert field != null; - final Fields fields = fields(); + final InvertedFields fields = fields(); if (fields != null) { final Terms terms = fields.terms(field); if (terms != null) { @@ -209,7 +206,7 @@ * in this reader. */ public final long getUniqueTermCount() throws IOException { - final Fields fields = fields(); + final InvertedFields fields = fields(); if (fields == null) { return 0; } Index: lucene/core/src/java/org/apache/lucene/index/CheckIndex.java =================================================================== --- lucene/core/src/java/org/apache/lucene/index/CheckIndex.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/index/CheckIndex.java (working copy) @@ -710,7 +710,7 @@ } int computedFieldCount = 0; - final Fields fields = reader.fields(); + final InvertedFields fields = reader.fields(); if (fields == null) { msg("OK [no fields/terms]"); return status; @@ -973,7 +973,7 @@ // make sure TermsEnum is empty: final Terms fieldTerms2 = fieldsEnum.terms(); if (fieldTerms2 != null && fieldTerms2.iterator(null).next() != null) { - throw new RuntimeException("Fields.terms(field=" + field + ") returned null yet the field appears to have terms"); + throw new RuntimeException("InvertedFields.terms(field=" + field + ") returned null yet the field appears to have terms"); } } else { if (fieldTerms instanceof BlockTreeTermsReader.FieldReader) { @@ -1292,7 +1292,7 @@ final Bits liveDocs = reader.getLiveDocs(); - final Fields postingsFields; + final InvertedFields postingsFields; // TODO: testTermsIndex if (crossCheckTermVectors) { postingsFields = reader.fields(); @@ -1306,7 +1306,7 @@ for (int j = 0; j < info.docCount; ++j) { if (liveDocs == null || liveDocs.get(j)) { status.docCount++; - Fields tfv = reader.getTermVectors(j); + InvertedFields tfv = reader.getTermVectors(j); if (tfv != null) { int tfvComputedFieldCount = 0; long tfvComputedTermCount = 0; Index: lucene/core/src/java/org/apache/lucene/index/BaseMultiReader.java =================================================================== --- lucene/core/src/java/org/apache/lucene/index/BaseMultiReader.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/index/BaseMultiReader.java (working copy) @@ -49,7 +49,7 @@ } @Override - public final Fields getTermVectors(int docID) throws IOException { + public final InvertedFields getTermVectors(int docID) throws IOException { ensureOpen(); final int i = readerIndex(docID); // find segment num return subReaders[i].getTermVectors(docID - starts[i]); // dispatch to segment Index: lucene/core/src/java/org/apache/lucene/index/Terms.java =================================================================== --- lucene/core/src/java/org/apache/lucene/index/Terms.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/index/Terms.java (working copy) @@ -24,7 +24,7 @@ import org.apache.lucene.util.automaton.CompiledAutomaton; /** - * Access to the terms in a specific field. See {@link Fields}. + * Access to the terms in a specific field. See {@link InvertedFields}. * @lucene.experimental */ Index: lucene/core/src/java/org/apache/lucene/index/SlowCompositeReaderWrapper.java =================================================================== --- lucene/core/src/java/org/apache/lucene/index/SlowCompositeReaderWrapper.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/index/SlowCompositeReaderWrapper.java (working copy) @@ -24,8 +24,6 @@ import org.apache.lucene.util.Bits; import org.apache.lucene.util.ReaderUtil; // javadoc -import org.apache.lucene.index.DirectoryReader; // javadoc -import org.apache.lucene.index.MultiReader; // javadoc /** * This class forces a composite reader (eg a {@link @@ -47,7 +45,7 @@ private final CompositeReader in; private final Map normsCache = new HashMap(); - private final Fields fields; + private final InvertedFields fields; private final Bits liveDocs; /** This method is sugar for getting an {@link AtomicReader} from @@ -76,7 +74,7 @@ } @Override - public Fields fields() throws IOException { + public InvertedFields fields() throws IOException { ensureOpen(); in.ensureOpen(); // as we cached the fields, we better check the original reader return fields; @@ -100,7 +98,7 @@ } @Override - public Fields getTermVectors(int docID) + public InvertedFields getTermVectors(int docID) throws IOException { ensureOpen(); return in.getTermVectors(docID); Index: lucene/core/src/java/org/apache/lucene/index/SegmentMerger.java =================================================================== --- lucene/core/src/java/org/apache/lucene/index/SegmentMerger.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/index/SegmentMerger.java (working copy) @@ -23,12 +23,8 @@ import java.util.List; import java.util.Map; -import org.apache.lucene.codecs.Codec; -import org.apache.lucene.codecs.FieldInfosWriter; -import org.apache.lucene.codecs.FieldsConsumer; -import org.apache.lucene.codecs.PerDocConsumer; -import org.apache.lucene.codecs.StoredFieldsWriter; -import org.apache.lucene.codecs.TermVectorsWriter; +import org.apache.lucene.codecs.*; +import org.apache.lucene.codecs.InvertedFieldsConsumer; import org.apache.lucene.store.Directory; import org.apache.lucene.store.IOContext; import org.apache.lucene.util.Bits; @@ -334,14 +330,14 @@ private final void mergeTerms(SegmentWriteState segmentWriteState) throws CorruptIndexException, IOException { - final List fields = new ArrayList(); + final List fields = new ArrayList(); final List slices = new ArrayList(); int docBase = 0; for(int readerIndex=0;readerIndex IMPL: FormatPostingsTermsDictWriter -> TermsConsumer -> IMPL: FormatPostingsTermsDictWriter.TermsWriter Index: lucene/core/src/java/org/apache/lucene/index/FreqProxTermsWriterPerField.java =================================================================== --- lucene/core/src/java/org/apache/lucene/index/FreqProxTermsWriterPerField.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/index/FreqProxTermsWriterPerField.java (working copy) @@ -23,7 +23,7 @@ import org.apache.lucene.analysis.tokenattributes.OffsetAttribute; import org.apache.lucene.analysis.tokenattributes.PayloadAttribute; -import org.apache.lucene.codecs.FieldsConsumer; +import org.apache.lucene.codecs.InvertedFieldsConsumer; import org.apache.lucene.codecs.PostingsConsumer; import org.apache.lucene.codecs.TermStats; import org.apache.lucene.codecs.TermsConsumer; @@ -316,7 +316,7 @@ /* Walk through all unique text tokens (Posting * instances) found in this field and serialize them * into a single RAM segment. */ - void flush(String fieldName, FieldsConsumer consumer, final SegmentWriteState state) + void flush(String fieldName, InvertedFieldsConsumer consumer, final SegmentWriteState state) throws CorruptIndexException, IOException { final TermsConsumer termsConsumer = consumer.addField(fieldInfo); Index: lucene/core/src/java/org/apache/lucene/index/ParallelAtomicReader.java =================================================================== --- lucene/core/src/java/org/apache/lucene/index/ParallelAtomicReader.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/index/ParallelAtomicReader.java (working copy) @@ -126,9 +126,9 @@ private final class ParallelFieldsEnum extends FieldsEnum { private String currentField; private final Iterator keys; - private final Fields fields; + private final InvertedFields fields; - ParallelFieldsEnum(Fields fields) { + ParallelFieldsEnum(InvertedFields fields) { this.fields = fields; keys = fieldToReader.keySet().iterator(); } @@ -151,7 +151,7 @@ } // Single instance of this, per ParallelReader instance - private final class ParallelFields extends Fields { + private final class ParallelFields extends InvertedFields { final HashMap fields = new HashMap(); ParallelFields() { @@ -189,7 +189,7 @@ } @Override - public Fields fields() { + public InvertedFields fields() { ensureOpen(); return fields; } @@ -222,7 +222,7 @@ // get all vectors @Override - public Fields getTermVectors(int docID) throws IOException { + public InvertedFields getTermVectors(int docID) throws IOException { ensureOpen(); ParallelFields fields = new ParallelFields(); for (Map.Entry ent : fieldToReader.entrySet()) { Index: lucene/core/src/java/org/apache/lucene/index/DocValues.java =================================================================== --- lucene/core/src/java/org/apache/lucene/index/DocValues.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/index/DocValues.java (working copy) @@ -74,7 +74,7 @@ *

* {@link Source} instances obtained from this method are closed / released * from the cache once this {@link DocValues} instance is closed by the - * {@link IndexReader}, {@link Fields} or {@link FieldsEnum} the + * {@link IndexReader}, {@link InvertedFields} or {@link FieldsEnum} the * {@link DocValues} was created from. */ public Source getSource() throws IOException { Index: lucene/core/src/java/org/apache/lucene/index/IndexReader.java =================================================================== --- lucene/core/src/java/org/apache/lucene/index/IndexReader.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/index/IndexReader.java (working copy) @@ -26,11 +26,9 @@ import org.apache.lucene.document.Document; import org.apache.lucene.document.DocumentStoredFieldVisitor; -import org.apache.lucene.search.SearcherManager; // javadocs import org.apache.lucene.store.*; import org.apache.lucene.util.Bits; import org.apache.lucene.util.BytesRef; -import org.apache.lucene.util.ReaderUtil; // for javadocs /** IndexReader is an abstract class, providing an interface for accessing an index. Search of an index is done entirely through this abstract interface, @@ -314,19 +312,19 @@ } /** Retrieve term vectors for this document, or null if - * term vectors were not indexed. The returned Fields + * term vectors were not indexed. The returned InvertedFields * instance acts like a single-document inverted index * (the docID will be 0). */ - public abstract Fields getTermVectors(int docID) + public abstract InvertedFields getTermVectors(int docID) throws IOException; /** Retrieve term vector for this document and field, or * null if term vectors were not indexed. The returned - * Fields instance acts like a single-document inverted + * InvertedFields instance acts like a single-document inverted * index (the docID will be 0). */ public final Terms getTermVector(int docID, String field) throws IOException { - Fields vectors = getTermVectors(docID); + InvertedFields vectors = getTermVectors(docID); if (vectors == null) { return null; } Index: lucene/core/src/java/org/apache/lucene/index/FilterAtomicReader.java =================================================================== --- lucene/core/src/java/org/apache/lucene/index/FilterAtomicReader.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/index/FilterAtomicReader.java (working copy) @@ -34,12 +34,12 @@ */ public class FilterAtomicReader extends AtomicReader { - /** Base class for filtering {@link Fields} + /** Base class for filtering {@link InvertedFields} * implementations. */ - public static class FilterFields extends Fields { - protected Fields in; + public static class FilterFields extends InvertedFields { + protected InvertedFields in; - public FilterFields(Fields in) { + public FilterFields(InvertedFields in) { this.in = in; } @@ -296,7 +296,7 @@ } @Override - public Fields getTermVectors(int docID) + public InvertedFields getTermVectors(int docID) throws IOException { ensureOpen(); return in.getTermVectors(docID); @@ -338,7 +338,7 @@ } @Override - public Fields fields() throws IOException { + public InvertedFields fields() throws IOException { ensureOpen(); return in.fields(); } Index: lucene/core/src/java/org/apache/lucene/index/SegmentCoreReaders.java =================================================================== --- lucene/core/src/java/org/apache/lucene/index/SegmentCoreReaders.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/index/SegmentCoreReaders.java (working copy) @@ -23,12 +23,8 @@ import java.util.Set; import java.util.concurrent.atomic.AtomicInteger; -import org.apache.lucene.codecs.Codec; -import org.apache.lucene.codecs.FieldsProducer; -import org.apache.lucene.codecs.PerDocProducer; -import org.apache.lucene.codecs.PostingsFormat; -import org.apache.lucene.codecs.StoredFieldsReader; -import org.apache.lucene.codecs.TermVectorsReader; +import org.apache.lucene.codecs.*; +import org.apache.lucene.codecs.InvertedFieldsProducer; import org.apache.lucene.index.SegmentReader.CoreClosedListener; import org.apache.lucene.store.CompoundFileDirectory; import org.apache.lucene.store.Directory; @@ -50,7 +46,7 @@ final FieldInfos fieldInfos; - final FieldsProducer fields; + final InvertedFieldsProducer fields; final PerDocProducer perDocProducer; final PerDocProducer norms; @@ -104,7 +100,7 @@ this.termsIndexDivisor = termsIndexDivisor; final PostingsFormat format = codec.postingsFormat(); final SegmentReadState segmentReadState = new SegmentReadState(cfsDir, si, fieldInfos, context, termsIndexDivisor); - // Ask codec for its Fields + // Ask codec for its InvertedFields fields = format.fieldsProducer(segmentReadState); assert fields != null; // ask codec for its Norms: Index: lucene/core/src/java/org/apache/lucene/index/MultiFields.java =================================================================== --- lucene/core/src/java/org/apache/lucene/index/MultiFields.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/index/MultiFields.java (working copy) @@ -46,12 +46,12 @@ * @lucene.experimental */ -public final class MultiFields extends Fields { - private final Fields[] subs; +public final class MultiFields extends InvertedFields { + private final InvertedFields[] subs; private final ReaderUtil.Slice[] subSlices; private final Map terms = new ConcurrentHashMap(); - /** Returns a single {@link Fields} instance for this + /** Returns a single {@link InvertedFields} instance for this * reader, merging fields/terms/docs/positions on the * fly. This method will return null if the reader * has no postings. @@ -60,7 +60,7 @@ * It's better to get the sub-readers (using {@link * Gather}) and iterate through them * yourself. */ - public static Fields getFields(IndexReader r) throws IOException { + public static InvertedFields getFields(IndexReader r) throws IOException { if (r instanceof AtomicReader) { // already an atomic reader return ((AtomicReader) r).fields(); @@ -71,13 +71,13 @@ // no fields return null; } else { - final List fields = new ArrayList(); + final List fields = new ArrayList(); final List slices = new ArrayList(); new ReaderUtil.Gather(r) { @Override protected void add(int base, AtomicReader r) throws IOException { - final Fields f = r.fields(); + final InvertedFields f = r.fields(); if (f != null) { fields.add(f); slices.add(new ReaderUtil.Slice(base, r.maxDoc(), fields.size()-1)); @@ -90,7 +90,7 @@ } else if (fields.size() == 1) { return fields.get(0); } else { - return new MultiFields(fields.toArray(Fields.EMPTY_ARRAY), + return new MultiFields(fields.toArray(InvertedFields.EMPTY_ARRAY), slices.toArray(ReaderUtil.Slice.EMPTY_ARRAY)); } } @@ -130,7 +130,7 @@ /** This method may return null if the field does not exist.*/ public static Terms getTerms(IndexReader r, String field) throws IOException { - final Fields fields = getFields(r); + final InvertedFields fields = getFields(r); if (fields == null) { return null; } else { @@ -170,7 +170,7 @@ return null; } - public MultiFields(Fields[] subs, ReaderUtil.Slice[] subSlices) { + public MultiFields(InvertedFields[] subs, ReaderUtil.Slice[] subSlices) { this.subs = subs; this.subSlices = subSlices; } Index: lucene/core/src/java/org/apache/lucene/util/TermContext.java =================================================================== --- lucene/core/src/java/org/apache/lucene/util/TermContext.java (revision 1245181) +++ lucene/core/src/java/org/apache/lucene/util/TermContext.java (working copy) @@ -20,14 +20,8 @@ import java.io.IOException; import java.util.Arrays; -import org.apache.lucene.index.AtomicReaderContext; -import org.apache.lucene.index.Fields; -import org.apache.lucene.index.IndexReader; -import org.apache.lucene.index.IndexReaderContext; -import org.apache.lucene.index.Term; -import org.apache.lucene.index.TermState; -import org.apache.lucene.index.Terms; -import org.apache.lucene.index.TermsEnum; +import org.apache.lucene.index.*; +import org.apache.lucene.index.InvertedFields; /** * Maintains a {@link IndexReader} {@link TermState} view over @@ -89,7 +83,7 @@ //if (DEBUG) System.out.println("prts.build term=" + term); for (int i = 0; i < leaves.length; i++) { //if (DEBUG) System.out.println(" r=" + leaves[i].reader); - final Fields fields = leaves[i].reader().fields(); + final InvertedFields fields = leaves[i].reader().fields(); if (fields != null) { final Terms terms = fields.terms(field); if (terms != null) { Index: lucene/contrib/misc/src/java/org/apache/lucene/misc/HighFreqTerms.java =================================================================== --- lucene/contrib/misc/src/java/org/apache/lucene/misc/HighFreqTerms.java (revision 1245181) +++ lucene/contrib/misc/src/java/org/apache/lucene/misc/HighFreqTerms.java (working copy) @@ -17,14 +17,8 @@ * limitations under the License. */ -import org.apache.lucene.index.AtomicReader; -import org.apache.lucene.index.IndexReader; -import org.apache.lucene.index.MultiFields; -import org.apache.lucene.index.Fields; -import org.apache.lucene.index.TermsEnum; -import org.apache.lucene.index.FieldsEnum; -import org.apache.lucene.index.Terms; -import org.apache.lucene.index.DocsEnum; +import org.apache.lucene.index.*; +import org.apache.lucene.index.InvertedFields; import org.apache.lucene.search.DocIdSetIterator; import org.apache.lucene.store.FSDirectory; import org.apache.lucene.util.PriorityQueue; @@ -115,7 +109,7 @@ TermStatsQueue tiq = null; if (field != null) { - Fields fields = MultiFields.getFields(reader); + InvertedFields fields = MultiFields.getFields(reader); if (fields == null) { throw new RuntimeException("field " + field + " not found"); } @@ -126,7 +120,7 @@ fillQueue(termsEnum, tiq, field); } } else { - Fields fields = MultiFields.getFields(reader); + InvertedFields fields = MultiFields.getFields(reader); if (fields == null) { throw new RuntimeException("no fields found for this index"); } Index: lucene/contrib/highlighter/src/java/org/apache/lucene/search/highlight/TokenSources.java =================================================================== --- lucene/contrib/highlighter/src/java/org/apache/lucene/search/highlight/TokenSources.java (revision 1245181) +++ lucene/contrib/highlighter/src/java/org/apache/lucene/search/highlight/TokenSources.java (working copy) @@ -32,11 +32,8 @@ import org.apache.lucene.analysis.tokenattributes.OffsetAttribute; import org.apache.lucene.analysis.tokenattributes.PositionIncrementAttribute; import org.apache.lucene.document.Document; -import org.apache.lucene.index.DocsAndPositionsEnum; -import org.apache.lucene.index.Fields; -import org.apache.lucene.index.IndexReader; -import org.apache.lucene.index.Terms; -import org.apache.lucene.index.TermsEnum; +import org.apache.lucene.index.*; +import org.apache.lucene.index.InvertedFields; import org.apache.lucene.util.ArrayUtil; import org.apache.lucene.util.BytesRef; @@ -70,7 +67,7 @@ String field, Document doc, Analyzer analyzer) throws IOException { TokenStream ts = null; - Fields vectors = reader.getTermVectors(docId); + InvertedFields vectors = reader.getTermVectors(docId); if (vectors != null) { Terms vector = vectors.terms(field); if (vector != null) { @@ -102,7 +99,7 @@ String field, Analyzer analyzer) throws IOException { TokenStream ts = null; - Fields vectors = reader.getTermVectors(docId); + InvertedFields vectors = reader.getTermVectors(docId); if (vectors != null) { Terms vector = vectors.terms(field); if (vector != null) { @@ -275,7 +272,7 @@ public static TokenStream getTokenStream(IndexReader reader, int docId, String field) throws IOException { - Fields vectors = reader.getTermVectors(docId); + InvertedFields vectors = reader.getTermVectors(docId); if (vectors == null) { throw new IllegalArgumentException(field + " in doc #" + docId + "does not have any term position data stored"); Index: lucene/contrib/highlighter/src/java/org/apache/lucene/search/vectorhighlight/FieldTermStack.java =================================================================== --- lucene/contrib/highlighter/src/java/org/apache/lucene/search/vectorhighlight/FieldTermStack.java (revision 1245181) +++ lucene/contrib/highlighter/src/java/org/apache/lucene/search/vectorhighlight/FieldTermStack.java (working copy) @@ -22,7 +22,7 @@ import java.util.Set; import org.apache.lucene.index.DocsAndPositionsEnum; -import org.apache.lucene.index.Fields; +import org.apache.lucene.index.InvertedFields; import org.apache.lucene.index.IndexReader; import org.apache.lucene.index.Terms; import org.apache.lucene.index.TermsEnum; @@ -78,7 +78,7 @@ // just return to make null snippet if un-matched fieldName specified when fieldMatch == true if( termSet == null ) return; - final Fields vectors = reader.getTermVectors(docId); + final InvertedFields vectors = reader.getTermVectors(docId); if (vectors == null) { // null snippet return; Index: lucene/contrib/memory/src/java/org/apache/lucene/index/memory/MemoryIndex.java =================================================================== --- lucene/contrib/memory/src/java/org/apache/lucene/index/memory/MemoryIndex.java (revision 1245181) +++ lucene/contrib/memory/src/java/org/apache/lucene/index/memory/MemoryIndex.java (working copy) @@ -33,23 +33,8 @@ import org.apache.lucene.analysis.tokenattributes.PositionIncrementAttribute; import org.apache.lucene.analysis.tokenattributes.TermToBytesRefAttribute; import org.apache.lucene.document.Document; -import org.apache.lucene.index.AtomicReader; -import org.apache.lucene.index.AtomicReaderContext; -import org.apache.lucene.index.Norm; -import org.apache.lucene.index.DocValues; -import org.apache.lucene.index.DocsAndPositionsEnum; -import org.apache.lucene.index.DocsEnum; -import org.apache.lucene.index.FieldInfos; -import org.apache.lucene.index.FieldInvertState; -import org.apache.lucene.index.Fields; -import org.apache.lucene.index.FieldsEnum; -import org.apache.lucene.index.IndexReader; -import org.apache.lucene.index.OrdTermState; -import org.apache.lucene.index.StoredFieldVisitor; -import org.apache.lucene.index.Term; -import org.apache.lucene.index.TermState; -import org.apache.lucene.index.Terms; -import org.apache.lucene.index.TermsEnum; +import org.apache.lucene.index.*; +import org.apache.lucene.index.InvertedFields; import org.apache.lucene.index.memory.MemoryIndexNormDocValues.SingleValueSource; import org.apache.lucene.search.Collector; import org.apache.lucene.search.IndexSearcher; @@ -775,7 +760,7 @@ return fieldInfos; } - private class MemoryFields extends Fields { + private class MemoryFields extends InvertedFields { @Override public FieldsEnum iterator() { return new FieldsEnum() { @@ -850,7 +835,7 @@ } @Override - public Fields fields() { + public InvertedFields fields() { sortFields(); return new MemoryFields(); } @@ -1076,7 +1061,7 @@ } @Override - public Fields getTermVectors(int docID) { + public InvertedFields getTermVectors(int docID) { if (docID == 0) { return fields(); } else { Index: lucene/test-framework/src/java/org/apache/lucene/codecs/mockintblock/MockFixedIntBlockPostingsFormat.java =================================================================== --- lucene/test-framework/src/java/org/apache/lucene/codecs/mockintblock/MockFixedIntBlockPostingsFormat.java (revision 1245181) +++ lucene/test-framework/src/java/org/apache/lucene/codecs/mockintblock/MockFixedIntBlockPostingsFormat.java (working copy) @@ -20,17 +20,8 @@ import java.io.IOException; import java.util.Set; -import org.apache.lucene.codecs.BlockTermsReader; -import org.apache.lucene.codecs.BlockTermsWriter; -import org.apache.lucene.codecs.FieldsConsumer; -import org.apache.lucene.codecs.FieldsProducer; -import org.apache.lucene.codecs.FixedGapTermsIndexReader; -import org.apache.lucene.codecs.FixedGapTermsIndexWriter; -import org.apache.lucene.codecs.PostingsFormat; -import org.apache.lucene.codecs.PostingsReaderBase; -import org.apache.lucene.codecs.PostingsWriterBase; -import org.apache.lucene.codecs.TermsIndexReaderBase; -import org.apache.lucene.codecs.TermsIndexWriterBase; +import org.apache.lucene.codecs.*; +import org.apache.lucene.codecs.InvertedFieldsConsumer; import org.apache.lucene.codecs.intblock.FixedIntBlockIndexInput; import org.apache.lucene.codecs.intblock.FixedIntBlockIndexOutput; import org.apache.lucene.codecs.lucene40.Lucene40PostingsFormat; @@ -125,7 +116,7 @@ } @Override - public FieldsConsumer fieldsConsumer(SegmentWriteState state) throws IOException { + public InvertedFieldsConsumer fieldsConsumer(SegmentWriteState state) throws IOException { PostingsWriterBase postingsWriter = new SepPostingsWriter(state, new MockIntFactory(blockSize)); boolean success = false; @@ -141,7 +132,7 @@ success = false; try { - FieldsConsumer ret = new BlockTermsWriter(indexWriter, state, postingsWriter); + InvertedFieldsConsumer ret = new BlockTermsWriter(indexWriter, state, postingsWriter); success = true; return ret; } finally { @@ -156,7 +147,7 @@ } @Override - public FieldsProducer fieldsProducer(SegmentReadState state) throws IOException { + public InvertedFieldsProducer fieldsProducer(SegmentReadState state) throws IOException { PostingsReaderBase postingsReader = new SepPostingsReader(state.dir, state.segmentInfo, state.context, @@ -180,7 +171,7 @@ success = false; try { - FieldsProducer ret = new BlockTermsReader(indexReader, + InvertedFieldsProducer ret = new BlockTermsReader(indexReader, state.dir, state.fieldInfos, state.segmentInfo.name, Index: lucene/test-framework/src/java/org/apache/lucene/codecs/mockintblock/MockVariableIntBlockPostingsFormat.java =================================================================== --- lucene/test-framework/src/java/org/apache/lucene/codecs/mockintblock/MockVariableIntBlockPostingsFormat.java (revision 1245181) +++ lucene/test-framework/src/java/org/apache/lucene/codecs/mockintblock/MockVariableIntBlockPostingsFormat.java (working copy) @@ -20,17 +20,8 @@ import java.io.IOException; import java.util.Set; -import org.apache.lucene.codecs.BlockTermsReader; -import org.apache.lucene.codecs.BlockTermsWriter; -import org.apache.lucene.codecs.FieldsConsumer; -import org.apache.lucene.codecs.FieldsProducer; -import org.apache.lucene.codecs.FixedGapTermsIndexReader; -import org.apache.lucene.codecs.FixedGapTermsIndexWriter; -import org.apache.lucene.codecs.PostingsFormat; -import org.apache.lucene.codecs.PostingsReaderBase; -import org.apache.lucene.codecs.PostingsWriterBase; -import org.apache.lucene.codecs.TermsIndexReaderBase; -import org.apache.lucene.codecs.TermsIndexWriterBase; +import org.apache.lucene.codecs.*; +import org.apache.lucene.codecs.InvertedFieldsProducer; import org.apache.lucene.codecs.intblock.VariableIntBlockIndexInput; import org.apache.lucene.codecs.intblock.VariableIntBlockIndexOutput; import org.apache.lucene.codecs.lucene40.Lucene40PostingsFormat; @@ -148,7 +139,7 @@ } @Override - public FieldsConsumer fieldsConsumer(SegmentWriteState state) throws IOException { + public InvertedFieldsConsumer fieldsConsumer(SegmentWriteState state) throws IOException { PostingsWriterBase postingsWriter = new SepPostingsWriter(state, new MockIntFactory(baseBlockSize)); boolean success = false; @@ -164,7 +155,7 @@ success = false; try { - FieldsConsumer ret = new BlockTermsWriter(indexWriter, state, postingsWriter); + InvertedFieldsConsumer ret = new BlockTermsWriter(indexWriter, state, postingsWriter); success = true; return ret; } finally { @@ -179,7 +170,7 @@ } @Override - public FieldsProducer fieldsProducer(SegmentReadState state) throws IOException { + public InvertedFieldsProducer fieldsProducer(SegmentReadState state) throws IOException { PostingsReaderBase postingsReader = new SepPostingsReader(state.dir, state.segmentInfo, state.context, @@ -203,7 +194,7 @@ success = false; try { - FieldsProducer ret = new BlockTermsReader(indexReader, + InvertedFieldsProducer ret = new BlockTermsReader(indexReader, state.dir, state.fieldInfos, state.segmentInfo.name, Index: lucene/test-framework/src/java/org/apache/lucene/codecs/ramonly/RAMOnlyPostingsFormat.java =================================================================== --- lucene/test-framework/src/java/org/apache/lucene/codecs/ramonly/RAMOnlyPostingsFormat.java (revision 1245181) +++ lucene/test-framework/src/java/org/apache/lucene/codecs/ramonly/RAMOnlyPostingsFormat.java (working copy) @@ -29,12 +29,8 @@ import java.util.TreeMap; import java.util.concurrent.atomic.AtomicInteger; -import org.apache.lucene.codecs.FieldsConsumer; -import org.apache.lucene.codecs.FieldsProducer; -import org.apache.lucene.codecs.PostingsConsumer; -import org.apache.lucene.codecs.PostingsFormat; -import org.apache.lucene.codecs.TermStats; -import org.apache.lucene.codecs.TermsConsumer; +import org.apache.lucene.codecs.*; +import org.apache.lucene.codecs.InvertedFieldsConsumer; import org.apache.lucene.index.DocsAndPositionsEnum; import org.apache.lucene.index.DocsEnum; import org.apache.lucene.index.FieldInfo.IndexOptions; @@ -100,7 +96,7 @@ } // Postings state: - static class RAMPostings extends FieldsProducer { + static class RAMPostings extends InvertedFieldsProducer { final Map fieldToTerms = new TreeMap(); @Override @@ -186,7 +182,7 @@ } // Classes for writing to the postings state - private static class RAMFieldsConsumer extends FieldsConsumer { + private static class RAMFieldsConsumer extends InvertedFieldsConsumer { private final RAMPostings postings; private final RAMTermsConsumer termsConsumer = new RAMTermsConsumer(); @@ -534,7 +530,7 @@ private static final String ID_EXTENSION = "id"; @Override - public FieldsConsumer fieldsConsumer(SegmentWriteState writeState) throws IOException { + public InvertedFieldsConsumer fieldsConsumer(SegmentWriteState writeState) throws IOException { final int id = nextID.getAndIncrement(); // TODO -- ok to do this up front instead of @@ -565,7 +561,7 @@ } @Override - public FieldsProducer fieldsProducer(SegmentReadState readState) + public InvertedFieldsProducer fieldsProducer(SegmentReadState readState) throws IOException { // Load our ID: Index: lucene/test-framework/src/java/org/apache/lucene/codecs/mocksep/MockSepPostingsFormat.java =================================================================== --- lucene/test-framework/src/java/org/apache/lucene/codecs/mocksep/MockSepPostingsFormat.java (revision 1245181) +++ lucene/test-framework/src/java/org/apache/lucene/codecs/mocksep/MockSepPostingsFormat.java (working copy) @@ -20,17 +20,8 @@ import java.io.IOException; import java.util.Set; -import org.apache.lucene.codecs.BlockTermsReader; -import org.apache.lucene.codecs.BlockTermsWriter; -import org.apache.lucene.codecs.FieldsConsumer; -import org.apache.lucene.codecs.FieldsProducer; -import org.apache.lucene.codecs.FixedGapTermsIndexReader; -import org.apache.lucene.codecs.FixedGapTermsIndexWriter; -import org.apache.lucene.codecs.PostingsFormat; -import org.apache.lucene.codecs.PostingsReaderBase; -import org.apache.lucene.codecs.PostingsWriterBase; -import org.apache.lucene.codecs.TermsIndexReaderBase; -import org.apache.lucene.codecs.TermsIndexWriterBase; +import org.apache.lucene.codecs.*; +import org.apache.lucene.codecs.InvertedFieldsConsumer; import org.apache.lucene.codecs.lucene40.Lucene40PostingsFormat; import org.apache.lucene.codecs.sep.SepPostingsReader; import org.apache.lucene.codecs.sep.SepPostingsWriter; @@ -52,7 +43,7 @@ } @Override - public FieldsConsumer fieldsConsumer(SegmentWriteState state) throws IOException { + public InvertedFieldsConsumer fieldsConsumer(SegmentWriteState state) throws IOException { PostingsWriterBase postingsWriter = new SepPostingsWriter(state, new MockSingleIntFactory()); @@ -69,7 +60,7 @@ success = false; try { - FieldsConsumer ret = new BlockTermsWriter(indexWriter, state, postingsWriter); + InvertedFieldsConsumer ret = new BlockTermsWriter(indexWriter, state, postingsWriter); success = true; return ret; } finally { @@ -84,7 +75,7 @@ } @Override - public FieldsProducer fieldsProducer(SegmentReadState state) throws IOException { + public InvertedFieldsProducer fieldsProducer(SegmentReadState state) throws IOException { PostingsReaderBase postingsReader = new SepPostingsReader(state.dir, state.segmentInfo, state.context, new MockSingleIntFactory(), state.segmentSuffix); @@ -107,7 +98,7 @@ success = false; try { - FieldsProducer ret = new BlockTermsReader(indexReader, + InvertedFieldsProducer ret = new BlockTermsReader(indexReader, state.dir, state.fieldInfos, state.segmentInfo.name, Index: lucene/test-framework/src/java/org/apache/lucene/codecs/nestedpulsing/NestedPulsingPostingsFormat.java =================================================================== --- lucene/test-framework/src/java/org/apache/lucene/codecs/nestedpulsing/NestedPulsingPostingsFormat.java (revision 1245181) +++ lucene/test-framework/src/java/org/apache/lucene/codecs/nestedpulsing/NestedPulsingPostingsFormat.java (working copy) @@ -20,13 +20,8 @@ import java.io.IOException; import java.util.Set; -import org.apache.lucene.codecs.BlockTreeTermsReader; -import org.apache.lucene.codecs.BlockTreeTermsWriter; -import org.apache.lucene.codecs.FieldsConsumer; -import org.apache.lucene.codecs.FieldsProducer; -import org.apache.lucene.codecs.PostingsFormat; -import org.apache.lucene.codecs.PostingsReaderBase; -import org.apache.lucene.codecs.PostingsWriterBase; +import org.apache.lucene.codecs.*; +import org.apache.lucene.codecs.InvertedFieldsConsumer; import org.apache.lucene.codecs.lucene40.Lucene40PostingsReader; import org.apache.lucene.codecs.lucene40.Lucene40PostingsWriter; import org.apache.lucene.codecs.pulsing.PulsingPostingsReader; @@ -48,7 +43,7 @@ } @Override - public FieldsConsumer fieldsConsumer(SegmentWriteState state) throws IOException { + public InvertedFieldsConsumer fieldsConsumer(SegmentWriteState state) throws IOException { PostingsWriterBase docsWriter = new Lucene40PostingsWriter(state); PostingsWriterBase pulsingWriterInner = new PulsingPostingsWriter(2, docsWriter); @@ -57,7 +52,7 @@ // Terms dict boolean success = false; try { - FieldsConsumer ret = new BlockTreeTermsWriter(state, pulsingWriter, + InvertedFieldsConsumer ret = new BlockTreeTermsWriter(state, pulsingWriter, BlockTreeTermsWriter.DEFAULT_MIN_BLOCK_SIZE, BlockTreeTermsWriter.DEFAULT_MAX_BLOCK_SIZE); success = true; return ret; @@ -69,13 +64,13 @@ } @Override - public FieldsProducer fieldsProducer(SegmentReadState state) throws IOException { + public InvertedFieldsProducer fieldsProducer(SegmentReadState state) throws IOException { PostingsReaderBase docsReader = new Lucene40PostingsReader(state.dir, state.segmentInfo, state.context, state.segmentSuffix); PostingsReaderBase pulsingReaderInner = new PulsingPostingsReader(docsReader); PostingsReaderBase pulsingReader = new PulsingPostingsReader(pulsingReaderInner); boolean success = false; try { - FieldsProducer ret = new BlockTreeTermsReader( + InvertedFieldsProducer ret = new BlockTreeTermsReader( state.dir, state.fieldInfos, state.segmentInfo.name, pulsingReader, state.context, Index: lucene/test-framework/src/java/org/apache/lucene/codecs/lucene3x/PreFlexRWFieldsWriter.java =================================================================== --- lucene/test-framework/src/java/org/apache/lucene/codecs/lucene3x/PreFlexRWFieldsWriter.java (revision 1245181) +++ lucene/test-framework/src/java/org/apache/lucene/codecs/lucene3x/PreFlexRWFieldsWriter.java (working copy) @@ -20,7 +20,7 @@ import java.io.IOException; import java.util.Comparator; -import org.apache.lucene.codecs.FieldsConsumer; +import org.apache.lucene.codecs.InvertedFieldsConsumer; import org.apache.lucene.codecs.PostingsConsumer; import org.apache.lucene.codecs.TermStats; import org.apache.lucene.codecs.TermsConsumer; @@ -34,7 +34,7 @@ import org.apache.lucene.util.BytesRef; import org.apache.lucene.util.IOUtils; -class PreFlexRWFieldsWriter extends FieldsConsumer { +class PreFlexRWFieldsWriter extends InvertedFieldsConsumer { private final TermInfosWriter termsOut; private final IndexOutput freqOut; Index: lucene/test-framework/src/java/org/apache/lucene/codecs/lucene3x/PreFlexRWPostingsFormat.java =================================================================== --- lucene/test-framework/src/java/org/apache/lucene/codecs/lucene3x/PreFlexRWPostingsFormat.java (revision 1245181) +++ lucene/test-framework/src/java/org/apache/lucene/codecs/lucene3x/PreFlexRWPostingsFormat.java (working copy) @@ -19,8 +19,8 @@ import java.io.IOException; -import org.apache.lucene.codecs.FieldsConsumer; -import org.apache.lucene.codecs.FieldsProducer; +import org.apache.lucene.codecs.InvertedFieldsConsumer; +import org.apache.lucene.codecs.InvertedFieldsProducer; import org.apache.lucene.index.SegmentWriteState; import org.apache.lucene.index.SegmentReadState; import org.apache.lucene.util.LuceneTestCase; @@ -38,12 +38,12 @@ } @Override - public FieldsConsumer fieldsConsumer(SegmentWriteState state) throws IOException { + public InvertedFieldsConsumer fieldsConsumer(SegmentWriteState state) throws IOException { return new PreFlexRWFieldsWriter(state); } @Override - public FieldsProducer fieldsProducer(SegmentReadState state) throws IOException { + public InvertedFieldsProducer fieldsProducer(SegmentReadState state) throws IOException { // Whenever IW opens readers, eg for merging, we have to // keep terms order in UTF16: Index: lucene/test-framework/src/java/org/apache/lucene/codecs/lucene40ords/Lucene40WithOrds.java =================================================================== --- lucene/test-framework/src/java/org/apache/lucene/codecs/lucene40ords/Lucene40WithOrds.java (revision 1245181) +++ lucene/test-framework/src/java/org/apache/lucene/codecs/lucene40ords/Lucene40WithOrds.java (working copy) @@ -22,8 +22,8 @@ import org.apache.lucene.codecs.BlockTermsReader; import org.apache.lucene.codecs.BlockTermsWriter; -import org.apache.lucene.codecs.FieldsConsumer; -import org.apache.lucene.codecs.FieldsProducer; +import org.apache.lucene.codecs.InvertedFieldsConsumer; +import org.apache.lucene.codecs.InvertedFieldsProducer; import org.apache.lucene.codecs.FixedGapTermsIndexReader; import org.apache.lucene.codecs.FixedGapTermsIndexWriter; import org.apache.lucene.codecs.PostingsFormat; @@ -48,7 +48,7 @@ } @Override - public FieldsConsumer fieldsConsumer(SegmentWriteState state) throws IOException { + public InvertedFieldsConsumer fieldsConsumer(SegmentWriteState state) throws IOException { PostingsWriterBase docs = new Lucene40PostingsWriter(state); // TODO: should we make the terms index more easily @@ -70,7 +70,7 @@ try { // Must use BlockTermsWriter (not BlockTree) because // BlockTree doens't support ords (yet)... - FieldsConsumer ret = new BlockTermsWriter(indexWriter, state, docs); + InvertedFieldsConsumer ret = new BlockTermsWriter(indexWriter, state, docs); success = true; return ret; } finally { @@ -87,7 +87,7 @@ public final static int TERMS_CACHE_SIZE = 1024; @Override - public FieldsProducer fieldsProducer(SegmentReadState state) throws IOException { + public InvertedFieldsProducer fieldsProducer(SegmentReadState state) throws IOException { PostingsReaderBase postings = new Lucene40PostingsReader(state.dir, state.segmentInfo, state.context, state.segmentSuffix); TermsIndexReaderBase indexReader; @@ -108,7 +108,7 @@ success = false; try { - FieldsProducer ret = new BlockTermsReader(indexReader, + InvertedFieldsProducer ret = new BlockTermsReader(indexReader, state.dir, state.fieldInfos, state.segmentInfo.name, Index: lucene/test-framework/src/java/org/apache/lucene/codecs/mockrandom/MockRandomPostingsFormat.java =================================================================== --- lucene/test-framework/src/java/org/apache/lucene/codecs/mockrandom/MockRandomPostingsFormat.java (revision 1245181) +++ lucene/test-framework/src/java/org/apache/lucene/codecs/mockrandom/MockRandomPostingsFormat.java (working copy) @@ -24,22 +24,8 @@ import java.util.Random; import java.util.Set; -import org.apache.lucene.codecs.BlockTermsReader; -import org.apache.lucene.codecs.BlockTermsWriter; -import org.apache.lucene.codecs.BlockTreeTermsReader; -import org.apache.lucene.codecs.BlockTreeTermsWriter; -import org.apache.lucene.codecs.FieldsConsumer; -import org.apache.lucene.codecs.FieldsProducer; -import org.apache.lucene.codecs.FixedGapTermsIndexReader; -import org.apache.lucene.codecs.FixedGapTermsIndexWriter; -import org.apache.lucene.codecs.PostingsFormat; -import org.apache.lucene.codecs.PostingsReaderBase; -import org.apache.lucene.codecs.PostingsWriterBase; -import org.apache.lucene.codecs.TermStats; -import org.apache.lucene.codecs.TermsIndexReaderBase; -import org.apache.lucene.codecs.TermsIndexWriterBase; -import org.apache.lucene.codecs.VariableGapTermsIndexReader; -import org.apache.lucene.codecs.VariableGapTermsIndexWriter; +import org.apache.lucene.codecs.*; +import org.apache.lucene.codecs.InvertedFieldsConsumer; import org.apache.lucene.codecs.lucene40.Lucene40PostingsReader; import org.apache.lucene.codecs.lucene40.Lucene40PostingsWriter; import org.apache.lucene.codecs.mockintblock.MockFixedIntBlockPostingsFormat; @@ -126,7 +112,7 @@ } @Override - public FieldsConsumer fieldsConsumer(SegmentWriteState state) throws IOException { + public InvertedFieldsConsumer fieldsConsumer(SegmentWriteState state) throws IOException { // we pull this before the seed intentionally: because its not consumed at runtime // (the skipInterval is written into postings header) int skipInterval = _TestUtil.nextInt(seedRandom, 2, 10); @@ -171,7 +157,7 @@ postingsWriter = new PulsingPostingsWriter(totTFCutoff, postingsWriter); } - final FieldsConsumer fields; + final InvertedFieldsConsumer fields; if (random.nextBoolean()) { // Use BlockTree terms dict @@ -270,7 +256,7 @@ } @Override - public FieldsProducer fieldsProducer(SegmentReadState state) throws IOException { + public InvertedFieldsProducer fieldsProducer(SegmentReadState state) throws IOException { final String seedFileName = IndexFileNames.segmentFileName(state.segmentInfo.name, state.segmentSuffix, SEED_EXT); final IndexInput in = state.dir.openInput(seedFileName, state.context); @@ -310,7 +296,7 @@ postingsReader = new PulsingPostingsReader(postingsReader); } - final FieldsProducer fields; + final InvertedFieldsProducer fields; if (random.nextBoolean()) { // Use BlockTree terms dict Index: lucene/test-framework/src/java/org/apache/lucene/index/ThreadedIndexingAndSearchingTestCase.java =================================================================== --- lucene/test-framework/src/java/org/apache/lucene/index/ThreadedIndexingAndSearchingTestCase.java (revision 1245181) +++ lucene/test-framework/src/java/org/apache/lucene/index/ThreadedIndexingAndSearchingTestCase.java (working copy) @@ -340,7 +340,7 @@ try { if (s.getIndexReader().numDocs() > 0) { smokeTestSearcher(s); - Fields fields = MultiFields.getFields(s.getIndexReader()); + InvertedFields fields = MultiFields.getFields(s.getIndexReader()); if (fields == null) { continue; }