Index: solr/src/java/org/apache/solr/search/SolrIndexReader.java =================================================================== --- solr/src/java/org/apache/solr/search/SolrIndexReader.java (revision 1056829) +++ solr/src/java/org/apache/solr/search/SolrIndexReader.java (working copy) @@ -288,11 +288,6 @@ } @Override - public void norms(String f, byte[] bytes, int offset) throws IOException { - in.norms(f, bytes, offset); - } - - @Override protected void doSetNorm(int d, String f, byte b) throws CorruptIndexException, IOException { in.setNorm(d, f, b); } @@ -473,11 +468,6 @@ } @Override - public void setNorm(int doc, String field, float value) throws StaleReaderException, CorruptIndexException, LockObtainFailedException, IOException { - in.setNorm(doc, field, value); - } - - @Override public void undeleteAll() throws StaleReaderException, CorruptIndexException, LockObtainFailedException, IOException { in.undeleteAll(); } Index: lucene/src/test/org/apache/lucene/TestExternalCodecs.java =================================================================== --- lucene/src/test/org/apache/lucene/TestExternalCodecs.java (revision 1056829) +++ lucene/src/test/org/apache/lucene/TestExternalCodecs.java (working copy) @@ -88,6 +88,11 @@ public FieldsEnum iterator() { return new RAMFieldsEnum(this); } + + @Override + public boolean hasField(String field) { + return fieldToTerms.containsKey(field); + } @Override public void close() { Index: lucene/src/test/org/apache/lucene/search/TestSetNorm.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestSetNorm.java (revision 1056829) +++ lucene/src/test/org/apache/lucene/search/TestSetNorm.java (working copy) @@ -50,10 +50,10 @@ // reset the boost of each instance of this document IndexReader reader = IndexReader.open(store, false); - reader.setNorm(0, "field", 1.0f); - reader.setNorm(1, "field", 2.0f); - reader.setNorm(2, "field", 4.0f); - reader.setNorm(3, "field", 16.0f); + reader.setNorm(0, "field", Similarity.getDefault().encodeNormValue(1.0f)); + reader.setNorm(1, "field", Similarity.getDefault().encodeNormValue(2.0f)); + reader.setNorm(2, "field", Similarity.getDefault().encodeNormValue(4.0f)); + reader.setNorm(3, "field", Similarity.getDefault().encodeNormValue(16.0f)); reader.close(); // check that searches are ordered by this boost Index: lucene/src/test/org/apache/lucene/search/TestMatchAllDocsQuery.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestMatchAllDocsQuery.java (revision 1056829) +++ lucene/src/test/org/apache/lucene/search/TestMatchAllDocsQuery.java (working copy) @@ -69,7 +69,7 @@ assertEquals("one", ir.document(hits[2].doc).get("key")); // change norm & retest - ir.setNorm(0, "key", 400f); + ir.setNorm(0, "key", Similarity.getDefault().encodeNormValue(400f)); normsQuery = new MatchAllDocsQuery("key"); hits = is.search(normsQuery, null, 1000).scoreDocs; assertEquals(3, hits.length); Index: lucene/src/test/org/apache/lucene/index/TestIndexReaderClone.java =================================================================== --- lucene/src/test/org/apache/lucene/index/TestIndexReaderClone.java (revision 1056829) +++ lucene/src/test/org/apache/lucene/index/TestIndexReaderClone.java (working copy) @@ -276,7 +276,7 @@ IndexReader pr1Clone = (IndexReader) r1.clone(); pr1Clone.deleteDocument(10); - pr1Clone.setNorm(4, "field1", 0.5f); + pr1Clone.setNorm(4, "field1", Similarity.getDefault().encodeNormValue(0.5f)); assertTrue(Similarity.getDefault().decodeNormValue(MultiNorms.norms(r1, "field1")[4]) == norm1); assertTrue(Similarity.getDefault().decodeNormValue(MultiNorms.norms(pr1Clone, "field1")[4]) != norm1); @@ -327,7 +327,7 @@ TestIndexReaderReopen.createIndex(random, dir1, false); SegmentReader origSegmentReader = getOnlySegmentReader(IndexReader.open(dir1, false)); origSegmentReader.deleteDocument(1); - origSegmentReader.setNorm(4, "field1", 0.5f); + origSegmentReader.setNorm(4, "field1", Similarity.getDefault().encodeNormValue(0.5f)); SegmentReader clonedSegmentReader = (SegmentReader) origSegmentReader .clone(); @@ -426,7 +426,7 @@ final Directory dir1 = newDirectory(); TestIndexReaderReopen.createIndex(random, dir1, false); IndexReader orig = IndexReader.open(dir1, false); - orig.setNorm(1, "field1", 17.0f); + orig.setNorm(1, "field1", Similarity.getDefault().encodeNormValue(17.0f)); final byte encoded = Similarity.getDefault().encodeNormValue(17.0f); assertEquals(encoded, MultiNorms.norms(orig, "field1")[1]); Index: lucene/src/test/org/apache/lucene/index/TestIndexReader.java =================================================================== --- lucene/src/test/org/apache/lucene/index/TestIndexReader.java (revision 1056829) +++ lucene/src/test/org/apache/lucene/index/TestIndexReader.java (working copy) @@ -43,6 +43,7 @@ import org.apache.lucene.search.FieldCache; import org.apache.lucene.search.IndexSearcher; import org.apache.lucene.search.ScoreDoc; +import org.apache.lucene.search.Similarity; import org.apache.lucene.search.TermQuery; import org.apache.lucene.store.AlreadyClosedException; import org.apache.lucene.store.Directory; @@ -463,7 +464,7 @@ } try { - reader.setNorm(5, "aaa", 2.0f); + reader.setNorm(5, "aaa", Similarity.getDefault().encodeNormValue(2.0f)); fail("setNorm after close failed to throw IOException"); } catch (AlreadyClosedException e) { // expected @@ -503,7 +504,7 @@ } try { - reader.setNorm(5, "aaa", 2.0f); + reader.setNorm(5, "aaa", Similarity.getDefault().encodeNormValue(2.0f)); fail("setNorm should have hit LockObtainFailedException"); } catch (LockObtainFailedException e) { // expected @@ -533,7 +534,7 @@ // now open reader & set norm for doc 0 IndexReader reader = IndexReader.open(dir, false); - reader.setNorm(0, "content", (float) 2.0); + reader.setNorm(0, "content", Similarity.getDefault().encodeNormValue(2.0f)); // we should be holding the write lock now: assertTrue("locked", IndexWriter.isLocked(dir)); @@ -547,7 +548,7 @@ IndexReader reader2 = IndexReader.open(dir, false); // set norm again for doc 0 - reader.setNorm(0, "content", (float) 3.0); + reader.setNorm(0, "content", Similarity.getDefault().encodeNormValue(3.0f)); assertTrue("locked", IndexWriter.isLocked(dir)); reader.close(); @@ -580,12 +581,12 @@ // now open reader & set norm for doc 0 (writes to // _0_1.s0) reader = IndexReader.open(dir, false); - reader.setNorm(0, "content", (float) 2.0); + reader.setNorm(0, "content", Similarity.getDefault().encodeNormValue(2.0f)); reader.close(); // now open reader again & set norm for doc 0 (writes to _0_2.s0) reader = IndexReader.open(dir, false); - reader.setNorm(0, "content", (float) 2.0); + reader.setNorm(0, "content", Similarity.getDefault().encodeNormValue(2.0f)); reader.close(); assertFalse("failed to remove first generation norms file on writing second generation", dir.fileExists("_0_1.s0")); @@ -953,7 +954,7 @@ int docId = 12; for(int i=0;i<13;i++) { reader.deleteDocument(docId); - reader.setNorm(docId, "content", (float) 2.0); + reader.setNorm(docId, "content", Similarity.getDefault().encodeNormValue(2.0f)); docId += 12; } } @@ -1112,7 +1113,7 @@ reader = IndexReader.open(dir, false); try { - reader.setNorm(1, "content", (float) 2.0); + reader.setNorm(1, "content", Similarity.getDefault().encodeNormValue(2.0f)); fail("did not hit exception when calling setNorm on an invalid doc number"); } catch (ArrayIndexOutOfBoundsException e) { // expected Index: lucene/src/test/org/apache/lucene/index/TestIndexFileDeleter.java =================================================================== --- lucene/src/test/org/apache/lucene/index/TestIndexFileDeleter.java (revision 1056829) +++ lucene/src/test/org/apache/lucene/index/TestIndexFileDeleter.java (working copy) @@ -18,6 +18,7 @@ */ import org.apache.lucene.util.LuceneTestCase; +import org.apache.lucene.search.Similarity; import org.apache.lucene.store.Directory; import org.apache.lucene.store.IndexInput; import org.apache.lucene.store.IndexOutput; @@ -68,7 +69,7 @@ assertEquals("didn't delete the right number of documents", 1, delCount); // Set one norm so we get a .s0 file: - reader.setNorm(21, "content", (float) 1.5); + reader.setNorm(21, "content", Similarity.getDefault().encodeNormValue(1.5f)); reader.close(); // Now, artificially create an extra .del file & extra Index: lucene/src/test/org/apache/lucene/index/TestIndexReaderCloneNorms.java =================================================================== --- lucene/src/test/org/apache/lucene/index/TestIndexReaderCloneNorms.java (revision 1056829) +++ lucene/src/test/org/apache/lucene/index/TestIndexReaderCloneNorms.java (working copy) @@ -203,11 +203,11 @@ IndexReader reader4C = (IndexReader) reader3C.clone(); SegmentReader segmentReader4C = getOnlySegmentReader(reader4C); assertEquals(4, reader3CCNorm.bytesRef().get()); - reader4C.setNorm(5, "field1", 0.33f); + reader4C.setNorm(5, "field1", Similarity.getDefault().encodeNormValue(0.33f)); // generate a cannot update exception in reader1 try { - reader3C.setNorm(1, "field1", 0.99f); + reader3C.setNorm(1, "field1", Similarity.getDefault().encodeNormValue(0.99f)); fail("did not hit expected exception"); } catch (Exception ex) { // expected @@ -223,7 +223,7 @@ IndexReader reader5C = (IndexReader) reader4C.clone(); SegmentReader segmentReader5C = getOnlySegmentReader(reader5C); Norm reader5CCNorm = segmentReader5C.norms.get("field1"); - reader5C.setNorm(5, "field1", 0.7f); + reader5C.setNorm(5, "field1", Similarity.getDefault().encodeNormValue(0.7f)); assertEquals(1, reader5CCNorm.bytesRef().get()); reader5C.close(); @@ -256,8 +256,8 @@ // System.out.println(" and: for "+k+" from "+newNorm+" to "+origNorm); modifiedNorms.set(i, Float.valueOf(newNorm)); modifiedNorms.set(k, Float.valueOf(origNorm)); - ir.setNorm(i, "f" + 1, newNorm); - ir.setNorm(k, "f" + 1, origNorm); + ir.setNorm(i, "f" + 1, Similarity.getDefault().encodeNormValue(newNorm)); + ir.setNorm(k, "f" + 1, Similarity.getDefault().encodeNormValue(origNorm)); // System.out.println("setNorm i: "+i); // break; } Index: lucene/src/test/org/apache/lucene/index/TestOmitNorms.java =================================================================== --- lucene/src/test/org/apache/lucene/index/TestOmitNorms.java (revision 0) +++ lucene/src/test/org/apache/lucene/index/TestOmitNorms.java (revision 0) @@ -0,0 +1,216 @@ +package org.apache.lucene.index; + +/** + * Licensed to the Apache Software Foundation (ASF) under one or more + * contributor license agreements. See the NOTICE file distributed with + * this work for additional information regarding copyright ownership. + * The ASF licenses this file to You under the Apache License, Version 2.0 + * (the "License"); you may not use this file except in compliance with + * the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +import org.apache.lucene.util.LuceneTestCase; +import org.apache.lucene.util._TestUtil; +import org.apache.lucene.analysis.Analyzer; +import org.apache.lucene.analysis.MockAnalyzer; +import org.apache.lucene.document.Document; +import org.apache.lucene.document.Field; +import org.apache.lucene.store.Directory; + +public class TestOmitNorms extends LuceneTestCase { + // Tests whether the DocumentWriter correctly enable the + // omitNorms bit in the FieldInfo + public void testOmitNorms() throws Exception { + Directory ram = newDirectory(); + Analyzer analyzer = new MockAnalyzer(); + IndexWriter writer = new IndexWriter(ram, newIndexWriterConfig( TEST_VERSION_CURRENT, analyzer)); + Document d = new Document(); + + // this field will have norms + Field f1 = newField("f1", "This field has norms", Field.Store.NO, Field.Index.ANALYZED); + d.add(f1); + + // this field will NOT have norms + Field f2 = newField("f2", "This field has NO norms in all docs", Field.Store.NO, Field.Index.ANALYZED); + f2.setOmitNorms(true); + d.add(f2); + + writer.addDocument(d); + writer.optimize(); + // now we add another document which has term freq for field f2 and not for f1 and verify if the SegmentMerger + // keep things constant + d = new Document(); + + // Reverse + f1.setOmitNorms(true); + d.add(f1); + + f2.setOmitNorms(false); + d.add(f2); + + writer.addDocument(d); + + // force merge + writer.optimize(); + // flush + writer.close(); + _TestUtil.checkIndex(ram); + + SegmentReader reader = getOnlySegmentReader(IndexReader.open(ram, false)); + FieldInfos fi = reader.fieldInfos(); + assertTrue("OmitNorms field bit should be set.", fi.fieldInfo("f1").omitNorms); + assertTrue("OmitNorms field bit should be set.", fi.fieldInfo("f2").omitNorms); + + reader.close(); + ram.close(); + } + + // Tests whether merging of docs that have different + // omitNorms for the same field works + public void testMixedMerge() throws Exception { + Directory ram = newDirectory(); + Analyzer analyzer = new MockAnalyzer(); + IndexWriter writer = new IndexWriter( + ram, + newIndexWriterConfig(TEST_VERSION_CURRENT, analyzer). + setMaxBufferedDocs(3). + setMergePolicy(newLogMergePolicy(2)) + ); + Document d = new Document(); + + // this field will have norms + Field f1 = newField("f1", "This field has norms", Field.Store.NO, Field.Index.ANALYZED); + d.add(f1); + + // this field will NOT have norms + Field f2 = newField("f2", "This field has NO norms in all docs", Field.Store.NO, Field.Index.ANALYZED); + f2.setOmitNorms(true); + d.add(f2); + + for(int i=0;i<30;i++) + writer.addDocument(d); + + // now we add another document which has norms for field f2 and not for f1 and verify if the SegmentMerger + // keep things constant + d = new Document(); + + // Reverese + f1.setOmitNorms(true); + d.add(f1); + + f2.setOmitNorms(false); + d.add(f2); + + for(int i=0;i<30;i++) + writer.addDocument(d); + + // force merge + writer.optimize(); + // flush + writer.close(); + + _TestUtil.checkIndex(ram); + + SegmentReader reader = getOnlySegmentReader(IndexReader.open(ram, false)); + FieldInfos fi = reader.fieldInfos(); + assertTrue("OmitNorms field bit should be set.", fi.fieldInfo("f1").omitNorms); + assertTrue("OmitNorms field bit should be set.", fi.fieldInfo("f2").omitNorms); + + reader.close(); + ram.close(); + } + + // Make sure first adding docs that do not omitNorms for + // field X, then adding docs that do omitNorms for that same + // field, + public void testMixedRAM() throws Exception { + Directory ram = newDirectory(); + Analyzer analyzer = new MockAnalyzer(); + IndexWriter writer = new IndexWriter( + ram, + newIndexWriterConfig(TEST_VERSION_CURRENT, analyzer). + setMaxBufferedDocs(10). + setMergePolicy(newLogMergePolicy(2)) + ); + Document d = new Document(); + + // this field will have norms + Field f1 = newField("f1", "This field has norms", Field.Store.NO, Field.Index.ANALYZED); + d.add(f1); + + // this field will NOT have norms + Field f2 = newField("f2", "This field has NO norms in all docs", Field.Store.NO, Field.Index.ANALYZED); + d.add(f2); + + for(int i=0;i<5;i++) + writer.addDocument(d); + + f2.setOmitNorms(true); + + for(int i=0;i<20;i++) + writer.addDocument(d); + + // force merge + writer.optimize(); + + // flush + writer.close(); + + _TestUtil.checkIndex(ram); + + SegmentReader reader = getOnlySegmentReader(IndexReader.open(ram, false)); + FieldInfos fi = reader.fieldInfos(); + assertTrue("OmitNorms field bit should not be set.", !fi.fieldInfo("f1").omitNorms); + assertTrue("OmitNorms field bit should be set.", fi.fieldInfo("f2").omitNorms); + + reader.close(); + ram.close(); + } + + private void assertNoNrm(Directory dir) throws Throwable { + final String[] files = dir.listAll(); + for(int i=0;i it = codecs.values().iterator(); IOException err = null; Index: lucene/src/java/org/apache/lucene/index/IndexReader.java =================================================================== --- lucene/src/java/org/apache/lucene/index/IndexReader.java (revision 1056829) +++ lucene/src/java/org/apache/lucene/index/IndexReader.java (working copy) @@ -931,14 +931,6 @@ */ public abstract byte[] norms(String field) throws IOException; - /** Reads the byte-encoded normalization factor for the named field of every - * document. This is used by the search code to score documents. - * - * @see org.apache.lucene.document.Field#setBoost(float) - */ - public abstract void norms(String field, byte[] bytes, int offset) - throws IOException; - /** Expert: Resets the normalization factor for the named field of the named * document. The norm represents the product of the field's {@link * org.apache.lucene.document.Fieldable#setBoost(float) boost} and its {@link Similarity#lengthNorm(String, @@ -970,26 +962,6 @@ protected abstract void doSetNorm(int doc, String field, byte value) throws CorruptIndexException, IOException; - /** Expert: Resets the normalization factor for the named field of the named - * document. - * - * @see #norms(String) - * @see Similarity#decodeNormValue(byte) - * - * @throws StaleReaderException if the index has changed - * since this reader was opened - * @throws CorruptIndexException if the index is corrupt - * @throws LockObtainFailedException if another writer - * has this index open (write.lock could not - * be obtained) - * @throws IOException if there is a low-level IO error - */ - public void setNorm(int doc, String field, float value) - throws StaleReaderException, CorruptIndexException, LockObtainFailedException, IOException { - ensureOpen(); - setNorm(doc, field, Similarity.getDefault().encodeNormValue(value)); - } - /** Flex API: returns {@link Fields} for this reader. * This method may return null if the reader has no * postings. Index: lucene/src/java/org/apache/lucene/index/CheckIndex.java =================================================================== --- lucene/src/java/org/apache/lucene/index/CheckIndex.java (revision 1056829) +++ lucene/src/java/org/apache/lucene/index/CheckIndex.java (working copy) @@ -548,10 +548,10 @@ if (infoStream != null) { infoStream.print(" test: field norms........."); } - final byte[] b = new byte[reader.maxDoc()]; + byte[] b; for (final String fieldName : fieldNames) { if (reader.hasNorms(fieldName)) { - reader.norms(fieldName, b, 0); + b = reader.norms(fieldName); ++status.totFields; } } Index: lucene/src/java/org/apache/lucene/index/Fields.java =================================================================== --- lucene/src/java/org/apache/lucene/index/Fields.java (revision 1056829) +++ lucene/src/java/org/apache/lucene/index/Fields.java (working copy) @@ -28,6 +28,9 @@ * names. This will not return null. */ public abstract FieldsEnum iterator() throws IOException; + /** Returns true if this field exists */ + public abstract boolean hasField(String field) throws IOException; + /** Get the {@link Terms} for this field. This may return * null if the field does not exist. */ public abstract Terms terms(String field) throws IOException; Index: lucene/src/java/org/apache/lucene/index/MultiNorms.java =================================================================== --- lucene/src/java/org/apache/lucene/index/MultiNorms.java (revision 1056829) +++ lucene/src/java/org/apache/lucene/index/MultiNorms.java (working copy) @@ -22,7 +22,6 @@ import java.util.Arrays; import java.util.List; -import org.apache.lucene.search.Similarity; import org.apache.lucene.util.ReaderUtil; /** @@ -61,26 +60,25 @@ ReaderUtil.gatherSubReaders(leaves, r); int end = 0; for (IndexReader leaf : leaves) { + boolean hasField = false; + Fields fields = leaf.fields(); + if (fields != null && fields.hasField(field)) + hasField = true; + int start = end; - leaf.norms(field, norms, start); + byte leafNorms[] = leaf.norms(field); + if (leafNorms == null) { + if (hasField) // omitted norms + return null; + // doesn't have field, fill bytes + leafNorms = new byte[leaf.maxDoc()]; + Arrays.fill(leafNorms, (byte) 0); + } + + System.arraycopy(leafNorms, 0, norms, start, leafNorms.length); end += leaf.maxDoc(); } return norms; } } - - /** - * Warning: this is heavy! Do not use in a loop, or implement norms() - * in your own reader with this (you should likely cache the result). - */ - public static void norms(IndexReader r, String field, byte[] bytes, int offset) - throws IOException { - // TODO: optimize more maybe - byte[] norms = norms(r, field); - if (norms == null) { - Arrays.fill(bytes, offset, bytes.length, Similarity.getDefault().encodeNormValue(1.0f)); - } else { - System.arraycopy(norms, 0, bytes, offset, r.maxDoc()); - } - } } Index: lucene/src/java/org/apache/lucene/index/codecs/simpletext/SimpleTextFieldsReader.java =================================================================== --- lucene/src/java/org/apache/lucene/index/codecs/simpletext/SimpleTextFieldsReader.java (revision 1056829) +++ lucene/src/java/org/apache/lucene/index/codecs/simpletext/SimpleTextFieldsReader.java (working copy) @@ -513,6 +513,23 @@ public FieldsEnum iterator() throws IOException { return new SimpleTextFieldsEnum(); } + + @Override + public boolean hasField(String field) throws IOException { + Terms terms = termsCache.get(field); + if (terms != null) + return true; + + SimpleTextFieldsEnum fe = (SimpleTextFieldsEnum) iterator(); + String fieldUpto; + while((fieldUpto = fe.next()) != null) { + if (fieldUpto.equals(field)) { + return true; + } + } + + return false; + } private final Map termsCache = new HashMap(); Index: lucene/src/java/org/apache/lucene/index/codecs/preflex/PreFlexFields.java =================================================================== --- lucene/src/java/org/apache/lucene/index/codecs/preflex/PreFlexFields.java (revision 1056829) +++ lucene/src/java/org/apache/lucene/index/codecs/preflex/PreFlexFields.java (working copy) @@ -161,6 +161,11 @@ return preTerms.get(field); } + @Override + public boolean hasField(String field) throws IOException { + return fields.containsKey(field); + } + synchronized private TermInfosReader getTermsDict() { if (tis != null) { return tis; Index: lucene/src/java/org/apache/lucene/index/codecs/PrefixCodedTermsReader.java =================================================================== --- lucene/src/java/org/apache/lucene/index/codecs/PrefixCodedTermsReader.java (revision 1056829) +++ lucene/src/java/org/apache/lucene/index/codecs/PrefixCodedTermsReader.java (working copy) @@ -219,6 +219,11 @@ return fields.get(field); } + @Override + public boolean hasField(String field) throws IOException { + return fields.containsKey(field); + } + // Iterates through all fields private class TermFieldsEnum extends FieldsEnum { final Iterator it; Index: lucene/src/java/org/apache/lucene/index/MultiFields.java =================================================================== --- lucene/src/java/org/apache/lucene/index/MultiFields.java (revision 1056829) +++ lucene/src/java/org/apache/lucene/index/MultiFields.java (working copy) @@ -239,6 +239,19 @@ } @Override + public boolean hasField(String field) throws IOException { + if (terms.containsKey(field)) + return true; + + for (Fields f : subs) { + if (f.hasField(field)) + return true; + } + + return false; + } + + @Override public Terms terms(String field) throws IOException { Terms result = terms.get(field); Index: lucene/src/java/org/apache/lucene/index/SlowMultiReaderWrapper.java =================================================================== --- lucene/src/java/org/apache/lucene/index/SlowMultiReaderWrapper.java (revision 1056829) +++ lucene/src/java/org/apache/lucene/index/SlowMultiReaderWrapper.java (working copy) @@ -18,19 +18,14 @@ */ import java.io.IOException; -import java.util.Arrays; import java.util.HashMap; -import java.util.List; -import java.util.ArrayList; import java.util.Map; -import org.apache.lucene.search.Similarity; import org.apache.lucene.util.Bits; import org.apache.lucene.util.ReaderUtil; // javadoc import org.apache.lucene.index.DirectoryReader; // javadoc import org.apache.lucene.index.MultiReader; // javadoc -import org.apache.lucene.index.IndexReader.ReaderContext; /** * This class forces a composite reader (eg a {@link @@ -88,23 +83,13 @@ return bytes; if (!hasNorms(field)) return null; - + if (normsCache.containsKey(field)) // cached omitNorms, not missing key + return null; + bytes = MultiNorms.norms(in, field); normsCache.put(field, bytes); return bytes; } - - @Override - public synchronized void norms(String field, byte[] bytes, int offset) throws IOException { - // TODO: maybe optimize - ensureOpen(); - byte[] norms = norms(field); - if (norms == null) { - Arrays.fill(bytes, offset, bytes.length, Similarity.getDefault().encodeNormValue(1.0f)); - } else { - System.arraycopy(norms, 0, bytes, offset, maxDoc()); - } - } @Override public ReaderContext getTopReaderContext() { Index: lucene/src/java/org/apache/lucene/index/NormsWriter.java =================================================================== --- lucene/src/java/org/apache/lucene/index/NormsWriter.java (revision 1056829) +++ lucene/src/java/org/apache/lucene/index/NormsWriter.java (working copy) @@ -26,7 +26,6 @@ import java.util.ArrayList; import org.apache.lucene.store.IndexOutput; -import org.apache.lucene.search.Similarity; // TODO FI: norms could actually be stored as doc store @@ -37,7 +36,7 @@ final class NormsWriter extends InvertedDocEndConsumer { - private final byte defaultNorm = Similarity.getDefault().encodeNormValue(1.0f); + private final byte defaultNorm = 0; private FieldInfos fieldInfos; @Override public InvertedDocEndConsumerPerThread addThread(DocInverterPerThread docInverterPerThread) { @@ -62,6 +61,10 @@ final Map> byField = new HashMap>(); + if (!fieldInfos.hasNorms()) { + return; + } + // Typically, each thread will have encountered the same // field. So first we collate by field, ie, all // per-thread field instances that correspond to the Index: lucene/contrib/instantiated/src/test/org/apache/lucene/store/instantiated/TestEmptyIndex.java =================================================================== --- lucene/contrib/instantiated/src/test/org/apache/lucene/store/instantiated/TestEmptyIndex.java (revision 1056829) +++ lucene/contrib/instantiated/src/test/org/apache/lucene/store/instantiated/TestEmptyIndex.java (working copy) @@ -71,12 +71,6 @@ byte[] norms = MultiNorms.norms(r, "foo"); if (norms != null) { assertEquals(0, norms.length); - norms = new byte[10]; - Arrays.fill(norms, (byte)10); - r.norms("foo", norms, 10); - for (byte b : norms) { - assertEquals((byte)10, b); - } } } Index: lucene/contrib/instantiated/src/test/org/apache/lucene/store/instantiated/TestIndicesEquals.java =================================================================== --- lucene/contrib/instantiated/src/test/org/apache/lucene/store/instantiated/TestIndicesEquals.java (revision 1056829) +++ lucene/contrib/instantiated/src/test/org/apache/lucene/store/instantiated/TestIndicesEquals.java (working copy) @@ -351,35 +351,6 @@ for (int i = 0; i < aprioriNorms.length; i++) { assertEquals("norms does not equals for field " + field + " in document " + i, aprioriNorms[i], testNorms[i]); } - - // test norms as used by multireader - - aprioriNorms = new byte[aprioriReader.maxDoc()]; - MultiNorms.norms(aprioriReader, (String) field, aprioriNorms, 0); - - testNorms = new byte[testReader.maxDoc()]; - MultiNorms.norms(testReader, (String) field, testNorms, 0); - - assertEquals(aprioriNorms.length, testNorms.length); - - for (int i = 0; i < aprioriNorms.length; i++) { - assertEquals("norms does not equals for field " + field + " in document " + i, aprioriNorms[i], testNorms[i]); - } - - - // test norms as used by multireader - - aprioriNorms = new byte[aprioriReader.maxDoc() + 10]; - MultiNorms.norms(aprioriReader, (String) field, aprioriNorms, 10); - - testNorms = new byte[testReader.maxDoc() + 10]; - MultiNorms.norms(testReader, (String) field, testNorms, 10); - - assertEquals(aprioriNorms.length, testNorms.length); - - for (int i = 0; i < aprioriNorms.length; i++) { - assertEquals("norms does not equals for field " + field + " in document " + i, aprioriNorms[i], testNorms[i]); - } } } Index: lucene/contrib/instantiated/src/java/org/apache/lucene/store/instantiated/InstantiatedIndexReader.java =================================================================== --- lucene/contrib/instantiated/src/java/org/apache/lucene/store/instantiated/InstantiatedIndexReader.java (revision 1056829) +++ lucene/contrib/instantiated/src/java/org/apache/lucene/store/instantiated/InstantiatedIndexReader.java (working copy) @@ -334,15 +334,6 @@ } @Override - public void norms(String field, byte[] bytes, int offset) throws IOException { - byte[] norms = getIndex().getNormsByFieldNameAndDocumentNumber().get(field); - if (norms == null) { - return; - } - System.arraycopy(norms, 0, bytes, offset, norms.length); - } - - @Override protected void doSetNorm(int doc, String field, byte value) throws IOException { if (uncommittedNormsByFieldNameAndDocumentNumber == null) { uncommittedNormsByFieldNameAndDocumentNumber = new HashMap>(getIndex().getNormsByFieldNameAndDocumentNumber().size()); @@ -399,6 +390,20 @@ } }; } + + @Override + public boolean hasField(String field) throws IOException { + final InstantiatedTerm[] orderedTerms = getIndex().getOrderedTerms(); + int i = Arrays.binarySearch(orderedTerms, new Term(field), InstantiatedTerm.termComparator); + if (i < 0) { + i = -i - 1; + } + if (i >= orderedTerms.length || !orderedTerms[i].field().equals(field)) { + // field does not exist + return false; + } + return true; + } @Override public Terms terms(final String field) { Index: lucene/contrib/instantiated/src/java/org/apache/lucene/store/instantiated/InstantiatedIndexWriter.java =================================================================== --- lucene/contrib/instantiated/src/java/org/apache/lucene/store/instantiated/InstantiatedIndexWriter.java (revision 1056829) +++ lucene/contrib/instantiated/src/java/org/apache/lucene/store/instantiated/InstantiatedIndexWriter.java (working copy) @@ -201,9 +201,9 @@ byte[] oldNorms = index.getNormsByFieldNameAndDocumentNumber().get(field); if (oldNorms != null) { System.arraycopy(oldNorms, 0, norms, 0, oldNorms.length); - Arrays.fill(norms, oldNorms.length, norms.length, similarity.encodeNormValue(1.0f)); + Arrays.fill(norms, oldNorms.length, norms.length, (byte) 0); } else { - Arrays.fill(norms, 0, norms.length, similarity.encodeNormValue(1.0f)); + Arrays.fill(norms, 0, norms.length, (byte) 0); } normsByFieldNameAndDocumentNumber.put(field, norms); fieldNames.remove(field); @@ -211,7 +211,7 @@ for (String field : fieldNames) { //System.out.println(field); byte[] norms = new byte[index.getDocumentsByNumber().length + termDocumentInformationFactoryByDocument.size()]; - Arrays.fill(norms, 0, norms.length, similarity.encodeNormValue(1.0f)); + Arrays.fill(norms, 0, norms.length, (byte) 0); normsByFieldNameAndDocumentNumber.put(field, norms); } fieldNames.clear(); Index: lucene/contrib/memory/src/java/org/apache/lucene/index/memory/MemoryIndex.java =================================================================== --- lucene/contrib/memory/src/java/org/apache/lucene/index/memory/MemoryIndex.java (revision 1056829) +++ lucene/contrib/memory/src/java/org/apache/lucene/index/memory/MemoryIndex.java (working copy) @@ -800,6 +800,11 @@ } @Override + public boolean hasField(String field) throws IOException { + return Arrays.binarySearch(sortedFields, field, termComparator) >= 0; + } + + @Override public Terms terms(final String field) { int i = Arrays.binarySearch(sortedFields, field, termComparator); if (i < 0) { @@ -1174,13 +1179,6 @@ } @Override - public void norms(String fieldName, byte[] bytes, int offset) { - if (DEBUG) System.err.println("MemoryIndexReader.norms*: " + fieldName); - byte[] norms = norms(fieldName); - System.arraycopy(norms, 0, bytes, offset, norms.length); - } - - @Override protected void doSetNorm(int doc, String fieldName, byte value) { throw new UnsupportedOperationException(); }