Index: lucene/src/test/org/apache/lucene/TestMergeSchedulerExternal.java =================================================================== --- lucene/src/test/org/apache/lucene/TestMergeSchedulerExternal.java (revision 995772) +++ lucene/src/test/org/apache/lucene/TestMergeSchedulerExternal.java (working copy) @@ -86,7 +86,7 @@ dir.failOn(new FailOnlyOnMerge()); Document doc = new Document(); - Field idField = new Field("id", "", Field.Store.YES, Field.Index.NOT_ANALYZED); + Field idField = newField("id", "", Field.Store.YES, Field.Index.NOT_ANALYZED); doc.add(idField); IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig( Index: lucene/src/test/org/apache/lucene/queryParser/TestMultiFieldQueryParser.java =================================================================== --- lucene/src/test/org/apache/lucene/queryParser/TestMultiFieldQueryParser.java (revision 995772) +++ lucene/src/test/org/apache/lucene/queryParser/TestMultiFieldQueryParser.java (working copy) @@ -284,7 +284,7 @@ Directory ramDir = newDirectory(); IndexWriter iw = new IndexWriter(ramDir, newIndexWriterConfig(TEST_VERSION_CURRENT, analyzer)); Document doc = new Document(); - doc.add(new Field("body", "blah the footest blah", Field.Store.NO, Field.Index.ANALYZED)); + doc.add(newField("body", "blah the footest blah", Field.Store.NO, Field.Index.ANALYZED)); iw.addDocument(doc); iw.close(); Index: lucene/src/test/org/apache/lucene/queryParser/TestQueryParser.java =================================================================== --- lucene/src/test/org/apache/lucene/queryParser/TestQueryParser.java (revision 995772) +++ lucene/src/test/org/apache/lucene/queryParser/TestQueryParser.java (working copy) @@ -571,7 +571,7 @@ Directory ramDir = newDirectory(); IndexWriter iw = new IndexWriter(ramDir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(MockTokenizer.WHITESPACE, false))); Document doc = new Document(); - doc.add(new Field("content","\u0633\u0627\u0628", + doc.add(newField("content","\u0633\u0627\u0628", Field.Store.YES, Field.Index.NOT_ANALYZED)); iw.addDocument(doc); iw.close(); @@ -1131,13 +1131,13 @@ assertEquals(expected, hits.length); } - private static void addDateDoc(String content, int year, int month, + private void addDateDoc(String content, int year, int month, int day, int hour, int minute, int second, IndexWriter iw) throws IOException { Document d = new Document(); - d.add(new Field("f", content, Field.Store.YES, Field.Index.ANALYZED)); + d.add(newField("f", content, Field.Store.YES, Field.Index.ANALYZED)); Calendar cal = Calendar.getInstance(Locale.ENGLISH); cal.set(year, month-1, day, hour, minute, second); - d.add(new Field("date", DateField.dateToString(cal.getTime()), Field.Store.YES, Field.Index.NOT_ANALYZED)); + d.add(newField("date", DateField.dateToString(cal.getTime()), Field.Store.YES, Field.Index.NOT_ANALYZED)); iw.addDocument(d); } @@ -1155,7 +1155,7 @@ Analyzer a = new MockAnalyzer(MockTokenizer.SIMPLE, true, MockTokenFilter.ENGLISH_STOPSET, true); IndexWriter w = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, a)); Document doc = new Document(); - doc.add(new Field("f", "the wizard of ozzy", Field.Store.NO, Field.Index.ANALYZED)); + doc.add(newField("f", "the wizard of ozzy", Field.Store.NO, Field.Index.ANALYZED)); w.addDocument(doc); IndexReader r = w.getReader(); w.close(); Index: lucene/src/test/org/apache/lucene/TestExternalCodecs.java =================================================================== --- lucene/src/test/org/apache/lucene/TestExternalCodecs.java (revision 995772) +++ lucene/src/test/org/apache/lucene/TestExternalCodecs.java (working copy) @@ -607,11 +607,11 @@ w.setMergeFactor(3); Document doc = new Document(); // uses default codec: - doc.add(new Field("field1", "this field uses the standard codec as the test", Field.Store.NO, Field.Index.ANALYZED)); + doc.add(newField("field1", "this field uses the standard codec as the test", Field.Store.NO, Field.Index.ANALYZED)); // uses pulsing codec: - doc.add(new Field("field2", "this field uses the pulsing codec as the test", Field.Store.NO, Field.Index.ANALYZED)); + doc.add(newField("field2", "this field uses the pulsing codec as the test", Field.Store.NO, Field.Index.ANALYZED)); - Field idField = new Field("id", "", Field.Store.NO, Field.Index.NOT_ANALYZED); + Field idField = newField("id", "", Field.Store.NO, Field.Index.NOT_ANALYZED); doc.add(idField); for(int i=0;i> docs = new ArrayList>(); Document d = new Document(); - Field f = new Field("f", "", Field.Store.NO, Field.Index.ANALYZED); + Field f = newField("f", "", Field.Store.NO, Field.Index.ANALYZED); d.add(f); Random r = random; Index: lucene/src/test/org/apache/lucene/search/TestPositionIncrement.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestPositionIncrement.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestPositionIncrement.java (working copy) @@ -93,7 +93,7 @@ Directory store = newDirectory(); RandomIndexWriter writer = new RandomIndexWriter(random, store, analyzer); Document d = new Document(); - d.add(new Field("field", "bogus", Field.Store.YES, Field.Index.ANALYZED)); + d.add(newField("field", "bogus", Field.Store.YES, Field.Index.ANALYZED)); writer.addDocument(d); IndexReader reader = writer.getReader(); writer.close(); Index: lucene/src/test/org/apache/lucene/search/TestTermRangeQuery.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestTermRangeQuery.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestTermRangeQuery.java (working copy) @@ -355,8 +355,8 @@ private void insertDoc(IndexWriter writer, String content) throws IOException { Document doc = new Document(); - doc.add(new Field("id", "id" + docCount, Field.Store.YES, Field.Index.NOT_ANALYZED)); - doc.add(new Field("content", content, Field.Store.NO, Field.Index.ANALYZED)); + doc.add(newField("id", "id" + docCount, Field.Store.YES, Field.Index.NOT_ANALYZED)); + doc.add(newField("content", content, Field.Store.NO, Field.Index.ANALYZED)); writer.addDocument(doc); docCount++; Index: lucene/src/test/org/apache/lucene/search/TestBooleanMinShouldMatch.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestBooleanMinShouldMatch.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestBooleanMinShouldMatch.java (working copy) @@ -56,10 +56,10 @@ for (int i = 0; i < data.length; i++) { Document doc = new Document(); - doc.add(new Field("id", String.valueOf(i), Field.Store.YES, Field.Index.NOT_ANALYZED));//Field.Keyword("id",String.valueOf(i))); - doc.add(new Field("all", "all", Field.Store.YES, Field.Index.NOT_ANALYZED));//Field.Keyword("all","all")); + doc.add(newField("id", String.valueOf(i), Field.Store.YES, Field.Index.NOT_ANALYZED));//Field.Keyword("id",String.valueOf(i))); + doc.add(newField("all", "all", Field.Store.YES, Field.Index.NOT_ANALYZED));//Field.Keyword("all","all")); if (null != data[i]) { - doc.add(new Field("data", data[i], Field.Store.YES, Field.Index.ANALYZED));//Field.Text("data",data[i])); + doc.add(newField("data", data[i], Field.Store.YES, Field.Index.ANALYZED));//Field.Text("data",data[i])); } w.addDocument(doc); } Index: lucene/src/test/org/apache/lucene/search/TestBooleanOr.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestBooleanOr.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestBooleanOr.java (working copy) @@ -143,12 +143,12 @@ // Document d = new Document(); - d.add(new Field( + d.add(newField( FIELD_T, "Optimize not deleting all files", Field.Store.YES, Field.Index.ANALYZED)); - d.add(new Field( + d.add(newField( FIELD_C, "Deleted When I run an optimize in our production environment.", Field.Store.YES, Index: lucene/src/test/org/apache/lucene/search/TestDateSort.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestDateSort.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestDateSort.java (working copy) @@ -106,16 +106,16 @@ assertEquals(Arrays.asList(expectedOrder), Arrays.asList(actualOrder)); } - private static Document createDocument(String text, long time) { + private Document createDocument(String text, long time) { Document document = new Document(); // Add the text field. - Field textField = new Field(TEXT_FIELD, text, Field.Store.YES, Field.Index.ANALYZED); + Field textField = newField(TEXT_FIELD, text, Field.Store.YES, Field.Index.ANALYZED); document.add(textField); // Add the date/time field. String dateTimeString = DateTools.timeToString(time, DateTools.Resolution.SECOND); - Field dateTimeField = new Field(DATE_TIME_FIELD, dateTimeString, Field.Store.YES, + Field dateTimeField = newField(DATE_TIME_FIELD, dateTimeString, Field.Store.YES, Field.Index.NOT_ANALYZED); document.add(dateTimeField); Index: lucene/src/test/org/apache/lucene/search/TestSort.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestSort.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestSort.java (working copy) @@ -1099,12 +1099,12 @@ IndexWriter w = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer())); Document doc = new Document(); - doc.add(new Field("f", "", Field.Store.NO, Field.Index.NOT_ANALYZED)); - doc.add(new Field("t", "1", Field.Store.NO, Field.Index.NOT_ANALYZED)); + doc.add(newField("f", "", Field.Store.NO, Field.Index.NOT_ANALYZED)); + doc.add(newField("t", "1", Field.Store.NO, Field.Index.NOT_ANALYZED)); w.addDocument(doc); w.commit(); doc = new Document(); - doc.add(new Field("t", "1", Field.Store.NO, Field.Index.NOT_ANALYZED)); + doc.add(newField("t", "1", Field.Store.NO, Field.Index.NOT_ANALYZED)); w.addDocument(doc); IndexReader r = w.getReader(); Index: lucene/src/test/org/apache/lucene/search/TestMultiTermConstantScore.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestMultiTermConstantScore.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestMultiTermConstantScore.java (working copy) @@ -59,13 +59,13 @@ for (int i = 0; i < data.length; i++) { Document doc = new Document(); - doc.add(new Field("id", String.valueOf(i), Field.Store.YES, + doc.add(newField("id", String.valueOf(i), Field.Store.YES, Field.Index.NOT_ANALYZED));// Field.Keyword("id",String.valueOf(i))); doc - .add(new Field("all", "all", Field.Store.YES, + .add(newField("all", "all", Field.Store.YES, Field.Index.NOT_ANALYZED));// Field.Keyword("all","all")); if (null != data[i]) { - doc.add(new Field("data", data[i], Field.Store.YES, + doc.add(newField("data", data[i], Field.Store.YES, Field.Index.ANALYZED));// Field.Text("data",data[i])); } writer.addDocument(doc); @@ -611,10 +611,10 @@ Directory farsiIndex = newDirectory(); RandomIndexWriter writer = new RandomIndexWriter(random, farsiIndex, new MockAnalyzer(MockTokenizer.SIMPLE, true)); Document doc = new Document(); - doc.add(new Field("content", "\u0633\u0627\u0628", Field.Store.YES, + doc.add(newField("content", "\u0633\u0627\u0628", Field.Store.YES, Field.Index.NOT_ANALYZED)); doc - .add(new Field("body", "body", Field.Store.YES, + .add(newField("body", "body", Field.Store.YES, Field.Index.NOT_ANALYZED)); writer.addDocument(doc); @@ -656,9 +656,9 @@ String[] words = { "H\u00D8T", "H\u00C5T", "MAND" }; for (int docnum = 0 ; docnum < words.length ; ++docnum) { Document doc = new Document(); - doc.add(new Field("content", words[docnum], + doc.add(newField("content", words[docnum], Field.Store.YES, Field.Index.NOT_ANALYZED)); - doc.add(new Field("body", "body", + doc.add(newField("body", "body", Field.Store.YES, Field.Index.NOT_ANALYZED)); writer.addDocument(doc); } Index: lucene/src/test/org/apache/lucene/search/TestWildcardRandom.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestWildcardRandom.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestWildcardRandom.java (working copy) @@ -47,7 +47,7 @@ RandomIndexWriter writer = new RandomIndexWriter(random, dir); Document doc = new Document(); - Field field = new Field("field", "", Field.Store.NO, Field.Index.ANALYZED); + Field field = newField("field", "", Field.Store.NO, Field.Index.ANALYZED); doc.add(field); NumberFormat df = new DecimalFormat("0000", new DecimalFormatSymbols(Locale.ENGLISH)); Index: lucene/src/test/org/apache/lucene/search/payloads/TestPayloadNearQuery.java =================================================================== --- lucene/src/test/org/apache/lucene/search/payloads/TestPayloadNearQuery.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/payloads/TestPayloadNearQuery.java (working copy) @@ -113,9 +113,9 @@ //writer.infoStream = System.out; for (int i = 0; i < 1000; i++) { Document doc = new Document(); - doc.add(new Field("field", English.intToEnglish(i), Field.Store.YES, Field.Index.ANALYZED)); + doc.add(newField("field", English.intToEnglish(i), Field.Store.YES, Field.Index.ANALYZED)); String txt = English.intToEnglish(i) +' '+English.intToEnglish(i+1); - doc.add(new Field("field2", txt, Field.Store.YES, Field.Index.ANALYZED)); + doc.add(newField("field2", txt, Field.Store.YES, Field.Index.ANALYZED)); writer.addDocument(doc); } reader = writer.getReader(); Index: lucene/src/test/org/apache/lucene/search/payloads/TestPayloadTermQuery.java =================================================================== --- lucene/src/test/org/apache/lucene/search/payloads/TestPayloadTermQuery.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/payloads/TestPayloadTermQuery.java (working copy) @@ -118,11 +118,11 @@ //writer.infoStream = System.out; for (int i = 0; i < 1000; i++) { Document doc = new Document(); - Field noPayloadField = new Field(PayloadHelper.NO_PAYLOAD_FIELD, English.intToEnglish(i), Field.Store.YES, Field.Index.ANALYZED); + Field noPayloadField = newField(PayloadHelper.NO_PAYLOAD_FIELD, English.intToEnglish(i), Field.Store.YES, Field.Index.ANALYZED); //noPayloadField.setBoost(0); doc.add(noPayloadField); - doc.add(new Field("field", English.intToEnglish(i), Field.Store.YES, Field.Index.ANALYZED)); - doc.add(new Field("multiField", English.intToEnglish(i) + " " + English.intToEnglish(i), Field.Store.YES, Field.Index.ANALYZED)); + doc.add(newField("field", English.intToEnglish(i), Field.Store.YES, Field.Index.ANALYZED)); + doc.add(newField("multiField", English.intToEnglish(i) + " " + English.intToEnglish(i), Field.Store.YES, Field.Index.ANALYZED)); writer.addDocument(doc); } reader = writer.getReader(); Index: lucene/src/test/org/apache/lucene/search/TestBooleanScorer.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestBooleanScorer.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestBooleanScorer.java (working copy) @@ -46,7 +46,7 @@ RandomIndexWriter writer = new RandomIndexWriter(random, directory); for (int i = 0; i < values.length; i++) { Document doc = new Document(); - doc.add(new Field(FIELD, values[i], Field.Store.YES, Field.Index.NOT_ANALYZED)); + doc.add(newField(FIELD, values[i], Field.Store.YES, Field.Index.NOT_ANALYZED)); writer.addDocument(doc); } IndexReader ir = writer.getReader(); Index: lucene/src/test/org/apache/lucene/search/TestDocIdSet.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestDocIdSet.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestDocIdSet.java (working copy) @@ -103,7 +103,7 @@ Directory dir = newDirectory(); RandomIndexWriter writer = new RandomIndexWriter(random, dir); Document doc = new Document(); - doc.add(new Field("c", "val", Store.NO, Index.NOT_ANALYZED_NO_NORMS)); + doc.add(newField("c", "val", Store.NO, Index.NOT_ANALYZED_NO_NORMS)); writer.addDocument(doc); IndexReader reader = writer.getReader(); writer.close(); Index: lucene/src/test/org/apache/lucene/search/spans/TestNearSpansOrdered.java =================================================================== --- lucene/src/test/org/apache/lucene/search/spans/TestNearSpansOrdered.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/spans/TestNearSpansOrdered.java (working copy) @@ -57,7 +57,7 @@ RandomIndexWriter writer= new RandomIndexWriter(random, directory); for (int i = 0; i < docFields.length; i++) { Document doc = new Document(); - doc.add(new Field(FIELD, docFields[i], Field.Store.NO, Field.Index.ANALYZED)); + doc.add(newField(FIELD, docFields[i], Field.Store.NO, Field.Index.ANALYZED)); writer.addDocument(doc); } reader = writer.getReader(); Index: lucene/src/test/org/apache/lucene/search/spans/TestBasics.java =================================================================== --- lucene/src/test/org/apache/lucene/search/spans/TestBasics.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/spans/TestBasics.java (working copy) @@ -65,7 +65,7 @@ //writer.infoStream = System.out; for (int i = 0; i < 1000; i++) { Document doc = new Document(); - doc.add(new Field("field", English.intToEnglish(i), Field.Store.YES, Field.Index.ANALYZED)); + doc.add(newField("field", English.intToEnglish(i), Field.Store.YES, Field.Index.ANALYZED)); writer.addDocument(doc); } reader = writer.getReader(); Index: lucene/src/test/org/apache/lucene/search/spans/TestSpans.java =================================================================== --- lucene/src/test/org/apache/lucene/search/spans/TestSpans.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/spans/TestSpans.java (working copy) @@ -53,7 +53,7 @@ RandomIndexWriter writer= new RandomIndexWriter(random, directory); for (int i = 0; i < docFields.length; i++) { Document doc = new Document(); - doc.add(new Field(field, docFields[i], Field.Store.YES, Field.Index.ANALYZED)); + doc.add(newField(field, docFields[i], Field.Store.YES, Field.Index.ANALYZED)); writer.addDocument(doc); } reader = writer.getReader(); @@ -433,8 +433,8 @@ // LUCENE-1404 private void addDoc(IndexWriter writer, String id, String text) throws IOException { final Document doc = new Document(); - doc.add( new Field("id", id, Field.Store.YES, Field.Index.NOT_ANALYZED) ); - doc.add( new Field("text", text, Field.Store.YES, Field.Index.ANALYZED) ); + doc.add( newField("id", id, Field.Store.YES, Field.Index.NOT_ANALYZED) ); + doc.add( newField("text", text, Field.Store.YES, Field.Index.ANALYZED) ); writer.addDocument(doc); } Index: lucene/src/test/org/apache/lucene/search/spans/TestFieldMaskingSpanQuery.java =================================================================== --- lucene/src/test/org/apache/lucene/search/spans/TestFieldMaskingSpanQuery.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/spans/TestFieldMaskingSpanQuery.java (working copy) @@ -43,8 +43,8 @@ return doc; } - protected static Field field(String name, String value) { - return new Field(name, value, Field.Store.NO, Field.Index.ANALYZED); + protected Field field(String name, String value) { + return newField(name, value, Field.Store.NO, Field.Index.ANALYZED); } protected IndexSearcher searcher; Index: lucene/src/test/org/apache/lucene/search/spans/TestSpansAdvanced.java =================================================================== --- lucene/src/test/org/apache/lucene/search/spans/TestSpansAdvanced.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/spans/TestSpansAdvanced.java (working copy) @@ -89,9 +89,9 @@ final String text) throws IOException { final Document document = new Document(); - document.add(new Field(FIELD_ID, id, Field.Store.YES, + document.add(newField(FIELD_ID, id, Field.Store.YES, Field.Index.NOT_ANALYZED)); - document.add(new Field(FIELD_TEXT, text, Field.Store.YES, + document.add(newField(FIELD_TEXT, text, Field.Store.YES, Field.Index.ANALYZED)); writer.addDocument(document); } Index: lucene/src/test/org/apache/lucene/search/spans/TestPayloadSpans.java =================================================================== --- lucene/src/test/org/apache/lucene/search/spans/TestPayloadSpans.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/spans/TestPayloadSpans.java (working copy) @@ -118,7 +118,7 @@ newIndexWriterConfig(TEST_VERSION_CURRENT, new PayloadAnalyzer()).setSimilarity(similarity)); Document doc = new Document(); - doc.add(new Field(PayloadHelper.FIELD, "one two three one four three", + doc.add(newField(PayloadHelper.FIELD, "one two three one four three", Field.Store.YES, Field.Index.ANALYZED)); writer.addDocument(doc); IndexReader reader = writer.getReader(); @@ -377,7 +377,7 @@ newIndexWriterConfig(TEST_VERSION_CURRENT, new PayloadAnalyzer()).setSimilarity(similarity)); Document doc = new Document(); - doc.add(new Field(PayloadHelper.FIELD,"xx rr yy mm pp", Field.Store.YES, Field.Index.ANALYZED)); + doc.add(newField(PayloadHelper.FIELD,"xx rr yy mm pp", Field.Store.YES, Field.Index.ANALYZED)); writer.addDocument(doc); IndexReader reader = writer.getReader(); @@ -440,7 +440,7 @@ for(int i = 0; i < docs.length; i++) { doc = new Document(); String docText = docs[i]; - doc.add(new Field(PayloadHelper.FIELD,docText, Field.Store.YES, Field.Index.ANALYZED)); + doc.add(newField(PayloadHelper.FIELD,docText, Field.Store.YES, Field.Index.ANALYZED)); writer.addDocument(doc); } Index: lucene/src/test/org/apache/lucene/search/TestMultiSearcher.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestMultiSearcher.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestMultiSearcher.java (working copy) @@ -63,23 +63,23 @@ // creating a document to store Document lDoc = new Document(); - lDoc.add(new Field("fulltext", "Once upon a time.....", Field.Store.YES, Field.Index.ANALYZED)); - lDoc.add(new Field("id", "doc1", Field.Store.YES, Field.Index.NOT_ANALYZED)); - lDoc.add(new Field("handle", "1", Field.Store.YES, Field.Index.NOT_ANALYZED)); + lDoc.add(newField("fulltext", "Once upon a time.....", Field.Store.YES, Field.Index.ANALYZED)); + lDoc.add(newField("id", "doc1", Field.Store.YES, Field.Index.NOT_ANALYZED)); + lDoc.add(newField("handle", "1", Field.Store.YES, Field.Index.NOT_ANALYZED)); // creating a document to store Document lDoc2 = new Document(); - lDoc2.add(new Field("fulltext", "in a galaxy far far away.....", + lDoc2.add(newField("fulltext", "in a galaxy far far away.....", Field.Store.YES, Field.Index.ANALYZED)); - lDoc2.add(new Field("id", "doc2", Field.Store.YES, Field.Index.NOT_ANALYZED)); - lDoc2.add(new Field("handle", "1", Field.Store.YES, Field.Index.NOT_ANALYZED)); + lDoc2.add(newField("id", "doc2", Field.Store.YES, Field.Index.NOT_ANALYZED)); + lDoc2.add(newField("handle", "1", Field.Store.YES, Field.Index.NOT_ANALYZED)); // creating a document to store Document lDoc3 = new Document(); - lDoc3.add(new Field("fulltext", "a bizarre bug manifested itself....", + lDoc3.add(newField("fulltext", "a bizarre bug manifested itself....", Field.Store.YES, Field.Index.ANALYZED)); - lDoc3.add(new Field("id", "doc3", Field.Store.YES, Field.Index.NOT_ANALYZED)); - lDoc3.add(new Field("handle", "1", Field.Store.YES, Field.Index.NOT_ANALYZED)); + lDoc3.add(newField("id", "doc3", Field.Store.YES, Field.Index.NOT_ANALYZED)); + lDoc3.add(newField("handle", "1", Field.Store.YES, Field.Index.NOT_ANALYZED)); // creating an index writer for the first index IndexWriter writerA = new IndexWriter(indexStoreA, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer())); @@ -205,19 +205,19 @@ indexStoreB.close(); } - private static Document createDocument(String contents1, String contents2) { + private Document createDocument(String contents1, String contents2) { Document document=new Document(); - document.add(new Field("contents", contents1, Field.Store.YES, Field.Index.NOT_ANALYZED)); - document.add(new Field("other", "other contents", Field.Store.YES, Field.Index.NOT_ANALYZED)); + document.add(newField("contents", contents1, Field.Store.YES, Field.Index.NOT_ANALYZED)); + document.add(newField("other", "other contents", Field.Store.YES, Field.Index.NOT_ANALYZED)); if (contents2!=null) { - document.add(new Field("contents", contents2, Field.Store.YES, Field.Index.NOT_ANALYZED)); + document.add(newField("contents", contents2, Field.Store.YES, Field.Index.NOT_ANALYZED)); } return document; } - private static void initIndex(Random random, Directory directory, int nDocs, boolean create, String contents2) throws IOException { + private void initIndex(Random random, Directory directory, int nDocs, boolean create, String contents2) throws IOException { IndexWriter indexWriter=null; try { Index: lucene/src/test/org/apache/lucene/search/TestSubScorerFreqs.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestSubScorerFreqs.java (revision 995688) +++ lucene/src/test/org/apache/lucene/search/TestSubScorerFreqs.java (working copy) @@ -41,17 +41,18 @@ @BeforeClass public static void makeIndex() throws Exception { dir = new RAMDirectory(); + Random random = newStaticRandom(TestSubScorerFreqs.class); RandomIndexWriter w = new RandomIndexWriter( - newStaticRandom(TestSubScorerFreqs.class), dir); + random, dir); // make sure we have more than one segment occationally for (int i = 0; i < 31 * RANDOM_MULTIPLIER; i++) { Document doc = new Document(); - doc.add(new Field("f", "a b c d b c d c d d", Field.Store.NO, + doc.add(newField(random, "f", "a b c d b c d c d d", Field.Store.NO, Field.Index.ANALYZED)); w.addDocument(doc); doc = new Document(); - doc.add(new Field("f", "a b c d", Field.Store.NO, Field.Index.ANALYZED)); + doc.add(newField(random, "f", "a b c d", Field.Store.NO, Field.Index.ANALYZED)); w.addDocument(doc); } Index: lucene/src/test/org/apache/lucene/search/TestQueryWrapperFilter.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestQueryWrapperFilter.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestQueryWrapperFilter.java (working copy) @@ -34,7 +34,7 @@ Directory dir = newDirectory(); RandomIndexWriter writer = new RandomIndexWriter(random, dir); Document doc = new Document(); - doc.add(new Field("field", "value", Store.NO, Index.ANALYZED)); + doc.add(newField("field", "value", Store.NO, Index.ANALYZED)); writer.addDocument(doc); IndexReader reader = writer.getReader(); writer.close(); Index: lucene/src/test/org/apache/lucene/search/TestSimpleExplanations.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestSimpleExplanations.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestSimpleExplanations.java (working copy) @@ -309,11 +309,11 @@ Directory indexStoreB = newDirectory(); Document lDoc = new Document(); - lDoc.add(new Field("handle", "1 2", Field.Store.YES, Field.Index.ANALYZED)); + lDoc.add(newField("handle", "1 2", Field.Store.YES, Field.Index.ANALYZED)); Document lDoc2 = new Document(); - lDoc2.add(new Field("handle", "1 2", Field.Store.YES, Field.Index.ANALYZED)); + lDoc2.add(newField("handle", "1 2", Field.Store.YES, Field.Index.ANALYZED)); Document lDoc3 = new Document(); - lDoc3.add(new Field("handle", "1 2", Field.Store.YES, Field.Index.ANALYZED)); + lDoc3.add(newField("handle", "1 2", Field.Store.YES, Field.Index.ANALYZED)); IndexWriter writerA = new IndexWriter(indexStoreA, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer())); Index: lucene/src/test/org/apache/lucene/search/TestMultiValuedNumericRangeQuery.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestMultiValuedNumericRangeQuery.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestMultiValuedNumericRangeQuery.java (working copy) @@ -47,7 +47,7 @@ Document doc = new Document(); for (int m=0, c=random.nextInt(10); m<=c; m++) { int value = random.nextInt(Integer.MAX_VALUE); - doc.add(new Field("asc", format.format(value), Field.Store.NO, Field.Index.NOT_ANALYZED)); + doc.add(newField("asc", format.format(value), Field.Store.NO, Field.Index.NOT_ANALYZED)); doc.add(new NumericField("trie", Field.Store.NO, true).setIntValue(value)); } writer.addDocument(doc); Index: lucene/src/test/org/apache/lucene/search/TestTermScorer.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestTermScorer.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestTermScorer.java (working copy) @@ -52,7 +52,7 @@ for (int i = 0; i < values.length; i++) { Document doc = new Document(); doc - .add(new Field(FIELD, values[i], Field.Store.YES, + .add(newField(FIELD, values[i], Field.Store.YES, Field.Index.ANALYZED)); writer.addDocument(doc); } Index: lucene/src/test/org/apache/lucene/search/TestPhrasePrefixQuery.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestPhrasePrefixQuery.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestPhrasePrefixQuery.java (working copy) @@ -50,15 +50,15 @@ Document doc3 = new Document(); Document doc4 = new Document(); Document doc5 = new Document(); - doc1.add(new Field("body", "blueberry pie", Field.Store.YES, + doc1.add(newField("body", "blueberry pie", Field.Store.YES, Field.Index.ANALYZED)); - doc2.add(new Field("body", "blueberry strudel", Field.Store.YES, + doc2.add(newField("body", "blueberry strudel", Field.Store.YES, Field.Index.ANALYZED)); - doc3.add(new Field("body", "blueberry pizza", Field.Store.YES, + doc3.add(newField("body", "blueberry pizza", Field.Store.YES, Field.Index.ANALYZED)); - doc4.add(new Field("body", "blueberry chewing gum", Field.Store.YES, + doc4.add(newField("body", "blueberry chewing gum", Field.Store.YES, Field.Index.ANALYZED)); - doc5.add(new Field("body", "piccadilly circus", Field.Store.YES, + doc5.add(newField("body", "piccadilly circus", Field.Store.YES, Field.Index.ANALYZED)); writer.addDocument(doc1); writer.addDocument(doc2); Index: lucene/src/test/org/apache/lucene/search/TestSetNorm.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestSetNorm.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestSetNorm.java (working copy) @@ -42,7 +42,7 @@ IndexWriter writer = new IndexWriter(store, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer())); // add the same document four times - Fieldable f1 = new Field("field", "word", Field.Store.YES, Field.Index.ANALYZED); + Fieldable f1 = newField("field", "word", Field.Store.YES, Field.Index.ANALYZED); Document d1 = new Document(); d1.add(f1); writer.addDocument(d1); Index: lucene/src/test/org/apache/lucene/search/TestWildcard.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestWildcard.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestWildcard.java (working copy) @@ -211,7 +211,7 @@ RandomIndexWriter writer = new RandomIndexWriter(random, indexStore); for (int i = 0; i < contents.length; ++i) { Document doc = new Document(); - doc.add(new Field(field, contents[i], Field.Store.YES, Field.Index.ANALYZED)); + doc.add(newField(field, contents[i], Field.Store.YES, Field.Index.ANALYZED)); writer.addDocument(doc); } writer.close(); @@ -267,7 +267,7 @@ RandomIndexWriter iw = new RandomIndexWriter(random, dir); for (int i = 0; i < docs.length; i++) { Document doc = new Document(); - doc.add(new Field(field,docs[i],Store.NO,Index.ANALYZED)); + doc.add(newField(field,docs[i],Store.NO,Index.ANALYZED)); iw.addDocument(doc); } iw.close(); Index: lucene/src/test/org/apache/lucene/search/BaseTestRangeFilter.java =================================================================== --- lucene/src/test/org/apache/lucene/search/BaseTestRangeFilter.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/BaseTestRangeFilter.java (working copy) @@ -111,7 +111,7 @@ for (int d = minId; d <= maxId; d++) { Document doc = new Document(); - doc.add(new Field("id", pad(d), Field.Store.YES, + doc.add(newField("id", pad(d), Field.Store.YES, Field.Index.NOT_ANALYZED)); int r = index.allowNegativeRandomInts ? random.nextInt() : random .nextInt(Integer.MAX_VALUE); @@ -121,9 +121,9 @@ if (r < index.minR) { index.minR = r; } - doc.add(new Field("rand", pad(r), Field.Store.YES, + doc.add(newField("rand", pad(r), Field.Store.YES, Field.Index.NOT_ANALYZED)); - doc.add(new Field("body", "body", Field.Store.YES, + doc.add(newField("body", "body", Field.Store.YES, Field.Index.NOT_ANALYZED)); writer.addDocument(doc); } Index: lucene/src/test/org/apache/lucene/search/TestAutomatonQuery.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestAutomatonQuery.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestAutomatonQuery.java (working copy) @@ -43,11 +43,11 @@ directory = newDirectory(); RandomIndexWriter writer = new RandomIndexWriter(random, directory); Document doc = new Document(); - Field titleField = new Field("title", "some title", Field.Store.NO, + Field titleField = newField("title", "some title", Field.Store.NO, Field.Index.ANALYZED); - Field field = new Field(FN, "this is document one 2345", Field.Store.NO, + Field field = newField(FN, "this is document one 2345", Field.Store.NO, Field.Index.ANALYZED); - Field footerField = new Field("footer", "a footer", Field.Store.NO, + Field footerField = newField("footer", "a footer", Field.Store.NO, Field.Index.ANALYZED); doc.add(titleField); doc.add(field); Index: lucene/src/test/org/apache/lucene/search/TestBoolean2.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestBoolean2.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestBoolean2.java (working copy) @@ -54,7 +54,7 @@ RandomIndexWriter writer= new RandomIndexWriter(random, directory); for (int i = 0; i < docFields.length; i++) { Document doc = new Document(); - doc.add(new Field(field, docFields[i], Field.Store.NO, Field.Index.ANALYZED)); + doc.add(newField(field, docFields[i], Field.Store.NO, Field.Index.ANALYZED)); writer.addDocument(doc); } writer.close(); @@ -77,12 +77,12 @@ RandomIndexWriter w = new RandomIndexWriter(random, dir2); Document doc = new Document(); - doc.add(new Field("field2", "xxx", Field.Store.NO, Field.Index.ANALYZED)); + doc.add(newField("field2", "xxx", Field.Store.NO, Field.Index.ANALYZED)); for(int i=0;i terms = new ArrayList(); int num = 2000 * RANDOM_MULTIPLIER; Index: lucene/src/test/org/apache/lucene/search/TestMultiThreadTermVectors.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestMultiThreadTermVectors.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestMultiThreadTermVectors.java (working copy) @@ -47,7 +47,7 @@ //writer.infoStream = System.out; for (int i = 0; i < numDocs; i++) { Document doc = new Document(); - Fieldable fld = new Field("field", English.intToEnglish(i), Field.Store.YES, Field.Index.NOT_ANALYZED, Field.TermVector.YES); + Fieldable fld = newField("field", English.intToEnglish(i), Field.Store.YES, Field.Index.NOT_ANALYZED, Field.TermVector.YES); doc.add(fld); writer.addDocument(doc); } Index: lucene/src/test/org/apache/lucene/search/TestRegexpQuery.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestRegexpQuery.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestRegexpQuery.java (working copy) @@ -47,7 +47,7 @@ directory = newDirectory(); RandomIndexWriter writer = new RandomIndexWriter(random, directory); Document doc = new Document(); - doc.add(new Field(FN, + doc.add(newField(FN, "the quick brown fox jumps over the lazy ??? dog 493432 49344", Field.Store.NO, Field.Index.ANALYZED)); writer.addDocument(doc); Index: lucene/src/test/org/apache/lucene/search/TestDateFilter.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestDateFilter.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestDateFilter.java (working copy) @@ -51,10 +51,10 @@ Document doc = new Document(); // add time that is in the past - doc.add(new Field("datefield", DateTools.timeToString(now - 1000, + doc.add(newField("datefield", DateTools.timeToString(now - 1000, DateTools.Resolution.MILLISECOND), Field.Store.YES, Field.Index.NOT_ANALYZED)); - doc.add(new Field("body", "Today is a very sunny day in New York City", + doc.add(newField("body", "Today is a very sunny day in New York City", Field.Store.YES, Field.Index.ANALYZED)); writer.addDocument(doc); @@ -117,10 +117,10 @@ Document doc = new Document(); // add time that is in the future - doc.add(new Field("datefield", DateTools.timeToString(now + 888888, + doc.add(newField("datefield", DateTools.timeToString(now + 888888, DateTools.Resolution.MILLISECOND), Field.Store.YES, Field.Index.NOT_ANALYZED)); - doc.add(new Field("body", "Today is a very sunny day in New York City", + doc.add(newField("body", "Today is a very sunny day in New York City", Field.Store.YES, Field.Index.ANALYZED)); writer.addDocument(doc); Index: lucene/src/test/org/apache/lucene/search/TestFilteredSearch.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestFilteredSearch.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestFilteredSearch.java (working copy) @@ -67,7 +67,7 @@ try { for (int i = 0; i < 60; i++) {//Simple docs Document doc = new Document(); - doc.add(new Field(FIELD, Integer.toString(i), Field.Store.YES, Field.Index.NOT_ANALYZED)); + doc.add(newField(FIELD, Integer.toString(i), Field.Store.YES, Field.Index.NOT_ANALYZED)); writer.addDocument(doc); } if(optimize) Index: lucene/src/test/org/apache/lucene/search/TestCachingSpanFilter.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestCachingSpanFilter.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestCachingSpanFilter.java (working copy) @@ -42,7 +42,7 @@ // add a doc, refresh the reader, and check that its there Document doc = new Document(); - doc.add(new Field("id", "1", Field.Store.YES, Field.Index.NOT_ANALYZED)); + doc.add(newField("id", "1", Field.Store.YES, Field.Index.NOT_ANALYZED)); writer.addDocument(doc); reader = refreshReader(reader); Index: lucene/src/test/org/apache/lucene/search/TestMatchAllDocsQuery.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestMatchAllDocsQuery.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestMatchAllDocsQuery.java (working copy) @@ -124,7 +124,7 @@ private void addDoc(String text, IndexWriter iw, float boost) throws IOException { Document doc = new Document(); - Field f = new Field("key", text, Field.Store.YES, Field.Index.ANALYZED); + Field f = newField("key", text, Field.Store.YES, Field.Index.ANALYZED); f.setBoost(boost); doc.add(f); iw.addDocument(doc); Index: lucene/src/test/org/apache/lucene/search/TestDisjunctionMaxQuery.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestDisjunctionMaxQuery.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestDisjunctionMaxQuery.java (working copy) @@ -88,13 +88,13 @@ // d1 is an "ok" match for: albino elephant { Document d1 = new Document(); - d1.add(new Field("id", "d1", Field.Store.YES, Field.Index.NOT_ANALYZED));// Field.Keyword("id", + d1.add(newField("id", "d1", Field.Store.YES, Field.Index.NOT_ANALYZED));// Field.Keyword("id", // "d1")); d1 - .add(new Field("hed", "elephant", Field.Store.YES, + .add(newField("hed", "elephant", Field.Store.YES, Field.Index.ANALYZED));// Field.Text("hed", "elephant")); d1 - .add(new Field("dek", "elephant", Field.Store.YES, + .add(newField("dek", "elephant", Field.Store.YES, Field.Index.ANALYZED));// Field.Text("dek", "elephant")); writer.addDocument(d1); } @@ -102,15 +102,15 @@ // d2 is a "good" match for: albino elephant { Document d2 = new Document(); - d2.add(new Field("id", "d2", Field.Store.YES, Field.Index.NOT_ANALYZED));// Field.Keyword("id", + d2.add(newField("id", "d2", Field.Store.YES, Field.Index.NOT_ANALYZED));// Field.Keyword("id", // "d2")); d2 - .add(new Field("hed", "elephant", Field.Store.YES, + .add(newField("hed", "elephant", Field.Store.YES, Field.Index.ANALYZED));// Field.Text("hed", "elephant")); - d2.add(new Field("dek", "albino", Field.Store.YES, Field.Index.ANALYZED));// Field.Text("dek", + d2.add(newField("dek", "albino", Field.Store.YES, Field.Index.ANALYZED));// Field.Text("dek", // "albino")); d2 - .add(new Field("dek", "elephant", Field.Store.YES, + .add(newField("dek", "elephant", Field.Store.YES, Field.Index.ANALYZED));// Field.Text("dek", "elephant")); writer.addDocument(d2); } @@ -118,12 +118,12 @@ // d3 is a "better" match for: albino elephant { Document d3 = new Document(); - d3.add(new Field("id", "d3", Field.Store.YES, Field.Index.NOT_ANALYZED));// Field.Keyword("id", + d3.add(newField("id", "d3", Field.Store.YES, Field.Index.NOT_ANALYZED));// Field.Keyword("id", // "d3")); - d3.add(new Field("hed", "albino", Field.Store.YES, Field.Index.ANALYZED));// Field.Text("hed", + d3.add(newField("hed", "albino", Field.Store.YES, Field.Index.ANALYZED));// Field.Text("hed", // "albino")); d3 - .add(new Field("hed", "elephant", Field.Store.YES, + .add(newField("hed", "elephant", Field.Store.YES, Field.Index.ANALYZED));// Field.Text("hed", "elephant")); writer.addDocument(d3); } @@ -131,14 +131,14 @@ // d4 is the "best" match for: albino elephant { Document d4 = new Document(); - d4.add(new Field("id", "d4", Field.Store.YES, Field.Index.NOT_ANALYZED));// Field.Keyword("id", + d4.add(newField("id", "d4", Field.Store.YES, Field.Index.NOT_ANALYZED));// Field.Keyword("id", // "d4")); - d4.add(new Field("hed", "albino", Field.Store.YES, Field.Index.ANALYZED));// Field.Text("hed", + d4.add(newField("hed", "albino", Field.Store.YES, Field.Index.ANALYZED));// Field.Text("hed", // "albino")); d4 - .add(new Field("hed", "elephant", Field.Store.YES, + .add(newField("hed", "elephant", Field.Store.YES, Field.Index.ANALYZED));// Field.Text("hed", "elephant")); - d4.add(new Field("dek", "albino", Field.Store.YES, Field.Index.ANALYZED));// Field.Text("dek", + d4.add(newField("dek", "albino", Field.Store.YES, Field.Index.ANALYZED));// Field.Text("dek", // "albino")); writer.addDocument(d4); } Index: lucene/src/test/org/apache/lucene/search/TestSimilarity.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestSimilarity.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestSimilarity.java (working copy) @@ -68,10 +68,10 @@ .setSimilarity(new SimpleSimilarity())); Document d1 = new Document(); - d1.add(new Field("field", "a c", Field.Store.YES, Field.Index.ANALYZED)); + d1.add(newField("field", "a c", Field.Store.YES, Field.Index.ANALYZED)); Document d2 = new Document(); - d2.add(new Field("field", "a b c", Field.Store.YES, Field.Index.ANALYZED)); + d2.add(newField("field", "a b c", Field.Store.YES, Field.Index.ANALYZED)); writer.addDocument(d1); writer.addDocument(d2); Index: lucene/src/test/org/apache/lucene/search/TestFieldCacheRangeFilter.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestFieldCacheRangeFilter.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestFieldCacheRangeFilter.java (working copy) @@ -520,8 +520,8 @@ for (int d = -20; d <= 20; d++) { Document doc = new Document(); - doc.add(new Field("id",Integer.toString(d), Field.Store.NO, Field.Index.NOT_ANALYZED)); - doc.add(new Field("body","body", Field.Store.NO, Field.Index.NOT_ANALYZED)); + doc.add(newField("id",Integer.toString(d), Field.Store.NO, Field.Index.NOT_ANALYZED)); + doc.add(newField("body","body", Field.Store.NO, Field.Index.NOT_ANALYZED)); writer.addDocument(doc); } Index: lucene/src/test/org/apache/lucene/search/TestRegexpRandom.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestRegexpRandom.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestRegexpRandom.java (working copy) @@ -47,7 +47,7 @@ RandomIndexWriter writer = new RandomIndexWriter(random, dir); Document doc = new Document(); - Field field = new Field("field", "", Field.Store.NO, Field.Index.ANALYZED); + Field field = newField("field", "", Field.Store.NO, Field.Index.ANALYZED); doc.add(field); NumberFormat df = new DecimalFormat("0000", new DecimalFormatSymbols(Locale.ENGLISH)); Index: lucene/src/test/org/apache/lucene/search/TestSpanQueryFilter.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestSpanQueryFilter.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestSpanQueryFilter.java (working copy) @@ -41,7 +41,7 @@ RandomIndexWriter writer = new RandomIndexWriter(random, dir); for (int i = 0; i < 500; i++) { Document document = new Document(); - document.add(new Field("field", English.intToEnglish(i) + " equals " + English.intToEnglish(i), + document.add(newField("field", English.intToEnglish(i) + " equals " + English.intToEnglish(i), Field.Store.NO, Field.Index.ANALYZED)); writer.addDocument(document); } Index: lucene/src/test/org/apache/lucene/search/TestPrefixRandom.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestPrefixRandom.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestPrefixRandom.java (working copy) @@ -48,7 +48,7 @@ RandomIndexWriter writer = new RandomIndexWriter(random, dir, new MockAnalyzer(MockTokenizer.KEYWORD, false)); Document doc = new Document(); - Field field = new Field("field", "", Field.Store.NO, Field.Index.NOT_ANALYZED); + Field field = newField("field", "", Field.Store.NO, Field.Index.NOT_ANALYZED); doc.add(field); int num = 2000 * RANDOM_MULTIPLIER; Index: lucene/src/test/org/apache/lucene/search/TestCustomSearcherSort.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestCustomSearcherSort.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestCustomSearcherSort.java (working copy) @@ -56,15 +56,15 @@ Document doc = new Document(); if ((i % 5) != 0) { // some documents must not have an entry in the first // sort field - doc.add(new Field("publicationDate_", random.getLuceneDate(), + doc.add(newField("publicationDate_", random.getLuceneDate(), Field.Store.YES, Field.Index.NOT_ANALYZED)); } if ((i % 7) == 0) { // some documents to match the query (see below) - doc.add(new Field("content", "test", Field.Store.YES, + doc.add(newField("content", "test", Field.Store.YES, Field.Index.ANALYZED)); } // every document has a defined 'mandant' field - doc.add(new Field("mandant", Integer.toString(i % 3), Field.Store.YES, + doc.add(newField("mandant", Integer.toString(i % 3), Field.Store.YES, Field.Index.NOT_ANALYZED)); writer.addDocument(doc); } Index: lucene/src/test/org/apache/lucene/search/TestPrefixInBooleanQuery.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestPrefixInBooleanQuery.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestPrefixInBooleanQuery.java (working copy) @@ -50,26 +50,26 @@ for (int i = 0; i < 5137; ++i) { Document doc = new Document(); - doc.add(new Field(FIELD, "meaninglessnames", Field.Store.YES, + doc.add(newField(FIELD, "meaninglessnames", Field.Store.YES, Field.Index.NOT_ANALYZED)); writer.addDocument(doc); } { Document doc = new Document(); - doc.add(new Field(FIELD, "tangfulin", Field.Store.YES, + doc.add(newField(FIELD, "tangfulin", Field.Store.YES, Field.Index.NOT_ANALYZED)); writer.addDocument(doc); } for (int i = 5138; i < 11377; ++i) { Document doc = new Document(); - doc.add(new Field(FIELD, "meaninglessnames", Field.Store.YES, + doc.add(newField(FIELD, "meaninglessnames", Field.Store.YES, Field.Index.NOT_ANALYZED)); writer.addDocument(doc); } { Document doc = new Document(); - doc.add(new Field(FIELD, "tangfulin", Field.Store.YES, + doc.add(newField(FIELD, "tangfulin", Field.Store.YES, Field.Index.NOT_ANALYZED)); writer.addDocument(doc); } Index: lucene/src/test/org/apache/lucene/search/TestDocBoost.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestDocBoost.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestDocBoost.java (working copy) @@ -40,8 +40,8 @@ Directory store = newDirectory(); RandomIndexWriter writer = new RandomIndexWriter(random, store); - Fieldable f1 = new Field("field", "word", Field.Store.YES, Field.Index.ANALYZED); - Fieldable f2 = new Field("field", "word", Field.Store.YES, Field.Index.ANALYZED); + Fieldable f1 = newField("field", "word", Field.Store.YES, Field.Index.ANALYZED); + Fieldable f2 = newField("field", "word", Field.Store.YES, Field.Index.ANALYZED); f2.setBoost(2.0f); Document d1 = new Document(); Index: lucene/src/test/org/apache/lucene/search/TestElevationComparator.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestElevationComparator.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestElevationComparator.java (working copy) @@ -119,7 +119,7 @@ private Document adoc(String[] vals) { Document doc = new Document(); for (int i = 0; i < vals.length - 2; i += 2) { - doc.add(new Field(vals[i], vals[i + 1], Field.Store.YES, Field.Index.ANALYZED)); + doc.add(newField(vals[i], vals[i + 1], Field.Store.YES, Field.Index.ANALYZED)); } return doc; } Index: lucene/src/test/org/apache/lucene/search/TestMultiSearcherRanking.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestMultiSearcherRanking.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestMultiSearcherRanking.java (working copy) @@ -166,7 +166,7 @@ private void add(String value, IndexWriter iw) throws IOException { Document d = new Document(); - d.add(new Field(FIELD_NAME, value, Field.Store.YES, Field.Index.ANALYZED)); + d.add(newField(FIELD_NAME, value, Field.Store.YES, Field.Index.ANALYZED)); iw.addDocument(d); } Index: lucene/src/test/org/apache/lucene/search/TestPrefixQuery.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestPrefixQuery.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestPrefixQuery.java (working copy) @@ -39,7 +39,7 @@ RandomIndexWriter writer = new RandomIndexWriter(random, directory); for (int i = 0; i < categories.length; i++) { Document doc = new Document(); - doc.add(new Field("category", categories[i], Field.Store.YES, Field.Index.NOT_ANALYZED)); + doc.add(newField("category", categories[i], Field.Store.YES, Field.Index.NOT_ANALYZED)); writer.addDocument(doc); } IndexReader reader = writer.getReader(); Index: lucene/src/test/org/apache/lucene/search/TestMultiPhraseQuery.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestMultiPhraseQuery.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestMultiPhraseQuery.java (working copy) @@ -155,7 +155,7 @@ private void add(String s, RandomIndexWriter writer) throws IOException { Document doc = new Document(); - doc.add(new Field("body", s, Field.Store.YES, Field.Index.ANALYZED)); + doc.add(newField("body", s, Field.Store.YES, Field.Index.ANALYZED)); writer.addDocument(doc); } @@ -278,8 +278,8 @@ private void add(String s, String type, RandomIndexWriter writer) throws IOException { Document doc = new Document(); - doc.add(new Field("body", s, Field.Store.YES, Field.Index.ANALYZED)); - doc.add(new Field("type", type, Field.Store.YES, Field.Index.NOT_ANALYZED)); + doc.add(newField("body", s, Field.Store.YES, Field.Index.ANALYZED)); + doc.add(newField("type", type, Field.Store.YES, Field.Index.NOT_ANALYZED)); writer.addDocument(doc); } Index: lucene/src/test/org/apache/lucene/search/TestBooleanPrefixQuery.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestBooleanPrefixQuery.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestBooleanPrefixQuery.java (working copy) @@ -78,7 +78,7 @@ RandomIndexWriter writer = new RandomIndexWriter(random, directory); for (int i = 0; i < categories.length; i++) { Document doc = new Document(); - doc.add(new Field("category", categories[i], Field.Store.YES, Field.Index.NOT_ANALYZED)); + doc.add(newField("category", categories[i], Field.Store.YES, Field.Index.NOT_ANALYZED)); writer.addDocument(doc); } reader = writer.getReader(); Index: lucene/src/test/org/apache/lucene/search/TestFilteredQuery.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestFilteredQuery.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestFilteredQuery.java (working copy) @@ -51,23 +51,23 @@ RandomIndexWriter writer = new RandomIndexWriter (random, directory); Document doc = new Document(); - doc.add (new Field("field", "one two three four five", Field.Store.YES, Field.Index.ANALYZED)); - doc.add (new Field("sorter", "b", Field.Store.YES, Field.Index.ANALYZED)); + doc.add (newField("field", "one two three four five", Field.Store.YES, Field.Index.ANALYZED)); + doc.add (newField("sorter", "b", Field.Store.YES, Field.Index.ANALYZED)); writer.addDocument (doc); doc = new Document(); - doc.add (new Field("field", "one two three four", Field.Store.YES, Field.Index.ANALYZED)); - doc.add (new Field("sorter", "d", Field.Store.YES, Field.Index.ANALYZED)); + doc.add (newField("field", "one two three four", Field.Store.YES, Field.Index.ANALYZED)); + doc.add (newField("sorter", "d", Field.Store.YES, Field.Index.ANALYZED)); writer.addDocument (doc); doc = new Document(); - doc.add (new Field("field", "one two three y", Field.Store.YES, Field.Index.ANALYZED)); - doc.add (new Field("sorter", "a", Field.Store.YES, Field.Index.ANALYZED)); + doc.add (newField("field", "one two three y", Field.Store.YES, Field.Index.ANALYZED)); + doc.add (newField("sorter", "a", Field.Store.YES, Field.Index.ANALYZED)); writer.addDocument (doc); doc = new Document(); - doc.add (new Field("field", "one two x", Field.Store.YES, Field.Index.ANALYZED)); - doc.add (new Field("sorter", "c", Field.Store.YES, Field.Index.ANALYZED)); + doc.add (newField("field", "one two x", Field.Store.YES, Field.Index.ANALYZED)); + doc.add (newField("sorter", "c", Field.Store.YES, Field.Index.ANALYZED)); writer.addDocument (doc); // tests here require single segment (eg try seed Index: lucene/src/test/org/apache/lucene/search/TestFieldCacheTermsFilter.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestFieldCacheTermsFilter.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestFieldCacheTermsFilter.java (working copy) @@ -41,7 +41,7 @@ for (int i = 0; i < 100; i++) { Document doc = new Document(); int term = i * 10; //terms are units of 10; - doc.add(new Field(fieldName, "" + term, Field.Store.YES, Field.Index.NOT_ANALYZED)); + doc.add(newField(fieldName, "" + term, Field.Store.YES, Field.Index.NOT_ANALYZED)); w.addDocument(doc); } IndexReader reader = w.getReader(); Index: lucene/src/test/org/apache/lucene/search/TestTermRangeFilter.java =================================================================== --- lucene/src/test/org/apache/lucene/search/TestTermRangeFilter.java (revision 995772) +++ lucene/src/test/org/apache/lucene/search/TestTermRangeFilter.java (working copy) @@ -401,10 +401,10 @@ Directory farsiIndex = newDirectory(); RandomIndexWriter writer = new RandomIndexWriter(random, farsiIndex); Document doc = new Document(); - doc.add(new Field("content", "\u0633\u0627\u0628", Field.Store.YES, + doc.add(newField("content", "\u0633\u0627\u0628", Field.Store.YES, Field.Index.NOT_ANALYZED)); doc - .add(new Field("body", "body", Field.Store.YES, + .add(newField("body", "body", Field.Store.YES, Field.Index.NOT_ANALYZED)); writer.addDocument(doc); @@ -445,9 +445,9 @@ String[] words = {"H\u00D8T", "H\u00C5T", "MAND"}; for (int docnum = 0; docnum < words.length; ++docnum) { Document doc = new Document(); - doc.add(new Field("content", words[docnum], Field.Store.YES, + doc.add(newField("content", words[docnum], Field.Store.YES, Field.Index.NOT_ANALYZED)); - doc.add(new Field("body", "body", Field.Store.YES, + doc.add(newField("body", "body", Field.Store.YES, Field.Index.NOT_ANALYZED)); writer.addDocument(doc); } Index: lucene/src/test/org/apache/lucene/TestSearch.java =================================================================== --- lucene/src/test/org/apache/lucene/TestSearch.java (revision 995772) +++ lucene/src/test/org/apache/lucene/TestSearch.java (working copy) @@ -91,7 +91,7 @@ }; for (int j = 0; j < docs.length; j++) { Document d = new Document(); - d.add(new Field("contents", docs[j], Field.Store.YES, Field.Index.ANALYZED)); + d.add(newField("contents", docs[j], Field.Store.YES, Field.Index.ANALYZED)); writer.addDocument(d); } writer.close(); Index: lucene/src/test/org/apache/lucene/index/TestSnapshotDeletionPolicy.java =================================================================== --- lucene/src/test/org/apache/lucene/index/TestSnapshotDeletionPolicy.java (revision 995772) +++ lucene/src/test/org/apache/lucene/index/TestSnapshotDeletionPolicy.java (working copy) @@ -111,7 +111,7 @@ @Override public void run() { Document doc = new Document(); - doc.add(new Field("content", "aaa", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.WITH_POSITIONS_OFFSETS)); + doc.add(newField("content", "aaa", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.WITH_POSITIONS_OFFSETS)); do { for(int i=0;i<27;i++) { try { @@ -152,7 +152,7 @@ // final segment, so deletion policy has a chance to // delete again: Document doc = new Document(); - doc.add(new Field("content", "aaa", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.WITH_POSITIONS_OFFSETS)); + doc.add(newField("content", "aaa", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.WITH_POSITIONS_OFFSETS)); writer.addDocument(doc); // Make sure we don't have any leftover files in the Index: lucene/src/test/org/apache/lucene/index/TestPayloadProcessorProvider.java =================================================================== --- lucene/src/test/org/apache/lucene/index/TestPayloadProcessorProvider.java (revision 995772) +++ lucene/src/test/org/apache/lucene/index/TestPayloadProcessorProvider.java (working copy) @@ -140,8 +140,8 @@ TokenStream payloadTS2 = new PayloadTokenStream("p2"); for (int i = 0; i < NUM_DOCS; i++) { Document doc = new Document(); - doc.add(new Field("id", "doc" + i, Store.NO, Index.NOT_ANALYZED_NO_NORMS)); - doc.add(new Field("content", "doc content " + i, Store.NO, Index.ANALYZED)); + doc.add(newField("id", "doc" + i, Store.NO, Index.NOT_ANALYZED_NO_NORMS)); + doc.add(newField("content", "doc content " + i, Store.NO, Index.ANALYZED)); doc.add(new Field("p", payloadTS1)); doc.add(new Field("p", payloadTS2)); writer.addDocument(doc); Index: lucene/src/test/org/apache/lucene/index/TestParallelTermEnum.java =================================================================== --- lucene/src/test/org/apache/lucene/index/TestParallelTermEnum.java (revision 995772) +++ lucene/src/test/org/apache/lucene/index/TestParallelTermEnum.java (working copy) @@ -42,11 +42,11 @@ IndexWriter iw1 = new IndexWriter(rd1, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer())); doc = new Document(); - doc.add(new Field("field1", "the quick brown fox jumps", Store.YES, + doc.add(newField("field1", "the quick brown fox jumps", Store.YES, Index.ANALYZED)); - doc.add(new Field("field2", "the quick brown fox jumps", Store.YES, + doc.add(newField("field2", "the quick brown fox jumps", Store.YES, Index.ANALYZED)); - doc.add(new Field("field4", "", Store.NO, Index.ANALYZED)); + doc.add(newField("field4", "", Store.NO, Index.ANALYZED)); iw1.addDocument(doc); iw1.close(); @@ -54,10 +54,10 @@ IndexWriter iw2 = new IndexWriter(rd2, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer())); doc = new Document(); - doc.add(new Field("field0", "", Store.NO, Index.ANALYZED)); - doc.add(new Field("field1", "the fox jumps over the lazy dog", + doc.add(newField("field0", "", Store.NO, Index.ANALYZED)); + doc.add(newField("field1", "the fox jumps over the lazy dog", Store.YES, Index.ANALYZED)); - doc.add(new Field("field3", "the fox jumps over the lazy dog", + doc.add(newField("field3", "the fox jumps over the lazy dog", Store.YES, Index.ANALYZED)); iw2.addDocument(doc); Index: lucene/src/test/org/apache/lucene/index/TestIndexReaderClone.java =================================================================== --- lucene/src/test/org/apache/lucene/index/TestIndexReaderClone.java (revision 995772) +++ lucene/src/test/org/apache/lucene/index/TestIndexReaderClone.java (working copy) @@ -495,7 +495,7 @@ ((LogMergePolicy) w.getConfig().getMergePolicy()).setUseCompoundFile(false); ((LogMergePolicy) w.getConfig().getMergePolicy()).setUseCompoundDocStore(false); Document doc = new Document(); - doc.add(new Field("field", "yes it's stored", Field.Store.YES, Field.Index.ANALYZED)); + doc.add(newField("field", "yes it's stored", Field.Store.YES, Field.Index.ANALYZED)); w.addDocument(doc); w.close(); IndexReader r1 = IndexReader.open(dir, false); Index: lucene/src/test/org/apache/lucene/index/TestIsCurrent.java =================================================================== --- lucene/src/test/org/apache/lucene/index/TestIsCurrent.java (revision 995772) +++ lucene/src/test/org/apache/lucene/index/TestIsCurrent.java (working copy) @@ -45,7 +45,7 @@ // write document Document doc = new Document(); - doc.add(new Field("UUID", "1", Store.YES, Index.ANALYZED)); + doc.add(newField("UUID", "1", Store.YES, Index.ANALYZED)); writer.addDocument(doc); writer.commit(); } Index: lucene/src/test/org/apache/lucene/index/TestSegmentTermEnum.java =================================================================== --- lucene/src/test/org/apache/lucene/index/TestSegmentTermEnum.java (revision 995772) +++ lucene/src/test/org/apache/lucene/index/TestSegmentTermEnum.java (working copy) @@ -127,7 +127,7 @@ private void addDoc(IndexWriter writer, String value) throws IOException { Document doc = new Document(); - doc.add(new Field("content", value, Field.Store.NO, Field.Index.ANALYZED)); + doc.add(newField("content", value, Field.Store.NO, Field.Index.ANALYZED)); writer.addDocument(doc); } } Index: lucene/src/test/org/apache/lucene/index/TestRollback.java =================================================================== --- lucene/src/test/org/apache/lucene/index/TestRollback.java (revision 995772) +++ lucene/src/test/org/apache/lucene/index/TestRollback.java (working copy) @@ -33,7 +33,7 @@ RandomIndexWriter rw = new RandomIndexWriter(random, dir); for (int i = 0; i < 5; i++) { Document doc = new Document(); - doc.add(new Field("pk", Integer.toString(i), Store.YES, Index.ANALYZED_NO_NORMS)); + doc.add(newField("pk", Integer.toString(i), Store.YES, Index.ANALYZED_NO_NORMS)); rw.addDocument(doc); } rw.close(); @@ -45,8 +45,8 @@ for (int i = 0; i < 3; i++) { Document doc = new Document(); String value = Integer.toString(i); - doc.add(new Field("pk", value, Store.YES, Index.ANALYZED_NO_NORMS)); - doc.add(new Field("text", "foo", Store.YES, Index.ANALYZED_NO_NORMS)); + doc.add(newField("pk", value, Store.YES, Index.ANALYZED_NO_NORMS)); + doc.add(newField("text", "foo", Store.YES, Index.ANALYZED_NO_NORMS)); w.updateDocument(pkTerm.createTerm(value), doc); } w.rollback(); Index: lucene/src/test/org/apache/lucene/index/TestIndexReader.java =================================================================== --- lucene/src/test/org/apache/lucene/index/TestIndexReader.java (revision 995772) +++ lucene/src/test/org/apache/lucene/index/TestIndexReader.java (working copy) @@ -163,7 +163,14 @@ // set up writer IndexWriter writer = new IndexWriter(d, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer())); - addDocumentWithFields(writer); + + Document doc = new Document(); + doc.add(new Field("keyword","test1", Field.Store.YES, Field.Index.NOT_ANALYZED)); + doc.add(new Field("text","test1", Field.Store.YES, Field.Index.ANALYZED)); + doc.add(new Field("unindexed","test1", Field.Store.YES, Field.Index.NO)); + doc.add(new Field("unstored","test1", Field.Store.NO, Field.Index.ANALYZED)); + writer.addDocument(doc); + writer.close(); // set up reader IndexReader reader = IndexReader.open(d, false); @@ -179,15 +186,31 @@ // want to get some more segments here int mergeFactor = ((LogMergePolicy) writer.getConfig().getMergePolicy()).getMergeFactor(); for (int i = 0; i < 5*mergeFactor; i++) { - addDocumentWithFields(writer); + doc = new Document(); + doc.add(new Field("keyword","test1", Field.Store.YES, Field.Index.NOT_ANALYZED)); + doc.add(new Field("text","test1", Field.Store.YES, Field.Index.ANALYZED)); + doc.add(new Field("unindexed","test1", Field.Store.YES, Field.Index.NO)); + doc.add(new Field("unstored","test1", Field.Store.NO, Field.Index.ANALYZED)); + writer.addDocument(doc); } // new fields are in some different segments (we hope) for (int i = 0; i < 5*mergeFactor; i++) { - addDocumentWithDifferentFields(writer); + doc = new Document(); + doc.add(new Field("keyword2","test1", Field.Store.YES, Field.Index.NOT_ANALYZED)); + doc.add(new Field("text2","test1", Field.Store.YES, Field.Index.ANALYZED)); + doc.add(new Field("unindexed2","test1", Field.Store.YES, Field.Index.NO)); + doc.add(new Field("unstored2","test1", Field.Store.NO, Field.Index.ANALYZED)); + writer.addDocument(doc); } // new termvector fields for (int i = 0; i < 5*mergeFactor; i++) { - addDocumentWithTermVectorFields(writer); + doc = new Document(); + doc.add(new Field("tvnot","tvnot", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.NO)); + doc.add(new Field("termvector","termvector", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.YES)); + doc.add(new Field("tvoffset","tvoffset", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.WITH_OFFSETS)); + doc.add(new Field("tvposition","tvposition", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.WITH_POSITIONS)); + doc.add(newField("tvpositionoffset","tvpositionoffset", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.WITH_POSITIONS_OFFSETS)); + writer.addDocument(doc); } writer.close(); @@ -869,8 +892,8 @@ IndexWriter writer = new IndexWriter(startDir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer())); for(int i=0;i<157;i++) { Document d = new Document(); - d.add(new Field("id", Integer.toString(i), Field.Store.YES, Field.Index.NOT_ANALYZED)); - d.add(new Field("content", "aaa " + i, Field.Store.NO, Field.Index.ANALYZED)); + d.add(newField("id", Integer.toString(i), Field.Store.YES, Field.Index.NOT_ANALYZED)); + d.add(newField("content", "aaa " + i, Field.Store.NO, Field.Index.ANALYZED)); writer.addDocument(d); if (0==i%10) writer.commit(); @@ -1137,11 +1160,11 @@ Directory dir = newDirectory(); RandomIndexWriter w = new RandomIndexWriter(random, dir); Document doc = new Document(); - doc.add(new Field("f", "doctor", Field.Store.NO, Field.Index.NOT_ANALYZED)); + doc.add(newField("f", "doctor", Field.Store.NO, Field.Index.NOT_ANALYZED)); w.addDocument(doc); doc = new Document(); w.commit(); - doc.add(new Field("f", "who", Field.Store.NO, Field.Index.NOT_ANALYZED)); + doc.add(newField("f", "who", Field.Store.NO, Field.Index.NOT_ANALYZED)); w.addDocument(doc); IndexReader r = w.getReader(); IndexReader wr = SlowMultiReaderWrapper.wrap(r); @@ -1267,31 +1290,31 @@ private void addDocumentWithFields(IndexWriter writer) throws IOException { Document doc = new Document(); - doc.add(new Field("keyword","test1", Field.Store.YES, Field.Index.NOT_ANALYZED)); - doc.add(new Field("text","test1", Field.Store.YES, Field.Index.ANALYZED)); - doc.add(new Field("unindexed","test1", Field.Store.YES, Field.Index.NO)); - doc.add(new Field("unstored","test1", Field.Store.NO, Field.Index.ANALYZED)); + doc.add(newField("keyword","test1", Field.Store.YES, Field.Index.NOT_ANALYZED)); + doc.add(newField("text","test1", Field.Store.YES, Field.Index.ANALYZED)); + doc.add(newField("unindexed","test1", Field.Store.YES, Field.Index.NO)); + doc.add(newField("unstored","test1", Field.Store.NO, Field.Index.ANALYZED)); writer.addDocument(doc); } private void addDocumentWithDifferentFields(IndexWriter writer) throws IOException { Document doc = new Document(); - doc.add(new Field("keyword2","test1", Field.Store.YES, Field.Index.NOT_ANALYZED)); - doc.add(new Field("text2","test1", Field.Store.YES, Field.Index.ANALYZED)); - doc.add(new Field("unindexed2","test1", Field.Store.YES, Field.Index.NO)); - doc.add(new Field("unstored2","test1", Field.Store.NO, Field.Index.ANALYZED)); + doc.add(newField("keyword2","test1", Field.Store.YES, Field.Index.NOT_ANALYZED)); + doc.add(newField("text2","test1", Field.Store.YES, Field.Index.ANALYZED)); + doc.add(newField("unindexed2","test1", Field.Store.YES, Field.Index.NO)); + doc.add(newField("unstored2","test1", Field.Store.NO, Field.Index.ANALYZED)); writer.addDocument(doc); } private void addDocumentWithTermVectorFields(IndexWriter writer) throws IOException { Document doc = new Document(); - doc.add(new Field("tvnot","tvnot", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.NO)); - doc.add(new Field("termvector","termvector", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.YES)); - doc.add(new Field("tvoffset","tvoffset", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.WITH_OFFSETS)); - doc.add(new Field("tvposition","tvposition", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.WITH_POSITIONS)); - doc.add(new Field("tvpositionoffset","tvpositionoffset", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.WITH_POSITIONS_OFFSETS)); + doc.add(newField("tvnot","tvnot", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.NO)); + doc.add(newField("termvector","termvector", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.YES)); + doc.add(newField("tvoffset","tvoffset", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.WITH_OFFSETS)); + doc.add(newField("tvposition","tvposition", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.WITH_POSITIONS)); + doc.add(newField("tvpositionoffset","tvpositionoffset", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.WITH_POSITIONS_OFFSETS)); writer.addDocument(doc); } @@ -1299,7 +1322,7 @@ private void addDoc(IndexWriter writer, String value) throws IOException { Document doc = new Document(); - doc.add(new Field("content", value, Field.Store.NO, Field.Index.ANALYZED)); + doc.add(newField("content", value, Field.Store.NO, Field.Index.ANALYZED)); writer.addDocument(doc); } private void rmDir(File dir) { @@ -1558,7 +1581,7 @@ private Document createDocument(String id) { Document doc = new Document(); - doc.add(new Field("id", id, Field.Store.YES, Field.Index.NOT_ANALYZED_NO_NORMS)); + doc.add(newField("id", id, Field.Store.YES, Field.Index.NOT_ANALYZED_NO_NORMS)); return doc; } @@ -1608,7 +1631,7 @@ Directory dir = newDirectory(); IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer())); Document doc = new Document(); - doc.add(new Field("number", "17", Field.Store.NO, Field.Index.NOT_ANALYZED)); + doc.add(newField("number", "17", Field.Store.NO, Field.Index.NOT_ANALYZED)); writer.addDocument(doc); writer.close(); @@ -1639,7 +1662,7 @@ Directory dir = newDirectory(); IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer())); Document doc = new Document(); - doc.add(new Field("number", "17", Field.Store.NO, Field.Index.NOT_ANALYZED)); + doc.add(newField("number", "17", Field.Store.NO, Field.Index.NOT_ANALYZED)); ((LogMergePolicy) writer.getMergePolicy()).setMergeFactor(10); writer.addDocument(doc); writer.commit(); @@ -1673,7 +1696,7 @@ IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer()).setMaxBufferedDocs(-1)); ((LogMergePolicy) writer.getConfig().getMergePolicy()).setMergeFactor(10); Document doc = new Document(); - doc.add(new Field("number", "17", Field.Store.NO, Field.Index.NOT_ANALYZED)); + doc.add(newField("number", "17", Field.Store.NO, Field.Index.NOT_ANALYZED)); writer.addDocument(doc); writer.commit(); @@ -1714,8 +1737,8 @@ Directory dir = newDirectory(); IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer()).setCodecProvider(_TestUtil.alwaysCodec("Standard"))); Document doc = new Document(); - doc.add(new Field("field", "a b c d e f g h i j k l m n o p q r s t u v w x y z", Field.Store.NO, Field.Index.ANALYZED)); - doc.add(new Field("number", "0 1 2 3 4 5 6 7 8 9", Field.Store.NO, Field.Index.ANALYZED)); + doc.add(newField("field", "a b c d e f g h i j k l m n o p q r s t u v w x y z", Field.Store.NO, Field.Index.ANALYZED)); + doc.add(newField("number", "0 1 2 3 4 5 6 7 8 9", Field.Store.NO, Field.Index.ANALYZED)); writer.addDocument(doc); writer.addDocument(doc); writer.commit(); @@ -1747,8 +1770,8 @@ Directory dir = newDirectory(); IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer()).setCodecProvider(_TestUtil.alwaysCodec("Standard"))); Document doc = new Document(); - doc.add(new Field("field", "a b c d e f g h i j k l m n o p q r s t u v w x y z", Field.Store.NO, Field.Index.ANALYZED)); - doc.add(new Field("number", "0 1 2 3 4 5 6 7 8 9", Field.Store.NO, Field.Index.ANALYZED)); + doc.add(newField("field", "a b c d e f g h i j k l m n o p q r s t u v w x y z", Field.Store.NO, Field.Index.ANALYZED)); + doc.add(newField("number", "0 1 2 3 4 5 6 7 8 9", Field.Store.NO, Field.Index.ANALYZED)); writer.addDocument(doc); writer.addDocument(doc); writer.close(); Index: lucene/src/test/org/apache/lucene/index/TestStressIndexing2.java =================================================================== --- lucene/src/test/org/apache/lucene/index/TestStressIndexing2.java (revision 995772) +++ lucene/src/test/org/apache/lucene/index/TestStressIndexing2.java (working copy) @@ -530,20 +530,26 @@ for(int j=0;j fields = new ArrayList(); String idString = getIdString(); - Field idField = new Field(idTerm.field(), idString, Field.Store.YES, Field.Index.NOT_ANALYZED_NO_NORMS); + Field idField = newField(idTerm.field(), idString, Field.Store.YES, Field.Index.NOT_ANALYZED_NO_NORMS); fields.add(idField); int nFields = nextInt(maxFields); @@ -663,16 +669,16 @@ switch (nextInt(4)) { case 0: - fields.add(new Field("f" + nextInt(100), getString(1), Field.Store.YES, Field.Index.NOT_ANALYZED_NO_NORMS, tvVal)); + fields.add(newField("f" + nextInt(100), getString(1), Field.Store.YES, Field.Index.NOT_ANALYZED_NO_NORMS, tvVal)); break; case 1: - fields.add(new Field("f" + nextInt(100), getString(0), Field.Store.NO, Field.Index.ANALYZED, tvVal)); + fields.add(newField("f" + nextInt(100), getString(0), Field.Store.NO, Field.Index.ANALYZED, tvVal)); break; case 2: - fields.add(new Field("f" + nextInt(100), getString(0), Field.Store.YES, Field.Index.NO, Field.TermVector.NO)); + fields.add(newField("f" + nextInt(100), getString(0), Field.Store.YES, Field.Index.NO, Field.TermVector.NO)); break; case 3: - fields.add(new Field("f" + nextInt(100), getString(bigFieldSize), Field.Store.YES, Field.Index.ANALYZED, tvVal)); + fields.add(newField("f" + nextInt(100), getString(bigFieldSize), Field.Store.YES, Field.Index.ANALYZED, tvVal)); break; } } Index: lucene/src/test/org/apache/lucene/index/TestFlex.java =================================================================== --- lucene/src/test/org/apache/lucene/index/TestFlex.java (revision 995772) +++ lucene/src/test/org/apache/lucene/index/TestFlex.java (working copy) @@ -37,10 +37,10 @@ if (iter == 0) { w.setMaxBufferedDocs(7); Document doc = new Document(); - doc.add(new Field("field1", "this is field1", Field.Store.NO, Field.Index.ANALYZED)); - doc.add(new Field("field2", "this is field2", Field.Store.NO, Field.Index.ANALYZED)); - doc.add(new Field("field3", "aaa", Field.Store.NO, Field.Index.ANALYZED)); - doc.add(new Field("field4", "bbb", Field.Store.NO, Field.Index.ANALYZED)); + doc.add(newField("field1", "this is field1", Field.Store.NO, Field.Index.ANALYZED)); + doc.add(newField("field2", "this is field2", Field.Store.NO, Field.Index.ANALYZED)); + doc.add(newField("field3", "aaa", Field.Store.NO, Field.Index.ANALYZED)); + doc.add(newField("field4", "bbb", Field.Store.NO, Field.Index.ANALYZED)); for(int i=0;i allTerms = new HashSet(); @@ -4799,7 +4799,7 @@ s.append(' ').append(""+i); } Document d = new Document(); - Field f = new Field("field", s.toString(), Field.Store.NO, Field.Index.ANALYZED); + Field f = newField("field", s.toString(), Field.Store.NO, Field.Index.ANALYZED); d.add(f); w.addDocument(d); IndexReader r = w.getReader(2).getSequentialSubReaders()[0]; @@ -4824,7 +4824,7 @@ IndexWriter w = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer())); ((LogMergePolicy) w.getMergePolicy()).setUseCompoundFile(true); Document doc = new Document(); - doc.add(new Field("field", "go", Field.Store.NO, Field.Index.ANALYZED)); + doc.add(newField("field", "go", Field.Store.NO, Field.Index.ANALYZED)); w.addDocument(doc); IndexReader r; if (iter == 0) { @@ -4887,7 +4887,7 @@ // First commit Document doc = new Document(); - doc.add(new Field("c", "val", Store.YES, Index.ANALYZED, TermVector.WITH_POSITIONS_OFFSETS)); + doc.add(newField("c", "val", Store.YES, Index.ANALYZED, TermVector.WITH_POSITIONS_OFFSETS)); writer.addDocument(doc); writer.commit(); assertEquals(1, IndexReader.listCommits(dir).size()); @@ -4897,7 +4897,7 @@ // Second commit - now KeepOnlyLastCommit cannot delete the prev commit. doc = new Document(); - doc.add(new Field("c", "val", Store.YES, Index.ANALYZED, TermVector.WITH_POSITIONS_OFFSETS)); + doc.add(newField("c", "val", Store.YES, Index.ANALYZED, TermVector.WITH_POSITIONS_OFFSETS)); writer.addDocument(doc); writer.commit(); assertEquals(2, IndexReader.listCommits(dir).size()); @@ -4928,7 +4928,7 @@ FlushCountingIndexWriter w = new FlushCountingIndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer()).setRAMBufferSizeMB(0.5).setMaxBufferedDocs(-1).setMaxBufferedDeleteTerms(-1)); //w.setInfoStream(System.out); Document doc = new Document(); - doc.add(new Field("field", "go 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20", Field.Store.NO, Field.Index.ANALYZED)); + doc.add(newField("field", "go 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20", Field.Store.NO, Field.Index.ANALYZED)); int num = 6 * RANDOM_MULTIPLIER; for (int iter = 0; iter < num; iter++) { int count = 0; @@ -5004,13 +5004,13 @@ Document doc = new Document(); // create as many files as possible - doc.add(new Field("c", "val", Store.YES, Index.ANALYZED, TermVector.WITH_POSITIONS_OFFSETS)); + doc.add(newField("c", "val", Store.YES, Index.ANALYZED, TermVector.WITH_POSITIONS_OFFSETS)); writer.addDocument(doc); // Adding just one document does not call flush yet. assertEquals("only the stored and term vector files should exist in the directory", 5 + extraFileCount, dir.listAll().length); doc = new Document(); - doc.add(new Field("c", "val", Store.YES, Index.ANALYZED, TermVector.WITH_POSITIONS_OFFSETS)); + doc.add(newField("c", "val", Store.YES, Index.ANALYZED, TermVector.WITH_POSITIONS_OFFSETS)); writer.addDocument(doc); // The second document should cause a flush. assertTrue("flush should have occurred and files created", dir.listAll().length > 5 + extraFileCount); @@ -5034,7 +5034,7 @@ TEST_VERSION_CURRENT, new MockAnalyzer()).setMaxBufferedDocs(2)); Document doc = new Document(); - doc.add(new Field("c", "val", Store.YES, Index.ANALYZED, TermVector.WITH_POSITIONS_OFFSETS)); + doc.add(newField("c", "val", Store.YES, Index.ANALYZED, TermVector.WITH_POSITIONS_OFFSETS)); w.addDocument(doc); w.addDocument(doc); IndexWriter w2 = new IndexWriter(dir, newIndexWriterConfig( @@ -5140,7 +5140,7 @@ final List fieldIDs = new ArrayList(); - Field idField = new Field("id", "", Field.Store.YES, Field.Index.NOT_ANALYZED); + Field idField = newField("id", "", Field.Store.YES, Field.Index.NOT_ANALYZED); for(int i=0;i0) { int k = i-1; @@ -1196,7 +1196,7 @@ ((LogMergePolicy) writer.getConfig().getMergePolicy()).setMergeFactor(10); for(int i=0;i<4;i++) { Document doc = new Document(); - doc.add(new Field("id", ""+i, Field.Store.NO, Field.Index.NOT_ANALYZED)); + doc.add(newField("id", ""+i, Field.Store.NO, Field.Index.NOT_ANALYZED)); writer.addDocument(doc); Map data = new HashMap(); data.put("index", i+""); Index: lucene/src/test/org/apache/lucene/index/TestAddIndexes.java =================================================================== --- lucene/src/test/org/apache/lucene/index/TestAddIndexes.java (revision 995772) +++ lucene/src/test/org/apache/lucene/index/TestAddIndexes.java (working copy) @@ -150,8 +150,8 @@ // docs, so 10 pending deletes: for (int i = 0; i < 20; i++) { Document doc = new Document(); - doc.add(new Field("id", "" + (i % 10), Field.Store.NO, Field.Index.NOT_ANALYZED)); - doc.add(new Field("content", "bbb " + i, Field.Store.NO, + doc.add(newField("id", "" + (i % 10), Field.Store.NO, Field.Index.NOT_ANALYZED)); + doc.add(newField("content", "bbb " + i, Field.Store.NO, Field.Index.ANALYZED)); writer.updateDocument(new Term("id", "" + (i%10)), doc); } @@ -186,8 +186,8 @@ // docs, so 10 pending deletes: for (int i = 0; i < 20; i++) { Document doc = new Document(); - doc.add(new Field("id", "" + (i % 10), Field.Store.NO, Field.Index.NOT_ANALYZED)); - doc.add(new Field("content", "bbb " + i, Field.Store.NO, Field.Index.ANALYZED)); + doc.add(newField("id", "" + (i % 10), Field.Store.NO, Field.Index.NOT_ANALYZED)); + doc.add(newField("content", "bbb " + i, Field.Store.NO, Field.Index.ANALYZED)); writer.updateDocument(new Term("id", "" + (i%10)), doc); } @@ -224,8 +224,8 @@ // docs, so 10 pending deletes: for (int i = 0; i < 20; i++) { Document doc = new Document(); - doc.add(new Field("id", "" + (i % 10), Field.Store.NO, Field.Index.NOT_ANALYZED)); - doc.add(new Field("content", "bbb " + i, Field.Store.NO, + doc.add(newField("id", "" + (i % 10), Field.Store.NO, Field.Index.NOT_ANALYZED)); + doc.add(newField("content", "bbb " + i, Field.Store.NO, Field.Index.ANALYZED)); writer.updateDocument(new Term("id", "" + (i%10)), doc); } @@ -456,7 +456,7 @@ private void addDocs(IndexWriter writer, int numDocs) throws IOException { for (int i = 0; i < numDocs; i++) { Document doc = new Document(); - doc.add(new Field("content", "aaa", Field.Store.NO, + doc.add(newField("content", "aaa", Field.Store.NO, Field.Index.ANALYZED)); writer.addDocument(doc); } @@ -465,7 +465,7 @@ private void addDocs2(IndexWriter writer, int numDocs) throws IOException { for (int i = 0; i < numDocs; i++) { Document doc = new Document(); - doc.add(new Field("content", "bbb", Field.Store.NO, + doc.add(newField("content", "bbb", Field.Store.NO, Field.Index.ANALYZED)); writer.addDocument(doc); } @@ -530,19 +530,19 @@ .setMaxBufferedDocs(5).setMergePolicy(lmp)); Document doc = new Document(); - doc.add(new Field("content", "aaa bbb ccc ddd eee fff ggg hhh iii", Field.Store.YES, + doc.add(newField("content", "aaa bbb ccc ddd eee fff ggg hhh iii", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.WITH_POSITIONS_OFFSETS)); for(int i=0;i<60;i++) writer.addDocument(doc); Document doc2 = new Document(); - doc2.add(new Field("content", "aaa bbb ccc ddd eee fff ggg hhh iii", Field.Store.YES, + doc2.add(newField("content", "aaa bbb ccc ddd eee fff ggg hhh iii", Field.Store.YES, Field.Index.NO)); - doc2.add(new Field("content", "aaa bbb ccc ddd eee fff ggg hhh iii", Field.Store.YES, + doc2.add(newField("content", "aaa bbb ccc ddd eee fff ggg hhh iii", Field.Store.YES, Field.Index.NO)); - doc2.add(new Field("content", "aaa bbb ccc ddd eee fff ggg hhh iii", Field.Store.YES, + doc2.add(newField("content", "aaa bbb ccc ddd eee fff ggg hhh iii", Field.Store.YES, Field.Index.NO)); - doc2.add(new Field("content", "aaa bbb ccc ddd eee fff ggg hhh iii", Field.Store.YES, + doc2.add(newField("content", "aaa bbb ccc ddd eee fff ggg hhh iii", Field.Store.YES, Field.Index.NO)); for(int i=0;i<10;i++) writer.addDocument(doc2); Index: lucene/src/test/org/apache/lucene/index/TestStressIndexing.java =================================================================== --- lucene/src/test/org/apache/lucene/index/TestStressIndexing.java (revision 995772) +++ lucene/src/test/org/apache/lucene/index/TestStressIndexing.java (working copy) @@ -80,8 +80,8 @@ for(int j=0; j<10; j++) { Document d = new Document(); int n = random.nextInt(); - d.add(new Field("id", Integer.toString(nextID++), Field.Store.YES, Field.Index.NOT_ANALYZED)); - d.add(new Field("contents", English.intToEnglish(n), Field.Store.NO, Field.Index.ANALYZED)); + d.add(newField("id", Integer.toString(nextID++), Field.Store.YES, Field.Index.NOT_ANALYZED)); + d.add(newField("contents", English.intToEnglish(n), Field.Store.NO, Field.Index.ANALYZED)); writer.addDocument(d); } Index: lucene/src/test/org/apache/lucene/index/TestTermdocPerf.java =================================================================== --- lucene/src/test/org/apache/lucene/index/TestTermdocPerf.java (revision 995772) +++ lucene/src/test/org/apache/lucene/index/TestTermdocPerf.java (working copy) @@ -69,7 +69,7 @@ }; Document doc = new Document(); - doc.add(new Field(field,val, Field.Store.NO, Field.Index.NOT_ANALYZED_NO_NORMS)); + doc.add(newField(field,val, Field.Store.NO, Field.Index.NOT_ANALYZED_NO_NORMS)); IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, analyzer) .setOpenMode(OpenMode.CREATE).setMaxBufferedDocs(100)); Index: lucene/src/test/org/apache/lucene/index/TestPayloads.java =================================================================== --- lucene/src/test/org/apache/lucene/index/TestPayloads.java (revision 995772) +++ lucene/src/test/org/apache/lucene/index/TestPayloads.java (working copy) @@ -103,15 +103,15 @@ IndexWriter writer = new IndexWriter(ram, newIndexWriterConfig( TEST_VERSION_CURRENT, analyzer)); Document d = new Document(); // this field won't have any payloads - d.add(new Field("f1", "This field has no payloads", Field.Store.NO, Field.Index.ANALYZED)); + d.add(newField("f1", "This field has no payloads", Field.Store.NO, Field.Index.ANALYZED)); // this field will have payloads in all docs, however not for all term positions, // so this field is used to check if the DocumentWriter correctly enables the payloads bit // even if only some term positions have payloads - d.add(new Field("f2", "This field has payloads in all docs", Field.Store.NO, Field.Index.ANALYZED)); - d.add(new Field("f2", "This field has payloads in all docs", Field.Store.NO, Field.Index.ANALYZED)); + d.add(newField("f2", "This field has payloads in all docs", Field.Store.NO, Field.Index.ANALYZED)); + d.add(newField("f2", "This field has payloads in all docs", Field.Store.NO, Field.Index.ANALYZED)); // this field is used to verify if the SegmentMerger enables payloads for a field if it has payloads // enabled in only some documents - d.add(new Field("f3", "This field has payloads in some docs", Field.Store.NO, Field.Index.ANALYZED)); + d.add(newField("f3", "This field has payloads in some docs", Field.Store.NO, Field.Index.ANALYZED)); // only add payload data for field f2 analyzer.setPayloadData("f2", 1, "somedata".getBytes(), 0, 1); writer.addDocument(d); @@ -130,10 +130,10 @@ writer = new IndexWriter(ram, newIndexWriterConfig( TEST_VERSION_CURRENT, analyzer).setOpenMode(OpenMode.CREATE)); d = new Document(); - d.add(new Field("f1", "This field has no payloads", Field.Store.NO, Field.Index.ANALYZED)); - d.add(new Field("f2", "This field has payloads in all docs", Field.Store.NO, Field.Index.ANALYZED)); - d.add(new Field("f2", "This field has payloads in all docs", Field.Store.NO, Field.Index.ANALYZED)); - d.add(new Field("f3", "This field has payloads in some docs", Field.Store.NO, Field.Index.ANALYZED)); + d.add(newField("f1", "This field has no payloads", Field.Store.NO, Field.Index.ANALYZED)); + d.add(newField("f2", "This field has payloads in all docs", Field.Store.NO, Field.Index.ANALYZED)); + d.add(newField("f2", "This field has payloads in all docs", Field.Store.NO, Field.Index.ANALYZED)); + d.add(newField("f3", "This field has payloads in some docs", Field.Store.NO, Field.Index.ANALYZED)); // add payload data for field f2 and f3 analyzer.setPayloadData("f2", "somedata".getBytes(), 0, 1); analyzer.setPayloadData("f3", "somedata".getBytes(), 0, 3); @@ -196,7 +196,7 @@ byte[] payloadData = generateRandomData(payloadDataLength); Document d = new Document(); - d.add(new Field(fieldName, content, Field.Store.NO, Field.Index.ANALYZED)); + d.add(newField(fieldName, content, Field.Store.NO, Field.Index.ANALYZED)); // add the same document multiple times to have the same payload lengths for all // occurrences within two consecutive skip intervals int offset = 0; @@ -317,7 +317,7 @@ String singleTerm = "lucene"; d = new Document(); - d.add(new Field(fieldName, singleTerm, Field.Store.NO, Field.Index.ANALYZED)); + d.add(newField(fieldName, singleTerm, Field.Store.NO, Field.Index.ANALYZED)); // add a payload whose length is greater than the buffer size of BufferedIndexOutput payloadData = generateRandomData(2000); analyzer.setPayloadData(fieldName, payloadData, 100, 1500); Index: lucene/src/test/org/apache/lucene/index/TestLazyBug.java =================================================================== --- lucene/src/test/org/apache/lucene/index/TestLazyBug.java (revision 995772) +++ lucene/src/test/org/apache/lucene/index/TestLazyBug.java (working copy) @@ -78,7 +78,7 @@ for (int d = 1; d <= NUM_DOCS; d++) { Document doc = new Document(); for (int f = 1; f <= NUM_FIELDS; f++ ) { - doc.add(new Field("f"+f, + doc.add(newField("f"+f, data[f % data.length] + '#' + data[random.nextInt(data.length)], Field.Store.YES, Index: lucene/src/test/org/apache/lucene/index/TestIndexWriterMerging.java =================================================================== --- lucene/src/test/org/apache/lucene/index/TestIndexWriterMerging.java (revision 995772) +++ lucene/src/test/org/apache/lucene/index/TestIndexWriterMerging.java (working copy) @@ -103,7 +103,7 @@ for (int i = start; i < (start + numDocs); i++) { Document temp = new Document(); - temp.add(new Field("count", (""+i), Field.Store.YES, Field.Index.NOT_ANALYZED)); + temp.add(newField("count", (""+i), Field.Store.YES, Field.Index.NOT_ANALYZED)); writer.addDocument(temp); } Index: lucene/src/test/org/apache/lucene/index/TestParallelReader.java =================================================================== --- lucene/src/test/org/apache/lucene/index/TestParallelReader.java (revision 995772) +++ lucene/src/test/org/apache/lucene/index/TestParallelReader.java (working copy) @@ -123,7 +123,7 @@ Directory dir2 = newDirectory(); IndexWriter w2 = new IndexWriter(dir2, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer())); Document d3 = new Document(); - d3.add(new Field("f3", "v1", Field.Store.YES, Field.Index.ANALYZED)); + d3.add(newField("f3", "v1", Field.Store.YES, Field.Index.ANALYZED)); w2.addDocument(d3); w2.close(); @@ -177,14 +177,14 @@ IndexWriter modifier = new IndexWriter(dir1, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer())); ((LogMergePolicy) modifier.getMergePolicy()).setMergeFactor(10); Document d = new Document(); - d.add(new Field("f1", "v1", Field.Store.YES, Field.Index.ANALYZED)); + d.add(newField("f1", "v1", Field.Store.YES, Field.Index.ANALYZED)); modifier.addDocument(d); modifier.close(); modifier = new IndexWriter(dir2, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer())); ((LogMergePolicy) modifier.getMergePolicy()).setMergeFactor(10); d = new Document(); - d.add(new Field("f2", "v2", Field.Store.YES, Field.Index.ANALYZED)); + d.add(newField("f2", "v2", Field.Store.YES, Field.Index.ANALYZED)); modifier.addDocument(d); modifier.close(); @@ -241,16 +241,16 @@ dir = newDirectory(); IndexWriter w = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer())); Document d1 = new Document(); - d1.add(new Field("f1", "v1", Field.Store.YES, Field.Index.ANALYZED)); - d1.add(new Field("f2", "v1", Field.Store.YES, Field.Index.ANALYZED)); - d1.add(new Field("f3", "v1", Field.Store.YES, Field.Index.ANALYZED)); - d1.add(new Field("f4", "v1", Field.Store.YES, Field.Index.ANALYZED)); + d1.add(newField("f1", "v1", Field.Store.YES, Field.Index.ANALYZED)); + d1.add(newField("f2", "v1", Field.Store.YES, Field.Index.ANALYZED)); + d1.add(newField("f3", "v1", Field.Store.YES, Field.Index.ANALYZED)); + d1.add(newField("f4", "v1", Field.Store.YES, Field.Index.ANALYZED)); w.addDocument(d1); Document d2 = new Document(); - d2.add(new Field("f1", "v2", Field.Store.YES, Field.Index.ANALYZED)); - d2.add(new Field("f2", "v2", Field.Store.YES, Field.Index.ANALYZED)); - d2.add(new Field("f3", "v2", Field.Store.YES, Field.Index.ANALYZED)); - d2.add(new Field("f4", "v2", Field.Store.YES, Field.Index.ANALYZED)); + d2.add(newField("f1", "v2", Field.Store.YES, Field.Index.ANALYZED)); + d2.add(newField("f2", "v2", Field.Store.YES, Field.Index.ANALYZED)); + d2.add(newField("f3", "v2", Field.Store.YES, Field.Index.ANALYZED)); + d2.add(newField("f4", "v2", Field.Store.YES, Field.Index.ANALYZED)); w.addDocument(d2); w.close(); @@ -271,12 +271,12 @@ Directory dir1 = newDirectory(); IndexWriter w1 = new IndexWriter(dir1, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer())); Document d1 = new Document(); - d1.add(new Field("f1", "v1", Field.Store.YES, Field.Index.ANALYZED)); - d1.add(new Field("f2", "v1", Field.Store.YES, Field.Index.ANALYZED)); + d1.add(newField("f1", "v1", Field.Store.YES, Field.Index.ANALYZED)); + d1.add(newField("f2", "v1", Field.Store.YES, Field.Index.ANALYZED)); w1.addDocument(d1); Document d2 = new Document(); - d2.add(new Field("f1", "v2", Field.Store.YES, Field.Index.ANALYZED)); - d2.add(new Field("f2", "v2", Field.Store.YES, Field.Index.ANALYZED)); + d2.add(newField("f1", "v2", Field.Store.YES, Field.Index.ANALYZED)); + d2.add(newField("f2", "v2", Field.Store.YES, Field.Index.ANALYZED)); w1.addDocument(d2); w1.close(); return dir1; @@ -286,12 +286,12 @@ Directory dir2 = newDirectory(); IndexWriter w2 = new IndexWriter(dir2, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer())); Document d3 = new Document(); - d3.add(new Field("f3", "v1", Field.Store.YES, Field.Index.ANALYZED)); - d3.add(new Field("f4", "v1", Field.Store.YES, Field.Index.ANALYZED)); + d3.add(newField("f3", "v1", Field.Store.YES, Field.Index.ANALYZED)); + d3.add(newField("f4", "v1", Field.Store.YES, Field.Index.ANALYZED)); w2.addDocument(d3); Document d4 = new Document(); - d4.add(new Field("f3", "v2", Field.Store.YES, Field.Index.ANALYZED)); - d4.add(new Field("f4", "v2", Field.Store.YES, Field.Index.ANALYZED)); + d4.add(newField("f3", "v2", Field.Store.YES, Field.Index.ANALYZED)); + d4.add(newField("f4", "v2", Field.Store.YES, Field.Index.ANALYZED)); w2.addDocument(d4); w2.close(); return dir2; Index: lucene/src/test/org/apache/lucene/index/codecs/preflex/TestSurrogates.java =================================================================== --- lucene/src/test/org/apache/lucene/index/codecs/preflex/TestSurrogates.java (revision 995772) +++ lucene/src/test/org/apache/lucene/index/codecs/preflex/TestSurrogates.java (working copy) @@ -297,7 +297,7 @@ uniqueTerms.add(term); fieldTerms.add(new Term(field, term)); Document doc = new Document(); - doc.add(new Field(field, term, Field.Store.NO, Field.Index.NOT_ANALYZED)); + doc.add(newField(field, term, Field.Store.NO, Field.Index.NOT_ANALYZED)); w.addDocument(doc); } uniqueTermCount += uniqueTerms.size(); Index: lucene/src/test/org/apache/lucene/index/TestDeletionPolicy.java =================================================================== --- lucene/src/test/org/apache/lucene/index/TestDeletionPolicy.java (revision 995772) +++ lucene/src/test/org/apache/lucene/index/TestDeletionPolicy.java (working copy) @@ -795,7 +795,7 @@ private void addDoc(IndexWriter writer) throws IOException { Document doc = new Document(); - doc.add(new Field("content", "aaa", Field.Store.NO, Field.Index.ANALYZED)); + doc.add(newField("content", "aaa", Field.Store.NO, Field.Index.ANALYZED)); writer.addDocument(doc); } } Index: lucene/src/test/org/apache/lucene/index/TestParallelReaderEmptyIndex.java =================================================================== --- lucene/src/test/org/apache/lucene/index/TestParallelReaderEmptyIndex.java (revision 995772) +++ lucene/src/test/org/apache/lucene/index/TestParallelReaderEmptyIndex.java (working copy) @@ -79,10 +79,10 @@ { IndexWriter iw = new IndexWriter(rd1, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer())); Document doc = new Document(); - doc.add(new Field("test", "", Store.NO, Index.ANALYZED, + doc.add(newField("test", "", Store.NO, Index.ANALYZED, TermVector.YES)); iw.addDocument(doc); - doc.add(new Field("test", "", Store.NO, Index.ANALYZED, + doc.add(newField("test", "", Store.NO, Index.ANALYZED, TermVector.NO)); iw.addDocument(doc); iw.close(); Index: lucene/src/test/org/apache/lucene/index/TestBackwardsCompatibility.java =================================================================== --- lucene/src/test/org/apache/lucene/index/TestBackwardsCompatibility.java (revision 995772) +++ lucene/src/test/org/apache/lucene/index/TestBackwardsCompatibility.java (working copy) @@ -479,7 +479,7 @@ dirName = fullDir(dirName); Directory dir = FSDirectory.open(new File(dirName)); - IndexWriterConfig conf = newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer()).setMaxBufferedDocs(10); + IndexWriterConfig conf = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer()).setMaxBufferedDocs(10); ((LogMergePolicy) conf.getMergePolicy()).setUseCompoundFile(doCFS); ((LogMergePolicy) conf.getMergePolicy()).setUseCompoundDocStore(doCFS); IndexWriter writer = new IndexWriter(dir, conf); @@ -491,7 +491,7 @@ writer.close(); // open fresh writer so we get no prx file in the added segment - conf = newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer()).setMaxBufferedDocs(10); + conf = new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer()).setMaxBufferedDocs(10); ((LogMergePolicy) conf.getMergePolicy()).setUseCompoundFile(doCFS); ((LogMergePolicy) conf.getMergePolicy()).setUseCompoundDocStore(doCFS); writer = new IndexWriter(dir, conf); Index: lucene/src/test/org/apache/lucene/index/TestLazyProxSkipping.java =================================================================== --- lucene/src/test/org/apache/lucene/index/TestLazyProxSkipping.java (revision 995772) +++ lucene/src/test/org/apache/lucene/index/TestLazyProxSkipping.java (working copy) @@ -85,7 +85,7 @@ content = this.term3 + " " + this.term2; } - doc.add(new Field(this.field, content, Field.Store.YES, Field.Index.ANALYZED)); + doc.add(newField(this.field, content, Field.Store.YES, Field.Index.ANALYZED)); writer.addDocument(doc); } @@ -129,7 +129,7 @@ IndexWriter writer = new IndexWriter(directory, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer())); for (int i = 0; i < 10; i++) { Document doc = new Document(); - doc.add(new Field(this.field, "a b", Field.Store.YES, Field.Index.ANALYZED)); + doc.add(newField(this.field, "a b", Field.Store.YES, Field.Index.ANALYZED)); writer.addDocument(doc); } Index: lucene/src/test/org/apache/lucene/index/TestNorms.java =================================================================== --- lucene/src/test/org/apache/lucene/index/TestNorms.java (revision 995772) +++ lucene/src/test/org/apache/lucene/index/TestNorms.java (working copy) @@ -208,7 +208,7 @@ Document d = new Document(); float boost = nextNorm(); for (int i = 0; i < 10; i++) { - Field f = new Field("f"+i,"v"+i,Store.NO,Index.NOT_ANALYZED); + Field f = newField("f"+i,"v"+i,Store.NO,Index.NOT_ANALYZED); f.setBoost(boost); d.add(f); } Index: lucene/src/test/org/apache/lucene/TestSearchForDuplicates.java =================================================================== --- lucene/src/test/org/apache/lucene/TestSearchForDuplicates.java (revision 995772) +++ lucene/src/test/org/apache/lucene/TestSearchForDuplicates.java (working copy) @@ -89,11 +89,11 @@ for (int j = 0; j < MAX_DOCS; j++) { Document d = new Document(); - d.add(new Field(PRIORITY_FIELD, HIGH_PRIORITY, Field.Store.YES, Field.Index.ANALYZED)); + d.add(newField(PRIORITY_FIELD, HIGH_PRIORITY, Field.Store.YES, Field.Index.ANALYZED)); // NOTE: this ID_FIELD produces no tokens since // MockAnalyzer discards numbers - d.add(new Field(ID_FIELD, Integer.toString(j), Field.Store.YES, Field.Index.ANALYZED)); + d.add(newField(ID_FIELD, Integer.toString(j), Field.Store.YES, Field.Index.ANALYZED)); writer.addDocument(d); } writer.close(); Index: lucene/src/test/org/apache/lucene/store/TestLockFactory.java =================================================================== --- lucene/src/test/org/apache/lucene/store/TestLockFactory.java (revision 995688) +++ lucene/src/test/org/apache/lucene/store/TestLockFactory.java (working copy) @@ -405,7 +405,7 @@ private void addDoc(IndexWriter writer) throws IOException { Document doc = new Document(); - doc.add(new Field("content", "aaa", Field.Store.NO, Field.Index.ANALYZED)); + doc.add(newField("content", "aaa", Field.Store.NO, Field.Index.ANALYZED)); writer.addDocument(doc); } } Index: lucene/src/test/org/apache/lucene/store/TestRAMDirectory.java =================================================================== --- lucene/src/test/org/apache/lucene/store/TestRAMDirectory.java (revision 995688) +++ lucene/src/test/org/apache/lucene/store/TestRAMDirectory.java (working copy) @@ -58,7 +58,7 @@ Document doc = null; for (int i = 0; i < docsToAdd; i++) { doc = new Document(); - doc.add(new Field("content", English.intToEnglish(i).trim(), Field.Store.YES, Field.Index.NOT_ANALYZED)); + doc.add(newField("content", English.intToEnglish(i).trim(), Field.Store.YES, Field.Index.NOT_ANALYZED)); writer.addDocument(doc); } assertEquals(docsToAdd, writer.maxDoc()); @@ -118,7 +118,7 @@ public void run() { for (int j=1; j