Index: solr/core/src/java/org/apache/solr/schema/FieldTypePluginLoader.java =================================================================== --- solr/core/src/java/org/apache/solr/schema/FieldTypePluginLoader.java (revision 1368355) +++ solr/core/src/java/org/apache/solr/schema/FieldTypePluginLoader.java (working copy) @@ -293,10 +293,8 @@ protected void init(CharFilterFactory plugin, Node node) throws Exception { if( plugin != null ) { final Map params = DOMUtil.toMapExcept(node.getAttributes(),"class"); + validateConfiguredVersion(params, plugin.getClass().getSimpleName()); - String configuredVersion = params.remove(LUCENE_MATCH_VERSION_PARAM); - plugin.setLuceneMatchVersion(parseConfiguredVersion(configuredVersion, plugin.getClass().getSimpleName())); - plugin.init( params ); charFilters.add( plugin ); } @@ -327,10 +325,8 @@ "The schema defines multiple tokenizers for: "+node ); } final Map params = DOMUtil.toMapExcept(node.getAttributes(),"class"); + validateConfiguredVersion(params, plugin.getClass().getSimpleName()); - String configuredVersion = params.remove(LUCENE_MATCH_VERSION_PARAM); - plugin.setLuceneMatchVersion(parseConfiguredVersion(configuredVersion, plugin.getClass().getSimpleName())); - plugin.init( params ); tokenizers.add( plugin ); } @@ -360,10 +356,8 @@ protected void init(TokenFilterFactory plugin, Node node) throws Exception { if( plugin != null ) { final Map params = DOMUtil.toMapExcept(node.getAttributes(),"class"); + validateConfiguredVersion(params, plugin.getClass().getSimpleName()); - String configuredVersion = params.remove(LUCENE_MATCH_VERSION_PARAM); - plugin.setLuceneMatchVersion(parseConfiguredVersion(configuredVersion, plugin.getClass().getSimpleName())); - plugin.init( params ); filters.add( plugin ); } @@ -380,7 +374,8 @@ tokenizers.get(0), filters.toArray(new TokenFilterFactory[filters.size()])); } - private Version parseConfiguredVersion(String configuredVersion, String pluginClassName) { + private void validateConfiguredVersion(Map params, String pluginClassName) { + String configuredVersion = params.get(LUCENE_MATCH_VERSION_PARAM); Version version = (configuredVersion != null) ? Config.parseLuceneVersionString(configuredVersion) : schema.getDefaultLuceneMatchVersion(); @@ -389,7 +384,10 @@ " emulation. You should at some point declare and reindex to at least 4.0, because " + "3.x emulation is deprecated and will be removed in 5.0"); } - return version; + + if (configuredVersion == null) { + params.put(LUCENE_MATCH_VERSION_PARAM, version.toString()); + } } } Index: lucene/analysis/kuromoji/src/test/org/apache/lucene/analysis/ja/TestJapaneseBaseFormFilterFactory.java =================================================================== --- lucene/analysis/kuromoji/src/test/org/apache/lucene/analysis/ja/TestJapaneseBaseFormFilterFactory.java (revision 1368355) +++ lucene/analysis/kuromoji/src/test/org/apache/lucene/analysis/ja/TestJapaneseBaseFormFilterFactory.java (working copy) @@ -31,9 +31,7 @@ public class TestJapaneseBaseFormFilterFactory extends BaseTokenStreamTestCase { public void testBasics() throws IOException { JapaneseTokenizerFactory tokenizerFactory = new JapaneseTokenizerFactory(); - tokenizerFactory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - Map args = Collections.emptyMap(); - tokenizerFactory.init(args); + tokenizerFactory.init(TEST_VERSION_CURRENT_MAP()); tokenizerFactory.inform(new StringMockResourceLoader("")); TokenStream ts = tokenizerFactory.create(new StringReader("それはまだ実験段階にあります")); JapaneseBaseFormFilterFactory factory = new JapaneseBaseFormFilterFactory(); Index: lucene/analysis/kuromoji/src/test/org/apache/lucene/analysis/ja/TestJapaneseTokenizerFactory.java =================================================================== --- lucene/analysis/kuromoji/src/test/org/apache/lucene/analysis/ja/TestJapaneseTokenizerFactory.java (revision 1368355) +++ lucene/analysis/kuromoji/src/test/org/apache/lucene/analysis/ja/TestJapaneseTokenizerFactory.java (working copy) @@ -32,9 +32,7 @@ public class TestJapaneseTokenizerFactory extends BaseTokenStreamTestCase { public void testSimple() throws IOException { JapaneseTokenizerFactory factory = new JapaneseTokenizerFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - Map args = Collections.emptyMap(); - factory.init(args); + factory.init(TEST_VERSION_CURRENT_MAP()); factory.inform(new StringMockResourceLoader("")); TokenStream ts = factory.create(new StringReader("これは本ではない")); assertTokenStreamContents(ts, @@ -49,9 +47,7 @@ */ public void testDefaults() throws IOException { JapaneseTokenizerFactory factory = new JapaneseTokenizerFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - Map args = Collections.emptyMap(); - factory.init(args); + factory.init(TEST_VERSION_CURRENT_MAP()); factory.inform(new StringMockResourceLoader("")); TokenStream ts = factory.create(new StringReader("シニアソフトウェアエンジニア")); assertTokenStreamContents(ts, Index: lucene/analysis/kuromoji/src/test/org/apache/lucene/analysis/ja/TestJapanesePartOfSpeechStopFilterFactory.java =================================================================== --- lucene/analysis/kuromoji/src/test/org/apache/lucene/analysis/ja/TestJapanesePartOfSpeechStopFilterFactory.java (revision 1368355) +++ lucene/analysis/kuromoji/src/test/org/apache/lucene/analysis/ja/TestJapanesePartOfSpeechStopFilterFactory.java (working copy) @@ -36,15 +36,13 @@ "動詞-自立\n"; JapaneseTokenizerFactory tokenizerFactory = new JapaneseTokenizerFactory(); - tokenizerFactory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - Map tokenizerArgs = Collections.emptyMap(); + Map tokenizerArgs = TEST_VERSION_CURRENT_MAP(); tokenizerFactory.init(tokenizerArgs); tokenizerFactory.inform(new StringMockResourceLoader("")); TokenStream ts = tokenizerFactory.create(new StringReader("私は制限スピードを超える。")); JapanesePartOfSpeechStopFilterFactory factory = new JapanesePartOfSpeechStopFilterFactory(); - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); args.put("tags", "stoptags.txt"); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init(args); factory.inform(new StringMockResourceLoader(tags)); ts = factory.create(ts); Index: lucene/analysis/icu/src/test/org/apache/lucene/analysis/icu/TestICUFoldingFilterFactory.java =================================================================== --- lucene/analysis/icu/src/test/org/apache/lucene/analysis/icu/TestICUFoldingFilterFactory.java (revision 1368355) +++ lucene/analysis/icu/src/test/org/apache/lucene/analysis/icu/TestICUFoldingFilterFactory.java (working copy) @@ -32,7 +32,7 @@ public void test() throws Exception { Reader reader = new StringReader("Résumé"); ICUFoldingFilterFactory factory = new ICUFoldingFilterFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); + factory.init(TEST_VERSION_CURRENT_MAP()); Tokenizer tokenizer = new WhitespaceTokenizer(TEST_VERSION_CURRENT, reader); TokenStream stream = factory.create(tokenizer); assertTokenStreamContents(stream, new String[] { "resume" }); Index: lucene/analysis/icu/src/test/org/apache/lucene/analysis/icu/TestICUNormalizer2FilterFactory.java =================================================================== --- lucene/analysis/icu/src/test/org/apache/lucene/analysis/icu/TestICUNormalizer2FilterFactory.java (revision 1368355) +++ lucene/analysis/icu/src/test/org/apache/lucene/analysis/icu/TestICUNormalizer2FilterFactory.java (working copy) @@ -34,9 +34,7 @@ public void testDefaults() throws Exception { Reader reader = new StringReader("This is a Test"); ICUNormalizer2FilterFactory factory = new ICUNormalizer2FilterFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - Map args = Collections.emptyMap(); - factory.init(args); + factory.init(TEST_VERSION_CURRENT_MAP()); Tokenizer tokenizer = new WhitespaceTokenizer(TEST_VERSION_CURRENT, reader); TokenStream stream = factory.create(tokenizer); assertTokenStreamContents(stream, new String[] { "this", "is", "a", "test" }); Index: lucene/analysis/common/src/test/org/apache/lucene/analysis/commongrams/TestCommonGramsFilterFactory.java =================================================================== --- lucene/analysis/common/src/test/org/apache/lucene/analysis/commongrams/TestCommonGramsFilterFactory.java (revision 1368355) +++ lucene/analysis/common/src/test/org/apache/lucene/analysis/commongrams/TestCommonGramsFilterFactory.java (working copy) @@ -42,10 +42,9 @@ ResourceLoader loader = new ClasspathResourceLoader(TestStopFilter.class); assertTrue("loader is null and it shouldn't be", loader != null); CommonGramsFilterFactory factory = new CommonGramsFilterFactory(); - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); args.put("words", "stop-1.txt"); args.put("ignoreCase", "true"); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init(args); factory.inform(loader); CharArraySet words = factory.getCommonWords(); @@ -57,7 +56,6 @@ factory = new CommonGramsFilterFactory(); args.put("words", "stop-1.txt, stop-2.txt"); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init(args); factory.inform(loader); words = factory.getCommonWords(); @@ -70,7 +68,6 @@ factory = new CommonGramsFilterFactory(); args.put("words", "stop-snowball.txt"); args.put("format", "snowball"); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init(args); factory.inform(loader); words = factory.getCommonWords(); @@ -92,9 +89,7 @@ ResourceLoader loader = new ClasspathResourceLoader(TestStopFilter.class); assertTrue("loader is null and it shouldn't be", loader != null); CommonGramsFilterFactory factory = new CommonGramsFilterFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - Map args = Collections.emptyMap(); - factory.init(args); + factory.init(TEST_VERSION_CURRENT_MAP()); factory.inform(loader); CharArraySet words = factory.getCommonWords(); assertTrue("words is null and it shouldn't be", words != null); Index: lucene/analysis/common/src/test/org/apache/lucene/analysis/commongrams/TestCommonGramsQueryFilterFactory.java =================================================================== --- lucene/analysis/common/src/test/org/apache/lucene/analysis/commongrams/TestCommonGramsQueryFilterFactory.java (revision 1368355) +++ lucene/analysis/common/src/test/org/apache/lucene/analysis/commongrams/TestCommonGramsQueryFilterFactory.java (working copy) @@ -42,10 +42,9 @@ ResourceLoader loader = new ClasspathResourceLoader(TestStopFilter.class); assertTrue("loader is null and it shouldn't be", loader != null); CommonGramsQueryFilterFactory factory = new CommonGramsQueryFilterFactory(); - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); args.put("words", "stop-1.txt"); args.put("ignoreCase", "true"); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init(args); factory.inform(loader); CharArraySet words = factory.getCommonWords(); @@ -57,7 +56,6 @@ factory = new CommonGramsQueryFilterFactory(); args.put("words", "stop-1.txt, stop-2.txt"); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init(args); factory.inform(loader); words = factory.getCommonWords(); @@ -68,7 +66,6 @@ .isIgnoreCase() == true); factory = new CommonGramsQueryFilterFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); args.put("words", "stop-snowball.txt"); args.put("format", "snowball"); factory.init(args); @@ -92,9 +89,7 @@ ResourceLoader loader = new ClasspathResourceLoader(TestStopFilter.class); assertTrue("loader is null and it shouldn't be", loader != null); CommonGramsQueryFilterFactory factory = new CommonGramsQueryFilterFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - Map args = Collections.emptyMap(); - factory.init(args); + factory.init(TEST_VERSION_CURRENT_MAP()); factory.inform(loader); CharArraySet words = factory.getCommonWords(); assertTrue("words is null and it shouldn't be", words != null); Index: lucene/analysis/common/src/test/org/apache/lucene/analysis/compound/TestDictionaryCompoundWordTokenFilterFactory.java =================================================================== --- lucene/analysis/common/src/test/org/apache/lucene/analysis/compound/TestDictionaryCompoundWordTokenFilterFactory.java (revision 1368355) +++ lucene/analysis/common/src/test/org/apache/lucene/analysis/compound/TestDictionaryCompoundWordTokenFilterFactory.java (working copy) @@ -41,9 +41,8 @@ Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); DictionaryCompoundWordTokenFilterFactory factory = new DictionaryCompoundWordTokenFilterFactory(); ResourceLoader loader = new ClasspathResourceLoader(getClass()); - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); args.put("dictionary", "compoundDictionary.txt"); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init(args); factory.inform(loader); TokenStream stream = factory.create(tokenizer); Index: lucene/analysis/common/src/test/org/apache/lucene/analysis/compound/TestHyphenationCompoundWordTokenFilterFactory.java =================================================================== --- lucene/analysis/common/src/test/org/apache/lucene/analysis/compound/TestHyphenationCompoundWordTokenFilterFactory.java (revision 1368355) +++ lucene/analysis/common/src/test/org/apache/lucene/analysis/compound/TestHyphenationCompoundWordTokenFilterFactory.java (working copy) @@ -41,10 +41,9 @@ Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); HyphenationCompoundWordTokenFilterFactory factory = new HyphenationCompoundWordTokenFilterFactory(); ResourceLoader loader = new ClasspathResourceLoader(getClass()); - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); args.put("hyphenator", "da_UTF8.xml"); args.put("dictionary", "da_compoundDictionary.txt"); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init(args); factory.inform(loader); TokenStream stream = factory.create(tokenizer); @@ -65,11 +64,10 @@ Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); HyphenationCompoundWordTokenFilterFactory factory = new HyphenationCompoundWordTokenFilterFactory(); ResourceLoader loader = new ClasspathResourceLoader(getClass()); - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); args.put("hyphenator", "da_UTF8.xml"); args.put("minSubwordSize", "2"); args.put("maxSubwordSize", "4"); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init(args); factory.inform(loader); TokenStream stream = factory.create(tokenizer); Index: lucene/analysis/common/src/test/org/apache/lucene/analysis/core/TestAllAnalyzersHaveFactories.java =================================================================== --- lucene/analysis/common/src/test/org/apache/lucene/analysis/core/TestAllAnalyzersHaveFactories.java (revision 1368355) +++ lucene/analysis/common/src/test/org/apache/lucene/analysis/core/TestAllAnalyzersHaveFactories.java (working copy) @@ -27,6 +27,7 @@ import java.util.Map; import java.util.Set; +import org.apache.lucene.analysis.BaseTokenStreamTestCase; import org.apache.lucene.analysis.CachingTokenFilter; import org.apache.lucene.analysis.CharFilter; import org.apache.lucene.analysis.EmptyTokenizer; @@ -124,8 +125,7 @@ TokenizerFactory instance = TokenizerFactory.forName(simpleName); assertNotNull(instance); try { - instance.setLuceneMatchVersion(TEST_VERSION_CURRENT); - instance.init(Collections.emptyMap()); + instance.init(BaseTokenStreamTestCase.TEST_VERSION_CURRENT_MAP()); if (instance instanceof ResourceLoaderAware) { ((ResourceLoaderAware) instance).inform(loader); } @@ -140,8 +140,7 @@ TokenFilterFactory instance = TokenFilterFactory.forName(simpleName); assertNotNull(instance); try { - instance.setLuceneMatchVersion(TEST_VERSION_CURRENT); - instance.init(Collections.emptyMap()); + instance.init(BaseTokenStreamTestCase.TEST_VERSION_CURRENT_MAP()); if (instance instanceof ResourceLoaderAware) { ((ResourceLoaderAware) instance).inform(loader); } @@ -160,8 +159,7 @@ CharFilterFactory instance = CharFilterFactory.forName(simpleName); assertNotNull(instance); try { - instance.setLuceneMatchVersion(TEST_VERSION_CURRENT); - instance.init(Collections.emptyMap()); + instance.init(BaseTokenStreamTestCase.TEST_VERSION_CURRENT_MAP()); if (instance instanceof ResourceLoaderAware) { ((ResourceLoaderAware) instance).inform(loader); } Index: lucene/analysis/common/src/test/org/apache/lucene/analysis/core/TestFactories.java =================================================================== --- lucene/analysis/common/src/test/org/apache/lucene/analysis/core/TestFactories.java (revision 1368355) +++ lucene/analysis/common/src/test/org/apache/lucene/analysis/core/TestFactories.java (working copy) @@ -117,8 +117,7 @@ private boolean initialize(AbstractAnalysisFactory factory) { boolean success = false; try { - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - factory.init(Collections.emptyMap()); + factory.init(BaseTokenStreamTestCase.TEST_VERSION_CURRENT_MAP()); success = true; } catch (IllegalArgumentException ignored) { // its ok if we dont provide the right parameters to throw this Index: lucene/analysis/common/src/test/org/apache/lucene/analysis/core/TestTypeTokenFilterFactory.java =================================================================== --- lucene/analysis/common/src/test/org/apache/lucene/analysis/core/TestTypeTokenFilterFactory.java (revision 1368355) +++ lucene/analysis/common/src/test/org/apache/lucene/analysis/core/TestTypeTokenFilterFactory.java (working copy) @@ -36,10 +36,9 @@ public void testInform() throws Exception { ResourceLoader loader = new ClasspathResourceLoader(getClass()); TypeTokenFilterFactory factory = new TypeTokenFilterFactory(); - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); args.put("types", "stoptypes-1.txt"); args.put("enablePositionIncrements", "true"); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init(args); factory.inform(loader); Set types = factory.getStopTypes(); @@ -62,10 +61,9 @@ @Test public void testCreationWithBlackList() throws Exception { TypeTokenFilterFactory typeTokenFilterFactory = new TypeTokenFilterFactory(); - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); args.put("types", "stoptypes-1.txt, stoptypes-2.txt"); args.put("enablePositionIncrements", "false"); - typeTokenFilterFactory.setLuceneMatchVersion(TEST_VERSION_CURRENT); typeTokenFilterFactory.init(args); NumericTokenStream input = new NumericTokenStream(); input.setIntValue(123); @@ -75,11 +73,10 @@ @Test public void testCreationWithWhiteList() throws Exception { TypeTokenFilterFactory typeTokenFilterFactory = new TypeTokenFilterFactory(); - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); args.put("types", "stoptypes-1.txt, stoptypes-2.txt"); args.put("enablePositionIncrements", "false"); args.put("useWhitelist","true"); - typeTokenFilterFactory.setLuceneMatchVersion(TEST_VERSION_CURRENT); typeTokenFilterFactory.init(args); NumericTokenStream input = new NumericTokenStream(); input.setIntValue(123); @@ -90,9 +87,8 @@ public void testMissingTypesParameter() throws Exception { try { TypeTokenFilterFactory typeTokenFilterFactory = new TypeTokenFilterFactory(); - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); args.put("enablePositionIncrements", "false"); - typeTokenFilterFactory.setLuceneMatchVersion(TEST_VERSION_CURRENT); typeTokenFilterFactory.init(args); typeTokenFilterFactory.inform(new ClasspathResourceLoader(getClass())); fail("not supplying 'types' parameter should cause an IllegalArgumentException"); Index: lucene/analysis/common/src/test/org/apache/lucene/analysis/core/TestStopFilterFactory.java =================================================================== --- lucene/analysis/common/src/test/org/apache/lucene/analysis/core/TestStopFilterFactory.java (revision 1368355) +++ lucene/analysis/common/src/test/org/apache/lucene/analysis/core/TestStopFilterFactory.java (working copy) @@ -35,10 +35,9 @@ ResourceLoader loader = new ClasspathResourceLoader(getClass()); assertTrue("loader is null and it shouldn't be", loader != null); StopFilterFactory factory = new StopFilterFactory(); - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); args.put("words", "stop-1.txt"); args.put("ignoreCase", "true"); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init(args); factory.inform(loader); CharArraySet words = factory.getStopWords(); @@ -48,7 +47,6 @@ factory = new StopFilterFactory(); args.put("words", "stop-1.txt, stop-2.txt"); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init(args); factory.inform(loader); words = factory.getStopWords(); @@ -57,7 +55,6 @@ assertTrue(factory.isIgnoreCase() + " does not equal: " + true, factory.isIgnoreCase() == true); factory = new StopFilterFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); args.put("words", "stop-snowball.txt"); args.put("format", "snowball"); factory.init(args); Index: lucene/analysis/common/src/test/org/apache/lucene/analysis/hi/TestHindiFilters.java =================================================================== --- lucene/analysis/common/src/test/org/apache/lucene/analysis/hi/TestHindiFilters.java (revision 1368355) +++ lucene/analysis/common/src/test/org/apache/lucene/analysis/hi/TestHindiFilters.java (working copy) @@ -38,10 +38,8 @@ public void testIndicNormalizer() throws Exception { Reader reader = new StringReader("ত্‍ अाैर"); StandardTokenizerFactory factory = new StandardTokenizerFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); IndicNormalizationFilterFactory filterFactory = new IndicNormalizationFilterFactory(); - filterFactory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - Map args = Collections.emptyMap(); + Map args = TEST_VERSION_CURRENT_MAP(); factory.init(args); filterFactory.init(args); Tokenizer tokenizer = factory.create(reader); @@ -55,11 +53,9 @@ public void testHindiNormalizer() throws Exception { Reader reader = new StringReader("क़िताब"); StandardTokenizerFactory factory = new StandardTokenizerFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); IndicNormalizationFilterFactory indicFilterFactory = new IndicNormalizationFilterFactory(); HindiNormalizationFilterFactory hindiFilterFactory = new HindiNormalizationFilterFactory(); - hindiFilterFactory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - Map args = Collections.emptyMap(); + Map args = TEST_VERSION_CURRENT_MAP(); factory.init(args); hindiFilterFactory.init(args); Tokenizer tokenizer = factory.create(reader); @@ -74,12 +70,10 @@ public void testStemmer() throws Exception { Reader reader = new StringReader("किताबें"); StandardTokenizerFactory factory = new StandardTokenizerFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); IndicNormalizationFilterFactory indicFilterFactory = new IndicNormalizationFilterFactory(); HindiNormalizationFilterFactory hindiFilterFactory = new HindiNormalizationFilterFactory(); HindiStemFilterFactory stemFactory = new HindiStemFilterFactory(); - stemFactory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - Map args = Collections.emptyMap(); + Map args = TEST_VERSION_CURRENT_MAP(); factory.init(args); stemFactory.init(args); Tokenizer tokenizer = factory.create(reader); Index: lucene/analysis/common/src/test/org/apache/lucene/analysis/standard/TestStandardFactories.java =================================================================== --- lucene/analysis/common/src/test/org/apache/lucene/analysis/standard/TestStandardFactories.java (revision 1368355) +++ lucene/analysis/common/src/test/org/apache/lucene/analysis/standard/TestStandardFactories.java (working copy) @@ -43,9 +43,7 @@ public void testStandardTokenizer() throws Exception { Reader reader = new StringReader("Wha\u0301t's this thing do?"); StandardTokenizerFactory factory = new StandardTokenizerFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - Map args = Collections.emptyMap(); - factory.init(args); + factory.init(TEST_VERSION_CURRENT_MAP()); Tokenizer stream = factory.create(reader); assertTokenStreamContents(stream, new String[] {"Wha\u0301t's", "this", "thing", "do" }); @@ -59,10 +57,9 @@ String longWord = builder.toString(); String content = "one two three " + longWord + " four five six"; Reader reader = new StringReader(content); - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); args.put("maxTokenLength", "1000"); StandardTokenizerFactory factory = new StandardTokenizerFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init(args); Tokenizer stream = factory.create(reader); assertTokenStreamContents(stream, @@ -75,9 +72,7 @@ public void testClassicTokenizer() throws Exception { Reader reader = new StringReader("What's this thing do?"); ClassicTokenizerFactory factory = new ClassicTokenizerFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - Map args = Collections.emptyMap(); - factory.init(args); + factory.init(TEST_VERSION_CURRENT_MAP()); Tokenizer stream = factory.create(reader); assertTokenStreamContents(stream, new String[] {"What's", "this", "thing", "do" }); @@ -91,10 +86,9 @@ String longWord = builder.toString(); String content = "one two three " + longWord + " four five six"; Reader reader = new StringReader(content); - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); args.put("maxTokenLength", "1000"); ClassicTokenizerFactory factory = new ClassicTokenizerFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init(args); Tokenizer stream = factory.create(reader); assertTokenStreamContents(stream, @@ -107,12 +101,9 @@ public void testStandardFilter() throws Exception { Reader reader = new StringReader("What's this thing do?"); ClassicTokenizerFactory factory = new ClassicTokenizerFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - Map args = Collections.emptyMap(); - factory.init(args); + factory.init(TEST_VERSION_CURRENT_MAP()); ClassicFilterFactory filterFactory = new ClassicFilterFactory(); - filterFactory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - filterFactory.init(args); + filterFactory.init(TEST_VERSION_CURRENT_MAP()); Tokenizer tokenizer = factory.create(reader); TokenStream stream = filterFactory.create(tokenizer); assertTokenStreamContents(stream, @@ -125,9 +116,7 @@ public void testKeywordTokenizer() throws Exception { Reader reader = new StringReader("What's this thing do?"); KeywordTokenizerFactory factory = new KeywordTokenizerFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - Map args = Collections.emptyMap(); - factory.init(args); + factory.init(TEST_VERSION_CURRENT_MAP()); Tokenizer stream = factory.create(reader); assertTokenStreamContents(stream, new String[] {"What's this thing do?"}); @@ -139,9 +128,7 @@ public void testWhitespaceTokenizer() throws Exception { Reader reader = new StringReader("What's this thing do?"); WhitespaceTokenizerFactory factory = new WhitespaceTokenizerFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - Map args = Collections.emptyMap(); - factory.init(args); + factory.init(TEST_VERSION_CURRENT_MAP()); Tokenizer stream = factory.create(reader); assertTokenStreamContents(stream, new String[] {"What's", "this", "thing", "do?"}); @@ -153,9 +140,7 @@ public void testLetterTokenizer() throws Exception { Reader reader = new StringReader("What's this thing do?"); LetterTokenizerFactory factory = new LetterTokenizerFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - Map args = Collections.emptyMap(); - factory.init(args); + factory.init(TEST_VERSION_CURRENT_MAP()); Tokenizer stream = factory.create(reader); assertTokenStreamContents(stream, new String[] {"What", "s", "this", "thing", "do"}); @@ -167,9 +152,7 @@ public void testLowerCaseTokenizer() throws Exception { Reader reader = new StringReader("What's this thing do?"); LowerCaseTokenizerFactory factory = new LowerCaseTokenizerFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - Map args = Collections.emptyMap(); - factory.init(args); + factory.init(TEST_VERSION_CURRENT_MAP()); Tokenizer stream = factory.create(reader); assertTokenStreamContents(stream, new String[] {"what", "s", "this", "thing", "do"}); @@ -182,9 +165,7 @@ Reader reader = new StringReader("Česká"); Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); ASCIIFoldingFilterFactory factory = new ASCIIFoldingFilterFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - Map args = Collections.emptyMap(); - factory.init(args); + factory.init(TEST_VERSION_CURRENT_MAP()); TokenStream stream = factory.create(tokenizer); assertTokenStreamContents(stream, new String[] { "Ceska" }); } Index: lucene/analysis/common/src/test/org/apache/lucene/analysis/standard/TestUAX29URLEmailTokenizerFactory.java =================================================================== --- lucene/analysis/common/src/test/org/apache/lucene/analysis/standard/TestUAX29URLEmailTokenizerFactory.java (revision 1368355) +++ lucene/analysis/common/src/test/org/apache/lucene/analysis/standard/TestUAX29URLEmailTokenizerFactory.java (working copy) @@ -35,9 +35,7 @@ public void testUAX29URLEmailTokenizer() throws Exception { Reader reader = new StringReader("Wha\u0301t's this thing do?"); UAX29URLEmailTokenizerFactory factory = new UAX29URLEmailTokenizerFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - Map args = Collections.emptyMap(); - factory.init(args); + factory.init(TEST_VERSION_CURRENT_MAP()); Tokenizer stream = factory.create(reader); assertTokenStreamContents(stream, new String[] {"Wha\u0301t's", "this", "thing", "do" }); @@ -46,9 +44,7 @@ public void testArabic() throws Exception { Reader reader = new StringReader("الفيلم الوثائقي الأول عن ويكيبيديا يسمى \"الحقيقة بالأرقام: قصة ويكيبيديا\" (بالإنجليزية: Truth in Numbers: The Wikipedia Story)، سيتم إطلاقه في 2008."); UAX29URLEmailTokenizerFactory factory = new UAX29URLEmailTokenizerFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - Map args = Collections.emptyMap(); - factory.init(args); + factory.init(TEST_VERSION_CURRENT_MAP()); Tokenizer stream = factory.create(reader); assertTokenStreamContents(stream, new String[] {"الفيلم", "الوثائقي", "الأول", "عن", "ويكيبيديا", "يسمى", "الحقيقة", "بالأرقام", "قصة", "ويكيبيديا", @@ -58,9 +54,7 @@ public void testChinese() throws Exception { Reader reader = new StringReader("我是中国人。 1234 Tests "); UAX29URLEmailTokenizerFactory factory = new UAX29URLEmailTokenizerFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - Map args = Collections.emptyMap(); - factory.init(args); + factory.init(TEST_VERSION_CURRENT_MAP()); Tokenizer stream = factory.create(reader); assertTokenStreamContents(stream, new String[] {"我", "是", "中", "国", "人", "1234", "Tests"}); @@ -69,9 +63,7 @@ public void testKorean() throws Exception { Reader reader = new StringReader("안녕하세요 한글입니다"); UAX29URLEmailTokenizerFactory factory = new UAX29URLEmailTokenizerFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - Map args = Collections.emptyMap(); - factory.init(args); + factory.init(TEST_VERSION_CURRENT_MAP()); Tokenizer stream = factory.create(reader); assertTokenStreamContents(stream, new String[] {"안녕하세요", "한글입니다"}); @@ -80,9 +72,7 @@ public void testHyphen() throws Exception { Reader reader = new StringReader("some-dashed-phrase"); UAX29URLEmailTokenizerFactory factory = new UAX29URLEmailTokenizerFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - Map args = Collections.emptyMap(); - factory.init(args); + factory.init(TEST_VERSION_CURRENT_MAP()); Tokenizer stream = factory.create(reader); assertTokenStreamContents(stream, new String[] {"some", "dashed", "phrase"}); @@ -106,9 +96,7 @@ + "http://[a42:a7b6::]/qSmxSUU4z/%52qVl4\n"; Reader reader = new StringReader(textWithURLs); UAX29URLEmailTokenizerFactory factory = new UAX29URLEmailTokenizerFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - Map args = Collections.emptyMap(); - factory.init(args); + factory.init(TEST_VERSION_CURRENT_MAP()); Tokenizer stream = factory.create(reader); assertTokenStreamContents(stream, new String[] { @@ -148,9 +136,7 @@ + "lv'p@tqk.vj5s0tgl.0dlu7su3iyiaz.dqso.494.3hb76.XN--MGBAAM7A8H\n"; Reader reader = new StringReader(textWithEmails); UAX29URLEmailTokenizerFactory factory = new UAX29URLEmailTokenizerFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - Map args = Collections.emptyMap(); - factory.init(args); + factory.init(TEST_VERSION_CURRENT_MAP()); Tokenizer stream = factory.create(reader); assertTokenStreamContents(stream, new String[] { @@ -180,10 +166,9 @@ String longWord = builder.toString(); String content = "one two three " + longWord + " four five six"; Reader reader = new StringReader(content); - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); args.put("maxTokenLength", "1000"); UAX29URLEmailTokenizerFactory factory = new UAX29URLEmailTokenizerFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init(args); Tokenizer stream = factory.create(reader); assertTokenStreamContents(stream, Index: lucene/analysis/common/src/test/org/apache/lucene/analysis/miscellaneous/TestCapitalizationFilterFactory.java =================================================================== --- lucene/analysis/common/src/test/org/apache/lucene/analysis/miscellaneous/TestCapitalizationFilterFactory.java (revision 1368355) +++ lucene/analysis/common/src/test/org/apache/lucene/analysis/miscellaneous/TestCapitalizationFilterFactory.java (working copy) @@ -33,12 +33,11 @@ public void testCapitalization() throws Exception { - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); args.put( CapitalizationFilterFactory.KEEP, "and the it BIG" ); args.put( CapitalizationFilterFactory.ONLY_FIRST_WORD, "true" ); CapitalizationFilterFactory factory = new CapitalizationFilterFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init( args ); assertTokenStreamContents(factory.create( new MockTokenizer(new StringReader("kiTTEN"), MockTokenizer.WHITESPACE, false)), @@ -95,7 +94,6 @@ // Now try some prefixes factory = new CapitalizationFilterFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); args.put( "okPrefix", "McK" ); // all words factory.init( args ); assertTokenStreamContents(factory.create( @@ -116,13 +114,12 @@ } public void testKeepIgnoreCase() throws Exception { - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); args.put( CapitalizationFilterFactory.KEEP, "kitten" ); args.put( CapitalizationFilterFactory.KEEP_IGNORE_CASE, "true" ); args.put( CapitalizationFilterFactory.ONLY_FIRST_WORD, "true" ); CapitalizationFilterFactory factory = new CapitalizationFilterFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init( args ); factory.forceFirstLetter = true; assertTokenStreamContents(factory.create( @@ -146,11 +143,10 @@ * This is very weird when combined with ONLY_FIRST_WORD!!! */ public void testMinWordLength() throws Exception { - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); args.put(CapitalizationFilterFactory.ONLY_FIRST_WORD, "true"); args.put(CapitalizationFilterFactory.MIN_WORD_LENGTH, "5"); CapitalizationFilterFactory factory = new CapitalizationFilterFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init(args); Tokenizer tokenizer = new MockTokenizer(new StringReader( "helo testing"), MockTokenizer.WHITESPACE, false); @@ -163,10 +159,9 @@ * in each token (it should do nothing) */ public void testMaxWordCount() throws Exception { - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); args.put(CapitalizationFilterFactory.MAX_WORD_COUNT, "2"); CapitalizationFilterFactory factory = new CapitalizationFilterFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init(args); Tokenizer tokenizer = new MockTokenizer(new StringReader( "one two three four"), MockTokenizer.WHITESPACE, false); @@ -178,10 +173,9 @@ * Test CapitalizationFilterFactory's maxWordCount option when exceeded */ public void testMaxWordCount2() throws Exception { - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); args.put(CapitalizationFilterFactory.MAX_WORD_COUNT, "2"); CapitalizationFilterFactory factory = new CapitalizationFilterFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init(args); Tokenizer tokenizer = new MockTokenizer(new StringReader( "one two three four"), MockTokenizer.KEYWORD, false); @@ -195,10 +189,9 @@ * This is weird, it is not really a max, but inclusive (look at 'is') */ public void testMaxTokenLength() throws Exception { - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); args.put(CapitalizationFilterFactory.MAX_TOKEN_LENGTH, "2"); CapitalizationFilterFactory factory = new CapitalizationFilterFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init(args); Tokenizer tokenizer = new MockTokenizer(new StringReader( "this is a test"), MockTokenizer.WHITESPACE, false); @@ -210,11 +203,10 @@ * Test CapitalizationFilterFactory's forceFirstLetter option */ public void testForceFirstLetter() throws Exception { - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); args.put(CapitalizationFilterFactory.KEEP, "kitten"); args.put(CapitalizationFilterFactory.FORCE_FIRST_LETTER, "true"); CapitalizationFilterFactory factory = new CapitalizationFilterFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init(args); Tokenizer tokenizer = new MockTokenizer(new StringReader("kitten"), MockTokenizer.WHITESPACE, false); TokenStream ts = factory.create(tokenizer); Index: lucene/analysis/common/src/test/org/apache/lucene/analysis/miscellaneous/TestStemmerOverrideFilterFactory.java =================================================================== --- lucene/analysis/common/src/test/org/apache/lucene/analysis/miscellaneous/TestStemmerOverrideFilterFactory.java (revision 1368355) +++ lucene/analysis/common/src/test/org/apache/lucene/analysis/miscellaneous/TestStemmerOverrideFilterFactory.java (working copy) @@ -40,10 +40,9 @@ Reader reader = new StringReader("testing dogs"); Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); StemmerOverrideFilterFactory factory = new StemmerOverrideFilterFactory(); - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); ResourceLoader loader = new StringMockResourceLoader("dogs\tcat"); args.put("dictionary", "stemdict.txt"); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init(args); factory.inform(loader); @@ -55,11 +54,10 @@ Reader reader = new StringReader("testing DoGs"); Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); StemmerOverrideFilterFactory factory = new StemmerOverrideFilterFactory(); - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); ResourceLoader loader = new StringMockResourceLoader("dogs\tcat"); args.put("dictionary", "stemdict.txt"); args.put("ignoreCase", "true"); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init(args); factory.inform(loader); Index: lucene/analysis/common/src/test/org/apache/lucene/analysis/miscellaneous/TestKeepFilterFactory.java =================================================================== --- lucene/analysis/common/src/test/org/apache/lucene/analysis/miscellaneous/TestKeepFilterFactory.java (revision 1368355) +++ lucene/analysis/common/src/test/org/apache/lucene/analysis/miscellaneous/TestKeepFilterFactory.java (working copy) @@ -35,10 +35,9 @@ ResourceLoader loader = new ClasspathResourceLoader(getClass()); assertTrue("loader is null and it shouldn't be", loader != null); KeepWordFilterFactory factory = new KeepWordFilterFactory(); - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); args.put("words", "keep-1.txt"); args.put("ignoreCase", "true"); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init(args); factory.inform(loader); CharArraySet words = factory.getWords(); @@ -48,7 +47,6 @@ factory = new KeepWordFilterFactory(); args.put("words", "keep-1.txt, keep-2.txt"); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init(args); factory.inform(loader); words = factory.getWords(); Index: lucene/analysis/common/src/test/org/apache/lucene/analysis/miscellaneous/TestKeywordMarkerFilterFactory.java =================================================================== --- lucene/analysis/common/src/test/org/apache/lucene/analysis/miscellaneous/TestKeywordMarkerFilterFactory.java (revision 1368355) +++ lucene/analysis/common/src/test/org/apache/lucene/analysis/miscellaneous/TestKeywordMarkerFilterFactory.java (working copy) @@ -39,10 +39,9 @@ Reader reader = new StringReader("dogs cats"); Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); KeywordMarkerFilterFactory factory = new KeywordMarkerFilterFactory(); - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); ResourceLoader loader = new StringMockResourceLoader("cats"); args.put("protected", "protwords.txt"); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init(args); factory.inform(loader); @@ -54,11 +53,10 @@ Reader reader = new StringReader("dogs cats Cats"); Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); KeywordMarkerFilterFactory factory = new KeywordMarkerFilterFactory(); - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); ResourceLoader loader = new StringMockResourceLoader("cats"); args.put("protected", "protwords.txt"); args.put("ignoreCase", "true"); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init(args); factory.inform(loader); Index: lucene/analysis/common/src/test/org/apache/lucene/analysis/reverse/TestReverseStringFilterFactory.java =================================================================== --- lucene/analysis/common/src/test/org/apache/lucene/analysis/reverse/TestReverseStringFilterFactory.java (revision 1368355) +++ lucene/analysis/common/src/test/org/apache/lucene/analysis/reverse/TestReverseStringFilterFactory.java (working copy) @@ -38,9 +38,7 @@ Reader reader = new StringReader("simple test"); Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); ReverseStringFilterFactory factory = new ReverseStringFilterFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - Map args = Collections.emptyMap(); - factory.init(args); + factory.init(TEST_VERSION_CURRENT_MAP()); TokenStream stream = factory.create(tokenizer); assertTokenStreamContents(stream, new String[] { "elpmis", "tset" }); } Index: lucene/analysis/common/src/test/org/apache/lucene/analysis/th/TestThaiWordFilterFactory.java =================================================================== --- lucene/analysis/common/src/test/org/apache/lucene/analysis/th/TestThaiWordFilterFactory.java (revision 1368355) +++ lucene/analysis/common/src/test/org/apache/lucene/analysis/th/TestThaiWordFilterFactory.java (working copy) @@ -40,9 +40,7 @@ Reader reader = new StringReader("การที่ได้ต้องแสดงว่างานดี"); Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); ThaiWordFilterFactory factory = new ThaiWordFilterFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - Map args = Collections.emptyMap(); - factory.init(args); + factory.init(TEST_VERSION_CURRENT_MAP()); TokenStream stream = factory.create(tokenizer); assertTokenStreamContents(stream, new String[] {"การ", "ที่", "ได้", "ต้อง", "แสดง", "ว่า", "งาน", "ดี"}); Index: lucene/analysis/common/src/test/org/apache/lucene/analysis/snowball/TestSnowballPorterFilterFactory.java =================================================================== --- lucene/analysis/common/src/test/org/apache/lucene/analysis/snowball/TestSnowballPorterFilterFactory.java (revision 1368355) +++ lucene/analysis/common/src/test/org/apache/lucene/analysis/snowball/TestSnowballPorterFilterFactory.java (working copy) @@ -45,10 +45,9 @@ } SnowballPorterFilterFactory factory = new SnowballPorterFilterFactory(); - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); args.put("language", "English"); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init(args); factory.inform(new StringMockResourceLoader("")); Tokenizer tokenizer = new MockTokenizer( @@ -74,10 +73,9 @@ public void testProtected() throws Exception { SnowballPorterFilterFactory factory = new SnowballPorterFilterFactory(); ResourceLoader loader = new StringMockResourceLoader("ridding"); - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); args.put("protected", "protwords.txt"); args.put("language", "English"); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init(args); factory.inform(loader); Reader reader = new StringReader("ridding of some stemming"); Index: lucene/analysis/common/src/test/org/apache/lucene/analysis/hunspell/TestHunspellStemFilterFactory.java =================================================================== --- lucene/analysis/common/src/test/org/apache/lucene/analysis/hunspell/TestHunspellStemFilterFactory.java (revision 1368355) +++ lucene/analysis/common/src/test/org/apache/lucene/analysis/hunspell/TestHunspellStemFilterFactory.java (working copy) @@ -33,10 +33,9 @@ public class TestHunspellStemFilterFactory extends BaseTokenStreamTestCase { public void testStemming() throws Exception { HunspellStemFilterFactory factory = new HunspellStemFilterFactory(); - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); args.put("dictionary", "test.dic"); args.put("affix", "test.aff"); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init(args); factory.inform(new ClasspathResourceLoader(getClass())); Index: lucene/analysis/common/src/test/org/apache/lucene/analysis/el/TestGreekLowerCaseFilterFactory.java =================================================================== --- lucene/analysis/common/src/test/org/apache/lucene/analysis/el/TestGreekLowerCaseFilterFactory.java (revision 1368355) +++ lucene/analysis/common/src/test/org/apache/lucene/analysis/el/TestGreekLowerCaseFilterFactory.java (working copy) @@ -38,9 +38,7 @@ Reader reader = new StringReader("Μάϊος ΜΆΪΟΣ"); Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); GreekLowerCaseFilterFactory factory = new GreekLowerCaseFilterFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - Map args = Collections.emptyMap(); - factory.init(args); + factory.init(TEST_VERSION_CURRENT_MAP()); TokenStream stream = factory.create(tokenizer); assertTokenStreamContents(stream, new String[] { "μαιοσ", "μαιοσ" }); } Index: lucene/analysis/common/src/test/org/apache/lucene/analysis/ar/TestArabicFilters.java =================================================================== --- lucene/analysis/common/src/test/org/apache/lucene/analysis/ar/TestArabicFilters.java (revision 1368355) +++ lucene/analysis/common/src/test/org/apache/lucene/analysis/ar/TestArabicFilters.java (working copy) @@ -39,12 +39,9 @@ public void testNormalizer() throws Exception { Reader reader = new StringReader("الذين مَلكت أيمانكم"); StandardTokenizerFactory factory = new StandardTokenizerFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); + factory.init(TEST_VERSION_CURRENT_MAP()); ArabicNormalizationFilterFactory filterFactory = new ArabicNormalizationFilterFactory(); - filterFactory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - Map args = Collections.emptyMap(); - factory.init(args); - filterFactory.init(args); + filterFactory.init(TEST_VERSION_CURRENT_MAP()); Tokenizer tokenizer = factory.create(reader); TokenStream stream = filterFactory.create(tokenizer); assertTokenStreamContents(stream, new String[] {"الذين", "ملكت", "ايمانكم"}); @@ -56,13 +53,10 @@ public void testStemmer() throws Exception { Reader reader = new StringReader("الذين مَلكت أيمانكم"); StandardTokenizerFactory factory = new StandardTokenizerFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); ArabicNormalizationFilterFactory normFactory = new ArabicNormalizationFilterFactory(); - normFactory.setLuceneMatchVersion(TEST_VERSION_CURRENT); ArabicStemFilterFactory stemFactory = new ArabicStemFilterFactory(); - Map args = Collections.emptyMap(); - factory.init(args); - normFactory.init(args); + factory.init(TEST_VERSION_CURRENT_MAP()); + normFactory.init(TEST_VERSION_CURRENT_MAP()); Tokenizer tokenizer = factory.create(reader); TokenStream stream = normFactory.create(tokenizer); stream = stemFactory.create(stream); @@ -76,9 +70,7 @@ Reader reader = new StringReader("می‌خورد"); PersianCharFilterFactory charfilterFactory = new PersianCharFilterFactory(); StandardTokenizerFactory tokenizerFactory = new StandardTokenizerFactory(); - tokenizerFactory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - Map args = Collections.emptyMap(); - tokenizerFactory.init(args); + tokenizerFactory.init(TEST_VERSION_CURRENT_MAP()); TokenStream stream = tokenizerFactory.create(charfilterFactory.create(reader)); assertTokenStreamContents(stream, new String[] { "می", "خورد" }); } Index: lucene/analysis/common/src/test/org/apache/lucene/analysis/cjk/TestCJKBigramFilterFactory.java =================================================================== --- lucene/analysis/common/src/test/org/apache/lucene/analysis/cjk/TestCJKBigramFilterFactory.java (revision 1368355) +++ lucene/analysis/common/src/test/org/apache/lucene/analysis/cjk/TestCJKBigramFilterFactory.java (working copy) @@ -34,9 +34,7 @@ public void testDefaults() throws Exception { Reader reader = new StringReader("多くの学生が試験に落ちた。"); CJKBigramFilterFactory factory = new CJKBigramFilterFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - Map args = Collections.emptyMap(); - factory.init(args); + factory.init(TEST_VERSION_CURRENT_MAP()); TokenStream stream = factory.create(new StandardTokenizer(TEST_VERSION_CURRENT, reader)); assertTokenStreamContents(stream, new String[] { "多く", "くの", "の学", "学生", "生が", "が試", "試験", "験に", "に落", "落ち", "ちた" }); Index: lucene/analysis/common/src/test/org/apache/lucene/analysis/synonym/TestMultiWordSynonyms.java =================================================================== --- lucene/analysis/common/src/test/org/apache/lucene/analysis/synonym/TestMultiWordSynonyms.java (revision 1368355) +++ lucene/analysis/common/src/test/org/apache/lucene/analysis/synonym/TestMultiWordSynonyms.java (working copy) @@ -35,9 +35,8 @@ public void testMultiWordSynonyms() throws IOException { SynonymFilterFactory factory = new SynonymFilterFactory(); - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); args.put("synonyms", "synonyms.txt"); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init(args); factory.inform(new StringMockResourceLoader("a b c,d")); TokenStream ts = factory.create(new MockTokenizer(new StringReader("a e"), MockTokenizer.WHITESPACE, false)); Index: lucene/analysis/common/src/test/org/apache/lucene/analysis/synonym/TestSynonymFilterFactory.java =================================================================== --- lucene/analysis/common/src/test/org/apache/lucene/analysis/synonym/TestSynonymFilterFactory.java (revision 1368355) +++ lucene/analysis/common/src/test/org/apache/lucene/analysis/synonym/TestSynonymFilterFactory.java (working copy) @@ -32,9 +32,8 @@ /** test that we can parse and use the solr syn file */ public void testSynonyms() throws Exception { SynonymFilterFactory factory = new SynonymFilterFactory(); - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); args.put("synonyms", "synonyms.txt"); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init(args); factory.inform(new ClasspathResourceLoader(getClass())); TokenStream ts = factory.create(new MockTokenizer(new StringReader("GB"), MockTokenizer.WHITESPACE, false)); @@ -47,9 +46,8 @@ /** if the synonyms are completely empty, test that we still analyze correctly */ public void testEmptySynonyms() throws Exception { SynonymFilterFactory factory = new SynonymFilterFactory(); - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); args.put("synonyms", "synonyms.txt"); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init(args); factory.inform(new StringMockResourceLoader("")); // empty file! TokenStream ts = factory.create(new MockTokenizer(new StringReader("GB"), MockTokenizer.WHITESPACE, false)); Index: lucene/analysis/common/src/test/org/apache/lucene/analysis/util/TestElisionFilterFactory.java =================================================================== --- lucene/analysis/common/src/test/org/apache/lucene/analysis/util/TestElisionFilterFactory.java (revision 1368355) +++ lucene/analysis/common/src/test/org/apache/lucene/analysis/util/TestElisionFilterFactory.java (working copy) @@ -41,9 +41,8 @@ Reader reader = new StringReader("l'avion"); Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); ElisionFilterFactory factory = new ElisionFilterFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); ResourceLoader loader = new ClasspathResourceLoader(getClass()); - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); args.put("articles", "frenchArticles.txt"); factory.init(args); factory.inform(loader); @@ -58,9 +57,7 @@ Reader reader = new StringReader("l'avion"); Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); ElisionFilterFactory factory = new ElisionFilterFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - Map args = Collections.emptyMap(); - factory.init(args); + factory.init(TEST_VERSION_CURRENT_MAP()); ResourceLoader loader = new ClasspathResourceLoader(getClass()); factory.inform(loader); TokenStream stream = factory.create(tokenizer); @@ -74,9 +71,8 @@ Reader reader = new StringReader("L'avion"); Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); ElisionFilterFactory factory = new ElisionFilterFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); ResourceLoader loader = new ClasspathResourceLoader(getClass()); - Map args = new HashMap(); + Map args = TEST_VERSION_CURRENT_MAP(); args.put("articles", "frenchArticles.txt"); args.put("ignoreCase", "true"); factory.init(args); Index: lucene/analysis/common/src/java/org/apache/lucene/analysis/core/LowerCaseTokenizerFactory.java =================================================================== --- lucene/analysis/common/src/java/org/apache/lucene/analysis/core/LowerCaseTokenizerFactory.java (revision 1368355) +++ lucene/analysis/common/src/java/org/apache/lucene/analysis/core/LowerCaseTokenizerFactory.java (working copy) @@ -49,7 +49,6 @@ @Override public AbstractAnalysisFactory getMultiTermComponent() { LowerCaseFilterFactory filt = new LowerCaseFilterFactory(); - filt.setLuceneMatchVersion(luceneMatchVersion); filt.init(args); return filt; } Index: lucene/analysis/common/src/java/org/apache/lucene/analysis/synonym/SynonymFilterFactory.java =================================================================== --- lucene/analysis/common/src/java/org/apache/lucene/analysis/synonym/SynonymFilterFactory.java (revision 1368355) +++ lucene/analysis/common/src/java/org/apache/lucene/analysis/synonym/SynonymFilterFactory.java (working copy) @@ -155,7 +155,6 @@ // (there are no tests for this functionality) private TokenizerFactory loadTokenizerFactory(ResourceLoader loader, String cname) throws IOException { TokenizerFactory tokFactory = loader.newInstance(cname, TokenizerFactory.class); - tokFactory.setLuceneMatchVersion(luceneMatchVersion); tokFactory.init(args); if (tokFactory instanceof ResourceLoaderAware) { ((ResourceLoaderAware) tokFactory).inform(loader); Index: lucene/analysis/common/src/java/org/apache/lucene/analysis/util/AbstractAnalysisFactory.java =================================================================== --- lucene/analysis/common/src/java/org/apache/lucene/analysis/util/AbstractAnalysisFactory.java (revision 1368355) +++ lucene/analysis/common/src/java/org/apache/lucene/analysis/util/AbstractAnalysisFactory.java (working copy) @@ -48,6 +48,8 @@ public void init(Map args) { this.args = args; + String luceneMatchVersionArg = args.get("luceneMatchVersion"); + this.luceneMatchVersion = luceneMatchVersionArg == null ? null : Version.parseLeniently(luceneMatchVersionArg); } public Map getArgs() { @@ -64,10 +66,6 @@ } } - public void setLuceneMatchVersion(Version luceneMatchVersion) { - this.luceneMatchVersion = luceneMatchVersion; - } - public Version getLuceneMatchVersion() { return this.luceneMatchVersion; } Index: lucene/analysis/morfologik/src/test/org/apache/lucene/analysis/morfologik/TestMorfologikFilterFactory.java =================================================================== --- lucene/analysis/morfologik/src/test/org/apache/lucene/analysis/morfologik/TestMorfologikFilterFactory.java (revision 1368355) +++ lucene/analysis/morfologik/src/test/org/apache/lucene/analysis/morfologik/TestMorfologikFilterFactory.java (working copy) @@ -31,11 +31,10 @@ public class TestMorfologikFilterFactory extends BaseTokenStreamTestCase { public void testCreateDictionary() throws Exception { StringReader reader = new StringReader("rowery bilety"); - Map initParams = new HashMap(); + Map initParams = TEST_VERSION_CURRENT_MAP(); initParams.put(MorfologikFilterFactory.DICTIONARY_SCHEMA_ATTRIBUTE, "morfologik"); MorfologikFilterFactory factory = new MorfologikFilterFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); factory.init(initParams); TokenStream ts = factory.create(new WhitespaceTokenizer(TEST_VERSION_CURRENT, reader)); Index: lucene/analysis/phonetic/src/test/org/apache/lucene/analysis/phonetic/TestBeiderMorseFilterFactory.java =================================================================== --- lucene/analysis/phonetic/src/test/org/apache/lucene/analysis/phonetic/TestBeiderMorseFilterFactory.java (revision 1368355) +++ lucene/analysis/phonetic/src/test/org/apache/lucene/analysis/phonetic/TestBeiderMorseFilterFactory.java (working copy) @@ -30,9 +30,7 @@ public class TestBeiderMorseFilterFactory extends BaseTokenStreamTestCase { public void testBasics() throws Exception { BeiderMorseFilterFactory factory = new BeiderMorseFilterFactory(); - factory.setLuceneMatchVersion(TEST_VERSION_CURRENT); - Map args = Collections.emptyMap(); - factory.init(args); + factory.init(TEST_VERSION_CURRENT_MAP()); TokenStream ts = factory.create(new MockTokenizer(new StringReader("Weinberg"), MockTokenizer.WHITESPACE, false)); assertTokenStreamContents(ts, new String[] { "vDnbirk", "vanbirk", "vinbirk", "wDnbirk", "wanbirk", "winbirk" }, Index: lucene/test-framework/src/java/org/apache/lucene/analysis/BaseTokenStreamTestCase.java =================================================================== --- lucene/test-framework/src/java/org/apache/lucene/analysis/BaseTokenStreamTestCase.java (revision 1368355) +++ lucene/test-framework/src/java/org/apache/lucene/analysis/BaseTokenStreamTestCase.java (working copy) @@ -906,4 +906,10 @@ } return ret; } + + public static Map TEST_VERSION_CURRENT_MAP() { + Map map = new HashMap(); + map.put("luceneMatchVersion", TEST_VERSION_CURRENT.toString()); + return map; + } }