Index: src/java/org/apache/lucene/analysis/CharTokenizer.java =================================================================== --- src/java/org/apache/lucene/analysis/CharTokenizer.java (revision 606689) +++ src/java/org/apache/lucene/analysis/CharTokenizer.java (working copy) @@ -83,6 +83,7 @@ token.termLength = length; token.startOffset = start; token.endOffset = start+length; + token.positionIncrement = 1; return token; } Index: src/java/org/apache/lucene/analysis/KeywordTokenizer.java =================================================================== --- src/java/org/apache/lucene/analysis/KeywordTokenizer.java (revision 606689) +++ src/java/org/apache/lucene/analysis/KeywordTokenizer.java (working copy) @@ -51,6 +51,7 @@ buffer = result.resizeTermBuffer(1+buffer.length); } result.termLength = upto; + result.positionIncrement = 1; return result; } return null; Index: src/java/org/apache/lucene/analysis/standard/StandardTokenizer.java =================================================================== --- src/java/org/apache/lucene/analysis/standard/StandardTokenizer.java (revision 606689) +++ src/java/org/apache/lucene/analysis/standard/StandardTokenizer.java (working copy) @@ -73,6 +73,7 @@ result.setStartOffset(start); result.setEndOffset(start+result.termLength()); result.setType(StandardTokenizerImpl.TOKEN_TYPES[tokenType]); + result.setPositionIncrement(1); return result; } Index: src/java/org/apache/lucene/analysis/Tokenizer.java =================================================================== --- src/java/org/apache/lucene/analysis/Tokenizer.java (revision 606865) +++ src/java/org/apache/lucene/analysis/Tokenizer.java (working copy) @@ -23,8 +23,12 @@ /** A Tokenizer is a TokenStream whose input is a Reader.

This is an abstract class. +

NOTE: subclasses must override at least one of {@link #next()} or {@link #next(Token)}. +

+ NOTE: subclasses overiding {@link #next(Token)} must usually + reset to 1 the position increment of the reused {@link Token}. */ public abstract class Tokenizer extends TokenStream {