Uploaded image for project: 'Lucene - Core'
  1. Lucene - Core
  2. LUCENE-2407

make CharTokenizer.MAX_WORD_LEN parametrizable

Details

    • Improvement
    • Status: Open
    • Minor
    • Resolution: Unresolved
    • 3.0.1
    • 4.9, 6.0
    • modules/analysis
    • New

    Description

      as discussed here http://n3.nabble.com/are-long-words-split-into-up-to-256-long-tokens-tp739914p739914.html it would be nice to be able to parametrize that value.

      Attachments

        Activity

          People

            Unassigned Unassigned
            jmwap jmwap
            Votes:
            0 Vote for this issue
            Watchers:
            0 Start watching this issue

            Dates

              Created:
              Updated: