| org.apache.lucene.analysis.Tokenizer org.apache.lucene.analysis.ngram.NGramTokenizer
NGramTokenizer | public class NGramTokenizer extends Tokenizer (Code) | | Tokenizes the input into n-grams of the given size(s).
author: Otis Gospodnetic |
Constructor Summary | |
public | NGramTokenizer(Reader input, int minGram, int maxGram) Creates NGramTokenizer with given min and max n-grams. | public | NGramTokenizer(Reader input) Creates NGramTokenizer with default min and max n-grams. |
Method Summary | |
final public Token | next() Returns the next token in the stream, or null at EOS. |
DEFAULT_MAX_NGRAM_SIZE | final public static int DEFAULT_MAX_NGRAM_SIZE(Code) | | |
DEFAULT_MIN_NGRAM_SIZE | final public static int DEFAULT_MIN_NGRAM_SIZE(Code) | | |
NGramTokenizer | public NGramTokenizer(Reader input, int minGram, int maxGram)(Code) | | Creates NGramTokenizer with given min and max n-grams.
Parameters: input - Reader holding the input to be tokenized Parameters: minGram - the smallest n-gram to generate Parameters: maxGram - the largest n-gram to generate |
NGramTokenizer | public NGramTokenizer(Reader input)(Code) | | Creates NGramTokenizer with default min and max n-grams.
Parameters: input - Reader holding the input to be tokenized |
|
|