|
||||||||||
PREV NEXT | FRAMES NO FRAMES |
Packages that use TokenFilter | |
---|---|
org.apache.lucene.analysis | API and code to convert text into indexable/searchable tokens. |
org.apache.lucene.analysis.br | Analyzer for Brazilian. |
org.apache.lucene.analysis.cn | Analyzer for Chinese. |
org.apache.lucene.analysis.de | Analyzer for German. |
org.apache.lucene.analysis.el | Analyzer for Greek. |
org.apache.lucene.analysis.fr | Analyzer for French. |
org.apache.lucene.analysis.ngram | |
org.apache.lucene.analysis.nl | Analyzer for Dutch. |
org.apache.lucene.analysis.payloads | Provides various convenience classes for creating payloads on Tokens. |
org.apache.lucene.analysis.ru | Analyzer for Russian. |
org.apache.lucene.analysis.snowball | TokenFilter and Analyzer implementations that use Snowball
stemmers. |
org.apache.lucene.analysis.standard | A fast grammar-based tokenizer constructed with JFlex. |
org.apache.lucene.analysis.th | |
org.apache.lucene.index.memory | High-performance single-document main memory Apache Lucene fulltext search index. |
Uses of TokenFilter in org.apache.lucene.analysis |
---|
Subclasses of TokenFilter in org.apache.lucene.analysis | |
---|---|
class |
CachingTokenFilter
This class can be used if the Tokens of a TokenStream are intended to be consumed more than once. |
class |
ISOLatin1AccentFilter
A filter that replaces accented characters in the ISO Latin 1 character set (ISO-8859-1) by their unaccented equivalent. |
class |
LengthFilter
Removes words that are too long and too short from the stream. |
class |
LowerCaseFilter
Normalizes token text to lower case. |
class |
PorterStemFilter
Transforms the token stream as per the Porter stemming algorithm. |
class |
StopFilter
Removes stop words from a token stream. |
class |
TeeTokenFilter
Works in conjunction with the SinkTokenizer to provide the ability to set aside tokens that have already been analyzed. |
Uses of TokenFilter in org.apache.lucene.analysis.br |
---|
Subclasses of TokenFilter in org.apache.lucene.analysis.br | |
---|---|
class |
BrazilianStemFilter
Based on GermanStemFilter |
Uses of TokenFilter in org.apache.lucene.analysis.cn |
---|
Subclasses of TokenFilter in org.apache.lucene.analysis.cn | |
---|---|
class |
ChineseFilter
Title: ChineseFilter Description: Filter with a stop word table Rule: No digital is allowed. |
Uses of TokenFilter in org.apache.lucene.analysis.de |
---|
Subclasses of TokenFilter in org.apache.lucene.analysis.de | |
---|---|
class |
GermanStemFilter
A filter that stems German words. |
Uses of TokenFilter in org.apache.lucene.analysis.el |
---|
Subclasses of TokenFilter in org.apache.lucene.analysis.el | |
---|---|
class |
GreekLowerCaseFilter
Normalizes token text to lower case, analyzing given ("greek") charset. |
Uses of TokenFilter in org.apache.lucene.analysis.fr |
---|
Subclasses of TokenFilter in org.apache.lucene.analysis.fr | |
---|---|
class |
ElisionFilter
Removes elisions from a token stream. |
class |
FrenchStemFilter
A filter that stemms french words. |
Uses of TokenFilter in org.apache.lucene.analysis.ngram |
---|
Subclasses of TokenFilter in org.apache.lucene.analysis.ngram | |
---|---|
class |
EdgeNGramTokenFilter
Tokenizes the given token into n-grams of given size(s). |
class |
NGramTokenFilter
Tokenizes the input into n-grams of the given size(s). |
Uses of TokenFilter in org.apache.lucene.analysis.nl |
---|
Subclasses of TokenFilter in org.apache.lucene.analysis.nl | |
---|---|
class |
DutchStemFilter
A filter that stems Dutch words. |
Uses of TokenFilter in org.apache.lucene.analysis.payloads |
---|
Subclasses of TokenFilter in org.apache.lucene.analysis.payloads | |
---|---|
class |
NumericPayloadTokenFilter
Assigns a payload to a token based on the Token.type() |
class |
TokenOffsetPayloadTokenFilter
Adds the Token.setStartOffset(int)
and Token.setEndOffset(int)
First 4 bytes are the start |
class |
TypeAsPayloadTokenFilter
Makes the Token.type() a payload. |
Uses of TokenFilter in org.apache.lucene.analysis.ru |
---|
Subclasses of TokenFilter in org.apache.lucene.analysis.ru | |
---|---|
class |
RussianLowerCaseFilter
Normalizes token text to lower case, analyzing given ("russian") charset. |
class |
RussianStemFilter
A filter that stems Russian words. |
Uses of TokenFilter in org.apache.lucene.analysis.snowball |
---|
Subclasses of TokenFilter in org.apache.lucene.analysis.snowball | |
---|---|
class |
SnowballFilter
A filter that stems words using a Snowball-generated stemmer. |
Uses of TokenFilter in org.apache.lucene.analysis.standard |
---|
Subclasses of TokenFilter in org.apache.lucene.analysis.standard | |
---|---|
class |
StandardFilter
Normalizes tokens extracted with StandardTokenizer . |
Uses of TokenFilter in org.apache.lucene.analysis.th |
---|
Subclasses of TokenFilter in org.apache.lucene.analysis.th | |
---|---|
class |
ThaiWordFilter
TokenFilter that use java.text.BreakIterator to break each Token that is Thai into separate Token(s) for each Thai word. |
Uses of TokenFilter in org.apache.lucene.index.memory |
---|
Subclasses of TokenFilter in org.apache.lucene.index.memory | |
---|---|
class |
SynonymTokenFilter
Injects additional tokens for synonyms of token terms fetched from the underlying child stream; the child stream must deliver lowercase tokens for synonyms to be found. |
|
||||||||||
PREV NEXT | FRAMES NO FRAMES |