Uses of Class
org.apache.lucene.analysis.CharArraySet

Packages that use CharArraySet
Package
Description
Text analysis.
Analyzer for Arabic.
Analyzer for Bulgarian.
Analyzer for Bengali Language.
Analyzer for Brazilian Portuguese.
Analyzer for Catalan.
Normalization of text before the tokenizer.
Analyzer for Chinese, Japanese, and Korean, which indexes bigrams.
Analyzer for Sorani Kurdish.
Fast, general-purpose grammar-based tokenizers.
Analyzer for Simplified Chinese, which indexes words.
Construct n-grams for frequently occurring terms and phrases.
A filter that decomposes compound words you find in many Germanic languages into the word parts.
Basic, general-purpose analysis components.
Analyzer for Czech.
Analyzer for Danish.
Analyzer for German.
Analyzer for Greek.
Fast, general-purpose URLs and email addresses tokenizers.
Analyzer for English.
Analyzer for Spanish.
Analyzer for Estonian.
Analyzer for Basque.
Analyzer for Persian.
Analyzer for Finnish.
Analyzer for French.
Analyzer for Irish.
Analyzer for Galician.
Analyzer for Hindi.
Analyzer for Hungarian.
Analyzer for Armenian.
Analyzer for Indonesian.
Analyzer for Italian.
Analyzer for Japanese.
Analyzer for Lithuanian.
Analyzer for Latvian.
Miscellaneous Tokenstreams.
Analyzer for Nepali.
Analyzer for Dutch.
Analyzer for Norwegian.
Analyzer for Polish.
Analyzer for Portuguese.
Analyzer for Romanian.
Analyzer for Russian.
TokenFilter and Analyzer implementations that use a modified version of Snowball stemmers.
Analyzer for Serbian.
Fast, general-purpose grammar-based tokenizer StandardTokenizer implements the Word Break rules from the Unicode Text Segmentation algorithm, as specified in Unicode Standard Annex #29.
Analyzer for Swedish.
Analyzer for Tamil.
Analyzer for Telugu Language.
Analyzer for Thai.
Analyzer for Turkish.
Utility functions for text analysis.
Monitoring framework
Analyzer based autosuggest.
Support for document suggestion