Lucene 2.9.4 contrib-analyzers API

Packages 
Package Description
org.apache.lucene.analysis.ar
Analyzer for Arabic.
org.apache.lucene.analysis.br
Analyzer for Brazilian Portuguese.
org.apache.lucene.analysis.cjk
Analyzer for Chinese, Japanese, and Korean, which indexes bigrams (overlapping groups of two adjacent Han characters).
org.apache.lucene.analysis.cn
Analyzer for Chinese, which indexes unigrams (individual chinese characters).
org.apache.lucene.analysis.compound
A filter that decomposes compound words you find in many Germanic languages into the word parts.
org.apache.lucene.analysis.compound.hyphenation
The code for the compound word hyphenation is taken from the Apache FOP project.
org.apache.lucene.analysis.cz
Analyzer for Czech.
org.apache.lucene.analysis.de
Analyzer for German.
org.apache.lucene.analysis.el
Analyzer for Greek.
org.apache.lucene.analysis.fa
Analyzer for Persian.
org.apache.lucene.analysis.fr
Analyzer for French.
org.apache.lucene.analysis.miscellaneous
Miscellaneous TokenStreams
org.apache.lucene.analysis.ngram
Character n-gram tokenizers and filters.
org.apache.lucene.analysis.nl
Analyzer for Dutch.
org.apache.lucene.analysis.payloads
Provides various convenience classes for creating payloads on Tokens.
org.apache.lucene.analysis.position
Filter for assigning position increments.
org.apache.lucene.analysis.query
Automatically filter high-frequency stopwords.
org.apache.lucene.analysis.reverse
Filter to reverse token text.
org.apache.lucene.analysis.ru
Analyzer for Russian.
org.apache.lucene.analysis.shingle
Word n-gram filters
org.apache.lucene.analysis.sinks
Implementations of the SinkTokenizer that might be useful.
org.apache.lucene.analysis.th
Analyzer for Thai.

Copyright © 2000-2014 Apache Software Foundation. All Rights Reserved.