- Which search engines does this cover?
- The analyzer simulates Elasticsearch/OpenSearch analyzer chains including character filters, tokenizers, and token filters. Output is equivalent to the Analyze API.
- What's the difference between a tokenizer and an analyzer?
- A tokenizer splits text into tokens. An analyzer is the full pipeline: character filters (pre-process) → tokenizer → token filters (lowercase, stem, stopwords). An analyzer always contains exactly one tokenizer.
- Does it support language-specific analyzers?
- Yes — English, French, German, Spanish, and other built-in language analyzers are included, with their respective stemmer and stopword configurations pre-loaded.