You're viewing version 2.13 of the OpenSearch documentation. This version is no longer maintained. For the latest version, see the current documentation. For information about OpenSearch version maintenance, see Release Schedule and Maintenance Policy.
Text analysis
When you are searching documents using a full-text search, you want to receive all relevant results. If you’re looking for “walk”, you’re interested in results that contain any form of the word, like “Walk”, “walked”, or “walking”. To facilitate full-text search, OpenSearch uses text analysis.
The objective of text analysis is to split the unstructured free text content of the source document into a sequence of terms, which are then stored in an inverted index. Subsequently, when a similar text analysis is applied to a user’s query, the resulting sequence of terms facilitates the matching of relevant source documents.
From a technical point of view, the text analysis process consists of several steps, some of which are optional:
-
Before the free text content can be split into individual words, it may be beneficial to refine the text at the character level. The primary aim of this optional step is to help the tokenizer (the subsequent stage in the analysis process) generate better tokens. This can include removal of markup tags (such as HTML) or handling specific character patterns (like replacing the 🙂 emoji with the text
:slightly_smiling_face:
). -
The next step is to split the free text into individual words—tokens. This is performed by a tokenizer. For example, after tokenization, the sentence
Actions speak louder than words
is split into tokensActions
,speak
,louder
,than
, andwords
. -
The last step is to process individual tokens by applying a series of token filters. The aim is to convert each token into a predictable form that is directly stored in the index, for example, by converting them to lowercase or performing stemming (reducing the word to its root). For example, the token
Actions
becomesaction
,louder
becomesloud
, andwords
becomesword
.
Although the terms token and term may sound similar and are occasionally used interchangeably, it is helpful to understand the difference between the two. In the context of Apache Lucene, each holds a distinct role. A token is created by a tokenizer during text analysis and often undergoes a number of additional modifications as it passes through the chain of token filters. Each token is associated with metadata that can be further used during the text analysis process. A term is a data value that is directly stored in the inverted index and is associated with much less metadata. During search, matching operates at the term level.
Analyzers
In OpenSearch, the abstraction that encompasses text analysis is referred to as an analyzer. Each analyzer contains the following sequentially applied components:
-
Character filters: First, a character filter receives the original text as a stream of characters and adds, removes, or modifies characters in the text. For example, a character filter can strip HTML characters from a string so that the text
<p><b>Actions</b> speak louder than <em>words</em></p>
becomes\nActions speak louder than words\n
. The output of a character filter is a stream of characters. -
Tokenizer: Next, a tokenizer receives the stream of characters that has been processed by the character filter and splits the text into individual tokens (usually, words). For example, a tokenizer can split text on white space so that the preceding text becomes [
Actions
,speak
,louder
,than
,words
]. Tokenizers also maintain metadata about tokens, such as their starting and ending positions in the text. The output of a tokenizer is a stream of tokens. -
Token filters: Last, a token filter receives the stream of tokens from the tokenizer and adds, removes, or modifies tokens. For example, a token filter may lowercase the tokens so that
Actions
becomesaction
, remove stopwords likethan
, or add synonyms liketalk
for the wordspeak
.
An analyzer must contain exactly one tokenizer and may contain zero or more character filters and zero or more token filters.
There is also a special type of analyzer called a normalizer. A normalizer is similar to an analyzer except that it does not contain a tokenizer and can only include specific types of character filters and token filters. These filters can perform only character-level operations, such as character or pattern replacement, and cannot perform operations on the token as a whole. This means that replacing a token with a synonym or stemming is not supported. See Normalizers for further details.
Built-in analyzers
The following table lists the built-in analyzers that OpenSearch provides. The last column of the table contains the result of applying the analyzer to the string It’s fun to contribute a brand-new PR or 2 to OpenSearch!
.
Analyzer | Analysis performed | Analyzer output |
---|---|---|
Standard (default) | - Parses strings into tokens at word boundaries - Removes most punctuation - Converts tokens to lowercase | [it’s , fun , to , contribute , a ,brand , new , pr , or , 2 , to , opensearch ] |
Simple | - Parses strings into tokens on any non-letter character - Removes non-letter characters - Converts tokens to lowercase | [it , s , fun , to , contribute , a ,brand , new , pr , or , to , opensearch ] |
Whitespace | - Parses strings into tokens on white space | [It’s , fun , to , contribute , a ,brand-new , PR , or , 2 , to , OpenSearch! ] |
Stop | - Parses strings into tokens on any non-letter character - Removes non-letter characters - Removes stop words - Converts tokens to lowercase | [s , fun , contribute , brand , new , pr , opensearch ] |
Keyword (no-op) | - Outputs the entire string unchanged | [It’s fun to contribute a brand-new PR or 2 to OpenSearch! ] |
Pattern | - Parses strings into tokens using regular expressions - Supports converting strings to lowercase - Supports removing stop words | [it , s , fun , to , contribute , a ,brand , new , pr , or , 2 , to , opensearch ] |
Language | Performs analysis specific to a certain language (for example, english ). | [fun , contribut , brand , new , pr , 2 , opensearch ] |
Fingerprint | - Parses strings on any non-letter character - Normalizes characters by converting them to ASCII - Converts tokens to lowercase - Sorts, deduplicates, and concatenates tokens into a single token - Supports removing stop words | [2 a brand contribute fun it's new opensearch or pr to ] Note that the apostrophe was converted to its ASCII counterpart. |
Custom analyzers
If needed, you can combine tokenizers, token filters, and character filters to create a custom analyzer.
Text analysis at indexing time and query time
OpenSearch performs text analysis on text fields when you index a document and when you send a search request. Depending on the time of text analysis, the analyzers used for it are classified as follows:
-
An index analyzer performs analysis at indexing time: When you are indexing a text field, OpenSearch analyzes it before indexing it. For more information about ways to specify index analyzers, see Index analyzers.
-
A search analyzer performs analysis at query time: OpenSearch analyzes the query string when you run a full-text query on a text field. For more information about ways to specify search analyzers, see Search analyzers.
In most cases, you should use the same analyzer at both indexing and search time because the text field and the query string will be analyzed in the same way and the resulting tokens will match as expected.
Example
When you index a document that has a text field with the text Actions speak louder than words
, OpenSearch analyzes the text and produces the following list of tokens:
Text field tokens = [action
, speak
, loud
, than
, word
]
When you search for documents that match the query speaking loudly
, OpenSearch analyzes the query string and produces the following list of tokens:
Query string tokens = [speak
, loud
]
Then OpenSearch compares each token in the query string against the list of text field tokens and finds that both lists contain the tokens speak
and loud
, so OpenSearch returns this document as part of the search results that match the query.
Testing an analyzer
To test a built-in analyzer and view the list of tokens it generates when a document is indexed, you can use the Analyze API.
Specify the analyzer and the text to be analyzed in the request:
GET /_analyze
{
"analyzer" : "standard",
"text" : "Let’s contribute to OpenSearch!"
}
The following image shows the query string.
The response contains each token and its start and end offsets that correspond to the starting index in the original string (inclusive) and the ending index (exclusive):
{
"tokens": [
{
"token": "let’s",
"start_offset": 0,
"end_offset": 5,
"type": "<ALPHANUM>",
"position": 0
},
{
"token": "contribute",
"start_offset": 6,
"end_offset": 16,
"type": "<ALPHANUM>",
"position": 1
},
{
"token": "to",
"start_offset": 17,
"end_offset": 19,
"type": "<ALPHANUM>",
"position": 2
},
{
"token": "opensearch",
"start_offset": 20,
"end_offset": 30,
"type": "<ALPHANUM>",
"position": 3
}
]
}
Verifying analyzer settings
To verify which analyzer is associated with which field, you can use the get mapping API operation:
GET /testindex/_mapping
The response provides information about the analyzers for each field:
{
"testindex": {
"mappings": {
"properties": {
"text_entry": {
"type": "text",
"analyzer": "simple",
"search_analyzer": "whitespace"
}
}
}
}
}
Next steps
- Learn more about specifying index analyzers and search analyzers.