Percolator field type

The percolator field type parses a json structure into a native query and stores that query, so that the percolate query can use it to match provided documents.

Any field that contains a json object can be configured to be a percolator field. The percolator field type has no settings. Just configuring the percolator field type is sufficient to instruct Elasticsearch to treat a field as a query.

If the following mapping configures the percolator field type for the query field:

  1. PUT my-index-000001
  2. {
  3. "mappings": {
  4. "properties": {
  5. "query": {
  6. "type": "percolator"
  7. },
  8. "field": {
  9. "type": "text"
  10. }
  11. }
  12. }
  13. }

Then you can index a query:

  1. PUT my-index-000001/_doc/match_value
  2. {
  3. "query": {
  4. "match": {
  5. "field": "value"
  6. }
  7. }
  8. }

Fields referred to in a percolator query must already exist in the mapping associated with the index used for percolation. In order to make sure these fields exist, add or update a mapping via the create index or put mapping APIs.

Reindexing your percolator queries

Reindexing percolator queries is sometimes required to benefit from improvements made to the percolator field type in new releases.

Reindexing percolator queries can be reindexed by using the reindex api. Lets take a look at the following index with a percolator field type:

  1. PUT index
  2. {
  3. "mappings": {
  4. "properties": {
  5. "query" : {
  6. "type" : "percolator"
  7. },
  8. "body" : {
  9. "type": "text"
  10. }
  11. }
  12. }
  13. }
  14. POST _aliases
  15. {
  16. "actions": [
  17. {
  18. "add": {
  19. "index": "index",
  20. "alias": "queries"
  21. }
  22. }
  23. ]
  24. }
  25. PUT queries/_doc/1?refresh
  26. {
  27. "query" : {
  28. "match" : {
  29. "body" : "quick brown fox"
  30. }
  31. }
  32. }

It is always recommended to define an alias for your index, so that in case of a reindex systems / applications don’t need to be changed to know that the percolator queries are now in a different index.

Lets say you’re going to upgrade to a new major version and in order for the new Elasticsearch version to still be able to read your queries you need to reindex your queries into a new index on the current Elasticsearch version:

  1. PUT new_index
  2. {
  3. "mappings": {
  4. "properties": {
  5. "query" : {
  6. "type" : "percolator"
  7. },
  8. "body" : {
  9. "type": "text"
  10. }
  11. }
  12. }
  13. }
  14. POST /_reindex?refresh
  15. {
  16. "source": {
  17. "index": "index"
  18. },
  19. "dest": {
  20. "index": "new_index"
  21. }
  22. }
  23. POST _aliases
  24. {
  25. "actions": [
  26. {
  27. "remove": {
  28. "index" : "index",
  29. "alias": "queries"
  30. }
  31. },
  32. {
  33. "add": {
  34. "index": "new_index",
  35. "alias": "queries"
  36. }
  37. }
  38. ]
  39. }

If you have an alias don’t forget to point it to the new index.

Executing the percolate query via the queries alias:

  1. GET /queries/_search
  2. {
  3. "query": {
  4. "percolate" : {
  5. "field" : "query",
  6. "document" : {
  7. "body" : "fox jumps over the lazy dog"
  8. }
  9. }
  10. }
  11. }

now returns matches from the new index:

  1. {
  2. "took": 3,
  3. "timed_out": false,
  4. "_shards": {
  5. "total": 1,
  6. "successful": 1,
  7. "skipped" : 0,
  8. "failed": 0
  9. },
  10. "hits": {
  11. "total" : {
  12. "value": 1,
  13. "relation": "eq"
  14. },
  15. "max_score": 0.13076457,
  16. "hits": [
  17. {
  18. "_index": "new_index",
  19. "_type": "_doc",
  20. "_id": "1",
  21. "_score": 0.13076457,
  22. "_source": {
  23. "query": {
  24. "match": {
  25. "body": "quick brown fox"
  26. }
  27. }
  28. },
  29. "fields" : {
  30. "_percolator_document_slot" : [0]
  31. }
  32. }
  33. ]
  34. }
  35. }

Percolator query hit is now being presented from the new index.

Optimizing query time text analysis

When the percolator verifies a percolator candidate match it is going to parse, perform query time text analysis and actually run the percolator query on the document being percolated. This is done for each candidate match and every time the percolate query executes. If your query time text analysis is relatively expensive part of query parsing then text analysis can become the dominating factor time is being spent on when percolating. This query parsing overhead can become noticeable when the percolator ends up verifying many candidate percolator query matches.

To avoid the most expensive part of text analysis at percolate time. One can choose to do the expensive part of text analysis when indexing the percolator query. This requires using two different analyzers. The first analyzer actually performs text analysis that needs be performed (expensive part). The second analyzer (usually whitespace) just splits the generated tokens that the first analyzer has produced. Then before indexing a percolator query, the analyze api should be used to analyze the query text with the more expensive analyzer. The result of the analyze api, the tokens, should be used to substitute the original query text in the percolator query. It is important that the query should now be configured to override the analyzer from the mapping and just the second analyzer. Most text based queries support an analyzer option (match, query_string, simple_query_string). Using this approach the expensive text analysis is performed once instead of many times.

Lets demonstrate this workflow via a simplified example.

Lets say we want to index the following percolator query:

  1. {
  2. "query" : {
  3. "match" : {
  4. "body" : {
  5. "query" : "missing bicycles"
  6. }
  7. }
  8. }
  9. }

with these settings and mapping:

  1. PUT /test_index
  2. {
  3. "settings": {
  4. "analysis": {
  5. "analyzer": {
  6. "my_analyzer" : {
  7. "tokenizer": "standard",
  8. "filter" : ["lowercase", "porter_stem"]
  9. }
  10. }
  11. }
  12. },
  13. "mappings": {
  14. "properties": {
  15. "query" : {
  16. "type": "percolator"
  17. },
  18. "body" : {
  19. "type": "text",
  20. "analyzer": "my_analyzer"
  21. }
  22. }
  23. }
  24. }

For the purpose of this example, this analyzer is considered expensive.

First we need to use the analyze api to perform the text analysis prior to indexing:

  1. POST /test_index/_analyze
  2. {
  3. "analyzer" : "my_analyzer",
  4. "text" : "missing bicycles"
  5. }

This results the following response:

  1. {
  2. "tokens": [
  3. {
  4. "token": "miss",
  5. "start_offset": 0,
  6. "end_offset": 7,
  7. "type": "<ALPHANUM>",
  8. "position": 0
  9. },
  10. {
  11. "token": "bicycl",
  12. "start_offset": 8,
  13. "end_offset": 16,
  14. "type": "<ALPHANUM>",
  15. "position": 1
  16. }
  17. ]
  18. }

All the tokens in the returned order need to replace the query text in the percolator query:

  1. PUT /test_index/_doc/1?refresh
  2. {
  3. "query" : {
  4. "match" : {
  5. "body" : {
  6. "query" : "miss bicycl",
  7. "analyzer" : "whitespace"
  8. }
  9. }
  10. }
  11. }

It is important to select a whitespace analyzer here, otherwise the analyzer defined in the mapping will be used, which defeats the point of using this workflow. Note that whitespace is a built-in analyzer, if a different analyzer needs to be used, it needs to be configured first in the index’s settings.

The analyze api prior to the indexing the percolator flow should be done for each percolator query.

At percolate time nothing changes and the percolate query can be defined normally:

  1. GET /test_index/_search
  2. {
  3. "query": {
  4. "percolate" : {
  5. "field" : "query",
  6. "document" : {
  7. "body" : "Bycicles are missing"
  8. }
  9. }
  10. }
  11. }

This results in a response like this:

  1. {
  2. "took": 6,
  3. "timed_out": false,
  4. "_shards": {
  5. "total": 1,
  6. "successful": 1,
  7. "skipped" : 0,
  8. "failed": 0
  9. },
  10. "hits": {
  11. "total" : {
  12. "value": 1,
  13. "relation": "eq"
  14. },
  15. "max_score": 0.13076457,
  16. "hits": [
  17. {
  18. "_index": "test_index",
  19. "_type": "_doc",
  20. "_id": "1",
  21. "_score": 0.13076457,
  22. "_source": {
  23. "query": {
  24. "match": {
  25. "body": {
  26. "query": "miss bicycl",
  27. "analyzer": "whitespace"
  28. }
  29. }
  30. }
  31. },
  32. "fields" : {
  33. "_percolator_document_slot" : [0]
  34. }
  35. }
  36. ]
  37. }
  38. }

Optimizing wildcard queries.

Wildcard queries are more expensive than other queries for the percolator, especially if the wildcard expressions are large.

In the case of wildcard queries with prefix wildcard expressions or just the prefix query, the edge_ngram token filter can be used to replace these queries with regular term query on a field where the edge_ngram token filter is configured.

Creating an index with custom analysis settings:

  1. PUT my_queries1
  2. {
  3. "settings": {
  4. "analysis": {
  5. "analyzer": {
  6. "wildcard_prefix": {
  7. "type": "custom",
  8. "tokenizer": "standard",
  9. "filter": [
  10. "lowercase",
  11. "wildcard_edge_ngram"
  12. ]
  13. }
  14. },
  15. "filter": {
  16. "wildcard_edge_ngram": {
  17. "type": "edge_ngram",
  18. "min_gram": 1,
  19. "max_gram": 32
  20. }
  21. }
  22. }
  23. },
  24. "mappings": {
  25. "properties": {
  26. "query": {
  27. "type": "percolator"
  28. },
  29. "my_field": {
  30. "type": "text",
  31. "fields": {
  32. "prefix": {
  33. "type": "text",
  34. "analyzer": "wildcard_prefix",
  35. "search_analyzer": "standard"
  36. }
  37. }
  38. }
  39. }
  40. }
  41. }

The analyzer that generates the prefix tokens to be used at index time only.

Increase the min_gram and decrease max_gram settings based on your prefix search needs.

This multifield should be used to do the prefix search with a term or match query instead of a prefix or wildcard query.

Then instead of indexing the following query:

  1. {
  2. "query": {
  3. "wildcard": {
  4. "my_field": "abc*"
  5. }
  6. }
  7. }

this query below should be indexed:

  1. PUT /my_queries1/_doc/1?refresh
  2. {
  3. "query": {
  4. "term": {
  5. "my_field.prefix": "abc"
  6. }
  7. }
  8. }

This way can handle the second query more efficiently than the first query.

The following search request will match with the previously indexed percolator query:

  1. GET /my_queries1/_search
  2. {
  3. "query": {
  4. "percolate": {
  5. "field": "query",
  6. "document": {
  7. "my_field": "abcd"
  8. }
  9. }
  10. }
  11. }
  1. {
  2. "took": 6,
  3. "timed_out": false,
  4. "_shards": {
  5. "total": 1,
  6. "successful": 1,
  7. "skipped": 0,
  8. "failed": 0
  9. },
  10. "hits": {
  11. "total" : {
  12. "value": 1,
  13. "relation": "eq"
  14. },
  15. "max_score": 0.18864399,
  16. "hits": [
  17. {
  18. "_index": "my_queries1",
  19. "_type": "_doc",
  20. "_id": "1",
  21. "_score": 0.18864399,
  22. "_source": {
  23. "query": {
  24. "term": {
  25. "my_field.prefix": "abc"
  26. }
  27. }
  28. },
  29. "fields": {
  30. "_percolator_document_slot": [
  31. 0
  32. ]
  33. }
  34. }
  35. ]
  36. }
  37. }

The same technique can also be used to speed up suffix wildcard searches. By using the reverse token filter before the edge_ngram token filter.

  1. PUT my_queries2
  2. {
  3. "settings": {
  4. "analysis": {
  5. "analyzer": {
  6. "wildcard_suffix": {
  7. "type": "custom",
  8. "tokenizer": "standard",
  9. "filter": [
  10. "lowercase",
  11. "reverse",
  12. "wildcard_edge_ngram"
  13. ]
  14. },
  15. "wildcard_suffix_search_time": {
  16. "type": "custom",
  17. "tokenizer": "standard",
  18. "filter": [
  19. "lowercase",
  20. "reverse"
  21. ]
  22. }
  23. },
  24. "filter": {
  25. "wildcard_edge_ngram": {
  26. "type": "edge_ngram",
  27. "min_gram": 1,
  28. "max_gram": 32
  29. }
  30. }
  31. }
  32. },
  33. "mappings": {
  34. "properties": {
  35. "query": {
  36. "type": "percolator"
  37. },
  38. "my_field": {
  39. "type": "text",
  40. "fields": {
  41. "suffix": {
  42. "type": "text",
  43. "analyzer": "wildcard_suffix",
  44. "search_analyzer": "wildcard_suffix_search_time"
  45. }
  46. }
  47. }
  48. }
  49. }
  50. }

A custom analyzer is needed at search time too, because otherwise the query terms are not being reversed and would otherwise not match with the reserved suffix tokens.

Then instead of indexing the following query:

  1. {
  2. "query": {
  3. "wildcard": {
  4. "my_field": "*xyz"
  5. }
  6. }
  7. }

the following query below should be indexed:

  1. PUT /my_queries2/_doc/2?refresh
  2. {
  3. "query": {
  4. "match": {
  5. "my_field.suffix": "xyz"
  6. }
  7. }
  8. }

The match query should be used instead of the term query, because text analysis needs to reverse the query terms.

The following search request will match with the previously indexed percolator query:

  1. GET /my_queries2/_search
  2. {
  3. "query": {
  4. "percolate": {
  5. "field": "query",
  6. "document": {
  7. "my_field": "wxyz"
  8. }
  9. }
  10. }
  11. }

Dedicated Percolator Index

Percolate queries can be added to any index. Instead of adding percolate queries to the index the data resides in, these queries can also be added to a dedicated index. The advantage of this is that this dedicated percolator index can have its own index settings (For example the number of primary and replica shards). If you choose to have a dedicated percolate index, you need to make sure that the mappings from the normal index are also available on the percolate index. Otherwise percolate queries can be parsed incorrectly.

Forcing Unmapped Fields to be Handled as Strings

In certain cases it is unknown what kind of percolator queries do get registered, and if no field mapping exists for fields that are referred by percolator queries then adding a percolator query fails. This means the mapping needs to be updated to have the field with the appropriate settings, and then the percolator query can be added. But sometimes it is sufficient if all unmapped fields are handled as if these were default text fields. In those cases one can configure the index.percolator.map_unmapped_fields_as_text setting to true (default to false) and then if a field referred in a percolator query does not exist, it will be handled as a default text field so that adding the percolator query doesn’t fail.

Limitations

Parent/child

Because the percolate query is processing one document at a time, it doesn’t support queries and filters that run against child documents such as has_child and has_parent.

Fetching queries

There are a number of queries that fetch data via a get call during query parsing. For example the terms query when using terms lookup, template query when using indexed scripts and geo_shape when using pre-indexed shapes. When these queries are indexed by the percolator field type then the get call is executed once. So each time the percolator query evaluates these queries, the fetches terms, shapes etc. as the were upon index time will be used. Important to note is that fetching of terms that these queries do, happens both each time the percolator query gets indexed on both primary and replica shards, so the terms that are actually indexed can be different between shard copies, if the source index changed while indexing.

Script query

The script inside a script query can only access doc values fields. The percolate query indexes the provided document into an in-memory index. This in-memory index doesn’t support stored fields and because of that the _source field and other stored fields are not stored. This is the reason why in the script query the _source and other stored fields aren’t available.

Field aliases

Percolator queries that contain field aliases may not always behave as expected. In particular, if a percolator query is registered that contains a field alias, and then that alias is updated in the mappings to refer to a different field, the stored query will still refer to the original target field. To pick up the change to the field alias, the percolator query must be explicitly reindexed.