Grok processor

The grok processor is used to parse and structure unstructured data using pattern matching. You can use the grok processor to extract fields from log messages, web server access logs, application logs, and other log data that follows a consistent format.

Grok basics

The grok processor uses a set of predefined patterns to match parts of the input text. Each pattern consists of a name and a regular expression. For example, the pattern %{IP:ip_address} matches an IP address and assigns it to the field ip_address. You can combine multiple patterns to create more complex expressions. For example, the pattern %{IP:client} %{WORD:method} %{URIPATHPARM:request} %{NUMBER:bytes %NUMBER:duration} matches a line from a web server access log and extracts the client IP address, the HTTP method, the request URI, the number of bytes sent, and the duration of the request.

The grok processor is built on the Oniguruma regular expression library and supports all the patterns from that library. You can use the Grok Debugger tool to test and debug your grok expressions.

Syntax

The following is the basic syntax for the grok processor:

  1. {
  2. "grok": {
  3. "field": "your_message",
  4. "patterns": ["your_patterns"]
  5. }
  6. }

copy

Configuration parameters

To configure the grok processor, you have various options that allow you to define patterns, match specific keys, and control the processor’s behavior. The following table lists the required and optional parameters for the grok processor.

ParameterRequired/OptionalDescription
fieldRequiredThe name of the field containing the text to be parsed.
patternsRequiredA list of grok expressions used to match and extract named captures. The first matching expression in the list is returned.
pattern_definitionsOptionalA dictionary of pattern names and pattern tuples used to define custom patterns for the current processor. If a pattern matches an existing name, it overrides the pre-existing definition.
trace_matchOptionalWhen the parameter is set to true, the processor adds a field named _grok_match_index to the processed document. This field contains the index of the pattern within the patterns array that successfully matched the document. This information can be useful for debugging and understanding which pattern was applied to the document. Default is false.
descriptionOptionalA brief description of the processor.
ifOptionalA condition for running the processor.
ignore_failureOptionalSpecifies whether the processor continues execution even if it encounters errors. If set to true, failures are ignored. Default is false.
ignore_missingOptionalSpecifies whether the processor should ignore documents that do not contain the specified field. If set to true, the processor does not modify the document if the field does not exist or is null. Default is false.
on_failureOptionalA list of processors to run if the processor fails.
tagOptionalAn identifier tag for the processor. Useful for debugging in order to distinguish between processors of the same type.

Creating a pipeline

The following steps guide you through creating an ingest pipeline with the grok processor.

Step 1: Create a pipeline

The following query creates a pipeline, named log_line. It extracts fields from the message field of the document using the specified pattern. In this case, it extracts the clientip, timestamp, and response_status fields:

  1. PUT _ingest/pipeline/log_line
  2. {
  3. "description": "Extract fields from a log line",
  4. "processors": [
  5. {
  6. "grok": {
  7. "field": "message",
  8. "patterns": ["%{IPORHOST:clientip} %{HTTPDATE:timestamp} %{NUMBER:response_status:int}"]
  9. }
  10. }
  11. ]
  12. }

copy

Step 2 (Optional): Test the pipeline.

alert icon NOTE
It is recommended that you test your pipeline before you ingest documents.

To test the pipeline, run the following query:

  1. POST _ingest/pipeline/log_line/_simulate
  2. {
  3. "docs": [
  4. {
  5. "_source": {
  6. "message": "127.0.0.1 198.126.12 10/Oct/2000:13:55:36 -0700 200"
  7. }
  8. }
  9. ]
  10. }

copy

Response

The following response confirms that the pipeline is working as expected:

  1. {
  2. "docs": [
  3. {
  4. "doc": {
  5. "_index": "_index",
  6. "_id": "_id",
  7. "_source": {
  8. "message": "127.0.0.1 198.126.12 10/Oct/2000:13:55:36 -0700 200",
  9. "response_status": 200,
  10. "clientip": "198.126.12",
  11. "timestamp": "10/Oct/2000:13:55:36 -0700"
  12. },
  13. "_ingest": {
  14. "timestamp": "2023-09-13T21:41:52.064540505Z"
  15. }
  16. }
  17. }
  18. ]
  19. }

Step 3: Ingest a document

The following query ingests a document into an index named testindex1:

  1. PUT testindex1/_doc/1?pipeline=log_line
  2. {
  3. "message": "127.0.0.1 198.126.12 10/Oct/2000:13:55:36 -0700 200"
  4. }

copy

Step 4 (Optional): Retrieve the document

To retrieve the document, run the following query:

  1. GET testindex1/_doc/1

copy

Custom patterns

You can use default patterns, or you can add custom patterns to your pipelines using the patterns_definitions parameter. Custom grok patterns can be used in a pipeline to extract structured data from log messages that do not match the built-in grok patterns. This can be useful for parsing log messages from custom applications or for parsing log messages that have been modified in some way. Custom patterns adhere to a straightforward structure: each pattern has a unique name and the corresponding regular expression that defines its matching behavior.

The following is an example of how to include a custom pattern in your configuration. In this example, the issue number is between 3 and 4 digits and is parsed into the issue_number field and the status is parsed into the status field:

  1. PUT _ingest/pipeline/log_line
  2. {
  3. "processors": [
  4. {
  5. "grok": {
  6. "field": "message",
  7. "patterns": ["The issue number %{NUMBER:issue_number} is %{STATUS:status}"],
  8. "pattern_definitions" : {
  9. "NUMBER" : "\\d{3,4}",
  10. "STATUS" : "open|closed"
  11. }
  12. }
  13. }
  14. ]
  15. }

copy

Tracing which patterns matched

To trace which patterns matched and populated the fields, you can use the trace_match parameter. The following is an example of how to include this parameter in your configuration:

  1. PUT _ingest/pipeline/log_line
  2. {
  3. "description": "Extract fields from a log line",
  4. "processors": [
  5. {
  6. "grok": {
  7. "field": "message",
  8. "patterns": ["%{HTTPDATE:timestamp} %{IPORHOST:clientip}", "%{IPORHOST:clientip} %{HTTPDATE:timestamp} %{NUMBER:response_status:int}"],
  9. "trace_match": true
  10. }
  11. }
  12. ]
  13. }

copy

When you simulate the pipeline, OpenSearch returns the _ingest metadata that includes the grok_match_index, as shown in the following output:

  1. {
  2. "docs": [
  3. {
  4. "doc": {
  5. "_index": "_index",
  6. "_id": "_id",
  7. "_source": {
  8. "message": "127.0.0.1 198.126.12 10/Oct/2000:13:55:36 -0700 200",
  9. "response_status": 200,
  10. "clientip": "198.126.12",
  11. "timestamp": "10/Oct/2000:13:55:36 -0700"
  12. },
  13. "_ingest": {
  14. "_grok_match_index": "1",
  15. "timestamp": "2023-11-02T18:48:40.455619084Z"
  16. }
  17. }
  18. }
  19. ]
  20. }