Getting started with logging 5.7

This overview of the logging deployment process is provided for ease of reference. It is not a substitute for full documentation. For new installations, Vector and LokiStack are recommended.

As of logging version 5.5, you have the option of choosing from Fluentd or Vector collector implementations, and Elasticsearch or LokiStack as log stores. Documentation for logging is in the process of being updated to reflect these underlying component changes.

Logging is provided as an installable component, with a distinct release cycle from the core OKD. The Red Hat OpenShift Container Platform Life Cycle Policy outlines release compatibility.

Prerequisites

  • Log store preference: Elasticsearch or LokiStack

  • Collector implementation preference: Fluentd or Vector

  • Credentials for your log forwarding outputs

Procedure

As of logging version 5.4.3 the OpenShift Elasticsearch Operator is deprecated and is planned to be removed in a future release. Red Hat will provide bug fixes and support for this feature during the current release lifecycle, but this feature will no longer receive enhancements and will be removed. As an alternative to using the OpenShift Elasticsearch Operator to manage the default log storage, you can use the Loki Operator.

  1. Install the Operator for the log store you’d like to use.

    • For Elasticsearch, install the OpenShift Elasticsearch Operator.

    • For LokiStack, install the Loki Operator.

      If you have installed the Loki Operator, create a LokiStack custom resource (CR) instance as well.

  2. Install the Red Hat OpenShift Logging Operator.

  3. Create a ClusterLogging CR instance.

    • Select your collector implementation.