This overview of the logging deployment process is provided for ease of reference. It is not a substitute for full documentation. For new installations, Vector and LokiStack are recommended.
|
As of logging version 5.5, you have the option of choosing from Fluentd or Vector collector implementations, and Elasticsearch or LokiStack as log stores. Documentation for logging is in the process of being updated to reflect these underlying component changes.
|
Prerequisites
-
LogStore preference: Elasticsearch or LokiStack
-
Collector implementation preference: Fluentd or Vector
-
Credentials for your log forwarding outputs
Procedure
|
As of logging version 5.4.3 the Elasticsearch Operator is deprecated and is planned to be removed in a future release. Red Hat will provide bug fixes and support for this feature during the current release lifecycle, but this feature will no longer receive enhancements and will be removed. As an alternative to using the Elasticsearch Operator to manage the default log storage, you can use the Loki Operator.
|
-
Install the Operator for the logstore you’d like to use.
-
For Elasticsearch, install the OpenShift Elasticsearch Operator.
-
For LokiStack, install the Loki Operator.
-
Install the Red Hat OpenShift logging Operator.
-
Create a Clusterlogging
custom resource (CR) instance.
-
Select your Collector Implementation.
|
As of logging version 5.6 Fluentd is deprecated and is planned to be removed in a future release. Red Hat will provide bug fixes and support for this feature during the current release lifecycle, but this feature will no longer receive enhancements and will be removed. As an alternative to Fluentd, you can use Vector instead.
|
-
Create a ClusterLogForwarder
custom resource (CR) instance.
-
Create a secret for the selected output pipeline.