The logging subsystem for Red Hat OpenShift is provided as an installable component, with a distinct release cycle from the core OKD. The Red Hat OpenShift Container Platform Life Cycle Policy outlines release compatibility.
This release includes OpenShift Logging Bug Fix Release 5.4.9.
Before this update, the Fluentd collector would warn of unused configuration parameters. This update removes those configuration parameters and their warning messages. (LOG-3074)
Before this update, Kibana had a fixed 24h
OAuth cookie expiration time, which resulted in 401 errors in Kibana whenever the accessTokenInactivityTimeout
field was set to a value lower than 24h
. With this update, Kibana’s OAuth cookie expiration time synchronizes to the accessTokenInactivityTimeout
, with a default value of 24h
. (LOG-3306)
This release includes RHSA-2022:7435-OpenShift Logging Bug Fix Release 5.4.8.
This release includes OpenShift Logging Bug Fix Release 5.4.6.
Before this update, Fluentd would sometimes not recognize that the Kubernetes platform rotated the log file and would no longer read log messages. This update corrects that by setting the configuration parameter suggested by the upstream development team. (LOG-2792)
Before this update, each rollover job created empty indices when the ClusterLogForwarder
custom resource had JSON parsing defined. With this update, new indices are not empty. (LOG-2823)
Before this update, if you deleted the Kibana Custom Resource, the OKD web console continued displaying a link to Kibana. With this update, removing the Kibana Custom Resource also removes that link. (LOG-3054)
This release includes RHSA-2022:6183-OpenShift Logging Bug Fix Release 5.4.5.
Before this update, the Operator did not ensure that the pod was ready, which caused the cluster to reach an inoperable state during a cluster restart. With this update, the Operator marks new pods as ready before continuing to a new pod during a restart, which resolves the issue. (LOG-2881)
Before this update, the addition of multi-line error detection caused internal routing to change and forward records to the wrong destination. With this update, the internal routing is correct. (LOG-2946)
Before this update, the Operator could not decode index setting JSON responses with a quoted Boolean value and would result in an error. With this update, the Operator can properly decode this JSON response. (LOG-3009)
Before this update, Elasticsearch index templates defined the fields for labels with the wrong types. This change updates those templates to match the expected types forwarded by the log collector. (LOG-2972)
This release includes RHBA-2022:5907-OpenShift Logging Bug Fix Release 5.4.4.
Before this update, non-latin characters displayed incorrectly in Elasticsearch. With this update, Elasticsearch displays all valid UTF-8 symbols correctly. (LOG-2794)
Before this update, non-latin characters displayed incorrectly in Fluentd. With this update, Fluentd displays all valid UTF-8 symbols correctly. (LOG-2657)
Before this update, the metrics server for the collector attempted to bind to the address using a value exposed by an environment value. This change modifies the configuration to bind to any available interface. (LOG-2821)
Before this update, the cluster-logging
Operator relied on the cluster to create a secret. This cluster behavior changed in OKD 4.11, which caused logging deployments to fail. With this update, the cluster-logging
Operator resolves the issue by creating the secret if needed. (LOG-2840)
This release includes RHBA-2022:5556-OpenShift Logging Bug Fix Release 5.4.3.
In logging subsystem 5.4.3 the Elasticsearch Operator is deprecated and is planned to be removed in a future release. Red Hat will provide bug fixes and support for this feature during the current release lifecycle, but this feature will no longer receive enhancements and will be removed. As an alternative to using the Elasticsearch Operator to manage the default log storage, you can use the Loki Operator.
Before this update, the OpenShift Logging Dashboard showed the number of active primary shards instead of all active shards. With this update, the dashboard displays all active shards. (LOG-2781)
Before this update, a bug in a library used by elasticsearch-operator
contained a denial of service attack vulnerability. With this update, the library has been updated to a version that does not contain this vulnerability. (LOG-2816)
Before this update, when configuring Vector to forward logs to Loki, it was not possible to set a custom bearer token or use the default token if Loki had TLS enabled. With this update, Vector can forward logs to Loki using tokens with TLS enabled. (LOG-2786
Before this update, the ElasticSearch Operator omitted the referencePolicy
property of the ImageStream
custom resource when selecting an oauth-proxy
image. This omission caused the Kibana deployment to fail in specific environments. With this update, using referencePolicy
resolves the issue, and the Operator can deploy Kibana successfully. (LOG-2791)
Before this update, alerting rules for the ClusterLogForwarder
custom resource did not take multiple forward outputs into account. This update resolves the issue. (LOG-2640)
Before this update, clusters configured to forward logs to Amazon CloudWatch wrote rejected log files to temporary storage, causing cluster instability over time. With this update, chunk backup for CloudWatch has been disabled, resolving the issue. (LOG-2768)
This release includes RHBA-2022:4874-OpenShift Logging Bug Fix Release 5.4.2
Before this update, editing the Collector configuration using oc edit
was difficult because it had inconsistent use of white-space. This change introduces logic to normalize and format the configuration prior to any updates by the Operator so that it is easy to edit using oc edit
. (LOG-2319)
Before this update, the FluentdNodeDown
alert could not provide instance labels in the message section appropriately. This update resolves the issue by fixing the alert rule to provide instance labels in cases of partial instance failures. (LOG-2607)
Before this update, several log levels, such as`critical`, that were documented as supported by the product were not. This update fixes the discrepancy so the documented log levels are now supported by the product. (LOG-2033)
This release includes RHSA-2022:2216-OpenShift Logging Bug Fix Release 5.4.1.
Before this update, the log file metric exporter only reported logs created while the exporter was running, which resulted in inaccurate log growth data. This update resolves this issue by monitoring /var/log/pods
. (LOG-2442)
Before this update, the collector would be blocked because it continually tried to use a stale connection when forwarding logs to fluentd forward receivers. With this release, the keepalive_timeout
value has been set to 30 seconds (30s
) so that the collector recycles the connection and re-attempts to send failed messages within a reasonable amount of time. (LOG-2534)
Before this update, an error in the gateway component enforcing tenancy for reading logs limited access to logs with a Kubernetes namespace causing "audit" and some "infrastructure" logs to be unreadable. With this update, the proxy correctly detects users with admin access and allows access to logs without a namespace. (LOG-2448)
Before this update, the system:serviceaccount:openshift-monitoring:prometheus-k8s
service account had cluster level privileges as a clusterrole
and clusterrolebinding
. This update restricts the service account` to the openshift-logging
namespace with a role and rolebinding. (LOG-2437)
Before this update, Linux audit log time parsing relied on an ordinal position of a key/value pair. This update changes the parsing to use a regular expression to find the time entry. (LOG-2321)
The following advisories are available for logging 5.4: Logging subsystem for Red Hat OpenShift Release 5.4
Vector is a Technology Preview feature only. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process. For more information about the support scope of Red Hat Technology Preview features, see https://access.redhat.com/support/offerings/techpreview/. |
Vector is a log collector offered as a tech-preview alternative to the current default collector for the logging subsystem.
The following outputs are supported:
elasticsearch
. An external Elasticsearch instance. The elasticsearch
output can use a TLS connection.
kafka
. A Kafka broker. The kafka
output can use an unsecured or TLS connection.
loki
. Loki, a horizontally scalable, highly available, multi-tenant log aggregation system.
Vector is not enabled by default. Use the following steps to enable Vector on your OKD cluster.
Vector does not support FIPS Enabled Clusters. |
OKD: 4.10
Logging subsystem for Red Hat OpenShift: 5.4
FIPS disabled
Edit the ClusterLogging
custom resource (CR) in the openshift-logging
project:
$ oc -n openshift-logging edit ClusterLogging instance
Add a logging.openshift.io/preview-vector-collector: enabled
annotation to the ClusterLogging
custom resource (CR).
Add vector
as a collection type to the ClusterLogging
custom resource (CR).
apiVersion: "logging.openshift.io/v1"
kind: "ClusterLogging"
metadata:
name: "instance"
namespace: "openshift-logging"
annotations:
logging.openshift.io/preview-vector-collector: enabled
spec:
collection:
logs:
type: "vector"
vector: {}
Loki Operator is a Technology Preview feature only. Technology Preview features are not supported with Red Hat production service level agreements (SLAs) and might not be functionally complete. Red Hat does not recommend using them in production. These features provide early access to upcoming product features, enabling customers to test functionality and provide feedback during the development process. For more information about the support scope of Red Hat Technology Preview features, see https://access.redhat.com/support/offerings/techpreview/. |
Loki is a horizontally scalable, highly available, multi-tenant log aggregation system currently offered as an alternative to Elasticsearch as a log store for the logging subsystem.
You can use the OKD web console to install the LokiOperator.
OKD: 4.10
Logging subsystem for Red Hat OpenShift: 5.4
To install the LokiOperator using the OKD web console:
Install the LokiOperator:
In the OKD web console, click Operators → OperatorHub.
Choose LokiOperator from the list of available Operators, and click Install.
Under Installation Mode, select All namespaces on the cluster.
Under Installed Namespace, select openshift-operators-redhat.
You must specify the openshift-operators-redhat
namespace. The openshift-operators
namespace might contain Community Operators, which are untrusted and could publish
a metric with the same name as an OKD metric, which would cause
conflicts.
Select Enable operator recommended cluster monitoring on this namespace.
This option sets the openshift.io/cluster-monitoring: "true"
label in the Namespace object.
You must select this option to ensure that cluster monitoring
scrapes the openshift-operators-redhat
namespace.
Select an Approval Strategy.
The Automatic strategy allows Operator Lifecycle Manager (OLM) to automatically update the Operator when a new version is available.
The Manual strategy requires a user with appropriate credentials to approve the Operator update.
Click Install.
Verify that you installed the LokiOperator. Visit the Operators → Installed Operators page and look for "LokiOperator."
Ensure that LokiOperator is listed in all the projects whose Status is Succeeded.
Before this update, the cluster-logging-operator
used cluster scoped roles and bindings to establish permissions for the Prometheus service account to scrape metrics. These permissions were created when deploying the Operator using the console interface but were missing when deploying from the command line. This update fixes the issue by making the roles and bindings namespace-scoped. (LOG-2286)
Before this update, a prior change to fix dashboard reconciliation introduced a ownerReferences
field to the resource across namespaces. As a result, both the config map and dashboard were not created in the namespace. With this update, the removal of the ownerReferences
field resolves the issue, and the OpenShift Logging dashboard is available in the console. (LOG-2163)
Before this update, changes to the metrics dashboards did not deploy because the cluster-logging-operator
did not correctly compare existing and modified config maps that contain the dashboard. With this update, the addition of a unique hash value to object labels resolves the issue. (LOG-2071)
Before this update, the OpenShift Logging dashboard did not correctly display the pods and namespaces in the table, which displays the top producing containers collected over the last 24 hours. With this update, the pods and namespaces are displayed correctly. (LOG-2069)
Before this update, when the ClusterLogForwarder
was set up with Elasticsearch OutputDefault
and Elasticsearch outputs did not have structured keys, the generated configuration contained the incorrect values for authentication. This update corrects the secret and certificates used. (LOG-2056)
Before this update, the OpenShift Logging dashboard displayed an empty CPU graph because of a reference to an invalid metric. With this update, the correct data point has been selected, resolving the issue. (LOG-2026)
Before this update, the Fluentd container image included builder tools that were unnecessary at run time. This update removes those tools from the image.(LOG-1927)
Before this update, a name change of the deployed collector in the 5.3 release caused the logging collector to generate the FluentdNodeDown
alert. This update resolves the issue by fixing the job name for the Prometheus alert. (LOG-1918)
Before this update, the log collector was collecting its own logs due to a refactoring of the component name change. This lead to a potential feedback loop of the collector processing its own log that might result in memory and log message size issues. This update resolves the issue by excluding the collector logs from the collection. (LOG-1774)
Before this update, Elasticsearch generated the error Unable to create PersistentVolumeClaim due to forbidden: exceeded quota: infra-storage-quota.
if the PVC already existed. With this update, Elasticsearch checks for existing PVCs, resolving the issue. (LOG-2131)
Before this update, Elasticsearch was unable to return to the ready state when the elasticsearch-signing
secret was removed. With this update, Elasticsearch is able to go back to the ready state after that secret is removed. (LOG-2171)
Before this update, the change of the path from which the collector reads container logs caused the collector to forward some records to the wrong indices. With this update, the collector now uses the correct configuration to resolve the issue. (LOG-2160)
Before this update, clusters with a large number of namespaces caused Elasticsearch to stop serving requests because the list of namespaces reached the maximum header size limit. With this update, headers only include a list of namespace names, resolving the issue. (LOG-1899)
Before this update, the OKD Logging dashboard showed the number of shards 'x' times larger than the actual value when Elasticsearch had 'x' nodes. This issue occurred because it was printing all primary shards for each Elasticsearch pod and calculating a sum on it, although the output was always for the whole Elasticsearch cluster. With this update, the number of shards is now correctly calculated. (LOG-2156)
Before this update, the secrets kibana
and kibana-proxy
were not recreated if they were deleted manually. With this update, the elasticsearch-operator
will watch the resources and automatically recreate them if deleted. (LOG-2250)
Before this update, tuning the buffer chunk size could cause the collector to generate a warning about the chunk size exceeding the byte limit for the event stream. With this update, you can also tune the read line limit, resolving the issue. (LOG-2379)
Before this update, the logging console link in OpenShift web console was not removed with the ClusterLogging CR. With this update, deleting the CR or uninstalling the Cluster Logging Operator removes the link. (LOG-2373)
Before this update, a change to the container logs path caused the collection metric to always be zero with older releases configured with the original path. With this update, the plugin which exposes metrics about collected logs supports reading from either path to resolve the issue. (LOG-2462)
This release includes RHSA-2022:68828-OpenShift Logging Bug Fix Release 5.3.13.
Before this update, the log file size map generated by the log-file-metrics-exporter
component did not remove entries for deleted files, resulting in increased file size, and process memory. With this update, the log file size map does not contain entries for deleted files. (LOG-3293)
This release includes OpenShift Logging Bug Fix Release 5.3.12.
This release includes OpenShift Logging Bug Fix Release 5.3.11.
Before this update, the Operator did not ensure that the pod was ready, which caused the cluster to reach an inoperable state during a cluster restart. With this update, the Operator marks new pods as ready before continuing to a new pod during a restart, which resolves the issue. (LOG-2871)
This release includes RHSA-2022:5908-OpenShift Logging Bug Fix Release 5.3.10.
This release includes RHBA-2022:5557-OpenShift Logging Bug Fix Release 5.3.9.
Before this update, the logging collector included a path as a label for the metrics it produced. This path changed frequently and contributed to significant storage changes for the Prometheus server. With this update, the label has been dropped to resolve the issue and reduce storage consumption. (LOG-2682)
This release includes RHBA-2022:5010-OpenShift Logging Bug Fix Release 5.3.8
This release includes RHSA-2022:2217 OpenShift Logging Bug Fix Release 5.3.7
Before this update, Linux audit log time parsing relied on an ordinal position of key/value pair. This update changes the parsing to utilize a regex to find the time entry. (LOG-2322)
Before this update, some log forwarder outputs could re-order logs with the same time-stamp. With this update, a sequence number has been added to the log record to order entries that have matching timestamps. (LOG-2334)
Before this update, clusters with a large number of namespaces caused Elasticsearch to stop serving requests because the list of namespaces reached the maximum header size limit. With this update, headers only include a list of namespace names, resolving the issue. (LOG-2450)
Before this update, system:serviceaccount:openshift-monitoring:prometheus-k8s
had cluster level privileges as a clusterrole
and clusterrolebinding
. This update restricts the serviceaccount
to the openshift-logging
namespace with a role and rolebinding. (LOG-2481))
This release includes RHBA-2022:1377 OpenShift Logging Bug Fix Release 5.3.6
Before this update, defining a toleration with no key and the existing Operator caused the Operator to be unable to complete an upgrade. With this update, this toleration no longer blocks the upgrade from completing. (LOG-2126)
Before this change, it was possible for the collector to generate a warning where the chunk byte limit was exceeding an emitted event. With this change, you can tune the readline limit to resolve the issue as advised by the upstream documentation. (LOG-2380)
This release includes RHSA-2022:0721 OpenShift Logging Bug Fix Release 5.3.5
Before this update, if you removed OpenShift Logging from OKD, the web console continued displaying a link to the Logging page. With this update, removing or uninstalling OpenShift Logging also removes that link. (LOG-2182)
This release includes RHBA-2022:0411 OpenShift Logging Bug Fix Release 5.3.4
Before this update, changes to the metrics dashboards had not yet been deployed because the cluster-logging-operator
did not correctly compare existing and desired config maps that contained the dashboard. This update fixes the logic by adding a unique hash value to the object labels. (LOG-2066)
Before this update, Elasticsearch pods failed to start after updating with FIPS enabled. With this update, Elasticsearch pods start successfully. (LOG-1974)
Before this update, elasticsearch generated the error "Unable to create PersistentVolumeClaim due to forbidden: exceeded quota: infra-storage-quota." if the PVC already existed. With this update, elasticsearch checks for existing PVCs, resolving the issue. (LOG-2127)
This release includes RHSA-2022:0227 OpenShift Logging Bug Fix Release 5.3.3
Before this update, changes to the metrics dashboards had not yet been deployed because the cluster-logging-operator did not correctly compare existing and desired configmaps containing the dashboard. This update fixes the logic by adding a dashboard unique hash value to the object labels.(LOG-2066)
This update changes the log4j dependency to 2.17.1 to resolve CVE-2021-44832.(LOG-2102)
This release includes RHSA-2022:0044 OpenShift Logging Bug Fix Release 5.3.2
Before this update, Elasticsearch rejected logs from the Event Router due to a parsing error. This update changes the data model to resolve the parsing error. However, as a result, previous indices might cause warnings or errors within Kibana. The kubernetes.event.metadata.resourceVersion
field causes errors until existing indices are removed or reindexed. If this field is not used in Kibana, you can ignore the error messages. If you have a retention policy that deletes old indices, the policy eventually removes the old indices and stops the error messages. Otherwise, manually reindex to stop the error messages. (LOG-2087)
Before this update, the OpenShift Logging Dashboard displayed the wrong pod namespace in the table that displays top producing and collected containers over the last 24 hours. With this update, the OpenShift Logging Dashboard displays the correct pod namespace. (LOG-2051)
Before this update, if outputDefaults.elasticsearch.structuredTypeKey
in the ClusterLogForwarder
custom resource (CR) instance did not have a structured key, the CR replaced the output secret with the default secret used to communicate to the default log store. With this update, the defined output secret is correctly used. (LOG-2046)
This release includes RHSA-2021:5129 OpenShift Logging Bug Fix Release 5.3.1
Before this update, the Fluentd container image included builder tools that were unnecessary at run time. This update removes those tools from the image. (LOG-1998)
Before this update, the Logging dashboard displayed an empty CPU graph because of a reference to an invalid metric. With this update, the Logging dashboard displays CPU graphs correctly. (LOG-1925)
Before this update, the Elasticsearch Prometheus exporter plugin compiled index-level metrics using a high-cost query that impacted the Elasticsearch node performance. This update implements a lower-cost query that improves performance. (LOG-1897)
This release includes RHSA-2021:4627 OpenShift Logging Bug Fix Release 5.3.0
With this update, authorization options for Log Forwarding have been expanded. Outputs may now be configured with SASL, username/password, or TLS.
Before this update, if you forwarded logs using the syslog protocol, serializing a ruby hash encoded key/value pairs to contain a '⇒' character and replaced tabs with "#11". This update fixes the issue so that log messages are correctly serialized as valid JSON. (LOG-1494)
Before this update, application logs were not correctly configured to forward to the proper Cloudwatch stream with multi-line error detection enabled. (LOG-1939)
Before this update, a name change of the deployed collector in the 5.3 release caused the alert 'fluentnodedown' to generate. (LOG-1918)
Before this update, a regression introduced in a prior release configuration caused the collector to flush its buffered messages before shutdown, creating a delay the termination and restart of collector Pods. With this update, fluentd no longer flushes buffers at shutdown, resolving the issue. (LOG-1735)
Before this update, a regression introduced in a prior release intentionally disabled JSON message parsing. This update re-enables JSON parsing. It also sets the log entry "level" based on the "level" field in parsed JSON message or by using regex to extract a match from a message field. (LOG-1199)
Before this update, the ClusterLogging
custom resource (CR) applied the value of the totalLimitSize
field to the Fluentd total_limit_size
field, even if the required buffer space was not available. With this update, the CR applies the lesser of the two totalLimitSize
or 'default' values to the Fluentd total_limit_size
field, resolving the issue. (LOG-1776)
If you forward logs to an external Elasticsearch server and then change a configured value in the pipeline secret, such as the username and password, the Fluentd forwarder loads the new secret but uses the old value to connect to an external Elasticsearch server. This issue happens because the Red Hat OpenShift Logging Operator does not currently monitor secrets for content changes. (LOG-1652)
As a workaround, if you change the secret, you can force the Fluentd pods to redeploy by entering:
$ oc delete pod -l component=collector
Some features available in previous releases have been deprecated or removed.
Deprecated functionality is still included in OpenShift Logging and continues to be supported; however, it will be removed in a future release of this product and is not recommended for new deployments.
In OpenShift Logging 5.3, the legacy methods of forwarding logs to Syslog and Fluentd are removed. Bug fixes and support are provided through the end of the OpenShift Logging 5.2 life cycle. After which, no new feature enhancements are made.
Instead, use the following non-legacy methods:
This release includes RHSA-2022:5909-OpenShift Logging Bug Fix Release 5.2.13.
This release includes RHBA-2022:5558-OpenShift Logging Bug Fix Release 5.2.12.
This release includes RHBA-2022:5012-OpenShift Logging Bug Fix Release 5.2.11
Before this update, clusters configured to perform CloudWatch forwarding wrote rejected log files to temporary storage, causing cluster instability over time. With this update, chunk backup for CloudWatch has been disabled, resolving the issue. (LOG-2635)
This release includes OpenShift Logging Bug Fix Release 5.2.10]
Before this update some log forwarder outputs could re-order logs with the same time-stamp. With this update, a sequence number has been added to the log record to order entries that have matching timestamps.(LOG-2335)
Before this update, clusters with a large number of namespaces caused Elasticsearch to stop serving requests because the list of namespaces reached the maximum header size limit. With this update, headers only include a list of namespace names, resolving the issue. (LOG-2475)
Before this update, system:serviceaccount:openshift-monitoring:prometheus-k8s
had cluster level privileges as a clusterrole
and clusterrolebinding
. This update restricts the serviceaccount
to the openshift-logging
namespace with a role and rolebinding. (LOG-2480)
Before this update, the cluster-logging-operator
utilized cluster scoped roles and bindings to establish permissions for the Prometheus service account to scrape metrics. These permissions were only created when deploying the Operator using the console interface and were missing when the Operator was deployed from the command line. This fixes the issue by making this role and binding namespace scoped. (LOG-1972)
This release includes RHBA-2022:1375 OpenShift Logging Bug Fix Release 5.2.9]
Before this update, defining a toleration with no key and the existing Operator caused the Operator to be unable to complete an upgrade. With this update, this toleration no longer blocks the upgrade from completing. (LOG-2304)
This release includes RHSA-2022:0728 OpenShift Logging Bug Fix Release 5.2.8
Before this update, if you removed OpenShift Logging from OKD, the web console continued displaying a link to the Logging page. With this update, removing or uninstalling OpenShift Logging also removes that link. (LOG-2180)
This release includes RHBA-2022:0478 OpenShift Logging Bug Fix Release 5.2.7
Before this update, Elasticsearch pods with FIPS enabled failed to start after updating. With this update, Elasticsearch pods start successfully. (LOG-2000)
Before this update, if a persistent volume claim (PVC) already existed, Elasticsearch generated an error, "Unable to create PersistentVolumeClaim due to forbidden: exceeded quota: infra-storage-quota." With this update, Elasticsearch checks for existing PVCs, resolving the issue. (LOG-2118)
This release includes RHSA-2022:0230 OpenShift Logging Bug Fix Release 5.2.6
Before this update, the release did not include a filter change which caused Fluentd to crash. With this update, the missing filter has been corrected. (LOG-2104)
This update changes the log4j dependency to 2.17.1 to resolve CVE-2021-44832.(LOG-2101)
This release includes RHSA-2022:0043 OpenShift Logging Bug Fix Release 5.2.5
Before this update, Elasticsearch rejected logs from the Event Router due to a parsing error. This update changes the data model to resolve the parsing error. However, as a result, previous indices might cause warnings or errors within Kibana. The kubernetes.event.metadata.resourceVersion
field causes errors until existing indices are removed or reindexed. If this field is not used in Kibana, you can ignore the error messages. If you have a retention policy that deletes old indices, the policy eventually removes the old indices and stops the error messages. Otherwise, manually reindex to stop the error messages. LOG-2087)
This release includes RHSA-2021:5127 OpenShift Logging Bug Fix Release 5.2.4
Before this update, records shipped via syslog would serialize a ruby hash encoding key/value pairs to contain a '⇒' character, as well as replace tabs with "#11". This update serializes the message correctly as proper JSON. (LOG-1775)
Before this update, the Elasticsearch Prometheus exporter plugin compiled index-level metrics using a high-cost query that impacted the Elasticsearch node performance. This update implements a lower-cost query that improves performance. (LOG-1970)
Before this update, Elasticsearch sometimes rejected messages when Log Forwarding was configured with multiple outputs. This happened because configuring one of the outputs modified message content to be a single message. With this update, Log Forwarding duplicates the messages for each output so that output-specific processing does not affect the other outputs. (LOG-1824)
This release includes RHSA-2021:4032 OpenShift Logging Bug Fix Release 5.2.3
Before this update, some alerts did not include a namespace label. This omission does not comply with the OpenShift Monitoring Team’s guidelines for writing alerting rules in OKD. With this update, all the alerts in Elasticsearch Operator include a namespace label and follow all the guidelines for writing alerting rules in OKD. (LOG-1857)
Before this update, a regression introduced in a prior release intentionally disabled JSON message parsing. This update re-enables JSON parsing. It also sets the log entry level
based on the level
field in parsed JSON message or by using regex to extract a match from a message field. (LOG-1759)
This release includes RHBA-2021:3747 OpenShift Logging Bug Fix Release 5.2.2
Before this update, the ClusterLogging
custom resource (CR) applied the value of the totalLimitSize
field to the Fluentd total_limit_size
field, even if the required buffer space was not available. With this update, the CR applies the lesser of the two totalLimitSize
or 'default' values to the Fluentd total_limit_size
field, resolving the issue.(LOG-1738)
Before this update, a regression introduced in a prior release configuration caused the collector to flush its buffered messages before shutdown, creating a delay to the termination and restart of collector pods. With this update, Fluentd no longer flushes buffers at shutdown, resolving the issue. (LOG-1739)
Before this update, an issue in the bundle manifests prevented installation of the Elasticsearch Operator through OLM on OKD 4.9. With this update, a correction to bundle manifests re-enables installation and upgrade in 4.9.(LOG-1780)
This release includes RHBA-2021:3550 OpenShift Logging Bug Fix Release 5.2.1
Before this update, due to an issue in the release pipeline scripts, the value of the olm.skipRange
field remained unchanged at 5.2.0
instead of reflecting the current release number. This update fixes the pipeline scripts to update the value of this field when the release numbers change. (LOG-1743)
This release includes RHBA-2021:3393 OpenShift Logging Bug Fix Release 5.2.0
With this update, you can forward log data to Amazon CloudWatch, which provides application and infrastructure monitoring. For more information, see Forwarding logs to Amazon CloudWatch. (LOG-1173)
With this update, you can forward log data to Loki, a horizontally scalable, highly available, multi-tenant log aggregation system. For more information, see Forwarding logs to Loki. (LOG-684)
With this update, if you use the Fluentd forward protocol to forward log data over a TLS-encrypted connection, now you can use a password-encrypted private key file and specify the passphrase in the Cluster Log Forwarder configuration. For more information, see Forwarding logs using the Fluentd forward protocol. (LOG-1525)
This enhancement enables you to use a username and password to authenticate a log forwarding connection to an external Elasticsearch instance. For example, if you cannot use mutual TLS (mTLS) because a third-party operates the Elasticsearch instance, you can use HTTP or HTTPS and set a secret that contains the username and password. For more information, see Forwarding logs to an external Elasticsearch instance. (LOG-1022)
With this update, you can collect OVN network policy audit logs for forwarding to a logging server. (LOG-1526)
By default, the data model introduced in OKD 4.5 gave logs from different namespaces a single index in common. This change made it harder to see which namespaces produced the most logs.
The current release adds namespace metrics to the Logging dashboard in the OKD console. With these metrics, you can see which namespaces produce logs and how many logs each namespace produces for a given timestamp.
To see these metrics, open the Administrator perspective in the OKD web console, and navigate to Observe → Dashboards → Logging/Elasticsearch. (LOG-1680)
The current release, OpenShift Logging 5.2, enables two new metrics: For a given timestamp or duration, you can see the total logs produced or logged by individual containers, and the total logs collected by the collector. These metrics are labeled by namespace, pod, and container name so that you can see how many logs each namespace and pod collects and produces. (LOG-1213)
Before this update, when the OpenShift Elasticsearch Operator created index management cronjobs, it added the POLICY_MAPPING
environment variable twice, which caused the apiserver to report the duplication. This update fixes the issue so that the POLICY_MAPPING
environment variable is set only once per cronjob, and there is no duplication for the apiserver to report. (LOG-1130)
Before this update, suspending an Elasticsearch cluster to zero nodes did not suspend the index-management cronjobs, which put these cronjobs into maximum backoff. Then, after unsuspending the Elasticsearch cluster, these cronjobs stayed halted due to maximum backoff reached. This update resolves the issue by suspending the cronjobs and the cluster. (LOG-1268)
Before this update, in the Logging dashboard in the OKD console, the list of top 10 log-producing containers was missing the "chart namespace" label and provided the incorrect metric name, fluentd_input_status_total_bytes_logged
. With this update, the chart shows the namespace label and the correct metric name, log_logged_bytes_total
. (LOG-1271)
Before this update, if an index management cronjob terminated with an error, it did not report the error exit code: instead, its job status was "complete." This update resolves the issue by reporting the error exit codes of index management cronjobs that terminate with errors. (LOG-1273)
The priorityclasses.v1beta1.scheduling.k8s.io
was removed in 1.22 and replaced by priorityclasses.v1.scheduling.k8s.io
(v1beta1
was replaced by v1
). Before this update, APIRemovedInNextReleaseInUse
alerts were generated for priorityclasses
because v1beta1
was still present . This update resolves the issue by replacing v1beta1
with v1
. The alert is no longer generated. (LOG-1385)
Previously, the OpenShift Elasticsearch Operator and Red Hat OpenShift Logging Operator did not have the annotation that was required for them to appear in the OKD web console list of Operators that can run in a disconnected environment. This update adds the operators.openshift.io/infrastructure-features: '["Disconnected"]'
annotation to these two Operators so that they appear in the list of Operators that run in disconnected environments. (LOG-1420)
Before this update, Red Hat OpenShift Logging Operator pods were scheduled on CPU cores that were reserved for customer workloads on performance-optimized single-node clusters. With this update, cluster logging Operator pods are scheduled on the correct CPU cores. (LOG-1440)
Before this update, some log entries had unrecognized UTF-8 bytes, which caused Elasticsearch to reject the messages and block the entire buffered payload. With this update, rejected payloads drop the invalid log entries and resubmit the remaining entries to resolve the issue. (LOG-1499)
Before this update, the kibana-proxy
pod sometimes entered the CrashLoopBackoff
state and logged the following message Invalid configuration: cookie_secret must be 16, 24, or 32 bytes to create an AES cipher when pass_access_token == true or cookie_refresh != 0, but is 29 bytes.
The exact actual number of bytes could vary. With this update, the generation of the Kibana session secret has been corrected, and the kibana-proxy pod no longer enters a CrashLoopBackoff
state due to this error. (LOG-1446)
Before this update, the AWS CloudWatch Fluentd plugin logged its AWS API calls to the Fluentd log at all log levels, consuming additional OKD node resources. With this update, the AWS CloudWatch Fluentd plugin logs AWS API calls only at the "debug" and "trace" log levels. This way, at the default "warn" log level, Fluentd does not consume extra node resources. (LOG-1071)
Before this update, the Elasticsearch OpenDistro security plugin caused user index migrations to fail. This update resolves the issue by providing a newer version of the plugin. Now, index migrations proceed without errors. (LOG-1276)
Before this update, in the Logging dashboard in the OKD console, the list of top 10 log-producing containers lacked data points. This update resolves the issue, and the dashboard displays all data points. (LOG-1353)
Before this update, if you were tuning the performance of the Fluentd log forwarder by adjusting the chunkLimitSize
and totalLimitSize
values, the Setting queued_chunks_limit_size for each buffer to
message reported values that were too low. The current update fixes this issue so that this message reports the correct values. (LOG-1411)
Before this update, the Kibana OpenDistro security plugin caused user index migrations to fail. This update resolves the issue by providing a newer version of the plugin. Now, index migrations proceed without errors. (LOG-1558)
Before this update, using a namespace input filter prevented logs in that namespace from appearing in other inputs. With this update, logs are sent to all inputs that can accept them. (LOG-1570)
Before this update, a missing license file for the viaq/logerr
dependency caused license scanners to abort without success. With this update, the viaq/logerr
dependency is licensed under Apache 2.0 and the license scanners run successfully. (LOG-1590)
Before this update, an incorrect brew tag for curator5
within the elasticsearch-operator-bundle
build pipeline caused the pull of an image pinned to a dummy SHA1. With this update, the build pipeline uses the logging-curator5-rhel8
reference for curator5
, enabling index management cronjobs to pull the correct image from registry.redhat.io
. (LOG-1624)
Before this update, an issue with the ServiceAccount
permissions caused errors such as no permissions for [indices:admin/aliases/get]
. With this update, a permission fix resolves the issue. (LOG-1657)
Before this update, the Custom Resource Definition (CRD) for the Red Hat OpenShift Logging Operator was missing the Loki output type, which caused the admission controller to reject the ClusterLogForwarder
custom resource object. With this update, the CRD includes Loki as an output type so that administrators can configure ClusterLogForwarder
to send logs to a Loki server. (LOG-1683)
Before this update, OpenShift Elasticsearch Operator reconciliation of the ServiceAccounts
overwrote third-party-owned fields that contained secrets. This issue caused memory and CPU spikes due to frequent recreation of secrets. This update resolves the issue. Now, the OpenShift Elasticsearch Operator does not overwrite third-party-owned fields. (LOG-1714)
Before this update, in the ClusterLogging
custom resource (CR) definition, if you specified a flush_interval
value but did not set flush_mode
to interval
, the Red Hat OpenShift Logging Operator generated a Fluentd configuration. However, the Fluentd collector generated an error at runtime. With this update, the Red Hat OpenShift Logging Operator validates the ClusterLogging
CR definition and only generates the Fluentd configuration if both fields are specified. (LOG-1723)
If you forward logs to an external Elasticsearch server and then change a configured value in the pipeline secret, such as the username and password, the Fluentd forwarder loads the new secret but uses the old value to connect to an external Elasticsearch server. This issue happens because the Red Hat OpenShift Logging Operator does not currently monitor secrets for content changes. (LOG-1652)
As a workaround, if you change the secret, you can force the Fluentd pods to redeploy by entering:
$ oc delete pod -l component=collector
Some features available in previous releases have been deprecated or removed.
Deprecated functionality is still included in OpenShift Logging and continues to be supported; however, it will be removed in a future release of this product and is not recommended for new deployments.
From OKD 4.6 to the present, forwarding logs by using the following legacy methods have been deprecated and will be removed in a future release:
Forwarding logs using the legacy Fluentd method
Forwarding logs using the legacy syslog method
Instead, use the following non-legacy methods: