$ oc -n openshift-logging edit ClusterLogging instance
OpenShift Container Platform uses Kibana to display the log data collected by the logging subsystem.
You can scale Kibana for redundancy and configure the CPU and memory for your Kibana nodes.
The logging subsystem components allow for adjustments to both the CPU and memory limits.
Edit the ClusterLogging
custom resource (CR) in the openshift-logging
project:
$ oc -n openshift-logging edit ClusterLogging instance
apiVersion: "logging.openshift.io/v1"
kind: "ClusterLogging"
metadata:
name: "instance"
namespace: openshift-logging
...
spec:
managementState: "Managed"
logStore:
type: "elasticsearch"
elasticsearch:
nodeCount: 3
resources: (1)
limits:
memory: 16Gi
requests:
cpu: 200m
memory: 16Gi
storage:
storageClassName: "gp2"
size: "200G"
redundancyPolicy: "SingleRedundancy"
visualization:
type: "kibana"
kibana:
resources: (2)
limits:
memory: 1Gi
requests:
cpu: 500m
memory: 1Gi
proxy:
resources: (2)
limits:
memory: 100Mi
requests:
cpu: 100m
memory: 100Mi
replicas: 2
collection:
logs:
type: "fluentd"
fluentd:
resources: (3)
limits:
memory: 736Mi
requests:
cpu: 200m
memory: 736Mi
1 | Specify the CPU and memory limits and requests for the log store as needed. For Elasticsearch, you must adjust both the request value and the limit value. |
2 | Specify the CPU and memory limits and requests for the log visualizer as needed. |
3 | Specify the CPU and memory limits and requests for the log collector as needed. |
You can scale the pod that hosts the log visualizer for redundancy.
Edit the ClusterLogging
custom resource (CR) in the openshift-logging
project:
$ oc edit ClusterLogging instance
$ oc edit ClusterLogging instance
apiVersion: "logging.openshift.io/v1"
kind: "ClusterLogging"
metadata:
name: "instance"
....
spec:
visualization:
type: "kibana"
kibana:
replicas: 1 (1)
1 | Specify the number of Kibana nodes. |