You can use the Knative Serving - Scaling Debugging dashboard to examine detailed and visualized data for Knative Serving autoscaling. The dashboard is useful for several purposes:
Troubleshooting your autoscaled workloads
Improving understanding of how autoscaling works
Determining why an application was autoscaled
Evaluating resource footprint of an application, such as number of pods
Currently, this dashboard only supports the Knative pod autoscaler (KPA). It does not support the horizontal pod autoscaler (HPA).
The dashboard demonstrations in this section use an OpenShift Container Platform cluster with the
autoscale-go sample application installed. The load is generated using the
hey load generator.
The sample application has a concurrency limit of 5 requests. When the limit is exceeded, autoscaling requests additional pods for Knative from Kubernetes.
You can use the following steps to navigate to the autoscaling dashboard in the OpenShift Container Platform web console.
You have logged in to the OpenShift Container Platform web console.
You have installed the OpenShift Serverless Operator and Knative Serving.
In the Developer perspective, navigate to the Monitoring → Dashboards page.
In the Dashboard field, select the Knative Serving - Scaling Debugging dashboard.
Use the Namespace, Configuration, and Revision fields to specify the workload you want to examine.
The top of the Knative Serving - Scaling Debugging dashboard shows the counts of the requested pods, as well as of the pods in various stages of deployment. The Revision Pod Counts (Timeline) graph shows the same data visualized on the timeline. This information might be useful for general assessment of autoscaling by checking for problems with pod allocation.