Resource metrics pipeline
For Kubernetes, the Metrics API offers a basic set of metrics to support automatic scaling and similar use cases. This API makes information available about resource usage for node and pod, including metrics for CPU and memory. If you deploy the Metrics API into your cluster, clients of the Kubernetes API can then query for this information, and you can use Kubernetes' access control mechanisms to manage permissions to do so.
The HorizontalPodAutoscaler (HPA) and
You can also view the resource metrics using the
Figure 1 illustrates the architecture of the resource metrics pipeline. Figure 1. Resource Metrics Pipeline The architecture components, from right to left in the figure, consist of the following: kubelet: Node agent for managing container
resources. Resource metrics are accessible using the Summary API: API provided by the kubelet for discovering and retrieving
per-node summarized stats available through the metrics-server: Cluster addon component that collects and aggregates resource
metrics pulled from each kubelet. The API server serves Metrics API for use by HPA, VPA, and by
the Metrics API: Kubernetes API supporting access to CPU and memory used for
workload autoscaling. To make this work in your cluster, you need an API extension server that
provides the Metrics API. The metrics-server implements the Metrics API. This API allows you to access CPU and memory usage
for the nodes and pods in your cluster. Its primary role is to feed resource usage metrics to K8s
autoscaler components. Here is an example of the Metrics API request for a Here is the same API call using Sample response: Here is an example of the Metrics API request for a Here is the same API call using Sample response: The Metrics API is defined in the
repository. You must enable the API aggregation layer
and register an APIService
for the To learn more about the Metrics API, see ,
the . CPU is reported as the average core usage measured in cpu units. One cpu, in Kubernetes, is
equivalent to 1 vCPU/Core for cloud providers, and 1 hyper-thread on bare-metal Intel processors. This value is derived by taking a rate over a cumulative CPU counter provided by the kernel (in
both Linux and Windows kernels). The time window used to calculate CPU is shown under window field
in Metrics API. To learn more about how Kubernetes allocates and measures CPU resources, see
meaning of CPU. Memory is reported as the working set, measured in bytes, at the instant the metric was collected. In an ideal world, the "working set" is the amount of memory in-use that cannot be freed under
memory pressure. However, calculation of the working set varies by host OS, and generally makes
heavy use of heuristics to produce an estimate. The Kubernetes model for a container's working set expects that the container runtime counts
anonymous memory associated with the container in question. The working set metric typically also
includes some cached (file-backed) memory, because the host OS cannot always reclaim pages. To learn more about how Kubernetes allocates and measures memory resources, see
meaning of memory. The metrics-server fetches resource metrics from the kubelets and exposes them in the Kubernetes
API server through the Metrics API for use by the HPA and VPA. You can also view these metrics
using the The metrics-server uses the Kubernetes API to track nodes and pods in your cluster. The
metrics-server queries each node over HTTP to fetch metrics. The metrics-server also builds an
internal view of pod metadata, and keeps a cache of pod health. That cached pod health information
is available via the extension API that the metrics-server makes available. For example with an HPA query, the metrics-server needs to identify which pods fulfill the label
selectors in the deployment. The metrics-server calls the kubelet API
to collect metrics from each node. Depending on the metrics-server version it uses: To learn more about the metrics-server, see the
. You can also check out the following: The kubelet gathers stats at the node,
volume, pod and container level, and emits this information in
the
for consumers to read. Here is an example of a Summary API request for a Here is the same API call using kubectl top
command.
]
A[Metrics-
Server]
subgraph B[Nodes]
direction TB
D[cAdvisor] --> C[kubelet]
E[Container
runtime] --> D
E1[Container
runtime] --> D
P[pod data] -.- C
end
L[API
server]
W[HPA]
C ---->|Summary
API| A -->|metrics
API| L --> W
end
L ---> K[kubectl
top]
classDef box fill:#fff,stroke:#000,stroke-width:1px,color:#000;
class W,B,P,K,cluster,D,E,E1 box
classDef spacewhite fill:#ffffff,stroke:#fff,stroke-width:0px,color:#000
class S spacewhite
classDef k8s fill:#326ce5,stroke:#fff,stroke-width:1px,color:#fff;
class A,L,C k8s
/metrics/resource
and /stats
kubelet
API endpoints./stats
endpoint.kubectl top
command. Metrics Server is a reference implementation of the Metrics API.Metrics API
Kubernetes 1.8 [beta]
minikube
node piped through jq
for easier
reading:kubectl get --raw "/apis/metrics.k8s.io/v1beta1/nodes/minikube" | jq '.'
curl
:curl http://localhost:8080/apis/metrics.k8s.io/v1beta1/nodes/minikube
{
"kind": "NodeMetrics",
"apiVersion": "metrics.k8s.io/v1beta1",
"metadata": {
"name": "minikube",
"selfLink": "/apis/metrics.k8s.io/v1beta1/nodes/minikube",
"creationTimestamp": "2022-01-27T18:48:43Z"
},
"timestamp": "2022-01-27T18:48:33Z",
"window": "30s",
"usage": {
"cpu": "487558164n",
"memory": "732212Ki"
}
}
kube-scheduler-minikube
pod contained in the
kube-system
namespace and piped through jq
for easier reading:kubectl get --raw "/apis/metrics.k8s.io/v1beta1/namespaces/kube-system/pods/kube-scheduler-minikube" | jq '.'
curl
:curl http://localhost:8080/apis/metrics.k8s.io/v1beta1/namespaces/kube-system/pods/kube-scheduler-minikube
{
"kind": "PodMetrics",
"apiVersion": "metrics.k8s.io/v1beta1",
"metadata": {
"name": "kube-scheduler-minikube",
"namespace": "kube-system",
"selfLink": "/apis/metrics.k8s.io/v1beta1/namespaces/kube-system/pods/kube-scheduler-minikube",
"creationTimestamp": "2022-01-27T19:25:00Z"
},
"timestamp": "2022-01-27T19:24:31Z",
"window": "30s",
"containers": [
{
"name": "kube-scheduler",
"usage": {
"cpu": "9559630n",
"memory": "22244Ki"
}
}
]
}
metrics.k8s.io
API.Measuring resource usage
CPU
Memory
Metrics Server
kubectl top
command.
/metrics/resource
in version v0.6.0+ or/stats/summary
in older versionsSummary API source
minikube
node:kubectl get --raw "/api/v1/nodes/minikube/proxy/stats/summary"
curl
:curl http://localhost:8080/api/v1/nodes/minikube/proxy/stats/summary
/stats/summary
endpoint will be replaced by the /metrics/resource
endpoint
beginning with metrics-server 0.6.x.