1- name: version_info
2 namespace: etcd
3 help: Etcd server's binary version
4 type: Gauge
5 stabilityLevel: ALPHA
6 labels:
7 - binary_version
8- name: certificate_manager_client_ttl_seconds
9 subsystem: kubelet
10 help: Gauge of the TTL (time-to-live) of the Kubelet's client certificate. The value
11 is in seconds until certificate expiry (negative if already expired). If client
12 certificate is invalid or unused, the value will be +INF.
13 type: Gauge
14 stabilityLevel: ALPHA
15- name: addresses_skipped_per_sync
16 subsystem: endpoint_slice_mirroring_controller
17 help: Number of addresses skipped on each Endpoints sync due to being invalid or
18 exceeding MaxEndpointsPerSubset
19 type: Histogram
20 stabilityLevel: ALPHA
21 buckets:
22 - 2
23 - 4
24 - 8
25 - 16
26 - 32
27 - 64
28 - 128
29 - 256
30 - 512
31 - 1024
32 - 2048
33 - 4096
34 - 8192
35 - 16384
36 - 32768
37- name: changes
38 subsystem: endpoint_slice_mirroring_controller
39 help: Number of EndpointSlice changes
40 type: Counter
41 stabilityLevel: ALPHA
42 labels:
43 - operation
44- name: desired_endpoint_slices
45 subsystem: endpoint_slice_mirroring_controller
46 help: Number of EndpointSlices that would exist with perfect endpoint allocation
47 type: Gauge
48 stabilityLevel: ALPHA
49- name: endpoints_added_per_sync
50 subsystem: endpoint_slice_mirroring_controller
51 help: Number of endpoints added on each Endpoints sync
52 type: Histogram
53 stabilityLevel: ALPHA
54 buckets:
55 - 2
56 - 4
57 - 8
58 - 16
59 - 32
60 - 64
61 - 128
62 - 256
63 - 512
64 - 1024
65 - 2048
66 - 4096
67 - 8192
68 - 16384
69 - 32768
70- name: endpoints_desired
71 subsystem: endpoint_slice_mirroring_controller
72 help: Number of endpoints desired
73 type: Gauge
74 stabilityLevel: ALPHA
75- name: endpoints_removed_per_sync
76 subsystem: endpoint_slice_mirroring_controller
77 help: Number of endpoints removed on each Endpoints sync
78 type: Histogram
79 stabilityLevel: ALPHA
80 buckets:
81 - 2
82 - 4
83 - 8
84 - 16
85 - 32
86 - 64
87 - 128
88 - 256
89 - 512
90 - 1024
91 - 2048
92 - 4096
93 - 8192
94 - 16384
95 - 32768
96- name: endpoints_sync_duration
97 subsystem: endpoint_slice_mirroring_controller
98 help: Duration of syncEndpoints() in seconds
99 type: Histogram
100 stabilityLevel: ALPHA
101 buckets:
102 - 0.001
103 - 0.002
104 - 0.004
105 - 0.008
106 - 0.016
107 - 0.032
108 - 0.064
109 - 0.128
110 - 0.256
111 - 0.512
112 - 1.024
113 - 2.048
114 - 4.096
115 - 8.192
116 - 16.384
117- name: endpoints_updated_per_sync
118 subsystem: endpoint_slice_mirroring_controller
119 help: Number of endpoints updated on each Endpoints sync
120 type: Histogram
121 stabilityLevel: ALPHA
122 buckets:
123 - 2
124 - 4
125 - 8
126 - 16
127 - 32
128 - 64
129 - 128
130 - 256
131 - 512
132 - 1024
133 - 2048
134 - 4096
135 - 8192
136 - 16384
137 - 32768
138- name: num_endpoint_slices
139 subsystem: endpoint_slice_mirroring_controller
140 help: Number of EndpointSlices
141 type: Gauge
142 stabilityLevel: ALPHA
143- name: sync_duration_seconds
144 subsystem: root_ca_cert_publisher
145 help: Number of namespace syncs happened in root ca cert publisher.
146 type: Histogram
147 stabilityLevel: ALPHA
148 labels:
149 - code
150 buckets:
151 - 0.001
152 - 0.002
153 - 0.004
154 - 0.008
155 - 0.016
156 - 0.032
157 - 0.064
158 - 0.128
159 - 0.256
160 - 0.512
161 - 1.024
162 - 2.048
163 - 4.096
164 - 8.192
165 - 16.384
166- name: sync_total
167 subsystem: root_ca_cert_publisher
168 help: Number of namespace syncs happened in root ca cert publisher.
169 type: Counter
170 stabilityLevel: ALPHA
171 labels:
172 - code
173- name: job_creation_skew_duration_seconds
174 subsystem: cronjob_controller
175 help: Time between when a cronjob is scheduled to be run, and when the corresponding
176 job is created
177 type: Histogram
178 stabilityLevel: STABLE
179 buckets:
180 - 1
181 - 2
182 - 4
183 - 8
184 - 16
185 - 32
186 - 64
187 - 128
188 - 256
189 - 512
190- name: resources_sync_error_total
191 subsystem: garbagecollector_controller
192 help: Number of garbage collector resources sync errors
193 type: Counter
194 stabilityLevel: ALPHA
195- name: metric_computation_duration_seconds
196 subsystem: horizontal_pod_autoscaler_controller
197 help: The time(seconds) that the HPA controller takes to calculate one metric. The
198 label 'action' should be either 'scale_down', 'scale_up', or 'none'. The label
199 'error' should be either 'spec', 'internal', or 'none'. The label 'metric_type'
200 corresponds to HPA.spec.metrics[*].type
201 type: Histogram
202 stabilityLevel: ALPHA
203 labels:
204 - action
205 - error
206 - metric_type
207 buckets:
208 - 0.001
209 - 0.002
210 - 0.004
211 - 0.008
212 - 0.016
213 - 0.032
214 - 0.064
215 - 0.128
216 - 0.256
217 - 0.512
218 - 1.024
219 - 2.048
220 - 4.096
221 - 8.192
222 - 16.384
223- name: metric_computation_total
224 subsystem: horizontal_pod_autoscaler_controller
225 help: Number of metric computations. The label 'action' should be either 'scale_down',
226 'scale_up', or 'none'. Also, the label 'error' should be either 'spec', 'internal',
227 or 'none'. The label 'metric_type' corresponds to HPA.spec.metrics[*].type
228 type: Counter
229 stabilityLevel: ALPHA
230 labels:
231 - action
232 - error
233 - metric_type
234- name: reconciliation_duration_seconds
235 subsystem: horizontal_pod_autoscaler_controller
236 help: The time(seconds) that the HPA controller takes to reconcile once. The label
237 'action' should be either 'scale_down', 'scale_up', or 'none'. Also, the label
238 'error' should be either 'spec', 'internal', or 'none'. Note that if both spec
239 and internal errors happen during a reconciliation, the first one to occur is
240 reported in `error` label.
241 type: Histogram
242 stabilityLevel: ALPHA
243 labels:
244 - action
245 - error
246 buckets:
247 - 0.001
248 - 0.002
249 - 0.004
250 - 0.008
251 - 0.016
252 - 0.032
253 - 0.064
254 - 0.128
255 - 0.256
256 - 0.512
257 - 1.024
258 - 2.048
259 - 4.096
260 - 8.192
261 - 16.384
262- name: reconciliations_total
263 subsystem: horizontal_pod_autoscaler_controller
264 help: Number of reconciliations of HPA controller. The label 'action' should be
265 either 'scale_down', 'scale_up', or 'none'. Also, the label 'error' should be
266 either 'spec', 'internal', or 'none'. Note that if both spec and internal errors
267 happen during a reconciliation, the first one to occur is reported in `error`
268 label.
269 type: Counter
270 stabilityLevel: ALPHA
271 labels:
272 - action
273 - error
274- name: job_finished_indexes_total
275 subsystem: job_controller
276 help: "`The number of finished indexes. Possible values for the\n\t\t\tstatus label
277 are: \"succeeded\", \"failed\". Possible values for the\n\t\t\tbackoffLimit label
278 are: \"perIndex\" and \"global\"`"
279 type: Counter
280 stabilityLevel: ALPHA
281 labels:
282 - backoffLimit
283 - status
284- name: job_pods_creation_total
285 subsystem: job_controller
286 help: |-
287 `The number of Pods created by the Job controller labelled with a reason for the Pod creation.
288 This metric also distinguishes between Pods created using different PodReplacementPolicy settings.
289 Possible values of the "reason" label are:
290 "new", "recreate_terminating_or_failed", "recreate_failed".
291 Possible values of the "status" label are:
292 "succeeded", "failed".`
293 type: Counter
294 stabilityLevel: ALPHA
295 labels:
296 - reason
297 - status
298- name: pod_failures_handled_by_failure_policy_total
299 subsystem: job_controller
300 help: "`The number of failed Pods handled by failure policy with\n\t\t\trespect
301 to the failure policy action applied based on the matched\n\t\t\trule. Possible
302 values of the action label correspond to the\n\t\t\tpossible values for the failure
303 policy rule action, which are:\n\t\t\t\"FailJob\", \"Ignore\" and \"Count\".`"
304 type: Counter
305 stabilityLevel: ALPHA
306 labels:
307 - action
308- name: terminated_pods_tracking_finalizer_total
309 subsystem: job_controller
310 help: |-
311 `The number of terminated pods (phase=Failed|Succeeded)
312 that have the finalizer batch.kubernetes.io/job-tracking
313 The event label can be "add" or "delete".`
314 type: Counter
315 stabilityLevel: ALPHA
316 labels:
317 - event
318- name: unhealthy_nodes_in_zone
319 subsystem: node_collector
320 help: Gauge measuring number of not Ready Nodes per zones.
321 type: Gauge
322 stabilityLevel: ALPHA
323 labels:
324 - zone
325- name: update_all_nodes_health_duration_seconds
326 subsystem: node_collector
327 help: Duration in seconds for NodeController to update the health of all nodes.
328 type: Histogram
329 stabilityLevel: ALPHA
330 buckets:
331 - 0.01
332 - 0.04
333 - 0.16
334 - 0.64
335 - 2.56
336 - 10.24
337 - 40.96
338 - 163.84
339- name: update_node_health_duration_seconds
340 subsystem: node_collector
341 help: Duration in seconds for NodeController to update the health of a single node.
342 type: Histogram
343 stabilityLevel: ALPHA
344 buckets:
345 - 0.001
346 - 0.004
347 - 0.016
348 - 0.064
349 - 0.256
350 - 1.024
351 - 4.096
352 - 16.384
353- name: zone_health
354 subsystem: node_collector
355 help: Gauge measuring percentage of healthy nodes per zone.
356 type: Gauge
357 stabilityLevel: ALPHA
358 labels:
359 - zone
360- name: zone_size
361 subsystem: node_collector
362 help: Gauge measuring number of registered Nodes per zones.
363 type: Gauge
364 stabilityLevel: ALPHA
365 labels:
366 - zone
367- name: cidrset_allocation_tries_per_request
368 subsystem: node_ipam_controller
369 help: Number of endpoints added on each Service sync
370 type: Histogram
371 stabilityLevel: ALPHA
372 labels:
373 - clusterCIDR
374 buckets:
375 - 1
376 - 5
377 - 25
378 - 125
379 - 625
380- name: cidrset_cidrs_allocations_total
381 subsystem: node_ipam_controller
382 help: Counter measuring total number of CIDR allocations.
383 type: Counter
384 stabilityLevel: ALPHA
385 labels:
386 - clusterCIDR
387- name: cidrset_cidrs_releases_total
388 subsystem: node_ipam_controller
389 help: Counter measuring total number of CIDR releases.
390 type: Counter
391 stabilityLevel: ALPHA
392 labels:
393 - clusterCIDR
394- name: cidrset_usage_cidrs
395 subsystem: node_ipam_controller
396 help: Gauge measuring percentage of allocated CIDRs.
397 type: Gauge
398 stabilityLevel: ALPHA
399 labels:
400 - clusterCIDR
401- name: cirdset_max_cidrs
402 subsystem: node_ipam_controller
403 help: Maximum number of CIDRs that can be allocated.
404 type: Gauge
405 stabilityLevel: ALPHA
406 labels:
407 - clusterCIDR
408- name: force_delete_pod_errors_total
409 subsystem: pod_gc_collector
410 help: Number of errors encountered when forcefully deleting the pods since the Pod
411 GC Controller started.
412 type: Counter
413 stabilityLevel: ALPHA
414 labels:
415 - namespace
416 - reason
417- name: force_delete_pods_total
418 subsystem: pod_gc_collector
419 help: Number of pods that are being forcefully deleted since the Pod GC Controller
420 started.
421 type: Counter
422 stabilityLevel: ALPHA
423 labels:
424 - namespace
425 - reason
426- name: sorting_deletion_age_ratio
427 subsystem: replicaset_controller
428 help: The ratio of chosen deleted pod's ages to the current youngest pod's age (at
429 the time). Should be <2. The intent of this metric is to measure the rough efficacy
430 of the LogarithmicScaleDown feature gate's effect on the sorting (and deletion)
431 of pods when a replicaset scales down. This only considers Ready pods when calculating
432 and reporting.
433 type: Histogram
434 stabilityLevel: ALPHA
435 buckets:
436 - 0.25
437 - 0.5
438 - 1
439 - 2
440 - 4
441 - 8
442- name: create_attempts_total
443 subsystem: resourceclaim_controller
444 help: Number of ResourceClaims creation requests
445 type: Counter
446 stabilityLevel: ALPHA
447- name: create_failures_total
448 subsystem: resourceclaim_controller
449 help: Number of ResourceClaims creation request failures
450 type: Counter
451 stabilityLevel: ALPHA
452- name: pod_deletion_duration_seconds
453 subsystem: taint_eviction_controller
454 help: Latency, in seconds, between the time when a taint effect has been activated
455 for the Pod and its deletion via TaintEvictionController.
456 type: Histogram
457 stabilityLevel: ALPHA
458 buckets:
459 - 0.005
460 - 0.025
461 - 0.1
462 - 0.5
463 - 1
464 - 2.5
465 - 10
466 - 30
467 - 60
468 - 120
469 - 180
470 - 240
471- name: pod_deletions_total
472 subsystem: taint_eviction_controller
473 help: Total number of Pods deleted by TaintEvictionController since its start.
474 type: Counter
475 stabilityLevel: ALPHA
476- name: job_pods_finished_total
477 subsystem: job_controller
478 help: The number of finished Pods that are fully tracked
479 type: Counter
480 stabilityLevel: STABLE
481 labels:
482 - completion_mode
483 - result
484- name: job_sync_duration_seconds
485 subsystem: job_controller
486 help: The time it took to sync a job
487 type: Histogram
488 stabilityLevel: STABLE
489 labels:
490 - action
491 - completion_mode
492 - result
493 buckets:
494 - 0.004
495 - 0.008
496 - 0.016
497 - 0.032
498 - 0.064
499 - 0.128
500 - 0.256
501 - 0.512
502 - 1.024
503 - 2.048
504 - 4.096
505 - 8.192
506 - 16.384
507 - 32.768
508 - 65.536
509- name: job_syncs_total
510 subsystem: job_controller
511 help: The number of job syncs
512 type: Counter
513 stabilityLevel: STABLE
514 labels:
515 - action
516 - completion_mode
517 - result
518- name: jobs_finished_total
519 subsystem: job_controller
520 help: The number of finished jobs
521 type: Counter
522 stabilityLevel: STABLE
523 labels:
524 - completion_mode
525 - reason
526 - result
527- name: evictions_total
528 subsystem: node_collector
529 help: Number of Node evictions that happened since current instance of NodeController
530 started.
531 type: Counter
532 stabilityLevel: STABLE
533 labels:
534 - zone
535- name: attachdetach_controller_forced_detaches
536 subsystem: attach_detach_controller
537 help: Number of times the A/D Controller performed a forced detach
538 type: Counter
539 stabilityLevel: ALPHA
540 labels:
541 - reason
542- name: attachdetach_controller_total_volumes
543 help: Number of volumes in A/D Controller
544 type: Custom
545 stabilityLevel: ALPHA
546 labels:
547 - plugin_name
548 - state
549- name: create_failures_total
550 subsystem: ephemeral_volume_controller
551 help: Number of PersistenVolumeClaims creation requests
552 type: Counter
553 stabilityLevel: ALPHA
554- name: create_total
555 subsystem: ephemeral_volume_controller
556 help: Number of PersistenVolumeClaims creation requests
557 type: Counter
558 stabilityLevel: ALPHA
559- name: client_expiration_renew_errors
560 subsystem: certificate_manager
561 namespace: kubelet
562 help: Counter of certificate renewal errors.
563 type: Counter
564 stabilityLevel: ALPHA
565- name: certificate_manager_server_rotation_seconds
566 subsystem: kubelet
567 help: Histogram of the number of seconds the previous certificate lived before being
568 rotated.
569 type: Histogram
570 stabilityLevel: ALPHA
571 buckets:
572 - 60
573 - 3600
574 - 14400
575 - 86400
576 - 604800
577 - 2.592e+06
578 - 7.776e+06
579 - 1.5552e+07
580 - 3.1104e+07
581 - 1.24416e+08
582- name: certificate_manager_server_ttl_seconds
583 subsystem: kubelet
584 help: Gauge of the shortest TTL (time-to-live) of the Kubelet's serving certificate.
585 The value is in seconds until certificate expiry (negative if already expired).
586 If serving certificate is invalid or unused, the value will be +INF.
587 type: Gauge
588 stabilityLevel: ALPHA
589- name: credential_provider_plugin_duration
590 subsystem: kubelet
591 help: Duration of execution in seconds for credential provider plugin
592 type: Histogram
593 stabilityLevel: ALPHA
594 labels:
595 - plugin_name
596 buckets:
597 - 0.005
598 - 0.01
599 - 0.025
600 - 0.05
601 - 0.1
602 - 0.25
603 - 0.5
604 - 1
605 - 2.5
606 - 5
607 - 10
608- name: credential_provider_plugin_errors
609 subsystem: kubelet
610 help: Number of errors from credential provider plugin
611 type: Counter
612 stabilityLevel: ALPHA
613 labels:
614 - plugin_name
615- name: server_expiration_renew_errors
616 subsystem: kubelet
617 help: Counter of certificate renewal errors.
618 type: Counter
619 stabilityLevel: ALPHA
620- name: pv_collector_bound_pv_count
621 help: Gauge measuring number of persistent volume currently bound
622 type: Custom
623 stabilityLevel: ALPHA
624 labels:
625 - storage_class
626- name: pv_collector_bound_pvc_count
627 help: Gauge measuring number of persistent volume claim currently bound
628 type: Custom
629 stabilityLevel: ALPHA
630 labels:
631 - namespace
632- name: pv_collector_total_pv_count
633 help: Gauge measuring total number of persistent volumes
634 type: Custom
635 stabilityLevel: ALPHA
636 labels:
637 - plugin_name
638 - volume_mode
639- name: pv_collector_unbound_pv_count
640 help: Gauge measuring number of persistent volume currently unbound
641 type: Custom
642 stabilityLevel: ALPHA
643 labels:
644 - storage_class
645- name: pv_collector_unbound_pvc_count
646 help: Gauge measuring number of persistent volume claim currently unbound
647 type: Custom
648 stabilityLevel: ALPHA
649 labels:
650 - namespace
651- name: retroactive_storageclass_errors_total
652 help: Total number of failed retroactive StorageClass assignments to persistent
653 volume claim
654 type: Counter
655 stabilityLevel: ALPHA
656- name: retroactive_storageclass_total
657 help: Total number of retroactive StorageClass assignments to persistent volume
658 claim
659 type: Counter
660 stabilityLevel: ALPHA
661- name: storage_count_attachable_volumes_in_use
662 help: Measure number of volumes in use
663 type: Custom
664 stabilityLevel: ALPHA
665 labels:
666 - node
667 - volume_plugin
668- name: job_deletion_duration_seconds
669 subsystem: ttl_after_finished_controller
670 help: The time it took to delete the job since it became eligible for deletion
671 type: Histogram
672 stabilityLevel: ALPHA
673 buckets:
674 - 0.1
675 - 0.2
676 - 0.4
677 - 0.8
678 - 1.6
679 - 3.2
680 - 6.4
681 - 12.8
682 - 25.6
683 - 51.2
684 - 102.4
685 - 204.8
686 - 409.6
687 - 819.2
688- name: volume_operation_total_errors
689 help: Total volume operation errors
690 type: Counter
691 stabilityLevel: ALPHA
692 labels:
693 - operation_name
694 - plugin_name
695- name: container_swap_usage_bytes
696 help: Current amount of the container swap usage in bytes. Reported only on non-windows
697 systems
698 type: Custom
699 stabilityLevel: ALPHA
700 labels:
701 - container
702 - pod
703 - namespace
704- name: active_pods
705 subsystem: kubelet
706 help: The number of pods the kubelet considers active and which are being considered
707 when admitting new pods. static is true if the pod is not from the apiserver.
708 type: Gauge
709 stabilityLevel: ALPHA
710 labels:
711 - static
712- name: cgroup_manager_duration_seconds
713 subsystem: kubelet
714 help: Duration in seconds for cgroup manager operations. Broken down by method.
715 type: Histogram
716 stabilityLevel: ALPHA
717 labels:
718 - operation_type
719 buckets:
720 - 0.005
721 - 0.01
722 - 0.025
723 - 0.05
724 - 0.1
725 - 0.25
726 - 0.5
727 - 1
728 - 2.5
729 - 5
730 - 10
731- name: kubelet_container_log_filesystem_used_bytes
732 help: Bytes used by the container's logs on the filesystem.
733 type: Custom
734 stabilityLevel: ALPHA
735 labels:
736 - uid
737 - namespace
738 - pod
739 - container
740- name: containers_per_pod_count
741 subsystem: kubelet
742 help: The number of containers per pod.
743 type: Histogram
744 stabilityLevel: ALPHA
745 buckets:
746 - 1
747 - 2
748 - 4
749 - 8
750 - 16
751- name: cpu_manager_pinning_errors_total
752 subsystem: kubelet
753 help: The number of cpu core allocations which required pinning failed.
754 type: Counter
755 stabilityLevel: ALPHA
756- name: cpu_manager_pinning_requests_total
757 subsystem: kubelet
758 help: The number of cpu core allocations which required pinning.
759 type: Counter
760 stabilityLevel: ALPHA
761- name: desired_pods
762 subsystem: kubelet
763 help: The number of pods the kubelet is being instructed to run. static is true
764 if the pod is not from the apiserver.
765 type: Gauge
766 stabilityLevel: ALPHA
767 labels:
768 - static
769- name: device_plugin_alloc_duration_seconds
770 subsystem: kubelet
771 help: Duration in seconds to serve a device plugin Allocation request. Broken down
772 by resource name.
773 type: Histogram
774 stabilityLevel: ALPHA
775 labels:
776 - resource_name
777 buckets:
778 - 0.005
779 - 0.01
780 - 0.025
781 - 0.05
782 - 0.1
783 - 0.25
784 - 0.5
785 - 1
786 - 2.5
787 - 5
788 - 10
789- name: device_plugin_registration_total
790 subsystem: kubelet
791 help: Cumulative number of device plugin registrations. Broken down by resource
792 name.
793 type: Counter
794 stabilityLevel: ALPHA
795 labels:
796 - resource_name
797- name: evented_pleg_connection_error_count
798 subsystem: kubelet
799 help: The number of errors encountered during the establishment of streaming connection
800 with the CRI runtime.
801 type: Counter
802 stabilityLevel: ALPHA
803- name: evented_pleg_connection_latency_seconds
804 subsystem: kubelet
805 help: The latency of streaming connection with the CRI runtime, measured in seconds.
806 type: Histogram
807 stabilityLevel: ALPHA
808 buckets:
809 - 0.005
810 - 0.01
811 - 0.025
812 - 0.05
813 - 0.1
814 - 0.25
815 - 0.5
816 - 1
817 - 2.5
818 - 5
819 - 10
820- name: evented_pleg_connection_success_count
821 subsystem: kubelet
822 help: The number of times a streaming client was obtained to receive CRI Events.
823 type: Counter
824 stabilityLevel: ALPHA
825- name: eviction_stats_age_seconds
826 subsystem: kubelet
827 help: Time between when stats are collected, and when pod is evicted based on those
828 stats by eviction signal
829 type: Histogram
830 stabilityLevel: ALPHA
831 labels:
832 - eviction_signal
833 buckets:
834 - 0.005
835 - 0.01
836 - 0.025
837 - 0.05
838 - 0.1
839 - 0.25
840 - 0.5
841 - 1
842 - 2.5
843 - 5
844 - 10
845- name: evictions
846 subsystem: kubelet
847 help: Cumulative number of pod evictions by eviction signal
848 type: Counter
849 stabilityLevel: ALPHA
850 labels:
851 - eviction_signal
852- name: graceful_shutdown_end_time_seconds
853 subsystem: kubelet
854 help: Last graceful shutdown start time since unix epoch in seconds
855 type: Gauge
856 stabilityLevel: ALPHA
857- name: graceful_shutdown_start_time_seconds
858 subsystem: kubelet
859 help: Last graceful shutdown start time since unix epoch in seconds
860 type: Gauge
861 stabilityLevel: ALPHA
862- name: image_garbage_collected_total
863 subsystem: kubelet
864 help: Total number of images garbage collected by the kubelet, whether through disk
865 usage or image age.
866 type: Counter
867 stabilityLevel: ALPHA
868- name: lifecycle_handler_http_fallbacks_total
869 subsystem: kubelet
870 help: The number of times lifecycle handlers successfully fell back to http from
871 https.
872 type: Counter
873 stabilityLevel: ALPHA
874- name: managed_ephemeral_containers
875 subsystem: kubelet
876 help: Current number of ephemeral containers in pods managed by this kubelet.
877 type: Gauge
878 stabilityLevel: ALPHA
879- name: mirror_pods
880 subsystem: kubelet
881 help: The number of mirror pods the kubelet will try to create (one per admitted
882 static pod)
883 type: Gauge
884 stabilityLevel: ALPHA
885- name: node_name
886 subsystem: kubelet
887 help: The node's name. The count is always 1.
888 type: Gauge
889 stabilityLevel: ALPHA
890 labels:
891 - node
892- name: node_startup_duration_seconds
893 subsystem: kubelet
894 help: Duration in seconds of node startup in total.
895 type: Gauge
896 stabilityLevel: ALPHA
897- name: node_startup_post_registration_duration_seconds
898 subsystem: kubelet
899 help: Duration in seconds of node startup after registration.
900 type: Gauge
901 stabilityLevel: ALPHA
902- name: node_startup_pre_kubelet_duration_seconds
903 subsystem: kubelet
904 help: Duration in seconds of node startup before kubelet starts.
905 type: Gauge
906 stabilityLevel: ALPHA
907- name: node_startup_pre_registration_duration_seconds
908 subsystem: kubelet
909 help: Duration in seconds of node startup before registration.
910 type: Gauge
911 stabilityLevel: ALPHA
912- name: node_startup_registration_duration_seconds
913 subsystem: kubelet
914 help: Duration in seconds of node startup during registration.
915 type: Gauge
916 stabilityLevel: ALPHA
917- name: orphan_pod_cleaned_volumes
918 subsystem: kubelet
919 help: The total number of orphaned Pods whose volumes were cleaned in the last periodic
920 sweep.
921 type: Gauge
922 stabilityLevel: ALPHA
923- name: orphan_pod_cleaned_volumes_errors
924 subsystem: kubelet
925 help: The number of orphaned Pods whose volumes failed to be cleaned in the last
926 periodic sweep.
927 type: Gauge
928 stabilityLevel: ALPHA
929- name: orphaned_runtime_pods_total
930 subsystem: kubelet
931 help: Number of pods that have been detected in the container runtime without being
932 already known to the pod worker. This typically indicates the kubelet was restarted
933 while a pod was force deleted in the API or in the local configuration, which
934 is unusual.
935 type: Counter
936 stabilityLevel: ALPHA
937- name: pleg_discard_events
938 subsystem: kubelet
939 help: The number of discard events in PLEG.
940 type: Counter
941 stabilityLevel: ALPHA
942- name: pleg_last_seen_seconds
943 subsystem: kubelet
944 help: Timestamp in seconds when PLEG was last seen active.
945 type: Gauge
946 stabilityLevel: ALPHA
947- name: pleg_relist_duration_seconds
948 subsystem: kubelet
949 help: Duration in seconds for relisting pods in PLEG.
950 type: Histogram
951 stabilityLevel: ALPHA
952 buckets:
953 - 0.005
954 - 0.01
955 - 0.025
956 - 0.05
957 - 0.1
958 - 0.25
959 - 0.5
960 - 1
961 - 2.5
962 - 5
963 - 10
964- name: pleg_relist_interval_seconds
965 subsystem: kubelet
966 help: Interval in seconds between relisting in PLEG.
967 type: Histogram
968 stabilityLevel: ALPHA
969 buckets:
970 - 0.005
971 - 0.01
972 - 0.025
973 - 0.05
974 - 0.1
975 - 0.25
976 - 0.5
977 - 1
978 - 2.5
979 - 5
980 - 10
981- name: pod_resources_endpoint_errors_get
982 subsystem: kubelet
983 help: Number of requests to the PodResource Get endpoint which returned error. Broken
984 down by server api version.
985 type: Counter
986 stabilityLevel: ALPHA
987 labels:
988 - server_api_version
989- name: pod_resources_endpoint_errors_get_allocatable
990 subsystem: kubelet
991 help: Number of requests to the PodResource GetAllocatableResources endpoint which
992 returned error. Broken down by server api version.
993 type: Counter
994 stabilityLevel: ALPHA
995 labels:
996 - server_api_version
997- name: pod_resources_endpoint_errors_list
998 subsystem: kubelet
999 help: Number of requests to the PodResource List endpoint which returned error.
1000 Broken down by server api version.
1001 type: Counter
1002 stabilityLevel: ALPHA
1003 labels:
1004 - server_api_version
1005- name: pod_resources_endpoint_requests_get
1006 subsystem: kubelet
1007 help: Number of requests to the PodResource Get endpoint. Broken down by server
1008 api version.
1009 type: Counter
1010 stabilityLevel: ALPHA
1011 labels:
1012 - server_api_version
1013- name: pod_resources_endpoint_requests_get_allocatable
1014 subsystem: kubelet
1015 help: Number of requests to the PodResource GetAllocatableResources endpoint. Broken
1016 down by server api version.
1017 type: Counter
1018 stabilityLevel: ALPHA
1019 labels:
1020 - server_api_version
1021- name: pod_resources_endpoint_requests_list
1022 subsystem: kubelet
1023 help: Number of requests to the PodResource List endpoint. Broken down by server
1024 api version.
1025 type: Counter
1026 stabilityLevel: ALPHA
1027 labels:
1028 - server_api_version
1029- name: pod_resources_endpoint_requests_total
1030 subsystem: kubelet
1031 help: Cumulative number of requests to the PodResource endpoint. Broken down by
1032 server api version.
1033 type: Counter
1034 stabilityLevel: ALPHA
1035 labels:
1036 - server_api_version
1037- name: pod_start_duration_seconds
1038 subsystem: kubelet
1039 help: Duration in seconds from kubelet seeing a pod for the first time to the pod
1040 starting to run
1041 type: Histogram
1042 stabilityLevel: ALPHA
1043 buckets:
1044 - 0.5
1045 - 1
1046 - 2
1047 - 3
1048 - 4
1049 - 5
1050 - 6
1051 - 8
1052 - 10
1053 - 20
1054 - 30
1055 - 45
1056 - 60
1057 - 120
1058 - 180
1059 - 240
1060 - 300
1061 - 360
1062 - 480
1063 - 600
1064 - 900
1065 - 1200
1066 - 1800
1067 - 2700
1068 - 3600
1069- name: pod_start_sli_duration_seconds
1070 subsystem: kubelet
1071 help: Duration in seconds to start a pod, excluding time to pull images and run
1072 init containers, measured from pod creation timestamp to when all its containers
1073 are reported as started and observed via watch
1074 type: Histogram
1075 stabilityLevel: ALPHA
1076 buckets:
1077 - 0.5
1078 - 1
1079 - 2
1080 - 3
1081 - 4
1082 - 5
1083 - 6
1084 - 8
1085 - 10
1086 - 20
1087 - 30
1088 - 45
1089 - 60
1090 - 120
1091 - 180
1092 - 240
1093 - 300
1094 - 360
1095 - 480
1096 - 600
1097 - 900
1098 - 1200
1099 - 1800
1100 - 2700
1101 - 3600
1102- name: pod_start_total_duration_seconds
1103 subsystem: kubelet
1104 help: Duration in seconds to start a pod since creation, including time to pull
1105 images and run init containers, measured from pod creation timestamp to when all
1106 its containers are reported as started and observed via watch
1107 type: Histogram
1108 stabilityLevel: ALPHA
1109 buckets:
1110 - 0.5
1111 - 1
1112 - 2
1113 - 3
1114 - 4
1115 - 5
1116 - 6
1117 - 8
1118 - 10
1119 - 20
1120 - 30
1121 - 45
1122 - 60
1123 - 120
1124 - 180
1125 - 240
1126 - 300
1127 - 360
1128 - 480
1129 - 600
1130 - 900
1131 - 1200
1132 - 1800
1133 - 2700
1134 - 3600
1135- name: pod_status_sync_duration_seconds
1136 subsystem: kubelet
1137 help: Duration in seconds to sync a pod status update. Measures time from detection
1138 of a change to pod status until the API is successfully updated for that pod,
1139 even if multiple intevening changes to pod status occur.
1140 type: Histogram
1141 stabilityLevel: ALPHA
1142 buckets:
1143 - 0.01
1144 - 0.05
1145 - 0.1
1146 - 0.5
1147 - 1
1148 - 5
1149 - 10
1150 - 20
1151 - 30
1152 - 45
1153 - 60
1154- name: pod_worker_duration_seconds
1155 subsystem: kubelet
1156 help: 'Duration in seconds to sync a single pod. Broken down by operation type:
1157 create, update, or sync'
1158 type: Histogram
1159 stabilityLevel: ALPHA
1160 labels:
1161 - operation_type
1162 buckets:
1163 - 0.005
1164 - 0.01
1165 - 0.025
1166 - 0.05
1167 - 0.1
1168 - 0.25
1169 - 0.5
1170 - 1
1171 - 2.5
1172 - 5
1173 - 10
1174- name: pod_worker_start_duration_seconds
1175 subsystem: kubelet
1176 help: Duration in seconds from kubelet seeing a pod to starting a worker.
1177 type: Histogram
1178 stabilityLevel: ALPHA
1179 buckets:
1180 - 0.005
1181 - 0.01
1182 - 0.025
1183 - 0.05
1184 - 0.1
1185 - 0.25
1186 - 0.5
1187 - 1
1188 - 2.5
1189 - 5
1190 - 10
1191- name: preemptions
1192 subsystem: kubelet
1193 help: Cumulative number of pod preemptions by preemption resource
1194 type: Counter
1195 stabilityLevel: ALPHA
1196 labels:
1197 - preemption_signal
1198- name: restarted_pods_total
1199 subsystem: kubelet
1200 help: Number of pods that have been restarted because they were deleted and recreated
1201 with the same UID while the kubelet was watching them (common for static pods,
1202 extremely uncommon for API pods)
1203 type: Counter
1204 stabilityLevel: ALPHA
1205 labels:
1206 - static
1207- name: run_podsandbox_duration_seconds
1208 subsystem: kubelet
1209 help: Duration in seconds of the run_podsandbox operations. Broken down by RuntimeClass.Handler.
1210 type: Histogram
1211 stabilityLevel: ALPHA
1212 labels:
1213 - runtime_handler
1214 buckets:
1215 - 0.005
1216 - 0.01
1217 - 0.025
1218 - 0.05
1219 - 0.1
1220 - 0.25
1221 - 0.5
1222 - 1
1223 - 2.5
1224 - 5
1225 - 10
1226- name: run_podsandbox_errors_total
1227 subsystem: kubelet
1228 help: Cumulative number of the run_podsandbox operation errors by RuntimeClass.Handler.
1229 type: Counter
1230 stabilityLevel: ALPHA
1231 labels:
1232 - runtime_handler
1233- name: running_containers
1234 subsystem: kubelet
1235 help: Number of containers currently running
1236 type: Gauge
1237 stabilityLevel: ALPHA
1238 labels:
1239 - container_state
1240- name: running_pods
1241 subsystem: kubelet
1242 help: Number of pods that have a running pod sandbox
1243 type: Gauge
1244 stabilityLevel: ALPHA
1245- name: runtime_operations_duration_seconds
1246 subsystem: kubelet
1247 help: Duration in seconds of runtime operations. Broken down by operation type.
1248 type: Histogram
1249 stabilityLevel: ALPHA
1250 labels:
1251 - operation_type
1252 buckets:
1253 - 0.005
1254 - 0.0125
1255 - 0.03125
1256 - 0.078125
1257 - 0.1953125
1258 - 0.48828125
1259 - 1.220703125
1260 - 3.0517578125
1261 - 7.62939453125
1262 - 19.073486328125
1263 - 47.6837158203125
1264 - 119.20928955078125
1265 - 298.0232238769531
1266 - 745.0580596923828
1267- name: runtime_operations_errors_total
1268 subsystem: kubelet
1269 help: Cumulative number of runtime operation errors by operation type.
1270 type: Counter
1271 stabilityLevel: ALPHA
1272 labels:
1273 - operation_type
1274- name: runtime_operations_total
1275 subsystem: kubelet
1276 help: Cumulative number of runtime operations by operation type.
1277 type: Counter
1278 stabilityLevel: ALPHA
1279 labels:
1280 - operation_type
1281- name: started_containers_errors_total
1282 subsystem: kubelet
1283 help: Cumulative number of errors when starting containers
1284 type: Counter
1285 stabilityLevel: ALPHA
1286 labels:
1287 - code
1288 - container_type
1289- name: started_containers_total
1290 subsystem: kubelet
1291 help: Cumulative number of containers started
1292 type: Counter
1293 stabilityLevel: ALPHA
1294 labels:
1295 - container_type
1296- name: started_host_process_containers_errors_total
1297 subsystem: kubelet
1298 help: Cumulative number of errors when starting hostprocess containers. This metric
1299 will only be collected on Windows.
1300 type: Counter
1301 stabilityLevel: ALPHA
1302 labels:
1303 - code
1304 - container_type
1305- name: started_host_process_containers_total
1306 subsystem: kubelet
1307 help: Cumulative number of hostprocess containers started. This metric will only
1308 be collected on Windows.
1309 type: Counter
1310 stabilityLevel: ALPHA
1311 labels:
1312 - container_type
1313- name: started_pods_errors_total
1314 subsystem: kubelet
1315 help: Cumulative number of errors when starting pods
1316 type: Counter
1317 stabilityLevel: ALPHA
1318- name: started_pods_total
1319 subsystem: kubelet
1320 help: Cumulative number of pods started
1321 type: Counter
1322 stabilityLevel: ALPHA
1323- name: topology_manager_admission_duration_ms
1324 subsystem: kubelet
1325 help: Duration in milliseconds to serve a pod admission request.
1326 type: Histogram
1327 stabilityLevel: ALPHA
1328 buckets:
1329 - 0.05
1330 - 0.1
1331 - 0.2
1332 - 0.4
1333 - 0.8
1334 - 1.6
1335 - 3.2
1336 - 6.4
1337 - 12.8
1338 - 25.6
1339 - 51.2
1340 - 102.4
1341 - 204.8
1342 - 409.6
1343 - 819.2
1344- name: topology_manager_admission_errors_total
1345 subsystem: kubelet
1346 help: The number of admission request failures where resources could not be aligned.
1347 type: Counter
1348 stabilityLevel: ALPHA
1349- name: topology_manager_admission_requests_total
1350 subsystem: kubelet
1351 help: The number of admission requests where resources have to be aligned.
1352 type: Counter
1353 stabilityLevel: ALPHA
1354- name: kubelet_volume_stats_available_bytes
1355 help: Number of available bytes in the volume
1356 type: Custom
1357 stabilityLevel: ALPHA
1358 labels:
1359 - namespace
1360 - persistentvolumeclaim
1361- name: kubelet_volume_stats_capacity_bytes
1362 help: Capacity in bytes of the volume
1363 type: Custom
1364 stabilityLevel: ALPHA
1365 labels:
1366 - namespace
1367 - persistentvolumeclaim
1368- name: kubelet_volume_stats_health_status_abnormal
1369 help: Abnormal volume health status. The count is either 1 or 0. 1 indicates the
1370 volume is unhealthy, 0 indicates volume is healthy
1371 type: Custom
1372 stabilityLevel: ALPHA
1373 labels:
1374 - namespace
1375 - persistentvolumeclaim
1376- name: kubelet_volume_stats_inodes
1377 help: Maximum number of inodes in the volume
1378 type: Custom
1379 stabilityLevel: ALPHA
1380 labels:
1381 - namespace
1382 - persistentvolumeclaim
1383- name: kubelet_volume_stats_inodes_free
1384 help: Number of free inodes in the volume
1385 type: Custom
1386 stabilityLevel: ALPHA
1387 labels:
1388 - namespace
1389 - persistentvolumeclaim
1390- name: kubelet_volume_stats_inodes_used
1391 help: Number of used inodes in the volume
1392 type: Custom
1393 stabilityLevel: ALPHA
1394 labels:
1395 - namespace
1396 - persistentvolumeclaim
1397- name: kubelet_volume_stats_used_bytes
1398 help: Number of used bytes in the volume
1399 type: Custom
1400 stabilityLevel: ALPHA
1401 labels:
1402 - namespace
1403 - persistentvolumeclaim
1404- name: working_pods
1405 subsystem: kubelet
1406 help: Number of pods the kubelet is actually running, broken down by lifecycle phase,
1407 whether the pod is desired, orphaned, or runtime only (also orphaned), and whether
1408 the pod is static. An orphaned pod has been removed from local configuration or
1409 force deleted in the API and consumes resources that are not otherwise visible.
1410 type: Gauge
1411 stabilityLevel: ALPHA
1412 labels:
1413 - config
1414 - lifecycle
1415 - static
1416- name: node_swap_usage_bytes
1417 help: Current swap usage of the node in bytes. Reported only on non-windows systems
1418 type: Custom
1419 stabilityLevel: ALPHA
1420- name: pod_swap_usage_bytes
1421 help: Current amount of the pod swap usage in bytes. Reported only on non-windows
1422 systems
1423 type: Custom
1424 stabilityLevel: ALPHA
1425 labels:
1426 - pod
1427 - namespace
1428- name: scrape_error
1429 help: 1 if there was an error while getting container metrics, 0 otherwise
1430 type: Custom
1431 deprecatedVersion: 1.29.0
1432 stabilityLevel: ALPHA
1433- name: container_cpu_usage_seconds_total
1434 help: Cumulative cpu time consumed by the container in core-seconds
1435 type: Custom
1436 stabilityLevel: STABLE
1437 labels:
1438 - container
1439 - pod
1440 - namespace
1441- name: container_memory_working_set_bytes
1442 help: Current working set of the container in bytes
1443 type: Custom
1444 stabilityLevel: STABLE
1445 labels:
1446 - container
1447 - pod
1448 - namespace
1449- name: container_start_time_seconds
1450 help: Start time of the container since unix epoch in seconds
1451 type: Custom
1452 stabilityLevel: STABLE
1453 labels:
1454 - container
1455 - pod
1456 - namespace
1457- name: node_cpu_usage_seconds_total
1458 help: Cumulative cpu time consumed by the node in core-seconds
1459 type: Custom
1460 stabilityLevel: STABLE
1461- name: node_memory_working_set_bytes
1462 help: Current working set of the node in bytes
1463 type: Custom
1464 stabilityLevel: STABLE
1465- name: pod_cpu_usage_seconds_total
1466 help: Cumulative cpu time consumed by the pod in core-seconds
1467 type: Custom
1468 stabilityLevel: STABLE
1469 labels:
1470 - pod
1471 - namespace
1472- name: pod_memory_working_set_bytes
1473 help: Current working set of the pod in bytes
1474 type: Custom
1475 stabilityLevel: STABLE
1476 labels:
1477 - pod
1478 - namespace
1479- name: resource_scrape_error
1480 help: 1 if there was an error while getting container metrics, 0 otherwise
1481 type: Custom
1482 stabilityLevel: STABLE
1483- name: force_cleaned_failed_volume_operation_errors_total
1484 help: The number of volumes that failed force cleanup after their reconstruction
1485 failed during kubelet startup.
1486 type: Counter
1487 stabilityLevel: ALPHA
1488- name: force_cleaned_failed_volume_operations_total
1489 help: The number of volumes that were force cleaned after their reconstruction failed
1490 during kubelet startup. This includes both successful and failed cleanups.
1491 type: Counter
1492 stabilityLevel: ALPHA
1493- name: http_inflight_requests
1494 subsystem: kubelet
1495 help: Number of the inflight http requests
1496 type: Gauge
1497 stabilityLevel: ALPHA
1498 labels:
1499 - long_running
1500 - method
1501 - path
1502 - server_type
1503- name: http_requests_duration_seconds
1504 subsystem: kubelet
1505 help: Duration in seconds to serve http requests
1506 type: Histogram
1507 stabilityLevel: ALPHA
1508 labels:
1509 - long_running
1510 - method
1511 - path
1512 - server_type
1513 buckets:
1514 - 0.005
1515 - 0.01
1516 - 0.025
1517 - 0.05
1518 - 0.1
1519 - 0.25
1520 - 0.5
1521 - 1
1522 - 2.5
1523 - 5
1524 - 10
1525- name: http_requests_total
1526 subsystem: kubelet
1527 help: Number of the http requests received since the server started
1528 type: Counter
1529 stabilityLevel: ALPHA
1530 labels:
1531 - long_running
1532 - method
1533 - path
1534 - server_type
1535- name: volume_metric_collection_duration_seconds
1536 subsystem: kubelet
1537 help: Duration in seconds to calculate volume stats
1538 type: Histogram
1539 stabilityLevel: ALPHA
1540 labels:
1541 - metric_source
1542 buckets:
1543 - 0.005
1544 - 0.01
1545 - 0.025
1546 - 0.05
1547 - 0.1
1548 - 0.25
1549 - 0.5
1550 - 1
1551 - 2.5
1552 - 5
1553 - 10
1554- name: network_programming_duration_seconds
1555 subsystem: kubeproxy
1556 help: In Cluster Network Programming Latency in seconds
1557 type: Histogram
1558 stabilityLevel: ALPHA
1559 buckets:
1560 - 0.25
1561 - 0.5
1562 - 1
1563 - 2
1564 - 3
1565 - 4
1566 - 5
1567 - 6
1568 - 7
1569 - 8
1570 - 9
1571 - 10
1572 - 11
1573 - 12
1574 - 13
1575 - 14
1576 - 15
1577 - 16
1578 - 17
1579 - 18
1580 - 19
1581 - 20
1582 - 21
1583 - 22
1584 - 23
1585 - 24
1586 - 25
1587 - 26
1588 - 27
1589 - 28
1590 - 29
1591 - 30
1592 - 31
1593 - 32
1594 - 33
1595 - 34
1596 - 35
1597 - 36
1598 - 37
1599 - 38
1600 - 39
1601 - 40
1602 - 41
1603 - 42
1604 - 43
1605 - 44
1606 - 45
1607 - 46
1608 - 47
1609 - 48
1610 - 49
1611 - 50
1612 - 51
1613 - 52
1614 - 53
1615 - 54
1616 - 55
1617 - 56
1618 - 57
1619 - 58
1620 - 59
1621 - 60
1622 - 65
1623 - 70
1624 - 75
1625 - 80
1626 - 85
1627 - 90
1628 - 95
1629 - 100
1630 - 105
1631 - 110
1632 - 115
1633 - 120
1634 - 150
1635 - 180
1636 - 210
1637 - 240
1638 - 270
1639 - 300
1640- name: proxy_healthz_total
1641 subsystem: kubeproxy
1642 help: Cumulative proxy healthz HTTP status
1643 type: Counter
1644 stabilityLevel: ALPHA
1645 labels:
1646 - code
1647- name: proxy_livez_total
1648 subsystem: kubeproxy
1649 help: Cumulative proxy livez HTTP status
1650 type: Counter
1651 stabilityLevel: ALPHA
1652 labels:
1653 - code
1654- name: sync_full_proxy_rules_duration_seconds
1655 subsystem: kubeproxy
1656 help: SyncProxyRules latency in seconds for full resyncs
1657 type: Histogram
1658 stabilityLevel: ALPHA
1659 buckets:
1660 - 0.001
1661 - 0.002
1662 - 0.004
1663 - 0.008
1664 - 0.016
1665 - 0.032
1666 - 0.064
1667 - 0.128
1668 - 0.256
1669 - 0.512
1670 - 1.024
1671 - 2.048
1672 - 4.096
1673 - 8.192
1674 - 16.384
1675- name: sync_partial_proxy_rules_duration_seconds
1676 subsystem: kubeproxy
1677 help: SyncProxyRules latency in seconds for partial resyncs
1678 type: Histogram
1679 stabilityLevel: ALPHA
1680 buckets:
1681 - 0.001
1682 - 0.002
1683 - 0.004
1684 - 0.008
1685 - 0.016
1686 - 0.032
1687 - 0.064
1688 - 0.128
1689 - 0.256
1690 - 0.512
1691 - 1.024
1692 - 2.048
1693 - 4.096
1694 - 8.192
1695 - 16.384
1696- name: sync_proxy_rules_duration_seconds
1697 subsystem: kubeproxy
1698 help: SyncProxyRules latency in seconds
1699 type: Histogram
1700 stabilityLevel: ALPHA
1701 buckets:
1702 - 0.001
1703 - 0.002
1704 - 0.004
1705 - 0.008
1706 - 0.016
1707 - 0.032
1708 - 0.064
1709 - 0.128
1710 - 0.256
1711 - 0.512
1712 - 1.024
1713 - 2.048
1714 - 4.096
1715 - 8.192
1716 - 16.384
1717- name: sync_proxy_rules_endpoint_changes_pending
1718 subsystem: kubeproxy
1719 help: Pending proxy rules Endpoint changes
1720 type: Gauge
1721 stabilityLevel: ALPHA
1722- name: sync_proxy_rules_endpoint_changes_total
1723 subsystem: kubeproxy
1724 help: Cumulative proxy rules Endpoint changes
1725 type: Counter
1726 stabilityLevel: ALPHA
1727- name: sync_proxy_rules_iptables_last
1728 subsystem: kubeproxy
1729 help: Number of iptables rules written by kube-proxy in last sync
1730 type: Gauge
1731 stabilityLevel: ALPHA
1732 labels:
1733 - table
1734- name: sync_proxy_rules_iptables_partial_restore_failures_total
1735 subsystem: kubeproxy
1736 help: Cumulative proxy iptables partial restore failures
1737 type: Counter
1738 stabilityLevel: ALPHA
1739- name: sync_proxy_rules_iptables_restore_failures_total
1740 subsystem: kubeproxy
1741 help: Cumulative proxy iptables restore failures
1742 type: Counter
1743 stabilityLevel: ALPHA
1744- name: sync_proxy_rules_iptables_total
1745 subsystem: kubeproxy
1746 help: Total number of iptables rules owned by kube-proxy
1747 type: Gauge
1748 stabilityLevel: ALPHA
1749 labels:
1750 - table
1751- name: sync_proxy_rules_last_queued_timestamp_seconds
1752 subsystem: kubeproxy
1753 help: The last time a sync of proxy rules was queued
1754 type: Gauge
1755 stabilityLevel: ALPHA
1756- name: sync_proxy_rules_last_timestamp_seconds
1757 subsystem: kubeproxy
1758 help: The last time proxy rules were successfully synced
1759 type: Gauge
1760 stabilityLevel: ALPHA
1761- name: sync_proxy_rules_no_local_endpoints_total
1762 subsystem: kubeproxy
1763 help: Number of services with a Local traffic policy and no endpoints
1764 type: Gauge
1765 stabilityLevel: ALPHA
1766 labels:
1767 - traffic_policy
1768- name: sync_proxy_rules_service_changes_pending
1769 subsystem: kubeproxy
1770 help: Pending proxy rules Service changes
1771 type: Gauge
1772 stabilityLevel: ALPHA
1773- name: sync_proxy_rules_service_changes_total
1774 subsystem: kubeproxy
1775 help: Cumulative proxy rules Service changes
1776 type: Counter
1777 stabilityLevel: ALPHA
1778- name: plugin_manager_total_plugins
1779 help: Number of plugins in Plugin Manager
1780 type: Custom
1781 stabilityLevel: ALPHA
1782 labels:
1783 - socket_path
1784 - state
1785- name: probe_duration_seconds
1786 subsystem: prober
1787 help: Duration in seconds for a probe response.
1788 type: Histogram
1789 stabilityLevel: ALPHA
1790 labels:
1791 - container
1792 - namespace
1793 - pod
1794 - probe_type
1795- name: probe_total
1796 subsystem: prober
1797 help: Cumulative number of a liveness, readiness or startup probe for a container
1798 by result.
1799 type: Counter
1800 stabilityLevel: ALPHA
1801 labels:
1802 - container
1803 - namespace
1804 - pod
1805 - pod_uid
1806 - probe_type
1807 - result
1808- name: reconstruct_volume_operations_errors_total
1809 help: The number of volumes that failed reconstruction from the operating system
1810 during kubelet startup.
1811 type: Counter
1812 stabilityLevel: ALPHA
1813- name: reconstruct_volume_operations_total
1814 help: The number of volumes that were attempted to be reconstructed from the operating
1815 system during kubelet startup. This includes both successful and failed reconstruction.
1816 type: Counter
1817 stabilityLevel: ALPHA
1818- name: volume_manager_selinux_container_errors_total
1819 help: Number of errors when kubelet cannot compute SELinux context for a container.
1820 Kubelet can't start such a Pod then and it will retry, therefore value of this
1821 metric may not represent the actual nr. of containers.
1822 type: Gauge
1823 stabilityLevel: ALPHA
1824- name: volume_manager_selinux_container_warnings_total
1825 help: Number of errors when kubelet cannot compute SELinux context for a container
1826 that are ignored. They will become real errors when SELinuxMountReadWriteOncePod
1827 feature is expanded to all volume access modes.
1828 type: Gauge
1829 stabilityLevel: ALPHA
1830- name: volume_manager_selinux_pod_context_mismatch_errors_total
1831 help: Number of errors when a Pod defines different SELinux contexts for its containers
1832 that use the same volume. Kubelet can't start such a Pod then and it will retry,
1833 therefore value of this metric may not represent the actual nr. of Pods.
1834 type: Gauge
1835 stabilityLevel: ALPHA
1836- name: volume_manager_selinux_pod_context_mismatch_warnings_total
1837 help: Number of errors when a Pod defines different SELinux contexts for its containers
1838 that use the same volume. They are not errors yet, but they will become real errors
1839 when SELinuxMountReadWriteOncePod feature is expanded to all volume access modes.
1840 type: Gauge
1841 stabilityLevel: ALPHA
1842- name: volume_manager_selinux_volume_context_mismatch_errors_total
1843 help: Number of errors when a Pod uses a volume that is already mounted with a different
1844 SELinux context than the Pod needs. Kubelet can't start such a Pod then and it
1845 will retry, therefore value of this metric may not represent the actual nr. of
1846 Pods.
1847 type: Gauge
1848 stabilityLevel: ALPHA
1849 labels:
1850 - volume_plugin
1851- name: volume_manager_selinux_volume_context_mismatch_warnings_total
1852 help: Number of errors when a Pod uses a volume that is already mounted with a different
1853 SELinux context than the Pod needs. They are not errors yet, but they will become
1854 real errors when SELinuxMountReadWriteOncePod feature is expanded to all volume
1855 access modes.
1856 type: Gauge
1857 stabilityLevel: ALPHA
1858 labels:
1859 - volume_plugin
1860- name: volume_manager_selinux_volumes_admitted_total
1861 help: Number of volumes whose SELinux context was fine and will be mounted with
1862 mount -o context option.
1863 type: Gauge
1864 stabilityLevel: ALPHA
1865 labels:
1866 - volume_plugin
1867- name: volume_manager_total_volumes
1868 help: Number of volumes in Volume Manager
1869 type: Custom
1870 stabilityLevel: ALPHA
1871 labels:
1872 - plugin_name
1873 - state
1874- name: csr_honored_duration_total
1875 subsystem: certificates_registry
1876 namespace: apiserver
1877 help: Total number of issued CSRs with a requested duration that was honored, sliced
1878 by signer (only kubernetes.io signer names are specifically identified)
1879 type: Counter
1880 stabilityLevel: ALPHA
1881 labels:
1882 - signerName
1883- name: csr_requested_duration_total
1884 subsystem: certificates_registry
1885 namespace: apiserver
1886 help: Total number of issued CSRs with a requested duration, sliced by signer (only
1887 kubernetes.io signer names are specifically identified)
1888 type: Counter
1889 stabilityLevel: ALPHA
1890 labels:
1891 - signerName
1892- name: ip_errors_total
1893 subsystem: clusterip_repair
1894 namespace: apiserver
1895 help: 'Number of errors detected on clusterips by the repair loop broken down by
1896 type of error: leak, repair, full, outOfRange, duplicate, unknown, invalid'
1897 type: Counter
1898 stabilityLevel: ALPHA
1899 labels:
1900 - type
1901- name: reconcile_errors_total
1902 subsystem: clusterip_repair
1903 namespace: apiserver
1904 help: Number of reconciliation failures on the clusterip repair reconcile loop
1905 type: Counter
1906 stabilityLevel: ALPHA
1907- name: port_errors_total
1908 subsystem: nodeport_repair
1909 namespace: apiserver
1910 help: 'Number of errors detected on ports by the repair loop broken down by type
1911 of error: leak, repair, full, outOfRange, duplicate, unknown'
1912 type: Counter
1913 stabilityLevel: ALPHA
1914 labels:
1915 - type
1916- name: allocated_ips
1917 subsystem: clusterip_allocator
1918 namespace: kube_apiserver
1919 help: Gauge measuring the number of allocated IPs for Services
1920 type: Gauge
1921 stabilityLevel: ALPHA
1922 labels:
1923 - cidr
1924- name: allocation_errors_total
1925 subsystem: clusterip_allocator
1926 namespace: kube_apiserver
1927 help: Number of errors trying to allocate Cluster IPs
1928 type: Counter
1929 stabilityLevel: ALPHA
1930 labels:
1931 - cidr
1932 - scope
1933- name: allocation_total
1934 subsystem: clusterip_allocator
1935 namespace: kube_apiserver
1936 help: Number of Cluster IPs allocations
1937 type: Counter
1938 stabilityLevel: ALPHA
1939 labels:
1940 - cidr
1941 - scope
1942- name: available_ips
1943 subsystem: clusterip_allocator
1944 namespace: kube_apiserver
1945 help: Gauge measuring the number of available IPs for Services
1946 type: Gauge
1947 stabilityLevel: ALPHA
1948 labels:
1949 - cidr
1950- name: allocated_ports
1951 subsystem: nodeport_allocator
1952 namespace: kube_apiserver
1953 help: Gauge measuring the number of allocated NodePorts for Services
1954 type: Gauge
1955 stabilityLevel: ALPHA
1956- name: available_ports
1957 subsystem: nodeport_allocator
1958 namespace: kube_apiserver
1959 help: Gauge measuring the number of available NodePorts for Services
1960 type: Gauge
1961 stabilityLevel: ALPHA
1962- name: backend_tls_failure_total
1963 subsystem: pod_logs
1964 namespace: kube_apiserver
1965 help: Total number of requests for pods/logs that failed due to kubelet server TLS
1966 verification
1967 type: Counter
1968 stabilityLevel: ALPHA
1969- name: insecure_backend_total
1970 subsystem: pod_logs
1971 namespace: kube_apiserver
1972 help: 'Total number of requests for pods/logs sliced by usage type: enforce_tls,
1973 skip_tls_allowed, skip_tls_denied'
1974 type: Counter
1975 stabilityLevel: ALPHA
1976 labels:
1977 - usage
1978- name: pods_logs_backend_tls_failure_total
1979 subsystem: pod_logs
1980 namespace: kube_apiserver
1981 help: Total number of requests for pods/logs that failed due to kubelet server TLS
1982 verification
1983 type: Counter
1984 deprecatedVersion: 1.27.0
1985 stabilityLevel: ALPHA
1986- name: pods_logs_insecure_backend_total
1987 subsystem: pod_logs
1988 namespace: kube_apiserver
1989 help: 'Total number of requests for pods/logs sliced by usage type: enforce_tls,
1990 skip_tls_allowed, skip_tls_denied'
1991 type: Counter
1992 deprecatedVersion: 1.27.0
1993 stabilityLevel: ALPHA
1994 labels:
1995 - usage
1996- name: goroutines
1997 subsystem: scheduler
1998 help: Number of running goroutines split by the work they do such as binding.
1999 type: Gauge
2000 stabilityLevel: ALPHA
2001 labels:
2002 - operation
2003- name: permit_wait_duration_seconds
2004 subsystem: scheduler
2005 help: Duration of waiting on permit.
2006 type: Histogram
2007 stabilityLevel: ALPHA
2008 labels:
2009 - result
2010 buckets:
2011 - 0.001
2012 - 0.002
2013 - 0.004
2014 - 0.008
2015 - 0.016
2016 - 0.032
2017 - 0.064
2018 - 0.128
2019 - 0.256
2020 - 0.512
2021 - 1.024
2022 - 2.048
2023 - 4.096
2024 - 8.192
2025 - 16.384
2026- name: plugin_evaluation_total
2027 subsystem: scheduler
2028 help: Number of attempts to schedule pods by each plugin and the extension point
2029 (available only in PreFilter and Filter.).
2030 type: Counter
2031 stabilityLevel: ALPHA
2032 labels:
2033 - extension_point
2034 - plugin
2035 - profile
2036- name: plugin_execution_duration_seconds
2037 subsystem: scheduler
2038 help: Duration for running a plugin at a specific extension point.
2039 type: Histogram
2040 stabilityLevel: ALPHA
2041 labels:
2042 - extension_point
2043 - plugin
2044 - status
2045 buckets:
2046 - 1e-05
2047 - 1.5000000000000002e-05
2048 - 2.2500000000000005e-05
2049 - 3.375000000000001e-05
2050 - 5.062500000000001e-05
2051 - 7.593750000000002e-05
2052 - 0.00011390625000000003
2053 - 0.00017085937500000006
2054 - 0.0002562890625000001
2055 - 0.00038443359375000017
2056 - 0.0005766503906250003
2057 - 0.0008649755859375004
2058 - 0.0012974633789062506
2059 - 0.0019461950683593758
2060 - 0.0029192926025390638
2061 - 0.004378938903808595
2062 - 0.006568408355712893
2063 - 0.009852612533569338
2064 - 0.014778918800354007
2065 - 0.02216837820053101
2066- name: scheduler_cache_size
2067 subsystem: scheduler
2068 help: Number of nodes, pods, and assumed (bound) pods in the scheduler cache.
2069 type: Gauge
2070 stabilityLevel: ALPHA
2071 labels:
2072 - type
2073- name: scheduling_algorithm_duration_seconds
2074 subsystem: scheduler
2075 help: Scheduling algorithm latency in seconds
2076 type: Histogram
2077 stabilityLevel: ALPHA
2078 buckets:
2079 - 0.001
2080 - 0.002
2081 - 0.004
2082 - 0.008
2083 - 0.016
2084 - 0.032
2085 - 0.064
2086 - 0.128
2087 - 0.256
2088 - 0.512
2089 - 1.024
2090 - 2.048
2091 - 4.096
2092 - 8.192
2093 - 16.384
2094- name: unschedulable_pods
2095 subsystem: scheduler
2096 help: The number of unschedulable pods broken down by plugin name. A pod will increment
2097 the gauge for all plugins that caused it to not schedule and so this metric have
2098 meaning only when broken down by plugin.
2099 type: Gauge
2100 stabilityLevel: ALPHA
2101 labels:
2102 - plugin
2103 - profile
2104- name: binder_cache_requests_total
2105 subsystem: scheduler_volume
2106 help: Total number for request volume binding cache
2107 type: Counter
2108 stabilityLevel: ALPHA
2109 labels:
2110 - operation
2111- name: scheduling_stage_error_total
2112 subsystem: scheduler_volume
2113 help: Volume scheduling stage error count
2114 type: Counter
2115 stabilityLevel: ALPHA
2116 labels:
2117 - operation
2118- name: invalid_legacy_auto_token_uses_total
2119 subsystem: serviceaccount
2120 help: Cumulative invalid auto-generated legacy tokens used
2121 type: Counter
2122 stabilityLevel: ALPHA
2123- name: legacy_auto_token_uses_total
2124 subsystem: serviceaccount
2125 help: Cumulative auto-generated legacy tokens used
2126 type: Counter
2127 stabilityLevel: ALPHA
2128- name: legacy_manual_token_uses_total
2129 subsystem: serviceaccount
2130 help: Cumulative manually created legacy tokens used
2131 type: Counter
2132 stabilityLevel: ALPHA
2133- name: legacy_tokens_total
2134 subsystem: serviceaccount
2135 help: Cumulative legacy service account tokens used
2136 type: Counter
2137 stabilityLevel: ALPHA
2138- name: stale_tokens_total
2139 subsystem: serviceaccount
2140 help: Cumulative stale projected service account tokens used
2141 type: Counter
2142 stabilityLevel: ALPHA
2143- name: valid_tokens_total
2144 subsystem: serviceaccount
2145 help: Cumulative valid projected service account tokens used
2146 type: Counter
2147 stabilityLevel: ALPHA
2148- name: pod_scheduling_sli_duration_seconds
2149 subsystem: scheduler
2150 help: E2e latency for a pod being scheduled, from the time the pod enters the scheduling
2151 queue an d might involve multiple scheduling attempts.
2152 type: Histogram
2153 stabilityLevel: BETA
2154 labels:
2155 - attempts
2156 buckets:
2157 - 0.01
2158 - 0.02
2159 - 0.04
2160 - 0.08
2161 - 0.16
2162 - 0.32
2163 - 0.64
2164 - 1.28
2165 - 2.56
2166 - 5.12
2167 - 10.24
2168 - 20.48
2169 - 40.96
2170 - 81.92
2171 - 163.84
2172 - 327.68
2173 - 655.36
2174 - 1310.72
2175 - 2621.44
2176 - 5242.88
2177- name: kube_pod_resource_limit
2178 help: Resources limit for workloads on the cluster, broken down by pod. This shows
2179 the resource usage the scheduler and kubelet expect per pod for resources along
2180 with the unit for the resource if any.
2181 type: Custom
2182 stabilityLevel: STABLE
2183 labels:
2184 - namespace
2185 - pod
2186 - node
2187 - scheduler
2188 - priority
2189 - resource
2190 - unit
2191- name: kube_pod_resource_request
2192 help: Resources requested by workloads on the cluster, broken down by pod. This
2193 shows the resource usage the scheduler and kubelet expect per pod for resources
2194 along with the unit for the resource if any.
2195 type: Custom
2196 stabilityLevel: STABLE
2197 labels:
2198 - namespace
2199 - pod
2200 - node
2201 - scheduler
2202 - priority
2203 - resource
2204 - unit
2205- name: framework_extension_point_duration_seconds
2206 subsystem: scheduler
2207 help: Latency for running all plugins of a specific extension point.
2208 type: Histogram
2209 stabilityLevel: STABLE
2210 labels:
2211 - extension_point
2212 - profile
2213 - status
2214 buckets:
2215 - 0.0001
2216 - 0.0002
2217 - 0.0004
2218 - 0.0008
2219 - 0.0016
2220 - 0.0032
2221 - 0.0064
2222 - 0.0128
2223 - 0.0256
2224 - 0.0512
2225 - 0.1024
2226 - 0.2048
2227- name: pending_pods
2228 subsystem: scheduler
2229 help: Number of pending pods, by the queue type. 'active' means number of pods in
2230 activeQ; 'backoff' means number of pods in backoffQ; 'unschedulable' means number
2231 of pods in unschedulablePods that the scheduler attempted to schedule and failed;
2232 'gated' is the number of unschedulable pods that the scheduler never attempted
2233 to schedule because they are gated.
2234 type: Gauge
2235 stabilityLevel: STABLE
2236 labels:
2237 - queue
2238- name: pod_scheduling_attempts
2239 subsystem: scheduler
2240 help: Number of attempts to successfully schedule a pod.
2241 type: Histogram
2242 stabilityLevel: STABLE
2243 buckets:
2244 - 1
2245 - 2
2246 - 4
2247 - 8
2248 - 16
2249- name: pod_scheduling_duration_seconds
2250 subsystem: scheduler
2251 help: E2e latency for a pod being scheduled which may include multiple scheduling
2252 attempts.
2253 type: Histogram
2254 deprecatedVersion: 1.28.0
2255 stabilityLevel: STABLE
2256 labels:
2257 - attempts
2258 buckets:
2259 - 0.01
2260 - 0.02
2261 - 0.04
2262 - 0.08
2263 - 0.16
2264 - 0.32
2265 - 0.64
2266 - 1.28
2267 - 2.56
2268 - 5.12
2269 - 10.24
2270 - 20.48
2271 - 40.96
2272 - 81.92
2273 - 163.84
2274 - 327.68
2275 - 655.36
2276 - 1310.72
2277 - 2621.44
2278 - 5242.88
2279- name: preemption_attempts_total
2280 subsystem: scheduler
2281 help: Total preemption attempts in the cluster till now
2282 type: Counter
2283 stabilityLevel: STABLE
2284- name: preemption_victims
2285 subsystem: scheduler
2286 help: Number of selected preemption victims
2287 type: Histogram
2288 stabilityLevel: STABLE
2289 buckets:
2290 - 1
2291 - 2
2292 - 4
2293 - 8
2294 - 16
2295 - 32
2296 - 64
2297- name: queue_incoming_pods_total
2298 subsystem: scheduler
2299 help: Number of pods added to scheduling queues by event and queue type.
2300 type: Counter
2301 stabilityLevel: STABLE
2302 labels:
2303 - event
2304 - queue
2305- name: schedule_attempts_total
2306 subsystem: scheduler
2307 help: Number of attempts to schedule pods, by the result. 'unschedulable' means
2308 a pod could not be scheduled, while 'error' means an internal scheduler problem.
2309 type: Counter
2310 stabilityLevel: STABLE
2311 labels:
2312 - profile
2313 - result
2314- name: scheduling_attempt_duration_seconds
2315 subsystem: scheduler
2316 help: Scheduling attempt latency in seconds (scheduling algorithm + binding)
2317 type: Histogram
2318 stabilityLevel: STABLE
2319 labels:
2320 - profile
2321 - result
2322 buckets:
2323 - 0.001
2324 - 0.002
2325 - 0.004
2326 - 0.008
2327 - 0.016
2328 - 0.032
2329 - 0.064
2330 - 0.128
2331 - 0.256
2332 - 0.512
2333 - 1.024
2334 - 2.048
2335 - 4.096
2336 - 8.192
2337 - 16.384
2338- name: operations_seconds
2339 subsystem: csi
2340 help: Container Storage Interface operation duration with gRPC error code status
2341 total
2342 type: Histogram
2343 stabilityLevel: ALPHA
2344 labels:
2345 - driver_name
2346 - grpc_status_code
2347 - method_name
2348 - migrated
2349 buckets:
2350 - 0.1
2351 - 0.25
2352 - 0.5
2353 - 1
2354 - 2.5
2355 - 5
2356 - 10
2357 - 15
2358 - 25
2359 - 50
2360 - 120
2361 - 300
2362 - 600
2363- name: storage_operation_duration_seconds
2364 help: Storage operation duration
2365 type: Histogram
2366 stabilityLevel: ALPHA
2367 labels:
2368 - migrated
2369 - operation_name
2370 - status
2371 - volume_plugin
2372 buckets:
2373 - 0.1
2374 - 0.25
2375 - 0.5
2376 - 1
2377 - 2.5
2378 - 5
2379 - 10
2380 - 15
2381 - 25
2382 - 50
2383 - 120
2384 - 300
2385 - 600
2386- name: volume_operation_total_seconds
2387 help: Storage operation end to end duration in seconds
2388 type: Histogram
2389 stabilityLevel: ALPHA
2390 labels:
2391 - operation_name
2392 - plugin_name
2393 buckets:
2394 - 0.1
2395 - 0.25
2396 - 0.5
2397 - 1
2398 - 2.5
2399 - 5
2400 - 10
2401 - 15
2402 - 25
2403 - 50
2404 - 120
2405 - 300
2406 - 600
2407- name: graph_actions_duration_seconds
2408 subsystem: node_authorizer
2409 help: Histogram of duration of graph actions in node authorizer.
2410 type: Histogram
2411 stabilityLevel: ALPHA
2412 labels:
2413 - operation
2414 buckets:
2415 - 0.0001
2416 - 0.0002
2417 - 0.0004
2418 - 0.0008
2419 - 0.0016
2420 - 0.0032
2421 - 0.0064
2422 - 0.0128
2423 - 0.0256
2424 - 0.0512
2425 - 0.1024
2426 - 0.2048
2427- name: ratcheting_seconds
2428 subsystem: validation
2429 namespace: apiextensions_apiserver
2430 help: Time for comparison of old to new for the purposes of CRDValidationRatcheting
2431 during an UPDATE in seconds.
2432 type: Histogram
2433 stabilityLevel: ALPHA
2434 buckets:
2435 - 1e-05
2436 - 4e-05
2437 - 0.00016
2438 - 0.00064
2439 - 0.00256
2440 - 0.01024
2441 - 0.04096
2442 - 0.16384
2443 - 0.65536
2444 - 2.62144
2445- name: conversion_webhook_duration_seconds
2446 namespace: apiserver
2447 help: Conversion webhook request latency
2448 type: Histogram
2449 stabilityLevel: ALPHA
2450 labels:
2451 - failure_type
2452 - result
2453 buckets:
2454 - 0.005
2455 - 0.01
2456 - 0.02
2457 - 0.05
2458 - 0.1
2459 - 0.2
2460 - 0.5
2461 - 1
2462 - 2
2463 - 5
2464 - 10
2465 - 20
2466 - 30
2467 - 45
2468 - 60
2469- name: conversion_webhook_request_total
2470 namespace: apiserver
2471 help: Counter for conversion webhook requests with success/failure and failure error
2472 type
2473 type: Counter
2474 stabilityLevel: ALPHA
2475 labels:
2476 - failure_type
2477 - result
2478- name: apiserver_crd_conversion_webhook_duration_seconds
2479 help: CRD webhook conversion duration in seconds
2480 type: Histogram
2481 stabilityLevel: ALPHA
2482 labels:
2483 - crd_name
2484 - from_version
2485 - succeeded
2486 - to_version
2487 buckets:
2488 - 0.001
2489 - 0.002
2490 - 0.004
2491 - 0.008
2492 - 0.016
2493 - 0.032
2494 - 0.064
2495 - 0.128
2496 - 0.256
2497 - 0.512
2498 - 1.024
2499 - 2.048
2500 - 4.096
2501 - 8.192
2502 - 16.384
2503- name: apiextensions_openapi_v2_regeneration_count
2504 help: Counter of OpenAPI v2 spec regeneration count broken down by causing CRD name
2505 and reason.
2506 type: Counter
2507 stabilityLevel: ALPHA
2508 labels:
2509 - crd
2510 - reason
2511- name: apiextensions_openapi_v3_regeneration_count
2512 help: Counter of OpenAPI v3 spec regeneration count broken down by group, version,
2513 causing CRD and reason.
2514 type: Counter
2515 stabilityLevel: ALPHA
2516 labels:
2517 - crd
2518 - group
2519 - reason
2520 - version
2521- name: match_condition_evaluation_errors_total
2522 subsystem: admission
2523 namespace: apiserver
2524 help: Admission match condition evaluation errors count, identified by name of resource
2525 containing the match condition and broken out for each kind containing matchConditions
2526 (webhook or policy), operation and admission type (validate or admit).
2527 type: Counter
2528 stabilityLevel: ALPHA
2529 labels:
2530 - kind
2531 - name
2532 - operation
2533 - type
2534- name: match_condition_evaluation_seconds
2535 subsystem: admission
2536 namespace: apiserver
2537 help: Admission match condition evaluation time in seconds, identified by name and
2538 broken out for each kind containing matchConditions (webhook or policy), operation
2539 and type (validate or admit).
2540 type: Histogram
2541 stabilityLevel: ALPHA
2542 labels:
2543 - kind
2544 - name
2545 - operation
2546 - type
2547 buckets:
2548 - 0.001
2549 - 0.005
2550 - 0.01
2551 - 0.025
2552 - 0.1
2553 - 0.2
2554 - 0.25
2555- name: match_condition_exclusions_total
2556 subsystem: admission
2557 namespace: apiserver
2558 help: Admission match condition evaluation exclusions count, identified by name
2559 of resource containing the match condition and broken out for each kind containing
2560 matchConditions (webhook or policy), operation and admission type (validate or
2561 admit).
2562 type: Counter
2563 stabilityLevel: ALPHA
2564 labels:
2565 - kind
2566 - name
2567 - operation
2568 - type
2569- name: step_admission_duration_seconds_summary
2570 subsystem: admission
2571 namespace: apiserver
2572 help: Admission sub-step latency summary in seconds, broken out for each operation
2573 and API resource and step type (validate or admit).
2574 type: Summary
2575 stabilityLevel: ALPHA
2576 labels:
2577 - operation
2578 - rejected
2579 - type
2580 maxAge: 18000000000000
2581- name: webhook_fail_open_count
2582 subsystem: admission
2583 namespace: apiserver
2584 help: Admission webhook fail open count, identified by name and broken out for each
2585 admission type (validating or mutating).
2586 type: Counter
2587 stabilityLevel: ALPHA
2588 labels:
2589 - name
2590 - type
2591- name: webhook_rejection_count
2592 subsystem: admission
2593 namespace: apiserver
2594 help: Admission webhook rejection count, identified by name and broken out for each
2595 admission type (validating or admit) and operation. Additional labels specify
2596 an error type (calling_webhook_error or apiserver_internal_error if an error occurred;
2597 no_error otherwise) and optionally a non-zero rejection code if the webhook rejects
2598 the request with an HTTP status code (honored by the apiserver when the code is
2599 greater or equal to 400). Codes greater than 600 are truncated to 600, to keep
2600 the metrics cardinality bounded.
2601 type: Counter
2602 stabilityLevel: ALPHA
2603 labels:
2604 - error_type
2605 - name
2606 - operation
2607 - rejection_code
2608 - type
2609- name: webhook_request_total
2610 subsystem: admission
2611 namespace: apiserver
2612 help: Admission webhook request total, identified by name and broken out for each
2613 admission type (validating or mutating) and operation. Additional labels specify
2614 whether the request was rejected or not and an HTTP status code. Codes greater
2615 than 600 are truncated to 600, to keep the metrics cardinality bounded.
2616 type: Counter
2617 stabilityLevel: ALPHA
2618 labels:
2619 - code
2620 - name
2621 - operation
2622 - rejected
2623 - type
2624- name: check_duration_seconds
2625 subsystem: validating_admission_policy
2626 namespace: apiserver
2627 help: Validation admission latency for individual validation expressions in seconds,
2628 labeled by policy and further including binding, state and enforcement action
2629 taken.
2630 type: Histogram
2631 stabilityLevel: ALPHA
2632 labels:
2633 - enforcement_action
2634 - policy
2635 - policy_binding
2636 - state
2637 buckets:
2638 - 5e-07
2639 - 0.001
2640 - 0.01
2641 - 0.1
2642 - 1
2643- name: check_total
2644 subsystem: validating_admission_policy
2645 namespace: apiserver
2646 help: Validation admission policy check total, labeled by policy and further identified
2647 by binding, enforcement action taken, and state.
2648 type: Counter
2649 stabilityLevel: ALPHA
2650 labels:
2651 - enforcement_action
2652 - policy
2653 - policy_binding
2654 - state
2655- name: definition_total
2656 subsystem: validating_admission_policy
2657 namespace: apiserver
2658 help: Validation admission policy count total, labeled by state and enforcement
2659 action.
2660 type: Counter
2661 stabilityLevel: ALPHA
2662 labels:
2663 - enforcement_action
2664 - state
2665- name: controller_admission_duration_seconds
2666 subsystem: admission
2667 namespace: apiserver
2668 help: Admission controller latency histogram in seconds, identified by name and
2669 broken out for each operation and API resource and type (validate or admit).
2670 type: Histogram
2671 stabilityLevel: STABLE
2672 labels:
2673 - name
2674 - operation
2675 - rejected
2676 - type
2677 buckets:
2678 - 0.005
2679 - 0.025
2680 - 0.1
2681 - 0.5
2682 - 1
2683 - 2.5
2684- name: step_admission_duration_seconds
2685 subsystem: admission
2686 namespace: apiserver
2687 help: Admission sub-step latency histogram in seconds, broken out for each operation
2688 and API resource and step type (validate or admit).
2689 type: Histogram
2690 stabilityLevel: STABLE
2691 labels:
2692 - operation
2693 - rejected
2694 - type
2695 buckets:
2696 - 0.005
2697 - 0.025
2698 - 0.1
2699 - 0.5
2700 - 1
2701 - 2.5
2702- name: webhook_admission_duration_seconds
2703 subsystem: admission
2704 namespace: apiserver
2705 help: Admission webhook latency histogram in seconds, identified by name and broken
2706 out for each operation and API resource and type (validate or admit).
2707 type: Histogram
2708 stabilityLevel: STABLE
2709 labels:
2710 - name
2711 - operation
2712 - rejected
2713 - type
2714 buckets:
2715 - 0.005
2716 - 0.025
2717 - 0.1
2718 - 0.5
2719 - 1
2720 - 2.5
2721 - 10
2722 - 25
2723- name: aggregator_discovery_aggregation_count_total
2724 help: Counter of number of times discovery was aggregated
2725 type: Counter
2726 stabilityLevel: ALPHA
2727- name: error_total
2728 subsystem: apiserver_audit
2729 help: Counter of audit events that failed to be audited properly. Plugin identifies
2730 the plugin affected by the error.
2731 type: Counter
2732 stabilityLevel: ALPHA
2733 labels:
2734 - plugin
2735- name: event_total
2736 subsystem: apiserver_audit
2737 help: Counter of audit events generated and sent to the audit backend.
2738 type: Counter
2739 stabilityLevel: ALPHA
2740- name: level_total
2741 subsystem: apiserver_audit
2742 help: Counter of policy levels for audit events (1 per request).
2743 type: Counter
2744 stabilityLevel: ALPHA
2745 labels:
2746 - level
2747- name: requests_rejected_total
2748 subsystem: apiserver_audit
2749 help: Counter of apiserver requests rejected due to an error in audit logging backend.
2750 type: Counter
2751 stabilityLevel: ALPHA
2752- name: compilation_duration_seconds
2753 subsystem: cel
2754 namespace: apiserver
2755 help: CEL compilation time in seconds.
2756 type: Histogram
2757 stabilityLevel: ALPHA
2758- name: evaluation_duration_seconds
2759 subsystem: cel
2760 namespace: apiserver
2761 help: CEL evaluation time in seconds.
2762 type: Histogram
2763 stabilityLevel: ALPHA
2764- name: certificate_expiration_seconds
2765 subsystem: client
2766 namespace: apiserver
2767 help: Distribution of the remaining lifetime on the certificate used to authenticate
2768 a request.
2769 type: Histogram
2770 stabilityLevel: ALPHA
2771 buckets:
2772 - 0
2773 - 1800
2774 - 3600
2775 - 7200
2776 - 21600
2777 - 43200
2778 - 86400
2779 - 172800
2780 - 345600
2781 - 604800
2782 - 2.592e+06
2783 - 7.776e+06
2784 - 1.5552e+07
2785 - 3.1104e+07
2786- name: apiserver_delegated_authn_request_duration_seconds
2787 help: Request latency in seconds. Broken down by status code.
2788 type: Histogram
2789 stabilityLevel: ALPHA
2790 labels:
2791 - code
2792 buckets:
2793 - 0.25
2794 - 0.5
2795 - 0.7
2796 - 1
2797 - 1.5
2798 - 3
2799 - 5
2800 - 10
2801- name: apiserver_delegated_authn_request_total
2802 help: Number of HTTP requests partitioned by status code.
2803 type: Counter
2804 stabilityLevel: ALPHA
2805 labels:
2806 - code
2807- name: apiserver_delegated_authz_request_duration_seconds
2808 help: Request latency in seconds. Broken down by status code.
2809 type: Histogram
2810 stabilityLevel: ALPHA
2811 labels:
2812 - code
2813 buckets:
2814 - 0.25
2815 - 0.5
2816 - 0.7
2817 - 1
2818 - 1.5
2819 - 3
2820 - 5
2821 - 10
2822- name: apiserver_delegated_authz_request_total
2823 help: Number of HTTP requests partitioned by status code.
2824 type: Counter
2825 stabilityLevel: ALPHA
2826 labels:
2827 - code
2828- name: active_fetch_count
2829 subsystem: token_cache
2830 namespace: authentication
2831 type: Gauge
2832 stabilityLevel: ALPHA
2833 labels:
2834 - status
2835- name: fetch_total
2836 subsystem: token_cache
2837 namespace: authentication
2838 type: Counter
2839 stabilityLevel: ALPHA
2840 labels:
2841 - status
2842- name: request_duration_seconds
2843 subsystem: token_cache
2844 namespace: authentication
2845 type: Histogram
2846 stabilityLevel: ALPHA
2847 labels:
2848 - status
2849- name: request_total
2850 subsystem: token_cache
2851 namespace: authentication
2852 type: Counter
2853 stabilityLevel: ALPHA
2854 labels:
2855 - status
2856- name: cache_list_fetched_objects_total
2857 namespace: apiserver
2858 help: Number of objects read from watch cache in the course of serving a LIST request
2859 type: Counter
2860 stabilityLevel: ALPHA
2861 labels:
2862 - index
2863 - resource_prefix
2864- name: cache_list_returned_objects_total
2865 namespace: apiserver
2866 help: Number of objects returned for a LIST request from watch cache
2867 type: Counter
2868 stabilityLevel: ALPHA
2869 labels:
2870 - resource_prefix
2871- name: cache_list_total
2872 namespace: apiserver
2873 help: Number of LIST requests served from watch cache
2874 type: Counter
2875 stabilityLevel: ALPHA
2876 labels:
2877 - index
2878 - resource_prefix
2879- name: current_inqueue_requests
2880 subsystem: apiserver
2881 help: Maximal number of queued requests in this apiserver per request kind in last
2882 second.
2883 type: Gauge
2884 stabilityLevel: ALPHA
2885 labels:
2886 - request_kind
2887- name: dial_duration_seconds
2888 subsystem: egress_dialer
2889 namespace: apiserver
2890 help: Dial latency histogram in seconds, labeled by the protocol (http-connect or
2891 grpc), transport (tcp or uds)
2892 type: Histogram
2893 stabilityLevel: ALPHA
2894 labels:
2895 - protocol
2896 - transport
2897 buckets:
2898 - 0.005
2899 - 0.025
2900 - 0.1
2901 - 0.5
2902 - 2.5
2903 - 12.5
2904- name: dial_failure_count
2905 subsystem: egress_dialer
2906 namespace: apiserver
2907 help: Dial failure count, labeled by the protocol (http-connect or grpc), transport
2908 (tcp or uds), and stage (connect or proxy). The stage indicates at which stage
2909 the dial failed
2910 type: Counter
2911 stabilityLevel: ALPHA
2912 labels:
2913 - protocol
2914 - stage
2915 - transport
2916- name: dial_start_total
2917 subsystem: egress_dialer
2918 namespace: apiserver
2919 help: Dial starts, labeled by the protocol (http-connect or grpc) and transport
2920 (tcp or uds).
2921 type: Counter
2922 stabilityLevel: ALPHA
2923 labels:
2924 - protocol
2925 - transport
2926- name: automatic_reload_failures_total
2927 subsystem: encryption_config_controller
2928 namespace: apiserver
2929 help: Total number of failed automatic reloads of encryption configuration split
2930 by apiserver identity.
2931 type: Counter
2932 stabilityLevel: ALPHA
2933 labels:
2934 - apiserver_id_hash
2935- name: automatic_reload_last_timestamp_seconds
2936 subsystem: encryption_config_controller
2937 namespace: apiserver
2938 help: Timestamp of the last successful or failed automatic reload of encryption
2939 configuration split by apiserver identity.
2940 type: Gauge
2941 stabilityLevel: ALPHA
2942 labels:
2943 - apiserver_id_hash
2944 - status
2945- name: automatic_reload_success_total
2946 subsystem: encryption_config_controller
2947 namespace: apiserver
2948 help: Total number of successful automatic reloads of encryption configuration split
2949 by apiserver identity.
2950 type: Counter
2951 stabilityLevel: ALPHA
2952 labels:
2953 - apiserver_id_hash
2954- name: init_events_total
2955 namespace: apiserver
2956 help: Counter of init events processed in watch cache broken by resource type.
2957 type: Counter
2958 stabilityLevel: ALPHA
2959 labels:
2960 - resource
2961- name: request_aborts_total
2962 subsystem: apiserver
2963 help: Number of requests which apiserver aborted possibly due to a timeout, for
2964 each group, version, verb, resource, subresource and scope
2965 type: Counter
2966 stabilityLevel: ALPHA
2967 labels:
2968 - group
2969 - resource
2970 - scope
2971 - subresource
2972 - verb
2973 - version
2974- name: request_body_size_bytes
2975 subsystem: apiserver
2976 help: Apiserver request body size in bytes broken out by resource and verb.
2977 type: Histogram
2978 stabilityLevel: ALPHA
2979 labels:
2980 - resource
2981 - verb
2982 buckets:
2983 - 50000
2984 - 150000
2985 - 250000
2986 - 350000
2987 - 450000
2988 - 550000
2989 - 650000
2990 - 750000
2991 - 850000
2992 - 950000
2993 - 1.05e+06
2994 - 1.15e+06
2995 - 1.25e+06
2996 - 1.35e+06
2997 - 1.45e+06
2998 - 1.55e+06
2999 - 1.65e+06
3000 - 1.75e+06
3001 - 1.85e+06
3002 - 1.95e+06
3003 - 2.05e+06
3004 - 2.15e+06
3005 - 2.25e+06
3006 - 2.35e+06
3007 - 2.45e+06
3008 - 2.55e+06
3009 - 2.65e+06
3010 - 2.75e+06
3011 - 2.85e+06
3012 - 2.95e+06
3013 - 3.05e+06
3014- name: request_filter_duration_seconds
3015 subsystem: apiserver
3016 help: Request filter latency distribution in seconds, for each filter type
3017 type: Histogram
3018 stabilityLevel: ALPHA
3019 labels:
3020 - filter
3021 buckets:
3022 - 0.0001
3023 - 0.0003
3024 - 0.001
3025 - 0.003
3026 - 0.01
3027 - 0.03
3028 - 0.1
3029 - 0.3
3030 - 1
3031 - 5
3032 - 10
3033 - 15
3034 - 30
3035- name: request_post_timeout_total
3036 subsystem: apiserver
3037 help: Tracks the activity of the request handlers after the associated requests
3038 have been timed out by the apiserver
3039 type: Counter
3040 stabilityLevel: ALPHA
3041 labels:
3042 - source
3043 - status
3044- name: request_sli_duration_seconds
3045 subsystem: apiserver
3046 help: Response latency distribution (not counting webhook duration and priority
3047 & fairness queue wait times) in seconds for each verb, group, version, resource,
3048 subresource, scope and component.
3049 type: Histogram
3050 stabilityLevel: ALPHA
3051 labels:
3052 - component
3053 - group
3054 - resource
3055 - scope
3056 - subresource
3057 - verb
3058 - version
3059 buckets:
3060 - 0.05
3061 - 0.1
3062 - 0.2
3063 - 0.4
3064 - 0.6
3065 - 0.8
3066 - 1
3067 - 1.25
3068 - 1.5
3069 - 2
3070 - 3
3071 - 4
3072 - 5
3073 - 6
3074 - 8
3075 - 10
3076 - 15
3077 - 20
3078 - 30
3079 - 45
3080 - 60
3081- name: request_slo_duration_seconds
3082 subsystem: apiserver
3083 help: Response latency distribution (not counting webhook duration and priority
3084 & fairness queue wait times) in seconds for each verb, group, version, resource,
3085 subresource, scope and component.
3086 type: Histogram
3087 deprecatedVersion: 1.27.0
3088 stabilityLevel: ALPHA
3089 labels:
3090 - component
3091 - group
3092 - resource
3093 - scope
3094 - subresource
3095 - verb
3096 - version
3097 buckets:
3098 - 0.05
3099 - 0.1
3100 - 0.2
3101 - 0.4
3102 - 0.6
3103 - 0.8
3104 - 1
3105 - 1.25
3106 - 1.5
3107 - 2
3108 - 3
3109 - 4
3110 - 5
3111 - 6
3112 - 8
3113 - 10
3114 - 15
3115 - 20
3116 - 30
3117 - 45
3118 - 60
3119- name: request_terminations_total
3120 subsystem: apiserver
3121 help: Number of requests which apiserver terminated in self-defense.
3122 type: Counter
3123 stabilityLevel: ALPHA
3124 labels:
3125 - code
3126 - component
3127 - group
3128 - resource
3129 - scope
3130 - subresource
3131 - verb
3132 - version
3133- name: request_timestamp_comparison_time
3134 subsystem: apiserver
3135 help: Time taken for comparison of old vs new objects in UPDATE or PATCH requests
3136 type: Histogram
3137 stabilityLevel: ALPHA
3138 labels:
3139 - code_path
3140 buckets:
3141 - 0.0001
3142 - 0.0003
3143 - 0.001
3144 - 0.003
3145 - 0.01
3146 - 0.03
3147 - 0.1
3148 - 0.3
3149 - 1
3150 - 5
3151- name: selfrequest_total
3152 subsystem: apiserver
3153 help: Counter of apiserver self-requests broken out for each verb, API resource
3154 and subresource.
3155 type: Counter
3156 stabilityLevel: ALPHA
3157 labels:
3158 - resource
3159 - subresource
3160 - verb
3161- name: storage_db_total_size_in_bytes
3162 subsystem: apiserver
3163 help: Total size of the storage database file physically allocated in bytes.
3164 type: Gauge
3165 deprecatedVersion: 1.28.0
3166 stabilityLevel: ALPHA
3167 labels:
3168 - endpoint
3169- name: storage_decode_errors_total
3170 namespace: apiserver
3171 help: Number of stored object decode errors split by object type
3172 type: Counter
3173 stabilityLevel: ALPHA
3174 labels:
3175 - resource
3176- name: storage_events_received_total
3177 subsystem: apiserver
3178 help: Number of etcd events received split by kind.
3179 type: Counter
3180 stabilityLevel: ALPHA
3181 labels:
3182 - resource
3183- name: apiserver_storage_list_evaluated_objects_total
3184 help: Number of objects tested in the course of serving a LIST request from storage
3185 type: Counter
3186 stabilityLevel: ALPHA
3187 labels:
3188 - resource
3189- name: apiserver_storage_list_fetched_objects_total
3190 help: Number of objects read from storage in the course of serving a LIST request
3191 type: Counter
3192 stabilityLevel: ALPHA
3193 labels:
3194 - resource
3195- name: apiserver_storage_list_returned_objects_total
3196 help: Number of objects returned for a LIST request from storage
3197 type: Counter
3198 stabilityLevel: ALPHA
3199 labels:
3200 - resource
3201- name: apiserver_storage_list_total
3202 help: Number of LIST requests served from storage
3203 type: Counter
3204 stabilityLevel: ALPHA
3205 labels:
3206 - resource
3207- name: apiserver_storage_size_bytes
3208 help: Size of the storage database file physically allocated in bytes.
3209 type: Custom
3210 stabilityLevel: ALPHA
3211 labels:
3212 - storage_cluster_id
3213- name: terminated_watchers_total
3214 namespace: apiserver
3215 help: Counter of watchers closed due to unresponsiveness broken by resource type.
3216 type: Counter
3217 stabilityLevel: ALPHA
3218 labels:
3219 - resource
3220- name: tls_handshake_errors_total
3221 subsystem: apiserver
3222 help: Number of requests dropped with 'TLS handshake error from' error
3223 type: Counter
3224 stabilityLevel: ALPHA
3225- name: events_dispatched_total
3226 subsystem: watch_cache
3227 namespace: apiserver
3228 help: Counter of events dispatched in watch cache broken by resource type.
3229 type: Counter
3230 stabilityLevel: ALPHA
3231 labels:
3232 - resource
3233- name: events_received_total
3234 subsystem: watch_cache
3235 namespace: apiserver
3236 help: Counter of events received in watch cache broken by resource type.
3237 type: Counter
3238 stabilityLevel: ALPHA
3239 labels:
3240 - resource
3241- name: initializations_total
3242 subsystem: watch_cache
3243 namespace: apiserver
3244 help: Counter of watch cache initializations broken by resource type.
3245 type: Counter
3246 stabilityLevel: ALPHA
3247 labels:
3248 - resource
3249- name: watch_events_sizes
3250 subsystem: apiserver
3251 help: Watch event size distribution in bytes
3252 type: Histogram
3253 stabilityLevel: ALPHA
3254 labels:
3255 - group
3256 - kind
3257 - version
3258 buckets:
3259 - 1024
3260 - 2048
3261 - 4096
3262 - 8192
3263 - 16384
3264 - 32768
3265 - 65536
3266 - 131072
3267- name: watch_events_total
3268 subsystem: apiserver
3269 help: Number of events sent in watch clients
3270 type: Counter
3271 stabilityLevel: ALPHA
3272 labels:
3273 - group
3274 - kind
3275 - version
3276- name: watch_list_duration_seconds
3277 subsystem: apiserver
3278 help: Response latency distribution in seconds for watch list requests broken by
3279 group, version, resource and scope.
3280 type: Histogram
3281 stabilityLevel: ALPHA
3282 labels:
3283 - group
3284 - resource
3285 - scope
3286 - version
3287 buckets:
3288 - 0.05
3289 - 0.1
3290 - 0.2
3291 - 0.4
3292 - 0.6
3293 - 0.8
3294 - 1
3295 - 2
3296 - 4
3297 - 6
3298 - 8
3299 - 10
3300 - 15
3301 - 20
3302 - 30
3303 - 45
3304 - 60
3305- name: authenticated_user_requests
3306 help: Counter of authenticated requests broken out by username.
3307 type: Counter
3308 stabilityLevel: ALPHA
3309 labels:
3310 - username
3311- name: authentication_attempts
3312 help: Counter of authenticated attempts.
3313 type: Counter
3314 stabilityLevel: ALPHA
3315 labels:
3316 - result
3317- name: authentication_duration_seconds
3318 help: Authentication duration in seconds broken out by result.
3319 type: Histogram
3320 stabilityLevel: ALPHA
3321 labels:
3322 - result
3323 buckets:
3324 - 0.001
3325 - 0.002
3326 - 0.004
3327 - 0.008
3328 - 0.016
3329 - 0.032
3330 - 0.064
3331 - 0.128
3332 - 0.256
3333 - 0.512
3334 - 1.024
3335 - 2.048
3336 - 4.096
3337 - 8.192
3338 - 16.384
3339- name: authorization_attempts_total
3340 help: Counter of authorization attempts broken down by result. It can be either
3341 'allowed', 'denied', 'no-opinion' or 'error'.
3342 type: Counter
3343 stabilityLevel: ALPHA
3344 labels:
3345 - result
3346- name: authorization_duration_seconds
3347 help: Authorization duration in seconds broken out by result.
3348 type: Histogram
3349 stabilityLevel: ALPHA
3350 labels:
3351 - result
3352 buckets:
3353 - 0.001
3354 - 0.002
3355 - 0.004
3356 - 0.008
3357 - 0.016
3358 - 0.032
3359 - 0.064
3360 - 0.128
3361 - 0.256
3362 - 0.512
3363 - 1.024
3364 - 2.048
3365 - 4.096
3366 - 8.192
3367 - 16.384
3368- name: etcd_bookmark_counts
3369 help: Number of etcd bookmarks (progress notify events) split by kind.
3370 type: Gauge
3371 stabilityLevel: ALPHA
3372 labels:
3373 - resource
3374- name: etcd_lease_object_counts
3375 help: Number of objects attached to a single etcd lease.
3376 type: Histogram
3377 stabilityLevel: ALPHA
3378 buckets:
3379 - 10
3380 - 50
3381 - 100
3382 - 500
3383 - 1000
3384 - 2500
3385 - 5000
3386- name: etcd_request_duration_seconds
3387 help: Etcd request latency in seconds for each operation and object type.
3388 type: Histogram
3389 stabilityLevel: ALPHA
3390 labels:
3391 - operation
3392 - type
3393 buckets:
3394 - 0.005
3395 - 0.025
3396 - 0.05
3397 - 0.1
3398 - 0.2
3399 - 0.4
3400 - 0.6
3401 - 0.8
3402 - 1
3403 - 1.25
3404 - 1.5
3405 - 2
3406 - 3
3407 - 4
3408 - 5
3409 - 6
3410 - 8
3411 - 10
3412 - 15
3413 - 20
3414 - 30
3415 - 45
3416 - 60
3417- name: etcd_request_errors_total
3418 help: Etcd failed request counts for each operation and object type.
3419 type: Counter
3420 stabilityLevel: ALPHA
3421 labels:
3422 - operation
3423 - type
3424- name: etcd_requests_total
3425 help: Etcd request counts for each operation and object type.
3426 type: Counter
3427 stabilityLevel: ALPHA
3428 labels:
3429 - operation
3430 - type
3431- name: field_validation_request_duration_seconds
3432 help: Response latency distribution in seconds for each field validation value
3433 type: Histogram
3434 stabilityLevel: ALPHA
3435 labels:
3436 - field_validation
3437 buckets:
3438 - 0.05
3439 - 0.1
3440 - 0.2
3441 - 0.4
3442 - 0.6
3443 - 0.8
3444 - 1
3445 - 1.25
3446 - 1.5
3447 - 2
3448 - 3
3449 - 4
3450 - 5
3451 - 6
3452 - 8
3453 - 10
3454 - 15
3455 - 20
3456 - 30
3457 - 45
3458 - 60
3459- name: capacity
3460 subsystem: watch_cache
3461 help: Total capacity of watch cache broken by resource type.
3462 type: Gauge
3463 stabilityLevel: ALPHA
3464 labels:
3465 - resource
3466- name: capacity_decrease_total
3467 subsystem: watch_cache
3468 help: Total number of watch cache capacity decrease events broken by resource type.
3469 type: Counter
3470 stabilityLevel: ALPHA
3471 labels:
3472 - resource
3473- name: capacity_increase_total
3474 subsystem: watch_cache
3475 help: Total number of watch cache capacity increase events broken by resource type.
3476 type: Counter
3477 stabilityLevel: ALPHA
3478 labels:
3479 - resource
3480- name: current_inflight_requests
3481 subsystem: apiserver
3482 help: Maximal number of currently used inflight request limit of this apiserver
3483 per request kind in last second.
3484 type: Gauge
3485 stabilityLevel: STABLE
3486 labels:
3487 - request_kind
3488- name: longrunning_requests
3489 subsystem: apiserver
3490 help: Gauge of all active long-running apiserver requests broken out by verb, group,
3491 version, resource, scope and component. Not all requests are tracked this way.
3492 type: Gauge
3493 stabilityLevel: STABLE
3494 labels:
3495 - component
3496 - group
3497 - resource
3498 - scope
3499 - subresource
3500 - verb
3501 - version
3502- name: request_duration_seconds
3503 subsystem: apiserver
3504 help: Response latency distribution in seconds for each verb, dry run value, group,
3505 version, resource, subresource, scope and component.
3506 type: Histogram
3507 stabilityLevel: STABLE
3508 labels:
3509 - component
3510 - dry_run
3511 - group
3512 - resource
3513 - scope
3514 - subresource
3515 - verb
3516 - version
3517 buckets:
3518 - 0.005
3519 - 0.025
3520 - 0.05
3521 - 0.1
3522 - 0.2
3523 - 0.4
3524 - 0.6
3525 - 0.8
3526 - 1
3527 - 1.25
3528 - 1.5
3529 - 2
3530 - 3
3531 - 4
3532 - 5
3533 - 6
3534 - 8
3535 - 10
3536 - 15
3537 - 20
3538 - 30
3539 - 45
3540 - 60
3541- name: request_total
3542 subsystem: apiserver
3543 help: Counter of apiserver requests broken out for each verb, dry run value, group,
3544 version, resource, scope, component, and HTTP response code.
3545 type: Counter
3546 stabilityLevel: STABLE
3547 labels:
3548 - code
3549 - component
3550 - dry_run
3551 - group
3552 - resource
3553 - scope
3554 - subresource
3555 - verb
3556 - version
3557- name: requested_deprecated_apis
3558 subsystem: apiserver
3559 help: Gauge of deprecated APIs that have been requested, broken out by API group,
3560 version, resource, subresource, and removed_release.
3561 type: Gauge
3562 stabilityLevel: STABLE
3563 labels:
3564 - group
3565 - removed_release
3566 - resource
3567 - subresource
3568 - version
3569- name: response_sizes
3570 subsystem: apiserver
3571 help: Response size distribution in bytes for each group, version, verb, resource,
3572 subresource, scope and component.
3573 type: Histogram
3574 stabilityLevel: STABLE
3575 labels:
3576 - component
3577 - group
3578 - resource
3579 - scope
3580 - subresource
3581 - verb
3582 - version
3583 buckets:
3584 - 1000
3585 - 10000
3586 - 100000
3587 - 1e+06
3588 - 1e+07
3589 - 1e+08
3590 - 1e+09
3591- name: apiserver_storage_objects
3592 help: Number of stored objects at the time of last check split by kind. In case
3593 of a fetching error, the value will be -1.
3594 type: Gauge
3595 stabilityLevel: STABLE
3596 labels:
3597 - resource
3598- name: dek_cache_fill_percent
3599 subsystem: envelope_encryption
3600 namespace: apiserver
3601 help: Percent of the cache slots currently occupied by cached DEKs.
3602 type: Gauge
3603 stabilityLevel: ALPHA
3604- name: dek_cache_inter_arrival_time_seconds
3605 subsystem: envelope_encryption
3606 namespace: apiserver
3607 help: Time (in seconds) of inter arrival of transformation requests.
3608 type: Histogram
3609 stabilityLevel: ALPHA
3610 labels:
3611 - transformation_type
3612 buckets:
3613 - 60
3614 - 120
3615 - 240
3616 - 480
3617 - 960
3618 - 1920
3619 - 3840
3620 - 7680
3621 - 15360
3622 - 30720
3623- name: dek_source_cache_size
3624 subsystem: envelope_encryption
3625 namespace: apiserver
3626 help: Number of records in data encryption key (DEK) source cache. On a restart,
3627 this value is an approximation of the number of decrypt RPC calls the server will
3628 make to the KMS plugin.
3629 type: Gauge
3630 stabilityLevel: ALPHA
3631 labels:
3632 - provider_name
3633- name: invalid_key_id_from_status_total
3634 subsystem: envelope_encryption
3635 namespace: apiserver
3636 help: Number of times an invalid keyID is returned by the Status RPC call split
3637 by error.
3638 type: Counter
3639 stabilityLevel: ALPHA
3640 labels:
3641 - error
3642 - provider_name
3643- name: key_id_hash_last_timestamp_seconds
3644 subsystem: envelope_encryption
3645 namespace: apiserver
3646 help: The last time in seconds when a keyID was used.
3647 type: Gauge
3648 stabilityLevel: ALPHA
3649 labels:
3650 - apiserver_id_hash
3651 - key_id_hash
3652 - provider_name
3653 - transformation_type
3654- name: key_id_hash_status_last_timestamp_seconds
3655 subsystem: envelope_encryption
3656 namespace: apiserver
3657 help: The last time in seconds when a keyID was returned by the Status RPC call.
3658 type: Gauge
3659 stabilityLevel: ALPHA
3660 labels:
3661 - apiserver_id_hash
3662 - key_id_hash
3663 - provider_name
3664- name: key_id_hash_total
3665 subsystem: envelope_encryption
3666 namespace: apiserver
3667 help: Number of times a keyID is used split by transformation type, provider, and
3668 apiserver identity.
3669 type: Counter
3670 stabilityLevel: ALPHA
3671 labels:
3672 - apiserver_id_hash
3673 - key_id_hash
3674 - provider_name
3675 - transformation_type
3676- name: kms_operations_latency_seconds
3677 subsystem: envelope_encryption
3678 namespace: apiserver
3679 help: KMS operation duration with gRPC error code status total.
3680 type: Histogram
3681 stabilityLevel: ALPHA
3682 labels:
3683 - grpc_status_code
3684 - method_name
3685 - provider_name
3686 buckets:
3687 - 0.0001
3688 - 0.0002
3689 - 0.0004
3690 - 0.0008
3691 - 0.0016
3692 - 0.0032
3693 - 0.0064
3694 - 0.0128
3695 - 0.0256
3696 - 0.0512
3697 - 0.1024
3698 - 0.2048
3699 - 0.4096
3700 - 0.8192
3701 - 1.6384
3702 - 3.2768
3703 - 6.5536
3704 - 13.1072
3705 - 26.2144
3706 - 52.4288
3707- name: current_inqueue_seats
3708 subsystem: flowcontrol
3709 namespace: apiserver
3710 help: Number of seats currently pending in queues of the API Priority and Fairness
3711 subsystem
3712 type: Gauge
3713 stabilityLevel: ALPHA
3714 labels:
3715 - flow_schema
3716 - priority_level
3717- name: current_limit_seats
3718 subsystem: flowcontrol
3719 namespace: apiserver
3720 help: current derived number of execution seats available to each priority level
3721 type: Gauge
3722 stabilityLevel: ALPHA
3723 labels:
3724 - priority_level
3725- name: current_r
3726 subsystem: flowcontrol
3727 namespace: apiserver
3728 help: R(time of last change)
3729 type: Gauge
3730 stabilityLevel: ALPHA
3731 labels:
3732 - priority_level
3733- name: demand_seats
3734 subsystem: flowcontrol
3735 namespace: apiserver
3736 help: Observations, at the end of every nanosecond, of (the number of seats each
3737 priority level could use) / (nominal number of seats for that level)
3738 type: TimingRatioHistogram
3739 stabilityLevel: ALPHA
3740 labels:
3741 - priority_level
3742 buckets:
3743 - 0.2
3744 - 0.4
3745 - 0.6
3746 - 0.8
3747 - 1
3748 - 1.2
3749 - 1.4
3750 - 1.7
3751 - 2
3752 - 2.8
3753 - 4
3754 - 6
3755- name: demand_seats_average
3756 subsystem: flowcontrol
3757 namespace: apiserver
3758 help: Time-weighted average, over last adjustment period, of demand_seats
3759 type: Gauge
3760 stabilityLevel: ALPHA
3761 labels:
3762 - priority_level
3763- name: demand_seats_high_watermark
3764 subsystem: flowcontrol
3765 namespace: apiserver
3766 help: High watermark, over last adjustment period, of demand_seats
3767 type: Gauge
3768 stabilityLevel: ALPHA
3769 labels:
3770 - priority_level
3771- name: demand_seats_smoothed
3772 subsystem: flowcontrol
3773 namespace: apiserver
3774 help: Smoothed seat demands
3775 type: Gauge
3776 stabilityLevel: ALPHA
3777 labels:
3778 - priority_level
3779- name: demand_seats_stdev
3780 subsystem: flowcontrol
3781 namespace: apiserver
3782 help: Time-weighted standard deviation, over last adjustment period, of demand_seats
3783 type: Gauge
3784 stabilityLevel: ALPHA
3785 labels:
3786 - priority_level
3787- name: dispatch_r
3788 subsystem: flowcontrol
3789 namespace: apiserver
3790 help: R(time of last dispatch)
3791 type: Gauge
3792 stabilityLevel: ALPHA
3793 labels:
3794 - priority_level
3795- name: epoch_advance_total
3796 subsystem: flowcontrol
3797 namespace: apiserver
3798 help: Number of times the queueset's progress meter jumped backward
3799 type: Counter
3800 stabilityLevel: ALPHA
3801 labels:
3802 - priority_level
3803 - success
3804- name: latest_s
3805 subsystem: flowcontrol
3806 namespace: apiserver
3807 help: S(most recently dispatched request)
3808 type: Gauge
3809 stabilityLevel: ALPHA
3810 labels:
3811 - priority_level
3812- name: lower_limit_seats
3813 subsystem: flowcontrol
3814 namespace: apiserver
3815 help: Configured lower bound on number of execution seats available to each priority
3816 level
3817 type: Gauge
3818 stabilityLevel: ALPHA
3819 labels:
3820 - priority_level
3821- name: next_discounted_s_bounds
3822 subsystem: flowcontrol
3823 namespace: apiserver
3824 help: min and max, over queues, of S(oldest waiting request in queue) - estimated
3825 work in progress
3826 type: Gauge
3827 stabilityLevel: ALPHA
3828 labels:
3829 - bound
3830 - priority_level
3831- name: next_s_bounds
3832 subsystem: flowcontrol
3833 namespace: apiserver
3834 help: min and max, over queues, of S(oldest waiting request in queue)
3835 type: Gauge
3836 stabilityLevel: ALPHA
3837 labels:
3838 - bound
3839 - priority_level
3840- name: priority_level_request_utilization
3841 subsystem: flowcontrol
3842 namespace: apiserver
3843 help: Observations, at the end of every nanosecond, of number of requests (as a
3844 fraction of the relevant limit) waiting or in any stage of execution (but only
3845 initial stage for WATCHes)
3846 type: TimingRatioHistogram
3847 stabilityLevel: ALPHA
3848 labels:
3849 - phase
3850 - priority_level
3851 buckets:
3852 - 0
3853 - 0.001
3854 - 0.003
3855 - 0.01
3856 - 0.03
3857 - 0.1
3858 - 0.25
3859 - 0.5
3860 - 0.75
3861 - 1
3862- name: priority_level_seat_utilization
3863 subsystem: flowcontrol
3864 namespace: apiserver
3865 help: Observations, at the end of every nanosecond, of utilization of seats for
3866 any stage of execution (but only initial stage for WATCHes)
3867 type: TimingRatioHistogram
3868 stabilityLevel: ALPHA
3869 labels:
3870 - priority_level
3871 buckets:
3872 - 0
3873 - 0.1
3874 - 0.2
3875 - 0.3
3876 - 0.4
3877 - 0.5
3878 - 0.6
3879 - 0.7
3880 - 0.8
3881 - 0.9
3882 - 0.95
3883 - 0.99
3884 - 1
3885 constLabels:
3886 phase: executing
3887- name: read_vs_write_current_requests
3888 subsystem: flowcontrol
3889 namespace: apiserver
3890 help: Observations, at the end of every nanosecond, of the number of requests (as
3891 a fraction of the relevant limit) waiting or in regular stage of execution
3892 type: TimingRatioHistogram
3893 stabilityLevel: ALPHA
3894 labels:
3895 - phase
3896 - request_kind
3897 buckets:
3898 - 0
3899 - 0.001
3900 - 0.01
3901 - 0.1
3902 - 0.2
3903 - 0.3
3904 - 0.4
3905 - 0.5
3906 - 0.6
3907 - 0.7
3908 - 0.8
3909 - 0.9
3910 - 0.95
3911 - 0.99
3912 - 1
3913- name: request_concurrency_in_use
3914 subsystem: flowcontrol
3915 namespace: apiserver
3916 help: Concurrency (number of seats) occupied by the currently executing (initial
3917 stage for a WATCH, any stage otherwise) requests in the API Priority and Fairness
3918 subsystem
3919 type: Gauge
3920 deprecatedVersion: 1.31.0
3921 stabilityLevel: ALPHA
3922 labels:
3923 - flow_schema
3924 - priority_level
3925- name: request_concurrency_limit
3926 subsystem: flowcontrol
3927 namespace: apiserver
3928 help: Nominal number of execution seats configured for each priority level
3929 type: Gauge
3930 deprecatedVersion: 1.30.0
3931 stabilityLevel: ALPHA
3932 labels:
3933 - priority_level
3934- name: request_dispatch_no_accommodation_total
3935 subsystem: flowcontrol
3936 namespace: apiserver
3937 help: Number of times a dispatch attempt resulted in a non accommodation due to
3938 lack of available seats
3939 type: Counter
3940 stabilityLevel: ALPHA
3941 labels:
3942 - flow_schema
3943 - priority_level
3944- name: request_execution_seconds
3945 subsystem: flowcontrol
3946 namespace: apiserver
3947 help: Duration of initial stage (for a WATCH) or any (for a non-WATCH) stage of
3948 request execution in the API Priority and Fairness subsystem
3949 type: Histogram
3950 stabilityLevel: ALPHA
3951 labels:
3952 - flow_schema
3953 - priority_level
3954 - type
3955 buckets:
3956 - 0
3957 - 0.005
3958 - 0.02
3959 - 0.05
3960 - 0.1
3961 - 0.2
3962 - 0.5
3963 - 1
3964 - 2
3965 - 5
3966 - 10
3967 - 15
3968 - 30
3969- name: request_queue_length_after_enqueue
3970 subsystem: flowcontrol
3971 namespace: apiserver
3972 help: Length of queue in the API Priority and Fairness subsystem, as seen by each
3973 request after it is enqueued
3974 type: Histogram
3975 stabilityLevel: ALPHA
3976 labels:
3977 - flow_schema
3978 - priority_level
3979 buckets:
3980 - 0
3981 - 10
3982 - 25
3983 - 50
3984 - 100
3985 - 250
3986 - 500
3987 - 1000
3988- name: seat_fair_frac
3989 subsystem: flowcontrol
3990 namespace: apiserver
3991 help: Fair fraction of server's concurrency to allocate to each priority level that
3992 can use it
3993 type: Gauge
3994 stabilityLevel: ALPHA
3995- name: target_seats
3996 subsystem: flowcontrol
3997 namespace: apiserver
3998 help: Seat allocation targets
3999 type: Gauge
4000 stabilityLevel: ALPHA
4001 labels:
4002 - priority_level
4003- name: upper_limit_seats
4004 subsystem: flowcontrol
4005 namespace: apiserver
4006 help: Configured upper bound on number of execution seats available to each priority
4007 level
4008 type: Gauge
4009 stabilityLevel: ALPHA
4010 labels:
4011 - priority_level
4012- name: watch_count_samples
4013 subsystem: flowcontrol
4014 namespace: apiserver
4015 help: count of watchers for mutating requests in API Priority and Fairness
4016 type: Histogram
4017 stabilityLevel: ALPHA
4018 labels:
4019 - flow_schema
4020 - priority_level
4021 buckets:
4022 - 0
4023 - 1
4024 - 10
4025 - 100
4026 - 1000
4027 - 10000
4028- name: work_estimated_seats
4029 subsystem: flowcontrol
4030 namespace: apiserver
4031 help: Number of estimated seats (maximum of initial and final seats) associated
4032 with requests in API Priority and Fairness
4033 type: Histogram
4034 stabilityLevel: ALPHA
4035 labels:
4036 - flow_schema
4037 - priority_level
4038 buckets:
4039 - 1
4040 - 2
4041 - 4
4042 - 10
4043- name: rerouted_request_total
4044 subsystem: apiserver
4045 help: Total number of requests that were proxied to a peer kube apiserver because
4046 the local apiserver was not capable of serving it
4047 type: Counter
4048 stabilityLevel: ALPHA
4049 labels:
4050 - code
4051- name: data_key_generation_duration_seconds
4052 subsystem: storage
4053 namespace: apiserver
4054 help: Latencies in seconds of data encryption key(DEK) generation operations.
4055 type: Histogram
4056 stabilityLevel: ALPHA
4057 buckets:
4058 - 5e-06
4059 - 1e-05
4060 - 2e-05
4061 - 4e-05
4062 - 8e-05
4063 - 0.00016
4064 - 0.00032
4065 - 0.00064
4066 - 0.00128
4067 - 0.00256
4068 - 0.00512
4069 - 0.01024
4070 - 0.02048
4071 - 0.04096
4072- name: data_key_generation_failures_total
4073 subsystem: storage
4074 namespace: apiserver
4075 help: Total number of failed data encryption key(DEK) generation operations.
4076 type: Counter
4077 stabilityLevel: ALPHA
4078- name: envelope_transformation_cache_misses_total
4079 subsystem: storage
4080 namespace: apiserver
4081 help: Total number of cache misses while accessing key decryption key(KEK).
4082 type: Counter
4083 stabilityLevel: ALPHA
4084- name: transformation_duration_seconds
4085 subsystem: storage
4086 namespace: apiserver
4087 help: Latencies in seconds of value transformation operations.
4088 type: Histogram
4089 stabilityLevel: ALPHA
4090 labels:
4091 - transformation_type
4092 - transformer_prefix
4093 buckets:
4094 - 5e-06
4095 - 1e-05
4096 - 2e-05
4097 - 4e-05
4098 - 8e-05
4099 - 0.00016
4100 - 0.00032
4101 - 0.00064
4102 - 0.00128
4103 - 0.00256
4104 - 0.00512
4105 - 0.01024
4106 - 0.02048
4107 - 0.04096
4108 - 0.08192
4109 - 0.16384
4110 - 0.32768
4111 - 0.65536
4112 - 1.31072
4113 - 2.62144
4114 - 5.24288
4115 - 10.48576
4116 - 20.97152
4117 - 41.94304
4118 - 83.88608
4119- name: transformation_operations_total
4120 subsystem: storage
4121 namespace: apiserver
4122 help: Total number of transformations. Successful transformation will have a status
4123 'OK' and a varied status string when the transformation fails. This status and
4124 transformation_type fields may be used for alerting on encryption/decryption failure
4125 using transformation_type from_storage for decryption and to_storage for encryption
4126 type: Counter
4127 stabilityLevel: ALPHA
4128 labels:
4129 - status
4130 - transformation_type
4131 - transformer_prefix
4132- name: x509_insecure_sha1_total
4133 subsystem: webhooks
4134 namespace: apiserver
4135 help: Counts the number of requests to servers with insecure SHA1 signatures in
4136 their serving certificate OR the number of connection failures due to the insecure
4137 SHA1 signatures (either/or, based on the runtime environment)
4138 type: Counter
4139 stabilityLevel: ALPHA
4140- name: x509_missing_san_total
4141 subsystem: webhooks
4142 namespace: apiserver
4143 help: Counts the number of requests to servers missing SAN extension in their serving
4144 certificate OR the number of connection failures due to the lack of x509 certificate
4145 SAN extension missing (either/or, based on the runtime environment)
4146 type: Counter
4147 stabilityLevel: ALPHA
4148- name: current_executing_requests
4149 subsystem: flowcontrol
4150 namespace: apiserver
4151 help: Number of requests in initial (for a WATCH) or any (for a non-WATCH) execution
4152 stage in the API Priority and Fairness subsystem
4153 type: Gauge
4154 stabilityLevel: BETA
4155 labels:
4156 - flow_schema
4157 - priority_level
4158- name: current_executing_seats
4159 subsystem: flowcontrol
4160 namespace: apiserver
4161 help: Concurrency (number of seats) occupied by the currently executing (initial
4162 stage for a WATCH, any stage otherwise) requests in the API Priority and Fairness
4163 subsystem
4164 type: Gauge
4165 stabilityLevel: BETA
4166 labels:
4167 - flow_schema
4168 - priority_level
4169- name: current_inqueue_requests
4170 subsystem: flowcontrol
4171 namespace: apiserver
4172 help: Number of requests currently pending in queues of the API Priority and Fairness
4173 subsystem
4174 type: Gauge
4175 stabilityLevel: BETA
4176 labels:
4177 - flow_schema
4178 - priority_level
4179- name: dispatched_requests_total
4180 subsystem: flowcontrol
4181 namespace: apiserver
4182 help: Number of requests executed by API Priority and Fairness subsystem
4183 type: Counter
4184 stabilityLevel: BETA
4185 labels:
4186 - flow_schema
4187 - priority_level
4188- name: nominal_limit_seats
4189 subsystem: flowcontrol
4190 namespace: apiserver
4191 help: Nominal number of execution seats configured for each priority level
4192 type: Gauge
4193 stabilityLevel: BETA
4194 labels:
4195 - priority_level
4196- name: rejected_requests_total
4197 subsystem: flowcontrol
4198 namespace: apiserver
4199 help: Number of requests rejected by API Priority and Fairness subsystem
4200 type: Counter
4201 stabilityLevel: BETA
4202 labels:
4203 - flow_schema
4204 - priority_level
4205 - reason
4206- name: request_wait_duration_seconds
4207 subsystem: flowcontrol
4208 namespace: apiserver
4209 help: Length of time a request spent waiting in its queue
4210 type: Histogram
4211 stabilityLevel: BETA
4212 labels:
4213 - execute
4214 - flow_schema
4215 - priority_level
4216 buckets:
4217 - 0
4218 - 0.005
4219 - 0.02
4220 - 0.05
4221 - 0.1
4222 - 0.2
4223 - 0.5
4224 - 1
4225 - 2
4226 - 5
4227 - 10
4228 - 15
4229 - 30
4230- name: request_duration_seconds
4231 subsystem: cloud_provider_webhook
4232 help: Request latency in seconds. Broken down by status code.
4233 type: Histogram
4234 stabilityLevel: ALPHA
4235 labels:
4236 - code
4237 - webhook
4238 buckets:
4239 - 0.25
4240 - 0.5
4241 - 0.7
4242 - 1
4243 - 1.5
4244 - 3
4245 - 5
4246 - 10
4247- name: request_total
4248 subsystem: cloud_provider_webhook
4249 help: Number of HTTP requests partitioned by status code.
4250 type: Counter
4251 stabilityLevel: ALPHA
4252 labels:
4253 - code
4254 - webhook
4255- name: cloud_provider_taint_removal_delay_seconds
4256 subsystem: node_controller
4257 help: Number of seconds after node creation when NodeController removed the cloud-provider
4258 taint of a single node.
4259 type: Histogram
4260 stabilityLevel: ALPHA
4261 buckets:
4262 - 1
4263 - 4
4264 - 16
4265 - 64
4266 - 256
4267 - 1024
4268- name: initial_node_sync_delay_seconds
4269 subsystem: node_controller
4270 help: Number of seconds after node creation when NodeController finished the initial
4271 synchronization of a single node.
4272 type: Histogram
4273 stabilityLevel: ALPHA
4274 buckets:
4275 - 1
4276 - 4
4277 - 16
4278 - 64
4279 - 256
4280 - 1024
4281- name: loadbalancer_sync_total
4282 subsystem: service_controller
4283 help: A metric counting the amount of times any load balancer has been configured,
4284 as an effect of service/node changes on the cluster
4285 type: Counter
4286 stabilityLevel: ALPHA
4287- name: nodesync_error_total
4288 subsystem: service_controller
4289 help: A metric counting the amount of times any load balancer has been configured
4290 and errored, as an effect of node changes on the cluster
4291 type: Counter
4292 stabilityLevel: ALPHA
4293- name: nodesync_latency_seconds
4294 subsystem: service_controller
4295 help: A metric measuring the latency for nodesync which updates loadbalancer hosts
4296 on cluster node updates.
4297 type: Histogram
4298 stabilityLevel: ALPHA
4299 buckets:
4300 - 1
4301 - 2
4302 - 4
4303 - 8
4304 - 16
4305 - 32
4306 - 64
4307 - 128
4308 - 256
4309 - 512
4310 - 1024
4311 - 2048
4312 - 4096
4313 - 8192
4314 - 16384
4315- name: update_loadbalancer_host_latency_seconds
4316 subsystem: service_controller
4317 help: A metric measuring the latency for updating each load balancer hosts.
4318 type: Histogram
4319 stabilityLevel: ALPHA
4320 buckets:
4321 - 1
4322 - 2
4323 - 4
4324 - 8
4325 - 16
4326 - 32
4327 - 64
4328 - 128
4329 - 256
4330 - 512
4331 - 1024
4332 - 2048
4333 - 4096
4334 - 8192
4335 - 16384
4336- name: changes
4337 subsystem: endpoint_slice_controller
4338 help: Number of EndpointSlice changes
4339 type: Counter
4340 stabilityLevel: ALPHA
4341 labels:
4342 - operation
4343- name: desired_endpoint_slices
4344 subsystem: endpoint_slice_controller
4345 help: Number of EndpointSlices that would exist with perfect endpoint allocation
4346 type: Gauge
4347 stabilityLevel: ALPHA
4348- name: endpoints_added_per_sync
4349 subsystem: endpoint_slice_controller
4350 help: Number of endpoints added on each Service sync
4351 type: Histogram
4352 stabilityLevel: ALPHA
4353 buckets:
4354 - 2
4355 - 4
4356 - 8
4357 - 16
4358 - 32
4359 - 64
4360 - 128
4361 - 256
4362 - 512
4363 - 1024
4364 - 2048
4365 - 4096
4366 - 8192
4367 - 16384
4368 - 32768
4369- name: endpoints_desired
4370 subsystem: endpoint_slice_controller
4371 help: Number of endpoints desired
4372 type: Gauge
4373 stabilityLevel: ALPHA
4374- name: endpoints_removed_per_sync
4375 subsystem: endpoint_slice_controller
4376 help: Number of endpoints removed on each Service sync
4377 type: Histogram
4378 stabilityLevel: ALPHA
4379 buckets:
4380 - 2
4381 - 4
4382 - 8
4383 - 16
4384 - 32
4385 - 64
4386 - 128
4387 - 256
4388 - 512
4389 - 1024
4390 - 2048
4391 - 4096
4392 - 8192
4393 - 16384
4394 - 32768
4395- name: endpointslices_changed_per_sync
4396 subsystem: endpoint_slice_controller
4397 help: Number of EndpointSlices changed on each Service sync
4398 type: Histogram
4399 stabilityLevel: ALPHA
4400 labels:
4401 - topology
4402- name: num_endpoint_slices
4403 subsystem: endpoint_slice_controller
4404 help: Number of EndpointSlices
4405 type: Gauge
4406 stabilityLevel: ALPHA
4407- name: syncs
4408 subsystem: endpoint_slice_controller
4409 help: Number of EndpointSlice syncs
4410 type: Counter
4411 stabilityLevel: ALPHA
4412 labels:
4413 - result
4414- name: kubernetes_build_info
4415 help: A metric with a constant '1' value labeled by major, minor, git version, git
4416 commit, git tree state, build date, Go version, and compiler from which Kubernetes
4417 was built, and platform on which it is running.
4418 type: Gauge
4419 stabilityLevel: ALPHA
4420 labels:
4421 - build_date
4422 - compiler
4423 - git_commit
4424 - git_tree_state
4425 - git_version
4426 - go_version
4427 - major
4428 - minor
4429 - platform
4430- name: leader_election_master_status
4431 help: Gauge of if the reporting system is master of the relevant lease, 0 indicates
4432 backup, 1 indicates master. 'name' is the string used to identify the lease. Please
4433 make sure to group by name.
4434 type: Gauge
4435 stabilityLevel: ALPHA
4436 labels:
4437 - name
4438- name: leader_election_slowpath_total
4439 help: Total number of slow path exercised in renewing leader leases. 'name' is the
4440 string used to identify the lease. Please make sure to group by name.
4441 type: Counter
4442 stabilityLevel: ALPHA
4443 labels:
4444 - name
4445- name: rest_client_dns_resolution_duration_seconds
4446 help: DNS resolver latency in seconds. Broken down by host.
4447 type: Histogram
4448 stabilityLevel: ALPHA
4449 labels:
4450 - host
4451 buckets:
4452 - 0.005
4453 - 0.025
4454 - 0.1
4455 - 0.25
4456 - 0.5
4457 - 1
4458 - 2
4459 - 4
4460 - 8
4461 - 15
4462 - 30
4463- name: rest_client_exec_plugin_call_total
4464 help: Number of calls to an exec plugin, partitioned by the type of event encountered
4465 (no_error, plugin_execution_error, plugin_not_found_error, client_internal_error)
4466 and an optional exit code. The exit code will be set to 0 if and only if the plugin
4467 call was successful.
4468 type: Counter
4469 stabilityLevel: ALPHA
4470 labels:
4471 - call_status
4472 - code
4473- name: rest_client_exec_plugin_certificate_rotation_age
4474 help: Histogram of the number of seconds the last auth exec plugin client certificate
4475 lived before being rotated. If auth exec plugin client certificates are unused,
4476 histogram will contain no data.
4477 type: Histogram
4478 stabilityLevel: ALPHA
4479 buckets:
4480 - 600
4481 - 1800
4482 - 3600
4483 - 14400
4484 - 86400
4485 - 604800
4486 - 2.592e+06
4487 - 7.776e+06
4488 - 1.5552e+07
4489 - 3.1104e+07
4490 - 1.24416e+08
4491- name: rest_client_exec_plugin_ttl_seconds
4492 help: Gauge of the shortest TTL (time-to-live) of the client certificate(s) managed
4493 by the auth exec plugin. The value is in seconds until certificate expiry (negative
4494 if already expired). If auth exec plugins are unused or manage no TLS certificates,
4495 the value will be +INF.
4496 type: Gauge
4497 stabilityLevel: ALPHA
4498- name: rest_client_rate_limiter_duration_seconds
4499 help: Client side rate limiter latency in seconds. Broken down by verb, and host.
4500 type: Histogram
4501 stabilityLevel: ALPHA
4502 labels:
4503 - host
4504 - verb
4505 buckets:
4506 - 0.005
4507 - 0.025
4508 - 0.1
4509 - 0.25
4510 - 0.5
4511 - 1
4512 - 2
4513 - 4
4514 - 8
4515 - 15
4516 - 30
4517 - 60
4518- name: rest_client_request_duration_seconds
4519 help: Request latency in seconds. Broken down by verb, and host.
4520 type: Histogram
4521 stabilityLevel: ALPHA
4522 labels:
4523 - host
4524 - verb
4525 buckets:
4526 - 0.005
4527 - 0.025
4528 - 0.1
4529 - 0.25
4530 - 0.5
4531 - 1
4532 - 2
4533 - 4
4534 - 8
4535 - 15
4536 - 30
4537 - 60
4538- name: rest_client_request_retries_total
4539 help: Number of request retries, partitioned by status code, verb, and host.
4540 type: Counter
4541 stabilityLevel: ALPHA
4542 labels:
4543 - code
4544 - host
4545 - verb
4546- name: rest_client_request_size_bytes
4547 help: Request size in bytes. Broken down by verb and host.
4548 type: Histogram
4549 stabilityLevel: ALPHA
4550 labels:
4551 - host
4552 - verb
4553 buckets:
4554 - 64
4555 - 256
4556 - 512
4557 - 1024
4558 - 4096
4559 - 16384
4560 - 65536
4561 - 262144
4562 - 1.048576e+06
4563 - 4.194304e+06
4564 - 1.6777216e+07
4565- name: rest_client_requests_total
4566 help: Number of HTTP requests, partitioned by status code, method, and host.
4567 type: Counter
4568 stabilityLevel: ALPHA
4569 labels:
4570 - code
4571 - host
4572 - method
4573- name: rest_client_response_size_bytes
4574 help: Response size in bytes. Broken down by verb and host.
4575 type: Histogram
4576 stabilityLevel: ALPHA
4577 labels:
4578 - host
4579 - verb
4580 buckets:
4581 - 64
4582 - 256
4583 - 512
4584 - 1024
4585 - 4096
4586 - 16384
4587 - 65536
4588 - 262144
4589 - 1.048576e+06
4590 - 4.194304e+06
4591 - 1.6777216e+07
4592- name: rest_client_transport_cache_entries
4593 help: Number of transport entries in the internal cache.
4594 type: Gauge
4595 stabilityLevel: ALPHA
4596- name: rest_client_transport_create_calls_total
4597 help: 'Number of calls to get a new transport, partitioned by the result of the
4598 operation hit: obtained from the cache, miss: created and added to the cache,
4599 uncacheable: created and not cached'
4600 type: Counter
4601 stabilityLevel: ALPHA
4602 labels:
4603 - result
4604- name: running_managed_controllers
4605 help: Indicates where instances of a controller are currently running
4606 type: Gauge
4607 stabilityLevel: ALPHA
4608 labels:
4609 - manager
4610 - name
4611- name: adds_total
4612 subsystem: workqueue
4613 help: Total number of adds handled by workqueue
4614 type: Counter
4615 stabilityLevel: ALPHA
4616 labels:
4617 - name
4618- name: depth
4619 subsystem: workqueue
4620 help: Current depth of workqueue
4621 type: Gauge
4622 stabilityLevel: ALPHA
4623 labels:
4624 - name
4625- name: longest_running_processor_seconds
4626 subsystem: workqueue
4627 help: How many seconds has the longest running processor for workqueue been running.
4628 type: Gauge
4629 stabilityLevel: ALPHA
4630 labels:
4631 - name
4632- name: queue_duration_seconds
4633 subsystem: workqueue
4634 help: How long in seconds an item stays in workqueue before being requested.
4635 type: Histogram
4636 stabilityLevel: ALPHA
4637 labels:
4638 - name
4639 buckets:
4640 - 1e-08
4641 - 1e-07
4642 - 1e-06
4643 - 9.999999999999999e-06
4644 - 9.999999999999999e-05
4645 - 0.001
4646 - 0.01
4647 - 0.1
4648 - 1
4649 - 10
4650- name: retries_total
4651 subsystem: workqueue
4652 help: Total number of retries handled by workqueue
4653 type: Counter
4654 stabilityLevel: ALPHA
4655 labels:
4656 - name
4657- name: unfinished_work_seconds
4658 subsystem: workqueue
4659 help: How many seconds of work has done that is in progress and hasn't been observed
4660 by work_duration. Large values indicate stuck threads. One can deduce the number
4661 of stuck threads by observing the rate at which this increases.
4662 type: Gauge
4663 stabilityLevel: ALPHA
4664 labels:
4665 - name
4666- name: work_duration_seconds
4667 subsystem: workqueue
4668 help: How long in seconds processing an item from workqueue takes.
4669 type: Histogram
4670 stabilityLevel: ALPHA
4671 labels:
4672 - name
4673 buckets:
4674 - 1e-08
4675 - 1e-07
4676 - 1e-06
4677 - 9.999999999999999e-06
4678 - 9.999999999999999e-05
4679 - 0.001
4680 - 0.01
4681 - 0.1
4682 - 1
4683 - 10
4684- name: disabled_metrics_total
4685 help: The count of disabled metrics.
4686 type: Counter
4687 stabilityLevel: BETA
4688- name: hidden_metrics_total
4689 help: The count of hidden metrics.
4690 type: Counter
4691 stabilityLevel: BETA
4692- name: feature_enabled
4693 namespace: kubernetes
4694 help: This metric records the data about the stage and enablement of a k8s feature.
4695 type: Gauge
4696 stabilityLevel: BETA
4697 labels:
4698 - name
4699 - stage
4700- name: registered_metrics_total
4701 help: The count of registered metrics broken by stability level and deprecation
4702 version.
4703 type: Counter
4704 stabilityLevel: BETA
4705 labels:
4706 - deprecated_version
4707 - stability_level
4708- name: healthcheck
4709 namespace: kubernetes
4710 help: This metric records the result of a single healthcheck.
4711 type: Gauge
4712 stabilityLevel: STABLE
4713 labels:
4714 - name
4715 - type
4716- name: healthchecks_total
4717 namespace: kubernetes
4718 help: This metric records the results of all healthcheck.
4719 type: Counter
4720 stabilityLevel: STABLE
4721 labels:
4722 - name
4723 - status
4724 - type
4725- name: aggregator_openapi_v2_regeneration_count
4726 help: Counter of OpenAPI v2 spec regeneration count broken down by causing APIService
4727 name and reason.
4728 type: Counter
4729 stabilityLevel: ALPHA
4730 labels:
4731 - apiservice
4732 - reason
4733- name: aggregator_openapi_v2_regeneration_duration
4734 help: Gauge of OpenAPI v2 spec regeneration duration in seconds.
4735 type: Gauge
4736 stabilityLevel: ALPHA
4737 labels:
4738 - reason
4739- name: aggregator_unavailable_apiservice
4740 help: Gauge of APIServices which are marked as unavailable broken down by APIService
4741 name.
4742 type: Custom
4743 stabilityLevel: ALPHA
4744 labels:
4745 - name
4746- name: aggregator_unavailable_apiservice_total
4747 help: Counter of APIServices which are marked as unavailable broken down by APIService
4748 name and reason.
4749 type: Counter
4750 stabilityLevel: ALPHA
4751 labels:
4752 - name
4753 - reason
4754- name: x509_insecure_sha1_total
4755 subsystem: kube_aggregator
4756 namespace: apiserver
4757 help: Counts the number of requests to servers with insecure SHA1 signatures in
4758 their serving certificate OR the number of connection failures due to the insecure
4759 SHA1 signatures (either/or, based on the runtime environment)
4760 type: Counter
4761 stabilityLevel: ALPHA
4762- name: x509_missing_san_total
4763 subsystem: kube_aggregator
4764 namespace: apiserver
4765 help: Counts the number of requests to servers missing SAN extension in their serving
4766 certificate OR the number of connection failures due to the lack of x509 certificate
4767 SAN extension missing (either/or, based on the runtime environment)
4768 type: Counter
4769 stabilityLevel: ALPHA
4770- name: api_request_duration_seconds
4771 namespace: cloudprovider_azure
4772 help: Latency of an Azure API call
4773 type: Histogram
4774 stabilityLevel: ALPHA
4775 labels:
4776 - request
4777 - resource_group
4778 - source
4779 - subscription_id
4780 buckets:
4781 - 0.1
4782 - 0.25
4783 - 0.5
4784 - 1
4785 - 2.5
4786 - 5
4787 - 10
4788 - 15
4789 - 25
4790 - 50
4791 - 120
4792 - 300
4793 - 600
4794 - 1200
4795- name: api_request_errors
4796 namespace: cloudprovider_azure
4797 help: Number of errors for an Azure API call
4798 type: Counter
4799 stabilityLevel: ALPHA
4800 labels:
4801 - request
4802 - resource_group
4803 - source
4804 - subscription_id
4805- name: api_request_ratelimited_count
4806 namespace: cloudprovider_azure
4807 help: Number of rate limited Azure API calls
4808 type: Counter
4809 stabilityLevel: ALPHA
4810 labels:
4811 - request
4812 - resource_group
4813 - source
4814 - subscription_id
4815- name: api_request_throttled_count
4816 namespace: cloudprovider_azure
4817 help: Number of throttled Azure API calls
4818 type: Counter
4819 stabilityLevel: ALPHA
4820 labels:
4821 - request
4822 - resource_group
4823 - source
4824 - subscription_id
4825- name: op_duration_seconds
4826 namespace: cloudprovider_azure
4827 help: Latency of an Azure service operation
4828 type: Histogram
4829 stabilityLevel: ALPHA
4830 labels:
4831 - request
4832 - resource_group
4833 - source
4834 - subscription_id
4835 buckets:
4836 - 0.1
4837 - 0.2
4838 - 0.5
4839 - 1
4840 - 10
4841 - 20
4842 - 30
4843 - 40
4844 - 50
4845 - 60
4846 - 100
4847 - 200
4848 - 300
4849- name: op_failure_count
4850 namespace: cloudprovider_azure
4851 help: Number of failed Azure service operations
4852 type: Counter
4853 stabilityLevel: ALPHA
4854 labels:
4855 - request
4856 - resource_group
4857 - source
4858 - subscription_id
4859- name: cloudprovider_gce_api_request_duration_seconds
4860 help: Latency of a GCE API call
4861 type: Histogram
4862 stabilityLevel: ALPHA
4863 labels:
4864 - region
4865 - request
4866 - version
4867 - zone
4868- name: cloudprovider_gce_api_request_errors
4869 help: Number of errors for an API call
4870 type: Counter
4871 stabilityLevel: ALPHA
4872 labels:
4873 - region
4874 - request
4875 - version
4876 - zone
4877- name: cloudprovider_vsphere_api_request_duration_seconds
4878 help: Latency of vsphere api call
4879 type: Histogram
4880 stabilityLevel: ALPHA
4881 labels:
4882 - request
4883- name: cloudprovider_vsphere_api_request_errors
4884 help: vsphere Api errors
4885 type: Counter
4886 stabilityLevel: ALPHA
4887 labels:
4888 - request
4889- name: cloudprovider_vsphere_operation_duration_seconds
4890 help: Latency of vsphere operation call
4891 type: Histogram
4892 stabilityLevel: ALPHA
4893 labels:
4894 - operation
4895- name: cloudprovider_vsphere_operation_errors
4896 help: vsphere operation errors
4897 type: Counter
4898 stabilityLevel: ALPHA
4899 labels:
4900 - operation
4901- name: cloudprovider_vsphere_vcenter_versions
4902 help: Versions for connected vSphere vCenters
4903 type: Custom
4904 stabilityLevel: ALPHA
4905 labels:
4906 - hostname
4907 - version
4908 - build
4909- name: get_token_count
4910 help: Counter of total Token() requests to the alternate token source
4911 type: Counter
4912 stabilityLevel: ALPHA
4913- name: get_token_fail_count
4914 help: Counter of failed Token() requests to the alternate token source
4915 type: Counter
4916 stabilityLevel: ALPHA
4917- name: number_of_l4_ilbs
4918 help: Number of L4 ILBs
4919 type: Gauge
4920 stabilityLevel: ALPHA
4921 labels:
4922 - feature
4923- name: pod_security_errors_total
4924 help: Number of errors preventing normal evaluation. Non-fatal errors may result
4925 in the latest restricted profile being used for evaluation.
4926 type: Counter
4927 stabilityLevel: ALPHA
4928 labels:
4929 - fatal
4930 - request_operation
4931 - resource
4932 - subresource
4933- name: pod_security_evaluations_total
4934 help: Number of policy evaluations that occurred, not counting ignored or exempt
4935 requests.
4936 type: Counter
4937 stabilityLevel: ALPHA
4938 labels:
4939 - decision
4940 - mode
4941 - policy_level
4942 - policy_version
4943 - request_operation
4944 - resource
4945 - subresource
4946- name: pod_security_exemptions_total
4947 help: Number of exempt requests, not counting ignored or out of scope requests.
4948 type: Counter
4949 stabilityLevel: ALPHA
4950 labels:
4951 - request_operation
4952 - resource
4953 - subresource
View as plain text