...

Text file src/k8s.io/kubernetes/test/instrumentation/documentation/documentation-list.yaml

Documentation: k8s.io/kubernetes/test/instrumentation/documentation

     1- name: version_info
     2  namespace: etcd
     3  help: Etcd server's binary version
     4  type: Gauge
     5  stabilityLevel: ALPHA
     6  labels:
     7  - binary_version
     8- name: certificate_manager_client_ttl_seconds
     9  subsystem: kubelet
    10  help: Gauge of the TTL (time-to-live) of the Kubelet's client certificate. The value
    11    is in seconds until certificate expiry (negative if already expired). If client
    12    certificate is invalid or unused, the value will be +INF.
    13  type: Gauge
    14  stabilityLevel: ALPHA
    15- name: addresses_skipped_per_sync
    16  subsystem: endpoint_slice_mirroring_controller
    17  help: Number of addresses skipped on each Endpoints sync due to being invalid or
    18    exceeding MaxEndpointsPerSubset
    19  type: Histogram
    20  stabilityLevel: ALPHA
    21  buckets:
    22  - 2
    23  - 4
    24  - 8
    25  - 16
    26  - 32
    27  - 64
    28  - 128
    29  - 256
    30  - 512
    31  - 1024
    32  - 2048
    33  - 4096
    34  - 8192
    35  - 16384
    36  - 32768
    37- name: changes
    38  subsystem: endpoint_slice_mirroring_controller
    39  help: Number of EndpointSlice changes
    40  type: Counter
    41  stabilityLevel: ALPHA
    42  labels:
    43  - operation
    44- name: desired_endpoint_slices
    45  subsystem: endpoint_slice_mirroring_controller
    46  help: Number of EndpointSlices that would exist with perfect endpoint allocation
    47  type: Gauge
    48  stabilityLevel: ALPHA
    49- name: endpoints_added_per_sync
    50  subsystem: endpoint_slice_mirroring_controller
    51  help: Number of endpoints added on each Endpoints sync
    52  type: Histogram
    53  stabilityLevel: ALPHA
    54  buckets:
    55  - 2
    56  - 4
    57  - 8
    58  - 16
    59  - 32
    60  - 64
    61  - 128
    62  - 256
    63  - 512
    64  - 1024
    65  - 2048
    66  - 4096
    67  - 8192
    68  - 16384
    69  - 32768
    70- name: endpoints_desired
    71  subsystem: endpoint_slice_mirroring_controller
    72  help: Number of endpoints desired
    73  type: Gauge
    74  stabilityLevel: ALPHA
    75- name: endpoints_removed_per_sync
    76  subsystem: endpoint_slice_mirroring_controller
    77  help: Number of endpoints removed on each Endpoints sync
    78  type: Histogram
    79  stabilityLevel: ALPHA
    80  buckets:
    81  - 2
    82  - 4
    83  - 8
    84  - 16
    85  - 32
    86  - 64
    87  - 128
    88  - 256
    89  - 512
    90  - 1024
    91  - 2048
    92  - 4096
    93  - 8192
    94  - 16384
    95  - 32768
    96- name: endpoints_sync_duration
    97  subsystem: endpoint_slice_mirroring_controller
    98  help: Duration of syncEndpoints() in seconds
    99  type: Histogram
   100  stabilityLevel: ALPHA
   101  buckets:
   102  - 0.001
   103  - 0.002
   104  - 0.004
   105  - 0.008
   106  - 0.016
   107  - 0.032
   108  - 0.064
   109  - 0.128
   110  - 0.256
   111  - 0.512
   112  - 1.024
   113  - 2.048
   114  - 4.096
   115  - 8.192
   116  - 16.384
   117- name: endpoints_updated_per_sync
   118  subsystem: endpoint_slice_mirroring_controller
   119  help: Number of endpoints updated on each Endpoints sync
   120  type: Histogram
   121  stabilityLevel: ALPHA
   122  buckets:
   123  - 2
   124  - 4
   125  - 8
   126  - 16
   127  - 32
   128  - 64
   129  - 128
   130  - 256
   131  - 512
   132  - 1024
   133  - 2048
   134  - 4096
   135  - 8192
   136  - 16384
   137  - 32768
   138- name: num_endpoint_slices
   139  subsystem: endpoint_slice_mirroring_controller
   140  help: Number of EndpointSlices
   141  type: Gauge
   142  stabilityLevel: ALPHA
   143- name: sync_duration_seconds
   144  subsystem: root_ca_cert_publisher
   145  help: Number of namespace syncs happened in root ca cert publisher.
   146  type: Histogram
   147  stabilityLevel: ALPHA
   148  labels:
   149  - code
   150  buckets:
   151  - 0.001
   152  - 0.002
   153  - 0.004
   154  - 0.008
   155  - 0.016
   156  - 0.032
   157  - 0.064
   158  - 0.128
   159  - 0.256
   160  - 0.512
   161  - 1.024
   162  - 2.048
   163  - 4.096
   164  - 8.192
   165  - 16.384
   166- name: sync_total
   167  subsystem: root_ca_cert_publisher
   168  help: Number of namespace syncs happened in root ca cert publisher.
   169  type: Counter
   170  stabilityLevel: ALPHA
   171  labels:
   172  - code
   173- name: job_creation_skew_duration_seconds
   174  subsystem: cronjob_controller
   175  help: Time between when a cronjob is scheduled to be run, and when the corresponding
   176    job is created
   177  type: Histogram
   178  stabilityLevel: STABLE
   179  buckets:
   180  - 1
   181  - 2
   182  - 4
   183  - 8
   184  - 16
   185  - 32
   186  - 64
   187  - 128
   188  - 256
   189  - 512
   190- name: resources_sync_error_total
   191  subsystem: garbagecollector_controller
   192  help: Number of garbage collector resources sync errors
   193  type: Counter
   194  stabilityLevel: ALPHA
   195- name: metric_computation_duration_seconds
   196  subsystem: horizontal_pod_autoscaler_controller
   197  help: The time(seconds) that the HPA controller takes to calculate one metric. The
   198    label 'action' should be either 'scale_down', 'scale_up', or 'none'. The label
   199    'error' should be either 'spec', 'internal', or 'none'. The label 'metric_type'
   200    corresponds to HPA.spec.metrics[*].type
   201  type: Histogram
   202  stabilityLevel: ALPHA
   203  labels:
   204  - action
   205  - error
   206  - metric_type
   207  buckets:
   208  - 0.001
   209  - 0.002
   210  - 0.004
   211  - 0.008
   212  - 0.016
   213  - 0.032
   214  - 0.064
   215  - 0.128
   216  - 0.256
   217  - 0.512
   218  - 1.024
   219  - 2.048
   220  - 4.096
   221  - 8.192
   222  - 16.384
   223- name: metric_computation_total
   224  subsystem: horizontal_pod_autoscaler_controller
   225  help: Number of metric computations. The label 'action' should be either 'scale_down',
   226    'scale_up', or 'none'. Also, the label 'error' should be either 'spec', 'internal',
   227    or 'none'. The label 'metric_type' corresponds to HPA.spec.metrics[*].type
   228  type: Counter
   229  stabilityLevel: ALPHA
   230  labels:
   231  - action
   232  - error
   233  - metric_type
   234- name: reconciliation_duration_seconds
   235  subsystem: horizontal_pod_autoscaler_controller
   236  help: The time(seconds) that the HPA controller takes to reconcile once. The label
   237    'action' should be either 'scale_down', 'scale_up', or 'none'. Also, the label
   238    'error' should be either 'spec', 'internal', or 'none'. Note that if both spec
   239    and internal errors happen during a reconciliation, the first one to occur is
   240    reported in `error` label.
   241  type: Histogram
   242  stabilityLevel: ALPHA
   243  labels:
   244  - action
   245  - error
   246  buckets:
   247  - 0.001
   248  - 0.002
   249  - 0.004
   250  - 0.008
   251  - 0.016
   252  - 0.032
   253  - 0.064
   254  - 0.128
   255  - 0.256
   256  - 0.512
   257  - 1.024
   258  - 2.048
   259  - 4.096
   260  - 8.192
   261  - 16.384
   262- name: reconciliations_total
   263  subsystem: horizontal_pod_autoscaler_controller
   264  help: Number of reconciliations of HPA controller. The label 'action' should be
   265    either 'scale_down', 'scale_up', or 'none'. Also, the label 'error' should be
   266    either 'spec', 'internal', or 'none'. Note that if both spec and internal errors
   267    happen during a reconciliation, the first one to occur is reported in `error`
   268    label.
   269  type: Counter
   270  stabilityLevel: ALPHA
   271  labels:
   272  - action
   273  - error
   274- name: job_finished_indexes_total
   275  subsystem: job_controller
   276  help: "`The number of finished indexes. Possible values for the\n\t\t\tstatus label
   277    are: \"succeeded\", \"failed\". Possible values for the\n\t\t\tbackoffLimit label
   278    are: \"perIndex\" and \"global\"`"
   279  type: Counter
   280  stabilityLevel: ALPHA
   281  labels:
   282  - backoffLimit
   283  - status
   284- name: job_pods_creation_total
   285  subsystem: job_controller
   286  help: |-
   287    `The number of Pods created by the Job controller labelled with a reason for the Pod creation.
   288    This metric also distinguishes between Pods created using different PodReplacementPolicy settings.
   289    Possible values of the "reason" label are:
   290    "new", "recreate_terminating_or_failed", "recreate_failed".
   291    Possible values of the "status" label are:
   292    "succeeded", "failed".`
   293  type: Counter
   294  stabilityLevel: ALPHA
   295  labels:
   296  - reason
   297  - status
   298- name: pod_failures_handled_by_failure_policy_total
   299  subsystem: job_controller
   300  help: "`The number of failed Pods handled by failure policy with\n\t\t\trespect
   301    to the failure policy action applied based on the matched\n\t\t\trule. Possible
   302    values of the action label correspond to the\n\t\t\tpossible values for the failure
   303    policy rule action, which are:\n\t\t\t\"FailJob\", \"Ignore\" and \"Count\".`"
   304  type: Counter
   305  stabilityLevel: ALPHA
   306  labels:
   307  - action
   308- name: terminated_pods_tracking_finalizer_total
   309  subsystem: job_controller
   310  help: |-
   311    `The number of terminated pods (phase=Failed|Succeeded)
   312    that have the finalizer batch.kubernetes.io/job-tracking
   313    The event label can be "add" or "delete".`
   314  type: Counter
   315  stabilityLevel: ALPHA
   316  labels:
   317  - event
   318- name: unhealthy_nodes_in_zone
   319  subsystem: node_collector
   320  help: Gauge measuring number of not Ready Nodes per zones.
   321  type: Gauge
   322  stabilityLevel: ALPHA
   323  labels:
   324  - zone
   325- name: update_all_nodes_health_duration_seconds
   326  subsystem: node_collector
   327  help: Duration in seconds for NodeController to update the health of all nodes.
   328  type: Histogram
   329  stabilityLevel: ALPHA
   330  buckets:
   331  - 0.01
   332  - 0.04
   333  - 0.16
   334  - 0.64
   335  - 2.56
   336  - 10.24
   337  - 40.96
   338  - 163.84
   339- name: update_node_health_duration_seconds
   340  subsystem: node_collector
   341  help: Duration in seconds for NodeController to update the health of a single node.
   342  type: Histogram
   343  stabilityLevel: ALPHA
   344  buckets:
   345  - 0.001
   346  - 0.004
   347  - 0.016
   348  - 0.064
   349  - 0.256
   350  - 1.024
   351  - 4.096
   352  - 16.384
   353- name: zone_health
   354  subsystem: node_collector
   355  help: Gauge measuring percentage of healthy nodes per zone.
   356  type: Gauge
   357  stabilityLevel: ALPHA
   358  labels:
   359  - zone
   360- name: zone_size
   361  subsystem: node_collector
   362  help: Gauge measuring number of registered Nodes per zones.
   363  type: Gauge
   364  stabilityLevel: ALPHA
   365  labels:
   366  - zone
   367- name: cidrset_allocation_tries_per_request
   368  subsystem: node_ipam_controller
   369  help: Number of endpoints added on each Service sync
   370  type: Histogram
   371  stabilityLevel: ALPHA
   372  labels:
   373  - clusterCIDR
   374  buckets:
   375  - 1
   376  - 5
   377  - 25
   378  - 125
   379  - 625
   380- name: cidrset_cidrs_allocations_total
   381  subsystem: node_ipam_controller
   382  help: Counter measuring total number of CIDR allocations.
   383  type: Counter
   384  stabilityLevel: ALPHA
   385  labels:
   386  - clusterCIDR
   387- name: cidrset_cidrs_releases_total
   388  subsystem: node_ipam_controller
   389  help: Counter measuring total number of CIDR releases.
   390  type: Counter
   391  stabilityLevel: ALPHA
   392  labels:
   393  - clusterCIDR
   394- name: cidrset_usage_cidrs
   395  subsystem: node_ipam_controller
   396  help: Gauge measuring percentage of allocated CIDRs.
   397  type: Gauge
   398  stabilityLevel: ALPHA
   399  labels:
   400  - clusterCIDR
   401- name: cirdset_max_cidrs
   402  subsystem: node_ipam_controller
   403  help: Maximum number of CIDRs that can be allocated.
   404  type: Gauge
   405  stabilityLevel: ALPHA
   406  labels:
   407  - clusterCIDR
   408- name: force_delete_pod_errors_total
   409  subsystem: pod_gc_collector
   410  help: Number of errors encountered when forcefully deleting the pods since the Pod
   411    GC Controller started.
   412  type: Counter
   413  stabilityLevel: ALPHA
   414  labels:
   415  - namespace
   416  - reason
   417- name: force_delete_pods_total
   418  subsystem: pod_gc_collector
   419  help: Number of pods that are being forcefully deleted since the Pod GC Controller
   420    started.
   421  type: Counter
   422  stabilityLevel: ALPHA
   423  labels:
   424  - namespace
   425  - reason
   426- name: sorting_deletion_age_ratio
   427  subsystem: replicaset_controller
   428  help: The ratio of chosen deleted pod's ages to the current youngest pod's age (at
   429    the time). Should be <2. The intent of this metric is to measure the rough efficacy
   430    of the LogarithmicScaleDown feature gate's effect on the sorting (and deletion)
   431    of pods when a replicaset scales down. This only considers Ready pods when calculating
   432    and reporting.
   433  type: Histogram
   434  stabilityLevel: ALPHA
   435  buckets:
   436  - 0.25
   437  - 0.5
   438  - 1
   439  - 2
   440  - 4
   441  - 8
   442- name: create_attempts_total
   443  subsystem: resourceclaim_controller
   444  help: Number of ResourceClaims creation requests
   445  type: Counter
   446  stabilityLevel: ALPHA
   447- name: create_failures_total
   448  subsystem: resourceclaim_controller
   449  help: Number of ResourceClaims creation request failures
   450  type: Counter
   451  stabilityLevel: ALPHA
   452- name: pod_deletion_duration_seconds
   453  subsystem: taint_eviction_controller
   454  help: Latency, in seconds, between the time when a taint effect has been activated
   455    for the Pod and its deletion via TaintEvictionController.
   456  type: Histogram
   457  stabilityLevel: ALPHA
   458  buckets:
   459  - 0.005
   460  - 0.025
   461  - 0.1
   462  - 0.5
   463  - 1
   464  - 2.5
   465  - 10
   466  - 30
   467  - 60
   468  - 120
   469  - 180
   470  - 240
   471- name: pod_deletions_total
   472  subsystem: taint_eviction_controller
   473  help: Total number of Pods deleted by TaintEvictionController since its start.
   474  type: Counter
   475  stabilityLevel: ALPHA
   476- name: job_pods_finished_total
   477  subsystem: job_controller
   478  help: The number of finished Pods that are fully tracked
   479  type: Counter
   480  stabilityLevel: STABLE
   481  labels:
   482  - completion_mode
   483  - result
   484- name: job_sync_duration_seconds
   485  subsystem: job_controller
   486  help: The time it took to sync a job
   487  type: Histogram
   488  stabilityLevel: STABLE
   489  labels:
   490  - action
   491  - completion_mode
   492  - result
   493  buckets:
   494  - 0.004
   495  - 0.008
   496  - 0.016
   497  - 0.032
   498  - 0.064
   499  - 0.128
   500  - 0.256
   501  - 0.512
   502  - 1.024
   503  - 2.048
   504  - 4.096
   505  - 8.192
   506  - 16.384
   507  - 32.768
   508  - 65.536
   509- name: job_syncs_total
   510  subsystem: job_controller
   511  help: The number of job syncs
   512  type: Counter
   513  stabilityLevel: STABLE
   514  labels:
   515  - action
   516  - completion_mode
   517  - result
   518- name: jobs_finished_total
   519  subsystem: job_controller
   520  help: The number of finished jobs
   521  type: Counter
   522  stabilityLevel: STABLE
   523  labels:
   524  - completion_mode
   525  - reason
   526  - result
   527- name: evictions_total
   528  subsystem: node_collector
   529  help: Number of Node evictions that happened since current instance of NodeController
   530    started.
   531  type: Counter
   532  stabilityLevel: STABLE
   533  labels:
   534  - zone
   535- name: attachdetach_controller_forced_detaches
   536  subsystem: attach_detach_controller
   537  help: Number of times the A/D Controller performed a forced detach
   538  type: Counter
   539  stabilityLevel: ALPHA
   540  labels:
   541  - reason
   542- name: attachdetach_controller_total_volumes
   543  help: Number of volumes in A/D Controller
   544  type: Custom
   545  stabilityLevel: ALPHA
   546  labels:
   547  - plugin_name
   548  - state
   549- name: create_failures_total
   550  subsystem: ephemeral_volume_controller
   551  help: Number of PersistenVolumeClaims creation requests
   552  type: Counter
   553  stabilityLevel: ALPHA
   554- name: create_total
   555  subsystem: ephemeral_volume_controller
   556  help: Number of PersistenVolumeClaims creation requests
   557  type: Counter
   558  stabilityLevel: ALPHA
   559- name: client_expiration_renew_errors
   560  subsystem: certificate_manager
   561  namespace: kubelet
   562  help: Counter of certificate renewal errors.
   563  type: Counter
   564  stabilityLevel: ALPHA
   565- name: certificate_manager_server_rotation_seconds
   566  subsystem: kubelet
   567  help: Histogram of the number of seconds the previous certificate lived before being
   568    rotated.
   569  type: Histogram
   570  stabilityLevel: ALPHA
   571  buckets:
   572  - 60
   573  - 3600
   574  - 14400
   575  - 86400
   576  - 604800
   577  - 2.592e+06
   578  - 7.776e+06
   579  - 1.5552e+07
   580  - 3.1104e+07
   581  - 1.24416e+08
   582- name: certificate_manager_server_ttl_seconds
   583  subsystem: kubelet
   584  help: Gauge of the shortest TTL (time-to-live) of the Kubelet's serving certificate.
   585    The value is in seconds until certificate expiry (negative if already expired).
   586    If serving certificate is invalid or unused, the value will be +INF.
   587  type: Gauge
   588  stabilityLevel: ALPHA
   589- name: credential_provider_plugin_duration
   590  subsystem: kubelet
   591  help: Duration of execution in seconds for credential provider plugin
   592  type: Histogram
   593  stabilityLevel: ALPHA
   594  labels:
   595  - plugin_name
   596  buckets:
   597  - 0.005
   598  - 0.01
   599  - 0.025
   600  - 0.05
   601  - 0.1
   602  - 0.25
   603  - 0.5
   604  - 1
   605  - 2.5
   606  - 5
   607  - 10
   608- name: credential_provider_plugin_errors
   609  subsystem: kubelet
   610  help: Number of errors from credential provider plugin
   611  type: Counter
   612  stabilityLevel: ALPHA
   613  labels:
   614  - plugin_name
   615- name: server_expiration_renew_errors
   616  subsystem: kubelet
   617  help: Counter of certificate renewal errors.
   618  type: Counter
   619  stabilityLevel: ALPHA
   620- name: pv_collector_bound_pv_count
   621  help: Gauge measuring number of persistent volume currently bound
   622  type: Custom
   623  stabilityLevel: ALPHA
   624  labels:
   625  - storage_class
   626- name: pv_collector_bound_pvc_count
   627  help: Gauge measuring number of persistent volume claim currently bound
   628  type: Custom
   629  stabilityLevel: ALPHA
   630  labels:
   631  - namespace
   632- name: pv_collector_total_pv_count
   633  help: Gauge measuring total number of persistent volumes
   634  type: Custom
   635  stabilityLevel: ALPHA
   636  labels:
   637  - plugin_name
   638  - volume_mode
   639- name: pv_collector_unbound_pv_count
   640  help: Gauge measuring number of persistent volume currently unbound
   641  type: Custom
   642  stabilityLevel: ALPHA
   643  labels:
   644  - storage_class
   645- name: pv_collector_unbound_pvc_count
   646  help: Gauge measuring number of persistent volume claim currently unbound
   647  type: Custom
   648  stabilityLevel: ALPHA
   649  labels:
   650  - namespace
   651- name: retroactive_storageclass_errors_total
   652  help: Total number of failed retroactive StorageClass assignments to persistent
   653    volume claim
   654  type: Counter
   655  stabilityLevel: ALPHA
   656- name: retroactive_storageclass_total
   657  help: Total number of retroactive StorageClass assignments to persistent volume
   658    claim
   659  type: Counter
   660  stabilityLevel: ALPHA
   661- name: storage_count_attachable_volumes_in_use
   662  help: Measure number of volumes in use
   663  type: Custom
   664  stabilityLevel: ALPHA
   665  labels:
   666  - node
   667  - volume_plugin
   668- name: job_deletion_duration_seconds
   669  subsystem: ttl_after_finished_controller
   670  help: The time it took to delete the job since it became eligible for deletion
   671  type: Histogram
   672  stabilityLevel: ALPHA
   673  buckets:
   674  - 0.1
   675  - 0.2
   676  - 0.4
   677  - 0.8
   678  - 1.6
   679  - 3.2
   680  - 6.4
   681  - 12.8
   682  - 25.6
   683  - 51.2
   684  - 102.4
   685  - 204.8
   686  - 409.6
   687  - 819.2
   688- name: volume_operation_total_errors
   689  help: Total volume operation errors
   690  type: Counter
   691  stabilityLevel: ALPHA
   692  labels:
   693  - operation_name
   694  - plugin_name
   695- name: container_swap_usage_bytes
   696  help: Current amount of the container swap usage in bytes. Reported only on non-windows
   697    systems
   698  type: Custom
   699  stabilityLevel: ALPHA
   700  labels:
   701  - container
   702  - pod
   703  - namespace
   704- name: active_pods
   705  subsystem: kubelet
   706  help: The number of pods the kubelet considers active and which are being considered
   707    when admitting new pods. static is true if the pod is not from the apiserver.
   708  type: Gauge
   709  stabilityLevel: ALPHA
   710  labels:
   711  - static
   712- name: cgroup_manager_duration_seconds
   713  subsystem: kubelet
   714  help: Duration in seconds for cgroup manager operations. Broken down by method.
   715  type: Histogram
   716  stabilityLevel: ALPHA
   717  labels:
   718  - operation_type
   719  buckets:
   720  - 0.005
   721  - 0.01
   722  - 0.025
   723  - 0.05
   724  - 0.1
   725  - 0.25
   726  - 0.5
   727  - 1
   728  - 2.5
   729  - 5
   730  - 10
   731- name: kubelet_container_log_filesystem_used_bytes
   732  help: Bytes used by the container's logs on the filesystem.
   733  type: Custom
   734  stabilityLevel: ALPHA
   735  labels:
   736  - uid
   737  - namespace
   738  - pod
   739  - container
   740- name: containers_per_pod_count
   741  subsystem: kubelet
   742  help: The number of containers per pod.
   743  type: Histogram
   744  stabilityLevel: ALPHA
   745  buckets:
   746  - 1
   747  - 2
   748  - 4
   749  - 8
   750  - 16
   751- name: cpu_manager_pinning_errors_total
   752  subsystem: kubelet
   753  help: The number of cpu core allocations which required pinning failed.
   754  type: Counter
   755  stabilityLevel: ALPHA
   756- name: cpu_manager_pinning_requests_total
   757  subsystem: kubelet
   758  help: The number of cpu core allocations which required pinning.
   759  type: Counter
   760  stabilityLevel: ALPHA
   761- name: desired_pods
   762  subsystem: kubelet
   763  help: The number of pods the kubelet is being instructed to run. static is true
   764    if the pod is not from the apiserver.
   765  type: Gauge
   766  stabilityLevel: ALPHA
   767  labels:
   768  - static
   769- name: device_plugin_alloc_duration_seconds
   770  subsystem: kubelet
   771  help: Duration in seconds to serve a device plugin Allocation request. Broken down
   772    by resource name.
   773  type: Histogram
   774  stabilityLevel: ALPHA
   775  labels:
   776  - resource_name
   777  buckets:
   778  - 0.005
   779  - 0.01
   780  - 0.025
   781  - 0.05
   782  - 0.1
   783  - 0.25
   784  - 0.5
   785  - 1
   786  - 2.5
   787  - 5
   788  - 10
   789- name: device_plugin_registration_total
   790  subsystem: kubelet
   791  help: Cumulative number of device plugin registrations. Broken down by resource
   792    name.
   793  type: Counter
   794  stabilityLevel: ALPHA
   795  labels:
   796  - resource_name
   797- name: evented_pleg_connection_error_count
   798  subsystem: kubelet
   799  help: The number of errors encountered during the establishment of streaming connection
   800    with the CRI runtime.
   801  type: Counter
   802  stabilityLevel: ALPHA
   803- name: evented_pleg_connection_latency_seconds
   804  subsystem: kubelet
   805  help: The latency of streaming connection with the CRI runtime, measured in seconds.
   806  type: Histogram
   807  stabilityLevel: ALPHA
   808  buckets:
   809  - 0.005
   810  - 0.01
   811  - 0.025
   812  - 0.05
   813  - 0.1
   814  - 0.25
   815  - 0.5
   816  - 1
   817  - 2.5
   818  - 5
   819  - 10
   820- name: evented_pleg_connection_success_count
   821  subsystem: kubelet
   822  help: The number of times a streaming client was obtained to receive CRI Events.
   823  type: Counter
   824  stabilityLevel: ALPHA
   825- name: eviction_stats_age_seconds
   826  subsystem: kubelet
   827  help: Time between when stats are collected, and when pod is evicted based on those
   828    stats by eviction signal
   829  type: Histogram
   830  stabilityLevel: ALPHA
   831  labels:
   832  - eviction_signal
   833  buckets:
   834  - 0.005
   835  - 0.01
   836  - 0.025
   837  - 0.05
   838  - 0.1
   839  - 0.25
   840  - 0.5
   841  - 1
   842  - 2.5
   843  - 5
   844  - 10
   845- name: evictions
   846  subsystem: kubelet
   847  help: Cumulative number of pod evictions by eviction signal
   848  type: Counter
   849  stabilityLevel: ALPHA
   850  labels:
   851  - eviction_signal
   852- name: graceful_shutdown_end_time_seconds
   853  subsystem: kubelet
   854  help: Last graceful shutdown start time since unix epoch in seconds
   855  type: Gauge
   856  stabilityLevel: ALPHA
   857- name: graceful_shutdown_start_time_seconds
   858  subsystem: kubelet
   859  help: Last graceful shutdown start time since unix epoch in seconds
   860  type: Gauge
   861  stabilityLevel: ALPHA
   862- name: image_garbage_collected_total
   863  subsystem: kubelet
   864  help: Total number of images garbage collected by the kubelet, whether through disk
   865    usage or image age.
   866  type: Counter
   867  stabilityLevel: ALPHA
   868- name: lifecycle_handler_http_fallbacks_total
   869  subsystem: kubelet
   870  help: The number of times lifecycle handlers successfully fell back to http from
   871    https.
   872  type: Counter
   873  stabilityLevel: ALPHA
   874- name: managed_ephemeral_containers
   875  subsystem: kubelet
   876  help: Current number of ephemeral containers in pods managed by this kubelet.
   877  type: Gauge
   878  stabilityLevel: ALPHA
   879- name: mirror_pods
   880  subsystem: kubelet
   881  help: The number of mirror pods the kubelet will try to create (one per admitted
   882    static pod)
   883  type: Gauge
   884  stabilityLevel: ALPHA
   885- name: node_name
   886  subsystem: kubelet
   887  help: The node's name. The count is always 1.
   888  type: Gauge
   889  stabilityLevel: ALPHA
   890  labels:
   891  - node
   892- name: node_startup_duration_seconds
   893  subsystem: kubelet
   894  help: Duration in seconds of node startup in total.
   895  type: Gauge
   896  stabilityLevel: ALPHA
   897- name: node_startup_post_registration_duration_seconds
   898  subsystem: kubelet
   899  help: Duration in seconds of node startup after registration.
   900  type: Gauge
   901  stabilityLevel: ALPHA
   902- name: node_startup_pre_kubelet_duration_seconds
   903  subsystem: kubelet
   904  help: Duration in seconds of node startup before kubelet starts.
   905  type: Gauge
   906  stabilityLevel: ALPHA
   907- name: node_startup_pre_registration_duration_seconds
   908  subsystem: kubelet
   909  help: Duration in seconds of node startup before registration.
   910  type: Gauge
   911  stabilityLevel: ALPHA
   912- name: node_startup_registration_duration_seconds
   913  subsystem: kubelet
   914  help: Duration in seconds of node startup during registration.
   915  type: Gauge
   916  stabilityLevel: ALPHA
   917- name: orphan_pod_cleaned_volumes
   918  subsystem: kubelet
   919  help: The total number of orphaned Pods whose volumes were cleaned in the last periodic
   920    sweep.
   921  type: Gauge
   922  stabilityLevel: ALPHA
   923- name: orphan_pod_cleaned_volumes_errors
   924  subsystem: kubelet
   925  help: The number of orphaned Pods whose volumes failed to be cleaned in the last
   926    periodic sweep.
   927  type: Gauge
   928  stabilityLevel: ALPHA
   929- name: orphaned_runtime_pods_total
   930  subsystem: kubelet
   931  help: Number of pods that have been detected in the container runtime without being
   932    already known to the pod worker. This typically indicates the kubelet was restarted
   933    while a pod was force deleted in the API or in the local configuration, which
   934    is unusual.
   935  type: Counter
   936  stabilityLevel: ALPHA
   937- name: pleg_discard_events
   938  subsystem: kubelet
   939  help: The number of discard events in PLEG.
   940  type: Counter
   941  stabilityLevel: ALPHA
   942- name: pleg_last_seen_seconds
   943  subsystem: kubelet
   944  help: Timestamp in seconds when PLEG was last seen active.
   945  type: Gauge
   946  stabilityLevel: ALPHA
   947- name: pleg_relist_duration_seconds
   948  subsystem: kubelet
   949  help: Duration in seconds for relisting pods in PLEG.
   950  type: Histogram
   951  stabilityLevel: ALPHA
   952  buckets:
   953  - 0.005
   954  - 0.01
   955  - 0.025
   956  - 0.05
   957  - 0.1
   958  - 0.25
   959  - 0.5
   960  - 1
   961  - 2.5
   962  - 5
   963  - 10
   964- name: pleg_relist_interval_seconds
   965  subsystem: kubelet
   966  help: Interval in seconds between relisting in PLEG.
   967  type: Histogram
   968  stabilityLevel: ALPHA
   969  buckets:
   970  - 0.005
   971  - 0.01
   972  - 0.025
   973  - 0.05
   974  - 0.1
   975  - 0.25
   976  - 0.5
   977  - 1
   978  - 2.5
   979  - 5
   980  - 10
   981- name: pod_resources_endpoint_errors_get
   982  subsystem: kubelet
   983  help: Number of requests to the PodResource Get endpoint which returned error. Broken
   984    down by server api version.
   985  type: Counter
   986  stabilityLevel: ALPHA
   987  labels:
   988  - server_api_version
   989- name: pod_resources_endpoint_errors_get_allocatable
   990  subsystem: kubelet
   991  help: Number of requests to the PodResource GetAllocatableResources endpoint which
   992    returned error. Broken down by server api version.
   993  type: Counter
   994  stabilityLevel: ALPHA
   995  labels:
   996  - server_api_version
   997- name: pod_resources_endpoint_errors_list
   998  subsystem: kubelet
   999  help: Number of requests to the PodResource List endpoint which returned error.
  1000    Broken down by server api version.
  1001  type: Counter
  1002  stabilityLevel: ALPHA
  1003  labels:
  1004  - server_api_version
  1005- name: pod_resources_endpoint_requests_get
  1006  subsystem: kubelet
  1007  help: Number of requests to the PodResource Get endpoint. Broken down by server
  1008    api version.
  1009  type: Counter
  1010  stabilityLevel: ALPHA
  1011  labels:
  1012  - server_api_version
  1013- name: pod_resources_endpoint_requests_get_allocatable
  1014  subsystem: kubelet
  1015  help: Number of requests to the PodResource GetAllocatableResources endpoint. Broken
  1016    down by server api version.
  1017  type: Counter
  1018  stabilityLevel: ALPHA
  1019  labels:
  1020  - server_api_version
  1021- name: pod_resources_endpoint_requests_list
  1022  subsystem: kubelet
  1023  help: Number of requests to the PodResource List endpoint. Broken down by server
  1024    api version.
  1025  type: Counter
  1026  stabilityLevel: ALPHA
  1027  labels:
  1028  - server_api_version
  1029- name: pod_resources_endpoint_requests_total
  1030  subsystem: kubelet
  1031  help: Cumulative number of requests to the PodResource endpoint. Broken down by
  1032    server api version.
  1033  type: Counter
  1034  stabilityLevel: ALPHA
  1035  labels:
  1036  - server_api_version
  1037- name: pod_start_duration_seconds
  1038  subsystem: kubelet
  1039  help: Duration in seconds from kubelet seeing a pod for the first time to the pod
  1040    starting to run
  1041  type: Histogram
  1042  stabilityLevel: ALPHA
  1043  buckets:
  1044  - 0.5
  1045  - 1
  1046  - 2
  1047  - 3
  1048  - 4
  1049  - 5
  1050  - 6
  1051  - 8
  1052  - 10
  1053  - 20
  1054  - 30
  1055  - 45
  1056  - 60
  1057  - 120
  1058  - 180
  1059  - 240
  1060  - 300
  1061  - 360
  1062  - 480
  1063  - 600
  1064  - 900
  1065  - 1200
  1066  - 1800
  1067  - 2700
  1068  - 3600
  1069- name: pod_start_sli_duration_seconds
  1070  subsystem: kubelet
  1071  help: Duration in seconds to start a pod, excluding time to pull images and run
  1072    init containers, measured from pod creation timestamp to when all its containers
  1073    are reported as started and observed via watch
  1074  type: Histogram
  1075  stabilityLevel: ALPHA
  1076  buckets:
  1077  - 0.5
  1078  - 1
  1079  - 2
  1080  - 3
  1081  - 4
  1082  - 5
  1083  - 6
  1084  - 8
  1085  - 10
  1086  - 20
  1087  - 30
  1088  - 45
  1089  - 60
  1090  - 120
  1091  - 180
  1092  - 240
  1093  - 300
  1094  - 360
  1095  - 480
  1096  - 600
  1097  - 900
  1098  - 1200
  1099  - 1800
  1100  - 2700
  1101  - 3600
  1102- name: pod_start_total_duration_seconds
  1103  subsystem: kubelet
  1104  help: Duration in seconds to start a pod since creation, including time to pull
  1105    images and run init containers, measured from pod creation timestamp to when all
  1106    its containers are reported as started and observed via watch
  1107  type: Histogram
  1108  stabilityLevel: ALPHA
  1109  buckets:
  1110  - 0.5
  1111  - 1
  1112  - 2
  1113  - 3
  1114  - 4
  1115  - 5
  1116  - 6
  1117  - 8
  1118  - 10
  1119  - 20
  1120  - 30
  1121  - 45
  1122  - 60
  1123  - 120
  1124  - 180
  1125  - 240
  1126  - 300
  1127  - 360
  1128  - 480
  1129  - 600
  1130  - 900
  1131  - 1200
  1132  - 1800
  1133  - 2700
  1134  - 3600
  1135- name: pod_status_sync_duration_seconds
  1136  subsystem: kubelet
  1137  help: Duration in seconds to sync a pod status update. Measures time from detection
  1138    of a change to pod status until the API is successfully updated for that pod,
  1139    even if multiple intevening changes to pod status occur.
  1140  type: Histogram
  1141  stabilityLevel: ALPHA
  1142  buckets:
  1143  - 0.01
  1144  - 0.05
  1145  - 0.1
  1146  - 0.5
  1147  - 1
  1148  - 5
  1149  - 10
  1150  - 20
  1151  - 30
  1152  - 45
  1153  - 60
  1154- name: pod_worker_duration_seconds
  1155  subsystem: kubelet
  1156  help: 'Duration in seconds to sync a single pod. Broken down by operation type:
  1157    create, update, or sync'
  1158  type: Histogram
  1159  stabilityLevel: ALPHA
  1160  labels:
  1161  - operation_type
  1162  buckets:
  1163  - 0.005
  1164  - 0.01
  1165  - 0.025
  1166  - 0.05
  1167  - 0.1
  1168  - 0.25
  1169  - 0.5
  1170  - 1
  1171  - 2.5
  1172  - 5
  1173  - 10
  1174- name: pod_worker_start_duration_seconds
  1175  subsystem: kubelet
  1176  help: Duration in seconds from kubelet seeing a pod to starting a worker.
  1177  type: Histogram
  1178  stabilityLevel: ALPHA
  1179  buckets:
  1180  - 0.005
  1181  - 0.01
  1182  - 0.025
  1183  - 0.05
  1184  - 0.1
  1185  - 0.25
  1186  - 0.5
  1187  - 1
  1188  - 2.5
  1189  - 5
  1190  - 10
  1191- name: preemptions
  1192  subsystem: kubelet
  1193  help: Cumulative number of pod preemptions by preemption resource
  1194  type: Counter
  1195  stabilityLevel: ALPHA
  1196  labels:
  1197  - preemption_signal
  1198- name: restarted_pods_total
  1199  subsystem: kubelet
  1200  help: Number of pods that have been restarted because they were deleted and recreated
  1201    with the same UID while the kubelet was watching them (common for static pods,
  1202    extremely uncommon for API pods)
  1203  type: Counter
  1204  stabilityLevel: ALPHA
  1205  labels:
  1206  - static
  1207- name: run_podsandbox_duration_seconds
  1208  subsystem: kubelet
  1209  help: Duration in seconds of the run_podsandbox operations. Broken down by RuntimeClass.Handler.
  1210  type: Histogram
  1211  stabilityLevel: ALPHA
  1212  labels:
  1213  - runtime_handler
  1214  buckets:
  1215  - 0.005
  1216  - 0.01
  1217  - 0.025
  1218  - 0.05
  1219  - 0.1
  1220  - 0.25
  1221  - 0.5
  1222  - 1
  1223  - 2.5
  1224  - 5
  1225  - 10
  1226- name: run_podsandbox_errors_total
  1227  subsystem: kubelet
  1228  help: Cumulative number of the run_podsandbox operation errors by RuntimeClass.Handler.
  1229  type: Counter
  1230  stabilityLevel: ALPHA
  1231  labels:
  1232  - runtime_handler
  1233- name: running_containers
  1234  subsystem: kubelet
  1235  help: Number of containers currently running
  1236  type: Gauge
  1237  stabilityLevel: ALPHA
  1238  labels:
  1239  - container_state
  1240- name: running_pods
  1241  subsystem: kubelet
  1242  help: Number of pods that have a running pod sandbox
  1243  type: Gauge
  1244  stabilityLevel: ALPHA
  1245- name: runtime_operations_duration_seconds
  1246  subsystem: kubelet
  1247  help: Duration in seconds of runtime operations. Broken down by operation type.
  1248  type: Histogram
  1249  stabilityLevel: ALPHA
  1250  labels:
  1251  - operation_type
  1252  buckets:
  1253  - 0.005
  1254  - 0.0125
  1255  - 0.03125
  1256  - 0.078125
  1257  - 0.1953125
  1258  - 0.48828125
  1259  - 1.220703125
  1260  - 3.0517578125
  1261  - 7.62939453125
  1262  - 19.073486328125
  1263  - 47.6837158203125
  1264  - 119.20928955078125
  1265  - 298.0232238769531
  1266  - 745.0580596923828
  1267- name: runtime_operations_errors_total
  1268  subsystem: kubelet
  1269  help: Cumulative number of runtime operation errors by operation type.
  1270  type: Counter
  1271  stabilityLevel: ALPHA
  1272  labels:
  1273  - operation_type
  1274- name: runtime_operations_total
  1275  subsystem: kubelet
  1276  help: Cumulative number of runtime operations by operation type.
  1277  type: Counter
  1278  stabilityLevel: ALPHA
  1279  labels:
  1280  - operation_type
  1281- name: started_containers_errors_total
  1282  subsystem: kubelet
  1283  help: Cumulative number of errors when starting containers
  1284  type: Counter
  1285  stabilityLevel: ALPHA
  1286  labels:
  1287  - code
  1288  - container_type
  1289- name: started_containers_total
  1290  subsystem: kubelet
  1291  help: Cumulative number of containers started
  1292  type: Counter
  1293  stabilityLevel: ALPHA
  1294  labels:
  1295  - container_type
  1296- name: started_host_process_containers_errors_total
  1297  subsystem: kubelet
  1298  help: Cumulative number of errors when starting hostprocess containers. This metric
  1299    will only be collected on Windows.
  1300  type: Counter
  1301  stabilityLevel: ALPHA
  1302  labels:
  1303  - code
  1304  - container_type
  1305- name: started_host_process_containers_total
  1306  subsystem: kubelet
  1307  help: Cumulative number of hostprocess containers started. This metric will only
  1308    be collected on Windows.
  1309  type: Counter
  1310  stabilityLevel: ALPHA
  1311  labels:
  1312  - container_type
  1313- name: started_pods_errors_total
  1314  subsystem: kubelet
  1315  help: Cumulative number of errors when starting pods
  1316  type: Counter
  1317  stabilityLevel: ALPHA
  1318- name: started_pods_total
  1319  subsystem: kubelet
  1320  help: Cumulative number of pods started
  1321  type: Counter
  1322  stabilityLevel: ALPHA
  1323- name: topology_manager_admission_duration_ms
  1324  subsystem: kubelet
  1325  help: Duration in milliseconds to serve a pod admission request.
  1326  type: Histogram
  1327  stabilityLevel: ALPHA
  1328  buckets:
  1329  - 0.05
  1330  - 0.1
  1331  - 0.2
  1332  - 0.4
  1333  - 0.8
  1334  - 1.6
  1335  - 3.2
  1336  - 6.4
  1337  - 12.8
  1338  - 25.6
  1339  - 51.2
  1340  - 102.4
  1341  - 204.8
  1342  - 409.6
  1343  - 819.2
  1344- name: topology_manager_admission_errors_total
  1345  subsystem: kubelet
  1346  help: The number of admission request failures where resources could not be aligned.
  1347  type: Counter
  1348  stabilityLevel: ALPHA
  1349- name: topology_manager_admission_requests_total
  1350  subsystem: kubelet
  1351  help: The number of admission requests where resources have to be aligned.
  1352  type: Counter
  1353  stabilityLevel: ALPHA
  1354- name: kubelet_volume_stats_available_bytes
  1355  help: Number of available bytes in the volume
  1356  type: Custom
  1357  stabilityLevel: ALPHA
  1358  labels:
  1359  - namespace
  1360  - persistentvolumeclaim
  1361- name: kubelet_volume_stats_capacity_bytes
  1362  help: Capacity in bytes of the volume
  1363  type: Custom
  1364  stabilityLevel: ALPHA
  1365  labels:
  1366  - namespace
  1367  - persistentvolumeclaim
  1368- name: kubelet_volume_stats_health_status_abnormal
  1369  help: Abnormal volume health status. The count is either 1 or 0. 1 indicates the
  1370    volume is unhealthy, 0 indicates volume is healthy
  1371  type: Custom
  1372  stabilityLevel: ALPHA
  1373  labels:
  1374  - namespace
  1375  - persistentvolumeclaim
  1376- name: kubelet_volume_stats_inodes
  1377  help: Maximum number of inodes in the volume
  1378  type: Custom
  1379  stabilityLevel: ALPHA
  1380  labels:
  1381  - namespace
  1382  - persistentvolumeclaim
  1383- name: kubelet_volume_stats_inodes_free
  1384  help: Number of free inodes in the volume
  1385  type: Custom
  1386  stabilityLevel: ALPHA
  1387  labels:
  1388  - namespace
  1389  - persistentvolumeclaim
  1390- name: kubelet_volume_stats_inodes_used
  1391  help: Number of used inodes in the volume
  1392  type: Custom
  1393  stabilityLevel: ALPHA
  1394  labels:
  1395  - namespace
  1396  - persistentvolumeclaim
  1397- name: kubelet_volume_stats_used_bytes
  1398  help: Number of used bytes in the volume
  1399  type: Custom
  1400  stabilityLevel: ALPHA
  1401  labels:
  1402  - namespace
  1403  - persistentvolumeclaim
  1404- name: working_pods
  1405  subsystem: kubelet
  1406  help: Number of pods the kubelet is actually running, broken down by lifecycle phase,
  1407    whether the pod is desired, orphaned, or runtime only (also orphaned), and whether
  1408    the pod is static. An orphaned pod has been removed from local configuration or
  1409    force deleted in the API and consumes resources that are not otherwise visible.
  1410  type: Gauge
  1411  stabilityLevel: ALPHA
  1412  labels:
  1413  - config
  1414  - lifecycle
  1415  - static
  1416- name: node_swap_usage_bytes
  1417  help: Current swap usage of the node in bytes. Reported only on non-windows systems
  1418  type: Custom
  1419  stabilityLevel: ALPHA
  1420- name: pod_swap_usage_bytes
  1421  help: Current amount of the pod swap usage in bytes. Reported only on non-windows
  1422    systems
  1423  type: Custom
  1424  stabilityLevel: ALPHA
  1425  labels:
  1426  - pod
  1427  - namespace
  1428- name: scrape_error
  1429  help: 1 if there was an error while getting container metrics, 0 otherwise
  1430  type: Custom
  1431  deprecatedVersion: 1.29.0
  1432  stabilityLevel: ALPHA
  1433- name: container_cpu_usage_seconds_total
  1434  help: Cumulative cpu time consumed by the container in core-seconds
  1435  type: Custom
  1436  stabilityLevel: STABLE
  1437  labels:
  1438  - container
  1439  - pod
  1440  - namespace
  1441- name: container_memory_working_set_bytes
  1442  help: Current working set of the container in bytes
  1443  type: Custom
  1444  stabilityLevel: STABLE
  1445  labels:
  1446  - container
  1447  - pod
  1448  - namespace
  1449- name: container_start_time_seconds
  1450  help: Start time of the container since unix epoch in seconds
  1451  type: Custom
  1452  stabilityLevel: STABLE
  1453  labels:
  1454  - container
  1455  - pod
  1456  - namespace
  1457- name: node_cpu_usage_seconds_total
  1458  help: Cumulative cpu time consumed by the node in core-seconds
  1459  type: Custom
  1460  stabilityLevel: STABLE
  1461- name: node_memory_working_set_bytes
  1462  help: Current working set of the node in bytes
  1463  type: Custom
  1464  stabilityLevel: STABLE
  1465- name: pod_cpu_usage_seconds_total
  1466  help: Cumulative cpu time consumed by the pod in core-seconds
  1467  type: Custom
  1468  stabilityLevel: STABLE
  1469  labels:
  1470  - pod
  1471  - namespace
  1472- name: pod_memory_working_set_bytes
  1473  help: Current working set of the pod in bytes
  1474  type: Custom
  1475  stabilityLevel: STABLE
  1476  labels:
  1477  - pod
  1478  - namespace
  1479- name: resource_scrape_error
  1480  help: 1 if there was an error while getting container metrics, 0 otherwise
  1481  type: Custom
  1482  stabilityLevel: STABLE
  1483- name: force_cleaned_failed_volume_operation_errors_total
  1484  help: The number of volumes that failed force cleanup after their reconstruction
  1485    failed during kubelet startup.
  1486  type: Counter
  1487  stabilityLevel: ALPHA
  1488- name: force_cleaned_failed_volume_operations_total
  1489  help: The number of volumes that were force cleaned after their reconstruction failed
  1490    during kubelet startup. This includes both successful and failed cleanups.
  1491  type: Counter
  1492  stabilityLevel: ALPHA
  1493- name: http_inflight_requests
  1494  subsystem: kubelet
  1495  help: Number of the inflight http requests
  1496  type: Gauge
  1497  stabilityLevel: ALPHA
  1498  labels:
  1499  - long_running
  1500  - method
  1501  - path
  1502  - server_type
  1503- name: http_requests_duration_seconds
  1504  subsystem: kubelet
  1505  help: Duration in seconds to serve http requests
  1506  type: Histogram
  1507  stabilityLevel: ALPHA
  1508  labels:
  1509  - long_running
  1510  - method
  1511  - path
  1512  - server_type
  1513  buckets:
  1514  - 0.005
  1515  - 0.01
  1516  - 0.025
  1517  - 0.05
  1518  - 0.1
  1519  - 0.25
  1520  - 0.5
  1521  - 1
  1522  - 2.5
  1523  - 5
  1524  - 10
  1525- name: http_requests_total
  1526  subsystem: kubelet
  1527  help: Number of the http requests received since the server started
  1528  type: Counter
  1529  stabilityLevel: ALPHA
  1530  labels:
  1531  - long_running
  1532  - method
  1533  - path
  1534  - server_type
  1535- name: volume_metric_collection_duration_seconds
  1536  subsystem: kubelet
  1537  help: Duration in seconds to calculate volume stats
  1538  type: Histogram
  1539  stabilityLevel: ALPHA
  1540  labels:
  1541  - metric_source
  1542  buckets:
  1543  - 0.005
  1544  - 0.01
  1545  - 0.025
  1546  - 0.05
  1547  - 0.1
  1548  - 0.25
  1549  - 0.5
  1550  - 1
  1551  - 2.5
  1552  - 5
  1553  - 10
  1554- name: network_programming_duration_seconds
  1555  subsystem: kubeproxy
  1556  help: In Cluster Network Programming Latency in seconds
  1557  type: Histogram
  1558  stabilityLevel: ALPHA
  1559  buckets:
  1560  - 0.25
  1561  - 0.5
  1562  - 1
  1563  - 2
  1564  - 3
  1565  - 4
  1566  - 5
  1567  - 6
  1568  - 7
  1569  - 8
  1570  - 9
  1571  - 10
  1572  - 11
  1573  - 12
  1574  - 13
  1575  - 14
  1576  - 15
  1577  - 16
  1578  - 17
  1579  - 18
  1580  - 19
  1581  - 20
  1582  - 21
  1583  - 22
  1584  - 23
  1585  - 24
  1586  - 25
  1587  - 26
  1588  - 27
  1589  - 28
  1590  - 29
  1591  - 30
  1592  - 31
  1593  - 32
  1594  - 33
  1595  - 34
  1596  - 35
  1597  - 36
  1598  - 37
  1599  - 38
  1600  - 39
  1601  - 40
  1602  - 41
  1603  - 42
  1604  - 43
  1605  - 44
  1606  - 45
  1607  - 46
  1608  - 47
  1609  - 48
  1610  - 49
  1611  - 50
  1612  - 51
  1613  - 52
  1614  - 53
  1615  - 54
  1616  - 55
  1617  - 56
  1618  - 57
  1619  - 58
  1620  - 59
  1621  - 60
  1622  - 65
  1623  - 70
  1624  - 75
  1625  - 80
  1626  - 85
  1627  - 90
  1628  - 95
  1629  - 100
  1630  - 105
  1631  - 110
  1632  - 115
  1633  - 120
  1634  - 150
  1635  - 180
  1636  - 210
  1637  - 240
  1638  - 270
  1639  - 300
  1640- name: proxy_healthz_total
  1641  subsystem: kubeproxy
  1642  help: Cumulative proxy healthz HTTP status
  1643  type: Counter
  1644  stabilityLevel: ALPHA
  1645  labels:
  1646  - code
  1647- name: proxy_livez_total
  1648  subsystem: kubeproxy
  1649  help: Cumulative proxy livez HTTP status
  1650  type: Counter
  1651  stabilityLevel: ALPHA
  1652  labels:
  1653  - code
  1654- name: sync_full_proxy_rules_duration_seconds
  1655  subsystem: kubeproxy
  1656  help: SyncProxyRules latency in seconds for full resyncs
  1657  type: Histogram
  1658  stabilityLevel: ALPHA
  1659  buckets:
  1660  - 0.001
  1661  - 0.002
  1662  - 0.004
  1663  - 0.008
  1664  - 0.016
  1665  - 0.032
  1666  - 0.064
  1667  - 0.128
  1668  - 0.256
  1669  - 0.512
  1670  - 1.024
  1671  - 2.048
  1672  - 4.096
  1673  - 8.192
  1674  - 16.384
  1675- name: sync_partial_proxy_rules_duration_seconds
  1676  subsystem: kubeproxy
  1677  help: SyncProxyRules latency in seconds for partial resyncs
  1678  type: Histogram
  1679  stabilityLevel: ALPHA
  1680  buckets:
  1681  - 0.001
  1682  - 0.002
  1683  - 0.004
  1684  - 0.008
  1685  - 0.016
  1686  - 0.032
  1687  - 0.064
  1688  - 0.128
  1689  - 0.256
  1690  - 0.512
  1691  - 1.024
  1692  - 2.048
  1693  - 4.096
  1694  - 8.192
  1695  - 16.384
  1696- name: sync_proxy_rules_duration_seconds
  1697  subsystem: kubeproxy
  1698  help: SyncProxyRules latency in seconds
  1699  type: Histogram
  1700  stabilityLevel: ALPHA
  1701  buckets:
  1702  - 0.001
  1703  - 0.002
  1704  - 0.004
  1705  - 0.008
  1706  - 0.016
  1707  - 0.032
  1708  - 0.064
  1709  - 0.128
  1710  - 0.256
  1711  - 0.512
  1712  - 1.024
  1713  - 2.048
  1714  - 4.096
  1715  - 8.192
  1716  - 16.384
  1717- name: sync_proxy_rules_endpoint_changes_pending
  1718  subsystem: kubeproxy
  1719  help: Pending proxy rules Endpoint changes
  1720  type: Gauge
  1721  stabilityLevel: ALPHA
  1722- name: sync_proxy_rules_endpoint_changes_total
  1723  subsystem: kubeproxy
  1724  help: Cumulative proxy rules Endpoint changes
  1725  type: Counter
  1726  stabilityLevel: ALPHA
  1727- name: sync_proxy_rules_iptables_last
  1728  subsystem: kubeproxy
  1729  help: Number of iptables rules written by kube-proxy in last sync
  1730  type: Gauge
  1731  stabilityLevel: ALPHA
  1732  labels:
  1733  - table
  1734- name: sync_proxy_rules_iptables_partial_restore_failures_total
  1735  subsystem: kubeproxy
  1736  help: Cumulative proxy iptables partial restore failures
  1737  type: Counter
  1738  stabilityLevel: ALPHA
  1739- name: sync_proxy_rules_iptables_restore_failures_total
  1740  subsystem: kubeproxy
  1741  help: Cumulative proxy iptables restore failures
  1742  type: Counter
  1743  stabilityLevel: ALPHA
  1744- name: sync_proxy_rules_iptables_total
  1745  subsystem: kubeproxy
  1746  help: Total number of iptables rules owned by kube-proxy
  1747  type: Gauge
  1748  stabilityLevel: ALPHA
  1749  labels:
  1750  - table
  1751- name: sync_proxy_rules_last_queued_timestamp_seconds
  1752  subsystem: kubeproxy
  1753  help: The last time a sync of proxy rules was queued
  1754  type: Gauge
  1755  stabilityLevel: ALPHA
  1756- name: sync_proxy_rules_last_timestamp_seconds
  1757  subsystem: kubeproxy
  1758  help: The last time proxy rules were successfully synced
  1759  type: Gauge
  1760  stabilityLevel: ALPHA
  1761- name: sync_proxy_rules_no_local_endpoints_total
  1762  subsystem: kubeproxy
  1763  help: Number of services with a Local traffic policy and no endpoints
  1764  type: Gauge
  1765  stabilityLevel: ALPHA
  1766  labels:
  1767  - traffic_policy
  1768- name: sync_proxy_rules_service_changes_pending
  1769  subsystem: kubeproxy
  1770  help: Pending proxy rules Service changes
  1771  type: Gauge
  1772  stabilityLevel: ALPHA
  1773- name: sync_proxy_rules_service_changes_total
  1774  subsystem: kubeproxy
  1775  help: Cumulative proxy rules Service changes
  1776  type: Counter
  1777  stabilityLevel: ALPHA
  1778- name: plugin_manager_total_plugins
  1779  help: Number of plugins in Plugin Manager
  1780  type: Custom
  1781  stabilityLevel: ALPHA
  1782  labels:
  1783  - socket_path
  1784  - state
  1785- name: probe_duration_seconds
  1786  subsystem: prober
  1787  help: Duration in seconds for a probe response.
  1788  type: Histogram
  1789  stabilityLevel: ALPHA
  1790  labels:
  1791  - container
  1792  - namespace
  1793  - pod
  1794  - probe_type
  1795- name: probe_total
  1796  subsystem: prober
  1797  help: Cumulative number of a liveness, readiness or startup probe for a container
  1798    by result.
  1799  type: Counter
  1800  stabilityLevel: ALPHA
  1801  labels:
  1802  - container
  1803  - namespace
  1804  - pod
  1805  - pod_uid
  1806  - probe_type
  1807  - result
  1808- name: reconstruct_volume_operations_errors_total
  1809  help: The number of volumes that failed reconstruction from the operating system
  1810    during kubelet startup.
  1811  type: Counter
  1812  stabilityLevel: ALPHA
  1813- name: reconstruct_volume_operations_total
  1814  help: The number of volumes that were attempted to be reconstructed from the operating
  1815    system during kubelet startup. This includes both successful and failed reconstruction.
  1816  type: Counter
  1817  stabilityLevel: ALPHA
  1818- name: volume_manager_selinux_container_errors_total
  1819  help: Number of errors when kubelet cannot compute SELinux context for a container.
  1820    Kubelet can't start such a Pod then and it will retry, therefore value of this
  1821    metric may not represent the actual nr. of containers.
  1822  type: Gauge
  1823  stabilityLevel: ALPHA
  1824- name: volume_manager_selinux_container_warnings_total
  1825  help: Number of errors when kubelet cannot compute SELinux context for a container
  1826    that are ignored. They will become real errors when SELinuxMountReadWriteOncePod
  1827    feature is expanded to all volume access modes.
  1828  type: Gauge
  1829  stabilityLevel: ALPHA
  1830- name: volume_manager_selinux_pod_context_mismatch_errors_total
  1831  help: Number of errors when a Pod defines different SELinux contexts for its containers
  1832    that use the same volume. Kubelet can't start such a Pod then and it will retry,
  1833    therefore value of this metric may not represent the actual nr. of Pods.
  1834  type: Gauge
  1835  stabilityLevel: ALPHA
  1836- name: volume_manager_selinux_pod_context_mismatch_warnings_total
  1837  help: Number of errors when a Pod defines different SELinux contexts for its containers
  1838    that use the same volume. They are not errors yet, but they will become real errors
  1839    when SELinuxMountReadWriteOncePod feature is expanded to all volume access modes.
  1840  type: Gauge
  1841  stabilityLevel: ALPHA
  1842- name: volume_manager_selinux_volume_context_mismatch_errors_total
  1843  help: Number of errors when a Pod uses a volume that is already mounted with a different
  1844    SELinux context than the Pod needs. Kubelet can't start such a Pod then and it
  1845    will retry, therefore value of this metric may not represent the actual nr. of
  1846    Pods.
  1847  type: Gauge
  1848  stabilityLevel: ALPHA
  1849  labels:
  1850  - volume_plugin
  1851- name: volume_manager_selinux_volume_context_mismatch_warnings_total
  1852  help: Number of errors when a Pod uses a volume that is already mounted with a different
  1853    SELinux context than the Pod needs. They are not errors yet, but they will become
  1854    real errors when SELinuxMountReadWriteOncePod feature is expanded to all volume
  1855    access modes.
  1856  type: Gauge
  1857  stabilityLevel: ALPHA
  1858  labels:
  1859  - volume_plugin
  1860- name: volume_manager_selinux_volumes_admitted_total
  1861  help: Number of volumes whose SELinux context was fine and will be mounted with
  1862    mount -o context option.
  1863  type: Gauge
  1864  stabilityLevel: ALPHA
  1865  labels:
  1866  - volume_plugin
  1867- name: volume_manager_total_volumes
  1868  help: Number of volumes in Volume Manager
  1869  type: Custom
  1870  stabilityLevel: ALPHA
  1871  labels:
  1872  - plugin_name
  1873  - state
  1874- name: csr_honored_duration_total
  1875  subsystem: certificates_registry
  1876  namespace: apiserver
  1877  help: Total number of issued CSRs with a requested duration that was honored, sliced
  1878    by signer (only kubernetes.io signer names are specifically identified)
  1879  type: Counter
  1880  stabilityLevel: ALPHA
  1881  labels:
  1882  - signerName
  1883- name: csr_requested_duration_total
  1884  subsystem: certificates_registry
  1885  namespace: apiserver
  1886  help: Total number of issued CSRs with a requested duration, sliced by signer (only
  1887    kubernetes.io signer names are specifically identified)
  1888  type: Counter
  1889  stabilityLevel: ALPHA
  1890  labels:
  1891  - signerName
  1892- name: ip_errors_total
  1893  subsystem: clusterip_repair
  1894  namespace: apiserver
  1895  help: 'Number of errors detected on clusterips by the repair loop broken down by
  1896    type of error: leak, repair, full, outOfRange, duplicate, unknown, invalid'
  1897  type: Counter
  1898  stabilityLevel: ALPHA
  1899  labels:
  1900  - type
  1901- name: reconcile_errors_total
  1902  subsystem: clusterip_repair
  1903  namespace: apiserver
  1904  help: Number of reconciliation failures on the clusterip repair reconcile loop
  1905  type: Counter
  1906  stabilityLevel: ALPHA
  1907- name: port_errors_total
  1908  subsystem: nodeport_repair
  1909  namespace: apiserver
  1910  help: 'Number of errors detected on ports by the repair loop broken down by type
  1911    of error: leak, repair, full, outOfRange, duplicate, unknown'
  1912  type: Counter
  1913  stabilityLevel: ALPHA
  1914  labels:
  1915  - type
  1916- name: allocated_ips
  1917  subsystem: clusterip_allocator
  1918  namespace: kube_apiserver
  1919  help: Gauge measuring the number of allocated IPs for Services
  1920  type: Gauge
  1921  stabilityLevel: ALPHA
  1922  labels:
  1923  - cidr
  1924- name: allocation_errors_total
  1925  subsystem: clusterip_allocator
  1926  namespace: kube_apiserver
  1927  help: Number of errors trying to allocate Cluster IPs
  1928  type: Counter
  1929  stabilityLevel: ALPHA
  1930  labels:
  1931  - cidr
  1932  - scope
  1933- name: allocation_total
  1934  subsystem: clusterip_allocator
  1935  namespace: kube_apiserver
  1936  help: Number of Cluster IPs allocations
  1937  type: Counter
  1938  stabilityLevel: ALPHA
  1939  labels:
  1940  - cidr
  1941  - scope
  1942- name: available_ips
  1943  subsystem: clusterip_allocator
  1944  namespace: kube_apiserver
  1945  help: Gauge measuring the number of available IPs for Services
  1946  type: Gauge
  1947  stabilityLevel: ALPHA
  1948  labels:
  1949  - cidr
  1950- name: allocated_ports
  1951  subsystem: nodeport_allocator
  1952  namespace: kube_apiserver
  1953  help: Gauge measuring the number of allocated NodePorts for Services
  1954  type: Gauge
  1955  stabilityLevel: ALPHA
  1956- name: available_ports
  1957  subsystem: nodeport_allocator
  1958  namespace: kube_apiserver
  1959  help: Gauge measuring the number of available NodePorts for Services
  1960  type: Gauge
  1961  stabilityLevel: ALPHA
  1962- name: backend_tls_failure_total
  1963  subsystem: pod_logs
  1964  namespace: kube_apiserver
  1965  help: Total number of requests for pods/logs that failed due to kubelet server TLS
  1966    verification
  1967  type: Counter
  1968  stabilityLevel: ALPHA
  1969- name: insecure_backend_total
  1970  subsystem: pod_logs
  1971  namespace: kube_apiserver
  1972  help: 'Total number of requests for pods/logs sliced by usage type: enforce_tls,
  1973    skip_tls_allowed, skip_tls_denied'
  1974  type: Counter
  1975  stabilityLevel: ALPHA
  1976  labels:
  1977  - usage
  1978- name: pods_logs_backend_tls_failure_total
  1979  subsystem: pod_logs
  1980  namespace: kube_apiserver
  1981  help: Total number of requests for pods/logs that failed due to kubelet server TLS
  1982    verification
  1983  type: Counter
  1984  deprecatedVersion: 1.27.0
  1985  stabilityLevel: ALPHA
  1986- name: pods_logs_insecure_backend_total
  1987  subsystem: pod_logs
  1988  namespace: kube_apiserver
  1989  help: 'Total number of requests for pods/logs sliced by usage type: enforce_tls,
  1990    skip_tls_allowed, skip_tls_denied'
  1991  type: Counter
  1992  deprecatedVersion: 1.27.0
  1993  stabilityLevel: ALPHA
  1994  labels:
  1995  - usage
  1996- name: goroutines
  1997  subsystem: scheduler
  1998  help: Number of running goroutines split by the work they do such as binding.
  1999  type: Gauge
  2000  stabilityLevel: ALPHA
  2001  labels:
  2002  - operation
  2003- name: permit_wait_duration_seconds
  2004  subsystem: scheduler
  2005  help: Duration of waiting on permit.
  2006  type: Histogram
  2007  stabilityLevel: ALPHA
  2008  labels:
  2009  - result
  2010  buckets:
  2011  - 0.001
  2012  - 0.002
  2013  - 0.004
  2014  - 0.008
  2015  - 0.016
  2016  - 0.032
  2017  - 0.064
  2018  - 0.128
  2019  - 0.256
  2020  - 0.512
  2021  - 1.024
  2022  - 2.048
  2023  - 4.096
  2024  - 8.192
  2025  - 16.384
  2026- name: plugin_evaluation_total
  2027  subsystem: scheduler
  2028  help: Number of attempts to schedule pods by each plugin and the extension point
  2029    (available only in PreFilter and Filter.).
  2030  type: Counter
  2031  stabilityLevel: ALPHA
  2032  labels:
  2033  - extension_point
  2034  - plugin
  2035  - profile
  2036- name: plugin_execution_duration_seconds
  2037  subsystem: scheduler
  2038  help: Duration for running a plugin at a specific extension point.
  2039  type: Histogram
  2040  stabilityLevel: ALPHA
  2041  labels:
  2042  - extension_point
  2043  - plugin
  2044  - status
  2045  buckets:
  2046  - 1e-05
  2047  - 1.5000000000000002e-05
  2048  - 2.2500000000000005e-05
  2049  - 3.375000000000001e-05
  2050  - 5.062500000000001e-05
  2051  - 7.593750000000002e-05
  2052  - 0.00011390625000000003
  2053  - 0.00017085937500000006
  2054  - 0.0002562890625000001
  2055  - 0.00038443359375000017
  2056  - 0.0005766503906250003
  2057  - 0.0008649755859375004
  2058  - 0.0012974633789062506
  2059  - 0.0019461950683593758
  2060  - 0.0029192926025390638
  2061  - 0.004378938903808595
  2062  - 0.006568408355712893
  2063  - 0.009852612533569338
  2064  - 0.014778918800354007
  2065  - 0.02216837820053101
  2066- name: scheduler_cache_size
  2067  subsystem: scheduler
  2068  help: Number of nodes, pods, and assumed (bound) pods in the scheduler cache.
  2069  type: Gauge
  2070  stabilityLevel: ALPHA
  2071  labels:
  2072  - type
  2073- name: scheduling_algorithm_duration_seconds
  2074  subsystem: scheduler
  2075  help: Scheduling algorithm latency in seconds
  2076  type: Histogram
  2077  stabilityLevel: ALPHA
  2078  buckets:
  2079  - 0.001
  2080  - 0.002
  2081  - 0.004
  2082  - 0.008
  2083  - 0.016
  2084  - 0.032
  2085  - 0.064
  2086  - 0.128
  2087  - 0.256
  2088  - 0.512
  2089  - 1.024
  2090  - 2.048
  2091  - 4.096
  2092  - 8.192
  2093  - 16.384
  2094- name: unschedulable_pods
  2095  subsystem: scheduler
  2096  help: The number of unschedulable pods broken down by plugin name. A pod will increment
  2097    the gauge for all plugins that caused it to not schedule and so this metric have
  2098    meaning only when broken down by plugin.
  2099  type: Gauge
  2100  stabilityLevel: ALPHA
  2101  labels:
  2102  - plugin
  2103  - profile
  2104- name: binder_cache_requests_total
  2105  subsystem: scheduler_volume
  2106  help: Total number for request volume binding cache
  2107  type: Counter
  2108  stabilityLevel: ALPHA
  2109  labels:
  2110  - operation
  2111- name: scheduling_stage_error_total
  2112  subsystem: scheduler_volume
  2113  help: Volume scheduling stage error count
  2114  type: Counter
  2115  stabilityLevel: ALPHA
  2116  labels:
  2117  - operation
  2118- name: invalid_legacy_auto_token_uses_total
  2119  subsystem: serviceaccount
  2120  help: Cumulative invalid auto-generated legacy tokens used
  2121  type: Counter
  2122  stabilityLevel: ALPHA
  2123- name: legacy_auto_token_uses_total
  2124  subsystem: serviceaccount
  2125  help: Cumulative auto-generated legacy tokens used
  2126  type: Counter
  2127  stabilityLevel: ALPHA
  2128- name: legacy_manual_token_uses_total
  2129  subsystem: serviceaccount
  2130  help: Cumulative manually created legacy tokens used
  2131  type: Counter
  2132  stabilityLevel: ALPHA
  2133- name: legacy_tokens_total
  2134  subsystem: serviceaccount
  2135  help: Cumulative legacy service account tokens used
  2136  type: Counter
  2137  stabilityLevel: ALPHA
  2138- name: stale_tokens_total
  2139  subsystem: serviceaccount
  2140  help: Cumulative stale projected service account tokens used
  2141  type: Counter
  2142  stabilityLevel: ALPHA
  2143- name: valid_tokens_total
  2144  subsystem: serviceaccount
  2145  help: Cumulative valid projected service account tokens used
  2146  type: Counter
  2147  stabilityLevel: ALPHA
  2148- name: pod_scheduling_sli_duration_seconds
  2149  subsystem: scheduler
  2150  help: E2e latency for a pod being scheduled, from the time the pod enters the scheduling
  2151    queue an d might involve multiple scheduling attempts.
  2152  type: Histogram
  2153  stabilityLevel: BETA
  2154  labels:
  2155  - attempts
  2156  buckets:
  2157  - 0.01
  2158  - 0.02
  2159  - 0.04
  2160  - 0.08
  2161  - 0.16
  2162  - 0.32
  2163  - 0.64
  2164  - 1.28
  2165  - 2.56
  2166  - 5.12
  2167  - 10.24
  2168  - 20.48
  2169  - 40.96
  2170  - 81.92
  2171  - 163.84
  2172  - 327.68
  2173  - 655.36
  2174  - 1310.72
  2175  - 2621.44
  2176  - 5242.88
  2177- name: kube_pod_resource_limit
  2178  help: Resources limit for workloads on the cluster, broken down by pod. This shows
  2179    the resource usage the scheduler and kubelet expect per pod for resources along
  2180    with the unit for the resource if any.
  2181  type: Custom
  2182  stabilityLevel: STABLE
  2183  labels:
  2184  - namespace
  2185  - pod
  2186  - node
  2187  - scheduler
  2188  - priority
  2189  - resource
  2190  - unit
  2191- name: kube_pod_resource_request
  2192  help: Resources requested by workloads on the cluster, broken down by pod. This
  2193    shows the resource usage the scheduler and kubelet expect per pod for resources
  2194    along with the unit for the resource if any.
  2195  type: Custom
  2196  stabilityLevel: STABLE
  2197  labels:
  2198  - namespace
  2199  - pod
  2200  - node
  2201  - scheduler
  2202  - priority
  2203  - resource
  2204  - unit
  2205- name: framework_extension_point_duration_seconds
  2206  subsystem: scheduler
  2207  help: Latency for running all plugins of a specific extension point.
  2208  type: Histogram
  2209  stabilityLevel: STABLE
  2210  labels:
  2211  - extension_point
  2212  - profile
  2213  - status
  2214  buckets:
  2215  - 0.0001
  2216  - 0.0002
  2217  - 0.0004
  2218  - 0.0008
  2219  - 0.0016
  2220  - 0.0032
  2221  - 0.0064
  2222  - 0.0128
  2223  - 0.0256
  2224  - 0.0512
  2225  - 0.1024
  2226  - 0.2048
  2227- name: pending_pods
  2228  subsystem: scheduler
  2229  help: Number of pending pods, by the queue type. 'active' means number of pods in
  2230    activeQ; 'backoff' means number of pods in backoffQ; 'unschedulable' means number
  2231    of pods in unschedulablePods that the scheduler attempted to schedule and failed;
  2232    'gated' is the number of unschedulable pods that the scheduler never attempted
  2233    to schedule because they are gated.
  2234  type: Gauge
  2235  stabilityLevel: STABLE
  2236  labels:
  2237  - queue
  2238- name: pod_scheduling_attempts
  2239  subsystem: scheduler
  2240  help: Number of attempts to successfully schedule a pod.
  2241  type: Histogram
  2242  stabilityLevel: STABLE
  2243  buckets:
  2244  - 1
  2245  - 2
  2246  - 4
  2247  - 8
  2248  - 16
  2249- name: pod_scheduling_duration_seconds
  2250  subsystem: scheduler
  2251  help: E2e latency for a pod being scheduled which may include multiple scheduling
  2252    attempts.
  2253  type: Histogram
  2254  deprecatedVersion: 1.28.0
  2255  stabilityLevel: STABLE
  2256  labels:
  2257  - attempts
  2258  buckets:
  2259  - 0.01
  2260  - 0.02
  2261  - 0.04
  2262  - 0.08
  2263  - 0.16
  2264  - 0.32
  2265  - 0.64
  2266  - 1.28
  2267  - 2.56
  2268  - 5.12
  2269  - 10.24
  2270  - 20.48
  2271  - 40.96
  2272  - 81.92
  2273  - 163.84
  2274  - 327.68
  2275  - 655.36
  2276  - 1310.72
  2277  - 2621.44
  2278  - 5242.88
  2279- name: preemption_attempts_total
  2280  subsystem: scheduler
  2281  help: Total preemption attempts in the cluster till now
  2282  type: Counter
  2283  stabilityLevel: STABLE
  2284- name: preemption_victims
  2285  subsystem: scheduler
  2286  help: Number of selected preemption victims
  2287  type: Histogram
  2288  stabilityLevel: STABLE
  2289  buckets:
  2290  - 1
  2291  - 2
  2292  - 4
  2293  - 8
  2294  - 16
  2295  - 32
  2296  - 64
  2297- name: queue_incoming_pods_total
  2298  subsystem: scheduler
  2299  help: Number of pods added to scheduling queues by event and queue type.
  2300  type: Counter
  2301  stabilityLevel: STABLE
  2302  labels:
  2303  - event
  2304  - queue
  2305- name: schedule_attempts_total
  2306  subsystem: scheduler
  2307  help: Number of attempts to schedule pods, by the result. 'unschedulable' means
  2308    a pod could not be scheduled, while 'error' means an internal scheduler problem.
  2309  type: Counter
  2310  stabilityLevel: STABLE
  2311  labels:
  2312  - profile
  2313  - result
  2314- name: scheduling_attempt_duration_seconds
  2315  subsystem: scheduler
  2316  help: Scheduling attempt latency in seconds (scheduling algorithm + binding)
  2317  type: Histogram
  2318  stabilityLevel: STABLE
  2319  labels:
  2320  - profile
  2321  - result
  2322  buckets:
  2323  - 0.001
  2324  - 0.002
  2325  - 0.004
  2326  - 0.008
  2327  - 0.016
  2328  - 0.032
  2329  - 0.064
  2330  - 0.128
  2331  - 0.256
  2332  - 0.512
  2333  - 1.024
  2334  - 2.048
  2335  - 4.096
  2336  - 8.192
  2337  - 16.384
  2338- name: operations_seconds
  2339  subsystem: csi
  2340  help: Container Storage Interface operation duration with gRPC error code status
  2341    total
  2342  type: Histogram
  2343  stabilityLevel: ALPHA
  2344  labels:
  2345  - driver_name
  2346  - grpc_status_code
  2347  - method_name
  2348  - migrated
  2349  buckets:
  2350  - 0.1
  2351  - 0.25
  2352  - 0.5
  2353  - 1
  2354  - 2.5
  2355  - 5
  2356  - 10
  2357  - 15
  2358  - 25
  2359  - 50
  2360  - 120
  2361  - 300
  2362  - 600
  2363- name: storage_operation_duration_seconds
  2364  help: Storage operation duration
  2365  type: Histogram
  2366  stabilityLevel: ALPHA
  2367  labels:
  2368  - migrated
  2369  - operation_name
  2370  - status
  2371  - volume_plugin
  2372  buckets:
  2373  - 0.1
  2374  - 0.25
  2375  - 0.5
  2376  - 1
  2377  - 2.5
  2378  - 5
  2379  - 10
  2380  - 15
  2381  - 25
  2382  - 50
  2383  - 120
  2384  - 300
  2385  - 600
  2386- name: volume_operation_total_seconds
  2387  help: Storage operation end to end duration in seconds
  2388  type: Histogram
  2389  stabilityLevel: ALPHA
  2390  labels:
  2391  - operation_name
  2392  - plugin_name
  2393  buckets:
  2394  - 0.1
  2395  - 0.25
  2396  - 0.5
  2397  - 1
  2398  - 2.5
  2399  - 5
  2400  - 10
  2401  - 15
  2402  - 25
  2403  - 50
  2404  - 120
  2405  - 300
  2406  - 600
  2407- name: graph_actions_duration_seconds
  2408  subsystem: node_authorizer
  2409  help: Histogram of duration of graph actions in node authorizer.
  2410  type: Histogram
  2411  stabilityLevel: ALPHA
  2412  labels:
  2413  - operation
  2414  buckets:
  2415  - 0.0001
  2416  - 0.0002
  2417  - 0.0004
  2418  - 0.0008
  2419  - 0.0016
  2420  - 0.0032
  2421  - 0.0064
  2422  - 0.0128
  2423  - 0.0256
  2424  - 0.0512
  2425  - 0.1024
  2426  - 0.2048
  2427- name: ratcheting_seconds
  2428  subsystem: validation
  2429  namespace: apiextensions_apiserver
  2430  help: Time for comparison of old to new for the purposes of CRDValidationRatcheting
  2431    during an UPDATE in seconds.
  2432  type: Histogram
  2433  stabilityLevel: ALPHA
  2434  buckets:
  2435  - 1e-05
  2436  - 4e-05
  2437  - 0.00016
  2438  - 0.00064
  2439  - 0.00256
  2440  - 0.01024
  2441  - 0.04096
  2442  - 0.16384
  2443  - 0.65536
  2444  - 2.62144
  2445- name: conversion_webhook_duration_seconds
  2446  namespace: apiserver
  2447  help: Conversion webhook request latency
  2448  type: Histogram
  2449  stabilityLevel: ALPHA
  2450  labels:
  2451  - failure_type
  2452  - result
  2453  buckets:
  2454  - 0.005
  2455  - 0.01
  2456  - 0.02
  2457  - 0.05
  2458  - 0.1
  2459  - 0.2
  2460  - 0.5
  2461  - 1
  2462  - 2
  2463  - 5
  2464  - 10
  2465  - 20
  2466  - 30
  2467  - 45
  2468  - 60
  2469- name: conversion_webhook_request_total
  2470  namespace: apiserver
  2471  help: Counter for conversion webhook requests with success/failure and failure error
  2472    type
  2473  type: Counter
  2474  stabilityLevel: ALPHA
  2475  labels:
  2476  - failure_type
  2477  - result
  2478- name: apiserver_crd_conversion_webhook_duration_seconds
  2479  help: CRD webhook conversion duration in seconds
  2480  type: Histogram
  2481  stabilityLevel: ALPHA
  2482  labels:
  2483  - crd_name
  2484  - from_version
  2485  - succeeded
  2486  - to_version
  2487  buckets:
  2488  - 0.001
  2489  - 0.002
  2490  - 0.004
  2491  - 0.008
  2492  - 0.016
  2493  - 0.032
  2494  - 0.064
  2495  - 0.128
  2496  - 0.256
  2497  - 0.512
  2498  - 1.024
  2499  - 2.048
  2500  - 4.096
  2501  - 8.192
  2502  - 16.384
  2503- name: apiextensions_openapi_v2_regeneration_count
  2504  help: Counter of OpenAPI v2 spec regeneration count broken down by causing CRD name
  2505    and reason.
  2506  type: Counter
  2507  stabilityLevel: ALPHA
  2508  labels:
  2509  - crd
  2510  - reason
  2511- name: apiextensions_openapi_v3_regeneration_count
  2512  help: Counter of OpenAPI v3 spec regeneration count broken down by group, version,
  2513    causing CRD and reason.
  2514  type: Counter
  2515  stabilityLevel: ALPHA
  2516  labels:
  2517  - crd
  2518  - group
  2519  - reason
  2520  - version
  2521- name: match_condition_evaluation_errors_total
  2522  subsystem: admission
  2523  namespace: apiserver
  2524  help: Admission match condition evaluation errors count, identified by name of resource
  2525    containing the match condition and broken out for each kind containing matchConditions
  2526    (webhook or policy), operation and admission type (validate or admit).
  2527  type: Counter
  2528  stabilityLevel: ALPHA
  2529  labels:
  2530  - kind
  2531  - name
  2532  - operation
  2533  - type
  2534- name: match_condition_evaluation_seconds
  2535  subsystem: admission
  2536  namespace: apiserver
  2537  help: Admission match condition evaluation time in seconds, identified by name and
  2538    broken out for each kind containing matchConditions (webhook or policy), operation
  2539    and type (validate or admit).
  2540  type: Histogram
  2541  stabilityLevel: ALPHA
  2542  labels:
  2543  - kind
  2544  - name
  2545  - operation
  2546  - type
  2547  buckets:
  2548  - 0.001
  2549  - 0.005
  2550  - 0.01
  2551  - 0.025
  2552  - 0.1
  2553  - 0.2
  2554  - 0.25
  2555- name: match_condition_exclusions_total
  2556  subsystem: admission
  2557  namespace: apiserver
  2558  help: Admission match condition evaluation exclusions count, identified by name
  2559    of resource containing the match condition and broken out for each kind containing
  2560    matchConditions (webhook or policy), operation and admission type (validate or
  2561    admit).
  2562  type: Counter
  2563  stabilityLevel: ALPHA
  2564  labels:
  2565  - kind
  2566  - name
  2567  - operation
  2568  - type
  2569- name: step_admission_duration_seconds_summary
  2570  subsystem: admission
  2571  namespace: apiserver
  2572  help: Admission sub-step latency summary in seconds, broken out for each operation
  2573    and API resource and step type (validate or admit).
  2574  type: Summary
  2575  stabilityLevel: ALPHA
  2576  labels:
  2577  - operation
  2578  - rejected
  2579  - type
  2580  maxAge: 18000000000000
  2581- name: webhook_fail_open_count
  2582  subsystem: admission
  2583  namespace: apiserver
  2584  help: Admission webhook fail open count, identified by name and broken out for each
  2585    admission type (validating or mutating).
  2586  type: Counter
  2587  stabilityLevel: ALPHA
  2588  labels:
  2589  - name
  2590  - type
  2591- name: webhook_rejection_count
  2592  subsystem: admission
  2593  namespace: apiserver
  2594  help: Admission webhook rejection count, identified by name and broken out for each
  2595    admission type (validating or admit) and operation. Additional labels specify
  2596    an error type (calling_webhook_error or apiserver_internal_error if an error occurred;
  2597    no_error otherwise) and optionally a non-zero rejection code if the webhook rejects
  2598    the request with an HTTP status code (honored by the apiserver when the code is
  2599    greater or equal to 400). Codes greater than 600 are truncated to 600, to keep
  2600    the metrics cardinality bounded.
  2601  type: Counter
  2602  stabilityLevel: ALPHA
  2603  labels:
  2604  - error_type
  2605  - name
  2606  - operation
  2607  - rejection_code
  2608  - type
  2609- name: webhook_request_total
  2610  subsystem: admission
  2611  namespace: apiserver
  2612  help: Admission webhook request total, identified by name and broken out for each
  2613    admission type (validating or mutating) and operation. Additional labels specify
  2614    whether the request was rejected or not and an HTTP status code. Codes greater
  2615    than 600 are truncated to 600, to keep the metrics cardinality bounded.
  2616  type: Counter
  2617  stabilityLevel: ALPHA
  2618  labels:
  2619  - code
  2620  - name
  2621  - operation
  2622  - rejected
  2623  - type
  2624- name: check_duration_seconds
  2625  subsystem: validating_admission_policy
  2626  namespace: apiserver
  2627  help: Validation admission latency for individual validation expressions in seconds,
  2628    labeled by policy and further including binding, state and enforcement action
  2629    taken.
  2630  type: Histogram
  2631  stabilityLevel: ALPHA
  2632  labels:
  2633  - enforcement_action
  2634  - policy
  2635  - policy_binding
  2636  - state
  2637  buckets:
  2638  - 5e-07
  2639  - 0.001
  2640  - 0.01
  2641  - 0.1
  2642  - 1
  2643- name: check_total
  2644  subsystem: validating_admission_policy
  2645  namespace: apiserver
  2646  help: Validation admission policy check total, labeled by policy and further identified
  2647    by binding, enforcement action taken, and state.
  2648  type: Counter
  2649  stabilityLevel: ALPHA
  2650  labels:
  2651  - enforcement_action
  2652  - policy
  2653  - policy_binding
  2654  - state
  2655- name: definition_total
  2656  subsystem: validating_admission_policy
  2657  namespace: apiserver
  2658  help: Validation admission policy count total, labeled by state and enforcement
  2659    action.
  2660  type: Counter
  2661  stabilityLevel: ALPHA
  2662  labels:
  2663  - enforcement_action
  2664  - state
  2665- name: controller_admission_duration_seconds
  2666  subsystem: admission
  2667  namespace: apiserver
  2668  help: Admission controller latency histogram in seconds, identified by name and
  2669    broken out for each operation and API resource and type (validate or admit).
  2670  type: Histogram
  2671  stabilityLevel: STABLE
  2672  labels:
  2673  - name
  2674  - operation
  2675  - rejected
  2676  - type
  2677  buckets:
  2678  - 0.005
  2679  - 0.025
  2680  - 0.1
  2681  - 0.5
  2682  - 1
  2683  - 2.5
  2684- name: step_admission_duration_seconds
  2685  subsystem: admission
  2686  namespace: apiserver
  2687  help: Admission sub-step latency histogram in seconds, broken out for each operation
  2688    and API resource and step type (validate or admit).
  2689  type: Histogram
  2690  stabilityLevel: STABLE
  2691  labels:
  2692  - operation
  2693  - rejected
  2694  - type
  2695  buckets:
  2696  - 0.005
  2697  - 0.025
  2698  - 0.1
  2699  - 0.5
  2700  - 1
  2701  - 2.5
  2702- name: webhook_admission_duration_seconds
  2703  subsystem: admission
  2704  namespace: apiserver
  2705  help: Admission webhook latency histogram in seconds, identified by name and broken
  2706    out for each operation and API resource and type (validate or admit).
  2707  type: Histogram
  2708  stabilityLevel: STABLE
  2709  labels:
  2710  - name
  2711  - operation
  2712  - rejected
  2713  - type
  2714  buckets:
  2715  - 0.005
  2716  - 0.025
  2717  - 0.1
  2718  - 0.5
  2719  - 1
  2720  - 2.5
  2721  - 10
  2722  - 25
  2723- name: aggregator_discovery_aggregation_count_total
  2724  help: Counter of number of times discovery was aggregated
  2725  type: Counter
  2726  stabilityLevel: ALPHA
  2727- name: error_total
  2728  subsystem: apiserver_audit
  2729  help: Counter of audit events that failed to be audited properly. Plugin identifies
  2730    the plugin affected by the error.
  2731  type: Counter
  2732  stabilityLevel: ALPHA
  2733  labels:
  2734  - plugin
  2735- name: event_total
  2736  subsystem: apiserver_audit
  2737  help: Counter of audit events generated and sent to the audit backend.
  2738  type: Counter
  2739  stabilityLevel: ALPHA
  2740- name: level_total
  2741  subsystem: apiserver_audit
  2742  help: Counter of policy levels for audit events (1 per request).
  2743  type: Counter
  2744  stabilityLevel: ALPHA
  2745  labels:
  2746  - level
  2747- name: requests_rejected_total
  2748  subsystem: apiserver_audit
  2749  help: Counter of apiserver requests rejected due to an error in audit logging backend.
  2750  type: Counter
  2751  stabilityLevel: ALPHA
  2752- name: compilation_duration_seconds
  2753  subsystem: cel
  2754  namespace: apiserver
  2755  help: CEL compilation time in seconds.
  2756  type: Histogram
  2757  stabilityLevel: ALPHA
  2758- name: evaluation_duration_seconds
  2759  subsystem: cel
  2760  namespace: apiserver
  2761  help: CEL evaluation time in seconds.
  2762  type: Histogram
  2763  stabilityLevel: ALPHA
  2764- name: certificate_expiration_seconds
  2765  subsystem: client
  2766  namespace: apiserver
  2767  help: Distribution of the remaining lifetime on the certificate used to authenticate
  2768    a request.
  2769  type: Histogram
  2770  stabilityLevel: ALPHA
  2771  buckets:
  2772  - 0
  2773  - 1800
  2774  - 3600
  2775  - 7200
  2776  - 21600
  2777  - 43200
  2778  - 86400
  2779  - 172800
  2780  - 345600
  2781  - 604800
  2782  - 2.592e+06
  2783  - 7.776e+06
  2784  - 1.5552e+07
  2785  - 3.1104e+07
  2786- name: apiserver_delegated_authn_request_duration_seconds
  2787  help: Request latency in seconds. Broken down by status code.
  2788  type: Histogram
  2789  stabilityLevel: ALPHA
  2790  labels:
  2791  - code
  2792  buckets:
  2793  - 0.25
  2794  - 0.5
  2795  - 0.7
  2796  - 1
  2797  - 1.5
  2798  - 3
  2799  - 5
  2800  - 10
  2801- name: apiserver_delegated_authn_request_total
  2802  help: Number of HTTP requests partitioned by status code.
  2803  type: Counter
  2804  stabilityLevel: ALPHA
  2805  labels:
  2806  - code
  2807- name: apiserver_delegated_authz_request_duration_seconds
  2808  help: Request latency in seconds. Broken down by status code.
  2809  type: Histogram
  2810  stabilityLevel: ALPHA
  2811  labels:
  2812  - code
  2813  buckets:
  2814  - 0.25
  2815  - 0.5
  2816  - 0.7
  2817  - 1
  2818  - 1.5
  2819  - 3
  2820  - 5
  2821  - 10
  2822- name: apiserver_delegated_authz_request_total
  2823  help: Number of HTTP requests partitioned by status code.
  2824  type: Counter
  2825  stabilityLevel: ALPHA
  2826  labels:
  2827  - code
  2828- name: active_fetch_count
  2829  subsystem: token_cache
  2830  namespace: authentication
  2831  type: Gauge
  2832  stabilityLevel: ALPHA
  2833  labels:
  2834  - status
  2835- name: fetch_total
  2836  subsystem: token_cache
  2837  namespace: authentication
  2838  type: Counter
  2839  stabilityLevel: ALPHA
  2840  labels:
  2841  - status
  2842- name: request_duration_seconds
  2843  subsystem: token_cache
  2844  namespace: authentication
  2845  type: Histogram
  2846  stabilityLevel: ALPHA
  2847  labels:
  2848  - status
  2849- name: request_total
  2850  subsystem: token_cache
  2851  namespace: authentication
  2852  type: Counter
  2853  stabilityLevel: ALPHA
  2854  labels:
  2855  - status
  2856- name: cache_list_fetched_objects_total
  2857  namespace: apiserver
  2858  help: Number of objects read from watch cache in the course of serving a LIST request
  2859  type: Counter
  2860  stabilityLevel: ALPHA
  2861  labels:
  2862  - index
  2863  - resource_prefix
  2864- name: cache_list_returned_objects_total
  2865  namespace: apiserver
  2866  help: Number of objects returned for a LIST request from watch cache
  2867  type: Counter
  2868  stabilityLevel: ALPHA
  2869  labels:
  2870  - resource_prefix
  2871- name: cache_list_total
  2872  namespace: apiserver
  2873  help: Number of LIST requests served from watch cache
  2874  type: Counter
  2875  stabilityLevel: ALPHA
  2876  labels:
  2877  - index
  2878  - resource_prefix
  2879- name: current_inqueue_requests
  2880  subsystem: apiserver
  2881  help: Maximal number of queued requests in this apiserver per request kind in last
  2882    second.
  2883  type: Gauge
  2884  stabilityLevel: ALPHA
  2885  labels:
  2886  - request_kind
  2887- name: dial_duration_seconds
  2888  subsystem: egress_dialer
  2889  namespace: apiserver
  2890  help: Dial latency histogram in seconds, labeled by the protocol (http-connect or
  2891    grpc), transport (tcp or uds)
  2892  type: Histogram
  2893  stabilityLevel: ALPHA
  2894  labels:
  2895  - protocol
  2896  - transport
  2897  buckets:
  2898  - 0.005
  2899  - 0.025
  2900  - 0.1
  2901  - 0.5
  2902  - 2.5
  2903  - 12.5
  2904- name: dial_failure_count
  2905  subsystem: egress_dialer
  2906  namespace: apiserver
  2907  help: Dial failure count, labeled by the protocol (http-connect or grpc), transport
  2908    (tcp or uds), and stage (connect or proxy). The stage indicates at which stage
  2909    the dial failed
  2910  type: Counter
  2911  stabilityLevel: ALPHA
  2912  labels:
  2913  - protocol
  2914  - stage
  2915  - transport
  2916- name: dial_start_total
  2917  subsystem: egress_dialer
  2918  namespace: apiserver
  2919  help: Dial starts, labeled by the protocol (http-connect or grpc) and transport
  2920    (tcp or uds).
  2921  type: Counter
  2922  stabilityLevel: ALPHA
  2923  labels:
  2924  - protocol
  2925  - transport
  2926- name: automatic_reload_failures_total
  2927  subsystem: encryption_config_controller
  2928  namespace: apiserver
  2929  help: Total number of failed automatic reloads of encryption configuration split
  2930    by apiserver identity.
  2931  type: Counter
  2932  stabilityLevel: ALPHA
  2933  labels:
  2934  - apiserver_id_hash
  2935- name: automatic_reload_last_timestamp_seconds
  2936  subsystem: encryption_config_controller
  2937  namespace: apiserver
  2938  help: Timestamp of the last successful or failed automatic reload of encryption
  2939    configuration split by apiserver identity.
  2940  type: Gauge
  2941  stabilityLevel: ALPHA
  2942  labels:
  2943  - apiserver_id_hash
  2944  - status
  2945- name: automatic_reload_success_total
  2946  subsystem: encryption_config_controller
  2947  namespace: apiserver
  2948  help: Total number of successful automatic reloads of encryption configuration split
  2949    by apiserver identity.
  2950  type: Counter
  2951  stabilityLevel: ALPHA
  2952  labels:
  2953  - apiserver_id_hash
  2954- name: init_events_total
  2955  namespace: apiserver
  2956  help: Counter of init events processed in watch cache broken by resource type.
  2957  type: Counter
  2958  stabilityLevel: ALPHA
  2959  labels:
  2960  - resource
  2961- name: request_aborts_total
  2962  subsystem: apiserver
  2963  help: Number of requests which apiserver aborted possibly due to a timeout, for
  2964    each group, version, verb, resource, subresource and scope
  2965  type: Counter
  2966  stabilityLevel: ALPHA
  2967  labels:
  2968  - group
  2969  - resource
  2970  - scope
  2971  - subresource
  2972  - verb
  2973  - version
  2974- name: request_body_size_bytes
  2975  subsystem: apiserver
  2976  help: Apiserver request body size in bytes broken out by resource and verb.
  2977  type: Histogram
  2978  stabilityLevel: ALPHA
  2979  labels:
  2980  - resource
  2981  - verb
  2982  buckets:
  2983  - 50000
  2984  - 150000
  2985  - 250000
  2986  - 350000
  2987  - 450000
  2988  - 550000
  2989  - 650000
  2990  - 750000
  2991  - 850000
  2992  - 950000
  2993  - 1.05e+06
  2994  - 1.15e+06
  2995  - 1.25e+06
  2996  - 1.35e+06
  2997  - 1.45e+06
  2998  - 1.55e+06
  2999  - 1.65e+06
  3000  - 1.75e+06
  3001  - 1.85e+06
  3002  - 1.95e+06
  3003  - 2.05e+06
  3004  - 2.15e+06
  3005  - 2.25e+06
  3006  - 2.35e+06
  3007  - 2.45e+06
  3008  - 2.55e+06
  3009  - 2.65e+06
  3010  - 2.75e+06
  3011  - 2.85e+06
  3012  - 2.95e+06
  3013  - 3.05e+06
  3014- name: request_filter_duration_seconds
  3015  subsystem: apiserver
  3016  help: Request filter latency distribution in seconds, for each filter type
  3017  type: Histogram
  3018  stabilityLevel: ALPHA
  3019  labels:
  3020  - filter
  3021  buckets:
  3022  - 0.0001
  3023  - 0.0003
  3024  - 0.001
  3025  - 0.003
  3026  - 0.01
  3027  - 0.03
  3028  - 0.1
  3029  - 0.3
  3030  - 1
  3031  - 5
  3032  - 10
  3033  - 15
  3034  - 30
  3035- name: request_post_timeout_total
  3036  subsystem: apiserver
  3037  help: Tracks the activity of the request handlers after the associated requests
  3038    have been timed out by the apiserver
  3039  type: Counter
  3040  stabilityLevel: ALPHA
  3041  labels:
  3042  - source
  3043  - status
  3044- name: request_sli_duration_seconds
  3045  subsystem: apiserver
  3046  help: Response latency distribution (not counting webhook duration and priority
  3047    & fairness queue wait times) in seconds for each verb, group, version, resource,
  3048    subresource, scope and component.
  3049  type: Histogram
  3050  stabilityLevel: ALPHA
  3051  labels:
  3052  - component
  3053  - group
  3054  - resource
  3055  - scope
  3056  - subresource
  3057  - verb
  3058  - version
  3059  buckets:
  3060  - 0.05
  3061  - 0.1
  3062  - 0.2
  3063  - 0.4
  3064  - 0.6
  3065  - 0.8
  3066  - 1
  3067  - 1.25
  3068  - 1.5
  3069  - 2
  3070  - 3
  3071  - 4
  3072  - 5
  3073  - 6
  3074  - 8
  3075  - 10
  3076  - 15
  3077  - 20
  3078  - 30
  3079  - 45
  3080  - 60
  3081- name: request_slo_duration_seconds
  3082  subsystem: apiserver
  3083  help: Response latency distribution (not counting webhook duration and priority
  3084    & fairness queue wait times) in seconds for each verb, group, version, resource,
  3085    subresource, scope and component.
  3086  type: Histogram
  3087  deprecatedVersion: 1.27.0
  3088  stabilityLevel: ALPHA
  3089  labels:
  3090  - component
  3091  - group
  3092  - resource
  3093  - scope
  3094  - subresource
  3095  - verb
  3096  - version
  3097  buckets:
  3098  - 0.05
  3099  - 0.1
  3100  - 0.2
  3101  - 0.4
  3102  - 0.6
  3103  - 0.8
  3104  - 1
  3105  - 1.25
  3106  - 1.5
  3107  - 2
  3108  - 3
  3109  - 4
  3110  - 5
  3111  - 6
  3112  - 8
  3113  - 10
  3114  - 15
  3115  - 20
  3116  - 30
  3117  - 45
  3118  - 60
  3119- name: request_terminations_total
  3120  subsystem: apiserver
  3121  help: Number of requests which apiserver terminated in self-defense.
  3122  type: Counter
  3123  stabilityLevel: ALPHA
  3124  labels:
  3125  - code
  3126  - component
  3127  - group
  3128  - resource
  3129  - scope
  3130  - subresource
  3131  - verb
  3132  - version
  3133- name: request_timestamp_comparison_time
  3134  subsystem: apiserver
  3135  help: Time taken for comparison of old vs new objects in UPDATE or PATCH requests
  3136  type: Histogram
  3137  stabilityLevel: ALPHA
  3138  labels:
  3139  - code_path
  3140  buckets:
  3141  - 0.0001
  3142  - 0.0003
  3143  - 0.001
  3144  - 0.003
  3145  - 0.01
  3146  - 0.03
  3147  - 0.1
  3148  - 0.3
  3149  - 1
  3150  - 5
  3151- name: selfrequest_total
  3152  subsystem: apiserver
  3153  help: Counter of apiserver self-requests broken out for each verb, API resource
  3154    and subresource.
  3155  type: Counter
  3156  stabilityLevel: ALPHA
  3157  labels:
  3158  - resource
  3159  - subresource
  3160  - verb
  3161- name: storage_db_total_size_in_bytes
  3162  subsystem: apiserver
  3163  help: Total size of the storage database file physically allocated in bytes.
  3164  type: Gauge
  3165  deprecatedVersion: 1.28.0
  3166  stabilityLevel: ALPHA
  3167  labels:
  3168  - endpoint
  3169- name: storage_decode_errors_total
  3170  namespace: apiserver
  3171  help: Number of stored object decode errors split by object type
  3172  type: Counter
  3173  stabilityLevel: ALPHA
  3174  labels:
  3175  - resource
  3176- name: storage_events_received_total
  3177  subsystem: apiserver
  3178  help: Number of etcd events received split by kind.
  3179  type: Counter
  3180  stabilityLevel: ALPHA
  3181  labels:
  3182  - resource
  3183- name: apiserver_storage_list_evaluated_objects_total
  3184  help: Number of objects tested in the course of serving a LIST request from storage
  3185  type: Counter
  3186  stabilityLevel: ALPHA
  3187  labels:
  3188  - resource
  3189- name: apiserver_storage_list_fetched_objects_total
  3190  help: Number of objects read from storage in the course of serving a LIST request
  3191  type: Counter
  3192  stabilityLevel: ALPHA
  3193  labels:
  3194  - resource
  3195- name: apiserver_storage_list_returned_objects_total
  3196  help: Number of objects returned for a LIST request from storage
  3197  type: Counter
  3198  stabilityLevel: ALPHA
  3199  labels:
  3200  - resource
  3201- name: apiserver_storage_list_total
  3202  help: Number of LIST requests served from storage
  3203  type: Counter
  3204  stabilityLevel: ALPHA
  3205  labels:
  3206  - resource
  3207- name: apiserver_storage_size_bytes
  3208  help: Size of the storage database file physically allocated in bytes.
  3209  type: Custom
  3210  stabilityLevel: ALPHA
  3211  labels:
  3212  - storage_cluster_id
  3213- name: terminated_watchers_total
  3214  namespace: apiserver
  3215  help: Counter of watchers closed due to unresponsiveness broken by resource type.
  3216  type: Counter
  3217  stabilityLevel: ALPHA
  3218  labels:
  3219  - resource
  3220- name: tls_handshake_errors_total
  3221  subsystem: apiserver
  3222  help: Number of requests dropped with 'TLS handshake error from' error
  3223  type: Counter
  3224  stabilityLevel: ALPHA
  3225- name: events_dispatched_total
  3226  subsystem: watch_cache
  3227  namespace: apiserver
  3228  help: Counter of events dispatched in watch cache broken by resource type.
  3229  type: Counter
  3230  stabilityLevel: ALPHA
  3231  labels:
  3232  - resource
  3233- name: events_received_total
  3234  subsystem: watch_cache
  3235  namespace: apiserver
  3236  help: Counter of events received in watch cache broken by resource type.
  3237  type: Counter
  3238  stabilityLevel: ALPHA
  3239  labels:
  3240  - resource
  3241- name: initializations_total
  3242  subsystem: watch_cache
  3243  namespace: apiserver
  3244  help: Counter of watch cache initializations broken by resource type.
  3245  type: Counter
  3246  stabilityLevel: ALPHA
  3247  labels:
  3248  - resource
  3249- name: watch_events_sizes
  3250  subsystem: apiserver
  3251  help: Watch event size distribution in bytes
  3252  type: Histogram
  3253  stabilityLevel: ALPHA
  3254  labels:
  3255  - group
  3256  - kind
  3257  - version
  3258  buckets:
  3259  - 1024
  3260  - 2048
  3261  - 4096
  3262  - 8192
  3263  - 16384
  3264  - 32768
  3265  - 65536
  3266  - 131072
  3267- name: watch_events_total
  3268  subsystem: apiserver
  3269  help: Number of events sent in watch clients
  3270  type: Counter
  3271  stabilityLevel: ALPHA
  3272  labels:
  3273  - group
  3274  - kind
  3275  - version
  3276- name: watch_list_duration_seconds
  3277  subsystem: apiserver
  3278  help: Response latency distribution in seconds for watch list requests broken by
  3279    group, version, resource and scope.
  3280  type: Histogram
  3281  stabilityLevel: ALPHA
  3282  labels:
  3283  - group
  3284  - resource
  3285  - scope
  3286  - version
  3287  buckets:
  3288  - 0.05
  3289  - 0.1
  3290  - 0.2
  3291  - 0.4
  3292  - 0.6
  3293  - 0.8
  3294  - 1
  3295  - 2
  3296  - 4
  3297  - 6
  3298  - 8
  3299  - 10
  3300  - 15
  3301  - 20
  3302  - 30
  3303  - 45
  3304  - 60
  3305- name: authenticated_user_requests
  3306  help: Counter of authenticated requests broken out by username.
  3307  type: Counter
  3308  stabilityLevel: ALPHA
  3309  labels:
  3310  - username
  3311- name: authentication_attempts
  3312  help: Counter of authenticated attempts.
  3313  type: Counter
  3314  stabilityLevel: ALPHA
  3315  labels:
  3316  - result
  3317- name: authentication_duration_seconds
  3318  help: Authentication duration in seconds broken out by result.
  3319  type: Histogram
  3320  stabilityLevel: ALPHA
  3321  labels:
  3322  - result
  3323  buckets:
  3324  - 0.001
  3325  - 0.002
  3326  - 0.004
  3327  - 0.008
  3328  - 0.016
  3329  - 0.032
  3330  - 0.064
  3331  - 0.128
  3332  - 0.256
  3333  - 0.512
  3334  - 1.024
  3335  - 2.048
  3336  - 4.096
  3337  - 8.192
  3338  - 16.384
  3339- name: authorization_attempts_total
  3340  help: Counter of authorization attempts broken down by result. It can be either
  3341    'allowed', 'denied', 'no-opinion' or 'error'.
  3342  type: Counter
  3343  stabilityLevel: ALPHA
  3344  labels:
  3345  - result
  3346- name: authorization_duration_seconds
  3347  help: Authorization duration in seconds broken out by result.
  3348  type: Histogram
  3349  stabilityLevel: ALPHA
  3350  labels:
  3351  - result
  3352  buckets:
  3353  - 0.001
  3354  - 0.002
  3355  - 0.004
  3356  - 0.008
  3357  - 0.016
  3358  - 0.032
  3359  - 0.064
  3360  - 0.128
  3361  - 0.256
  3362  - 0.512
  3363  - 1.024
  3364  - 2.048
  3365  - 4.096
  3366  - 8.192
  3367  - 16.384
  3368- name: etcd_bookmark_counts
  3369  help: Number of etcd bookmarks (progress notify events) split by kind.
  3370  type: Gauge
  3371  stabilityLevel: ALPHA
  3372  labels:
  3373  - resource
  3374- name: etcd_lease_object_counts
  3375  help: Number of objects attached to a single etcd lease.
  3376  type: Histogram
  3377  stabilityLevel: ALPHA
  3378  buckets:
  3379  - 10
  3380  - 50
  3381  - 100
  3382  - 500
  3383  - 1000
  3384  - 2500
  3385  - 5000
  3386- name: etcd_request_duration_seconds
  3387  help: Etcd request latency in seconds for each operation and object type.
  3388  type: Histogram
  3389  stabilityLevel: ALPHA
  3390  labels:
  3391  - operation
  3392  - type
  3393  buckets:
  3394  - 0.005
  3395  - 0.025
  3396  - 0.05
  3397  - 0.1
  3398  - 0.2
  3399  - 0.4
  3400  - 0.6
  3401  - 0.8
  3402  - 1
  3403  - 1.25
  3404  - 1.5
  3405  - 2
  3406  - 3
  3407  - 4
  3408  - 5
  3409  - 6
  3410  - 8
  3411  - 10
  3412  - 15
  3413  - 20
  3414  - 30
  3415  - 45
  3416  - 60
  3417- name: etcd_request_errors_total
  3418  help: Etcd failed request counts for each operation and object type.
  3419  type: Counter
  3420  stabilityLevel: ALPHA
  3421  labels:
  3422  - operation
  3423  - type
  3424- name: etcd_requests_total
  3425  help: Etcd request counts for each operation and object type.
  3426  type: Counter
  3427  stabilityLevel: ALPHA
  3428  labels:
  3429  - operation
  3430  - type
  3431- name: field_validation_request_duration_seconds
  3432  help: Response latency distribution in seconds for each field validation value
  3433  type: Histogram
  3434  stabilityLevel: ALPHA
  3435  labels:
  3436  - field_validation
  3437  buckets:
  3438  - 0.05
  3439  - 0.1
  3440  - 0.2
  3441  - 0.4
  3442  - 0.6
  3443  - 0.8
  3444  - 1
  3445  - 1.25
  3446  - 1.5
  3447  - 2
  3448  - 3
  3449  - 4
  3450  - 5
  3451  - 6
  3452  - 8
  3453  - 10
  3454  - 15
  3455  - 20
  3456  - 30
  3457  - 45
  3458  - 60
  3459- name: capacity
  3460  subsystem: watch_cache
  3461  help: Total capacity of watch cache broken by resource type.
  3462  type: Gauge
  3463  stabilityLevel: ALPHA
  3464  labels:
  3465  - resource
  3466- name: capacity_decrease_total
  3467  subsystem: watch_cache
  3468  help: Total number of watch cache capacity decrease events broken by resource type.
  3469  type: Counter
  3470  stabilityLevel: ALPHA
  3471  labels:
  3472  - resource
  3473- name: capacity_increase_total
  3474  subsystem: watch_cache
  3475  help: Total number of watch cache capacity increase events broken by resource type.
  3476  type: Counter
  3477  stabilityLevel: ALPHA
  3478  labels:
  3479  - resource
  3480- name: current_inflight_requests
  3481  subsystem: apiserver
  3482  help: Maximal number of currently used inflight request limit of this apiserver
  3483    per request kind in last second.
  3484  type: Gauge
  3485  stabilityLevel: STABLE
  3486  labels:
  3487  - request_kind
  3488- name: longrunning_requests
  3489  subsystem: apiserver
  3490  help: Gauge of all active long-running apiserver requests broken out by verb, group,
  3491    version, resource, scope and component. Not all requests are tracked this way.
  3492  type: Gauge
  3493  stabilityLevel: STABLE
  3494  labels:
  3495  - component
  3496  - group
  3497  - resource
  3498  - scope
  3499  - subresource
  3500  - verb
  3501  - version
  3502- name: request_duration_seconds
  3503  subsystem: apiserver
  3504  help: Response latency distribution in seconds for each verb, dry run value, group,
  3505    version, resource, subresource, scope and component.
  3506  type: Histogram
  3507  stabilityLevel: STABLE
  3508  labels:
  3509  - component
  3510  - dry_run
  3511  - group
  3512  - resource
  3513  - scope
  3514  - subresource
  3515  - verb
  3516  - version
  3517  buckets:
  3518  - 0.005
  3519  - 0.025
  3520  - 0.05
  3521  - 0.1
  3522  - 0.2
  3523  - 0.4
  3524  - 0.6
  3525  - 0.8
  3526  - 1
  3527  - 1.25
  3528  - 1.5
  3529  - 2
  3530  - 3
  3531  - 4
  3532  - 5
  3533  - 6
  3534  - 8
  3535  - 10
  3536  - 15
  3537  - 20
  3538  - 30
  3539  - 45
  3540  - 60
  3541- name: request_total
  3542  subsystem: apiserver
  3543  help: Counter of apiserver requests broken out for each verb, dry run value, group,
  3544    version, resource, scope, component, and HTTP response code.
  3545  type: Counter
  3546  stabilityLevel: STABLE
  3547  labels:
  3548  - code
  3549  - component
  3550  - dry_run
  3551  - group
  3552  - resource
  3553  - scope
  3554  - subresource
  3555  - verb
  3556  - version
  3557- name: requested_deprecated_apis
  3558  subsystem: apiserver
  3559  help: Gauge of deprecated APIs that have been requested, broken out by API group,
  3560    version, resource, subresource, and removed_release.
  3561  type: Gauge
  3562  stabilityLevel: STABLE
  3563  labels:
  3564  - group
  3565  - removed_release
  3566  - resource
  3567  - subresource
  3568  - version
  3569- name: response_sizes
  3570  subsystem: apiserver
  3571  help: Response size distribution in bytes for each group, version, verb, resource,
  3572    subresource, scope and component.
  3573  type: Histogram
  3574  stabilityLevel: STABLE
  3575  labels:
  3576  - component
  3577  - group
  3578  - resource
  3579  - scope
  3580  - subresource
  3581  - verb
  3582  - version
  3583  buckets:
  3584  - 1000
  3585  - 10000
  3586  - 100000
  3587  - 1e+06
  3588  - 1e+07
  3589  - 1e+08
  3590  - 1e+09
  3591- name: apiserver_storage_objects
  3592  help: Number of stored objects at the time of last check split by kind. In case
  3593    of a fetching error, the value will be -1.
  3594  type: Gauge
  3595  stabilityLevel: STABLE
  3596  labels:
  3597  - resource
  3598- name: dek_cache_fill_percent
  3599  subsystem: envelope_encryption
  3600  namespace: apiserver
  3601  help: Percent of the cache slots currently occupied by cached DEKs.
  3602  type: Gauge
  3603  stabilityLevel: ALPHA
  3604- name: dek_cache_inter_arrival_time_seconds
  3605  subsystem: envelope_encryption
  3606  namespace: apiserver
  3607  help: Time (in seconds) of inter arrival of transformation requests.
  3608  type: Histogram
  3609  stabilityLevel: ALPHA
  3610  labels:
  3611  - transformation_type
  3612  buckets:
  3613  - 60
  3614  - 120
  3615  - 240
  3616  - 480
  3617  - 960
  3618  - 1920
  3619  - 3840
  3620  - 7680
  3621  - 15360
  3622  - 30720
  3623- name: dek_source_cache_size
  3624  subsystem: envelope_encryption
  3625  namespace: apiserver
  3626  help: Number of records in data encryption key (DEK) source cache. On a restart,
  3627    this value is an approximation of the number of decrypt RPC calls the server will
  3628    make to the KMS plugin.
  3629  type: Gauge
  3630  stabilityLevel: ALPHA
  3631  labels:
  3632  - provider_name
  3633- name: invalid_key_id_from_status_total
  3634  subsystem: envelope_encryption
  3635  namespace: apiserver
  3636  help: Number of times an invalid keyID is returned by the Status RPC call split
  3637    by error.
  3638  type: Counter
  3639  stabilityLevel: ALPHA
  3640  labels:
  3641  - error
  3642  - provider_name
  3643- name: key_id_hash_last_timestamp_seconds
  3644  subsystem: envelope_encryption
  3645  namespace: apiserver
  3646  help: The last time in seconds when a keyID was used.
  3647  type: Gauge
  3648  stabilityLevel: ALPHA
  3649  labels:
  3650  - apiserver_id_hash
  3651  - key_id_hash
  3652  - provider_name
  3653  - transformation_type
  3654- name: key_id_hash_status_last_timestamp_seconds
  3655  subsystem: envelope_encryption
  3656  namespace: apiserver
  3657  help: The last time in seconds when a keyID was returned by the Status RPC call.
  3658  type: Gauge
  3659  stabilityLevel: ALPHA
  3660  labels:
  3661  - apiserver_id_hash
  3662  - key_id_hash
  3663  - provider_name
  3664- name: key_id_hash_total
  3665  subsystem: envelope_encryption
  3666  namespace: apiserver
  3667  help: Number of times a keyID is used split by transformation type, provider, and
  3668    apiserver identity.
  3669  type: Counter
  3670  stabilityLevel: ALPHA
  3671  labels:
  3672  - apiserver_id_hash
  3673  - key_id_hash
  3674  - provider_name
  3675  - transformation_type
  3676- name: kms_operations_latency_seconds
  3677  subsystem: envelope_encryption
  3678  namespace: apiserver
  3679  help: KMS operation duration with gRPC error code status total.
  3680  type: Histogram
  3681  stabilityLevel: ALPHA
  3682  labels:
  3683  - grpc_status_code
  3684  - method_name
  3685  - provider_name
  3686  buckets:
  3687  - 0.0001
  3688  - 0.0002
  3689  - 0.0004
  3690  - 0.0008
  3691  - 0.0016
  3692  - 0.0032
  3693  - 0.0064
  3694  - 0.0128
  3695  - 0.0256
  3696  - 0.0512
  3697  - 0.1024
  3698  - 0.2048
  3699  - 0.4096
  3700  - 0.8192
  3701  - 1.6384
  3702  - 3.2768
  3703  - 6.5536
  3704  - 13.1072
  3705  - 26.2144
  3706  - 52.4288
  3707- name: current_inqueue_seats
  3708  subsystem: flowcontrol
  3709  namespace: apiserver
  3710  help: Number of seats currently pending in queues of the API Priority and Fairness
  3711    subsystem
  3712  type: Gauge
  3713  stabilityLevel: ALPHA
  3714  labels:
  3715  - flow_schema
  3716  - priority_level
  3717- name: current_limit_seats
  3718  subsystem: flowcontrol
  3719  namespace: apiserver
  3720  help: current derived number of execution seats available to each priority level
  3721  type: Gauge
  3722  stabilityLevel: ALPHA
  3723  labels:
  3724  - priority_level
  3725- name: current_r
  3726  subsystem: flowcontrol
  3727  namespace: apiserver
  3728  help: R(time of last change)
  3729  type: Gauge
  3730  stabilityLevel: ALPHA
  3731  labels:
  3732  - priority_level
  3733- name: demand_seats
  3734  subsystem: flowcontrol
  3735  namespace: apiserver
  3736  help: Observations, at the end of every nanosecond, of (the number of seats each
  3737    priority level could use) / (nominal number of seats for that level)
  3738  type: TimingRatioHistogram
  3739  stabilityLevel: ALPHA
  3740  labels:
  3741  - priority_level
  3742  buckets:
  3743  - 0.2
  3744  - 0.4
  3745  - 0.6
  3746  - 0.8
  3747  - 1
  3748  - 1.2
  3749  - 1.4
  3750  - 1.7
  3751  - 2
  3752  - 2.8
  3753  - 4
  3754  - 6
  3755- name: demand_seats_average
  3756  subsystem: flowcontrol
  3757  namespace: apiserver
  3758  help: Time-weighted average, over last adjustment period, of demand_seats
  3759  type: Gauge
  3760  stabilityLevel: ALPHA
  3761  labels:
  3762  - priority_level
  3763- name: demand_seats_high_watermark
  3764  subsystem: flowcontrol
  3765  namespace: apiserver
  3766  help: High watermark, over last adjustment period, of demand_seats
  3767  type: Gauge
  3768  stabilityLevel: ALPHA
  3769  labels:
  3770  - priority_level
  3771- name: demand_seats_smoothed
  3772  subsystem: flowcontrol
  3773  namespace: apiserver
  3774  help: Smoothed seat demands
  3775  type: Gauge
  3776  stabilityLevel: ALPHA
  3777  labels:
  3778  - priority_level
  3779- name: demand_seats_stdev
  3780  subsystem: flowcontrol
  3781  namespace: apiserver
  3782  help: Time-weighted standard deviation, over last adjustment period, of demand_seats
  3783  type: Gauge
  3784  stabilityLevel: ALPHA
  3785  labels:
  3786  - priority_level
  3787- name: dispatch_r
  3788  subsystem: flowcontrol
  3789  namespace: apiserver
  3790  help: R(time of last dispatch)
  3791  type: Gauge
  3792  stabilityLevel: ALPHA
  3793  labels:
  3794  - priority_level
  3795- name: epoch_advance_total
  3796  subsystem: flowcontrol
  3797  namespace: apiserver
  3798  help: Number of times the queueset's progress meter jumped backward
  3799  type: Counter
  3800  stabilityLevel: ALPHA
  3801  labels:
  3802  - priority_level
  3803  - success
  3804- name: latest_s
  3805  subsystem: flowcontrol
  3806  namespace: apiserver
  3807  help: S(most recently dispatched request)
  3808  type: Gauge
  3809  stabilityLevel: ALPHA
  3810  labels:
  3811  - priority_level
  3812- name: lower_limit_seats
  3813  subsystem: flowcontrol
  3814  namespace: apiserver
  3815  help: Configured lower bound on number of execution seats available to each priority
  3816    level
  3817  type: Gauge
  3818  stabilityLevel: ALPHA
  3819  labels:
  3820  - priority_level
  3821- name: next_discounted_s_bounds
  3822  subsystem: flowcontrol
  3823  namespace: apiserver
  3824  help: min and max, over queues, of S(oldest waiting request in queue) - estimated
  3825    work in progress
  3826  type: Gauge
  3827  stabilityLevel: ALPHA
  3828  labels:
  3829  - bound
  3830  - priority_level
  3831- name: next_s_bounds
  3832  subsystem: flowcontrol
  3833  namespace: apiserver
  3834  help: min and max, over queues, of S(oldest waiting request in queue)
  3835  type: Gauge
  3836  stabilityLevel: ALPHA
  3837  labels:
  3838  - bound
  3839  - priority_level
  3840- name: priority_level_request_utilization
  3841  subsystem: flowcontrol
  3842  namespace: apiserver
  3843  help: Observations, at the end of every nanosecond, of number of requests (as a
  3844    fraction of the relevant limit) waiting or in any stage of execution (but only
  3845    initial stage for WATCHes)
  3846  type: TimingRatioHistogram
  3847  stabilityLevel: ALPHA
  3848  labels:
  3849  - phase
  3850  - priority_level
  3851  buckets:
  3852  - 0
  3853  - 0.001
  3854  - 0.003
  3855  - 0.01
  3856  - 0.03
  3857  - 0.1
  3858  - 0.25
  3859  - 0.5
  3860  - 0.75
  3861  - 1
  3862- name: priority_level_seat_utilization
  3863  subsystem: flowcontrol
  3864  namespace: apiserver
  3865  help: Observations, at the end of every nanosecond, of utilization of seats for
  3866    any stage of execution (but only initial stage for WATCHes)
  3867  type: TimingRatioHistogram
  3868  stabilityLevel: ALPHA
  3869  labels:
  3870  - priority_level
  3871  buckets:
  3872  - 0
  3873  - 0.1
  3874  - 0.2
  3875  - 0.3
  3876  - 0.4
  3877  - 0.5
  3878  - 0.6
  3879  - 0.7
  3880  - 0.8
  3881  - 0.9
  3882  - 0.95
  3883  - 0.99
  3884  - 1
  3885  constLabels:
  3886    phase: executing
  3887- name: read_vs_write_current_requests
  3888  subsystem: flowcontrol
  3889  namespace: apiserver
  3890  help: Observations, at the end of every nanosecond, of the number of requests (as
  3891    a fraction of the relevant limit) waiting or in regular stage of execution
  3892  type: TimingRatioHistogram
  3893  stabilityLevel: ALPHA
  3894  labels:
  3895  - phase
  3896  - request_kind
  3897  buckets:
  3898  - 0
  3899  - 0.001
  3900  - 0.01
  3901  - 0.1
  3902  - 0.2
  3903  - 0.3
  3904  - 0.4
  3905  - 0.5
  3906  - 0.6
  3907  - 0.7
  3908  - 0.8
  3909  - 0.9
  3910  - 0.95
  3911  - 0.99
  3912  - 1
  3913- name: request_concurrency_in_use
  3914  subsystem: flowcontrol
  3915  namespace: apiserver
  3916  help: Concurrency (number of seats) occupied by the currently executing (initial
  3917    stage for a WATCH, any stage otherwise) requests in the API Priority and Fairness
  3918    subsystem
  3919  type: Gauge
  3920  deprecatedVersion: 1.31.0
  3921  stabilityLevel: ALPHA
  3922  labels:
  3923  - flow_schema
  3924  - priority_level
  3925- name: request_concurrency_limit
  3926  subsystem: flowcontrol
  3927  namespace: apiserver
  3928  help: Nominal number of execution seats configured for each priority level
  3929  type: Gauge
  3930  deprecatedVersion: 1.30.0
  3931  stabilityLevel: ALPHA
  3932  labels:
  3933  - priority_level
  3934- name: request_dispatch_no_accommodation_total
  3935  subsystem: flowcontrol
  3936  namespace: apiserver
  3937  help: Number of times a dispatch attempt resulted in a non accommodation due to
  3938    lack of available seats
  3939  type: Counter
  3940  stabilityLevel: ALPHA
  3941  labels:
  3942  - flow_schema
  3943  - priority_level
  3944- name: request_execution_seconds
  3945  subsystem: flowcontrol
  3946  namespace: apiserver
  3947  help: Duration of initial stage (for a WATCH) or any (for a non-WATCH) stage of
  3948    request execution in the API Priority and Fairness subsystem
  3949  type: Histogram
  3950  stabilityLevel: ALPHA
  3951  labels:
  3952  - flow_schema
  3953  - priority_level
  3954  - type
  3955  buckets:
  3956  - 0
  3957  - 0.005
  3958  - 0.02
  3959  - 0.05
  3960  - 0.1
  3961  - 0.2
  3962  - 0.5
  3963  - 1
  3964  - 2
  3965  - 5
  3966  - 10
  3967  - 15
  3968  - 30
  3969- name: request_queue_length_after_enqueue
  3970  subsystem: flowcontrol
  3971  namespace: apiserver
  3972  help: Length of queue in the API Priority and Fairness subsystem, as seen by each
  3973    request after it is enqueued
  3974  type: Histogram
  3975  stabilityLevel: ALPHA
  3976  labels:
  3977  - flow_schema
  3978  - priority_level
  3979  buckets:
  3980  - 0
  3981  - 10
  3982  - 25
  3983  - 50
  3984  - 100
  3985  - 250
  3986  - 500
  3987  - 1000
  3988- name: seat_fair_frac
  3989  subsystem: flowcontrol
  3990  namespace: apiserver
  3991  help: Fair fraction of server's concurrency to allocate to each priority level that
  3992    can use it
  3993  type: Gauge
  3994  stabilityLevel: ALPHA
  3995- name: target_seats
  3996  subsystem: flowcontrol
  3997  namespace: apiserver
  3998  help: Seat allocation targets
  3999  type: Gauge
  4000  stabilityLevel: ALPHA
  4001  labels:
  4002  - priority_level
  4003- name: upper_limit_seats
  4004  subsystem: flowcontrol
  4005  namespace: apiserver
  4006  help: Configured upper bound on number of execution seats available to each priority
  4007    level
  4008  type: Gauge
  4009  stabilityLevel: ALPHA
  4010  labels:
  4011  - priority_level
  4012- name: watch_count_samples
  4013  subsystem: flowcontrol
  4014  namespace: apiserver
  4015  help: count of watchers for mutating requests in API Priority and Fairness
  4016  type: Histogram
  4017  stabilityLevel: ALPHA
  4018  labels:
  4019  - flow_schema
  4020  - priority_level
  4021  buckets:
  4022  - 0
  4023  - 1
  4024  - 10
  4025  - 100
  4026  - 1000
  4027  - 10000
  4028- name: work_estimated_seats
  4029  subsystem: flowcontrol
  4030  namespace: apiserver
  4031  help: Number of estimated seats (maximum of initial and final seats) associated
  4032    with requests in API Priority and Fairness
  4033  type: Histogram
  4034  stabilityLevel: ALPHA
  4035  labels:
  4036  - flow_schema
  4037  - priority_level
  4038  buckets:
  4039  - 1
  4040  - 2
  4041  - 4
  4042  - 10
  4043- name: rerouted_request_total
  4044  subsystem: apiserver
  4045  help: Total number of requests that were proxied to a peer kube apiserver because
  4046    the local apiserver was not capable of serving it
  4047  type: Counter
  4048  stabilityLevel: ALPHA
  4049  labels:
  4050  - code
  4051- name: data_key_generation_duration_seconds
  4052  subsystem: storage
  4053  namespace: apiserver
  4054  help: Latencies in seconds of data encryption key(DEK) generation operations.
  4055  type: Histogram
  4056  stabilityLevel: ALPHA
  4057  buckets:
  4058  - 5e-06
  4059  - 1e-05
  4060  - 2e-05
  4061  - 4e-05
  4062  - 8e-05
  4063  - 0.00016
  4064  - 0.00032
  4065  - 0.00064
  4066  - 0.00128
  4067  - 0.00256
  4068  - 0.00512
  4069  - 0.01024
  4070  - 0.02048
  4071  - 0.04096
  4072- name: data_key_generation_failures_total
  4073  subsystem: storage
  4074  namespace: apiserver
  4075  help: Total number of failed data encryption key(DEK) generation operations.
  4076  type: Counter
  4077  stabilityLevel: ALPHA
  4078- name: envelope_transformation_cache_misses_total
  4079  subsystem: storage
  4080  namespace: apiserver
  4081  help: Total number of cache misses while accessing key decryption key(KEK).
  4082  type: Counter
  4083  stabilityLevel: ALPHA
  4084- name: transformation_duration_seconds
  4085  subsystem: storage
  4086  namespace: apiserver
  4087  help: Latencies in seconds of value transformation operations.
  4088  type: Histogram
  4089  stabilityLevel: ALPHA
  4090  labels:
  4091  - transformation_type
  4092  - transformer_prefix
  4093  buckets:
  4094  - 5e-06
  4095  - 1e-05
  4096  - 2e-05
  4097  - 4e-05
  4098  - 8e-05
  4099  - 0.00016
  4100  - 0.00032
  4101  - 0.00064
  4102  - 0.00128
  4103  - 0.00256
  4104  - 0.00512
  4105  - 0.01024
  4106  - 0.02048
  4107  - 0.04096
  4108  - 0.08192
  4109  - 0.16384
  4110  - 0.32768
  4111  - 0.65536
  4112  - 1.31072
  4113  - 2.62144
  4114  - 5.24288
  4115  - 10.48576
  4116  - 20.97152
  4117  - 41.94304
  4118  - 83.88608
  4119- name: transformation_operations_total
  4120  subsystem: storage
  4121  namespace: apiserver
  4122  help: Total number of transformations. Successful transformation will have a status
  4123    'OK' and a varied status string when the transformation fails. This status and
  4124    transformation_type fields may be used for alerting on encryption/decryption failure
  4125    using transformation_type from_storage for decryption and to_storage for encryption
  4126  type: Counter
  4127  stabilityLevel: ALPHA
  4128  labels:
  4129  - status
  4130  - transformation_type
  4131  - transformer_prefix
  4132- name: x509_insecure_sha1_total
  4133  subsystem: webhooks
  4134  namespace: apiserver
  4135  help: Counts the number of requests to servers with insecure SHA1 signatures in
  4136    their serving certificate OR the number of connection failures due to the insecure
  4137    SHA1 signatures (either/or, based on the runtime environment)
  4138  type: Counter
  4139  stabilityLevel: ALPHA
  4140- name: x509_missing_san_total
  4141  subsystem: webhooks
  4142  namespace: apiserver
  4143  help: Counts the number of requests to servers missing SAN extension in their serving
  4144    certificate OR the number of connection failures due to the lack of x509 certificate
  4145    SAN extension missing (either/or, based on the runtime environment)
  4146  type: Counter
  4147  stabilityLevel: ALPHA
  4148- name: current_executing_requests
  4149  subsystem: flowcontrol
  4150  namespace: apiserver
  4151  help: Number of requests in initial (for a WATCH) or any (for a non-WATCH) execution
  4152    stage in the API Priority and Fairness subsystem
  4153  type: Gauge
  4154  stabilityLevel: BETA
  4155  labels:
  4156  - flow_schema
  4157  - priority_level
  4158- name: current_executing_seats
  4159  subsystem: flowcontrol
  4160  namespace: apiserver
  4161  help: Concurrency (number of seats) occupied by the currently executing (initial
  4162    stage for a WATCH, any stage otherwise) requests in the API Priority and Fairness
  4163    subsystem
  4164  type: Gauge
  4165  stabilityLevel: BETA
  4166  labels:
  4167  - flow_schema
  4168  - priority_level
  4169- name: current_inqueue_requests
  4170  subsystem: flowcontrol
  4171  namespace: apiserver
  4172  help: Number of requests currently pending in queues of the API Priority and Fairness
  4173    subsystem
  4174  type: Gauge
  4175  stabilityLevel: BETA
  4176  labels:
  4177  - flow_schema
  4178  - priority_level
  4179- name: dispatched_requests_total
  4180  subsystem: flowcontrol
  4181  namespace: apiserver
  4182  help: Number of requests executed by API Priority and Fairness subsystem
  4183  type: Counter
  4184  stabilityLevel: BETA
  4185  labels:
  4186  - flow_schema
  4187  - priority_level
  4188- name: nominal_limit_seats
  4189  subsystem: flowcontrol
  4190  namespace: apiserver
  4191  help: Nominal number of execution seats configured for each priority level
  4192  type: Gauge
  4193  stabilityLevel: BETA
  4194  labels:
  4195  - priority_level
  4196- name: rejected_requests_total
  4197  subsystem: flowcontrol
  4198  namespace: apiserver
  4199  help: Number of requests rejected by API Priority and Fairness subsystem
  4200  type: Counter
  4201  stabilityLevel: BETA
  4202  labels:
  4203  - flow_schema
  4204  - priority_level
  4205  - reason
  4206- name: request_wait_duration_seconds
  4207  subsystem: flowcontrol
  4208  namespace: apiserver
  4209  help: Length of time a request spent waiting in its queue
  4210  type: Histogram
  4211  stabilityLevel: BETA
  4212  labels:
  4213  - execute
  4214  - flow_schema
  4215  - priority_level
  4216  buckets:
  4217  - 0
  4218  - 0.005
  4219  - 0.02
  4220  - 0.05
  4221  - 0.1
  4222  - 0.2
  4223  - 0.5
  4224  - 1
  4225  - 2
  4226  - 5
  4227  - 10
  4228  - 15
  4229  - 30
  4230- name: request_duration_seconds
  4231  subsystem: cloud_provider_webhook
  4232  help: Request latency in seconds. Broken down by status code.
  4233  type: Histogram
  4234  stabilityLevel: ALPHA
  4235  labels:
  4236  - code
  4237  - webhook
  4238  buckets:
  4239  - 0.25
  4240  - 0.5
  4241  - 0.7
  4242  - 1
  4243  - 1.5
  4244  - 3
  4245  - 5
  4246  - 10
  4247- name: request_total
  4248  subsystem: cloud_provider_webhook
  4249  help: Number of HTTP requests partitioned by status code.
  4250  type: Counter
  4251  stabilityLevel: ALPHA
  4252  labels:
  4253  - code
  4254  - webhook
  4255- name: cloud_provider_taint_removal_delay_seconds
  4256  subsystem: node_controller
  4257  help: Number of seconds after node creation when NodeController removed the cloud-provider
  4258    taint of a single node.
  4259  type: Histogram
  4260  stabilityLevel: ALPHA
  4261  buckets:
  4262  - 1
  4263  - 4
  4264  - 16
  4265  - 64
  4266  - 256
  4267  - 1024
  4268- name: initial_node_sync_delay_seconds
  4269  subsystem: node_controller
  4270  help: Number of seconds after node creation when NodeController finished the initial
  4271    synchronization of a single node.
  4272  type: Histogram
  4273  stabilityLevel: ALPHA
  4274  buckets:
  4275  - 1
  4276  - 4
  4277  - 16
  4278  - 64
  4279  - 256
  4280  - 1024
  4281- name: loadbalancer_sync_total
  4282  subsystem: service_controller
  4283  help: A metric counting the amount of times any load balancer has been configured,
  4284    as an effect of service/node changes on the cluster
  4285  type: Counter
  4286  stabilityLevel: ALPHA
  4287- name: nodesync_error_total
  4288  subsystem: service_controller
  4289  help: A metric counting the amount of times any load balancer has been configured
  4290    and errored, as an effect of node changes on the cluster
  4291  type: Counter
  4292  stabilityLevel: ALPHA
  4293- name: nodesync_latency_seconds
  4294  subsystem: service_controller
  4295  help: A metric measuring the latency for nodesync which updates loadbalancer hosts
  4296    on cluster node updates.
  4297  type: Histogram
  4298  stabilityLevel: ALPHA
  4299  buckets:
  4300  - 1
  4301  - 2
  4302  - 4
  4303  - 8
  4304  - 16
  4305  - 32
  4306  - 64
  4307  - 128
  4308  - 256
  4309  - 512
  4310  - 1024
  4311  - 2048
  4312  - 4096
  4313  - 8192
  4314  - 16384
  4315- name: update_loadbalancer_host_latency_seconds
  4316  subsystem: service_controller
  4317  help: A metric measuring the latency for updating each load balancer hosts.
  4318  type: Histogram
  4319  stabilityLevel: ALPHA
  4320  buckets:
  4321  - 1
  4322  - 2
  4323  - 4
  4324  - 8
  4325  - 16
  4326  - 32
  4327  - 64
  4328  - 128
  4329  - 256
  4330  - 512
  4331  - 1024
  4332  - 2048
  4333  - 4096
  4334  - 8192
  4335  - 16384
  4336- name: changes
  4337  subsystem: endpoint_slice_controller
  4338  help: Number of EndpointSlice changes
  4339  type: Counter
  4340  stabilityLevel: ALPHA
  4341  labels:
  4342  - operation
  4343- name: desired_endpoint_slices
  4344  subsystem: endpoint_slice_controller
  4345  help: Number of EndpointSlices that would exist with perfect endpoint allocation
  4346  type: Gauge
  4347  stabilityLevel: ALPHA
  4348- name: endpoints_added_per_sync
  4349  subsystem: endpoint_slice_controller
  4350  help: Number of endpoints added on each Service sync
  4351  type: Histogram
  4352  stabilityLevel: ALPHA
  4353  buckets:
  4354  - 2
  4355  - 4
  4356  - 8
  4357  - 16
  4358  - 32
  4359  - 64
  4360  - 128
  4361  - 256
  4362  - 512
  4363  - 1024
  4364  - 2048
  4365  - 4096
  4366  - 8192
  4367  - 16384
  4368  - 32768
  4369- name: endpoints_desired
  4370  subsystem: endpoint_slice_controller
  4371  help: Number of endpoints desired
  4372  type: Gauge
  4373  stabilityLevel: ALPHA
  4374- name: endpoints_removed_per_sync
  4375  subsystem: endpoint_slice_controller
  4376  help: Number of endpoints removed on each Service sync
  4377  type: Histogram
  4378  stabilityLevel: ALPHA
  4379  buckets:
  4380  - 2
  4381  - 4
  4382  - 8
  4383  - 16
  4384  - 32
  4385  - 64
  4386  - 128
  4387  - 256
  4388  - 512
  4389  - 1024
  4390  - 2048
  4391  - 4096
  4392  - 8192
  4393  - 16384
  4394  - 32768
  4395- name: endpointslices_changed_per_sync
  4396  subsystem: endpoint_slice_controller
  4397  help: Number of EndpointSlices changed on each Service sync
  4398  type: Histogram
  4399  stabilityLevel: ALPHA
  4400  labels:
  4401  - topology
  4402- name: num_endpoint_slices
  4403  subsystem: endpoint_slice_controller
  4404  help: Number of EndpointSlices
  4405  type: Gauge
  4406  stabilityLevel: ALPHA
  4407- name: syncs
  4408  subsystem: endpoint_slice_controller
  4409  help: Number of EndpointSlice syncs
  4410  type: Counter
  4411  stabilityLevel: ALPHA
  4412  labels:
  4413  - result
  4414- name: kubernetes_build_info
  4415  help: A metric with a constant '1' value labeled by major, minor, git version, git
  4416    commit, git tree state, build date, Go version, and compiler from which Kubernetes
  4417    was built, and platform on which it is running.
  4418  type: Gauge
  4419  stabilityLevel: ALPHA
  4420  labels:
  4421  - build_date
  4422  - compiler
  4423  - git_commit
  4424  - git_tree_state
  4425  - git_version
  4426  - go_version
  4427  - major
  4428  - minor
  4429  - platform
  4430- name: leader_election_master_status
  4431  help: Gauge of if the reporting system is master of the relevant lease, 0 indicates
  4432    backup, 1 indicates master. 'name' is the string used to identify the lease. Please
  4433    make sure to group by name.
  4434  type: Gauge
  4435  stabilityLevel: ALPHA
  4436  labels:
  4437  - name
  4438- name: leader_election_slowpath_total
  4439  help: Total number of slow path exercised in renewing leader leases. 'name' is the
  4440    string used to identify the lease. Please make sure to group by name.
  4441  type: Counter
  4442  stabilityLevel: ALPHA
  4443  labels:
  4444  - name
  4445- name: rest_client_dns_resolution_duration_seconds
  4446  help: DNS resolver latency in seconds. Broken down by host.
  4447  type: Histogram
  4448  stabilityLevel: ALPHA
  4449  labels:
  4450  - host
  4451  buckets:
  4452  - 0.005
  4453  - 0.025
  4454  - 0.1
  4455  - 0.25
  4456  - 0.5
  4457  - 1
  4458  - 2
  4459  - 4
  4460  - 8
  4461  - 15
  4462  - 30
  4463- name: rest_client_exec_plugin_call_total
  4464  help: Number of calls to an exec plugin, partitioned by the type of event encountered
  4465    (no_error, plugin_execution_error, plugin_not_found_error, client_internal_error)
  4466    and an optional exit code. The exit code will be set to 0 if and only if the plugin
  4467    call was successful.
  4468  type: Counter
  4469  stabilityLevel: ALPHA
  4470  labels:
  4471  - call_status
  4472  - code
  4473- name: rest_client_exec_plugin_certificate_rotation_age
  4474  help: Histogram of the number of seconds the last auth exec plugin client certificate
  4475    lived before being rotated. If auth exec plugin client certificates are unused,
  4476    histogram will contain no data.
  4477  type: Histogram
  4478  stabilityLevel: ALPHA
  4479  buckets:
  4480  - 600
  4481  - 1800
  4482  - 3600
  4483  - 14400
  4484  - 86400
  4485  - 604800
  4486  - 2.592e+06
  4487  - 7.776e+06
  4488  - 1.5552e+07
  4489  - 3.1104e+07
  4490  - 1.24416e+08
  4491- name: rest_client_exec_plugin_ttl_seconds
  4492  help: Gauge of the shortest TTL (time-to-live) of the client certificate(s) managed
  4493    by the auth exec plugin. The value is in seconds until certificate expiry (negative
  4494    if already expired). If auth exec plugins are unused or manage no TLS certificates,
  4495    the value will be +INF.
  4496  type: Gauge
  4497  stabilityLevel: ALPHA
  4498- name: rest_client_rate_limiter_duration_seconds
  4499  help: Client side rate limiter latency in seconds. Broken down by verb, and host.
  4500  type: Histogram
  4501  stabilityLevel: ALPHA
  4502  labels:
  4503  - host
  4504  - verb
  4505  buckets:
  4506  - 0.005
  4507  - 0.025
  4508  - 0.1
  4509  - 0.25
  4510  - 0.5
  4511  - 1
  4512  - 2
  4513  - 4
  4514  - 8
  4515  - 15
  4516  - 30
  4517  - 60
  4518- name: rest_client_request_duration_seconds
  4519  help: Request latency in seconds. Broken down by verb, and host.
  4520  type: Histogram
  4521  stabilityLevel: ALPHA
  4522  labels:
  4523  - host
  4524  - verb
  4525  buckets:
  4526  - 0.005
  4527  - 0.025
  4528  - 0.1
  4529  - 0.25
  4530  - 0.5
  4531  - 1
  4532  - 2
  4533  - 4
  4534  - 8
  4535  - 15
  4536  - 30
  4537  - 60
  4538- name: rest_client_request_retries_total
  4539  help: Number of request retries, partitioned by status code, verb, and host.
  4540  type: Counter
  4541  stabilityLevel: ALPHA
  4542  labels:
  4543  - code
  4544  - host
  4545  - verb
  4546- name: rest_client_request_size_bytes
  4547  help: Request size in bytes. Broken down by verb and host.
  4548  type: Histogram
  4549  stabilityLevel: ALPHA
  4550  labels:
  4551  - host
  4552  - verb
  4553  buckets:
  4554  - 64
  4555  - 256
  4556  - 512
  4557  - 1024
  4558  - 4096
  4559  - 16384
  4560  - 65536
  4561  - 262144
  4562  - 1.048576e+06
  4563  - 4.194304e+06
  4564  - 1.6777216e+07
  4565- name: rest_client_requests_total
  4566  help: Number of HTTP requests, partitioned by status code, method, and host.
  4567  type: Counter
  4568  stabilityLevel: ALPHA
  4569  labels:
  4570  - code
  4571  - host
  4572  - method
  4573- name: rest_client_response_size_bytes
  4574  help: Response size in bytes. Broken down by verb and host.
  4575  type: Histogram
  4576  stabilityLevel: ALPHA
  4577  labels:
  4578  - host
  4579  - verb
  4580  buckets:
  4581  - 64
  4582  - 256
  4583  - 512
  4584  - 1024
  4585  - 4096
  4586  - 16384
  4587  - 65536
  4588  - 262144
  4589  - 1.048576e+06
  4590  - 4.194304e+06
  4591  - 1.6777216e+07
  4592- name: rest_client_transport_cache_entries
  4593  help: Number of transport entries in the internal cache.
  4594  type: Gauge
  4595  stabilityLevel: ALPHA
  4596- name: rest_client_transport_create_calls_total
  4597  help: 'Number of calls to get a new transport, partitioned by the result of the
  4598    operation hit: obtained from the cache, miss: created and added to the cache,
  4599    uncacheable: created and not cached'
  4600  type: Counter
  4601  stabilityLevel: ALPHA
  4602  labels:
  4603  - result
  4604- name: running_managed_controllers
  4605  help: Indicates where instances of a controller are currently running
  4606  type: Gauge
  4607  stabilityLevel: ALPHA
  4608  labels:
  4609  - manager
  4610  - name
  4611- name: adds_total
  4612  subsystem: workqueue
  4613  help: Total number of adds handled by workqueue
  4614  type: Counter
  4615  stabilityLevel: ALPHA
  4616  labels:
  4617  - name
  4618- name: depth
  4619  subsystem: workqueue
  4620  help: Current depth of workqueue
  4621  type: Gauge
  4622  stabilityLevel: ALPHA
  4623  labels:
  4624  - name
  4625- name: longest_running_processor_seconds
  4626  subsystem: workqueue
  4627  help: How many seconds has the longest running processor for workqueue been running.
  4628  type: Gauge
  4629  stabilityLevel: ALPHA
  4630  labels:
  4631  - name
  4632- name: queue_duration_seconds
  4633  subsystem: workqueue
  4634  help: How long in seconds an item stays in workqueue before being requested.
  4635  type: Histogram
  4636  stabilityLevel: ALPHA
  4637  labels:
  4638  - name
  4639  buckets:
  4640  - 1e-08
  4641  - 1e-07
  4642  - 1e-06
  4643  - 9.999999999999999e-06
  4644  - 9.999999999999999e-05
  4645  - 0.001
  4646  - 0.01
  4647  - 0.1
  4648  - 1
  4649  - 10
  4650- name: retries_total
  4651  subsystem: workqueue
  4652  help: Total number of retries handled by workqueue
  4653  type: Counter
  4654  stabilityLevel: ALPHA
  4655  labels:
  4656  - name
  4657- name: unfinished_work_seconds
  4658  subsystem: workqueue
  4659  help: How many seconds of work has done that is in progress and hasn't been observed
  4660    by work_duration. Large values indicate stuck threads. One can deduce the number
  4661    of stuck threads by observing the rate at which this increases.
  4662  type: Gauge
  4663  stabilityLevel: ALPHA
  4664  labels:
  4665  - name
  4666- name: work_duration_seconds
  4667  subsystem: workqueue
  4668  help: How long in seconds processing an item from workqueue takes.
  4669  type: Histogram
  4670  stabilityLevel: ALPHA
  4671  labels:
  4672  - name
  4673  buckets:
  4674  - 1e-08
  4675  - 1e-07
  4676  - 1e-06
  4677  - 9.999999999999999e-06
  4678  - 9.999999999999999e-05
  4679  - 0.001
  4680  - 0.01
  4681  - 0.1
  4682  - 1
  4683  - 10
  4684- name: disabled_metrics_total
  4685  help: The count of disabled metrics.
  4686  type: Counter
  4687  stabilityLevel: BETA
  4688- name: hidden_metrics_total
  4689  help: The count of hidden metrics.
  4690  type: Counter
  4691  stabilityLevel: BETA
  4692- name: feature_enabled
  4693  namespace: kubernetes
  4694  help: This metric records the data about the stage and enablement of a k8s feature.
  4695  type: Gauge
  4696  stabilityLevel: BETA
  4697  labels:
  4698  - name
  4699  - stage
  4700- name: registered_metrics_total
  4701  help: The count of registered metrics broken by stability level and deprecation
  4702    version.
  4703  type: Counter
  4704  stabilityLevel: BETA
  4705  labels:
  4706  - deprecated_version
  4707  - stability_level
  4708- name: healthcheck
  4709  namespace: kubernetes
  4710  help: This metric records the result of a single healthcheck.
  4711  type: Gauge
  4712  stabilityLevel: STABLE
  4713  labels:
  4714  - name
  4715  - type
  4716- name: healthchecks_total
  4717  namespace: kubernetes
  4718  help: This metric records the results of all healthcheck.
  4719  type: Counter
  4720  stabilityLevel: STABLE
  4721  labels:
  4722  - name
  4723  - status
  4724  - type
  4725- name: aggregator_openapi_v2_regeneration_count
  4726  help: Counter of OpenAPI v2 spec regeneration count broken down by causing APIService
  4727    name and reason.
  4728  type: Counter
  4729  stabilityLevel: ALPHA
  4730  labels:
  4731  - apiservice
  4732  - reason
  4733- name: aggregator_openapi_v2_regeneration_duration
  4734  help: Gauge of OpenAPI v2 spec regeneration duration in seconds.
  4735  type: Gauge
  4736  stabilityLevel: ALPHA
  4737  labels:
  4738  - reason
  4739- name: aggregator_unavailable_apiservice
  4740  help: Gauge of APIServices which are marked as unavailable broken down by APIService
  4741    name.
  4742  type: Custom
  4743  stabilityLevel: ALPHA
  4744  labels:
  4745  - name
  4746- name: aggregator_unavailable_apiservice_total
  4747  help: Counter of APIServices which are marked as unavailable broken down by APIService
  4748    name and reason.
  4749  type: Counter
  4750  stabilityLevel: ALPHA
  4751  labels:
  4752  - name
  4753  - reason
  4754- name: x509_insecure_sha1_total
  4755  subsystem: kube_aggregator
  4756  namespace: apiserver
  4757  help: Counts the number of requests to servers with insecure SHA1 signatures in
  4758    their serving certificate OR the number of connection failures due to the insecure
  4759    SHA1 signatures (either/or, based on the runtime environment)
  4760  type: Counter
  4761  stabilityLevel: ALPHA
  4762- name: x509_missing_san_total
  4763  subsystem: kube_aggregator
  4764  namespace: apiserver
  4765  help: Counts the number of requests to servers missing SAN extension in their serving
  4766    certificate OR the number of connection failures due to the lack of x509 certificate
  4767    SAN extension missing (either/or, based on the runtime environment)
  4768  type: Counter
  4769  stabilityLevel: ALPHA
  4770- name: api_request_duration_seconds
  4771  namespace: cloudprovider_azure
  4772  help: Latency of an Azure API call
  4773  type: Histogram
  4774  stabilityLevel: ALPHA
  4775  labels:
  4776  - request
  4777  - resource_group
  4778  - source
  4779  - subscription_id
  4780  buckets:
  4781  - 0.1
  4782  - 0.25
  4783  - 0.5
  4784  - 1
  4785  - 2.5
  4786  - 5
  4787  - 10
  4788  - 15
  4789  - 25
  4790  - 50
  4791  - 120
  4792  - 300
  4793  - 600
  4794  - 1200
  4795- name: api_request_errors
  4796  namespace: cloudprovider_azure
  4797  help: Number of errors for an Azure API call
  4798  type: Counter
  4799  stabilityLevel: ALPHA
  4800  labels:
  4801  - request
  4802  - resource_group
  4803  - source
  4804  - subscription_id
  4805- name: api_request_ratelimited_count
  4806  namespace: cloudprovider_azure
  4807  help: Number of rate limited Azure API calls
  4808  type: Counter
  4809  stabilityLevel: ALPHA
  4810  labels:
  4811  - request
  4812  - resource_group
  4813  - source
  4814  - subscription_id
  4815- name: api_request_throttled_count
  4816  namespace: cloudprovider_azure
  4817  help: Number of throttled Azure API calls
  4818  type: Counter
  4819  stabilityLevel: ALPHA
  4820  labels:
  4821  - request
  4822  - resource_group
  4823  - source
  4824  - subscription_id
  4825- name: op_duration_seconds
  4826  namespace: cloudprovider_azure
  4827  help: Latency of an Azure service operation
  4828  type: Histogram
  4829  stabilityLevel: ALPHA
  4830  labels:
  4831  - request
  4832  - resource_group
  4833  - source
  4834  - subscription_id
  4835  buckets:
  4836  - 0.1
  4837  - 0.2
  4838  - 0.5
  4839  - 1
  4840  - 10
  4841  - 20
  4842  - 30
  4843  - 40
  4844  - 50
  4845  - 60
  4846  - 100
  4847  - 200
  4848  - 300
  4849- name: op_failure_count
  4850  namespace: cloudprovider_azure
  4851  help: Number of failed Azure service operations
  4852  type: Counter
  4853  stabilityLevel: ALPHA
  4854  labels:
  4855  - request
  4856  - resource_group
  4857  - source
  4858  - subscription_id
  4859- name: cloudprovider_gce_api_request_duration_seconds
  4860  help: Latency of a GCE API call
  4861  type: Histogram
  4862  stabilityLevel: ALPHA
  4863  labels:
  4864  - region
  4865  - request
  4866  - version
  4867  - zone
  4868- name: cloudprovider_gce_api_request_errors
  4869  help: Number of errors for an API call
  4870  type: Counter
  4871  stabilityLevel: ALPHA
  4872  labels:
  4873  - region
  4874  - request
  4875  - version
  4876  - zone
  4877- name: cloudprovider_vsphere_api_request_duration_seconds
  4878  help: Latency of vsphere api call
  4879  type: Histogram
  4880  stabilityLevel: ALPHA
  4881  labels:
  4882  - request
  4883- name: cloudprovider_vsphere_api_request_errors
  4884  help: vsphere Api errors
  4885  type: Counter
  4886  stabilityLevel: ALPHA
  4887  labels:
  4888  - request
  4889- name: cloudprovider_vsphere_operation_duration_seconds
  4890  help: Latency of vsphere operation call
  4891  type: Histogram
  4892  stabilityLevel: ALPHA
  4893  labels:
  4894  - operation
  4895- name: cloudprovider_vsphere_operation_errors
  4896  help: vsphere operation errors
  4897  type: Counter
  4898  stabilityLevel: ALPHA
  4899  labels:
  4900  - operation
  4901- name: cloudprovider_vsphere_vcenter_versions
  4902  help: Versions for connected vSphere vCenters
  4903  type: Custom
  4904  stabilityLevel: ALPHA
  4905  labels:
  4906  - hostname
  4907  - version
  4908  - build
  4909- name: get_token_count
  4910  help: Counter of total Token() requests to the alternate token source
  4911  type: Counter
  4912  stabilityLevel: ALPHA
  4913- name: get_token_fail_count
  4914  help: Counter of failed Token() requests to the alternate token source
  4915  type: Counter
  4916  stabilityLevel: ALPHA
  4917- name: number_of_l4_ilbs
  4918  help: Number of L4 ILBs
  4919  type: Gauge
  4920  stabilityLevel: ALPHA
  4921  labels:
  4922  - feature
  4923- name: pod_security_errors_total
  4924  help: Number of errors preventing normal evaluation. Non-fatal errors may result
  4925    in the latest restricted profile being used for evaluation.
  4926  type: Counter
  4927  stabilityLevel: ALPHA
  4928  labels:
  4929  - fatal
  4930  - request_operation
  4931  - resource
  4932  - subresource
  4933- name: pod_security_evaluations_total
  4934  help: Number of policy evaluations that occurred, not counting ignored or exempt
  4935    requests.
  4936  type: Counter
  4937  stabilityLevel: ALPHA
  4938  labels:
  4939  - decision
  4940  - mode
  4941  - policy_level
  4942  - policy_version
  4943  - request_operation
  4944  - resource
  4945  - subresource
  4946- name: pod_security_exemptions_total
  4947  help: Number of exempt requests, not counting ignored or out of scope requests.
  4948  type: Counter
  4949  stabilityLevel: ALPHA
  4950  labels:
  4951  - request_operation
  4952  - resource
  4953  - subresource

View as plain text