# HELP go_gc_duration_seconds A summary of the pause duration of garbage collection cycles. # TYPE go_gc_duration_seconds summary go_gc_duration_seconds{quantile="0"} 0.00013468 go_gc_duration_seconds{quantile="0.25"} 0.000310529 go_gc_duration_seconds{quantile="0.5"} 0.000357761 go_gc_duration_seconds{quantile="0.75"} 0.000435266 go_gc_duration_seconds{quantile="1"} 0.031703182 go_gc_duration_seconds_sum 7.068330635 go_gc_duration_seconds_count 12300 # HELP go_goroutines Number of goroutines that currently exist. # TYPE go_goroutines gauge go_goroutines 360 # HELP go_info Information about the Go environment. # TYPE go_info gauge go_info{version="go1.15.2"} 1 # HELP go_memstats_alloc_bytes Number of bytes allocated and still in use. # TYPE go_memstats_alloc_bytes gauge go_memstats_alloc_bytes 5.80621536e+08 # HELP go_memstats_alloc_bytes_total Total number of bytes allocated, even if freed. # TYPE go_memstats_alloc_bytes_total counter go_memstats_alloc_bytes_total 2.349701297216e+12 # HELP go_memstats_buck_hash_sys_bytes Number of bytes used by the profiling bucket hash table. # TYPE go_memstats_buck_hash_sys_bytes gauge go_memstats_buck_hash_sys_bytes 3.45999e+06 # HELP go_memstats_frees_total Total number of frees. # TYPE go_memstats_frees_total counter go_memstats_frees_total 7.343867142e+09 # HELP go_memstats_gc_cpu_fraction The fraction of this program's available CPU time used by the GC since the program started. # TYPE go_memstats_gc_cpu_fraction gauge go_memstats_gc_cpu_fraction 0.00064467047181741 # HELP go_memstats_gc_sys_bytes Number of bytes used for garbage collection system metadata. # TYPE go_memstats_gc_sys_bytes gauge go_memstats_gc_sys_bytes 7.3659968e+07 # HELP go_memstats_heap_alloc_bytes Number of heap bytes allocated and still in use. # TYPE go_memstats_heap_alloc_bytes gauge go_memstats_heap_alloc_bytes 5.80621536e+08 # HELP go_memstats_heap_idle_bytes Number of heap bytes waiting to be used. # TYPE go_memstats_heap_idle_bytes gauge go_memstats_heap_idle_bytes 9.6833536e+08 # HELP go_memstats_heap_inuse_bytes Number of heap bytes that are in use. # TYPE go_memstats_heap_inuse_bytes gauge go_memstats_heap_inuse_bytes 7.02865408e+08 # HELP go_memstats_heap_objects Number of allocated objects. # TYPE go_memstats_heap_objects gauge go_memstats_heap_objects 4.352088e+06 # HELP go_memstats_heap_released_bytes Number of heap bytes released to OS. # TYPE go_memstats_heap_released_bytes gauge go_memstats_heap_released_bytes 7.2568832e+08 # HELP go_memstats_heap_sys_bytes Number of heap bytes obtained from system. # TYPE go_memstats_heap_sys_bytes gauge go_memstats_heap_sys_bytes 1.671200768e+09 # HELP go_memstats_last_gc_time_seconds Number of seconds since 1970 of last garbage collection. # TYPE go_memstats_last_gc_time_seconds gauge go_memstats_last_gc_time_seconds 1.7749317018967757e+09 # HELP go_memstats_lookups_total Total number of pointer lookups. # TYPE go_memstats_lookups_total counter go_memstats_lookups_total 0 # HELP go_memstats_mallocs_total Total number of mallocs. # TYPE go_memstats_mallocs_total counter go_memstats_mallocs_total 7.34821923e+09 # HELP go_memstats_mcache_inuse_bytes Number of bytes in use by mcache structures. # TYPE go_memstats_mcache_inuse_bytes gauge go_memstats_mcache_inuse_bytes 13888 # HELP go_memstats_mcache_sys_bytes Number of bytes used for mcache structures obtained from system. # TYPE go_memstats_mcache_sys_bytes gauge go_memstats_mcache_sys_bytes 16384 # HELP go_memstats_mspan_inuse_bytes Number of bytes in use by mspan structures. # TYPE go_memstats_mspan_inuse_bytes gauge go_memstats_mspan_inuse_bytes 8.818104e+06 # HELP go_memstats_mspan_sys_bytes Number of bytes used for mspan structures obtained from system. # TYPE go_memstats_mspan_sys_bytes gauge go_memstats_mspan_sys_bytes 2.3740416e+07 # HELP go_memstats_next_gc_bytes Number of heap bytes when next garbage collection will take place. # TYPE go_memstats_next_gc_bytes gauge go_memstats_next_gc_bytes 1.15961312e+09 # HELP go_memstats_other_sys_bytes Number of bytes used for other system allocations. # TYPE go_memstats_other_sys_bytes gauge go_memstats_other_sys_bytes 3.654634e+06 # HELP go_memstats_stack_inuse_bytes Number of bytes in use by the stack allocator. # TYPE go_memstats_stack_inuse_bytes gauge go_memstats_stack_inuse_bytes 6.520832e+06 # HELP go_memstats_stack_sys_bytes Number of bytes obtained from system for stack allocator. # TYPE go_memstats_stack_sys_bytes gauge go_memstats_stack_sys_bytes 6.520832e+06 # HELP go_memstats_sys_bytes Number of bytes obtained from system. # TYPE go_memstats_sys_bytes gauge go_memstats_sys_bytes 1.782252992e+09 # HELP go_threads Number of OS threads created. # TYPE go_threads gauge go_threads 16 # HELP net_conntrack_dialer_conn_attempted_total Total number of connections attempted by the given dialer a given name. # TYPE net_conntrack_dialer_conn_attempted_total counter net_conntrack_dialer_conn_attempted_total{dialer_name="alertmanager"} 1 net_conntrack_dialer_conn_attempted_total{dialer_name="default"} 0 net_conntrack_dialer_conn_attempted_total{dialer_name="kubernetes-apiservers"} 3 net_conntrack_dialer_conn_attempted_total{dialer_name="kubernetes-nodes"} 10 net_conntrack_dialer_conn_attempted_total{dialer_name="kubernetes-nodes-cadvisor"} 23 net_conntrack_dialer_conn_attempted_total{dialer_name="kubernetes-pods"} 23507 net_conntrack_dialer_conn_attempted_total{dialer_name="kubernetes-pods-slow"} 0 net_conntrack_dialer_conn_attempted_total{dialer_name="kubernetes-service-endpoints"} 25 net_conntrack_dialer_conn_attempted_total{dialer_name="kubernetes-service-endpoints-slow"} 0 net_conntrack_dialer_conn_attempted_total{dialer_name="kubernetes-services"} 0 net_conntrack_dialer_conn_attempted_total{dialer_name="opsmx_ssd_metrics"} 635916 net_conntrack_dialer_conn_attempted_total{dialer_name="otel-collector-redica-instance"} 397 net_conntrack_dialer_conn_attempted_total{dialer_name="prometheus"} 1 net_conntrack_dialer_conn_attempted_total{dialer_name="prometheus-pushgateway"} 0 # HELP net_conntrack_dialer_conn_closed_total Total number of connections closed which originated from the dialer of a given name. # TYPE net_conntrack_dialer_conn_closed_total counter net_conntrack_dialer_conn_closed_total{dialer_name="alertmanager"} 0 net_conntrack_dialer_conn_closed_total{dialer_name="default"} 0 net_conntrack_dialer_conn_closed_total{dialer_name="kubernetes-apiservers"} 2 net_conntrack_dialer_conn_closed_total{dialer_name="kubernetes-nodes"} 3 net_conntrack_dialer_conn_closed_total{dialer_name="kubernetes-nodes-cadvisor"} 18 net_conntrack_dialer_conn_closed_total{dialer_name="kubernetes-pods"} 23498 net_conntrack_dialer_conn_closed_total{dialer_name="kubernetes-pods-slow"} 0 net_conntrack_dialer_conn_closed_total{dialer_name="kubernetes-service-endpoints"} 4 net_conntrack_dialer_conn_closed_total{dialer_name="kubernetes-service-endpoints-slow"} 0 net_conntrack_dialer_conn_closed_total{dialer_name="kubernetes-services"} 0 net_conntrack_dialer_conn_closed_total{dialer_name="opsmx_ssd_metrics"} 588791 net_conntrack_dialer_conn_closed_total{dialer_name="otel-collector-redica-instance"} 396 net_conntrack_dialer_conn_closed_total{dialer_name="prometheus"} 0 net_conntrack_dialer_conn_closed_total{dialer_name="prometheus-pushgateway"} 0 # HELP net_conntrack_dialer_conn_established_total Total number of connections successfully established by the given dialer a given name. # TYPE net_conntrack_dialer_conn_established_total counter net_conntrack_dialer_conn_established_total{dialer_name="alertmanager"} 1 net_conntrack_dialer_conn_established_total{dialer_name="default"} 0 net_conntrack_dialer_conn_established_total{dialer_name="kubernetes-apiservers"} 3 net_conntrack_dialer_conn_established_total{dialer_name="kubernetes-nodes"} 4 net_conntrack_dialer_conn_established_total{dialer_name="kubernetes-nodes-cadvisor"} 19 net_conntrack_dialer_conn_established_total{dialer_name="kubernetes-pods"} 23506 net_conntrack_dialer_conn_established_total{dialer_name="kubernetes-pods-slow"} 0 net_conntrack_dialer_conn_established_total{dialer_name="kubernetes-service-endpoints"} 16 net_conntrack_dialer_conn_established_total{dialer_name="kubernetes-service-endpoints-slow"} 0 net_conntrack_dialer_conn_established_total{dialer_name="kubernetes-services"} 0 net_conntrack_dialer_conn_established_total{dialer_name="opsmx_ssd_metrics"} 588831 net_conntrack_dialer_conn_established_total{dialer_name="otel-collector-redica-instance"} 397 net_conntrack_dialer_conn_established_total{dialer_name="prometheus"} 1 net_conntrack_dialer_conn_established_total{dialer_name="prometheus-pushgateway"} 0 # HELP net_conntrack_dialer_conn_failed_total Total number of connections failed to dial by the dialer a given name. # TYPE net_conntrack_dialer_conn_failed_total counter net_conntrack_dialer_conn_failed_total{dialer_name="alertmanager",reason="refused"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="alertmanager",reason="resolution"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="alertmanager",reason="timeout"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="alertmanager",reason="unknown"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="default",reason="refused"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="default",reason="resolution"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="default",reason="timeout"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="default",reason="unknown"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="kubernetes-apiservers",reason="refused"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="kubernetes-apiservers",reason="resolution"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="kubernetes-apiservers",reason="timeout"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="kubernetes-apiservers",reason="unknown"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="kubernetes-nodes",reason="refused"} 6 net_conntrack_dialer_conn_failed_total{dialer_name="kubernetes-nodes",reason="resolution"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="kubernetes-nodes",reason="timeout"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="kubernetes-nodes",reason="unknown"} 6 net_conntrack_dialer_conn_failed_total{dialer_name="kubernetes-nodes-cadvisor",reason="refused"} 4 net_conntrack_dialer_conn_failed_total{dialer_name="kubernetes-nodes-cadvisor",reason="resolution"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="kubernetes-nodes-cadvisor",reason="timeout"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="kubernetes-nodes-cadvisor",reason="unknown"} 4 net_conntrack_dialer_conn_failed_total{dialer_name="kubernetes-pods",reason="refused"} 1 net_conntrack_dialer_conn_failed_total{dialer_name="kubernetes-pods",reason="resolution"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="kubernetes-pods",reason="timeout"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="kubernetes-pods",reason="unknown"} 1 net_conntrack_dialer_conn_failed_total{dialer_name="kubernetes-pods-slow",reason="refused"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="kubernetes-pods-slow",reason="resolution"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="kubernetes-pods-slow",reason="timeout"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="kubernetes-pods-slow",reason="unknown"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="kubernetes-service-endpoints",reason="refused"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="kubernetes-service-endpoints",reason="resolution"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="kubernetes-service-endpoints",reason="timeout"} 9 net_conntrack_dialer_conn_failed_total{dialer_name="kubernetes-service-endpoints",reason="unknown"} 9 net_conntrack_dialer_conn_failed_total{dialer_name="kubernetes-service-endpoints-slow",reason="refused"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="kubernetes-service-endpoints-slow",reason="resolution"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="kubernetes-service-endpoints-slow",reason="timeout"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="kubernetes-service-endpoints-slow",reason="unknown"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="kubernetes-services",reason="refused"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="kubernetes-services",reason="resolution"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="kubernetes-services",reason="timeout"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="kubernetes-services",reason="unknown"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="opsmx_ssd_metrics",reason="refused"} 47085 net_conntrack_dialer_conn_failed_total{dialer_name="opsmx_ssd_metrics",reason="resolution"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="opsmx_ssd_metrics",reason="timeout"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="opsmx_ssd_metrics",reason="unknown"} 47085 net_conntrack_dialer_conn_failed_total{dialer_name="otel-collector-redica-instance",reason="refused"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="otel-collector-redica-instance",reason="resolution"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="otel-collector-redica-instance",reason="timeout"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="otel-collector-redica-instance",reason="unknown"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="prometheus",reason="refused"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="prometheus",reason="resolution"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="prometheus",reason="timeout"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="prometheus",reason="unknown"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="prometheus-pushgateway",reason="refused"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="prometheus-pushgateway",reason="resolution"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="prometheus-pushgateway",reason="timeout"} 0 net_conntrack_dialer_conn_failed_total{dialer_name="prometheus-pushgateway",reason="unknown"} 0 # HELP net_conntrack_listener_conn_accepted_total Total number of connections opened to the listener of a given name. # TYPE net_conntrack_listener_conn_accepted_total counter net_conntrack_listener_conn_accepted_total{listener_name="http"} 380911 # HELP net_conntrack_listener_conn_closed_total Total number of connections closed that were made to the listener of a given name. # TYPE net_conntrack_listener_conn_closed_total counter net_conntrack_listener_conn_closed_total{listener_name="http"} 380912 # HELP process_cpu_seconds_total Total user and system CPU time spent in seconds. # TYPE process_cpu_seconds_total counter process_cpu_seconds_total 53211.49 # HELP process_max_fds Maximum number of open file descriptors. # TYPE process_max_fds gauge process_max_fds 1.048576e+06 # HELP process_open_fds Number of open file descriptors. # TYPE process_open_fds gauge process_open_fds 131 # HELP process_resident_memory_bytes Resident memory size in bytes. # TYPE process_resident_memory_bytes gauge process_resident_memory_bytes 1.402626048e+09 # HELP process_start_time_seconds Start time of the process since unix epoch in seconds. # TYPE process_start_time_seconds gauge process_start_time_seconds 1.77352193327e+09 # HELP process_virtual_memory_bytes Virtual memory size in bytes. # TYPE process_virtual_memory_bytes gauge process_virtual_memory_bytes 9.10850048e+09 # HELP process_virtual_memory_max_bytes Maximum amount of virtual memory available in bytes. # TYPE process_virtual_memory_max_bytes gauge process_virtual_memory_max_bytes -1 # HELP prometheus_api_remote_read_queries The current number of remote read queries being executed or waiting. # TYPE prometheus_api_remote_read_queries gauge prometheus_api_remote_read_queries 0 # HELP prometheus_build_info A metric with a constant '1' value labeled by version, revision, branch, and goversion from which prometheus was built. # TYPE prometheus_build_info gauge prometheus_build_info{branch="HEAD",goversion="go1.15.2",revision="e83ef207b6c2398919b69cd87d2693cfc2fb4127",version="2.21.0"} 1 # HELP prometheus_config_last_reload_success_timestamp_seconds Timestamp of the last successful configuration reload. # TYPE prometheus_config_last_reload_success_timestamp_seconds gauge prometheus_config_last_reload_success_timestamp_seconds 1.7735219477140858e+09 # HELP prometheus_config_last_reload_successful Whether the last configuration reload attempt was successful. # TYPE prometheus_config_last_reload_successful gauge prometheus_config_last_reload_successful 1 # HELP prometheus_engine_queries The current number of queries being executed or waiting. # TYPE prometheus_engine_queries gauge prometheus_engine_queries 0 # HELP prometheus_engine_queries_concurrent_max The max number of concurrent queries. # TYPE prometheus_engine_queries_concurrent_max gauge prometheus_engine_queries_concurrent_max 20 # HELP prometheus_engine_query_duration_seconds Query timings # TYPE prometheus_engine_query_duration_seconds summary prometheus_engine_query_duration_seconds{slice="inner_eval",quantile="0.5"} 0.000460741 prometheus_engine_query_duration_seconds{slice="inner_eval",quantile="0.9"} 0.007404221 prometheus_engine_query_duration_seconds{slice="inner_eval",quantile="0.99"} 0.010090897 prometheus_engine_query_duration_seconds_sum{slice="inner_eval"} 486.51677401199936 prometheus_engine_query_duration_seconds_count{slice="inner_eval"} 164535 prometheus_engine_query_duration_seconds{slice="prepare_time",quantile="0.5"} 0.000206004 prometheus_engine_query_duration_seconds{slice="prepare_time",quantile="0.9"} 0.002913252 prometheus_engine_query_duration_seconds{slice="prepare_time",quantile="0.99"} 0.003735649 prometheus_engine_query_duration_seconds_sum{slice="prepare_time"} 159.6361400169998 prometheus_engine_query_duration_seconds_count{slice="prepare_time"} 164535 prometheus_engine_query_duration_seconds{slice="queue_time",quantile="0.5"} 4.9258e-05 prometheus_engine_query_duration_seconds{slice="queue_time",quantile="0.9"} 7.3521e-05 prometheus_engine_query_duration_seconds{slice="queue_time",quantile="0.99"} 0.000132802 prometheus_engine_query_duration_seconds_sum{slice="queue_time"} 8.107088265000067 prometheus_engine_query_duration_seconds_count{slice="queue_time"} 164535 prometheus_engine_query_duration_seconds{slice="result_sort",quantile="0.5"} NaN prometheus_engine_query_duration_seconds{slice="result_sort",quantile="0.9"} NaN prometheus_engine_query_duration_seconds{slice="result_sort",quantile="0.99"} NaN prometheus_engine_query_duration_seconds_sum{slice="result_sort"} 0.0033560160000000007 prometheus_engine_query_duration_seconds_count{slice="result_sort"} 56 # HELP prometheus_engine_query_log_enabled State of the query log. # TYPE prometheus_engine_query_log_enabled gauge prometheus_engine_query_log_enabled 0 # HELP prometheus_engine_query_log_failures_total The number of query log failures. # TYPE prometheus_engine_query_log_failures_total counter prometheus_engine_query_log_failures_total 0 # HELP prometheus_http_request_duration_seconds Histogram of latencies for HTTP requests. # TYPE prometheus_http_request_duration_seconds histogram prometheus_http_request_duration_seconds_bucket{handler="/",le="0.1"} 71 prometheus_http_request_duration_seconds_bucket{handler="/",le="0.2"} 71 prometheus_http_request_duration_seconds_bucket{handler="/",le="0.4"} 71 prometheus_http_request_duration_seconds_bucket{handler="/",le="1"} 71 prometheus_http_request_duration_seconds_bucket{handler="/",le="3"} 71 prometheus_http_request_duration_seconds_bucket{handler="/",le="8"} 71 prometheus_http_request_duration_seconds_bucket{handler="/",le="20"} 71 prometheus_http_request_duration_seconds_bucket{handler="/",le="60"} 71 prometheus_http_request_duration_seconds_bucket{handler="/",le="120"} 71 prometheus_http_request_duration_seconds_bucket{handler="/",le="+Inf"} 71 prometheus_http_request_duration_seconds_sum{handler="/"} 0.0019274080000000002 prometheus_http_request_duration_seconds_count{handler="/"} 71 prometheus_http_request_duration_seconds_bucket{handler="/-/healthy",le="0.1"} 93982 prometheus_http_request_duration_seconds_bucket{handler="/-/healthy",le="0.2"} 93982 prometheus_http_request_duration_seconds_bucket{handler="/-/healthy",le="0.4"} 93982 prometheus_http_request_duration_seconds_bucket{handler="/-/healthy",le="1"} 93982 prometheus_http_request_duration_seconds_bucket{handler="/-/healthy",le="3"} 93982 prometheus_http_request_duration_seconds_bucket{handler="/-/healthy",le="8"} 93982 prometheus_http_request_duration_seconds_bucket{handler="/-/healthy",le="20"} 93982 prometheus_http_request_duration_seconds_bucket{handler="/-/healthy",le="60"} 93982 prometheus_http_request_duration_seconds_bucket{handler="/-/healthy",le="120"} 93982 prometheus_http_request_duration_seconds_bucket{handler="/-/healthy",le="+Inf"} 93982 prometheus_http_request_duration_seconds_sum{handler="/-/healthy"} 1.7411757729999906 prometheus_http_request_duration_seconds_count{handler="/-/healthy"} 93982 prometheus_http_request_duration_seconds_bucket{handler="/-/ready",le="0.1"} 281948 prometheus_http_request_duration_seconds_bucket{handler="/-/ready",le="0.2"} 281948 prometheus_http_request_duration_seconds_bucket{handler="/-/ready",le="0.4"} 281948 prometheus_http_request_duration_seconds_bucket{handler="/-/ready",le="1"} 281948 prometheus_http_request_duration_seconds_bucket{handler="/-/ready",le="3"} 281948 prometheus_http_request_duration_seconds_bucket{handler="/-/ready",le="8"} 281948 prometheus_http_request_duration_seconds_bucket{handler="/-/ready",le="20"} 281948 prometheus_http_request_duration_seconds_bucket{handler="/-/ready",le="60"} 281948 prometheus_http_request_duration_seconds_bucket{handler="/-/ready",le="120"} 281948 prometheus_http_request_duration_seconds_bucket{handler="/-/ready",le="+Inf"} 281948 prometheus_http_request_duration_seconds_sum{handler="/-/ready"} 5.785699900000067 prometheus_http_request_duration_seconds_count{handler="/-/ready"} 281948 prometheus_http_request_duration_seconds_bucket{handler="/alerts",le="0.1"} 2 prometheus_http_request_duration_seconds_bucket{handler="/alerts",le="0.2"} 2 prometheus_http_request_duration_seconds_bucket{handler="/alerts",le="0.4"} 2 prometheus_http_request_duration_seconds_bucket{handler="/alerts",le="1"} 2 prometheus_http_request_duration_seconds_bucket{handler="/alerts",le="3"} 2 prometheus_http_request_duration_seconds_bucket{handler="/alerts",le="8"} 2 prometheus_http_request_duration_seconds_bucket{handler="/alerts",le="20"} 2 prometheus_http_request_duration_seconds_bucket{handler="/alerts",le="60"} 2 prometheus_http_request_duration_seconds_bucket{handler="/alerts",le="120"} 2 prometheus_http_request_duration_seconds_bucket{handler="/alerts",le="+Inf"} 2 prometheus_http_request_duration_seconds_sum{handler="/alerts"} 0.036841873 prometheus_http_request_duration_seconds_count{handler="/alerts"} 2 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/label/:name/values",le="0.1"} 11 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/label/:name/values",le="0.2"} 11 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/label/:name/values",le="0.4"} 13 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/label/:name/values",le="1"} 14 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/label/:name/values",le="3"} 14 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/label/:name/values",le="8"} 14 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/label/:name/values",le="20"} 14 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/label/:name/values",le="60"} 14 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/label/:name/values",le="120"} 14 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/label/:name/values",le="+Inf"} 14 prometheus_http_request_duration_seconds_sum{handler="/api/v1/label/:name/values"} 1.3626831839999998 prometheus_http_request_duration_seconds_count{handler="/api/v1/label/:name/values"} 14 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/labels",le="0.1"} 4 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/labels",le="0.2"} 4 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/labels",le="0.4"} 4 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/labels",le="1"} 4 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/labels",le="3"} 4 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/labels",le="8"} 4 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/labels",le="20"} 4 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/labels",le="60"} 4 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/labels",le="120"} 4 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/labels",le="+Inf"} 4 prometheus_http_request_duration_seconds_sum{handler="/api/v1/labels"} 0.029584043 prometheus_http_request_duration_seconds_count{handler="/api/v1/labels"} 4 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/query",le="0.1"} 7 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/query",le="0.2"} 7 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/query",le="0.4"} 7 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/query",le="1"} 7 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/query",le="3"} 7 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/query",le="8"} 7 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/query",le="20"} 7 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/query",le="60"} 7 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/query",le="120"} 7 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/query",le="+Inf"} 7 prometheus_http_request_duration_seconds_sum{handler="/api/v1/query"} 0.049615417 prometheus_http_request_duration_seconds_count{handler="/api/v1/query"} 7 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/query_range",le="0.1"} 40 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/query_range",le="0.2"} 45 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/query_range",le="0.4"} 56 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/query_range",le="1"} 56 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/query_range",le="3"} 56 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/query_range",le="8"} 56 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/query_range",le="20"} 56 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/query_range",le="60"} 56 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/query_range",le="120"} 56 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/query_range",le="+Inf"} 56 prometheus_http_request_duration_seconds_sum{handler="/api/v1/query_range"} 4.417942130999999 prometheus_http_request_duration_seconds_count{handler="/api/v1/query_range"} 56 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/status/config",le="0.1"} 4 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/status/config",le="0.2"} 4 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/status/config",le="0.4"} 4 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/status/config",le="1"} 4 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/status/config",le="3"} 4 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/status/config",le="8"} 4 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/status/config",le="20"} 4 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/status/config",le="60"} 4 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/status/config",le="120"} 4 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/status/config",le="+Inf"} 4 prometheus_http_request_duration_seconds_sum{handler="/api/v1/status/config"} 0.068532366 prometheus_http_request_duration_seconds_count{handler="/api/v1/status/config"} 4 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/targets",le="0.1"} 0 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/targets",le="0.2"} 7 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/targets",le="0.4"} 8 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/targets",le="1"} 8 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/targets",le="3"} 8 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/targets",le="8"} 8 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/targets",le="20"} 8 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/targets",le="60"} 8 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/targets",le="120"} 8 prometheus_http_request_duration_seconds_bucket{handler="/api/v1/targets",le="+Inf"} 8 prometheus_http_request_duration_seconds_sum{handler="/api/v1/targets"} 1.231552441 prometheus_http_request_duration_seconds_count{handler="/api/v1/targets"} 8 prometheus_http_request_duration_seconds_bucket{handler="/config",le="0.1"} 2 prometheus_http_request_duration_seconds_bucket{handler="/config",le="0.2"} 2 prometheus_http_request_duration_seconds_bucket{handler="/config",le="0.4"} 2 prometheus_http_request_duration_seconds_bucket{handler="/config",le="1"} 2 prometheus_http_request_duration_seconds_bucket{handler="/config",le="3"} 2 prometheus_http_request_duration_seconds_bucket{handler="/config",le="8"} 2 prometheus_http_request_duration_seconds_bucket{handler="/config",le="20"} 2 prometheus_http_request_duration_seconds_bucket{handler="/config",le="60"} 2 prometheus_http_request_duration_seconds_bucket{handler="/config",le="120"} 2 prometheus_http_request_duration_seconds_bucket{handler="/config",le="+Inf"} 2 prometheus_http_request_duration_seconds_sum{handler="/config"} 0.011226902 prometheus_http_request_duration_seconds_count{handler="/config"} 2 prometheus_http_request_duration_seconds_bucket{handler="/debug/*subpath",le="0.1"} 2 prometheus_http_request_duration_seconds_bucket{handler="/debug/*subpath",le="0.2"} 2 prometheus_http_request_duration_seconds_bucket{handler="/debug/*subpath",le="0.4"} 2 prometheus_http_request_duration_seconds_bucket{handler="/debug/*subpath",le="1"} 2 prometheus_http_request_duration_seconds_bucket{handler="/debug/*subpath",le="3"} 2 prometheus_http_request_duration_seconds_bucket{handler="/debug/*subpath",le="8"} 2 prometheus_http_request_duration_seconds_bucket{handler="/debug/*subpath",le="20"} 2 prometheus_http_request_duration_seconds_bucket{handler="/debug/*subpath",le="60"} 2 prometheus_http_request_duration_seconds_bucket{handler="/debug/*subpath",le="120"} 2 prometheus_http_request_duration_seconds_bucket{handler="/debug/*subpath",le="+Inf"} 2 prometheus_http_request_duration_seconds_sum{handler="/debug/*subpath"} 0.001374527 prometheus_http_request_duration_seconds_count{handler="/debug/*subpath"} 2 prometheus_http_request_duration_seconds_bucket{handler="/flags",le="0.1"} 1 prometheus_http_request_duration_seconds_bucket{handler="/flags",le="0.2"} 1 prometheus_http_request_duration_seconds_bucket{handler="/flags",le="0.4"} 1 prometheus_http_request_duration_seconds_bucket{handler="/flags",le="1"} 1 prometheus_http_request_duration_seconds_bucket{handler="/flags",le="3"} 1 prometheus_http_request_duration_seconds_bucket{handler="/flags",le="8"} 1 prometheus_http_request_duration_seconds_bucket{handler="/flags",le="20"} 1 prometheus_http_request_duration_seconds_bucket{handler="/flags",le="60"} 1 prometheus_http_request_duration_seconds_bucket{handler="/flags",le="120"} 1 prometheus_http_request_duration_seconds_bucket{handler="/flags",le="+Inf"} 1 prometheus_http_request_duration_seconds_sum{handler="/flags"} 0.006170064 prometheus_http_request_duration_seconds_count{handler="/flags"} 1 prometheus_http_request_duration_seconds_bucket{handler="/graph",le="0.1"} 66 prometheus_http_request_duration_seconds_bucket{handler="/graph",le="0.2"} 66 prometheus_http_request_duration_seconds_bucket{handler="/graph",le="0.4"} 66 prometheus_http_request_duration_seconds_bucket{handler="/graph",le="1"} 66 prometheus_http_request_duration_seconds_bucket{handler="/graph",le="3"} 66 prometheus_http_request_duration_seconds_bucket{handler="/graph",le="8"} 66 prometheus_http_request_duration_seconds_bucket{handler="/graph",le="20"} 66 prometheus_http_request_duration_seconds_bucket{handler="/graph",le="60"} 66 prometheus_http_request_duration_seconds_bucket{handler="/graph",le="120"} 66 prometheus_http_request_duration_seconds_bucket{handler="/graph",le="+Inf"} 66 prometheus_http_request_duration_seconds_sum{handler="/graph"} 0.32814468499999994 prometheus_http_request_duration_seconds_count{handler="/graph"} 66 prometheus_http_request_duration_seconds_bucket{handler="/metrics",le="0.1"} 23494 prometheus_http_request_duration_seconds_bucket{handler="/metrics",le="0.2"} 23497 prometheus_http_request_duration_seconds_bucket{handler="/metrics",le="0.4"} 23497 prometheus_http_request_duration_seconds_bucket{handler="/metrics",le="1"} 23497 prometheus_http_request_duration_seconds_bucket{handler="/metrics",le="3"} 23497 prometheus_http_request_duration_seconds_bucket{handler="/metrics",le="8"} 23497 prometheus_http_request_duration_seconds_bucket{handler="/metrics",le="20"} 23497 prometheus_http_request_duration_seconds_bucket{handler="/metrics",le="60"} 23497 prometheus_http_request_duration_seconds_bucket{handler="/metrics",le="120"} 23497 prometheus_http_request_duration_seconds_bucket{handler="/metrics",le="+Inf"} 23497 prometheus_http_request_duration_seconds_sum{handler="/metrics"} 213.64189823500027 prometheus_http_request_duration_seconds_count{handler="/metrics"} 23497 prometheus_http_request_duration_seconds_bucket{handler="/new/*filepath",le="0.1"} 9 prometheus_http_request_duration_seconds_bucket{handler="/new/*filepath",le="0.2"} 9 prometheus_http_request_duration_seconds_bucket{handler="/new/*filepath",le="0.4"} 9 prometheus_http_request_duration_seconds_bucket{handler="/new/*filepath",le="1"} 9 prometheus_http_request_duration_seconds_bucket{handler="/new/*filepath",le="3"} 9 prometheus_http_request_duration_seconds_bucket{handler="/new/*filepath",le="8"} 9 prometheus_http_request_duration_seconds_bucket{handler="/new/*filepath",le="20"} 9 prometheus_http_request_duration_seconds_bucket{handler="/new/*filepath",le="60"} 9 prometheus_http_request_duration_seconds_bucket{handler="/new/*filepath",le="120"} 9 prometheus_http_request_duration_seconds_bucket{handler="/new/*filepath",le="+Inf"} 9 prometheus_http_request_duration_seconds_sum{handler="/new/*filepath"} 0.0027630529999999997 prometheus_http_request_duration_seconds_count{handler="/new/*filepath"} 9 prometheus_http_request_duration_seconds_bucket{handler="/rules",le="0.1"} 3 prometheus_http_request_duration_seconds_bucket{handler="/rules",le="0.2"} 3 prometheus_http_request_duration_seconds_bucket{handler="/rules",le="0.4"} 3 prometheus_http_request_duration_seconds_bucket{handler="/rules",le="1"} 3 prometheus_http_request_duration_seconds_bucket{handler="/rules",le="3"} 3 prometheus_http_request_duration_seconds_bucket{handler="/rules",le="8"} 3 prometheus_http_request_duration_seconds_bucket{handler="/rules",le="20"} 3 prometheus_http_request_duration_seconds_bucket{handler="/rules",le="60"} 3 prometheus_http_request_duration_seconds_bucket{handler="/rules",le="120"} 3 prometheus_http_request_duration_seconds_bucket{handler="/rules",le="+Inf"} 3 prometheus_http_request_duration_seconds_sum{handler="/rules"} 0.067344969 prometheus_http_request_duration_seconds_count{handler="/rules"} 3 prometheus_http_request_duration_seconds_bucket{handler="/service-discovery",le="0.1"} 0 prometheus_http_request_duration_seconds_bucket{handler="/service-discovery",le="0.2"} 0 prometheus_http_request_duration_seconds_bucket{handler="/service-discovery",le="0.4"} 2 prometheus_http_request_duration_seconds_bucket{handler="/service-discovery",le="1"} 2 prometheus_http_request_duration_seconds_bucket{handler="/service-discovery",le="3"} 2 prometheus_http_request_duration_seconds_bucket{handler="/service-discovery",le="8"} 2 prometheus_http_request_duration_seconds_bucket{handler="/service-discovery",le="20"} 2 prometheus_http_request_duration_seconds_bucket{handler="/service-discovery",le="60"} 2 prometheus_http_request_duration_seconds_bucket{handler="/service-discovery",le="120"} 2 prometheus_http_request_duration_seconds_bucket{handler="/service-discovery",le="+Inf"} 2 prometheus_http_request_duration_seconds_sum{handler="/service-discovery"} 0.649227226 prometheus_http_request_duration_seconds_count{handler="/service-discovery"} 2 prometheus_http_request_duration_seconds_bucket{handler="/static/*filepath",le="0.1"} 124 prometheus_http_request_duration_seconds_bucket{handler="/static/*filepath",le="0.2"} 124 prometheus_http_request_duration_seconds_bucket{handler="/static/*filepath",le="0.4"} 124 prometheus_http_request_duration_seconds_bucket{handler="/static/*filepath",le="1"} 124 prometheus_http_request_duration_seconds_bucket{handler="/static/*filepath",le="3"} 124 prometheus_http_request_duration_seconds_bucket{handler="/static/*filepath",le="8"} 124 prometheus_http_request_duration_seconds_bucket{handler="/static/*filepath",le="20"} 124 prometheus_http_request_duration_seconds_bucket{handler="/static/*filepath",le="60"} 124 prometheus_http_request_duration_seconds_bucket{handler="/static/*filepath",le="120"} 124 prometheus_http_request_duration_seconds_bucket{handler="/static/*filepath",le="+Inf"} 124 prometheus_http_request_duration_seconds_sum{handler="/static/*filepath"} 0.16647659599999995 prometheus_http_request_duration_seconds_count{handler="/static/*filepath"} 124 prometheus_http_request_duration_seconds_bucket{handler="/status",le="0.1"} 2 prometheus_http_request_duration_seconds_bucket{handler="/status",le="0.2"} 2 prometheus_http_request_duration_seconds_bucket{handler="/status",le="0.4"} 2 prometheus_http_request_duration_seconds_bucket{handler="/status",le="1"} 2 prometheus_http_request_duration_seconds_bucket{handler="/status",le="3"} 2 prometheus_http_request_duration_seconds_bucket{handler="/status",le="8"} 2 prometheus_http_request_duration_seconds_bucket{handler="/status",le="20"} 2 prometheus_http_request_duration_seconds_bucket{handler="/status",le="60"} 2 prometheus_http_request_duration_seconds_bucket{handler="/status",le="120"} 2 prometheus_http_request_duration_seconds_bucket{handler="/status",le="+Inf"} 2 prometheus_http_request_duration_seconds_sum{handler="/status"} 0.037025153000000005 prometheus_http_request_duration_seconds_count{handler="/status"} 2 prometheus_http_request_duration_seconds_bucket{handler="/targets",le="0.1"} 1 prometheus_http_request_duration_seconds_bucket{handler="/targets",le="0.2"} 2 prometheus_http_request_duration_seconds_bucket{handler="/targets",le="0.4"} 2 prometheus_http_request_duration_seconds_bucket{handler="/targets",le="1"} 2 prometheus_http_request_duration_seconds_bucket{handler="/targets",le="3"} 2 prometheus_http_request_duration_seconds_bucket{handler="/targets",le="8"} 2 prometheus_http_request_duration_seconds_bucket{handler="/targets",le="20"} 2 prometheus_http_request_duration_seconds_bucket{handler="/targets",le="60"} 2 prometheus_http_request_duration_seconds_bucket{handler="/targets",le="120"} 2 prometheus_http_request_duration_seconds_bucket{handler="/targets",le="+Inf"} 2 prometheus_http_request_duration_seconds_sum{handler="/targets"} 0.19012327299999998 prometheus_http_request_duration_seconds_count{handler="/targets"} 2 prometheus_http_request_duration_seconds_bucket{handler="/version",le="0.1"} 1 prometheus_http_request_duration_seconds_bucket{handler="/version",le="0.2"} 1 prometheus_http_request_duration_seconds_bucket{handler="/version",le="0.4"} 1 prometheus_http_request_duration_seconds_bucket{handler="/version",le="1"} 1 prometheus_http_request_duration_seconds_bucket{handler="/version",le="3"} 1 prometheus_http_request_duration_seconds_bucket{handler="/version",le="8"} 1 prometheus_http_request_duration_seconds_bucket{handler="/version",le="20"} 1 prometheus_http_request_duration_seconds_bucket{handler="/version",le="60"} 1 prometheus_http_request_duration_seconds_bucket{handler="/version",le="120"} 1 prometheus_http_request_duration_seconds_bucket{handler="/version",le="+Inf"} 1 prometheus_http_request_duration_seconds_sum{handler="/version"} 0.008873218 prometheus_http_request_duration_seconds_count{handler="/version"} 1 # HELP prometheus_http_requests_total Counter of HTTP requests. # TYPE prometheus_http_requests_total counter prometheus_http_requests_total{code="200",handler="/-/healthy"} 93982 prometheus_http_requests_total{code="200",handler="/-/ready"} 281948 prometheus_http_requests_total{code="200",handler="/alerts"} 2 prometheus_http_requests_total{code="200",handler="/api/v1/label/:name/values"} 14 prometheus_http_requests_total{code="200",handler="/api/v1/labels"} 4 prometheus_http_requests_total{code="200",handler="/api/v1/query"} 7 prometheus_http_requests_total{code="200",handler="/api/v1/query_range"} 56 prometheus_http_requests_total{code="200",handler="/api/v1/status/config"} 4 prometheus_http_requests_total{code="200",handler="/api/v1/targets"} 8 prometheus_http_requests_total{code="200",handler="/config"} 2 prometheus_http_requests_total{code="200",handler="/flags"} 1 prometheus_http_requests_total{code="200",handler="/graph"} 66 prometheus_http_requests_total{code="200",handler="/metrics"} 23497 prometheus_http_requests_total{code="200",handler="/rules"} 3 prometheus_http_requests_total{code="200",handler="/service-discovery"} 2 prometheus_http_requests_total{code="200",handler="/static/*filepath"} 117 prometheus_http_requests_total{code="200",handler="/status"} 2 prometheus_http_requests_total{code="200",handler="/targets"} 2 prometheus_http_requests_total{code="200",handler="/version"} 1 prometheus_http_requests_total{code="301",handler="/static/*filepath"} 2 prometheus_http_requests_total{code="302",handler="/"} 71 prometheus_http_requests_total{code="404",handler="/debug/*subpath"} 2 prometheus_http_requests_total{code="404",handler="/new/*filepath"} 9 prometheus_http_requests_total{code="404",handler="/static/*filepath"} 5 # HELP prometheus_http_response_size_bytes Histogram of response size for HTTP requests. # TYPE prometheus_http_response_size_bytes histogram prometheus_http_response_size_bytes_bucket{handler="/",le="100"} 71 prometheus_http_response_size_bytes_bucket{handler="/",le="1000"} 71 prometheus_http_response_size_bytes_bucket{handler="/",le="10000"} 71 prometheus_http_response_size_bytes_bucket{handler="/",le="100000"} 71 prometheus_http_response_size_bytes_bucket{handler="/",le="1e+06"} 71 prometheus_http_response_size_bytes_bucket{handler="/",le="1e+07"} 71 prometheus_http_response_size_bytes_bucket{handler="/",le="1e+08"} 71 prometheus_http_response_size_bytes_bucket{handler="/",le="1e+09"} 71 prometheus_http_response_size_bytes_bucket{handler="/",le="+Inf"} 71 prometheus_http_response_size_bytes_sum{handler="/"} 2059 prometheus_http_response_size_bytes_count{handler="/"} 71 prometheus_http_response_size_bytes_bucket{handler="/-/healthy",le="100"} 93982 prometheus_http_response_size_bytes_bucket{handler="/-/healthy",le="1000"} 93982 prometheus_http_response_size_bytes_bucket{handler="/-/healthy",le="10000"} 93982 prometheus_http_response_size_bytes_bucket{handler="/-/healthy",le="100000"} 93982 prometheus_http_response_size_bytes_bucket{handler="/-/healthy",le="1e+06"} 93982 prometheus_http_response_size_bytes_bucket{handler="/-/healthy",le="1e+07"} 93982 prometheus_http_response_size_bytes_bucket{handler="/-/healthy",le="1e+08"} 93982 prometheus_http_response_size_bytes_bucket{handler="/-/healthy",le="1e+09"} 93982 prometheus_http_response_size_bytes_bucket{handler="/-/healthy",le="+Inf"} 93982 prometheus_http_response_size_bytes_sum{handler="/-/healthy"} 2.161586e+06 prometheus_http_response_size_bytes_count{handler="/-/healthy"} 93982 prometheus_http_response_size_bytes_bucket{handler="/-/ready",le="100"} 281948 prometheus_http_response_size_bytes_bucket{handler="/-/ready",le="1000"} 281948 prometheus_http_response_size_bytes_bucket{handler="/-/ready",le="10000"} 281948 prometheus_http_response_size_bytes_bucket{handler="/-/ready",le="100000"} 281948 prometheus_http_response_size_bytes_bucket{handler="/-/ready",le="1e+06"} 281948 prometheus_http_response_size_bytes_bucket{handler="/-/ready",le="1e+07"} 281948 prometheus_http_response_size_bytes_bucket{handler="/-/ready",le="1e+08"} 281948 prometheus_http_response_size_bytes_bucket{handler="/-/ready",le="1e+09"} 281948 prometheus_http_response_size_bytes_bucket{handler="/-/ready",le="+Inf"} 281948 prometheus_http_response_size_bytes_sum{handler="/-/ready"} 5.920908e+06 prometheus_http_response_size_bytes_count{handler="/-/ready"} 281948 prometheus_http_response_size_bytes_bucket{handler="/alerts",le="100"} 0 prometheus_http_response_size_bytes_bucket{handler="/alerts",le="1000"} 0 prometheus_http_response_size_bytes_bucket{handler="/alerts",le="10000"} 0 prometheus_http_response_size_bytes_bucket{handler="/alerts",le="100000"} 2 prometheus_http_response_size_bytes_bucket{handler="/alerts",le="1e+06"} 2 prometheus_http_response_size_bytes_bucket{handler="/alerts",le="1e+07"} 2 prometheus_http_response_size_bytes_bucket{handler="/alerts",le="1e+08"} 2 prometheus_http_response_size_bytes_bucket{handler="/alerts",le="1e+09"} 2 prometheus_http_response_size_bytes_bucket{handler="/alerts",le="+Inf"} 2 prometheus_http_response_size_bytes_sum{handler="/alerts"} 34592 prometheus_http_response_size_bytes_count{handler="/alerts"} 2 prometheus_http_response_size_bytes_bucket{handler="/api/v1/label/:name/values",le="100"} 4 prometheus_http_response_size_bytes_bucket{handler="/api/v1/label/:name/values",le="1000"} 8 prometheus_http_response_size_bytes_bucket{handler="/api/v1/label/:name/values",le="10000"} 14 prometheus_http_response_size_bytes_bucket{handler="/api/v1/label/:name/values",le="100000"} 14 prometheus_http_response_size_bytes_bucket{handler="/api/v1/label/:name/values",le="1e+06"} 14 prometheus_http_response_size_bytes_bucket{handler="/api/v1/label/:name/values",le="1e+07"} 14 prometheus_http_response_size_bytes_bucket{handler="/api/v1/label/:name/values",le="1e+08"} 14 prometheus_http_response_size_bytes_bucket{handler="/api/v1/label/:name/values",le="1e+09"} 14 prometheus_http_response_size_bytes_bucket{handler="/api/v1/label/:name/values",le="+Inf"} 14 prometheus_http_response_size_bytes_sum{handler="/api/v1/label/:name/values"} 56672 prometheus_http_response_size_bytes_count{handler="/api/v1/label/:name/values"} 14 prometheus_http_response_size_bytes_bucket{handler="/api/v1/labels",le="100"} 0 prometheus_http_response_size_bytes_bucket{handler="/api/v1/labels",le="1000"} 0 prometheus_http_response_size_bytes_bucket{handler="/api/v1/labels",le="10000"} 4 prometheus_http_response_size_bytes_bucket{handler="/api/v1/labels",le="100000"} 4 prometheus_http_response_size_bytes_bucket{handler="/api/v1/labels",le="1e+06"} 4 prometheus_http_response_size_bytes_bucket{handler="/api/v1/labels",le="1e+07"} 4 prometheus_http_response_size_bytes_bucket{handler="/api/v1/labels",le="1e+08"} 4 prometheus_http_response_size_bytes_bucket{handler="/api/v1/labels",le="1e+09"} 4 prometheus_http_response_size_bytes_bucket{handler="/api/v1/labels",le="+Inf"} 4 prometheus_http_response_size_bytes_sum{handler="/api/v1/labels"} 5408 prometheus_http_response_size_bytes_count{handler="/api/v1/labels"} 4 prometheus_http_response_size_bytes_bucket{handler="/api/v1/query",le="100"} 0 prometheus_http_response_size_bytes_bucket{handler="/api/v1/query",le="1000"} 6 prometheus_http_response_size_bytes_bucket{handler="/api/v1/query",le="10000"} 7 prometheus_http_response_size_bytes_bucket{handler="/api/v1/query",le="100000"} 7 prometheus_http_response_size_bytes_bucket{handler="/api/v1/query",le="1e+06"} 7 prometheus_http_response_size_bytes_bucket{handler="/api/v1/query",le="1e+07"} 7 prometheus_http_response_size_bytes_bucket{handler="/api/v1/query",le="1e+08"} 7 prometheus_http_response_size_bytes_bucket{handler="/api/v1/query",le="1e+09"} 7 prometheus_http_response_size_bytes_bucket{handler="/api/v1/query",le="+Inf"} 7 prometheus_http_response_size_bytes_sum{handler="/api/v1/query"} 3897 prometheus_http_response_size_bytes_count{handler="/api/v1/query"} 7 prometheus_http_response_size_bytes_bucket{handler="/api/v1/query_range",le="100"} 42 prometheus_http_response_size_bytes_bucket{handler="/api/v1/query_range",le="1000"} 56 prometheus_http_response_size_bytes_bucket{handler="/api/v1/query_range",le="10000"} 56 prometheus_http_response_size_bytes_bucket{handler="/api/v1/query_range",le="100000"} 56 prometheus_http_response_size_bytes_bucket{handler="/api/v1/query_range",le="1e+06"} 56 prometheus_http_response_size_bytes_bucket{handler="/api/v1/query_range",le="1e+07"} 56 prometheus_http_response_size_bytes_bucket{handler="/api/v1/query_range",le="1e+08"} 56 prometheus_http_response_size_bytes_bucket{handler="/api/v1/query_range",le="1e+09"} 56 prometheus_http_response_size_bytes_bucket{handler="/api/v1/query_range",le="+Inf"} 56 prometheus_http_response_size_bytes_sum{handler="/api/v1/query_range"} 15332 prometheus_http_response_size_bytes_count{handler="/api/v1/query_range"} 56 prometheus_http_response_size_bytes_bucket{handler="/api/v1/status/config",le="100"} 0 prometheus_http_response_size_bytes_bucket{handler="/api/v1/status/config",le="1000"} 0 prometheus_http_response_size_bytes_bucket{handler="/api/v1/status/config",le="10000"} 4 prometheus_http_response_size_bytes_bucket{handler="/api/v1/status/config",le="100000"} 4 prometheus_http_response_size_bytes_bucket{handler="/api/v1/status/config",le="1e+06"} 4 prometheus_http_response_size_bytes_bucket{handler="/api/v1/status/config",le="1e+07"} 4 prometheus_http_response_size_bytes_bucket{handler="/api/v1/status/config",le="1e+08"} 4 prometheus_http_response_size_bytes_bucket{handler="/api/v1/status/config",le="1e+09"} 4 prometheus_http_response_size_bytes_bucket{handler="/api/v1/status/config",le="+Inf"} 4 prometheus_http_response_size_bytes_sum{handler="/api/v1/status/config"} 5432 prometheus_http_response_size_bytes_count{handler="/api/v1/status/config"} 4 prometheus_http_response_size_bytes_bucket{handler="/api/v1/targets",le="100"} 0 prometheus_http_response_size_bytes_bucket{handler="/api/v1/targets",le="1000"} 0 prometheus_http_response_size_bytes_bucket{handler="/api/v1/targets",le="10000"} 0 prometheus_http_response_size_bytes_bucket{handler="/api/v1/targets",le="100000"} 0 prometheus_http_response_size_bytes_bucket{handler="/api/v1/targets",le="1e+06"} 8 prometheus_http_response_size_bytes_bucket{handler="/api/v1/targets",le="1e+07"} 8 prometheus_http_response_size_bytes_bucket{handler="/api/v1/targets",le="1e+08"} 8 prometheus_http_response_size_bytes_bucket{handler="/api/v1/targets",le="1e+09"} 8 prometheus_http_response_size_bytes_bucket{handler="/api/v1/targets",le="+Inf"} 8 prometheus_http_response_size_bytes_sum{handler="/api/v1/targets"} 1.229241e+06 prometheus_http_response_size_bytes_count{handler="/api/v1/targets"} 8 prometheus_http_response_size_bytes_bucket{handler="/config",le="100"} 0 prometheus_http_response_size_bytes_bucket{handler="/config",le="1000"} 0 prometheus_http_response_size_bytes_bucket{handler="/config",le="10000"} 0 prometheus_http_response_size_bytes_bucket{handler="/config",le="100000"} 2 prometheus_http_response_size_bytes_bucket{handler="/config",le="1e+06"} 2 prometheus_http_response_size_bytes_bucket{handler="/config",le="1e+07"} 2 prometheus_http_response_size_bytes_bucket{handler="/config",le="1e+08"} 2 prometheus_http_response_size_bytes_bucket{handler="/config",le="1e+09"} 2 prometheus_http_response_size_bytes_bucket{handler="/config",le="+Inf"} 2 prometheus_http_response_size_bytes_sum{handler="/config"} 33176 prometheus_http_response_size_bytes_count{handler="/config"} 2 prometheus_http_response_size_bytes_bucket{handler="/debug/*subpath",le="100"} 2 prometheus_http_response_size_bytes_bucket{handler="/debug/*subpath",le="1000"} 2 prometheus_http_response_size_bytes_bucket{handler="/debug/*subpath",le="10000"} 2 prometheus_http_response_size_bytes_bucket{handler="/debug/*subpath",le="100000"} 2 prometheus_http_response_size_bytes_bucket{handler="/debug/*subpath",le="1e+06"} 2 prometheus_http_response_size_bytes_bucket{handler="/debug/*subpath",le="1e+07"} 2 prometheus_http_response_size_bytes_bucket{handler="/debug/*subpath",le="1e+08"} 2 prometheus_http_response_size_bytes_bucket{handler="/debug/*subpath",le="1e+09"} 2 prometheus_http_response_size_bytes_bucket{handler="/debug/*subpath",le="+Inf"} 2 prometheus_http_response_size_bytes_sum{handler="/debug/*subpath"} 38 prometheus_http_response_size_bytes_count{handler="/debug/*subpath"} 2 prometheus_http_response_size_bytes_bucket{handler="/flags",le="100"} 0 prometheus_http_response_size_bytes_bucket{handler="/flags",le="1000"} 0 prometheus_http_response_size_bytes_bucket{handler="/flags",le="10000"} 1 prometheus_http_response_size_bytes_bucket{handler="/flags",le="100000"} 1 prometheus_http_response_size_bytes_bucket{handler="/flags",le="1e+06"} 1 prometheus_http_response_size_bytes_bucket{handler="/flags",le="1e+07"} 1 prometheus_http_response_size_bytes_bucket{handler="/flags",le="1e+08"} 1 prometheus_http_response_size_bytes_bucket{handler="/flags",le="1e+09"} 1 prometheus_http_response_size_bytes_bucket{handler="/flags",le="+Inf"} 1 prometheus_http_response_size_bytes_sum{handler="/flags"} 8272 prometheus_http_response_size_bytes_count{handler="/flags"} 1 prometheus_http_response_size_bytes_bucket{handler="/graph",le="100"} 0 prometheus_http_response_size_bytes_bucket{handler="/graph",le="1000"} 0 prometheus_http_response_size_bytes_bucket{handler="/graph",le="10000"} 66 prometheus_http_response_size_bytes_bucket{handler="/graph",le="100000"} 66 prometheus_http_response_size_bytes_bucket{handler="/graph",le="1e+06"} 66 prometheus_http_response_size_bytes_bucket{handler="/graph",le="1e+07"} 66 prometheus_http_response_size_bytes_bucket{handler="/graph",le="1e+08"} 66 prometheus_http_response_size_bytes_bucket{handler="/graph",le="1e+09"} 66 prometheus_http_response_size_bytes_bucket{handler="/graph",le="+Inf"} 66 prometheus_http_response_size_bytes_sum{handler="/graph"} 394284 prometheus_http_response_size_bytes_count{handler="/graph"} 66 prometheus_http_response_size_bytes_bucket{handler="/metrics",le="100"} 0 prometheus_http_response_size_bytes_bucket{handler="/metrics",le="1000"} 0 prometheus_http_response_size_bytes_bucket{handler="/metrics",le="10000"} 1 prometheus_http_response_size_bytes_bucket{handler="/metrics",le="100000"} 23497 prometheus_http_response_size_bytes_bucket{handler="/metrics",le="1e+06"} 23497 prometheus_http_response_size_bytes_bucket{handler="/metrics",le="1e+07"} 23497 prometheus_http_response_size_bytes_bucket{handler="/metrics",le="1e+08"} 23497 prometheus_http_response_size_bytes_bucket{handler="/metrics",le="1e+09"} 23497 prometheus_http_response_size_bytes_bucket{handler="/metrics",le="+Inf"} 23497 prometheus_http_response_size_bytes_sum{handler="/metrics"} 2.78875438e+08 prometheus_http_response_size_bytes_count{handler="/metrics"} 23497 prometheus_http_response_size_bytes_bucket{handler="/new/*filepath",le="100"} 9 prometheus_http_response_size_bytes_bucket{handler="/new/*filepath",le="1000"} 9 prometheus_http_response_size_bytes_bucket{handler="/new/*filepath",le="10000"} 9 prometheus_http_response_size_bytes_bucket{handler="/new/*filepath",le="100000"} 9 prometheus_http_response_size_bytes_bucket{handler="/new/*filepath",le="1e+06"} 9 prometheus_http_response_size_bytes_bucket{handler="/new/*filepath",le="1e+07"} 9 prometheus_http_response_size_bytes_bucket{handler="/new/*filepath",le="1e+08"} 9 prometheus_http_response_size_bytes_bucket{handler="/new/*filepath",le="1e+09"} 9 prometheus_http_response_size_bytes_bucket{handler="/new/*filepath",le="+Inf"} 9 prometheus_http_response_size_bytes_sum{handler="/new/*filepath"} 171 prometheus_http_response_size_bytes_count{handler="/new/*filepath"} 9 prometheus_http_response_size_bytes_bucket{handler="/rules",le="100"} 0 prometheus_http_response_size_bytes_bucket{handler="/rules",le="1000"} 0 prometheus_http_response_size_bytes_bucket{handler="/rules",le="10000"} 0 prometheus_http_response_size_bytes_bucket{handler="/rules",le="100000"} 3 prometheus_http_response_size_bytes_bucket{handler="/rules",le="1e+06"} 3 prometheus_http_response_size_bytes_bucket{handler="/rules",le="1e+07"} 3 prometheus_http_response_size_bytes_bucket{handler="/rules",le="1e+08"} 3 prometheus_http_response_size_bytes_bucket{handler="/rules",le="1e+09"} 3 prometheus_http_response_size_bytes_bucket{handler="/rules",le="+Inf"} 3 prometheus_http_response_size_bytes_sum{handler="/rules"} 45629 prometheus_http_response_size_bytes_count{handler="/rules"} 3 prometheus_http_response_size_bytes_bucket{handler="/service-discovery",le="100"} 0 prometheus_http_response_size_bytes_bucket{handler="/service-discovery",le="1000"} 0 prometheus_http_response_size_bytes_bucket{handler="/service-discovery",le="10000"} 0 prometheus_http_response_size_bytes_bucket{handler="/service-discovery",le="100000"} 0 prometheus_http_response_size_bytes_bucket{handler="/service-discovery",le="1e+06"} 0 prometheus_http_response_size_bytes_bucket{handler="/service-discovery",le="1e+07"} 2 prometheus_http_response_size_bytes_bucket{handler="/service-discovery",le="1e+08"} 2 prometheus_http_response_size_bytes_bucket{handler="/service-discovery",le="1e+09"} 2 prometheus_http_response_size_bytes_bucket{handler="/service-discovery",le="+Inf"} 2 prometheus_http_response_size_bytes_sum{handler="/service-discovery"} 1.0801013e+07 prometheus_http_response_size_bytes_count{handler="/service-discovery"} 2 prometheus_http_response_size_bytes_bucket{handler="/static/*filepath",le="100"} 7 prometheus_http_response_size_bytes_bucket{handler="/static/*filepath",le="1000"} 12 prometheus_http_response_size_bytes_bucket{handler="/static/*filepath",le="10000"} 32 prometheus_http_response_size_bytes_bucket{handler="/static/*filepath",le="100000"} 111 prometheus_http_response_size_bytes_bucket{handler="/static/*filepath",le="1e+06"} 124 prometheus_http_response_size_bytes_bucket{handler="/static/*filepath",le="1e+07"} 124 prometheus_http_response_size_bytes_bucket{handler="/static/*filepath",le="1e+08"} 124 prometheus_http_response_size_bytes_bucket{handler="/static/*filepath",le="1e+09"} 124 prometheus_http_response_size_bytes_bucket{handler="/static/*filepath",le="+Inf"} 124 prometheus_http_response_size_bytes_sum{handler="/static/*filepath"} 5.125873e+06 prometheus_http_response_size_bytes_count{handler="/static/*filepath"} 124 prometheus_http_response_size_bytes_bucket{handler="/status",le="100"} 0 prometheus_http_response_size_bytes_bucket{handler="/status",le="1000"} 0 prometheus_http_response_size_bytes_bucket{handler="/status",le="10000"} 0 prometheus_http_response_size_bytes_bucket{handler="/status",le="100000"} 2 prometheus_http_response_size_bytes_bucket{handler="/status",le="1e+06"} 2 prometheus_http_response_size_bytes_bucket{handler="/status",le="1e+07"} 2 prometheus_http_response_size_bytes_bucket{handler="/status",le="1e+08"} 2 prometheus_http_response_size_bytes_bucket{handler="/status",le="1e+09"} 2 prometheus_http_response_size_bytes_bucket{handler="/status",le="+Inf"} 2 prometheus_http_response_size_bytes_sum{handler="/status"} 25274 prometheus_http_response_size_bytes_count{handler="/status"} 2 prometheus_http_response_size_bytes_bucket{handler="/targets",le="100"} 0 prometheus_http_response_size_bytes_bucket{handler="/targets",le="1000"} 0 prometheus_http_response_size_bytes_bucket{handler="/targets",le="10000"} 0 prometheus_http_response_size_bytes_bucket{handler="/targets",le="100000"} 0 prometheus_http_response_size_bytes_bucket{handler="/targets",le="1e+06"} 2 prometheus_http_response_size_bytes_bucket{handler="/targets",le="1e+07"} 2 prometheus_http_response_size_bytes_bucket{handler="/targets",le="1e+08"} 2 prometheus_http_response_size_bytes_bucket{handler="/targets",le="1e+09"} 2 prometheus_http_response_size_bytes_bucket{handler="/targets",le="+Inf"} 2 prometheus_http_response_size_bytes_sum{handler="/targets"} 1.131864e+06 prometheus_http_response_size_bytes_count{handler="/targets"} 2 prometheus_http_response_size_bytes_bucket{handler="/version",le="100"} 0 prometheus_http_response_size_bytes_bucket{handler="/version",le="1000"} 1 prometheus_http_response_size_bytes_bucket{handler="/version",le="10000"} 1 prometheus_http_response_size_bytes_bucket{handler="/version",le="100000"} 1 prometheus_http_response_size_bytes_bucket{handler="/version",le="1e+06"} 1 prometheus_http_response_size_bytes_bucket{handler="/version",le="1e+07"} 1 prometheus_http_response_size_bytes_bucket{handler="/version",le="1e+08"} 1 prometheus_http_response_size_bytes_bucket{handler="/version",le="1e+09"} 1 prometheus_http_response_size_bytes_bucket{handler="/version",le="+Inf"} 1 prometheus_http_response_size_bytes_sum{handler="/version"} 178 prometheus_http_response_size_bytes_count{handler="/version"} 1 # HELP prometheus_notifications_alertmanagers_discovered The number of alertmanagers discovered and active. # TYPE prometheus_notifications_alertmanagers_discovered gauge prometheus_notifications_alertmanagers_discovered 1 # HELP prometheus_notifications_dropped_total Total number of alerts dropped due to errors when sending to Alertmanager. # TYPE prometheus_notifications_dropped_total counter prometheus_notifications_dropped_total 0 # HELP prometheus_notifications_errors_total Total number of errors sending alert notifications. # TYPE prometheus_notifications_errors_total counter prometheus_notifications_errors_total{alertmanager="http://10.152.6.17:9093/api/v1/alerts"} 0 # HELP prometheus_notifications_latency_seconds Latency quantiles for sending alert notifications. # TYPE prometheus_notifications_latency_seconds summary prometheus_notifications_latency_seconds{alertmanager="http://10.152.6.17:9093/api/v1/alerts",quantile="0.5"} 0.001802263 prometheus_notifications_latency_seconds{alertmanager="http://10.152.6.17:9093/api/v1/alerts",quantile="0.9"} 0.001935143 prometheus_notifications_latency_seconds{alertmanager="http://10.152.6.17:9093/api/v1/alerts",quantile="0.99"} 0.002185055 prometheus_notifications_latency_seconds_sum{alertmanager="http://10.152.6.17:9093/api/v1/alerts"} 57.04237067 prometheus_notifications_latency_seconds_count{alertmanager="http://10.152.6.17:9093/api/v1/alerts"} 27720 # HELP prometheus_notifications_queue_capacity The capacity of the alert notifications queue. # TYPE prometheus_notifications_queue_capacity gauge prometheus_notifications_queue_capacity 10000 # HELP prometheus_notifications_queue_length The number of alert notifications in the queue. # TYPE prometheus_notifications_queue_length gauge prometheus_notifications_queue_length 0 # HELP prometheus_notifications_sent_total Total number of alerts sent. # TYPE prometheus_notifications_sent_total counter prometheus_notifications_sent_total{alertmanager="http://10.152.6.17:9093/api/v1/alerts"} 33362 # HELP prometheus_remote_storage_highest_timestamp_in_seconds Highest timestamp that has come into the remote storage via the Appender interface, in seconds since epoch. # TYPE prometheus_remote_storage_highest_timestamp_in_seconds gauge prometheus_remote_storage_highest_timestamp_in_seconds 1.774931704e+09 # HELP prometheus_remote_storage_samples_in_total Samples in to remote storage, compare to samples out for queue managers. # TYPE prometheus_remote_storage_samples_in_total counter prometheus_remote_storage_samples_in_total 3.754248692e+09 # HELP prometheus_remote_storage_string_interner_zero_reference_releases_total The number of times release has been called for strings that are not interned. # TYPE prometheus_remote_storage_string_interner_zero_reference_releases_total counter prometheus_remote_storage_string_interner_zero_reference_releases_total 0 # HELP prometheus_rule_evaluation_duration_seconds The duration for a rule to execute. # TYPE prometheus_rule_evaluation_duration_seconds summary prometheus_rule_evaluation_duration_seconds{quantile="0.5"} 0.000952213 prometheus_rule_evaluation_duration_seconds{quantile="0.9"} 0.010551126 prometheus_rule_evaluation_duration_seconds{quantile="0.99"} 0.014204735 prometheus_rule_evaluation_duration_seconds_sum 721.5253399390001 prometheus_rule_evaluation_duration_seconds_count 164472 # HELP prometheus_rule_evaluation_failures_total The total number of rule evaluation failures. # TYPE prometheus_rule_evaluation_failures_total counter prometheus_rule_evaluation_failures_total{rule_group="/etc/alerts.d/node_alerting_rules.yml;container_cpu_usage_is_high"} 0 prometheus_rule_evaluation_failures_total{rule_group="/etc/alerts.d/node_alerting_rules.yml;container_memory_usage_is_high"} 0 prometheus_rule_evaluation_failures_total{rule_group="/etc/alerts.d/node_alerting_rules.yml;node_cpu_greater_than_80"} 0 prometheus_rule_evaluation_failures_total{rule_group="/etc/alerts.d/node_alerting_rules.yml;node_disk_space_too_low"} 0 prometheus_rule_evaluation_failures_total{rule_group="/etc/alerts.d/node_alerting_rules.yml;node_down"} 0 prometheus_rule_evaluation_failures_total{rule_group="/etc/alerts.d/node_alerting_rules.yml;node_memory_left_lessser_than_10"} 0 prometheus_rule_evaluation_failures_total{rule_group="/etc/alerts.d/node_alerting_rules.yml;prometheus-job-down"} 0 # HELP prometheus_rule_evaluations_total The total number of rule evaluations. # TYPE prometheus_rule_evaluations_total counter prometheus_rule_evaluations_total{rule_group="/etc/alerts.d/node_alerting_rules.yml;container_cpu_usage_is_high"} 23496 prometheus_rule_evaluations_total{rule_group="/etc/alerts.d/node_alerting_rules.yml;container_memory_usage_is_high"} 23496 prometheus_rule_evaluations_total{rule_group="/etc/alerts.d/node_alerting_rules.yml;node_cpu_greater_than_80"} 23496 prometheus_rule_evaluations_total{rule_group="/etc/alerts.d/node_alerting_rules.yml;node_disk_space_too_low"} 23496 prometheus_rule_evaluations_total{rule_group="/etc/alerts.d/node_alerting_rules.yml;node_down"} 23496 prometheus_rule_evaluations_total{rule_group="/etc/alerts.d/node_alerting_rules.yml;node_memory_left_lessser_than_10"} 23496 prometheus_rule_evaluations_total{rule_group="/etc/alerts.d/node_alerting_rules.yml;prometheus-job-down"} 23496 # HELP prometheus_rule_group_duration_seconds The duration of rule group evaluations. # TYPE prometheus_rule_group_duration_seconds summary prometheus_rule_group_duration_seconds{quantile="0.01"} 0.000323405 prometheus_rule_group_duration_seconds{quantile="0.05"} 0.0003692 prometheus_rule_group_duration_seconds{quantile="0.5"} 0.000962047 prometheus_rule_group_duration_seconds{quantile="0.9"} 0.010560728 prometheus_rule_group_duration_seconds{quantile="0.99"} 0.014214859 prometheus_rule_group_duration_seconds_sum 723.7478937230019 prometheus_rule_group_duration_seconds_count 164472 # HELP prometheus_rule_group_interval_seconds The interval of a rule group. # TYPE prometheus_rule_group_interval_seconds gauge prometheus_rule_group_interval_seconds{rule_group="/etc/alerts.d/node_alerting_rules.yml;container_cpu_usage_is_high"} 60 prometheus_rule_group_interval_seconds{rule_group="/etc/alerts.d/node_alerting_rules.yml;container_memory_usage_is_high"} 60 prometheus_rule_group_interval_seconds{rule_group="/etc/alerts.d/node_alerting_rules.yml;node_cpu_greater_than_80"} 60 prometheus_rule_group_interval_seconds{rule_group="/etc/alerts.d/node_alerting_rules.yml;node_disk_space_too_low"} 60 prometheus_rule_group_interval_seconds{rule_group="/etc/alerts.d/node_alerting_rules.yml;node_down"} 60 prometheus_rule_group_interval_seconds{rule_group="/etc/alerts.d/node_alerting_rules.yml;node_memory_left_lessser_than_10"} 60 prometheus_rule_group_interval_seconds{rule_group="/etc/alerts.d/node_alerting_rules.yml;prometheus-job-down"} 60 # HELP prometheus_rule_group_iterations_missed_total The total number of rule group evaluations missed due to slow rule group evaluation. # TYPE prometheus_rule_group_iterations_missed_total counter prometheus_rule_group_iterations_missed_total{rule_group="/etc/alerts.d/node_alerting_rules.yml;container_cpu_usage_is_high"} 0 prometheus_rule_group_iterations_missed_total{rule_group="/etc/alerts.d/node_alerting_rules.yml;container_memory_usage_is_high"} 0 prometheus_rule_group_iterations_missed_total{rule_group="/etc/alerts.d/node_alerting_rules.yml;node_cpu_greater_than_80"} 0 prometheus_rule_group_iterations_missed_total{rule_group="/etc/alerts.d/node_alerting_rules.yml;node_disk_space_too_low"} 0 prometheus_rule_group_iterations_missed_total{rule_group="/etc/alerts.d/node_alerting_rules.yml;node_down"} 0 prometheus_rule_group_iterations_missed_total{rule_group="/etc/alerts.d/node_alerting_rules.yml;node_memory_left_lessser_than_10"} 0 prometheus_rule_group_iterations_missed_total{rule_group="/etc/alerts.d/node_alerting_rules.yml;prometheus-job-down"} 0 # HELP prometheus_rule_group_iterations_total The total number of scheduled rule group evaluations, whether executed or missed. # TYPE prometheus_rule_group_iterations_total counter prometheus_rule_group_iterations_total{rule_group="/etc/alerts.d/node_alerting_rules.yml;container_cpu_usage_is_high"} 23496 prometheus_rule_group_iterations_total{rule_group="/etc/alerts.d/node_alerting_rules.yml;container_memory_usage_is_high"} 23496 prometheus_rule_group_iterations_total{rule_group="/etc/alerts.d/node_alerting_rules.yml;node_cpu_greater_than_80"} 23496 prometheus_rule_group_iterations_total{rule_group="/etc/alerts.d/node_alerting_rules.yml;node_disk_space_too_low"} 23496 prometheus_rule_group_iterations_total{rule_group="/etc/alerts.d/node_alerting_rules.yml;node_down"} 23496 prometheus_rule_group_iterations_total{rule_group="/etc/alerts.d/node_alerting_rules.yml;node_memory_left_lessser_than_10"} 23496 prometheus_rule_group_iterations_total{rule_group="/etc/alerts.d/node_alerting_rules.yml;prometheus-job-down"} 23496 # HELP prometheus_rule_group_last_duration_seconds The duration of the last rule group evaluation. # TYPE prometheus_rule_group_last_duration_seconds gauge prometheus_rule_group_last_duration_seconds{rule_group="/etc/alerts.d/node_alerting_rules.yml;container_cpu_usage_is_high"} 0.004913595 prometheus_rule_group_last_duration_seconds{rule_group="/etc/alerts.d/node_alerting_rules.yml;container_memory_usage_is_high"} 0.010343808 prometheus_rule_group_last_duration_seconds{rule_group="/etc/alerts.d/node_alerting_rules.yml;node_cpu_greater_than_80"} 0.001386853 prometheus_rule_group_last_duration_seconds{rule_group="/etc/alerts.d/node_alerting_rules.yml;node_disk_space_too_low"} 0.000939159 prometheus_rule_group_last_duration_seconds{rule_group="/etc/alerts.d/node_alerting_rules.yml;node_down"} 0.000323405 prometheus_rule_group_last_duration_seconds{rule_group="/etc/alerts.d/node_alerting_rules.yml;node_memory_left_lessser_than_10"} 0.000833739 prometheus_rule_group_last_duration_seconds{rule_group="/etc/alerts.d/node_alerting_rules.yml;prometheus-job-down"} 0.000403493 # HELP prometheus_rule_group_last_evaluation_timestamp_seconds The timestamp of the last rule group evaluation in seconds. # TYPE prometheus_rule_group_last_evaluation_timestamp_seconds gauge prometheus_rule_group_last_evaluation_timestamp_seconds{rule_group="/etc/alerts.d/node_alerting_rules.yml;container_cpu_usage_is_high"} 1.7749316798764362e+09 prometheus_rule_group_last_evaluation_timestamp_seconds{rule_group="/etc/alerts.d/node_alerting_rules.yml;container_memory_usage_is_high"} 1.7749316693091018e+09 prometheus_rule_group_last_evaluation_timestamp_seconds{rule_group="/etc/alerts.d/node_alerting_rules.yml;node_cpu_greater_than_80"} 1.774931694618296e+09 prometheus_rule_group_last_evaluation_timestamp_seconds{rule_group="/etc/alerts.d/node_alerting_rules.yml;node_disk_space_too_low"} 1.7749316994970784e+09 prometheus_rule_group_last_evaluation_timestamp_seconds{rule_group="/etc/alerts.d/node_alerting_rules.yml;node_down"} 1.7749316850611274e+09 prometheus_rule_group_last_evaluation_timestamp_seconds{rule_group="/etc/alerts.d/node_alerting_rules.yml;node_memory_left_lessser_than_10"} 1.7749316743457522e+09 prometheus_rule_group_last_evaluation_timestamp_seconds{rule_group="/etc/alerts.d/node_alerting_rules.yml;prometheus-job-down"} 1.7749317030687864e+09 # HELP prometheus_rule_group_rules The number of rules. # TYPE prometheus_rule_group_rules gauge prometheus_rule_group_rules{rule_group="/etc/alerts.d/node_alerting_rules.yml;container_cpu_usage_is_high"} 1 prometheus_rule_group_rules{rule_group="/etc/alerts.d/node_alerting_rules.yml;container_memory_usage_is_high"} 1 prometheus_rule_group_rules{rule_group="/etc/alerts.d/node_alerting_rules.yml;node_cpu_greater_than_80"} 1 prometheus_rule_group_rules{rule_group="/etc/alerts.d/node_alerting_rules.yml;node_disk_space_too_low"} 1 prometheus_rule_group_rules{rule_group="/etc/alerts.d/node_alerting_rules.yml;node_down"} 1 prometheus_rule_group_rules{rule_group="/etc/alerts.d/node_alerting_rules.yml;node_memory_left_lessser_than_10"} 1 prometheus_rule_group_rules{rule_group="/etc/alerts.d/node_alerting_rules.yml;prometheus-job-down"} 1 # HELP prometheus_sd_consul_rpc_duration_seconds The duration of a Consul RPC call in seconds. # TYPE prometheus_sd_consul_rpc_duration_seconds summary prometheus_sd_consul_rpc_duration_seconds{call="service",endpoint="catalog",quantile="0.5"} NaN prometheus_sd_consul_rpc_duration_seconds{call="service",endpoint="catalog",quantile="0.9"} NaN prometheus_sd_consul_rpc_duration_seconds{call="service",endpoint="catalog",quantile="0.99"} NaN prometheus_sd_consul_rpc_duration_seconds_sum{call="service",endpoint="catalog"} 0 prometheus_sd_consul_rpc_duration_seconds_count{call="service",endpoint="catalog"} 0 prometheus_sd_consul_rpc_duration_seconds{call="services",endpoint="catalog",quantile="0.5"} NaN prometheus_sd_consul_rpc_duration_seconds{call="services",endpoint="catalog",quantile="0.9"} NaN prometheus_sd_consul_rpc_duration_seconds{call="services",endpoint="catalog",quantile="0.99"} NaN prometheus_sd_consul_rpc_duration_seconds_sum{call="services",endpoint="catalog"} 0 prometheus_sd_consul_rpc_duration_seconds_count{call="services",endpoint="catalog"} 0 # HELP prometheus_sd_consul_rpc_failures_total The number of Consul RPC call failures. # TYPE prometheus_sd_consul_rpc_failures_total counter prometheus_sd_consul_rpc_failures_total 0 # HELP prometheus_sd_discovered_targets Current number of discovered targets. # TYPE prometheus_sd_discovered_targets gauge prometheus_sd_discovered_targets{config="config-0",name="notify"} 268 prometheus_sd_discovered_targets{config="kubernetes-apiservers",name="scrape"} 154 prometheus_sd_discovered_targets{config="kubernetes-nodes",name="scrape"} 6 prometheus_sd_discovered_targets{config="kubernetes-nodes-cadvisor",name="scrape"} 6 prometheus_sd_discovered_targets{config="kubernetes-pods",name="scrape"} 268 prometheus_sd_discovered_targets{config="kubernetes-pods-slow",name="scrape"} 268 prometheus_sd_discovered_targets{config="kubernetes-service-endpoints",name="scrape"} 154 prometheus_sd_discovered_targets{config="kubernetes-service-endpoints-slow",name="scrape"} 154 prometheus_sd_discovered_targets{config="kubernetes-services",name="scrape"} 361 prometheus_sd_discovered_targets{config="opsmx_ssd_metrics",name="scrape"} 154 prometheus_sd_discovered_targets{config="otel-collector-redica-instance",name="scrape"} 1 prometheus_sd_discovered_targets{config="prometheus",name="scrape"} 1 prometheus_sd_discovered_targets{config="prometheus-pushgateway",name="scrape"} 361 # HELP prometheus_sd_dns_lookup_failures_total The number of DNS-SD lookup failures. # TYPE prometheus_sd_dns_lookup_failures_total counter prometheus_sd_dns_lookup_failures_total 0 # HELP prometheus_sd_dns_lookups_total The number of DNS-SD lookups. # TYPE prometheus_sd_dns_lookups_total counter prometheus_sd_dns_lookups_total 0 # HELP prometheus_sd_failed_configs Current number of service discovery configurations that failed to load. # TYPE prometheus_sd_failed_configs gauge prometheus_sd_failed_configs{name="notify"} 0 prometheus_sd_failed_configs{name="scrape"} 0 # HELP prometheus_sd_file_read_errors_total The number of File-SD read errors. # TYPE prometheus_sd_file_read_errors_total counter prometheus_sd_file_read_errors_total 0 # HELP prometheus_sd_file_scan_duration_seconds The duration of the File-SD scan in seconds. # TYPE prometheus_sd_file_scan_duration_seconds summary prometheus_sd_file_scan_duration_seconds{quantile="0.5"} NaN prometheus_sd_file_scan_duration_seconds{quantile="0.9"} NaN prometheus_sd_file_scan_duration_seconds{quantile="0.99"} NaN prometheus_sd_file_scan_duration_seconds_sum 0 prometheus_sd_file_scan_duration_seconds_count 0 # HELP prometheus_sd_kubernetes_events_total The number of Kubernetes events handled. # TYPE prometheus_sd_kubernetes_events_total counter prometheus_sd_kubernetes_events_total{event="add",role="endpoints"} 230 prometheus_sd_kubernetes_events_total{event="add",role="endpointslice"} 0 prometheus_sd_kubernetes_events_total{event="add",role="ingress"} 0 prometheus_sd_kubernetes_events_total{event="add",role="node"} 9 prometheus_sd_kubernetes_events_total{event="add",role="pod"} 356 prometheus_sd_kubernetes_events_total{event="add",role="service"} 458 prometheus_sd_kubernetes_events_total{event="delete",role="endpoints"} 4 prometheus_sd_kubernetes_events_total{event="delete",role="endpointslice"} 0 prometheus_sd_kubernetes_events_total{event="delete",role="ingress"} 0 prometheus_sd_kubernetes_events_total{event="delete",role="node"} 3 prometheus_sd_kubernetes_events_total{event="delete",role="pod"} 114 prometheus_sd_kubernetes_events_total{event="delete",role="service"} 8 prometheus_sd_kubernetes_events_total{event="update",role="endpoints"} 531238 prometheus_sd_kubernetes_events_total{event="update",role="endpointslice"} 0 prometheus_sd_kubernetes_events_total{event="update",role="ingress"} 0 prometheus_sd_kubernetes_events_total{event="update",role="node"} 69945 prometheus_sd_kubernetes_events_total{event="update",role="pod"} 569776 prometheus_sd_kubernetes_events_total{event="update",role="service"} 1.05728e+06 # HELP prometheus_sd_kubernetes_http_request_duration_seconds Summary of latencies for HTTP requests to the Kubernetes API by endpoint. # TYPE prometheus_sd_kubernetes_http_request_duration_seconds summary prometheus_sd_kubernetes_http_request_duration_seconds_sum{endpoint="/%7Bprefix%7D"} 0.647673251 prometheus_sd_kubernetes_http_request_duration_seconds_count{endpoint="/%7Bprefix%7D"} 16 # HELP prometheus_sd_kubernetes_http_request_total Total number of HTTP requests to the Kubernetes API by status code. # TYPE prometheus_sd_kubernetes_http_request_total counter prometheus_sd_kubernetes_http_request_total{status_code="200"} 21987 prometheus_sd_kubernetes_http_request_total{status_code="429"} 9 # HELP prometheus_sd_kubernetes_workqueue_depth Current depth of the work queue. # TYPE prometheus_sd_kubernetes_workqueue_depth gauge prometheus_sd_kubernetes_workqueue_depth{queue_name="endpoints"} 0 prometheus_sd_kubernetes_workqueue_depth{queue_name="node"} 0 prometheus_sd_kubernetes_workqueue_depth{queue_name="pod"} 0 prometheus_sd_kubernetes_workqueue_depth{queue_name="service"} 0 # HELP prometheus_sd_kubernetes_workqueue_items_total Total number of items added to the work queue. # TYPE prometheus_sd_kubernetes_workqueue_items_total counter prometheus_sd_kubernetes_workqueue_items_total{queue_name="endpoints"} 1.049752e+06 prometheus_sd_kubernetes_workqueue_items_total{queue_name="node"} 69957 prometheus_sd_kubernetes_workqueue_items_total{queue_name="pod"} 570246 prometheus_sd_kubernetes_workqueue_items_total{queue_name="service"} 528986 # HELP prometheus_sd_kubernetes_workqueue_latency_seconds How long an item stays in the work queue. # TYPE prometheus_sd_kubernetes_workqueue_latency_seconds summary prometheus_sd_kubernetes_workqueue_latency_seconds_sum{queue_name="endpoints"} 12281.947860455748 prometheus_sd_kubernetes_workqueue_latency_seconds_count{queue_name="endpoints"} 1.049752e+06 prometheus_sd_kubernetes_workqueue_latency_seconds_sum{queue_name="node"} 8.60792083699996 prometheus_sd_kubernetes_workqueue_latency_seconds_count{queue_name="node"} 69957 prometheus_sd_kubernetes_workqueue_latency_seconds_sum{queue_name="pod"} 1870.188588965 prometheus_sd_kubernetes_workqueue_latency_seconds_count{queue_name="pod"} 570246 prometheus_sd_kubernetes_workqueue_latency_seconds_sum{queue_name="service"} 3443.7126424050566 prometheus_sd_kubernetes_workqueue_latency_seconds_count{queue_name="service"} 528986 # HELP prometheus_sd_kubernetes_workqueue_longest_running_processor_seconds Duration of the longest running processor in the work queue. # TYPE prometheus_sd_kubernetes_workqueue_longest_running_processor_seconds gauge prometheus_sd_kubernetes_workqueue_longest_running_processor_seconds{queue_name="endpoints"} 0 prometheus_sd_kubernetes_workqueue_longest_running_processor_seconds{queue_name="node"} 0 prometheus_sd_kubernetes_workqueue_longest_running_processor_seconds{queue_name="pod"} 0 prometheus_sd_kubernetes_workqueue_longest_running_processor_seconds{queue_name="service"} 0 # HELP prometheus_sd_kubernetes_workqueue_unfinished_work_seconds How long an item has remained unfinished in the work queue. # TYPE prometheus_sd_kubernetes_workqueue_unfinished_work_seconds gauge prometheus_sd_kubernetes_workqueue_unfinished_work_seconds{queue_name="endpoints"} 0 prometheus_sd_kubernetes_workqueue_unfinished_work_seconds{queue_name="node"} 0 prometheus_sd_kubernetes_workqueue_unfinished_work_seconds{queue_name="pod"} 0 prometheus_sd_kubernetes_workqueue_unfinished_work_seconds{queue_name="service"} 0 # HELP prometheus_sd_kubernetes_workqueue_work_duration_seconds How long processing an item from the work queue takes. # TYPE prometheus_sd_kubernetes_workqueue_work_duration_seconds summary prometheus_sd_kubernetes_workqueue_work_duration_seconds_sum{queue_name="endpoints"} 111.67648773200212 prometheus_sd_kubernetes_workqueue_work_duration_seconds_count{queue_name="endpoints"} 1.049752e+06 prometheus_sd_kubernetes_workqueue_work_duration_seconds_sum{queue_name="node"} 16.45751810599998 prometheus_sd_kubernetes_workqueue_work_duration_seconds_count{queue_name="node"} 69957 prometheus_sd_kubernetes_workqueue_work_duration_seconds_sum{queue_name="pod"} 34.21304588000061 prometheus_sd_kubernetes_workqueue_work_duration_seconds_count{queue_name="pod"} 570246 prometheus_sd_kubernetes_workqueue_work_duration_seconds_sum{queue_name="service"} 32.81268237800029 prometheus_sd_kubernetes_workqueue_work_duration_seconds_count{queue_name="service"} 528986 # HELP prometheus_sd_received_updates_total Total number of update events received from the SD providers. # TYPE prometheus_sd_received_updates_total counter prometheus_sd_received_updates_total{name="notify"} 285123 prometheus_sd_received_updates_total{name="scrape"} 1.933822e+06 # HELP prometheus_sd_updates_total Total number of update events sent to the SD consumers. # TYPE prometheus_sd_updates_total counter prometheus_sd_updates_total{name="notify"} 2478 prometheus_sd_updates_total{name="scrape"} 56501 # HELP prometheus_target_interval_length_seconds Actual intervals between scrapes. # TYPE prometheus_target_interval_length_seconds summary prometheus_target_interval_length_seconds{interval="1m0s",quantile="0.01"} 59.999359898 prometheus_target_interval_length_seconds{interval="1m0s",quantile="0.05"} 59.999880835 prometheus_target_interval_length_seconds{interval="1m0s",quantile="0.5"} 60.000022185 prometheus_target_interval_length_seconds{interval="1m0s",quantile="0.9"} 60.000102041 prometheus_target_interval_length_seconds{interval="1m0s",quantile="0.99"} 60.00069793 prometheus_target_interval_length_seconds_sum{interval="1m0s"} 1.353289562644102e+08 prometheus_target_interval_length_seconds_count{interval="1m0s"} 2.255482e+06 prometheus_target_interval_length_seconds{interval="5s",quantile="0.01"} 4.9993119539999995 prometheus_target_interval_length_seconds{interval="5s",quantile="0.05"} 4.999829805 prometheus_target_interval_length_seconds{interval="5s",quantile="0.5"} 5.000005195 prometheus_target_interval_length_seconds{interval="5s",quantile="0.9"} 5.000118538 prometheus_target_interval_length_seconds{interval="5s",quantile="0.99"} 5.000925298 prometheus_target_interval_length_seconds_sum{interval="5s"} 1.4097543095969465e+06 prometheus_target_interval_length_seconds_count{interval="5s"} 281950 # HELP prometheus_target_metadata_cache_bytes The number of bytes that are currently used for storing metric metadata in the cache # TYPE prometheus_target_metadata_cache_bytes gauge prometheus_target_metadata_cache_bytes{scrape_job="kubernetes-apiservers"} 41475 prometheus_target_metadata_cache_bytes{scrape_job="kubernetes-nodes"} 219910 prometheus_target_metadata_cache_bytes{scrape_job="kubernetes-nodes-cadvisor"} 24840 prometheus_target_metadata_cache_bytes{scrape_job="kubernetes-pods"} 28906 prometheus_target_metadata_cache_bytes{scrape_job="kubernetes-pods-slow"} 0 prometheus_target_metadata_cache_bytes{scrape_job="kubernetes-service-endpoints"} 94765 prometheus_target_metadata_cache_bytes{scrape_job="kubernetes-service-endpoints-slow"} 0 prometheus_target_metadata_cache_bytes{scrape_job="kubernetes-services"} 0 prometheus_target_metadata_cache_bytes{scrape_job="opsmx_ssd_metrics"} 0 prometheus_target_metadata_cache_bytes{scrape_job="otel-collector-redica-instance"} 449 prometheus_target_metadata_cache_bytes{scrape_job="prometheus"} 9889 prometheus_target_metadata_cache_bytes{scrape_job="prometheus-pushgateway"} 0 # HELP prometheus_target_metadata_cache_entries Total number of metric metadata entries in the cache # TYPE prometheus_target_metadata_cache_entries gauge prometheus_target_metadata_cache_entries{scrape_job="kubernetes-apiservers"} 300 prometheus_target_metadata_cache_entries{scrape_job="kubernetes-nodes"} 1536 prometheus_target_metadata_cache_entries{scrape_job="kubernetes-nodes-cadvisor"} 468 prometheus_target_metadata_cache_entries{scrape_job="kubernetes-pods"} 614 prometheus_target_metadata_cache_entries{scrape_job="kubernetes-pods-slow"} 0 prometheus_target_metadata_cache_entries{scrape_job="kubernetes-service-endpoints"} 2131 prometheus_target_metadata_cache_entries{scrape_job="kubernetes-service-endpoints-slow"} 0 prometheus_target_metadata_cache_entries{scrape_job="kubernetes-services"} 0 prometheus_target_metadata_cache_entries{scrape_job="opsmx_ssd_metrics"} 0 prometheus_target_metadata_cache_entries{scrape_job="otel-collector-redica-instance"} 11 prometheus_target_metadata_cache_entries{scrape_job="prometheus"} 171 prometheus_target_metadata_cache_entries{scrape_job="prometheus-pushgateway"} 0 # HELP prometheus_target_scrape_pool_exceeded_target_limit_total Total number of times scrape pools hit the target limit, during sync or config reload. # TYPE prometheus_target_scrape_pool_exceeded_target_limit_total counter prometheus_target_scrape_pool_exceeded_target_limit_total 0 # HELP prometheus_target_scrape_pool_reloads_failed_total Total number of failed scrape pool reloads. # TYPE prometheus_target_scrape_pool_reloads_failed_total counter prometheus_target_scrape_pool_reloads_failed_total 0 # HELP prometheus_target_scrape_pool_reloads_total Total number of scrape pool reloads. # TYPE prometheus_target_scrape_pool_reloads_total counter prometheus_target_scrape_pool_reloads_total 0 # HELP prometheus_target_scrape_pool_sync_total Total number of syncs that were executed on a scrape pool. # TYPE prometheus_target_scrape_pool_sync_total counter prometheus_target_scrape_pool_sync_total{scrape_job="kubernetes-apiservers"} 56501 prometheus_target_scrape_pool_sync_total{scrape_job="kubernetes-nodes"} 56501 prometheus_target_scrape_pool_sync_total{scrape_job="kubernetes-nodes-cadvisor"} 56501 prometheus_target_scrape_pool_sync_total{scrape_job="kubernetes-pods"} 56501 prometheus_target_scrape_pool_sync_total{scrape_job="kubernetes-pods-slow"} 56501 prometheus_target_scrape_pool_sync_total{scrape_job="kubernetes-service-endpoints"} 56501 prometheus_target_scrape_pool_sync_total{scrape_job="kubernetes-service-endpoints-slow"} 56501 prometheus_target_scrape_pool_sync_total{scrape_job="kubernetes-services"} 56501 prometheus_target_scrape_pool_sync_total{scrape_job="opsmx_ssd_metrics"} 56501 prometheus_target_scrape_pool_sync_total{scrape_job="otel-collector-redica-instance"} 56501 prometheus_target_scrape_pool_sync_total{scrape_job="prometheus"} 56501 prometheus_target_scrape_pool_sync_total{scrape_job="prometheus-pushgateway"} 56501 # HELP prometheus_target_scrape_pool_targets Current number of targets in this scrape pool. # TYPE prometheus_target_scrape_pool_targets gauge prometheus_target_scrape_pool_targets{scrape_job="kubernetes-apiservers"} 1 prometheus_target_scrape_pool_targets{scrape_job="kubernetes-nodes"} 6 prometheus_target_scrape_pool_targets{scrape_job="kubernetes-nodes-cadvisor"} 6 prometheus_target_scrape_pool_targets{scrape_job="kubernetes-pods"} 9 prometheus_target_scrape_pool_targets{scrape_job="kubernetes-pods-slow"} 0 prometheus_target_scrape_pool_targets{scrape_job="kubernetes-service-endpoints"} 12 prometheus_target_scrape_pool_targets{scrape_job="kubernetes-service-endpoints-slow"} 0 prometheus_target_scrape_pool_targets{scrape_job="kubernetes-services"} 0 prometheus_target_scrape_pool_targets{scrape_job="opsmx_ssd_metrics"} 61 prometheus_target_scrape_pool_targets{scrape_job="otel-collector-redica-instance"} 1 prometheus_target_scrape_pool_targets{scrape_job="prometheus"} 1 prometheus_target_scrape_pool_targets{scrape_job="prometheus-pushgateway"} 0 # HELP prometheus_target_scrape_pools_failed_total Total number of scrape pool creations that failed. # TYPE prometheus_target_scrape_pools_failed_total counter prometheus_target_scrape_pools_failed_total 0 # HELP prometheus_target_scrape_pools_total Total number of scrape pool creation attempts. # TYPE prometheus_target_scrape_pools_total counter prometheus_target_scrape_pools_total 12 # HELP prometheus_target_scrapes_cache_flush_forced_total How many times a scrape cache was flushed due to getting big while scrapes are failing. # TYPE prometheus_target_scrapes_cache_flush_forced_total counter prometheus_target_scrapes_cache_flush_forced_total 0 # HELP prometheus_target_scrapes_exceeded_sample_limit_total Total number of scrapes that hit the sample limit and were rejected. # TYPE prometheus_target_scrapes_exceeded_sample_limit_total counter prometheus_target_scrapes_exceeded_sample_limit_total 0 # HELP prometheus_target_scrapes_sample_duplicate_timestamp_total Total number of samples rejected due to duplicate timestamps but different values # TYPE prometheus_target_scrapes_sample_duplicate_timestamp_total counter prometheus_target_scrapes_sample_duplicate_timestamp_total 0 # HELP prometheus_target_scrapes_sample_out_of_bounds_total Total number of samples rejected due to timestamp falling outside of the time bounds # TYPE prometheus_target_scrapes_sample_out_of_bounds_total counter prometheus_target_scrapes_sample_out_of_bounds_total 0 # HELP prometheus_target_scrapes_sample_out_of_order_total Total number of samples rejected due to not being out of the expected order # TYPE prometheus_target_scrapes_sample_out_of_order_total counter prometheus_target_scrapes_sample_out_of_order_total 0 # HELP prometheus_target_sync_length_seconds Actual interval to sync the scrape pool. # TYPE prometheus_target_sync_length_seconds summary prometheus_target_sync_length_seconds{scrape_job="kubernetes-apiservers",quantile="0.01"} 0.005564193 prometheus_target_sync_length_seconds{scrape_job="kubernetes-apiservers",quantile="0.05"} 0.005604183 prometheus_target_sync_length_seconds{scrape_job="kubernetes-apiservers",quantile="0.5"} 0.006423813 prometheus_target_sync_length_seconds{scrape_job="kubernetes-apiservers",quantile="0.9"} 0.012973602 prometheus_target_sync_length_seconds{scrape_job="kubernetes-apiservers",quantile="0.99"} 0.016781364 prometheus_target_sync_length_seconds_sum{scrape_job="kubernetes-apiservers"} 757.6579938550045 prometheus_target_sync_length_seconds_count{scrape_job="kubernetes-apiservers"} 56501 prometheus_target_sync_length_seconds{scrape_job="kubernetes-nodes",quantile="0.01"} 0.001543408 prometheus_target_sync_length_seconds{scrape_job="kubernetes-nodes",quantile="0.05"} 0.001608847 prometheus_target_sync_length_seconds{scrape_job="kubernetes-nodes",quantile="0.5"} 0.001865536 prometheus_target_sync_length_seconds{scrape_job="kubernetes-nodes",quantile="0.9"} 0.002084919 prometheus_target_sync_length_seconds{scrape_job="kubernetes-nodes",quantile="0.99"} 0.007485247 prometheus_target_sync_length_seconds_sum{scrape_job="kubernetes-nodes"} 185.41446130300181 prometheus_target_sync_length_seconds_count{scrape_job="kubernetes-nodes"} 56501 prometheus_target_sync_length_seconds{scrape_job="kubernetes-nodes-cadvisor",quantile="0.01"} 0.001227488 prometheus_target_sync_length_seconds{scrape_job="kubernetes-nodes-cadvisor",quantile="0.05"} 0.001730345 prometheus_target_sync_length_seconds{scrape_job="kubernetes-nodes-cadvisor",quantile="0.5"} 0.001839612 prometheus_target_sync_length_seconds{scrape_job="kubernetes-nodes-cadvisor",quantile="0.9"} 0.004248961 prometheus_target_sync_length_seconds{scrape_job="kubernetes-nodes-cadvisor",quantile="0.99"} 0.009761866 prometheus_target_sync_length_seconds_sum{scrape_job="kubernetes-nodes-cadvisor"} 189.98502800400036 prometheus_target_sync_length_seconds_count{scrape_job="kubernetes-nodes-cadvisor"} 56501 prometheus_target_sync_length_seconds{scrape_job="kubernetes-pods",quantile="0.01"} 0.006236608 prometheus_target_sync_length_seconds{scrape_job="kubernetes-pods",quantile="0.05"} 0.007403268 prometheus_target_sync_length_seconds{scrape_job="kubernetes-pods",quantile="0.5"} 0.008802878 prometheus_target_sync_length_seconds{scrape_job="kubernetes-pods",quantile="0.9"} 0.02580196 prometheus_target_sync_length_seconds{scrape_job="kubernetes-pods",quantile="0.99"} 0.029770206 prometheus_target_sync_length_seconds_sum{scrape_job="kubernetes-pods"} 980.4117281039933 prometheus_target_sync_length_seconds_count{scrape_job="kubernetes-pods"} 56501 prometheus_target_sync_length_seconds{scrape_job="kubernetes-pods-slow",quantile="0.01"} 0.004863928 prometheus_target_sync_length_seconds{scrape_job="kubernetes-pods-slow",quantile="0.05"} 0.004975457 prometheus_target_sync_length_seconds{scrape_job="kubernetes-pods-slow",quantile="0.5"} 0.00570023 prometheus_target_sync_length_seconds{scrape_job="kubernetes-pods-slow",quantile="0.9"} 0.010262827 prometheus_target_sync_length_seconds{scrape_job="kubernetes-pods-slow",quantile="0.99"} 0.012808386 prometheus_target_sync_length_seconds_sum{scrape_job="kubernetes-pods-slow"} 687.5105179350111 prometheus_target_sync_length_seconds_count{scrape_job="kubernetes-pods-slow"} 56501 prometheus_target_sync_length_seconds{scrape_job="kubernetes-service-endpoints",quantile="0.01"} 0.008761625 prometheus_target_sync_length_seconds{scrape_job="kubernetes-service-endpoints",quantile="0.05"} 0.009231869 prometheus_target_sync_length_seconds{scrape_job="kubernetes-service-endpoints",quantile="0.5"} 0.010164053 prometheus_target_sync_length_seconds{scrape_job="kubernetes-service-endpoints",quantile="0.9"} 0.019082188 prometheus_target_sync_length_seconds{scrape_job="kubernetes-service-endpoints",quantile="0.99"} 0.030975993 prometheus_target_sync_length_seconds_sum{scrape_job="kubernetes-service-endpoints"} 1131.9145774360031 prometheus_target_sync_length_seconds_count{scrape_job="kubernetes-service-endpoints"} 56501 prometheus_target_sync_length_seconds{scrape_job="kubernetes-service-endpoints-slow",quantile="0.01"} 0.005446472 prometheus_target_sync_length_seconds{scrape_job="kubernetes-service-endpoints-slow",quantile="0.05"} 0.005495203 prometheus_target_sync_length_seconds{scrape_job="kubernetes-service-endpoints-slow",quantile="0.5"} 0.006125801 prometheus_target_sync_length_seconds{scrape_job="kubernetes-service-endpoints-slow",quantile="0.9"} 0.012811284 prometheus_target_sync_length_seconds{scrape_job="kubernetes-service-endpoints-slow",quantile="0.99"} 0.018345913 prometheus_target_sync_length_seconds_sum{scrape_job="kubernetes-service-endpoints-slow"} 729.6942464550028 prometheus_target_sync_length_seconds_count{scrape_job="kubernetes-service-endpoints-slow"} 56501 prometheus_target_sync_length_seconds{scrape_job="kubernetes-services",quantile="0.01"} 0.006473808 prometheus_target_sync_length_seconds{scrape_job="kubernetes-services",quantile="0.05"} 0.006780424 prometheus_target_sync_length_seconds{scrape_job="kubernetes-services",quantile="0.5"} 0.008042994 prometheus_target_sync_length_seconds{scrape_job="kubernetes-services",quantile="0.9"} 0.021169813 prometheus_target_sync_length_seconds{scrape_job="kubernetes-services",quantile="0.99"} 0.032384743 prometheus_target_sync_length_seconds_sum{scrape_job="kubernetes-services"} 898.1425661220018 prometheus_target_sync_length_seconds_count{scrape_job="kubernetes-services"} 56501 prometheus_target_sync_length_seconds{scrape_job="opsmx_ssd_metrics",quantile="0.01"} 0.008619879 prometheus_target_sync_length_seconds{scrape_job="opsmx_ssd_metrics",quantile="0.05"} 0.008750187 prometheus_target_sync_length_seconds{scrape_job="opsmx_ssd_metrics",quantile="0.5"} 0.010309016 prometheus_target_sync_length_seconds{scrape_job="opsmx_ssd_metrics",quantile="0.9"} 0.026916425 prometheus_target_sync_length_seconds{scrape_job="opsmx_ssd_metrics",quantile="0.99"} 0.030512687 prometheus_target_sync_length_seconds_sum{scrape_job="opsmx_ssd_metrics"} 1128.1923891589925 prometheus_target_sync_length_seconds_count{scrape_job="opsmx_ssd_metrics"} 56501 prometheus_target_sync_length_seconds{scrape_job="otel-collector-redica-instance",quantile="0.01"} 1.862e-05 prometheus_target_sync_length_seconds{scrape_job="otel-collector-redica-instance",quantile="0.05"} 2.4782e-05 prometheus_target_sync_length_seconds{scrape_job="otel-collector-redica-instance",quantile="0.5"} 5.5224e-05 prometheus_target_sync_length_seconds{scrape_job="otel-collector-redica-instance",quantile="0.9"} 8.4615e-05 prometheus_target_sync_length_seconds{scrape_job="otel-collector-redica-instance",quantile="0.99"} 0.000172218 prometheus_target_sync_length_seconds_sum{scrape_job="otel-collector-redica-instance"} 5.776238875000021 prometheus_target_sync_length_seconds_count{scrape_job="otel-collector-redica-instance"} 56501 prometheus_target_sync_length_seconds{scrape_job="prometheus",quantile="0.01"} 1.922e-05 prometheus_target_sync_length_seconds{scrape_job="prometheus",quantile="0.05"} 2.5615e-05 prometheus_target_sync_length_seconds{scrape_job="prometheus",quantile="0.5"} 6.5073e-05 prometheus_target_sync_length_seconds{scrape_job="prometheus",quantile="0.9"} 8.4802e-05 prometheus_target_sync_length_seconds{scrape_job="prometheus",quantile="0.99"} 0.000111287 prometheus_target_sync_length_seconds_sum{scrape_job="prometheus"} 5.054500084999991 prometheus_target_sync_length_seconds_count{scrape_job="prometheus"} 56501 prometheus_target_sync_length_seconds{scrape_job="prometheus-pushgateway",quantile="0.01"} 0.006626786 prometheus_target_sync_length_seconds{scrape_job="prometheus-pushgateway",quantile="0.05"} 0.006707513 prometheus_target_sync_length_seconds{scrape_job="prometheus-pushgateway",quantile="0.5"} 0.007377523 prometheus_target_sync_length_seconds{scrape_job="prometheus-pushgateway",quantile="0.9"} 0.021976054 prometheus_target_sync_length_seconds{scrape_job="prometheus-pushgateway",quantile="0.99"} 0.027580933 prometheus_target_sync_length_seconds_sum{scrape_job="prometheus-pushgateway"} 881.2676923289946 prometheus_target_sync_length_seconds_count{scrape_job="prometheus-pushgateway"} 56501 # HELP prometheus_template_text_expansion_failures_total The total number of template text expansion failures. # TYPE prometheus_template_text_expansion_failures_total counter prometheus_template_text_expansion_failures_total 0 # HELP prometheus_template_text_expansions_total The total number of template text expansions. # TYPE prometheus_template_text_expansions_total counter prometheus_template_text_expansions_total 123231 # HELP prometheus_treecache_watcher_goroutines The current number of watcher goroutines. # TYPE prometheus_treecache_watcher_goroutines gauge prometheus_treecache_watcher_goroutines 0 # HELP prometheus_treecache_zookeeper_failures_total The total number of ZooKeeper failures. # TYPE prometheus_treecache_zookeeper_failures_total counter prometheus_treecache_zookeeper_failures_total 0 # HELP prometheus_tsdb_blocks_loaded Number of currently loaded data blocks # TYPE prometheus_tsdb_blocks_loaded gauge prometheus_tsdb_blocks_loaded 23 # HELP prometheus_tsdb_checkpoint_creations_failed_total Total number of checkpoint creations that failed. # TYPE prometheus_tsdb_checkpoint_creations_failed_total counter prometheus_tsdb_checkpoint_creations_failed_total 0 # HELP prometheus_tsdb_checkpoint_creations_total Total number of checkpoint creations attempted. # TYPE prometheus_tsdb_checkpoint_creations_total counter prometheus_tsdb_checkpoint_creations_total 99 # HELP prometheus_tsdb_checkpoint_deletions_failed_total Total number of checkpoint deletions that failed. # TYPE prometheus_tsdb_checkpoint_deletions_failed_total counter prometheus_tsdb_checkpoint_deletions_failed_total 0 # HELP prometheus_tsdb_checkpoint_deletions_total Total number of checkpoint deletions attempted. # TYPE prometheus_tsdb_checkpoint_deletions_total counter prometheus_tsdb_checkpoint_deletions_total 99 # HELP prometheus_tsdb_compaction_chunk_range_seconds Final time range of chunks on their first compaction # TYPE prometheus_tsdb_compaction_chunk_range_seconds histogram prometheus_tsdb_compaction_chunk_range_seconds_bucket{le="100"} 6284 prometheus_tsdb_compaction_chunk_range_seconds_bucket{le="400"} 6284 prometheus_tsdb_compaction_chunk_range_seconds_bucket{le="1600"} 6284 prometheus_tsdb_compaction_chunk_range_seconds_bucket{le="6400"} 6284 prometheus_tsdb_compaction_chunk_range_seconds_bucket{le="25600"} 6286 prometheus_tsdb_compaction_chunk_range_seconds_bucket{le="102400"} 16112 prometheus_tsdb_compaction_chunk_range_seconds_bucket{le="409600"} 25088 prometheus_tsdb_compaction_chunk_range_seconds_bucket{le="1.6384e+06"} 823777 prometheus_tsdb_compaction_chunk_range_seconds_bucket{le="6.5536e+06"} 1.061914e+06 prometheus_tsdb_compaction_chunk_range_seconds_bucket{le="2.62144e+07"} 3.1503906e+07 prometheus_tsdb_compaction_chunk_range_seconds_bucket{le="+Inf"} 3.1503906e+07 prometheus_tsdb_compaction_chunk_range_seconds_sum 2.18663833257163e+14 prometheus_tsdb_compaction_chunk_range_seconds_count 3.1503906e+07 # HELP prometheus_tsdb_compaction_chunk_samples Final number of samples on their first compaction # TYPE prometheus_tsdb_compaction_chunk_samples histogram prometheus_tsdb_compaction_chunk_samples_bucket{le="4"} 19248 prometheus_tsdb_compaction_chunk_samples_bucket{le="6"} 22542 prometheus_tsdb_compaction_chunk_samples_bucket{le="9"} 25993 prometheus_tsdb_compaction_chunk_samples_bucket{le="13.5"} 32851 prometheus_tsdb_compaction_chunk_samples_bucket{le="20.25"} 39516 prometheus_tsdb_compaction_chunk_samples_bucket{le="30.375"} 48479 prometheus_tsdb_compaction_chunk_samples_bucket{le="45.5625"} 49058 prometheus_tsdb_compaction_chunk_samples_bucket{le="68.34375"} 242578 prometheus_tsdb_compaction_chunk_samples_bucket{le="102.515625"} 283702 prometheus_tsdb_compaction_chunk_samples_bucket{le="153.7734375"} 3.1503474e+07 prometheus_tsdb_compaction_chunk_samples_bucket{le="230.66015625"} 3.1503565e+07 prometheus_tsdb_compaction_chunk_samples_bucket{le="345.990234375"} 3.1503565e+07 prometheus_tsdb_compaction_chunk_samples_bucket{le="+Inf"} 3.1503906e+07 prometheus_tsdb_compaction_chunk_samples_sum 3.760533423e+09 prometheus_tsdb_compaction_chunk_samples_count 3.1503906e+07 # HELP prometheus_tsdb_compaction_chunk_size_bytes Final size of chunks on their first compaction # TYPE prometheus_tsdb_compaction_chunk_size_bytes histogram prometheus_tsdb_compaction_chunk_size_bytes_bucket{le="32"} 72566 prometheus_tsdb_compaction_chunk_size_bytes_bucket{le="48"} 146855 prometheus_tsdb_compaction_chunk_size_bytes_bucket{le="72"} 1.0421454e+07 prometheus_tsdb_compaction_chunk_size_bytes_bucket{le="108"} 1.7955166e+07 prometheus_tsdb_compaction_chunk_size_bytes_bucket{le="162"} 1.9491951e+07 prometheus_tsdb_compaction_chunk_size_bytes_bucket{le="243"} 2.2956607e+07 prometheus_tsdb_compaction_chunk_size_bytes_bucket{le="364.5"} 2.9554173e+07 prometheus_tsdb_compaction_chunk_size_bytes_bucket{le="546.75"} 3.0350517e+07 prometheus_tsdb_compaction_chunk_size_bytes_bucket{le="820.125"} 3.1013698e+07 prometheus_tsdb_compaction_chunk_size_bytes_bucket{le="1230.1875"} 3.1503687e+07 prometheus_tsdb_compaction_chunk_size_bytes_bucket{le="1845.28125"} 3.1503781e+07 prometheus_tsdb_compaction_chunk_size_bytes_bucket{le="2767.921875"} 3.1503906e+07 prometheus_tsdb_compaction_chunk_size_bytes_bucket{le="+Inf"} 3.1503906e+07 prometheus_tsdb_compaction_chunk_size_bytes_sum 5.437395253e+09 prometheus_tsdb_compaction_chunk_size_bytes_count 3.1503906e+07 # HELP prometheus_tsdb_compaction_duration_seconds Duration of compaction runs # TYPE prometheus_tsdb_compaction_duration_seconds histogram prometheus_tsdb_compaction_duration_seconds_bucket{le="1"} 0 prometheus_tsdb_compaction_duration_seconds_bucket{le="2"} 0 prometheus_tsdb_compaction_duration_seconds_bucket{le="4"} 0 prometheus_tsdb_compaction_duration_seconds_bucket{le="8"} 0 prometheus_tsdb_compaction_duration_seconds_bucket{le="16"} 199 prometheus_tsdb_compaction_duration_seconds_bucket{le="32"} 283 prometheus_tsdb_compaction_duration_seconds_bucket{le="64"} 284 prometheus_tsdb_compaction_duration_seconds_bucket{le="128"} 284 prometheus_tsdb_compaction_duration_seconds_bucket{le="256"} 284 prometheus_tsdb_compaction_duration_seconds_bucket{le="512"} 284 prometheus_tsdb_compaction_duration_seconds_bucket{le="+Inf"} 284 prometheus_tsdb_compaction_duration_seconds_sum 4002.746064759001 prometheus_tsdb_compaction_duration_seconds_count 284 # HELP prometheus_tsdb_compaction_populating_block Set to 1 when a block is currently being written to the disk. # TYPE prometheus_tsdb_compaction_populating_block gauge prometheus_tsdb_compaction_populating_block 0 # HELP prometheus_tsdb_compactions_failed_total Total number of compactions that failed for the partition. # TYPE prometheus_tsdb_compactions_failed_total counter prometheus_tsdb_compactions_failed_total 0 # HELP prometheus_tsdb_compactions_skipped_total Total number of skipped compactions due to disabled auto compaction. # TYPE prometheus_tsdb_compactions_skipped_total counter prometheus_tsdb_compactions_skipped_total 0 # HELP prometheus_tsdb_compactions_total Total number of compactions that were executed for the partition. # TYPE prometheus_tsdb_compactions_total counter prometheus_tsdb_compactions_total 284 # HELP prometheus_tsdb_compactions_triggered_total Total number of triggered compactions for the partition. # TYPE prometheus_tsdb_compactions_triggered_total counter prometheus_tsdb_compactions_triggered_total 23689 # HELP prometheus_tsdb_head_active_appenders Number of currently active appender transactions # TYPE prometheus_tsdb_head_active_appenders gauge prometheus_tsdb_head_active_appenders 0 # HELP prometheus_tsdb_head_chunks Total number of chunks in the head block. # TYPE prometheus_tsdb_head_chunks gauge prometheus_tsdb_head_chunks 326996 # HELP prometheus_tsdb_head_chunks_created_total Total number of chunks created in the head # TYPE prometheus_tsdb_head_chunks_created_total counter prometheus_tsdb_head_chunks_created_total 3.1830902e+07 # HELP prometheus_tsdb_head_chunks_removed_total Total number of chunks removed in the head # TYPE prometheus_tsdb_head_chunks_removed_total counter prometheus_tsdb_head_chunks_removed_total 3.1503906e+07 # HELP prometheus_tsdb_head_gc_duration_seconds Runtime of garbage collection in the head block. # TYPE prometheus_tsdb_head_gc_duration_seconds summary prometheus_tsdb_head_gc_duration_seconds_sum 53.07835568200003 prometheus_tsdb_head_gc_duration_seconds_count 196 # HELP prometheus_tsdb_head_max_time Maximum timestamp of the head block. The unit is decided by the library consumer. # TYPE prometheus_tsdb_head_max_time gauge prometheus_tsdb_head_max_time 1.774931704354e+12 # HELP prometheus_tsdb_head_max_time_seconds Maximum timestamp of the head block. # TYPE prometheus_tsdb_head_max_time_seconds gauge prometheus_tsdb_head_max_time_seconds 1.774931704e+09 # HELP prometheus_tsdb_head_min_time Minimum time bound of the head block. The unit is decided by the library consumer. # TYPE prometheus_tsdb_head_min_time gauge prometheus_tsdb_head_min_time 1.7749224e+12 # HELP prometheus_tsdb_head_min_time_seconds Minimum time bound of the head block. # TYPE prometheus_tsdb_head_min_time_seconds gauge prometheus_tsdb_head_min_time_seconds 1.7749224e+09 # HELP prometheus_tsdb_head_samples_appended_total Total number of appended samples. # TYPE prometheus_tsdb_head_samples_appended_total counter prometheus_tsdb_head_samples_appended_total 3.754248692e+09 # HELP prometheus_tsdb_head_series Total number of series in the head block. # TYPE prometheus_tsdb_head_series gauge prometheus_tsdb_head_series 161210 # HELP prometheus_tsdb_head_series_created_total Total number of series created in the head # TYPE prometheus_tsdb_head_series_created_total counter prometheus_tsdb_head_series_created_total 358665 # HELP prometheus_tsdb_head_series_not_found_total Total number of requests for series that were not found. # TYPE prometheus_tsdb_head_series_not_found_total counter prometheus_tsdb_head_series_not_found_total 0 # HELP prometheus_tsdb_head_series_removed_total Total number of series removed in the head # TYPE prometheus_tsdb_head_series_removed_total counter prometheus_tsdb_head_series_removed_total 197455 # HELP prometheus_tsdb_head_truncations_failed_total Total number of head truncations that failed. # TYPE prometheus_tsdb_head_truncations_failed_total counter prometheus_tsdb_head_truncations_failed_total 0 # HELP prometheus_tsdb_head_truncations_total Total number of head truncations attempted. # TYPE prometheus_tsdb_head_truncations_total counter prometheus_tsdb_head_truncations_total 196 # HELP prometheus_tsdb_isolation_high_watermark The highest TSDB append ID that has been given out. # TYPE prometheus_tsdb_isolation_high_watermark gauge prometheus_tsdb_isolation_high_watermark 2.79518e+06 # HELP prometheus_tsdb_isolation_low_watermark The lowest TSDB append ID that is still referenced. # TYPE prometheus_tsdb_isolation_low_watermark gauge prometheus_tsdb_isolation_low_watermark 2.79518e+06 # HELP prometheus_tsdb_lowest_timestamp Lowest timestamp value stored in the database. The unit is decided by the library consumer. # TYPE prometheus_tsdb_lowest_timestamp gauge prometheus_tsdb_lowest_timestamp 1.773576e+12 # HELP prometheus_tsdb_lowest_timestamp_seconds Lowest timestamp value stored in the database. # TYPE prometheus_tsdb_lowest_timestamp_seconds gauge prometheus_tsdb_lowest_timestamp_seconds 1.773576e+09 # HELP prometheus_tsdb_mmap_chunk_corruptions_total Total number of memory-mapped chunk corruptions. # TYPE prometheus_tsdb_mmap_chunk_corruptions_total counter prometheus_tsdb_mmap_chunk_corruptions_total 0 # HELP prometheus_tsdb_out_of_bound_samples_total Total number of out of bound samples ingestion failed attempts. # TYPE prometheus_tsdb_out_of_bound_samples_total counter prometheus_tsdb_out_of_bound_samples_total 0 # HELP prometheus_tsdb_out_of_order_samples_total Total number of out of order samples ingestion failed attempts. # TYPE prometheus_tsdb_out_of_order_samples_total counter prometheus_tsdb_out_of_order_samples_total 33702 # HELP prometheus_tsdb_reloads_failures_total Number of times the database failed to reload block data from disk. # TYPE prometheus_tsdb_reloads_failures_total counter prometheus_tsdb_reloads_failures_total 0 # HELP prometheus_tsdb_reloads_total Number of times the database reloaded block data from disk. # TYPE prometheus_tsdb_reloads_total counter prometheus_tsdb_reloads_total 285 # HELP prometheus_tsdb_retention_limit_bytes Max number of bytes to be retained in the tsdb blocks, configured 0 means disabled # TYPE prometheus_tsdb_retention_limit_bytes gauge prometheus_tsdb_retention_limit_bytes 0 # HELP prometheus_tsdb_size_retentions_total The number of times that blocks were deleted because the maximum number of bytes was exceeded. # TYPE prometheus_tsdb_size_retentions_total counter prometheus_tsdb_size_retentions_total 0 # HELP prometheus_tsdb_storage_blocks_bytes The number of bytes that are currently used for local storage by all blocks. # TYPE prometheus_tsdb_storage_blocks_bytes gauge prometheus_tsdb_storage_blocks_bytes 6.345077953e+09 # HELP prometheus_tsdb_symbol_table_size_bytes Size of symbol table on disk (in bytes) # TYPE prometheus_tsdb_symbol_table_size_bytes gauge prometheus_tsdb_symbol_table_size_bytes 31712 # HELP prometheus_tsdb_time_retentions_total The number of times that blocks were deleted because the maximum time limit was exceeded. # TYPE prometheus_tsdb_time_retentions_total counter prometheus_tsdb_time_retentions_total 22 # HELP prometheus_tsdb_tombstone_cleanup_seconds The time taken to recompact blocks to remove tombstones. # TYPE prometheus_tsdb_tombstone_cleanup_seconds histogram prometheus_tsdb_tombstone_cleanup_seconds_bucket{le="0.005"} 0 prometheus_tsdb_tombstone_cleanup_seconds_bucket{le="0.01"} 0 prometheus_tsdb_tombstone_cleanup_seconds_bucket{le="0.025"} 0 prometheus_tsdb_tombstone_cleanup_seconds_bucket{le="0.05"} 0 prometheus_tsdb_tombstone_cleanup_seconds_bucket{le="0.1"} 0 prometheus_tsdb_tombstone_cleanup_seconds_bucket{le="0.25"} 0 prometheus_tsdb_tombstone_cleanup_seconds_bucket{le="0.5"} 0 prometheus_tsdb_tombstone_cleanup_seconds_bucket{le="1"} 0 prometheus_tsdb_tombstone_cleanup_seconds_bucket{le="2.5"} 0 prometheus_tsdb_tombstone_cleanup_seconds_bucket{le="5"} 0 prometheus_tsdb_tombstone_cleanup_seconds_bucket{le="10"} 0 prometheus_tsdb_tombstone_cleanup_seconds_bucket{le="+Inf"} 0 prometheus_tsdb_tombstone_cleanup_seconds_sum 0 prometheus_tsdb_tombstone_cleanup_seconds_count 0 # HELP prometheus_tsdb_vertical_compactions_total Total number of compactions done on overlapping blocks. # TYPE prometheus_tsdb_vertical_compactions_total counter prometheus_tsdb_vertical_compactions_total 0 # HELP prometheus_tsdb_wal_completed_pages_total Total number of completed pages. # TYPE prometheus_tsdb_wal_completed_pages_total counter prometheus_tsdb_wal_completed_pages_total 636805 # HELP prometheus_tsdb_wal_corruptions_total Total number of WAL corruptions. # TYPE prometheus_tsdb_wal_corruptions_total counter prometheus_tsdb_wal_corruptions_total 0 # HELP prometheus_tsdb_wal_fsync_duration_seconds Duration of WAL fsync. # TYPE prometheus_tsdb_wal_fsync_duration_seconds summary prometheus_tsdb_wal_fsync_duration_seconds{quantile="0.5"} NaN prometheus_tsdb_wal_fsync_duration_seconds{quantile="0.9"} NaN prometheus_tsdb_wal_fsync_duration_seconds{quantile="0.99"} NaN prometheus_tsdb_wal_fsync_duration_seconds_sum 1.4105887620000004 prometheus_tsdb_wal_fsync_duration_seconds_count 196 # HELP prometheus_tsdb_wal_page_flushes_total Total number of page flushes. # TYPE prometheus_tsdb_wal_page_flushes_total counter prometheus_tsdb_wal_page_flushes_total 3.20254e+06 # HELP prometheus_tsdb_wal_segment_current WAL segment index that TSDB is currently writing to. # TYPE prometheus_tsdb_wal_segment_current gauge prometheus_tsdb_wal_segment_current 7018 # HELP prometheus_tsdb_wal_truncate_duration_seconds Duration of WAL truncation. # TYPE prometheus_tsdb_wal_truncate_duration_seconds summary prometheus_tsdb_wal_truncate_duration_seconds_sum 446.05968546900004 prometheus_tsdb_wal_truncate_duration_seconds_count 99 # HELP prometheus_tsdb_wal_truncations_failed_total Total number of WAL truncations that failed. # TYPE prometheus_tsdb_wal_truncations_failed_total counter prometheus_tsdb_wal_truncations_failed_total 0 # HELP prometheus_tsdb_wal_truncations_total Total number of WAL truncations attempted. # TYPE prometheus_tsdb_wal_truncations_total counter prometheus_tsdb_wal_truncations_total 99 # HELP prometheus_tsdb_wal_writes_failed_total Total number of WAL writes that failed. # TYPE prometheus_tsdb_wal_writes_failed_total counter prometheus_tsdb_wal_writes_failed_total 0 # HELP prometheus_web_federation_errors_total Total number of errors that occurred while sending federation responses. # TYPE prometheus_web_federation_errors_total counter prometheus_web_federation_errors_total 0 # HELP prometheus_web_federation_warnings_total Total number of warnings that occurred while sending federation responses. # TYPE prometheus_web_federation_warnings_total counter prometheus_web_federation_warnings_total 0 # HELP promhttp_metric_handler_requests_in_flight Current number of scrapes being served. # TYPE promhttp_metric_handler_requests_in_flight gauge promhttp_metric_handler_requests_in_flight 1 # HELP promhttp_metric_handler_requests_total Total number of scrapes by HTTP status code. # TYPE promhttp_metric_handler_requests_total counter promhttp_metric_handler_requests_total{code="200"} 23497 promhttp_metric_handler_requests_total{code="500"} 0 promhttp_metric_handler_requests_total{code="503"} 0