# HELP go_gc_duration_seconds A summary of the wall-time pause (stop-the-world) duration in garbage collection cycles. # TYPE go_gc_duration_seconds summary go_gc_duration_seconds{quantile="0"} 1.2854e-05 go_gc_duration_seconds{quantile="0.25"} 2.8844e-05 go_gc_duration_seconds{quantile="0.5"} 3.7379e-05 go_gc_duration_seconds{quantile="0.75"} 4.7348e-05 go_gc_duration_seconds{quantile="1"} 0.004108949 go_gc_duration_seconds_sum 0.124697936 go_gc_duration_seconds_count 1294 # HELP go_gc_gogc_percent Heap size target percentage configured by the user, otherwise 100. This value is set by the GOGC environment variable, and the runtime/debug.SetGCPercent function. Sourced from /gc/gogc:percent. # TYPE go_gc_gogc_percent gauge go_gc_gogc_percent 100 # HELP go_gc_gomemlimit_bytes Go runtime memory limit configured by the user, otherwise math.MaxInt64. This value is set by the GOMEMLIMIT environment variable, and the runtime/debug.SetMemoryLimit function. Sourced from /gc/gomemlimit:bytes. # TYPE go_gc_gomemlimit_bytes gauge go_gc_gomemlimit_bytes 9.223372036854776e+18 # HELP go_goroutines Number of goroutines that currently exist. # TYPE go_goroutines gauge go_goroutines 8 # HELP go_info Information about the Go environment. # TYPE go_info gauge go_info{version="go1.26.1"} 1 # HELP go_memstats_alloc_bytes Number of bytes allocated in heap and currently in use. Equals to /memory/classes/heap/objects:bytes. # TYPE go_memstats_alloc_bytes gauge go_memstats_alloc_bytes 1.627968e+06 # HELP go_memstats_alloc_bytes_total Total number of bytes allocated in heap until now, even if released already. Equals to /gc/heap/allocs:bytes. # TYPE go_memstats_alloc_bytes_total counter go_memstats_alloc_bytes_total 2.279970392e+09 # HELP go_memstats_buck_hash_sys_bytes Number of bytes used by the profiling bucket hash table. Equals to /memory/classes/profiling/buckets:bytes. # TYPE go_memstats_buck_hash_sys_bytes gauge go_memstats_buck_hash_sys_bytes 1.664922e+06 # HELP go_memstats_frees_total Total number of heap objects frees. Equals to /gc/heap/frees:objects + /gc/heap/tiny/allocs:objects. # TYPE go_memstats_frees_total counter go_memstats_frees_total 2.0796871e+07 # HELP go_memstats_gc_sys_bytes Number of bytes used for garbage collection system metadata. Equals to /memory/classes/metadata/other:bytes. # TYPE go_memstats_gc_sys_bytes gauge go_memstats_gc_sys_bytes 3.58168e+06 # HELP go_memstats_heap_alloc_bytes Number of heap bytes allocated and currently in use, same as go_memstats_alloc_bytes. Equals to /memory/classes/heap/objects:bytes. # TYPE go_memstats_heap_alloc_bytes gauge go_memstats_heap_alloc_bytes 1.627968e+06 # HELP go_memstats_heap_idle_bytes Number of heap bytes waiting to be used. Equals to /memory/classes/heap/released:bytes + /memory/classes/heap/free:bytes. # TYPE go_memstats_heap_idle_bytes gauge go_memstats_heap_idle_bytes 8.76544e+06 # HELP go_memstats_heap_inuse_bytes Number of heap bytes that are in use. Equals to /memory/classes/heap/objects:bytes + /memory/classes/heap/unused:bytes # TYPE go_memstats_heap_inuse_bytes gauge go_memstats_heap_inuse_bytes 3.227648e+06 # HELP go_memstats_heap_objects Number of currently allocated objects. Equals to /gc/heap/objects:objects. # TYPE go_memstats_heap_objects gauge go_memstats_heap_objects 5263 # HELP go_memstats_heap_released_bytes Number of heap bytes released to OS. Equals to /memory/classes/heap/released:bytes. # TYPE go_memstats_heap_released_bytes gauge go_memstats_heap_released_bytes 6.930432e+06 # HELP go_memstats_heap_sys_bytes Number of heap bytes obtained from system. Equals to /memory/classes/heap/objects:bytes + /memory/classes/heap/unused:bytes + /memory/classes/heap/released:bytes + /memory/classes/heap/free:bytes. # TYPE go_memstats_heap_sys_bytes gauge go_memstats_heap_sys_bytes 1.1993088e+07 # HELP go_memstats_last_gc_time_seconds Number of seconds since 1970 of last garbage collection. # TYPE go_memstats_last_gc_time_seconds gauge go_memstats_last_gc_time_seconds 1.7766148360954115e+09 # HELP go_memstats_mallocs_total Total number of heap objects allocated, both live and gc-ed. Semantically a counter version for go_memstats_heap_objects gauge. Equals to /gc/heap/allocs:objects + /gc/heap/tiny/allocs:objects. # TYPE go_memstats_mallocs_total counter go_memstats_mallocs_total 2.0802134e+07 # HELP go_memstats_mcache_inuse_bytes Number of bytes in use by mcache structures. Equals to /memory/classes/metadata/mcache/inuse:bytes. # TYPE go_memstats_mcache_inuse_bytes gauge go_memstats_mcache_inuse_bytes 2296 # HELP go_memstats_mcache_sys_bytes Number of bytes used for mcache structures obtained from system. Equals to /memory/classes/metadata/mcache/inuse:bytes + /memory/classes/metadata/mcache/free:bytes. # TYPE go_memstats_mcache_sys_bytes gauge go_memstats_mcache_sys_bytes 16072 # HELP go_memstats_mspan_inuse_bytes Number of bytes in use by mspan structures. Equals to /memory/classes/metadata/mspan/inuse:bytes. # TYPE go_memstats_mspan_inuse_bytes gauge go_memstats_mspan_inuse_bytes 60160 # HELP go_memstats_mspan_sys_bytes Number of bytes used for mspan structures obtained from system. Equals to /memory/classes/metadata/mspan/inuse:bytes + /memory/classes/metadata/mspan/free:bytes. # TYPE go_memstats_mspan_sys_bytes gauge go_memstats_mspan_sys_bytes 97920 # HELP go_memstats_next_gc_bytes Number of heap bytes when next garbage collection will take place. Equals to /gc/heap/goal:bytes. # TYPE go_memstats_next_gc_bytes gauge go_memstats_next_gc_bytes 4.194304e+06 # HELP go_memstats_other_sys_bytes Number of bytes used for other system allocations. Equals to /memory/classes/other:bytes. # TYPE go_memstats_other_sys_bytes gauge go_memstats_other_sys_bytes 765238 # HELP go_memstats_stack_inuse_bytes Number of bytes obtained from system for stack allocator in non-CGO environments. Equals to /memory/classes/heap/stacks:bytes. # TYPE go_memstats_stack_inuse_bytes gauge go_memstats_stack_inuse_bytes 589824 # HELP go_memstats_stack_sys_bytes Number of bytes obtained from system for stack allocator. Equals to /memory/classes/heap/stacks:bytes + /memory/classes/os-stacks:bytes. # TYPE go_memstats_stack_sys_bytes gauge go_memstats_stack_sys_bytes 589824 # HELP go_memstats_sys_bytes Number of bytes obtained from system. Equals to /memory/classes/total:byte. # TYPE go_memstats_sys_bytes gauge go_memstats_sys_bytes 1.8708744e+07 # HELP go_sched_gomaxprocs_threads The current runtime.GOMAXPROCS setting, or the number of operating system threads that can execute user-level Go code simultaneously. Sourced from /sched/gomaxprocs:threads. # TYPE go_sched_gomaxprocs_threads gauge go_sched_gomaxprocs_threads 1 # HELP go_threads Number of OS threads created. # TYPE go_threads gauge go_threads 7 # HELP node_arp_entries ARP entries by device # TYPE node_arp_entries gauge node_arp_entries{device="azv01f90a66eae"} 1 node_arp_entries{device="azv242b3abebdc"} 1 node_arp_entries{device="azv4cd4fa71bae"} 1 node_arp_entries{device="azv63a96d38340"} 1 node_arp_entries{device="azv6c83e846b95"} 1 node_arp_entries{device="azv8101e985a2a"} 1 node_arp_entries{device="azv870dd5437ed"} 1 node_arp_entries{device="azv99960e86340"} 1 node_arp_entries{device="azva068c73f524"} 1 node_arp_entries{device="azva34ff6c5426"} 1 node_arp_entries{device="azvd81ad55895f"} 1 node_arp_entries{device="azve117880b5fb"} 1 node_arp_entries{device="azve4dcd2566f8"} 1 node_arp_entries{device="azvf5d838218a5"} 1 node_arp_entries{device="eth0"} 6 # HELP node_boot_time_seconds Node boot time, in unixtime. # TYPE node_boot_time_seconds gauge node_boot_time_seconds 1.776557876e+09 # HELP node_context_switches_total Total number of context switches. # TYPE node_context_switches_total counter node_context_switches_total 3.85153333e+08 # HELP node_cooling_device_cur_state Current throttle state of the cooling device # TYPE node_cooling_device_cur_state gauge node_cooling_device_cur_state{name="0",type="Processor"} 0 node_cooling_device_cur_state{name="1",type="Processor"} 0 # HELP node_cooling_device_max_state Maximum throttle state of the cooling device # TYPE node_cooling_device_max_state gauge node_cooling_device_max_state{name="0",type="Processor"} 0 node_cooling_device_max_state{name="1",type="Processor"} 0 # HELP node_cpu_guest_seconds_total Seconds the CPUs spent in guests (VMs) for each mode. # TYPE node_cpu_guest_seconds_total counter node_cpu_guest_seconds_total{cpu="0",mode="nice"} 0 node_cpu_guest_seconds_total{cpu="0",mode="user"} 0 node_cpu_guest_seconds_total{cpu="1",mode="nice"} 0 node_cpu_guest_seconds_total{cpu="1",mode="user"} 0 # HELP node_cpu_seconds_total Seconds the CPUs spent in each mode. # TYPE node_cpu_seconds_total counter node_cpu_seconds_total{cpu="0",mode="idle"} 48372.4 node_cpu_seconds_total{cpu="0",mode="iowait"} 124.98 node_cpu_seconds_total{cpu="0",mode="irq"} 0 node_cpu_seconds_total{cpu="0",mode="nice"} 1.06 node_cpu_seconds_total{cpu="0",mode="softirq"} 211.69 node_cpu_seconds_total{cpu="0",mode="steal"} 0 node_cpu_seconds_total{cpu="0",mode="system"} 1985.35 node_cpu_seconds_total{cpu="0",mode="user"} 5352.03 node_cpu_seconds_total{cpu="1",mode="idle"} 48317.2 node_cpu_seconds_total{cpu="1",mode="iowait"} 129.4 node_cpu_seconds_total{cpu="1",mode="irq"} 0 node_cpu_seconds_total{cpu="1",mode="nice"} 1.22 node_cpu_seconds_total{cpu="1",mode="softirq"} 232.03 node_cpu_seconds_total{cpu="1",mode="steal"} 0 node_cpu_seconds_total{cpu="1",mode="system"} 2020.23 node_cpu_seconds_total{cpu="1",mode="user"} 5337.97 # HELP node_disk_discard_time_seconds_total This is the total number of seconds spent by all discards. # TYPE node_disk_discard_time_seconds_total counter node_disk_discard_time_seconds_total{device="sda"} 46.682 node_disk_discard_time_seconds_total{device="sdb"} 0 node_disk_discard_time_seconds_total{device="sr0"} 0 # HELP node_disk_discarded_sectors_total The total number of sectors discarded successfully. # TYPE node_disk_discarded_sectors_total counter node_disk_discarded_sectors_total{device="sda"} 2.31077924e+08 node_disk_discarded_sectors_total{device="sdb"} 0 node_disk_discarded_sectors_total{device="sr0"} 0 # HELP node_disk_discards_completed_total The total number of discards completed successfully. # TYPE node_disk_discards_completed_total counter node_disk_discards_completed_total{device="sda"} 18997 node_disk_discards_completed_total{device="sdb"} 0 node_disk_discards_completed_total{device="sr0"} 0 # HELP node_disk_discards_merged_total The total number of discards merged. # TYPE node_disk_discards_merged_total counter node_disk_discards_merged_total{device="sda"} 1357 node_disk_discards_merged_total{device="sdb"} 0 node_disk_discards_merged_total{device="sr0"} 0 # HELP node_disk_filesystem_info Info about disk filesystem. # TYPE node_disk_filesystem_info gauge node_disk_filesystem_info{device="sdb",type="ext4",usage="filesystem",uuid="312c34f6-ee4b-4f16-aa91-3f93f8985a7d",version="1.0"} 1 # HELP node_disk_flush_requests_time_seconds_total This is the total number of seconds spent by all flush requests. # TYPE node_disk_flush_requests_time_seconds_total counter node_disk_flush_requests_time_seconds_total{device="sda"} 0 node_disk_flush_requests_time_seconds_total{device="sdb"} 0 node_disk_flush_requests_time_seconds_total{device="sr0"} 0 # HELP node_disk_flush_requests_total The total number of flush requests completed successfully # TYPE node_disk_flush_requests_total counter node_disk_flush_requests_total{device="sda"} 0 node_disk_flush_requests_total{device="sdb"} 0 node_disk_flush_requests_total{device="sr0"} 0 # HELP node_disk_info Info of /sys/block/. # TYPE node_disk_info gauge node_disk_info{device="sda",major="8",minor="0",model="Virtual_Disk",path="acpi-MSFT1000:00-scsi-0:0:0:0",revision="1.0",rotational="1",serial="60022480b24f216c442803939c5d145b",wwn="0x60022480b24f216c"} 1 node_disk_info{device="sdb",major="8",minor="16",model="Virtual_Disk",path="acpi-MSFT1000:00-scsi-0:0:0:0",revision="1.0",rotational="1",serial="60022480d787258fdfa946766d1f4c53",wwn="0x60022480d787258f"} 1 node_disk_info{device="sr0",major="11",minor="0",model="Virtual_DVD-ROM",path="acpi-MSFT1000:00-scsi-0:0:0:2",revision="1.0",rotational="1",serial="4d534654202020207305e3437703544694957d7ced624a7d",wwn=""} 1 # HELP node_disk_io_now The number of I/Os currently in progress. # TYPE node_disk_io_now gauge node_disk_io_now{device="sda"} 0 node_disk_io_now{device="sdb"} 0 node_disk_io_now{device="sr0"} 0 # HELP node_disk_io_time_seconds_total Total seconds spent doing I/Os. # TYPE node_disk_io_time_seconds_total counter node_disk_io_time_seconds_total{device="sda"} 637.928 node_disk_io_time_seconds_total{device="sdb"} 152.38 node_disk_io_time_seconds_total{device="sr0"} 0.048 # HELP node_disk_io_time_weighted_seconds_total The weighted # of seconds spent doing I/Os. # TYPE node_disk_io_time_weighted_seconds_total counter node_disk_io_time_weighted_seconds_total{device="sda"} 2323.81 node_disk_io_time_weighted_seconds_total{device="sdb"} 441.022 node_disk_io_time_weighted_seconds_total{device="sr0"} 0.01 # HELP node_disk_read_bytes_total The total number of bytes read successfully. # TYPE node_disk_read_bytes_total counter node_disk_read_bytes_total{device="sda"} 3.355547648e+09 node_disk_read_bytes_total{device="sdb"} 2.05566976e+08 node_disk_read_bytes_total{device="sr0"} 282624 # HELP node_disk_read_time_seconds_total The total number of seconds spent by all reads. # TYPE node_disk_read_time_seconds_total counter node_disk_read_time_seconds_total{device="sda"} 371.04 node_disk_read_time_seconds_total{device="sdb"} 31.101 node_disk_read_time_seconds_total{device="sr0"} 0.01 # HELP node_disk_reads_completed_total The total number of reads completed successfully. # TYPE node_disk_reads_completed_total counter node_disk_reads_completed_total{device="sda"} 53415 node_disk_reads_completed_total{device="sdb"} 35543 node_disk_reads_completed_total{device="sr0"} 86 # HELP node_disk_reads_merged_total The total number of reads merged. # TYPE node_disk_reads_merged_total counter node_disk_reads_merged_total{device="sda"} 8099 node_disk_reads_merged_total{device="sdb"} 10686 node_disk_reads_merged_total{device="sr0"} 0 # HELP node_disk_write_time_seconds_total This is the total number of seconds spent by all writes. # TYPE node_disk_write_time_seconds_total counter node_disk_write_time_seconds_total{device="sda"} 1906.086 node_disk_write_time_seconds_total{device="sdb"} 409.921 node_disk_write_time_seconds_total{device="sr0"} 0 # HELP node_disk_writes_completed_total The total number of writes completed successfully. # TYPE node_disk_writes_completed_total counter node_disk_writes_completed_total{device="sda"} 247109 node_disk_writes_completed_total{device="sdb"} 48905 node_disk_writes_completed_total{device="sr0"} 0 # HELP node_disk_writes_merged_total The number of writes merged. # TYPE node_disk_writes_merged_total counter node_disk_writes_merged_total{device="sda"} 362613 node_disk_writes_merged_total{device="sdb"} 127280 node_disk_writes_merged_total{device="sr0"} 0 # HELP node_disk_written_bytes_total The total number of bytes written successfully. # TYPE node_disk_written_bytes_total counter node_disk_written_bytes_total{device="sda"} 1.6518288384e+10 node_disk_written_bytes_total{device="sdb"} 2.065334272e+09 node_disk_written_bytes_total{device="sr0"} 0 # HELP node_dmi_info A metric with a constant '1' value labeled by bios_date, bios_release, bios_vendor, bios_version, board_asset_tag, board_name, board_serial, board_vendor, board_version, chassis_asset_tag, chassis_serial, chassis_vendor, chassis_version, product_family, product_name, product_serial, product_sku, product_uuid, product_version, system_vendor if provided by DMI. # TYPE node_dmi_info gauge node_dmi_info{bios_date="01/08/2026",bios_release="4.1",bios_vendor="Microsoft Corporation",bios_version="Hyper-V UEFI Release v4.1",board_asset_tag="None",board_name="Virtual Machine",board_vendor="Microsoft Corporation",board_version="Hyper-V UEFI Release v4.1",chassis_asset_tag="7783-7084-3265-9085-8269-3286-77",chassis_vendor="Microsoft Corporation",chassis_version="Hyper-V UEFI Release v4.1",product_family="Virtual Machine",product_name="Virtual Machine",product_sku="None",product_uuid="e7458ee1-7a43-4c2a-adee-515b5f699ae2",product_version="Hyper-V UEFI Release v4.1",system_vendor="Microsoft Corporation"} 1 # HELP node_entropy_available_bits Bits of available entropy. # TYPE node_entropy_available_bits gauge node_entropy_available_bits 256 # HELP node_entropy_pool_size_bits Bits of entropy pool. # TYPE node_entropy_pool_size_bits gauge node_entropy_pool_size_bits 256 # HELP node_exporter_build_info A metric with a constant '1' value labeled by version, revision, branch, goversion from which node_exporter was built, and the goos and goarch for the build. # TYPE node_exporter_build_info gauge node_exporter_build_info{branch="HEAD",goarch="amd64",goos="linux",goversion="go1.26.1",revision="0dd664dece3f8319f6bec5a221acd2c7ad13a23d",tags="unknown",version="1.11.1"} 1 # HELP node_filefd_allocated File descriptor statistics: allocated. # TYPE node_filefd_allocated gauge node_filefd_allocated 4480 # HELP node_filefd_maximum File descriptor statistics: maximum. # TYPE node_filefd_maximum gauge node_filefd_maximum 9.223372036854776e+18 # HELP node_filesystem_avail_bytes Filesystem space available to non-root users in bytes. # TYPE node_filesystem_avail_bytes gauge node_filesystem_avail_bytes{device="/dev/root",device_error="",fstype="ext4",mountpoint="/"} 1.01378854912e+11 node_filesystem_avail_bytes{device="/dev/root",device_error="",fstype="ext4",mountpoint="/var/lib/kubelet"} 1.01378854912e+11 node_filesystem_avail_bytes{device="/dev/sda15",device_error="",fstype="vfat",mountpoint="/boot/efi"} 1.03053312e+08 node_filesystem_avail_bytes{device="none",device_error="",fstype="ramfs",mountpoint="/run/credentials/systemd-sysusers.service"} 0 node_filesystem_avail_bytes{device="tmpfs",device_error="",fstype="tmpfs",mountpoint="/run"} 1.658187776e+09 node_filesystem_avail_bytes{device="tmpfs",device_error="",fstype="tmpfs",mountpoint="/run/lock"} 5.24288e+06 # HELP node_filesystem_device_error Whether an error occurred while getting statistics for the given device. # TYPE node_filesystem_device_error gauge node_filesystem_device_error{device="/dev/root",device_error="",fstype="ext4",mountpoint="/"} 0 node_filesystem_device_error{device="/dev/root",device_error="",fstype="ext4",mountpoint="/var/lib/kubelet"} 0 node_filesystem_device_error{device="/dev/sda15",device_error="",fstype="vfat",mountpoint="/boot/efi"} 0 node_filesystem_device_error{device="none",device_error="",fstype="ramfs",mountpoint="/run/credentials/systemd-sysusers.service"} 0 node_filesystem_device_error{device="tmpfs",device_error="",fstype="tmpfs",mountpoint="/run"} 0 node_filesystem_device_error{device="tmpfs",device_error="",fstype="tmpfs",mountpoint="/run/lock"} 0 # HELP node_filesystem_files Filesystem total file nodes. # TYPE node_filesystem_files gauge node_filesystem_files{device="/dev/root",device_error="",fstype="ext4",mountpoint="/"} 1.6515072e+07 node_filesystem_files{device="/dev/root",device_error="",fstype="ext4",mountpoint="/var/lib/kubelet"} 1.6515072e+07 node_filesystem_files{device="/dev/sda15",device_error="",fstype="vfat",mountpoint="/boot/efi"} 0 node_filesystem_files{device="none",device_error="",fstype="ramfs",mountpoint="/run/credentials/systemd-sysusers.service"} 0 node_filesystem_files{device="tmpfs",device_error="",fstype="tmpfs",mountpoint="/run"} 819200 node_filesystem_files{device="tmpfs",device_error="",fstype="tmpfs",mountpoint="/run/lock"} 1.015713e+06 # HELP node_filesystem_files_free Filesystem total free file nodes. # TYPE node_filesystem_files_free gauge node_filesystem_files_free{device="/dev/root",device_error="",fstype="ext4",mountpoint="/"} 1.6206007e+07 node_filesystem_files_free{device="/dev/root",device_error="",fstype="ext4",mountpoint="/var/lib/kubelet"} 1.6206007e+07 node_filesystem_files_free{device="/dev/sda15",device_error="",fstype="vfat",mountpoint="/boot/efi"} 0 node_filesystem_files_free{device="none",device_error="",fstype="ramfs",mountpoint="/run/credentials/systemd-sysusers.service"} 0 node_filesystem_files_free{device="tmpfs",device_error="",fstype="tmpfs",mountpoint="/run"} 816739 node_filesystem_files_free{device="tmpfs",device_error="",fstype="tmpfs",mountpoint="/run/lock"} 1.015709e+06 # HELP node_filesystem_free_bytes Filesystem free space in bytes. # TYPE node_filesystem_free_bytes gauge node_filesystem_free_bytes{device="/dev/root",device_error="",fstype="ext4",mountpoint="/"} 1.01395632128e+11 node_filesystem_free_bytes{device="/dev/root",device_error="",fstype="ext4",mountpoint="/var/lib/kubelet"} 1.01395632128e+11 node_filesystem_free_bytes{device="/dev/sda15",device_error="",fstype="vfat",mountpoint="/boot/efi"} 1.03053312e+08 node_filesystem_free_bytes{device="none",device_error="",fstype="ramfs",mountpoint="/run/credentials/systemd-sysusers.service"} 0 node_filesystem_free_bytes{device="tmpfs",device_error="",fstype="tmpfs",mountpoint="/run"} 1.658187776e+09 node_filesystem_free_bytes{device="tmpfs",device_error="",fstype="tmpfs",mountpoint="/run/lock"} 5.24288e+06 # HELP node_filesystem_mount_info Filesystem mount information. # TYPE node_filesystem_mount_info gauge node_filesystem_mount_info{device="/dev/root",major="8",minor="1",mountpoint="/"} 1 node_filesystem_mount_info{device="/dev/root",major="8",minor="1",mountpoint="/var/lib/kubelet"} 1 node_filesystem_mount_info{device="/dev/sda15",major="8",minor="15",mountpoint="/boot/efi"} 1 node_filesystem_mount_info{device="none",major="0",minor="35",mountpoint="/run/credentials/systemd-sysusers.service"} 1 node_filesystem_mount_info{device="tmpfs",major="0",minor="26",mountpoint="/run"} 1 node_filesystem_mount_info{device="tmpfs",major="0",minor="27",mountpoint="/run/lock"} 1 # HELP node_filesystem_purgeable_bytes Filesystem space available including purgeable space (MacOS specific). # TYPE node_filesystem_purgeable_bytes gauge node_filesystem_purgeable_bytes{device="/dev/root",device_error="",fstype="ext4",mountpoint="/"} 0 node_filesystem_purgeable_bytes{device="/dev/root",device_error="",fstype="ext4",mountpoint="/var/lib/kubelet"} 0 node_filesystem_purgeable_bytes{device="/dev/sda15",device_error="",fstype="vfat",mountpoint="/boot/efi"} 0 node_filesystem_purgeable_bytes{device="none",device_error="",fstype="ramfs",mountpoint="/run/credentials/systemd-sysusers.service"} 0 node_filesystem_purgeable_bytes{device="tmpfs",device_error="",fstype="tmpfs",mountpoint="/run"} 0 node_filesystem_purgeable_bytes{device="tmpfs",device_error="",fstype="tmpfs",mountpoint="/run/lock"} 0 # HELP node_filesystem_readonly Filesystem read-only status. # TYPE node_filesystem_readonly gauge node_filesystem_readonly{device="/dev/root",device_error="",fstype="ext4",mountpoint="/"} 0 node_filesystem_readonly{device="/dev/root",device_error="",fstype="ext4",mountpoint="/var/lib/kubelet"} 0 node_filesystem_readonly{device="/dev/sda15",device_error="",fstype="vfat",mountpoint="/boot/efi"} 0 node_filesystem_readonly{device="none",device_error="",fstype="ramfs",mountpoint="/run/credentials/systemd-sysusers.service"} 0 node_filesystem_readonly{device="tmpfs",device_error="",fstype="tmpfs",mountpoint="/run"} 0 node_filesystem_readonly{device="tmpfs",device_error="",fstype="tmpfs",mountpoint="/run/lock"} 0 # HELP node_filesystem_size_bytes Filesystem size in bytes. # TYPE node_filesystem_size_bytes gauge node_filesystem_size_bytes{device="/dev/root",device_error="",fstype="ext4",mountpoint="/"} 1.33003395072e+11 node_filesystem_size_bytes{device="/dev/root",device_error="",fstype="ext4",mountpoint="/var/lib/kubelet"} 1.33003395072e+11 node_filesystem_size_bytes{device="/dev/sda15",device_error="",fstype="vfat",mountpoint="/boot/efi"} 1.09395456e+08 node_filesystem_size_bytes{device="none",device_error="",fstype="ramfs",mountpoint="/run/credentials/systemd-sysusers.service"} 0 node_filesystem_size_bytes{device="tmpfs",device_error="",fstype="tmpfs",mountpoint="/run"} 1.664147456e+09 node_filesystem_size_bytes{device="tmpfs",device_error="",fstype="tmpfs",mountpoint="/run/lock"} 5.24288e+06 # HELP node_forks_total Total number of forks. # TYPE node_forks_total counter node_forks_total 1.087593e+06 # HELP node_infiniband_duplicate_requests_packets_total The number of received packets. A duplicate request is a request that had been previously executed. # TYPE node_infiniband_duplicate_requests_packets_total counter node_infiniband_duplicate_requests_packets_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_excessive_buffer_overrun_errors_total Number of times that OverrunErrors consecutive flow control update periods occurred, each having at least one overrun error. # TYPE node_infiniband_excessive_buffer_overrun_errors_total counter node_infiniband_excessive_buffer_overrun_errors_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_implied_nak_seq_errors_total The number of time the requested decided an ACK. with a PSN larger than the expected PSN for an RDMA read or response. # TYPE node_infiniband_implied_nak_seq_errors_total counter node_infiniband_implied_nak_seq_errors_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_info Non-numeric data from /sys/class/infiniband/, value is always 1. # TYPE node_infiniband_info gauge node_infiniband_info{board_id="MSF0010110035",device="mlx5_0",firmware_version="14.30.5026",hca_type="MT4118"} 1 # HELP node_infiniband_lifespan_seconds The maximum period in ms which defines the aging of the counter reads. Two consecutive reads within this period might return the same values. # TYPE node_infiniband_lifespan_seconds gauge node_infiniband_lifespan_seconds{device="mlx5_0",port="1"} 0 # HELP node_infiniband_link_downed_total Number of times the link failed to recover from an error state and went down # TYPE node_infiniband_link_downed_total counter node_infiniband_link_downed_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_link_error_recovery_total Number of times the link successfully recovered from an error state # TYPE node_infiniband_link_error_recovery_total counter node_infiniband_link_error_recovery_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_local_ack_timeout_errors_total The number of times QP's ack timer expired for RC, XRC, DCT QPs at the sender side. The QP retry limit was not exceed, therefore it is still recoverable error. # TYPE node_infiniband_local_ack_timeout_errors_total counter node_infiniband_local_ack_timeout_errors_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_local_link_integrity_errors_total Number of times that the count of local physical errors exceeded the threshold specified by LocalPhyErrors. # TYPE node_infiniband_local_link_integrity_errors_total counter node_infiniband_local_link_integrity_errors_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_multicast_packets_received_total Number of multicast packets received (including errors) # TYPE node_infiniband_multicast_packets_received_total counter node_infiniband_multicast_packets_received_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_multicast_packets_transmitted_total Number of multicast packets transmitted (including errors) # TYPE node_infiniband_multicast_packets_transmitted_total counter node_infiniband_multicast_packets_transmitted_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_out_of_buffer_drops_total The number of drops occurred due to lack of WQE for the associated QPs. # TYPE node_infiniband_out_of_buffer_drops_total counter node_infiniband_out_of_buffer_drops_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_out_of_sequence_packets_received_total The number of out of sequence packets received. # TYPE node_infiniband_out_of_sequence_packets_received_total counter node_infiniband_out_of_sequence_packets_received_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_packet_sequence_errors_total The number of received NAK sequence error packets. The QP retry limit was not exceeded. # TYPE node_infiniband_packet_sequence_errors_total counter node_infiniband_packet_sequence_errors_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_physical_state_id Physical state of the InfiniBand port (0: no change, 1: sleep, 2: polling, 3: disable, 4: shift, 5: link up, 6: link error recover, 7: phytest) # TYPE node_infiniband_physical_state_id gauge node_infiniband_physical_state_id{device="mlx5_0",port="1"} 5 # HELP node_infiniband_port_constraint_errors_received_total Number of packets received on the switch physical port that are discarded # TYPE node_infiniband_port_constraint_errors_received_total counter node_infiniband_port_constraint_errors_received_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_port_constraint_errors_transmitted_total Number of packets not transmitted from the switch physical port # TYPE node_infiniband_port_constraint_errors_transmitted_total counter node_infiniband_port_constraint_errors_transmitted_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_port_data_received_bytes_total Number of data octets received on all links # TYPE node_infiniband_port_data_received_bytes_total counter node_infiniband_port_data_received_bytes_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_port_data_transmitted_bytes_total Number of data octets transmitted on all links # TYPE node_infiniband_port_data_transmitted_bytes_total counter node_infiniband_port_data_transmitted_bytes_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_port_discards_transmitted_total Number of outbound packets discarded by the port because the port is down or congested # TYPE node_infiniband_port_discards_transmitted_total counter node_infiniband_port_discards_transmitted_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_port_errors_received_total Number of packets containing an error that were received on this port # TYPE node_infiniband_port_errors_received_total counter node_infiniband_port_errors_received_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_port_packets_received_total Number of packets received on all VLs by this port (including errors) # TYPE node_infiniband_port_packets_received_total counter node_infiniband_port_packets_received_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_port_packets_transmitted_total Number of packets transmitted on all VLs from this port (including errors) # TYPE node_infiniband_port_packets_transmitted_total counter node_infiniband_port_packets_transmitted_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_port_receive_remote_physical_errors_total Number of packets marked with the EBP (End of Bad Packet) delimiter received on the port. # TYPE node_infiniband_port_receive_remote_physical_errors_total counter node_infiniband_port_receive_remote_physical_errors_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_port_receive_switch_relay_errors_total Number of packets that could not be forwarded by the switch. # TYPE node_infiniband_port_receive_switch_relay_errors_total counter node_infiniband_port_receive_switch_relay_errors_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_port_transmit_wait_total Number of ticks during which the port had data to transmit but no data was sent during the entire tick # TYPE node_infiniband_port_transmit_wait_total counter node_infiniband_port_transmit_wait_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_rate_bytes_per_second Maximum signal transfer rate # TYPE node_infiniband_rate_bytes_per_second gauge node_infiniband_rate_bytes_per_second{device="mlx5_0",port="1"} 5e+09 # HELP node_infiniband_req_cqes_errors_total The number of times requester detected CQEs completed with errors. The counters was added in MLNX_OFED 4.1 # TYPE node_infiniband_req_cqes_errors_total counter node_infiniband_req_cqes_errors_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_req_cqes_flush_errors_total The number of times requester detected CQEs completed with flushed errors. The counters was added in MLNX_OFED 4.1 # TYPE node_infiniband_req_cqes_flush_errors_total counter node_infiniband_req_cqes_flush_errors_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_req_remote_access_errors_total The number of times requester detected remote access errors. The counters was added in MLNX_OFED 4.1 # TYPE node_infiniband_req_remote_access_errors_total counter node_infiniband_req_remote_access_errors_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_req_remote_invalid_request_errors_total The number of times requester detected remote invalid request errors. The counters was added in MLNX_OFED 4.1 # TYPE node_infiniband_req_remote_invalid_request_errors_total counter node_infiniband_req_remote_invalid_request_errors_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_resp_cqes_errors_total The number of times responder detected CQEs completed with errors. The counters was added in MLNX_OFED 4.1 # TYPE node_infiniband_resp_cqes_errors_total counter node_infiniband_resp_cqes_errors_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_resp_cqes_flush_errors_total The number of times responder detected CQEs completed with flushed errors. The counters was added in MLNX_OFED 4.1 # TYPE node_infiniband_resp_cqes_flush_errors_total counter node_infiniband_resp_cqes_flush_errors_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_resp_local_length_errors_total The number of times responder detected local length errors. The counters was added in MLNX_OFED 4.1 # TYPE node_infiniband_resp_local_length_errors_total counter node_infiniband_resp_local_length_errors_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_resp_remote_access_errors_total The number of times responder detected remote access errors. The counters was added in MLNX_OFED 4.1 # TYPE node_infiniband_resp_remote_access_errors_total counter node_infiniband_resp_remote_access_errors_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_rnr_nak_retry_packets_received_total The number of received RNR NAK packets. The QP retry limit was not exceeded. # TYPE node_infiniband_rnr_nak_retry_packets_received_total counter node_infiniband_rnr_nak_retry_packets_received_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_rx_atomic_requests_total The number of received ATOMIC request for the associated QPs. # TYPE node_infiniband_rx_atomic_requests_total counter node_infiniband_rx_atomic_requests_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_rx_read_requests_total The number of received READ requests for the associated QPs. # TYPE node_infiniband_rx_read_requests_total counter node_infiniband_rx_read_requests_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_rx_write_requests_total The number of received WRITE requests for the associated QPs. # TYPE node_infiniband_rx_write_requests_total counter node_infiniband_rx_write_requests_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_state_id State of the InfiniBand port (0: no change, 1: down, 2: init, 3: armed, 4: active, 5: act defer) # TYPE node_infiniband_state_id gauge node_infiniband_state_id{device="mlx5_0",port="1"} 4 # HELP node_infiniband_symbol_error_total Number of minor link errors detected on one or more physical lanes. # TYPE node_infiniband_symbol_error_total counter node_infiniband_symbol_error_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_unicast_packets_received_total Number of unicast packets received (including errors) # TYPE node_infiniband_unicast_packets_received_total counter node_infiniband_unicast_packets_received_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_unicast_packets_transmitted_total Number of unicast packets transmitted (including errors) # TYPE node_infiniband_unicast_packets_transmitted_total counter node_infiniband_unicast_packets_transmitted_total{device="mlx5_0",port="1"} 0 # HELP node_infiniband_vl15_dropped_total Number of incoming VL15 packets dropped due to resource limitations. # TYPE node_infiniband_vl15_dropped_total counter node_infiniband_vl15_dropped_total{device="mlx5_0",port="1"} 0 # HELP node_intr_total Total number of interrupts serviced. # TYPE node_intr_total counter node_intr_total 6.7735857e+07 # HELP node_load1 1m load average. # TYPE node_load1 gauge node_load1 0.38 # HELP node_load15 15m load average. # TYPE node_load15 gauge node_load15 0.58 # HELP node_load5 5m load average. # TYPE node_load5 gauge node_load5 0.58 # HELP node_memory_Active_anon_bytes Memory information field Active_anon_bytes. # TYPE node_memory_Active_anon_bytes gauge node_memory_Active_anon_bytes 5.828608e+06 # HELP node_memory_Active_bytes Memory information field Active_bytes. # TYPE node_memory_Active_bytes gauge node_memory_Active_bytes 1.80549632e+09 # HELP node_memory_Active_file_bytes Memory information field Active_file_bytes. # TYPE node_memory_Active_file_bytes gauge node_memory_Active_file_bytes 1.799667712e+09 # HELP node_memory_AnonHugePages_bytes Memory information field AnonHugePages_bytes. # TYPE node_memory_AnonHugePages_bytes gauge node_memory_AnonHugePages_bytes 1.503657984e+09 # HELP node_memory_AnonPages_bytes Memory information field AnonPages_bytes. # TYPE node_memory_AnonPages_bytes gauge node_memory_AnonPages_bytes 2.453245952e+09 # HELP node_memory_Bounce_bytes Memory information field Bounce_bytes. # TYPE node_memory_Bounce_bytes gauge node_memory_Bounce_bytes 0 # HELP node_memory_Buffers_bytes Memory information field Buffers_bytes. # TYPE node_memory_Buffers_bytes gauge node_memory_Buffers_bytes 4.3825152e+08 # HELP node_memory_Cached_bytes Memory information field Cached_bytes. # TYPE node_memory_Cached_bytes gauge node_memory_Cached_bytes 3.789799424e+09 # HELP node_memory_CommitLimit_bytes Memory information field CommitLimit_bytes. # TYPE node_memory_CommitLimit_bytes gauge node_memory_CommitLimit_bytes 4.160360448e+09 # HELP node_memory_Committed_AS_bytes Memory information field Committed_AS_bytes. # TYPE node_memory_Committed_AS_bytes gauge node_memory_Committed_AS_bytes 1.1845791744e+10 # HELP node_memory_DirectMap1G_bytes Memory information field DirectMap1G_bytes. # TYPE node_memory_DirectMap1G_bytes gauge node_memory_DirectMap1G_bytes 3.221225472e+09 # HELP node_memory_DirectMap2M_bytes Memory information field DirectMap2M_bytes. # TYPE node_memory_DirectMap2M_bytes gauge node_memory_DirectMap2M_bytes 7.17225984e+09 # HELP node_memory_DirectMap4k_bytes Memory information field DirectMap4k_bytes. # TYPE node_memory_DirectMap4k_bytes gauge node_memory_DirectMap4k_bytes 3.3882112e+08 # HELP node_memory_Dirty_bytes Memory information field Dirty_bytes. # TYPE node_memory_Dirty_bytes gauge node_memory_Dirty_bytes 483328 # HELP node_memory_HardwareCorrupted_bytes Memory information field HardwareCorrupted_bytes. # TYPE node_memory_HardwareCorrupted_bytes gauge node_memory_HardwareCorrupted_bytes 0 # HELP node_memory_HugePages_Free Memory information field HugePages_Free. # TYPE node_memory_HugePages_Free gauge node_memory_HugePages_Free 0 # HELP node_memory_HugePages_Rsvd Memory information field HugePages_Rsvd. # TYPE node_memory_HugePages_Rsvd gauge node_memory_HugePages_Rsvd 0 # HELP node_memory_HugePages_Surp Memory information field HugePages_Surp. # TYPE node_memory_HugePages_Surp gauge node_memory_HugePages_Surp 0 # HELP node_memory_HugePages_Total Memory information field HugePages_Total. # TYPE node_memory_HugePages_Total gauge node_memory_HugePages_Total 0 # HELP node_memory_Hugepagesize_bytes Memory information field Hugepagesize_bytes. # TYPE node_memory_Hugepagesize_bytes gauge node_memory_Hugepagesize_bytes 2.097152e+06 # HELP node_memory_Inactive_anon_bytes Memory information field Inactive_anon_bytes. # TYPE node_memory_Inactive_anon_bytes gauge node_memory_Inactive_anon_bytes 2.63417856e+09 # HELP node_memory_Inactive_bytes Memory information field Inactive_bytes. # TYPE node_memory_Inactive_bytes gauge node_memory_Inactive_bytes 5.046976512e+09 # HELP node_memory_Inactive_file_bytes Memory information field Inactive_file_bytes. # TYPE node_memory_Inactive_file_bytes gauge node_memory_Inactive_file_bytes 2.412797952e+09 # HELP node_memory_KernelStack_bytes Memory information field KernelStack_bytes. # TYPE node_memory_KernelStack_bytes gauge node_memory_KernelStack_bytes 2.0840448e+07 # HELP node_memory_Mapped_bytes Memory information field Mapped_bytes. # TYPE node_memory_Mapped_bytes gauge node_memory_Mapped_bytes 1.540235264e+09 # HELP node_memory_MemAvailable_bytes Memory information field MemAvailable_bytes. # TYPE node_memory_MemAvailable_bytes gauge node_memory_MemAvailable_bytes 4.933636096e+09 # HELP node_memory_MemFree_bytes Memory information field MemFree_bytes. # TYPE node_memory_MemFree_bytes gauge node_memory_MemFree_bytes 4.37977088e+08 # HELP node_memory_MemTotal_bytes Memory information field MemTotal_bytes. # TYPE node_memory_MemTotal_bytes gauge node_memory_MemTotal_bytes 8.320724992e+09 # HELP node_memory_Mlocked_bytes Memory information field Mlocked_bytes. # TYPE node_memory_Mlocked_bytes gauge node_memory_Mlocked_bytes 2.8315648e+07 # HELP node_memory_NFS_Unstable_bytes Memory information field NFS_Unstable_bytes. # TYPE node_memory_NFS_Unstable_bytes gauge node_memory_NFS_Unstable_bytes 0 # HELP node_memory_PageTables_bytes Memory information field PageTables_bytes. # TYPE node_memory_PageTables_bytes gauge node_memory_PageTables_bytes 2.49856e+07 # HELP node_memory_Percpu_bytes Memory information field Percpu_bytes. # TYPE node_memory_Percpu_bytes gauge node_memory_Percpu_bytes 2.097152e+06 # HELP node_memory_SReclaimable_bytes Memory information field SReclaimable_bytes. # TYPE node_memory_SReclaimable_bytes gauge node_memory_SReclaimable_bytes 6.0409856e+08 # HELP node_memory_SUnreclaim_bytes Memory information field SUnreclaim_bytes. # TYPE node_memory_SUnreclaim_bytes gauge node_memory_SUnreclaim_bytes 1.27574016e+08 # HELP node_memory_ShmemHugePages_bytes Memory information field ShmemHugePages_bytes. # TYPE node_memory_ShmemHugePages_bytes gauge node_memory_ShmemHugePages_bytes 0 # HELP node_memory_ShmemPmdMapped_bytes Memory information field ShmemPmdMapped_bytes. # TYPE node_memory_ShmemPmdMapped_bytes gauge node_memory_ShmemPmdMapped_bytes 0 # HELP node_memory_Shmem_bytes Memory information field Shmem_bytes. # TYPE node_memory_Shmem_bytes gauge node_memory_Shmem_bytes 6.303744e+06 # HELP node_memory_Slab_bytes Memory information field Slab_bytes. # TYPE node_memory_Slab_bytes gauge node_memory_Slab_bytes 7.31672576e+08 # HELP node_memory_SwapCached_bytes Memory information field SwapCached_bytes. # TYPE node_memory_SwapCached_bytes gauge node_memory_SwapCached_bytes 0 # HELP node_memory_SwapFree_bytes Memory information field SwapFree_bytes. # TYPE node_memory_SwapFree_bytes gauge node_memory_SwapFree_bytes 0 # HELP node_memory_SwapTotal_bytes Memory information field SwapTotal_bytes. # TYPE node_memory_SwapTotal_bytes gauge node_memory_SwapTotal_bytes 0 # HELP node_memory_Unevictable_bytes Memory information field Unevictable_bytes. # TYPE node_memory_Unevictable_bytes gauge node_memory_Unevictable_bytes 2.8315648e+07 # HELP node_memory_VmallocChunk_bytes Memory information field VmallocChunk_bytes. # TYPE node_memory_VmallocChunk_bytes gauge node_memory_VmallocChunk_bytes 0 # HELP node_memory_VmallocTotal_bytes Memory information field VmallocTotal_bytes. # TYPE node_memory_VmallocTotal_bytes gauge node_memory_VmallocTotal_bytes 3.5184372087808e+13 # HELP node_memory_VmallocUsed_bytes Memory information field VmallocUsed_bytes. # TYPE node_memory_VmallocUsed_bytes gauge node_memory_VmallocUsed_bytes 1.12082944e+08 # HELP node_memory_WritebackTmp_bytes Memory information field WritebackTmp_bytes. # TYPE node_memory_WritebackTmp_bytes gauge node_memory_WritebackTmp_bytes 0 # HELP node_memory_Writeback_bytes Memory information field Writeback_bytes. # TYPE node_memory_Writeback_bytes gauge node_memory_Writeback_bytes 0 # HELP node_netstat_Icmp6_InErrors Statistic Icmp6InErrors. # TYPE node_netstat_Icmp6_InErrors untyped node_netstat_Icmp6_InErrors 0 # HELP node_netstat_Icmp6_InMsgs Statistic Icmp6InMsgs. # TYPE node_netstat_Icmp6_InMsgs untyped node_netstat_Icmp6_InMsgs 0 # HELP node_netstat_Icmp6_OutMsgs Statistic Icmp6OutMsgs. # TYPE node_netstat_Icmp6_OutMsgs untyped node_netstat_Icmp6_OutMsgs 183 # HELP node_netstat_Icmp_InErrors Statistic IcmpInErrors. # TYPE node_netstat_Icmp_InErrors untyped node_netstat_Icmp_InErrors 1 # HELP node_netstat_Icmp_InMsgs Statistic IcmpInMsgs. # TYPE node_netstat_Icmp_InMsgs untyped node_netstat_Icmp_InMsgs 123 # HELP node_netstat_Icmp_OutMsgs Statistic IcmpOutMsgs. # TYPE node_netstat_Icmp_OutMsgs untyped node_netstat_Icmp_OutMsgs 16995 # HELP node_netstat_Ip6_InOctets Statistic Ip6InOctets. # TYPE node_netstat_Ip6_InOctets untyped node_netstat_Ip6_InOctets 7896 # HELP node_netstat_Ip6_OutOctets Statistic Ip6OutOctets. # TYPE node_netstat_Ip6_OutOctets untyped node_netstat_Ip6_OutOctets 17672 # HELP node_netstat_IpExt_InOctets Statistic IpExtInOctets. # TYPE node_netstat_IpExt_InOctets untyped node_netstat_IpExt_InOctets 2.9175025245e+10 # HELP node_netstat_IpExt_OutOctets Statistic IpExtOutOctets. # TYPE node_netstat_IpExt_OutOctets untyped node_netstat_IpExt_OutOctets 5.4710726946e+10 # HELP node_netstat_Ip_Forwarding Statistic IpForwarding. # TYPE node_netstat_Ip_Forwarding untyped node_netstat_Ip_Forwarding 1 # HELP node_netstat_TcpExt_ListenDrops Statistic TcpExtListenDrops. # TYPE node_netstat_TcpExt_ListenDrops untyped node_netstat_TcpExt_ListenDrops 0 # HELP node_netstat_TcpExt_ListenOverflows Statistic TcpExtListenOverflows. # TYPE node_netstat_TcpExt_ListenOverflows untyped node_netstat_TcpExt_ListenOverflows 0 # HELP node_netstat_TcpExt_SyncookiesFailed Statistic TcpExtSyncookiesFailed. # TYPE node_netstat_TcpExt_SyncookiesFailed untyped node_netstat_TcpExt_SyncookiesFailed 0 # HELP node_netstat_TcpExt_SyncookiesRecv Statistic TcpExtSyncookiesRecv. # TYPE node_netstat_TcpExt_SyncookiesRecv untyped node_netstat_TcpExt_SyncookiesRecv 0 # HELP node_netstat_TcpExt_SyncookiesSent Statistic TcpExtSyncookiesSent. # TYPE node_netstat_TcpExt_SyncookiesSent untyped node_netstat_TcpExt_SyncookiesSent 0 # HELP node_netstat_TcpExt_TCPOFOQueue Statistic TcpExtTCPOFOQueue. # TYPE node_netstat_TcpExt_TCPOFOQueue untyped node_netstat_TcpExt_TCPOFOQueue 14757 # HELP node_netstat_TcpExt_TCPRcvQDrop Statistic TcpExtTCPRcvQDrop. # TYPE node_netstat_TcpExt_TCPRcvQDrop untyped node_netstat_TcpExt_TCPRcvQDrop 0 # HELP node_netstat_TcpExt_TCPSynRetrans Statistic TcpExtTCPSynRetrans. # TYPE node_netstat_TcpExt_TCPSynRetrans untyped node_netstat_TcpExt_TCPSynRetrans 0 # HELP node_netstat_TcpExt_TCPTimeouts Statistic TcpExtTCPTimeouts. # TYPE node_netstat_TcpExt_TCPTimeouts untyped node_netstat_TcpExt_TCPTimeouts 115 # HELP node_netstat_Tcp_ActiveOpens Statistic TcpActiveOpens. # TYPE node_netstat_Tcp_ActiveOpens untyped node_netstat_Tcp_ActiveOpens 324894 # HELP node_netstat_Tcp_CurrEstab Statistic TcpCurrEstab. # TYPE node_netstat_Tcp_CurrEstab untyped node_netstat_Tcp_CurrEstab 24 # HELP node_netstat_Tcp_InErrs Statistic TcpInErrs. # TYPE node_netstat_Tcp_InErrs untyped node_netstat_Tcp_InErrs 0 # HELP node_netstat_Tcp_InSegs Statistic TcpInSegs. # TYPE node_netstat_Tcp_InSegs untyped node_netstat_Tcp_InSegs 4.434276e+06 # HELP node_netstat_Tcp_OutRsts Statistic TcpOutRsts. # TYPE node_netstat_Tcp_OutRsts untyped node_netstat_Tcp_OutRsts 36996 # HELP node_netstat_Tcp_OutSegs Statistic TcpOutSegs. # TYPE node_netstat_Tcp_OutSegs untyped node_netstat_Tcp_OutSegs 5.187183e+06 # HELP node_netstat_Tcp_PassiveOpens Statistic TcpPassiveOpens. # TYPE node_netstat_Tcp_PassiveOpens untyped node_netstat_Tcp_PassiveOpens 121425 # HELP node_netstat_Tcp_RetransSegs Statistic TcpRetransSegs. # TYPE node_netstat_Tcp_RetransSegs untyped node_netstat_Tcp_RetransSegs 758 # HELP node_netstat_Udp6_InDatagrams Statistic Udp6InDatagrams. # TYPE node_netstat_Udp6_InDatagrams untyped node_netstat_Udp6_InDatagrams 0 # HELP node_netstat_Udp6_InErrors Statistic Udp6InErrors. # TYPE node_netstat_Udp6_InErrors untyped node_netstat_Udp6_InErrors 0 # HELP node_netstat_Udp6_NoPorts Statistic Udp6NoPorts. # TYPE node_netstat_Udp6_NoPorts untyped node_netstat_Udp6_NoPorts 0 # HELP node_netstat_Udp6_OutDatagrams Statistic Udp6OutDatagrams. # TYPE node_netstat_Udp6_OutDatagrams untyped node_netstat_Udp6_OutDatagrams 0 # HELP node_netstat_Udp6_RcvbufErrors Statistic Udp6RcvbufErrors. # TYPE node_netstat_Udp6_RcvbufErrors untyped node_netstat_Udp6_RcvbufErrors 0 # HELP node_netstat_Udp6_SndbufErrors Statistic Udp6SndbufErrors. # TYPE node_netstat_Udp6_SndbufErrors untyped node_netstat_Udp6_SndbufErrors 0 # HELP node_netstat_UdpLite6_InErrors Statistic UdpLite6InErrors. # TYPE node_netstat_UdpLite6_InErrors untyped node_netstat_UdpLite6_InErrors 0 # HELP node_netstat_UdpLite_InErrors Statistic UdpLiteInErrors. # TYPE node_netstat_UdpLite_InErrors untyped node_netstat_UdpLite_InErrors 0 # HELP node_netstat_Udp_InDatagrams Statistic UdpInDatagrams. # TYPE node_netstat_Udp_InDatagrams untyped node_netstat_Udp_InDatagrams 2783 # HELP node_netstat_Udp_InErrors Statistic UdpInErrors. # TYPE node_netstat_Udp_InErrors untyped node_netstat_Udp_InErrors 0 # HELP node_netstat_Udp_NoPorts Statistic UdpNoPorts. # TYPE node_netstat_Udp_NoPorts untyped node_netstat_Udp_NoPorts 4 # HELP node_netstat_Udp_OutDatagrams Statistic UdpOutDatagrams. # TYPE node_netstat_Udp_OutDatagrams untyped node_netstat_Udp_OutDatagrams 2787 # HELP node_netstat_Udp_RcvbufErrors Statistic UdpRcvbufErrors. # TYPE node_netstat_Udp_RcvbufErrors untyped node_netstat_Udp_RcvbufErrors 0 # HELP node_netstat_Udp_SndbufErrors Statistic UdpSndbufErrors. # TYPE node_netstat_Udp_SndbufErrors untyped node_netstat_Udp_SndbufErrors 0 # HELP node_network_address_assign_type Network device property: address_assign_type # TYPE node_network_address_assign_type gauge node_network_address_assign_type{device="azv01f90a66eae"} 3 node_network_address_assign_type{device="azv242b3abebdc"} 3 node_network_address_assign_type{device="azv4cd4fa71bae"} 3 node_network_address_assign_type{device="azv63a96d38340"} 3 node_network_address_assign_type{device="azv6c83e846b95"} 3 node_network_address_assign_type{device="azv8101e985a2a"} 3 node_network_address_assign_type{device="azv870dd5437ed"} 3 node_network_address_assign_type{device="azv99960e86340"} 3 node_network_address_assign_type{device="azva068c73f524"} 3 node_network_address_assign_type{device="azva34ff6c5426"} 3 node_network_address_assign_type{device="azvd81ad55895f"} 3 node_network_address_assign_type{device="azve117880b5fb"} 3 node_network_address_assign_type{device="azve4dcd2566f8"} 3 node_network_address_assign_type{device="azvf5d838218a5"} 3 node_network_address_assign_type{device="eth0"} 0 node_network_address_assign_type{device="eth1"} 0 node_network_address_assign_type{device="lo"} 0 # HELP node_network_carrier Network device property: carrier # TYPE node_network_carrier gauge node_network_carrier{device="azv01f90a66eae"} 1 node_network_carrier{device="azv242b3abebdc"} 1 node_network_carrier{device="azv4cd4fa71bae"} 1 node_network_carrier{device="azv63a96d38340"} 1 node_network_carrier{device="azv6c83e846b95"} 1 node_network_carrier{device="azv8101e985a2a"} 1 node_network_carrier{device="azv870dd5437ed"} 1 node_network_carrier{device="azv99960e86340"} 1 node_network_carrier{device="azva068c73f524"} 1 node_network_carrier{device="azva34ff6c5426"} 1 node_network_carrier{device="azvd81ad55895f"} 1 node_network_carrier{device="azve117880b5fb"} 1 node_network_carrier{device="azve4dcd2566f8"} 1 node_network_carrier{device="azvf5d838218a5"} 1 node_network_carrier{device="eth0"} 1 node_network_carrier{device="eth1"} 1 node_network_carrier{device="lo"} 1 # HELP node_network_carrier_changes_total Network device property: carrier_changes_total # TYPE node_network_carrier_changes_total counter node_network_carrier_changes_total{device="azv01f90a66eae"} 4 node_network_carrier_changes_total{device="azv242b3abebdc"} 4 node_network_carrier_changes_total{device="azv4cd4fa71bae"} 4 node_network_carrier_changes_total{device="azv63a96d38340"} 4 node_network_carrier_changes_total{device="azv6c83e846b95"} 4 node_network_carrier_changes_total{device="azv8101e985a2a"} 4 node_network_carrier_changes_total{device="azv870dd5437ed"} 4 node_network_carrier_changes_total{device="azv99960e86340"} 4 node_network_carrier_changes_total{device="azva068c73f524"} 4 node_network_carrier_changes_total{device="azva34ff6c5426"} 4 node_network_carrier_changes_total{device="azvd81ad55895f"} 4 node_network_carrier_changes_total{device="azve117880b5fb"} 4 node_network_carrier_changes_total{device="azve4dcd2566f8"} 4 node_network_carrier_changes_total{device="azvf5d838218a5"} 4 node_network_carrier_changes_total{device="eth0"} 3 node_network_carrier_changes_total{device="eth1"} 1 node_network_carrier_changes_total{device="lo"} 0 # HELP node_network_carrier_down_changes_total Network device property: carrier_down_changes_total # TYPE node_network_carrier_down_changes_total counter node_network_carrier_down_changes_total{device="azv01f90a66eae"} 2 node_network_carrier_down_changes_total{device="azv242b3abebdc"} 2 node_network_carrier_down_changes_total{device="azv4cd4fa71bae"} 2 node_network_carrier_down_changes_total{device="azv63a96d38340"} 2 node_network_carrier_down_changes_total{device="azv6c83e846b95"} 2 node_network_carrier_down_changes_total{device="azv8101e985a2a"} 2 node_network_carrier_down_changes_total{device="azv870dd5437ed"} 2 node_network_carrier_down_changes_total{device="azv99960e86340"} 2 node_network_carrier_down_changes_total{device="azva068c73f524"} 2 node_network_carrier_down_changes_total{device="azva34ff6c5426"} 2 node_network_carrier_down_changes_total{device="azvd81ad55895f"} 2 node_network_carrier_down_changes_total{device="azve117880b5fb"} 2 node_network_carrier_down_changes_total{device="azve4dcd2566f8"} 2 node_network_carrier_down_changes_total{device="azvf5d838218a5"} 2 node_network_carrier_down_changes_total{device="eth0"} 1 node_network_carrier_down_changes_total{device="eth1"} 0 node_network_carrier_down_changes_total{device="lo"} 0 # HELP node_network_carrier_up_changes_total Network device property: carrier_up_changes_total # TYPE node_network_carrier_up_changes_total counter node_network_carrier_up_changes_total{device="azv01f90a66eae"} 2 node_network_carrier_up_changes_total{device="azv242b3abebdc"} 2 node_network_carrier_up_changes_total{device="azv4cd4fa71bae"} 2 node_network_carrier_up_changes_total{device="azv63a96d38340"} 2 node_network_carrier_up_changes_total{device="azv6c83e846b95"} 2 node_network_carrier_up_changes_total{device="azv8101e985a2a"} 2 node_network_carrier_up_changes_total{device="azv870dd5437ed"} 2 node_network_carrier_up_changes_total{device="azv99960e86340"} 2 node_network_carrier_up_changes_total{device="azva068c73f524"} 2 node_network_carrier_up_changes_total{device="azva34ff6c5426"} 2 node_network_carrier_up_changes_total{device="azvd81ad55895f"} 2 node_network_carrier_up_changes_total{device="azve117880b5fb"} 2 node_network_carrier_up_changes_total{device="azve4dcd2566f8"} 2 node_network_carrier_up_changes_total{device="azvf5d838218a5"} 2 node_network_carrier_up_changes_total{device="eth0"} 2 node_network_carrier_up_changes_total{device="eth1"} 1 node_network_carrier_up_changes_total{device="lo"} 0 # HELP node_network_device_id Network device property: device_id # TYPE node_network_device_id gauge node_network_device_id{device="azv01f90a66eae"} 0 node_network_device_id{device="azv242b3abebdc"} 0 node_network_device_id{device="azv4cd4fa71bae"} 0 node_network_device_id{device="azv63a96d38340"} 0 node_network_device_id{device="azv6c83e846b95"} 0 node_network_device_id{device="azv8101e985a2a"} 0 node_network_device_id{device="azv870dd5437ed"} 0 node_network_device_id{device="azv99960e86340"} 0 node_network_device_id{device="azva068c73f524"} 0 node_network_device_id{device="azva34ff6c5426"} 0 node_network_device_id{device="azvd81ad55895f"} 0 node_network_device_id{device="azve117880b5fb"} 0 node_network_device_id{device="azve4dcd2566f8"} 0 node_network_device_id{device="azvf5d838218a5"} 0 node_network_device_id{device="eth0"} 0 node_network_device_id{device="eth1"} 0 node_network_device_id{device="lo"} 0 # HELP node_network_dormant Network device property: dormant # TYPE node_network_dormant gauge node_network_dormant{device="azv01f90a66eae"} 0 node_network_dormant{device="azv242b3abebdc"} 0 node_network_dormant{device="azv4cd4fa71bae"} 0 node_network_dormant{device="azv63a96d38340"} 0 node_network_dormant{device="azv6c83e846b95"} 0 node_network_dormant{device="azv8101e985a2a"} 0 node_network_dormant{device="azv870dd5437ed"} 0 node_network_dormant{device="azv99960e86340"} 0 node_network_dormant{device="azva068c73f524"} 0 node_network_dormant{device="azva34ff6c5426"} 0 node_network_dormant{device="azvd81ad55895f"} 0 node_network_dormant{device="azve117880b5fb"} 0 node_network_dormant{device="azve4dcd2566f8"} 0 node_network_dormant{device="azvf5d838218a5"} 0 node_network_dormant{device="eth0"} 0 node_network_dormant{device="eth1"} 0 node_network_dormant{device="lo"} 0 # HELP node_network_flags Network device property: flags # TYPE node_network_flags gauge node_network_flags{device="azv01f90a66eae"} 4099 node_network_flags{device="azv242b3abebdc"} 4099 node_network_flags{device="azv4cd4fa71bae"} 4099 node_network_flags{device="azv63a96d38340"} 4099 node_network_flags{device="azv6c83e846b95"} 4099 node_network_flags{device="azv8101e985a2a"} 4099 node_network_flags{device="azv870dd5437ed"} 4099 node_network_flags{device="azv99960e86340"} 4099 node_network_flags{device="azva068c73f524"} 4099 node_network_flags{device="azva34ff6c5426"} 4099 node_network_flags{device="azvd81ad55895f"} 4099 node_network_flags{device="azve117880b5fb"} 4099 node_network_flags{device="azve4dcd2566f8"} 4099 node_network_flags{device="azvf5d838218a5"} 4099 node_network_flags{device="eth0"} 4099 node_network_flags{device="eth1"} 6147 node_network_flags{device="lo"} 9 # HELP node_network_iface_id Network device property: iface_id # TYPE node_network_iface_id gauge node_network_iface_id{device="azv01f90a66eae"} 11 node_network_iface_id{device="azv242b3abebdc"} 7 node_network_iface_id{device="azv4cd4fa71bae"} 27 node_network_iface_id{device="azv63a96d38340"} 17 node_network_iface_id{device="azv6c83e846b95"} 35 node_network_iface_id{device="azv8101e985a2a"} 15 node_network_iface_id{device="azv870dd5437ed"} 21 node_network_iface_id{device="azv99960e86340"} 29 node_network_iface_id{device="azva068c73f524"} 31 node_network_iface_id{device="azva34ff6c5426"} 33 node_network_iface_id{device="azvd81ad55895f"} 5 node_network_iface_id{device="azve117880b5fb"} 9 node_network_iface_id{device="azve4dcd2566f8"} 13 node_network_iface_id{device="azvf5d838218a5"} 25 node_network_iface_id{device="eth0"} 2 node_network_iface_id{device="eth1"} 3 node_network_iface_id{device="lo"} 1 # HELP node_network_iface_link Network device property: iface_link # TYPE node_network_iface_link gauge node_network_iface_link{device="azv01f90a66eae"} 10 node_network_iface_link{device="azv242b3abebdc"} 6 node_network_iface_link{device="azv4cd4fa71bae"} 26 node_network_iface_link{device="azv63a96d38340"} 16 node_network_iface_link{device="azv6c83e846b95"} 34 node_network_iface_link{device="azv8101e985a2a"} 14 node_network_iface_link{device="azv870dd5437ed"} 20 node_network_iface_link{device="azv99960e86340"} 28 node_network_iface_link{device="azva068c73f524"} 30 node_network_iface_link{device="azva34ff6c5426"} 32 node_network_iface_link{device="azvd81ad55895f"} 4 node_network_iface_link{device="azve117880b5fb"} 8 node_network_iface_link{device="azve4dcd2566f8"} 12 node_network_iface_link{device="azvf5d838218a5"} 24 node_network_iface_link{device="eth0"} 2 node_network_iface_link{device="eth1"} 3 node_network_iface_link{device="lo"} 1 # HELP node_network_iface_link_mode Network device property: iface_link_mode # TYPE node_network_iface_link_mode gauge node_network_iface_link_mode{device="azv01f90a66eae"} 0 node_network_iface_link_mode{device="azv242b3abebdc"} 0 node_network_iface_link_mode{device="azv4cd4fa71bae"} 0 node_network_iface_link_mode{device="azv63a96d38340"} 0 node_network_iface_link_mode{device="azv6c83e846b95"} 0 node_network_iface_link_mode{device="azv8101e985a2a"} 0 node_network_iface_link_mode{device="azv870dd5437ed"} 0 node_network_iface_link_mode{device="azv99960e86340"} 0 node_network_iface_link_mode{device="azva068c73f524"} 0 node_network_iface_link_mode{device="azva34ff6c5426"} 0 node_network_iface_link_mode{device="azvd81ad55895f"} 0 node_network_iface_link_mode{device="azve117880b5fb"} 0 node_network_iface_link_mode{device="azve4dcd2566f8"} 0 node_network_iface_link_mode{device="azvf5d838218a5"} 0 node_network_iface_link_mode{device="eth0"} 0 node_network_iface_link_mode{device="eth1"} 0 node_network_iface_link_mode{device="lo"} 0 # HELP node_network_info Non-numeric data from /sys/class/net/, value is always 1. # TYPE node_network_info gauge node_network_info{address="00:00:00:00:00:00",adminstate="up",broadcast="00:00:00:00:00:00",device="lo",duplex="",ifalias="",operstate="unknown"} 1 node_network_info{address="7c:ed:8d:e5:37:8f",adminstate="up",broadcast="ff:ff:ff:ff:ff:ff",device="eth0",duplex="unknown",ifalias="",operstate="up"} 1 node_network_info{address="7c:ed:8d:e5:37:8f",adminstate="up",broadcast="ff:ff:ff:ff:ff:ff",device="eth1",duplex="unknown",ifalias="",operstate="up"} 1 node_network_info{address="aa:aa:aa:aa:aa:aa",adminstate="up",broadcast="ff:ff:ff:ff:ff:ff",device="azv01f90a66eae",duplex="full",ifalias="",operstate="up"} 1 node_network_info{address="aa:aa:aa:aa:aa:aa",adminstate="up",broadcast="ff:ff:ff:ff:ff:ff",device="azv242b3abebdc",duplex="full",ifalias="",operstate="up"} 1 node_network_info{address="aa:aa:aa:aa:aa:aa",adminstate="up",broadcast="ff:ff:ff:ff:ff:ff",device="azv4cd4fa71bae",duplex="full",ifalias="",operstate="up"} 1 node_network_info{address="aa:aa:aa:aa:aa:aa",adminstate="up",broadcast="ff:ff:ff:ff:ff:ff",device="azv63a96d38340",duplex="full",ifalias="",operstate="up"} 1 node_network_info{address="aa:aa:aa:aa:aa:aa",adminstate="up",broadcast="ff:ff:ff:ff:ff:ff",device="azv6c83e846b95",duplex="full",ifalias="",operstate="up"} 1 node_network_info{address="aa:aa:aa:aa:aa:aa",adminstate="up",broadcast="ff:ff:ff:ff:ff:ff",device="azv8101e985a2a",duplex="full",ifalias="",operstate="up"} 1 node_network_info{address="aa:aa:aa:aa:aa:aa",adminstate="up",broadcast="ff:ff:ff:ff:ff:ff",device="azv870dd5437ed",duplex="full",ifalias="",operstate="up"} 1 node_network_info{address="aa:aa:aa:aa:aa:aa",adminstate="up",broadcast="ff:ff:ff:ff:ff:ff",device="azv99960e86340",duplex="full",ifalias="",operstate="up"} 1 node_network_info{address="aa:aa:aa:aa:aa:aa",adminstate="up",broadcast="ff:ff:ff:ff:ff:ff",device="azva068c73f524",duplex="full",ifalias="",operstate="up"} 1 node_network_info{address="aa:aa:aa:aa:aa:aa",adminstate="up",broadcast="ff:ff:ff:ff:ff:ff",device="azva34ff6c5426",duplex="full",ifalias="",operstate="up"} 1 node_network_info{address="aa:aa:aa:aa:aa:aa",adminstate="up",broadcast="ff:ff:ff:ff:ff:ff",device="azvd81ad55895f",duplex="full",ifalias="",operstate="up"} 1 node_network_info{address="aa:aa:aa:aa:aa:aa",adminstate="up",broadcast="ff:ff:ff:ff:ff:ff",device="azve117880b5fb",duplex="full",ifalias="",operstate="up"} 1 node_network_info{address="aa:aa:aa:aa:aa:aa",adminstate="up",broadcast="ff:ff:ff:ff:ff:ff",device="azve4dcd2566f8",duplex="full",ifalias="",operstate="up"} 1 node_network_info{address="aa:aa:aa:aa:aa:aa",adminstate="up",broadcast="ff:ff:ff:ff:ff:ff",device="azvf5d838218a5",duplex="full",ifalias="",operstate="up"} 1 # HELP node_network_mtu_bytes Network device property: mtu_bytes # TYPE node_network_mtu_bytes gauge node_network_mtu_bytes{device="azv01f90a66eae"} 1500 node_network_mtu_bytes{device="azv242b3abebdc"} 1500 node_network_mtu_bytes{device="azv4cd4fa71bae"} 1500 node_network_mtu_bytes{device="azv63a96d38340"} 1500 node_network_mtu_bytes{device="azv6c83e846b95"} 1500 node_network_mtu_bytes{device="azv8101e985a2a"} 1500 node_network_mtu_bytes{device="azv870dd5437ed"} 1500 node_network_mtu_bytes{device="azv99960e86340"} 1500 node_network_mtu_bytes{device="azva068c73f524"} 1500 node_network_mtu_bytes{device="azva34ff6c5426"} 1500 node_network_mtu_bytes{device="azvd81ad55895f"} 1500 node_network_mtu_bytes{device="azve117880b5fb"} 1500 node_network_mtu_bytes{device="azve4dcd2566f8"} 1500 node_network_mtu_bytes{device="azvf5d838218a5"} 1500 node_network_mtu_bytes{device="eth0"} 1500 node_network_mtu_bytes{device="eth1"} 1500 node_network_mtu_bytes{device="lo"} 65536 # HELP node_network_name_assign_type Network device property: name_assign_type # TYPE node_network_name_assign_type gauge node_network_name_assign_type{device="azv01f90a66eae"} 3 node_network_name_assign_type{device="azv242b3abebdc"} 3 node_network_name_assign_type{device="azv4cd4fa71bae"} 3 node_network_name_assign_type{device="azv63a96d38340"} 3 node_network_name_assign_type{device="azv6c83e846b95"} 3 node_network_name_assign_type{device="azv8101e985a2a"} 3 node_network_name_assign_type{device="azv870dd5437ed"} 3 node_network_name_assign_type{device="azv99960e86340"} 3 node_network_name_assign_type{device="azva068c73f524"} 3 node_network_name_assign_type{device="azva34ff6c5426"} 3 node_network_name_assign_type{device="azvd81ad55895f"} 3 node_network_name_assign_type{device="azve117880b5fb"} 3 node_network_name_assign_type{device="azve4dcd2566f8"} 3 node_network_name_assign_type{device="azvf5d838218a5"} 3 node_network_name_assign_type{device="lo"} 2 # HELP node_network_net_dev_group Network device property: net_dev_group # TYPE node_network_net_dev_group gauge node_network_net_dev_group{device="azv01f90a66eae"} 0 node_network_net_dev_group{device="azv242b3abebdc"} 0 node_network_net_dev_group{device="azv4cd4fa71bae"} 0 node_network_net_dev_group{device="azv63a96d38340"} 0 node_network_net_dev_group{device="azv6c83e846b95"} 0 node_network_net_dev_group{device="azv8101e985a2a"} 0 node_network_net_dev_group{device="azv870dd5437ed"} 0 node_network_net_dev_group{device="azv99960e86340"} 0 node_network_net_dev_group{device="azva068c73f524"} 0 node_network_net_dev_group{device="azva34ff6c5426"} 0 node_network_net_dev_group{device="azvd81ad55895f"} 0 node_network_net_dev_group{device="azve117880b5fb"} 0 node_network_net_dev_group{device="azve4dcd2566f8"} 0 node_network_net_dev_group{device="azvf5d838218a5"} 0 node_network_net_dev_group{device="eth0"} 0 node_network_net_dev_group{device="eth1"} 0 node_network_net_dev_group{device="lo"} 0 # HELP node_network_protocol_type Network device property: protocol_type # TYPE node_network_protocol_type gauge node_network_protocol_type{device="azv01f90a66eae"} 1 node_network_protocol_type{device="azv242b3abebdc"} 1 node_network_protocol_type{device="azv4cd4fa71bae"} 1 node_network_protocol_type{device="azv63a96d38340"} 1 node_network_protocol_type{device="azv6c83e846b95"} 1 node_network_protocol_type{device="azv8101e985a2a"} 1 node_network_protocol_type{device="azv870dd5437ed"} 1 node_network_protocol_type{device="azv99960e86340"} 1 node_network_protocol_type{device="azva068c73f524"} 1 node_network_protocol_type{device="azva34ff6c5426"} 1 node_network_protocol_type{device="azvd81ad55895f"} 1 node_network_protocol_type{device="azve117880b5fb"} 1 node_network_protocol_type{device="azve4dcd2566f8"} 1 node_network_protocol_type{device="azvf5d838218a5"} 1 node_network_protocol_type{device="eth0"} 1 node_network_protocol_type{device="eth1"} 1 node_network_protocol_type{device="lo"} 772 # HELP node_network_receive_bytes_total Network device statistic receive_bytes. # TYPE node_network_receive_bytes_total counter node_network_receive_bytes_total{device="azv01f90a66eae"} 6.4588922e+07 node_network_receive_bytes_total{device="azv242b3abebdc"} 1.44519505e+08 node_network_receive_bytes_total{device="azv4cd4fa71bae"} 2.01814912e+08 node_network_receive_bytes_total{device="azv63a96d38340"} 1.50574746e+08 node_network_receive_bytes_total{device="azv6c83e846b95"} 1.17259671e+08 node_network_receive_bytes_total{device="azv8101e985a2a"} 9.79523936e+08 node_network_receive_bytes_total{device="azv870dd5437ed"} 1.59060153e+08 node_network_receive_bytes_total{device="azv99960e86340"} 5.155628e+07 node_network_receive_bytes_total{device="azva068c73f524"} 3.2510117e+08 node_network_receive_bytes_total{device="azva34ff6c5426"} 3.51944085e+08 node_network_receive_bytes_total{device="azvd81ad55895f"} 8.241535115e+09 node_network_receive_bytes_total{device="azve117880b5fb"} 2.341752e+06 node_network_receive_bytes_total{device="azve4dcd2566f8"} 1.15849143e+08 node_network_receive_bytes_total{device="azvf5d838218a5"} 1.4375446e+07 node_network_receive_bytes_total{device="eth0"} 1.8318014863e+10 node_network_receive_bytes_total{device="eth1"} 1.8796618633e+10 node_network_receive_bytes_total{device="lo"} 9.7130547e+07 # HELP node_network_receive_compressed_total Network device statistic receive_compressed. # TYPE node_network_receive_compressed_total counter node_network_receive_compressed_total{device="azv01f90a66eae"} 0 node_network_receive_compressed_total{device="azv242b3abebdc"} 0 node_network_receive_compressed_total{device="azv4cd4fa71bae"} 0 node_network_receive_compressed_total{device="azv63a96d38340"} 0 node_network_receive_compressed_total{device="azv6c83e846b95"} 0 node_network_receive_compressed_total{device="azv8101e985a2a"} 0 node_network_receive_compressed_total{device="azv870dd5437ed"} 0 node_network_receive_compressed_total{device="azv99960e86340"} 0 node_network_receive_compressed_total{device="azva068c73f524"} 0 node_network_receive_compressed_total{device="azva34ff6c5426"} 0 node_network_receive_compressed_total{device="azvd81ad55895f"} 0 node_network_receive_compressed_total{device="azve117880b5fb"} 0 node_network_receive_compressed_total{device="azve4dcd2566f8"} 0 node_network_receive_compressed_total{device="azvf5d838218a5"} 0 node_network_receive_compressed_total{device="eth0"} 0 node_network_receive_compressed_total{device="eth1"} 0 node_network_receive_compressed_total{device="lo"} 0 # HELP node_network_receive_drop_total Network device statistic receive_drop. # TYPE node_network_receive_drop_total counter node_network_receive_drop_total{device="azv01f90a66eae"} 0 node_network_receive_drop_total{device="azv242b3abebdc"} 0 node_network_receive_drop_total{device="azv4cd4fa71bae"} 0 node_network_receive_drop_total{device="azv63a96d38340"} 0 node_network_receive_drop_total{device="azv6c83e846b95"} 0 node_network_receive_drop_total{device="azv8101e985a2a"} 0 node_network_receive_drop_total{device="azv870dd5437ed"} 0 node_network_receive_drop_total{device="azv99960e86340"} 0 node_network_receive_drop_total{device="azva068c73f524"} 0 node_network_receive_drop_total{device="azva34ff6c5426"} 0 node_network_receive_drop_total{device="azvd81ad55895f"} 0 node_network_receive_drop_total{device="azve117880b5fb"} 0 node_network_receive_drop_total{device="azve4dcd2566f8"} 0 node_network_receive_drop_total{device="azvf5d838218a5"} 0 node_network_receive_drop_total{device="eth0"} 0 node_network_receive_drop_total{device="eth1"} 3009 node_network_receive_drop_total{device="lo"} 0 # HELP node_network_receive_errs_total Network device statistic receive_errs. # TYPE node_network_receive_errs_total counter node_network_receive_errs_total{device="azv01f90a66eae"} 0 node_network_receive_errs_total{device="azv242b3abebdc"} 0 node_network_receive_errs_total{device="azv4cd4fa71bae"} 0 node_network_receive_errs_total{device="azv63a96d38340"} 0 node_network_receive_errs_total{device="azv6c83e846b95"} 0 node_network_receive_errs_total{device="azv8101e985a2a"} 0 node_network_receive_errs_total{device="azv870dd5437ed"} 0 node_network_receive_errs_total{device="azv99960e86340"} 0 node_network_receive_errs_total{device="azva068c73f524"} 0 node_network_receive_errs_total{device="azva34ff6c5426"} 0 node_network_receive_errs_total{device="azvd81ad55895f"} 0 node_network_receive_errs_total{device="azve117880b5fb"} 0 node_network_receive_errs_total{device="azve4dcd2566f8"} 0 node_network_receive_errs_total{device="azvf5d838218a5"} 0 node_network_receive_errs_total{device="eth0"} 0 node_network_receive_errs_total{device="eth1"} 0 node_network_receive_errs_total{device="lo"} 0 # HELP node_network_receive_fifo_total Network device statistic receive_fifo. # TYPE node_network_receive_fifo_total counter node_network_receive_fifo_total{device="azv01f90a66eae"} 0 node_network_receive_fifo_total{device="azv242b3abebdc"} 0 node_network_receive_fifo_total{device="azv4cd4fa71bae"} 0 node_network_receive_fifo_total{device="azv63a96d38340"} 0 node_network_receive_fifo_total{device="azv6c83e846b95"} 0 node_network_receive_fifo_total{device="azv8101e985a2a"} 0 node_network_receive_fifo_total{device="azv870dd5437ed"} 0 node_network_receive_fifo_total{device="azv99960e86340"} 0 node_network_receive_fifo_total{device="azva068c73f524"} 0 node_network_receive_fifo_total{device="azva34ff6c5426"} 0 node_network_receive_fifo_total{device="azvd81ad55895f"} 0 node_network_receive_fifo_total{device="azve117880b5fb"} 0 node_network_receive_fifo_total{device="azve4dcd2566f8"} 0 node_network_receive_fifo_total{device="azvf5d838218a5"} 0 node_network_receive_fifo_total{device="eth0"} 0 node_network_receive_fifo_total{device="eth1"} 0 node_network_receive_fifo_total{device="lo"} 0 # HELP node_network_receive_frame_total Network device statistic receive_frame. # TYPE node_network_receive_frame_total counter node_network_receive_frame_total{device="azv01f90a66eae"} 0 node_network_receive_frame_total{device="azv242b3abebdc"} 0 node_network_receive_frame_total{device="azv4cd4fa71bae"} 0 node_network_receive_frame_total{device="azv63a96d38340"} 0 node_network_receive_frame_total{device="azv6c83e846b95"} 0 node_network_receive_frame_total{device="azv8101e985a2a"} 0 node_network_receive_frame_total{device="azv870dd5437ed"} 0 node_network_receive_frame_total{device="azv99960e86340"} 0 node_network_receive_frame_total{device="azva068c73f524"} 0 node_network_receive_frame_total{device="azva34ff6c5426"} 0 node_network_receive_frame_total{device="azvd81ad55895f"} 0 node_network_receive_frame_total{device="azve117880b5fb"} 0 node_network_receive_frame_total{device="azve4dcd2566f8"} 0 node_network_receive_frame_total{device="azvf5d838218a5"} 0 node_network_receive_frame_total{device="eth0"} 0 node_network_receive_frame_total{device="eth1"} 0 node_network_receive_frame_total{device="lo"} 0 # HELP node_network_receive_multicast_total Network device statistic receive_multicast. # TYPE node_network_receive_multicast_total counter node_network_receive_multicast_total{device="azv01f90a66eae"} 0 node_network_receive_multicast_total{device="azv242b3abebdc"} 0 node_network_receive_multicast_total{device="azv4cd4fa71bae"} 0 node_network_receive_multicast_total{device="azv63a96d38340"} 0 node_network_receive_multicast_total{device="azv6c83e846b95"} 0 node_network_receive_multicast_total{device="azv8101e985a2a"} 0 node_network_receive_multicast_total{device="azv870dd5437ed"} 0 node_network_receive_multicast_total{device="azv99960e86340"} 0 node_network_receive_multicast_total{device="azva068c73f524"} 0 node_network_receive_multicast_total{device="azva34ff6c5426"} 0 node_network_receive_multicast_total{device="azvd81ad55895f"} 0 node_network_receive_multicast_total{device="azve117880b5fb"} 0 node_network_receive_multicast_total{device="azve4dcd2566f8"} 0 node_network_receive_multicast_total{device="azvf5d838218a5"} 0 node_network_receive_multicast_total{device="eth0"} 0 node_network_receive_multicast_total{device="eth1"} 0 node_network_receive_multicast_total{device="lo"} 0 # HELP node_network_receive_nohandler_total Network device statistic receive_nohandler. # TYPE node_network_receive_nohandler_total counter node_network_receive_nohandler_total{device="azv01f90a66eae"} 0 node_network_receive_nohandler_total{device="azv242b3abebdc"} 0 node_network_receive_nohandler_total{device="azv4cd4fa71bae"} 0 node_network_receive_nohandler_total{device="azv63a96d38340"} 0 node_network_receive_nohandler_total{device="azv6c83e846b95"} 0 node_network_receive_nohandler_total{device="azv8101e985a2a"} 0 node_network_receive_nohandler_total{device="azv870dd5437ed"} 0 node_network_receive_nohandler_total{device="azv99960e86340"} 0 node_network_receive_nohandler_total{device="azva068c73f524"} 0 node_network_receive_nohandler_total{device="azva34ff6c5426"} 0 node_network_receive_nohandler_total{device="azvd81ad55895f"} 0 node_network_receive_nohandler_total{device="azve117880b5fb"} 0 node_network_receive_nohandler_total{device="azve4dcd2566f8"} 0 node_network_receive_nohandler_total{device="azvf5d838218a5"} 0 node_network_receive_nohandler_total{device="eth0"} 0 node_network_receive_nohandler_total{device="eth1"} 0 node_network_receive_nohandler_total{device="lo"} 0 # HELP node_network_receive_packets_total Network device statistic receive_packets. # TYPE node_network_receive_packets_total counter node_network_receive_packets_total{device="azv01f90a66eae"} 99973 node_network_receive_packets_total{device="azv242b3abebdc"} 356668 node_network_receive_packets_total{device="azv4cd4fa71bae"} 877398 node_network_receive_packets_total{device="azv63a96d38340"} 651660 node_network_receive_packets_total{device="azv6c83e846b95"} 421805 node_network_receive_packets_total{device="azv8101e985a2a"} 1.582387e+06 node_network_receive_packets_total{device="azv870dd5437ed"} 659575 node_network_receive_packets_total{device="azv99960e86340"} 481841 node_network_receive_packets_total{device="azva068c73f524"} 181998 node_network_receive_packets_total{device="azva34ff6c5426"} 1.562265e+06 node_network_receive_packets_total{device="azvd81ad55895f"} 2.442602e+06 node_network_receive_packets_total{device="azve117880b5fb"} 24187 node_network_receive_packets_total{device="azve4dcd2566f8"} 345408 node_network_receive_packets_total{device="azvf5d838218a5"} 161775 node_network_receive_packets_total{device="eth0"} 1.0827634e+07 node_network_receive_packets_total{device="eth1"} 1.8823779e+07 node_network_receive_packets_total{device="lo"} 855160 # HELP node_network_speed_bytes Network device property: speed_bytes # TYPE node_network_speed_bytes gauge node_network_speed_bytes{device="azv01f90a66eae"} 1.25e+09 node_network_speed_bytes{device="azv242b3abebdc"} 1.25e+09 node_network_speed_bytes{device="azv4cd4fa71bae"} 1.25e+09 node_network_speed_bytes{device="azv63a96d38340"} 1.25e+09 node_network_speed_bytes{device="azv6c83e846b95"} 1.25e+09 node_network_speed_bytes{device="azv8101e985a2a"} 1.25e+09 node_network_speed_bytes{device="azv870dd5437ed"} 1.25e+09 node_network_speed_bytes{device="azv99960e86340"} 1.25e+09 node_network_speed_bytes{device="azva068c73f524"} 1.25e+09 node_network_speed_bytes{device="azva34ff6c5426"} 1.25e+09 node_network_speed_bytes{device="azvd81ad55895f"} 1.25e+09 node_network_speed_bytes{device="azve117880b5fb"} 1.25e+09 node_network_speed_bytes{device="azve4dcd2566f8"} 1.25e+09 node_network_speed_bytes{device="azvf5d838218a5"} 1.25e+09 node_network_speed_bytes{device="eth0"} 6.25e+09 node_network_speed_bytes{device="eth1"} 6.25e+09 # HELP node_network_transmit_bytes_total Network device statistic transmit_bytes. # TYPE node_network_transmit_bytes_total counter node_network_transmit_bytes_total{device="azv01f90a66eae"} 9.981301e+06 node_network_transmit_bytes_total{device="azv242b3abebdc"} 1.17744174e+08 node_network_transmit_bytes_total{device="azv4cd4fa71bae"} 1.06893505e+08 node_network_transmit_bytes_total{device="azv63a96d38340"} 8.7989996e+07 node_network_transmit_bytes_total{device="azv6c83e846b95"} 1.44371895e+08 node_network_transmit_bytes_total{device="azv8101e985a2a"} 9.77334436e+08 node_network_transmit_bytes_total{device="azv870dd5437ed"} 2.964813201e+09 node_network_transmit_bytes_total{device="azv99960e86340"} 1.15643511e+08 node_network_transmit_bytes_total{device="azva068c73f524"} 2.0371001e+07 node_network_transmit_bytes_total{device="azva34ff6c5426"} 2.70645952e+09 node_network_transmit_bytes_total{device="azvd81ad55895f"} 8.846891749e+09 node_network_transmit_bytes_total{device="azve117880b5fb"} 2.752176e+06 node_network_transmit_bytes_total{device="azve4dcd2566f8"} 5.50066665e+08 node_network_transmit_bytes_total{device="azvf5d838218a5"} 1.6712142e+07 node_network_transmit_bytes_total{device="eth0"} 1.2578262774e+10 node_network_transmit_bytes_total{device="eth1"} 1.2943416456e+10 node_network_transmit_bytes_total{device="lo"} 9.7130547e+07 # HELP node_network_transmit_carrier_total Network device statistic transmit_carrier. # TYPE node_network_transmit_carrier_total counter node_network_transmit_carrier_total{device="azv01f90a66eae"} 0 node_network_transmit_carrier_total{device="azv242b3abebdc"} 0 node_network_transmit_carrier_total{device="azv4cd4fa71bae"} 0 node_network_transmit_carrier_total{device="azv63a96d38340"} 0 node_network_transmit_carrier_total{device="azv6c83e846b95"} 0 node_network_transmit_carrier_total{device="azv8101e985a2a"} 0 node_network_transmit_carrier_total{device="azv870dd5437ed"} 0 node_network_transmit_carrier_total{device="azv99960e86340"} 0 node_network_transmit_carrier_total{device="azva068c73f524"} 0 node_network_transmit_carrier_total{device="azva34ff6c5426"} 0 node_network_transmit_carrier_total{device="azvd81ad55895f"} 0 node_network_transmit_carrier_total{device="azve117880b5fb"} 0 node_network_transmit_carrier_total{device="azve4dcd2566f8"} 0 node_network_transmit_carrier_total{device="azvf5d838218a5"} 0 node_network_transmit_carrier_total{device="eth0"} 0 node_network_transmit_carrier_total{device="eth1"} 0 node_network_transmit_carrier_total{device="lo"} 0 # HELP node_network_transmit_colls_total Network device statistic transmit_colls. # TYPE node_network_transmit_colls_total counter node_network_transmit_colls_total{device="azv01f90a66eae"} 0 node_network_transmit_colls_total{device="azv242b3abebdc"} 0 node_network_transmit_colls_total{device="azv4cd4fa71bae"} 0 node_network_transmit_colls_total{device="azv63a96d38340"} 0 node_network_transmit_colls_total{device="azv6c83e846b95"} 0 node_network_transmit_colls_total{device="azv8101e985a2a"} 0 node_network_transmit_colls_total{device="azv870dd5437ed"} 0 node_network_transmit_colls_total{device="azv99960e86340"} 0 node_network_transmit_colls_total{device="azva068c73f524"} 0 node_network_transmit_colls_total{device="azva34ff6c5426"} 0 node_network_transmit_colls_total{device="azvd81ad55895f"} 0 node_network_transmit_colls_total{device="azve117880b5fb"} 0 node_network_transmit_colls_total{device="azve4dcd2566f8"} 0 node_network_transmit_colls_total{device="azvf5d838218a5"} 0 node_network_transmit_colls_total{device="eth0"} 0 node_network_transmit_colls_total{device="eth1"} 0 node_network_transmit_colls_total{device="lo"} 0 # HELP node_network_transmit_compressed_total Network device statistic transmit_compressed. # TYPE node_network_transmit_compressed_total counter node_network_transmit_compressed_total{device="azv01f90a66eae"} 0 node_network_transmit_compressed_total{device="azv242b3abebdc"} 0 node_network_transmit_compressed_total{device="azv4cd4fa71bae"} 0 node_network_transmit_compressed_total{device="azv63a96d38340"} 0 node_network_transmit_compressed_total{device="azv6c83e846b95"} 0 node_network_transmit_compressed_total{device="azv8101e985a2a"} 0 node_network_transmit_compressed_total{device="azv870dd5437ed"} 0 node_network_transmit_compressed_total{device="azv99960e86340"} 0 node_network_transmit_compressed_total{device="azva068c73f524"} 0 node_network_transmit_compressed_total{device="azva34ff6c5426"} 0 node_network_transmit_compressed_total{device="azvd81ad55895f"} 0 node_network_transmit_compressed_total{device="azve117880b5fb"} 0 node_network_transmit_compressed_total{device="azve4dcd2566f8"} 0 node_network_transmit_compressed_total{device="azvf5d838218a5"} 0 node_network_transmit_compressed_total{device="eth0"} 0 node_network_transmit_compressed_total{device="eth1"} 0 node_network_transmit_compressed_total{device="lo"} 0 # HELP node_network_transmit_drop_total Network device statistic transmit_drop. # TYPE node_network_transmit_drop_total counter node_network_transmit_drop_total{device="azv01f90a66eae"} 0 node_network_transmit_drop_total{device="azv242b3abebdc"} 0 node_network_transmit_drop_total{device="azv4cd4fa71bae"} 0 node_network_transmit_drop_total{device="azv63a96d38340"} 0 node_network_transmit_drop_total{device="azv6c83e846b95"} 0 node_network_transmit_drop_total{device="azv8101e985a2a"} 0 node_network_transmit_drop_total{device="azv870dd5437ed"} 0 node_network_transmit_drop_total{device="azv99960e86340"} 0 node_network_transmit_drop_total{device="azva068c73f524"} 0 node_network_transmit_drop_total{device="azva34ff6c5426"} 0 node_network_transmit_drop_total{device="azvd81ad55895f"} 0 node_network_transmit_drop_total{device="azve117880b5fb"} 0 node_network_transmit_drop_total{device="azve4dcd2566f8"} 0 node_network_transmit_drop_total{device="azvf5d838218a5"} 0 node_network_transmit_drop_total{device="eth0"} 0 node_network_transmit_drop_total{device="eth1"} 0 node_network_transmit_drop_total{device="lo"} 0 # HELP node_network_transmit_errs_total Network device statistic transmit_errs. # TYPE node_network_transmit_errs_total counter node_network_transmit_errs_total{device="azv01f90a66eae"} 0 node_network_transmit_errs_total{device="azv242b3abebdc"} 0 node_network_transmit_errs_total{device="azv4cd4fa71bae"} 0 node_network_transmit_errs_total{device="azv63a96d38340"} 0 node_network_transmit_errs_total{device="azv6c83e846b95"} 0 node_network_transmit_errs_total{device="azv8101e985a2a"} 0 node_network_transmit_errs_total{device="azv870dd5437ed"} 0 node_network_transmit_errs_total{device="azv99960e86340"} 0 node_network_transmit_errs_total{device="azva068c73f524"} 0 node_network_transmit_errs_total{device="azva34ff6c5426"} 0 node_network_transmit_errs_total{device="azvd81ad55895f"} 0 node_network_transmit_errs_total{device="azve117880b5fb"} 0 node_network_transmit_errs_total{device="azve4dcd2566f8"} 0 node_network_transmit_errs_total{device="azvf5d838218a5"} 0 node_network_transmit_errs_total{device="eth0"} 0 node_network_transmit_errs_total{device="eth1"} 0 node_network_transmit_errs_total{device="lo"} 0 # HELP node_network_transmit_fifo_total Network device statistic transmit_fifo. # TYPE node_network_transmit_fifo_total counter node_network_transmit_fifo_total{device="azv01f90a66eae"} 0 node_network_transmit_fifo_total{device="azv242b3abebdc"} 0 node_network_transmit_fifo_total{device="azv4cd4fa71bae"} 0 node_network_transmit_fifo_total{device="azv63a96d38340"} 0 node_network_transmit_fifo_total{device="azv6c83e846b95"} 0 node_network_transmit_fifo_total{device="azv8101e985a2a"} 0 node_network_transmit_fifo_total{device="azv870dd5437ed"} 0 node_network_transmit_fifo_total{device="azv99960e86340"} 0 node_network_transmit_fifo_total{device="azva068c73f524"} 0 node_network_transmit_fifo_total{device="azva34ff6c5426"} 0 node_network_transmit_fifo_total{device="azvd81ad55895f"} 0 node_network_transmit_fifo_total{device="azve117880b5fb"} 0 node_network_transmit_fifo_total{device="azve4dcd2566f8"} 0 node_network_transmit_fifo_total{device="azvf5d838218a5"} 0 node_network_transmit_fifo_total{device="eth0"} 0 node_network_transmit_fifo_total{device="eth1"} 0 node_network_transmit_fifo_total{device="lo"} 0 # HELP node_network_transmit_packets_total Network device statistic transmit_packets. # TYPE node_network_transmit_packets_total counter node_network_transmit_packets_total{device="azv01f90a66eae"} 74793 node_network_transmit_packets_total{device="azv242b3abebdc"} 314198 node_network_transmit_packets_total{device="azv4cd4fa71bae"} 907662 node_network_transmit_packets_total{device="azv63a96d38340"} 658286 node_network_transmit_packets_total{device="azv6c83e846b95"} 434618 node_network_transmit_packets_total{device="azv8101e985a2a"} 1.622302e+06 node_network_transmit_packets_total{device="azv870dd5437ed"} 681820 node_network_transmit_packets_total{device="azv99960e86340"} 469777 node_network_transmit_packets_total{device="azva068c73f524"} 255002 node_network_transmit_packets_total{device="azva34ff6c5426"} 2.056658e+06 node_network_transmit_packets_total{device="azvd81ad55895f"} 3.287265e+06 node_network_transmit_packets_total{device="azve117880b5fb"} 26827 node_network_transmit_packets_total{device="azve4dcd2566f8"} 253425 node_network_transmit_packets_total{device="azvf5d838218a5"} 243503 node_network_transmit_packets_total{device="eth0"} 9.351293e+06 node_network_transmit_packets_total{device="eth1"} 1.4869808e+07 node_network_transmit_packets_total{device="lo"} 855160 # HELP node_network_transmit_queue_length Network device property: transmit_queue_length # TYPE node_network_transmit_queue_length gauge node_network_transmit_queue_length{device="azv01f90a66eae"} 1000 node_network_transmit_queue_length{device="azv242b3abebdc"} 1000 node_network_transmit_queue_length{device="azv4cd4fa71bae"} 1000 node_network_transmit_queue_length{device="azv63a96d38340"} 1000 node_network_transmit_queue_length{device="azv6c83e846b95"} 1000 node_network_transmit_queue_length{device="azv8101e985a2a"} 1000 node_network_transmit_queue_length{device="azv870dd5437ed"} 1000 node_network_transmit_queue_length{device="azv99960e86340"} 1000 node_network_transmit_queue_length{device="azva068c73f524"} 1000 node_network_transmit_queue_length{device="azva34ff6c5426"} 1000 node_network_transmit_queue_length{device="azvd81ad55895f"} 1000 node_network_transmit_queue_length{device="azve117880b5fb"} 1000 node_network_transmit_queue_length{device="azve4dcd2566f8"} 1000 node_network_transmit_queue_length{device="azvf5d838218a5"} 1000 node_network_transmit_queue_length{device="eth0"} 1000 node_network_transmit_queue_length{device="eth1"} 1000 node_network_transmit_queue_length{device="lo"} 1000 # HELP node_network_up Value is 1 if operstate is 'up', 0 otherwise. # TYPE node_network_up gauge node_network_up{device="azv01f90a66eae"} 1 node_network_up{device="azv242b3abebdc"} 1 node_network_up{device="azv4cd4fa71bae"} 1 node_network_up{device="azv63a96d38340"} 1 node_network_up{device="azv6c83e846b95"} 1 node_network_up{device="azv8101e985a2a"} 1 node_network_up{device="azv870dd5437ed"} 1 node_network_up{device="azv99960e86340"} 1 node_network_up{device="azva068c73f524"} 1 node_network_up{device="azva34ff6c5426"} 1 node_network_up{device="azvd81ad55895f"} 1 node_network_up{device="azve117880b5fb"} 1 node_network_up{device="azve4dcd2566f8"} 1 node_network_up{device="azvf5d838218a5"} 1 node_network_up{device="eth0"} 1 node_network_up{device="eth1"} 1 node_network_up{device="lo"} 0 # HELP node_nf_conntrack_entries Number of currently allocated flow entries for connection tracking. # TYPE node_nf_conntrack_entries gauge node_nf_conntrack_entries 1964 # HELP node_nf_conntrack_entries_limit Maximum size of connection tracking table. # TYPE node_nf_conntrack_entries_limit gauge node_nf_conntrack_entries_limit 131072 # HELP node_os_info A metric with a constant '1' value labeled by build_id, id, id_like, image_id, image_version, name, pretty_name, variant, variant_id, version, version_codename, version_id. # TYPE node_os_info gauge node_os_info{build_id="",id="ubuntu",id_like="debian",image_id="",image_version="",name="Ubuntu",pretty_name="Ubuntu 22.04.5 LTS",variant="",variant_id="",version="22.04.5 LTS (Jammy Jellyfish)",version_codename="jammy",version_id="22.04"} 1 # HELP node_os_version Metric containing the major.minor part of the OS version. # TYPE node_os_version gauge node_os_version{id="ubuntu",id_like="debian",name="Ubuntu"} 22.04 # HELP node_pressure_cpu_waiting_seconds_total Total time in seconds that processes have waited for CPU time # TYPE node_pressure_cpu_waiting_seconds_total counter node_pressure_cpu_waiting_seconds_total 3080.8368029999997 # HELP node_pressure_io_stalled_seconds_total Total time in seconds no process could make progress due to IO congestion # TYPE node_pressure_io_stalled_seconds_total counter node_pressure_io_stalled_seconds_total 127.90378999999999 # HELP node_pressure_io_waiting_seconds_total Total time in seconds that processes have waited due to IO congestion # TYPE node_pressure_io_waiting_seconds_total counter node_pressure_io_waiting_seconds_total 196.762223 # HELP node_pressure_memory_stalled_seconds_total Total time in seconds no process could make progress due to memory congestion # TYPE node_pressure_memory_stalled_seconds_total counter node_pressure_memory_stalled_seconds_total 3.277221 # HELP node_pressure_memory_waiting_seconds_total Total time in seconds that processes have waited for memory # TYPE node_pressure_memory_waiting_seconds_total counter node_pressure_memory_waiting_seconds_total 4.4363339999999996 # HELP node_procs_blocked Number of processes blocked waiting for I/O to complete. # TYPE node_procs_blocked gauge node_procs_blocked 0 # HELP node_procs_running Number of processes in runnable state. # TYPE node_procs_running gauge node_procs_running 1 # HELP node_schedstat_running_seconds_total Number of seconds CPU spent running a process. # TYPE node_schedstat_running_seconds_total counter node_schedstat_running_seconds_total{cpu="0"} 8016.108237726 node_schedstat_running_seconds_total{cpu="1"} 8061.505187457 # HELP node_schedstat_timeslices_total Number of timeslices executed by CPU. # TYPE node_schedstat_timeslices_total counter node_schedstat_timeslices_total{cpu="0"} 1.20275984e+08 node_schedstat_timeslices_total{cpu="1"} 1.20081956e+08 # HELP node_schedstat_waiting_seconds_total Number of seconds spent by processing waiting for this CPU. # TYPE node_schedstat_waiting_seconds_total counter node_schedstat_waiting_seconds_total{cpu="0"} 6302.713207479 node_schedstat_waiting_seconds_total{cpu="1"} 6310.960472754 # HELP node_scrape_collector_duration_seconds node_exporter: Duration of a collector scrape. # TYPE node_scrape_collector_duration_seconds gauge node_scrape_collector_duration_seconds{collector="arp"} 0.000693775 node_scrape_collector_duration_seconds{collector="bcache"} 2.0708e-05 node_scrape_collector_duration_seconds{collector="bcachefs"} 7.514e-06 node_scrape_collector_duration_seconds{collector="bonding"} 8.847e-06 node_scrape_collector_duration_seconds{collector="btrfs"} 0.002392536 node_scrape_collector_duration_seconds{collector="conntrack"} 8.6241e-05 node_scrape_collector_duration_seconds{collector="cpu"} 0.000282747 node_scrape_collector_duration_seconds{collector="cpufreq"} 7.5932e-05 node_scrape_collector_duration_seconds{collector="diskstats"} 0.003267778 node_scrape_collector_duration_seconds{collector="dmi"} 1.0409e-05 node_scrape_collector_duration_seconds{collector="edac"} 2.8965e-05 node_scrape_collector_duration_seconds{collector="entropy"} 0.000113001 node_scrape_collector_duration_seconds{collector="fibrechannel"} 1.3515e-05 node_scrape_collector_duration_seconds{collector="filefd"} 3.2621e-05 node_scrape_collector_duration_seconds{collector="filesystem"} 0.034160905 node_scrape_collector_duration_seconds{collector="hwmon"} 1.5048e-05 node_scrape_collector_duration_seconds{collector="infiniband"} 0.004673652 node_scrape_collector_duration_seconds{collector="ipvs"} 1.2854e-05 node_scrape_collector_duration_seconds{collector="kernel_hung"} 1.3666e-05 node_scrape_collector_duration_seconds{collector="loadavg"} 2.6639e-05 node_scrape_collector_duration_seconds{collector="mdadm"} 7.2826e-05 node_scrape_collector_duration_seconds{collector="meminfo"} 0.000130123 node_scrape_collector_duration_seconds{collector="netclass"} 0.014407308 node_scrape_collector_duration_seconds{collector="netdev"} 0.000462313 node_scrape_collector_duration_seconds{collector="netstat"} 0.000591002 node_scrape_collector_duration_seconds{collector="nfs"} 2.2382e-05 node_scrape_collector_duration_seconds{collector="nfsd"} 1.9336e-05 node_scrape_collector_duration_seconds{collector="nvme"} 2.2321e-05 node_scrape_collector_duration_seconds{collector="os"} 4.0345e-05 node_scrape_collector_duration_seconds{collector="powersupplyclass"} 2.5287e-05 node_scrape_collector_duration_seconds{collector="pressure"} 0.000106679 node_scrape_collector_duration_seconds{collector="rapl"} 1.4386e-05 node_scrape_collector_duration_seconds{collector="schedstat"} 3.772e-05 node_scrape_collector_duration_seconds{collector="selinux"} 9.02e-07 node_scrape_collector_duration_seconds{collector="sockstat"} 0.000114133 node_scrape_collector_duration_seconds{collector="softnet"} 4.0536e-05 node_scrape_collector_duration_seconds{collector="stat"} 0.000122238 node_scrape_collector_duration_seconds{collector="tapestats"} 9.257e-06 node_scrape_collector_duration_seconds{collector="textfile"} 1.2283e-05 node_scrape_collector_duration_seconds{collector="thermal_zone"} 0.000205473 node_scrape_collector_duration_seconds{collector="time"} 0.000118371 node_scrape_collector_duration_seconds{collector="timex"} 1.1933e-05 node_scrape_collector_duration_seconds{collector="udp_queues"} 0.000106468 node_scrape_collector_duration_seconds{collector="uname"} 1.4236e-05 node_scrape_collector_duration_seconds{collector="vmstat"} 0.000184454 node_scrape_collector_duration_seconds{collector="watchdog"} 1.5458e-05 node_scrape_collector_duration_seconds{collector="xfs"} 8.827e-06 node_scrape_collector_duration_seconds{collector="zfs"} 1.1612e-05 # HELP node_scrape_collector_success node_exporter: Whether a collector succeeded. # TYPE node_scrape_collector_success gauge node_scrape_collector_success{collector="arp"} 1 node_scrape_collector_success{collector="bcache"} 1 node_scrape_collector_success{collector="bcachefs"} 0 node_scrape_collector_success{collector="bonding"} 0 node_scrape_collector_success{collector="btrfs"} 1 node_scrape_collector_success{collector="conntrack"} 0 node_scrape_collector_success{collector="cpu"} 1 node_scrape_collector_success{collector="cpufreq"} 1 node_scrape_collector_success{collector="diskstats"} 1 node_scrape_collector_success{collector="dmi"} 1 node_scrape_collector_success{collector="edac"} 1 node_scrape_collector_success{collector="entropy"} 1 node_scrape_collector_success{collector="fibrechannel"} 1 node_scrape_collector_success{collector="filefd"} 1 node_scrape_collector_success{collector="filesystem"} 1 node_scrape_collector_success{collector="hwmon"} 1 node_scrape_collector_success{collector="infiniband"} 1 node_scrape_collector_success{collector="ipvs"} 0 node_scrape_collector_success{collector="kernel_hung"} 0 node_scrape_collector_success{collector="loadavg"} 1 node_scrape_collector_success{collector="mdadm"} 1 node_scrape_collector_success{collector="meminfo"} 1 node_scrape_collector_success{collector="netclass"} 1 node_scrape_collector_success{collector="netdev"} 1 node_scrape_collector_success{collector="netstat"} 1 node_scrape_collector_success{collector="nfs"} 0 node_scrape_collector_success{collector="nfsd"} 0 node_scrape_collector_success{collector="nvme"} 1 node_scrape_collector_success{collector="os"} 1 node_scrape_collector_success{collector="powersupplyclass"} 1 node_scrape_collector_success{collector="pressure"} 1 node_scrape_collector_success{collector="rapl"} 0 node_scrape_collector_success{collector="schedstat"} 1 node_scrape_collector_success{collector="selinux"} 1 node_scrape_collector_success{collector="sockstat"} 1 node_scrape_collector_success{collector="softnet"} 1 node_scrape_collector_success{collector="stat"} 1 node_scrape_collector_success{collector="tapestats"} 0 node_scrape_collector_success{collector="textfile"} 1 node_scrape_collector_success{collector="thermal_zone"} 1 node_scrape_collector_success{collector="time"} 1 node_scrape_collector_success{collector="timex"} 1 node_scrape_collector_success{collector="udp_queues"} 1 node_scrape_collector_success{collector="uname"} 1 node_scrape_collector_success{collector="vmstat"} 1 node_scrape_collector_success{collector="watchdog"} 1 node_scrape_collector_success{collector="xfs"} 1 node_scrape_collector_success{collector="zfs"} 0 # HELP node_selinux_enabled SELinux is enabled, 1 is true, 0 is false # TYPE node_selinux_enabled gauge node_selinux_enabled 0 # HELP node_sockstat_FRAG6_inuse Number of FRAG6 sockets in state inuse. # TYPE node_sockstat_FRAG6_inuse gauge node_sockstat_FRAG6_inuse 0 # HELP node_sockstat_FRAG6_memory Number of FRAG6 sockets in state memory. # TYPE node_sockstat_FRAG6_memory gauge node_sockstat_FRAG6_memory 0 # HELP node_sockstat_FRAG_inuse Number of FRAG sockets in state inuse. # TYPE node_sockstat_FRAG_inuse gauge node_sockstat_FRAG_inuse 0 # HELP node_sockstat_FRAG_memory Number of FRAG sockets in state memory. # TYPE node_sockstat_FRAG_memory gauge node_sockstat_FRAG_memory 0 # HELP node_sockstat_RAW6_inuse Number of RAW6 sockets in state inuse. # TYPE node_sockstat_RAW6_inuse gauge node_sockstat_RAW6_inuse 1 # HELP node_sockstat_RAW_inuse Number of RAW sockets in state inuse. # TYPE node_sockstat_RAW_inuse gauge node_sockstat_RAW_inuse 0 # HELP node_sockstat_TCP6_inuse Number of TCP6 sockets in state inuse. # TYPE node_sockstat_TCP6_inuse gauge node_sockstat_TCP6_inuse 20 # HELP node_sockstat_TCP_alloc Number of TCP sockets in state alloc. # TYPE node_sockstat_TCP_alloc gauge node_sockstat_TCP_alloc 592 # HELP node_sockstat_TCP_inuse Number of TCP sockets in state inuse. # TYPE node_sockstat_TCP_inuse gauge node_sockstat_TCP_inuse 14 # HELP node_sockstat_TCP_mem Number of TCP sockets in state mem. # TYPE node_sockstat_TCP_mem gauge node_sockstat_TCP_mem 16 # HELP node_sockstat_TCP_mem_bytes Number of TCP sockets in state mem_bytes. # TYPE node_sockstat_TCP_mem_bytes gauge node_sockstat_TCP_mem_bytes 65536 # HELP node_sockstat_TCP_orphan Number of TCP sockets in state orphan. # TYPE node_sockstat_TCP_orphan gauge node_sockstat_TCP_orphan 0 # HELP node_sockstat_TCP_tw Number of TCP sockets in state tw. # TYPE node_sockstat_TCP_tw gauge node_sockstat_TCP_tw 214 # HELP node_sockstat_UDP6_inuse Number of UDP6 sockets in state inuse. # TYPE node_sockstat_UDP6_inuse gauge node_sockstat_UDP6_inuse 2 # HELP node_sockstat_UDPLITE6_inuse Number of UDPLITE6 sockets in state inuse. # TYPE node_sockstat_UDPLITE6_inuse gauge node_sockstat_UDPLITE6_inuse 0 # HELP node_sockstat_UDPLITE_inuse Number of UDPLITE sockets in state inuse. # TYPE node_sockstat_UDPLITE_inuse gauge node_sockstat_UDPLITE_inuse 0 # HELP node_sockstat_UDP_inuse Number of UDP sockets in state inuse. # TYPE node_sockstat_UDP_inuse gauge node_sockstat_UDP_inuse 4 # HELP node_sockstat_UDP_mem Number of UDP sockets in state mem. # TYPE node_sockstat_UDP_mem gauge node_sockstat_UDP_mem 4 # HELP node_sockstat_UDP_mem_bytes Number of UDP sockets in state mem_bytes. # TYPE node_sockstat_UDP_mem_bytes gauge node_sockstat_UDP_mem_bytes 16384 # HELP node_sockstat_sockets_used Number of IPv4 sockets in use. # TYPE node_sockstat_sockets_used gauge node_sockstat_sockets_used 414 # HELP node_softnet_backlog_len Softnet backlog status # TYPE node_softnet_backlog_len gauge node_softnet_backlog_len{cpu="0"} 0 node_softnet_backlog_len{cpu="1"} 0 # HELP node_softnet_cpu_collision_total Number of collision occur while obtaining device lock while transmitting # TYPE node_softnet_cpu_collision_total counter node_softnet_cpu_collision_total{cpu="0"} 0 node_softnet_cpu_collision_total{cpu="1"} 0 # HELP node_softnet_dropped_total Number of dropped packets # TYPE node_softnet_dropped_total counter node_softnet_dropped_total{cpu="0"} 0 node_softnet_dropped_total{cpu="1"} 0 # HELP node_softnet_flow_limit_count_total Number of times flow limit has been reached # TYPE node_softnet_flow_limit_count_total counter node_softnet_flow_limit_count_total{cpu="0"} 0 node_softnet_flow_limit_count_total{cpu="1"} 0 # HELP node_softnet_processed_total Number of processed packets # TYPE node_softnet_processed_total counter node_softnet_processed_total{cpu="0"} 2.0386074e+07 node_softnet_processed_total{cpu="1"} 2.6689292e+07 # HELP node_softnet_received_rps_total Number of times cpu woken up received_rps # TYPE node_softnet_received_rps_total counter node_softnet_received_rps_total{cpu="0"} 0 node_softnet_received_rps_total{cpu="1"} 0 # HELP node_softnet_times_squeezed_total Number of times processing packets ran out of quota # TYPE node_softnet_times_squeezed_total counter node_softnet_times_squeezed_total{cpu="0"} 0 node_softnet_times_squeezed_total{cpu="1"} 0 # HELP node_textfile_scrape_error 1 if there was an error opening or reading a file, 0 otherwise # TYPE node_textfile_scrape_error gauge node_textfile_scrape_error 0 # HELP node_time_clocksource_available_info Available clocksources read from '/sys/devices/system/clocksource'. # TYPE node_time_clocksource_available_info gauge node_time_clocksource_available_info{clocksource="acpi_pm",device="0"} 1 node_time_clocksource_available_info{clocksource="hyperv_clocksource_tsc_page",device="0"} 1 node_time_clocksource_available_info{clocksource="tsc",device="0"} 1 # HELP node_time_clocksource_current_info Current clocksource read from '/sys/devices/system/clocksource'. # TYPE node_time_clocksource_current_info gauge node_time_clocksource_current_info{clocksource="tsc",device="0"} 1 # HELP node_time_seconds System time in seconds since epoch (1970). # TYPE node_time_seconds gauge node_time_seconds 1.7766148482934856e+09 # HELP node_time_zone_offset_seconds System time zone offset in seconds. # TYPE node_time_zone_offset_seconds gauge node_time_zone_offset_seconds{time_zone="UTC"} 0 # HELP node_timex_estimated_error_seconds Estimated error in seconds. # TYPE node_timex_estimated_error_seconds gauge node_timex_estimated_error_seconds 0 # HELP node_timex_frequency_adjustment_ratio Local clock frequency adjustment. # TYPE node_timex_frequency_adjustment_ratio gauge node_timex_frequency_adjustment_ratio 1.0000017159118653 # HELP node_timex_loop_time_constant Phase-locked loop time constant. # TYPE node_timex_loop_time_constant gauge node_timex_loop_time_constant 2 # HELP node_timex_maxerror_seconds Maximum error in seconds. # TYPE node_timex_maxerror_seconds gauge node_timex_maxerror_seconds 0.002006 # HELP node_timex_offset_seconds Time offset in between local system and reference clock. # TYPE node_timex_offset_seconds gauge node_timex_offset_seconds 0 # HELP node_timex_pps_calibration_total Pulse per second count of calibration intervals. # TYPE node_timex_pps_calibration_total counter node_timex_pps_calibration_total 0 # HELP node_timex_pps_error_total Pulse per second count of calibration errors. # TYPE node_timex_pps_error_total counter node_timex_pps_error_total 0 # HELP node_timex_pps_frequency_hertz Pulse per second frequency. # TYPE node_timex_pps_frequency_hertz gauge node_timex_pps_frequency_hertz 0 # HELP node_timex_pps_jitter_seconds Pulse per second jitter. # TYPE node_timex_pps_jitter_seconds gauge node_timex_pps_jitter_seconds 0 # HELP node_timex_pps_jitter_total Pulse per second count of jitter limit exceeded events. # TYPE node_timex_pps_jitter_total counter node_timex_pps_jitter_total 0 # HELP node_timex_pps_shift_seconds Pulse per second interval duration. # TYPE node_timex_pps_shift_seconds gauge node_timex_pps_shift_seconds 0 # HELP node_timex_pps_stability_exceeded_total Pulse per second count of stability limit exceeded events. # TYPE node_timex_pps_stability_exceeded_total counter node_timex_pps_stability_exceeded_total 0 # HELP node_timex_pps_stability_hertz Pulse per second stability, average of recent frequency changes. # TYPE node_timex_pps_stability_hertz gauge node_timex_pps_stability_hertz 0 # HELP node_timex_status Value of the status array bits. # TYPE node_timex_status gauge node_timex_status 0 # HELP node_timex_sync_status Is clock synchronized to a reliable server (1 = yes, 0 = no). # TYPE node_timex_sync_status gauge node_timex_sync_status 1 # HELP node_timex_tai_offset_seconds International Atomic Time (TAI) offset. # TYPE node_timex_tai_offset_seconds gauge node_timex_tai_offset_seconds 0 # HELP node_timex_tick_seconds Seconds between clock ticks. # TYPE node_timex_tick_seconds gauge node_timex_tick_seconds 0.01 # HELP node_udp_queues Number of allocated memory in the kernel for UDP datagrams in bytes. # TYPE node_udp_queues gauge node_udp_queues{ip="v4",queue="rx"} 0 node_udp_queues{ip="v4",queue="tx"} 0 node_udp_queues{ip="v6",queue="rx"} 0 node_udp_queues{ip="v6",queue="tx"} 0 # HELP node_uname_info Labeled system information as provided by the uname system call. # TYPE node_uname_info gauge node_uname_info{domainname="(none)",machine="x86_64",nodename="aks-systemnodes1-25099092-vmss000001",release="5.15.0-1102-azure",sysname="Linux",version="#111-Ubuntu SMP Fri Nov 21 22:22:11 UTC 2025"} 1 # HELP node_vmstat_oom_kill /proc/vmstat information field oom_kill. # TYPE node_vmstat_oom_kill untyped node_vmstat_oom_kill 0 # HELP node_vmstat_pgfault /proc/vmstat information field pgfault. # TYPE node_vmstat_pgfault untyped node_vmstat_pgfault 2.31128293e+08 # HELP node_vmstat_pgmajfault /proc/vmstat information field pgmajfault. # TYPE node_vmstat_pgmajfault untyped node_vmstat_pgmajfault 13706 # HELP node_vmstat_pgpgin /proc/vmstat information field pgpgin. # TYPE node_vmstat_pgpgin untyped node_vmstat_pgpgin 3.477927e+06 # HELP node_vmstat_pgpgout /proc/vmstat information field pgpgout. # TYPE node_vmstat_pgpgout untyped node_vmstat_pgpgout 1.8148069e+07 # HELP node_vmstat_pswpin /proc/vmstat information field pswpin. # TYPE node_vmstat_pswpin untyped node_vmstat_pswpin 0 # HELP node_vmstat_pswpout /proc/vmstat information field pswpout. # TYPE node_vmstat_pswpout untyped node_vmstat_pswpout 0 # HELP process_cpu_seconds_total Total user and system CPU time spent in seconds. # TYPE process_cpu_seconds_total counter process_cpu_seconds_total 23.99 # HELP process_max_fds Maximum number of open file descriptors. # TYPE process_max_fds gauge process_max_fds 1.048576e+06 # HELP process_network_receive_bytes_total Number of bytes received by the process over the network. # TYPE process_network_receive_bytes_total counter process_network_receive_bytes_total 2.9175025245e+10 # HELP process_network_transmit_bytes_total Number of bytes sent by the process over the network. # TYPE process_network_transmit_bytes_total counter process_network_transmit_bytes_total 5.4710726946e+10 # HELP process_open_fds Number of open file descriptors. # TYPE process_open_fds gauge process_open_fds 10 # HELP process_resident_memory_bytes Resident memory size in bytes. # TYPE process_resident_memory_bytes gauge process_resident_memory_bytes 2.9249536e+07 # HELP process_start_time_seconds Start time of the process since unix epoch in seconds. # TYPE process_start_time_seconds gauge process_start_time_seconds 1.77660100958e+09 # HELP process_virtual_memory_bytes Virtual memory size in bytes. # TYPE process_virtual_memory_bytes gauge process_virtual_memory_bytes 1.307598848e+09 # HELP process_virtual_memory_max_bytes Maximum amount of virtual memory available in bytes. # TYPE process_virtual_memory_max_bytes gauge process_virtual_memory_max_bytes 1.8446744073709552e+19 # HELP promhttp_metric_handler_errors_total Total number of internal errors encountered by the promhttp metric handler. # TYPE promhttp_metric_handler_errors_total counter promhttp_metric_handler_errors_total{cause="encoding"} 0 promhttp_metric_handler_errors_total{cause="gathering"} 0 # HELP promhttp_metric_handler_requests_in_flight Current number of scrapes being served. # TYPE promhttp_metric_handler_requests_in_flight gauge promhttp_metric_handler_requests_in_flight 1 # HELP promhttp_metric_handler_requests_total Total number of scrapes by HTTP status code. # TYPE promhttp_metric_handler_requests_total counter promhttp_metric_handler_requests_total{code="200"} 463 promhttp_metric_handler_requests_total{code="500"} 0 promhttp_metric_handler_requests_total{code="503"} 0