Targets


default/pi-stock-service-monitor/0 (1/1 up)

Endpoint State Labels Last Scrape Scrape Duration Error
http://10.42.0.68:3001/metrics
up endpoint="pi-stock" instance="10.42.0.68:3001" job="pi-stock" namespace="default" pod="pi-stock-5d8dd7f4f7-qxqr8" service="pi-stock" 19.947s ago 12.24ms

kube-system/traefik-monitoring/0 (3/3 up)

Endpoint State Labels Last Scrape Scrape Duration Error
http://10.42.0.59:8081/metrics
up container="traefik" endpoint="metrics" instance="10.42.0.59:8081" job="kube-system/traefik-monitoring" namespace="kube-system" pod="traefik-4vlnq" 14.184s ago 6.369ms
http://10.42.1.207:8081/metrics
up container="traefik" endpoint="metrics" instance="10.42.1.207:8081" job="kube-system/traefik-monitoring" namespace="kube-system" pod="traefik-mgcdm" 59.659s ago 64.04ms
http://10.42.2.38:8081/metrics
up container="traefik" endpoint="metrics" instance="10.42.2.38:8081" job="kube-system/traefik-monitoring" namespace="kube-system" pod="traefik-s8p27" 26.315s ago 11.67ms

monitoring/alertmanager/0 (1/1 up)

Endpoint State Labels Last Scrape Scrape Duration Error
http://10.42.2.36:9093/metrics
up endpoint="web" instance="10.42.2.36:9093" job="alertmanager-main" namespace="monitoring" pod="alertmanager-main-0" service="alertmanager-main" 22.338s ago 15.02ms

monitoring/arm-exporter/0 (3/3 up)

Endpoint State Labels Last Scrape Scrape Duration Error
https://10.42.0.56:9243/metrics
up endpoint="https" instance="main" job="arm-exporter" namespace="monitoring" pod="arm-exporter-gjg9w" service="arm-exporter" 29.157s ago 64.69ms
https://10.42.1.208:9243/metrics
up endpoint="https" instance="node1" job="arm-exporter" namespace="monitoring" pod="arm-exporter-jzhqp" service="arm-exporter" 28.907s ago 133ms
https://10.42.2.37:9243/metrics
up endpoint="https" instance="worker2" job="arm-exporter" namespace="monitoring" pod="arm-exporter-7w8n2" service="arm-exporter" 21.363s ago 844.4ms

monitoring/coredns/0 (2/2 up)

Endpoint State Labels Last Scrape Scrape Duration Error
http://10.42.0.74:9153/metrics
up endpoint="metrics" instance="10.42.0.74:9153" job="kube-dns" namespace="kube-system" pod="coredns-7796b77cd4-mt5dq" service="kube-dns" 14.459s ago 9.638ms
http://10.42.0.74:9153/metrics
up endpoint="metrics" instance="10.42.0.74:9153" job="kube-dns" namespace="kube-system" pod="coredns-7796b77cd4-mt5dq" service="kube-dns-prometheus-discovery" 4.898s ago 12.24ms

monitoring/grafana/0 (1/1 up)

Endpoint State Labels Last Scrape Scrape Duration Error
http://10.42.0.80:3000/metrics
up endpoint="http" instance="10.42.0.80:3000" job="grafana" namespace="monitoring" pod="grafana-594fc7f587-v7wlr" service="grafana" 9.224s ago 13.56ms

monitoring/kube-apiserver/0 (1/1 up)

Endpoint State Labels Last Scrape Scrape Duration Error
https://192.168.178.170:6443/metrics
up endpoint="https" instance="192.168.178.170:6443" job="apiserver" namespace="default" service="kubernetes" 5.632s ago 712.1ms

monitoring/kube-controller-manager/0 (0/1 up)

Endpoint State Labels Last Scrape Scrape Duration Error
http://192.168.178.170:10252/metrics
down endpoint="http-metrics" instance="192.168.178.170:10252" job="kube-controller-manager" namespace="kube-system" service="kube-controller-manager-prometheus-discovery" 11.477s ago 1.028ms Get "http://192.168.178.170:10252/metrics": dial tcp 192.168.178.170:10252: connect: connection refused

monitoring/kube-scheduler/0 (1/1 up)

Endpoint State Labels Last Scrape Scrape Duration Error
http://192.168.178.170:10251/metrics
up endpoint="http-metrics" instance="192.168.178.170:10251" job="kube-scheduler" namespace="kube-system" service="kube-scheduler-prometheus-discovery" 3.028s ago 721.7ms

monitoring/kube-state-metrics/0 (1/1 up)

Endpoint State Labels Last Scrape Scrape Duration Error
https://10.42.0.58:8443/metrics
up instance="10.42.0.58:8443" job="kube-state-metrics" 8.475s ago 69.42ms

monitoring/kube-state-metrics/1 (1/1 up)

Endpoint State Labels Last Scrape Scrape Duration Error
https://10.42.0.58:9443/metrics
up endpoint="https-self" instance="10.42.0.58:9443" job="kube-state-metrics" namespace="monitoring" pod="kube-state-metrics-6cb6df5d4-hbsn7" service="kube-state-metrics" 2.644s ago 8.454ms

monitoring/kubelet/0 (3/3 up)

Endpoint State Labels Last Scrape Scrape Duration Error
https://192.168.178.170:10250/metrics
up endpoint="https-metrics" instance="192.168.178.170:10250" job="kubelet" metrics_path="/metrics" namespace="kube-system" node="main" service="kubelet" 13.314s ago 812ms
https://192.168.178.27:10250/metrics
up endpoint="https-metrics" instance="192.168.178.27:10250" job="kubelet" metrics_path="/metrics" namespace="kube-system" node="worker2" service="kubelet" 13.126s ago 76.7ms
https://192.168.178.94:10250/metrics
up endpoint="https-metrics" instance="192.168.178.94:10250" job="kubelet" metrics_path="/metrics" namespace="kube-system" node="node1" service="kubelet" 6.776s ago 81.61ms

monitoring/kubelet/1 (3/3 up)

Endpoint State Labels Last Scrape Scrape Duration Error
https://192.168.178.170:10250/metrics/cadvisor
up endpoint="https-metrics" instance="192.168.178.170:10250" job="kubelet" metrics_path="/metrics/cadvisor" namespace="kube-system" node="main" service="kubelet" 3.923s ago 558.9ms
https://192.168.178.27:10250/metrics/cadvisor
up endpoint="https-metrics" instance="192.168.178.27:10250" job="kubelet" metrics_path="/metrics/cadvisor" namespace="kube-system" node="worker2" service="kubelet" 8.293s ago 208.3ms
https://192.168.178.94:10250/metrics/cadvisor
up endpoint="https-metrics" instance="192.168.178.94:10250" job="kubelet" metrics_path="/metrics/cadvisor" namespace="kube-system" node="node1" service="kubelet" 690ms ago 176.9ms

monitoring/node-exporter/0 (3/3 up)

Endpoint State Labels Last Scrape Scrape Duration Error
https://192.168.178.170:9100/metrics
up endpoint="https" instance="main" job="node-exporter" namespace="monitoring" pod="node-exporter-kgllz" service="node-exporter" 5.293s ago 615.3ms
https://192.168.178.94:9100/metrics
up endpoint="https" instance="node1" job="node-exporter" namespace="monitoring" pod="node-exporter-4xk97" service="node-exporter" 12.048s ago 178ms
https://192.168.178.27:9100/metrics
up endpoint="https" instance="worker2" job="node-exporter" namespace="monitoring" pod="node-exporter-t242n" service="node-exporter" 16.299s ago 391ms

monitoring/prometheus-operator/0 (1/1 up)

Endpoint State Labels Last Scrape Scrape Duration Error
https://10.42.0.55:8443/metrics
up endpoint="https" instance="10.42.0.55:8443" job="prometheus-operator" namespace="monitoring" pod="prometheus-operator-67755f959-g8phz" service="prometheus-operator" 15.051s ago 25.52ms

monitoring/prometheus/0 (1/1 up)

Endpoint State Labels Last Scrape Scrape Duration Error
http://10.42.0.54:9090/metrics
up endpoint="web" instance="10.42.0.54:9090" job="prometheus-k8s" namespace="monitoring" pod="prometheus-k8s-0" service="prometheus-k8s" 1.214s ago 31.16ms