prometheus 监控案例-Tomcat、Redis、Mysql、Haproxy、Nginx

  服务监控

  可以根据prometheus官网找到对应服务的exporter

  https://prometheus.io/docs/instrumenting/exporters/

  监控tomcat容器

  基于第三方 exporter 实现对目的服务的监控

  github项目地址:https://github.com/nlighten/tomcat_exporter

  所需要的Jar包获取

  选择对应版本采集服务程序包,点击下载按钮,下载则是最新版本

  如果想下载以往版本解决兼容某些tomcat问题,则点击版本后进入详细版本页面进行下载

  根据测试,simpleclient程序包使用0.12.0开始以后版本不对8.5.x 的tomcat兼容,无法显示metrics页面,但tomcat-exporter程序包可以使用最新版本

  Jar监控程序包列表:

  simpleclient-0.8.0.jar

  simpleclient_common-0.8.0.jar

  simpleclient_hotspot-0.8.0.jar

  simpleclient_servlet-0.8.0.jar

  tomcat_exporter_client-0.0.17.jar

  tomcat_exporter_servlet-0.0.17.war

  并将tomcat_exporter_servlet-0.0.17.war修改为metrics进行tomcat页面发布

 

  监控 tomcat 的活跃连接数、堆栈内存等信息:

# TYPE tomcat_connections_active_total gauge
tomcat_connections_active_total{name="http-nio-8080",} 2.0
# TYPE jvm_memory_bytes_used gauge
jvm_memory_bytes_used{area="heap",} 2.4451216E7

 

  制作tomcat镜像,将包含metrics监控指标的jar包导入tomcat镜像内

root@deploy:~/dockerfile# cat Dockerfile
FROM harbor.cncf.net/web/tomcat:8.5.43
MAINTAINER LXH
LABEL Description="tomcat-8.5.43-app1"
ADD metrics.war /usr/local/tomcat/webapps/
ADD simpleclient-0.8.0.jar /usr/local/tomcat/lib
ADD simpleclient_common-0.8.0.jar /usr/local/tomcat/lib
ADD simpleclient_hotspot-0.8.0.jar /usr/local/tomcat/lib
ADD simpleclient_servlet-0.8.0.jar /usr/local/tomcat/lib
ADD tomcat_exporter_client-0.0.17.jar /usr/local/tomcat/lib
EXPOSE 8080 8443

 

  容器镜像构建脚本

root@deploy:~/dockerfile# cat build.sh
#!/bin/bash
version=$1
docker build -t tomcatapp:${version} ./
docker tag tomcatapp:${version} harbor.cncf.net/web/tomcatapp:${version}
docker push harbor.cncf.net/web/tomcatapp:${version}

 

  编写tomcat k8s yaml文件

  k8s部署监控相关的组件的时候需要初始化很多linux的内核参数。 但是文件系统挂载到pod容器中就会变成read-only,难以进行操作实现需求。所以需要给POD privileged权限,然后在容器的初始化脚本或代码中去修改sysctl参数。给容器的spec指定`securityContext.privileged=true`参数

  prometheus-server如果使用k8s内部部署并配置服务发现,则需要添加prometheus的scrape注解`annotations.prometheus.io/scrape='true'`

root@master1:~/yaml/tomcat-exporter# cat tomcat-deploy.yaml
apiVersion: apps/v1
kind: Deployment
metadata:
name: tomcat-deployment
namespace: default
spec:
selector:
matchLabels:
app: tomcat
replicas: 1
template:
metadata:
labels:
app: tomcat
annotations:
prometheus.io/scrape: 'true' #添加prometheus服务发现的抓取注解
spec:
containers:
- name: tomcat
image: harbor.cncf.net/web/tomcatapp:1.0
imagePullPolicy: Always
ports:
- containerPort: 8080
securityContext:
privileged: true #给POD privileged权限

 

  部署tomcat server,如果svc也想被prometheus-server服务发现采集,也许要添加prometheus-server服务注解

root@master1:~/yaml/tomcat-exporter# cat tomcat-svc.yaml
kind: Service
apiVersion: v1
metadata:
annotations:
prometheus.io/scrape: 'true' #添加prometheus服务发现的抓取注解
name: tomcat-service
spec:
selector:
app: tomcat
ports:
- nodePort: 31080
port: 80
protocol: TCP
targetPort: 8080
type: NodePort

  

  验证:

root@master1:~/yaml/tomcat-exporter# kubectl get pods -n default
NAME READY STATUS RESTARTS AGE
tomcat-deployment-b8c8457cf-fgztx 1/1 Running 0 14s
root@master1:~/yaml/tomcat-exporter# kubectl get svc -n default
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE
kubernetes ClusterIP 10.100.0.1 <none> 443/TCP 17d

 

  访问tomcat页面和指标

 

  配置prometheus-server配置文件,添加tomcat-exporter

root@prometheus:/usr/local/prometheus# vim prometheus.yml
- job_name: "tomcat"
static_configs:
- targets: ["192.168.100.3:31080"]
root@prometheus:/usr/local/prometheus# systemctl restart prometheus

 

  部署导入grafana官方模板

  https://github.com/nlighten/tomcat_exporter/tree/master/dashboard

  

  复制example.json导入grafana

  

  监控redis

  通过 redis_exporter 监控 redis 服务状态。

   官方exporter:

  prometheus官网 https://prometheus.io/docs/instrumenting/exporters/ ,找到redis_exporter

  redis_exporter的GIthub项目地址 https://github.com/oliver006/redis_exporter

 

  下载redis-exporter

  https://github.com/oliver006/redis_exporter/releases

 

  解压部署redis_exporter

[root@redis1 local]# tar xf redis_exporter-v1.43.1.linux-amd64.tar.gz

  

  redis-exporter会默认监听本机localhost:6379端口,来返回指标数据/metrics。启动默认监听9121端口

  命令帮助信息

root@redis1 redis_exporter]# ./redis_exporter --help
Usage of ./redis_exporter:
-check-key-groups string
Comma separated list of lua regex for grouping keys
-check-keys string
Comma separated list of key-patterns to export value and length/size, searched for with SCAN
-check-keys-batch-size int
Approximate number of keys to process in each execution, larger value speeds up scanning.
WARNING: Still Redis is a single-threaded app, huge COUNT can affect production environment. (default 1000)
-check-single-keys string
Comma separated list of single keys to export value and length/size
-check-single-streams string
Comma separated list of single streams to export info about streams, groups and consumers
-check-streams string
Comma separated list of stream-patterns to export info about streams, groups and consumers, searched for with SCAN
-config-command string
What to use for the CONFIG command (default "CONFIG")
-connection-timeout string
Timeout for connection to Redis instance (default "15s")
-count-keys string
Comma separated list of patterns to count (eg: 'db0=production_*,db3=sessions:*'), searched for with SCAN
-debug
Output verbose debug information
-export-client-list
Whether to scrape Client List specific metrics
-export-client-port
Whether to include the client's port when exporting the client list. Warning: including the port increases the number of metrics generated and will make your Prometheus server take up more memory
-include-config-metrics
Whether to include all config settings as metrics
-include-system-metrics
Whether to include system metrics like e.g. redis_total_system_memory_bytes
-is-cluster
Whether this is a redis cluster (Enable this if you need to fetch key level data on a Redis Cluster).
-is-tile38
Whether to scrape Tile38 specific metrics
-log-format string
Log format, valid options are txt and json (default "txt")
-max-distinct-key-groups int
The maximum number of distinct key groups with the most memory utilization to present as distinct metrics per database, the leftover key groups will be aggregated in the 'overflow' bucket (default 100)
-namespace string
Namespace for metrics (default "redis")
-ping-on-connect
Whether to ping the redis instance after connecting
-redact-config-metrics
Whether to redact config settings that include potentially sensitive information like passwords (default true)
-redis-only-metrics
Whether to also export go runtime metrics
-redis.addr string
Address of the Redis instance to scrape (default "redis://localhost:6379")
-redis.password string
Password of the Redis instance to scrape
-redis.password-file string
Password file of the Redis instance to scrape
-redis.user string
User name to use for authentication (Redis ACL for Redis 6.0 and newer)
-script string
Path to Lua Redis script for collecting extra metrics
-set-client-name
Whether to set client name to redis_exporter (default true)
-skip-tls-verification
Whether to to skip TLS verification
-tls-ca-cert-file string
Name of the CA certificate file (including full path) if the server requires TLS client authentication
-tls-client-cert-file string
Name of the client certificate file (including full path) if the server requires TLS client authentication
-tls-client-key-file string
Name of the client key file (including full path) if the server requires TLS client authentication
-tls-server-ca-cert-file string
Name of the CA certificate file (including full path) if the web interface and telemetry should require TLS client authentication
-tls-server-cert-file string
Name of the server certificate file (including full path) if the web interface and telemetry should use TLS
-tls-server-key-file string
Name of the server key file (including full path) if the web interface and telemetry should use TLS
-tls-server-min-version string
Minimum TLS version that is acceptable by the web interface and telemetry when using TLS (default "TLS1.2")
-version
Show version information and exit
-web.listen-address string
Address to listen on for web interface and telemetry. (default ":9121")
-web.telemetry-path string
Path under which to expose metrics. (default "/metrics")

 

  常用选项:

  -redis.addr: 连接redis地址:端口

  -redis.password: 连接redis的密码

  -web.listen-address: 指定redis_exporter的服务端口,格式为-web.listen-address ":9121

 

  启动脚本

  使redis_exporter监听在自定义端口9333,默认端口监听为9121

[root@redis1 redis_exporter]# cat start.sh
#!/bin/bash
cmd=/usr/local/redis_exporter
ip=192.168.100.104 #reidis地址
password=123456 #redis密码
nohup $cmd/redis_exporter -redis.addr $ip:6379 -redis.password $password -web.listen-address ":9333" > /dev/null 2>&1 &
[ $? -eq 0 ] && echo "start success!" || echo "start failed"

 

  停止脚本

root@redis1 redis_exporter]# cat stop.sh
#!/bin/bash
PID=$(ps -ef|grep redis_exporter|grep -v grep|awk '{print $2}')
if [ -z ${PID} ];then
echo "redis_expoter isn't running"
else
kill $PID
sleep 2s
[ ! -z ${PID} ]
echo "redis_expoter is stopped"
fi

 

  监控redis单实例

  编辑启动脚本后,运行启动脚本

[root@redis1 redis_exporter]# cat start.sh
#!/bin/bash
cmd=/usr/local/redis_exporter
ip=192.168.100.104 #reidis地址
password=123456 #redis密码
nohup $cmd/redis_exporter -redis.addr $ip:6379 -redis.password $password -web.listen-address ":9333" > /dev/null 2>&1 &
[ $? -eq 0 ] && echo "start success!" || echo "start failed"
[root@redis1 redis_exporter]# ./start.sh
start success!

  

  访问验证

  prometheus-server配置

scrape_configs:
- job_name: redis_exporter
static_configs:
- targets: ['192.168.100.104:9333']

 

  监控redis多实例集群

  查看集群信息

  分别为192.168.100.104上的6379-6384端口的6个redis为集群

 

  官网示例

  https://github.com/oliver006/redis_exporter

  Using below config instead of the /metric endpoint the /scrape endpoint will be used by prometheus. As an example the first target will be queried with this web request:

  http://exporterhost:9121/scrape?target=first-redis-host:6379

 

  解释

  1、创建一个名称为redis_exporter_targets job,来获取每个redis节点的实例。

  将采集redis_expoter Http URI修改为/scrape,原因是默认prometheus会收集目标target的/metrics URI,所以需要在prometheus-server配置文件中指定metrics_path 的路径,并进行redis_exporter的标签重写,将\__address\__ 标签内容为`redis实例ip:6379、6380、6381`节点实例,替换为替换为instance标签,并将value修改为`redis_exporter http`的访问地址。最终prometheus-server就能获得每个redis的target:`http://exporterhost:9121/scrape?target=first-redis-host:6379`

  2、还需要单独添加一个名称为redis_exporter job来抓取redis_exporter的地址

scrape_configs:
## config for the multiple Redis targets that the exporter will scrape
- job_name: 'redis_exporter_targets'
static_configs:
- targets:
- redis://first-redis-host:6379
- redis://second-redis-host:6379
- redis://second-redis-host:6380
- redis://second-redis-host:6381
metrics_path: /scrape
relabel_configs:
- source_labels: [__address__]
target_label: __param_target
- source_labels: [__param_target]
target_label: instance
- target_label: __address__
replacement: <<REDIS-EXPORTER-HOSTNAME>>:9121
## config for scraping the exporter itself
- job_name: 'redis_exporter'
static_configs:
- targets:
- <<REDIS-EXPORTER-HOSTNAME>>:9121

 

  redis_expoter配置

  redis_expoter只要能连接到一个集群的一个节点,因此只需要配置连接集群中一个redis节点就能查询其他节点的指标

[root@redis1 redis_exporter]\# cat start.sh
#!/bin/bash
cmd=/usr/local/redis_exporter
ip=192.168.100.104
password=123456
nohup $cmd/redis_exporter -redis.addr $ip:6379 -redis.password $password -web.listen-address ":9333" > /dev/null 2>&1 &
[ $? -eq 0 ] && echo "start success!" || echo "start failed"

 

  prometheus-server配置

root@prometheus:~\# cat /usr/local/prometheus/prometheus.yml
# my global config
global:
scrape_interval: 15s # Set the scrape interval to every 15 seconds. Default is every 1 minute.
evaluation_interval: 15s # Evaluate rules every 15 seconds. The default is every 1 minute.
# scrape_timeout is set to the global default (10s).
alerting:
alertmanagers:
- static_configs:
- targets:
# - alertmanager:9093
rule_files:
# - "first_rules.yml"
# - "second_rules.yml"
scrape_configs:
- job_name: 'redis_exporter_targets'
static_configs:
- targets:
- redis://192.168.100.104:6379
- redis://192.168.100.104:6380
- redis://192.168.100.104:6381
- redis://192.168.100.104:6382
- redis://192.168.100.104:6383
- redis://192.168.100.104:6384
metrics_path: /scrape
relabel_configs:
- source_labels: [__address__]
target_label: __param_target
- source_labels: [__param_target]
target_label: instance
- target_label: __address__
replacement: 192.168.100.104:9333
## config for scraping the exporter itself
- job_name: 'redis_exporter'
static_configs:
- targets:
- 192.168.100.104:9333

 

  访问prometheus-server web页面

  将标签重写后,URI修改为/scripe?target=redis节点:6379

  配置grafana导入redis-exporter模板

  ID:11835

 

  

监控 k8s redis集群

  这里每个redis节点使用statefulset部署,并暴露集群端口供redis_exporter使用。redis_expoter单独使用deployment部署

  部署redis集群参考kubernetes 案例redis集群配置

 

  验证redis集群

  查看redis集群pod

  查看集群状态

redis_exporter镜像构建

root@deploy:~/dockerfile/redis-exporter\# cat Dockerfile
FROM harbor.cncf.net/os/ubuntu:20.04
ADD redis_exporter.tar.gz /usr/local/
ENTRYPOINT ["/usr/local/redis_exporter/start.sh"]

 

准备redis_exporter程序,下载后解压并配置

root@deploy:~/dockerfile/redis-exporter\# ll redis_exporter/
total 8892
drwxr-xr-x 2 root root 91 Sep 24 03:29 ./
drwxr-xr-x 3 root root 91 Sep 24 03:30 ../
-rw-r--r-- 1 root root 1063 Sep 24 03:10 LICENSE
-rw-r--r-- 1 root root 43382 Sep 24 03:10 README.md
-rwxr-xr-x 1 root root 9048064 Sep 24 03:10 redis_exporter*
-rwx------ 1 root root 202 Sep 24 03:29 start.sh*
-rwx------ 1 root root 224 Sep 24 03:10 stop.sh*

 

准备redis_exporter启动脚本

  redis的ip地址修改为k8s中redis集群中的master节点的pod服务名称,redis-0.redis.redis-test.svc.cluster.local

root@deploy:~/dockerfile/redis-exporter\# cat redis_exporter/start.sh
#!/bin/bash
cmd=/usr/local/redis_exporter
ip="redis-0.redis.redis-test.svc.cluster.local"
password=123456
$cmd/redis_exporter -redis.addr $ip:6379 -redis.password $password -web.listen-address ":9121"

 

  准备镜像构建脚本

root@deploy:~/dockerfile/redis-exporter\# cat build.sh
#!/bin/bash
version=$1
docker build -t harbor.cncf.net/project/redis_exporter:$version ./
docker push harbor.cncf.net/project/redis_exporter:$version

 

  部署redis_exporter

  创建deployment,与redis集群在同一个ns下

root@master1:~/yaml/redis-exporter# cat redis-exporter.yaml
apiVersion: apps/v1
kind: Deployment
metadata:
name: redis-exporter
namespace: redis-test
spec:
replicas: 1
selector:
matchLabels:
app: redis-exporter
template:
metadata:
labels:
app: redis-exporter
spec:
containers:
- name: redis-exporter
image: harbor.cncf.net/project/redis_exporter:1.7
imagePullPolicy: Always
resources:
requests:
cpu: 100m
memory: 100Mi
ports:
- containerPort: 9121

  

  创建svc

root@master1:~/yaml/redis-exporter# cat redis-exporter-svc.yaml
kind: Service #service 类型
apiVersion: v1
metadata:
name: redis-redis-service
namespace: redis-test
spec:
selector:
app: redis-exporter
ports:
- nodePort: 39121
name: redis-exporter
port: 9121
protocol: TCP
targetPort: 9121
type: NodePort

 

  验证部署

root@master1:~/yaml\# kubectl get svc -n redis-test redis-redis-service
NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE
redis-redis-service NodePort 10.100.135.230 <none> 9121:39121/TCP 99m
root@master1:~/yaml\# kubectl get pods -n redis-test
NAME READY STATUS RESTARTS AGE
redis-0 1/1 Running 0 134m
redis-1 1/1 Running 0 142m
redis-2 1/1 Running 0 142m
redis-3 1/1 Running 0 142m
redis-4 1/1 Running 0 142m
redis-5 1/1 Running 0 126m
redis-exporter-bf6bf8576-k9p4x 1/1 Running 0 101m

 

  验证查看redis_exporter web界面

  配置k8s中的prometheus-server,添加redis-exporter服务发现

  target为redis-exporter的svc名称

- job_name: 'redis_exporter'
static_configs:
- targets:
- redis-redis-service.redis-test:9121

 

  配置后重新应用promentheus configmap并删除prometheus-server pod重启

root@master1:~/yaml\# kubectl apply -f prometheus-cfg.yaml
configmap/prometheus-config configured
root@master1:~/yaml\# kubectl delete -f prometheus-deployment.yaml
deployment.apps "prometheus-server" deleted
root@master1:~/yaml\# kubectl apply -f prometheus-deployment.yaml

 

  查看prometheus target界面验证target

  查看grafana导入dashboard

 

  监控mysql

  通过 mysqld_exporter 监控 MySQL 服务的运行状态

  官方exporter:

  prometheus官网: https://prometheus.io/docs/instrumenting/exporters/,找到mysql_exporer

  Github项目地址:https://github.com/prometheus/mysqld_exporter

 

  监控mysql单示例

  1、确保监控的mysql实例正常运行,添加授权监控账号权限。

  已经安装mysql服务,端口监听在3307

CREATE USER 'exporter'@'localhost' IDENTIFIED BY '123456' WITH MAX_USER_CONNECTIONS 3;
GRANT PROCESS, REPLICATION CLIENT, SELECT ON *.* TO 'exporter'@'localhost';

 

  测试账号权限

root@mysql1:~\# mysql -u exporter -p
#输入密码
mysql> select user();
+--------------------+
| user() |
+--------------------+
| exporter@localhost |
+--------------------+
1 row in set (0.00 sec)

 

  2、部署mysql_expoter,安装在mysql服务本地

下载地址:https://github.com/prometheus/mysqld_exporter/releases

  解压安装

root@mysql1:/usr/local\# tar xf mysqld_exporter-0.14.0.linux-amd64.tar.gz
root@mysql1:/usr/local\# ln -s mysqld_exporter-0.14.0.linux-amd64 mysql_exporter

  

  添加mysql客户端文件,添加监控账户密码信息和mysql服务端信息,使mysql_exporter免交互登录mysql

root@mysql1:~\# vim /root/.my.cnf
[client]
host=localhost
port=3307
user = exporter
password = 123456

 

  测试权限,直接本地执行mysql命令登录mysql

root@mysql1:~\# mysql
Welcome to the MySQL monitor. Commands end with ; or \g.
Your MySQL connection id is 5
Server version: 5.7.36 MySQL Community Server (GPL)
Copyright (c) 2000, 2021, Oracle and/or its affiliates.
Oracle is a registered trademark of Oracle Corporation and/or its
affiliates. Other names may be trademarks of their respective
owners.
Type 'help;' or '\h' for help. Type '\c' to clear the current input statement.
mysql>

 

  3、创建mysql_exporter服务启动文件

  设置端口监听在本机出口网卡的ip:9105,并指定mysql的客户端配置文件。默认端口为9104

root@mysql1:~\# vim /etc/systemd/system/mysqld_exporter.service
[Unit]
Description=Prometheus Node Exporter
After=network.target
[Service]
ExecStart=/usr/local/bin/mysqld_exporter --config.my-cnf=/root/.my.cnf --web.listen-address=192.168.0.29:9105
[Install]
WantedBy=multi-user.target

 

  mysqld_exporter的主要参数介绍

Name Description
mysqld.address Hostname and port used for connecting to MySQL server, format: host:port. (default: locahost:3306)
mysqld.username Username to be used for connecting to MySQL Server
config.my-cnf Path to .my.cnf file to read MySQL credentials from. (default: ~/.my.cnf)
log.level Logging verbosity (default: info)
exporter.lock_wait_timeout Set a lock_wait_timeout (in seconds) on the connection to avoid long metadata locking. (default: 2)
exporter.log_slow_filter Add a log_slow_filter to avoid slow query logging of scrapes. NOTE: Not supported by Oracle MySQL.
tls.insecure-skip-verify Ignore tls verification errors.
web.config.file Path to a web configuration file
web.listen-address Address to listen on for web interface and telemetry.
web.telemetry-path Path under which to expose metrics.
version Print the version information.

 

  命令帮助信息

root@mysql1:/usr/local\# mysqld_exporter --help
usage: mysqld_exporter [<flags>]
Flags:
-h, --help Show context-sensitive help (also try --help-long and --help-man).
--exporter.lock_wait_timeout=2
Set a lock_wait_timeout (in seconds) on the connection to avoid long metadata locking.
--exporter.log_slow_filter
Add a log_slow_filter to avoid slow query logging of scrapes. NOTE: Not supported by
Oracle MySQL.
--collect.heartbeat.database="heartbeat"
Database from where to collect heartbeat data
--collect.heartbeat.table="heartbeat"
Table from where to collect heartbeat data
--collect.heartbeat.utc Use UTC for timestamps of the current server (`pt-heartbeat` is called with `--utc`)
--collect.info_schema.processlist.min_time=0
Minimum time a thread must be in each state to be counted
--collect.info_schema.processlist.processes_by_user
Enable collecting the number of processes by user
--collect.info_schema.processlist.processes_by_host
Enable collecting the number of processes by host
--collect.info_schema.tables.databases="*"
The list of databases to collect table stats for, or '*' for all
--collect.mysql.user.privileges
Enable collecting user privileges from mysql.user
--collect.perf_schema.eventsstatements.limit=250
Limit the number of events statements digests by response time
--collect.perf_schema.eventsstatements.timelimit=86400
Limit how old the 'last_seen' events statements can be, in seconds
--collect.perf_schema.eventsstatements.digest_text_limit=120
Maximum length of the normalized statement text
--collect.perf_schema.file_instances.filter=".*"
RegEx file_name filter for performance_schema.file_summary_by_instance
--collect.perf_schema.file_instances.remove_prefix="/var/lib/mysql/"
Remove path prefix in performance_schema.file_summary_by_instance
--collect.perf_schema.memory_events.remove_prefix="memory/"
Remove instrument prefix in performance_schema.memory_summary_global_by_event_name
--web.config.file="" [EXPERIMENTAL] Path to configuration file that can enable TLS or authentication.
--web.listen-address=":9104"
Address to listen on for web interface and telemetry.
--web.telemetry-path="/metrics"
Path under which to expose metrics.
--timeout-offset=0.25 Offset to subtract from timeout in seconds.
--config.my-cnf="/root/.my.cnf"
Path to .my.cnf file to read MySQL credentials from.
--tls.insecure-skip-verify
Ignore certificate and server verification when using a tls connection.
--collect.mysql.user Collect data from mysql.user
--collect.info_schema.tables
Collect metrics from information_schema.tables
--collect.info_schema.innodb_tablespaces
Collect metrics from information_schema.innodb_sys_tablespaces
--collect.info_schema.innodb_metrics
Collect metrics from information_schema.innodb_metrics
--collect.global_status Collect from SHOW GLOBAL STATUS
--collect.global_variables
Collect from SHOW GLOBAL VARIABLES
--collect.slave_status Collect from SHOW SLAVE STATUS
--collect.info_schema.processlist
Collect current thread state counts from the information_schema.processlist
--collect.perf_schema.tablelocks
Collect metrics from performance_schema.table_lock_waits_summary_by_table
--collect.perf_schema.eventsstatements
Collect metrics from performance_schema.events_statements_summary_by_digest
--collect.perf_schema.eventsstatementssum
Collect metrics of grand sums from
performance_schema.events_statements_summary_by_digest
--collect.perf_schema.eventswaits
Collect metrics from performance_schema.events_waits_summary_global_by_event_name
--collect.auto_increment.columns
Collect auto_increment columns and max values from information_schema
--collect.binlog_size Collect the current size of all registered binlog files
--collect.perf_schema.tableiowaits
Collect metrics from performance_schema.table_io_waits_summary_by_table
--collect.perf_schema.indexiowaits
Collect metrics from performance_schema.table_io_waits_summary_by_index_usage
--collect.perf_schema.replication_group_member_stats
Collect metrics from performance_schema.replication_group_member_stats
--collect.perf_schema.replication_applier_status_by_worker
Collect metrics from performance_schema.replication_applier_status_by_worker
--collect.info_schema.userstats
If running with userstat=1, set to true to collect user statistics
--collect.info_schema.clientstats
If running with userstat=1, set to true to collect client statistics
--collect.perf_schema.file_events
Collect metrics from performance_schema.file_summary_by_event_name
--collect.perf_schema.file_instances
Collect metrics from performance_schema.file_summary_by_instance
--collect.perf_schema.memory_events
Collect metrics from performance_schema.memory_summary_global_by_event_name
--collect.perf_schema.replication_group_members
Collect metrics from performance_schema.replication_group_members
--collect.info_schema.query_response_time
Collect query response time distribution if query_response_time_stats is ON.
--collect.engine_tokudb_status
Collect from SHOW ENGINE TOKUDB STATUS
--collect.engine_innodb_status
Collect from SHOW ENGINE INNODB STATUS
--collect.heartbeat Collect from heartbeat
--collect.info_schema.tablestats
If running with userstat=1, set to true to collect table statistics
--collect.info_schema.schemastats
If running with userstat=1, set to true to collect schema statistics
--collect.info_schema.innodb_cmp
Collect metrics from information_schema.innodb_cmp
--collect.info_schema.innodb_cmpmem
Collect metrics from information_schema.innodb_cmpmem
--collect.slave_hosts Scrape information from 'SHOW SLAVE HOSTS'
--collect.info_schema.replica_host
Collect metrics from information_schema.replica_host_status
--log.level=info Only log messages with the given severity or above. One of: [debug, info, warn, error]
--log.format=logfmt Output format of log messages. One of: [logfmt, json]
--version Show application version.

 

  启动mysqld_exporter

systemctl daemon-reload && systemctl enable mysqld_exporter && systemctl start mysqld_exporter

 

  浏览器访问mysqld_exporter,端口修改为自定义的端口9105

  访问URI /metrics

  4、配置prometheus-server配置文件,添加mysqld_exporter服务采集

scrape_configs:
- job_name: "mysqld_exporter"
static_configs:
- targets: ["192.168.0.29:9105"]

 

  浏览器访问prometheus-server/targets验证

  5、配置grafana添加myqld_exporter的dashboard

  ID-13106、ID-11323

 

 

  监控Haproxy

  通过 haproxy_exporter 监控 haproxy

  prometheus官网:https://prometheus.io/docs/instrumenting/exporters/,找到haproxy_exporer

Github项目地址:https://github.com/prometheus/haproxy_exporter

 

  haproxy_exporter的监控方式

  1、通过haproxy 的状态页URL进行监控

  Github示例:

  2、通过haproxy socket文件进行监控

  配置haproxy 添加状态页与添加socket文件

global
stats socket /run/haproxy/admin.sock mode 660 level admin expose-fd listeners
listen statistics
#自定义监听端口
bind *:8009
#启用基于程序编译时默认设置的统计报告
stats enable
#统计页面用户名和密码设置
stats auth admin:admin
#自定义统计页面的URL,默认为/haproxy?stats
stats uri /admin?stats
#隐藏统计页面上HAProxy的版本信息
stats hide-version
#统计页面自动刷新时间
stats refresh 30s
#如果认证通过就做管理功能,可以管理后端的服务器
stats admin if TRUE
#统计页面密码框上提示文本,默认为Haproxy\ Statistics
stats realm Hapadmin

 

  浏览器访问验证:

  haproxy 状态页URI:/admin?stats

  验证haproxy服务器本地socket文件

root@haproxyA:~\# ll /run/haproxy/admin.sock
srw-rw---- 1 root haproxy 0 Sep 26 08:44 /run/haproxy/admin.sock=

 

  部署 haproxy_exporter

  Github下载页面:

  https://github.com/prometheus/haproxy_exporter/releases

  解压部署

root@haproxyA:/usr/local\# tar xf haproxy_exporter-0.13.0.linux-amd64.tar.gz
root@haproxyA:/usr/local\# cd haproxy_exporter-0.13.0.linux-amd64/
root@haproxyA:/usr/local/haproxy_exporter-0.13.0.linux-amd64\# cp haproxy_exporter /usr/local/bin/

 

  haproxy_exporter命令帮助

root@haproxyA:~\# haproxy_exporter --help
usage: haproxy_exporter [<flags>]
Flags:
-h, --help Show context-sensitive help (also try --help-long and --help-man).
--web.config.file="" [EXPERIMENTAL] Path to configuration file that can enable TLS or authentication.
--web.listen-address=":9101"
Address to listen on for web interface and telemetry.
--web.telemetry-path="/metrics"
Path under which to expose metrics.
--haproxy.scrape-uri="http://localhost/;csv"
URI on which to scrape HAProxy.
--haproxy.ssl-verify Flag that enables SSL certificate verification for the scrape URI
--haproxy.server-metric-fields="2,3,4,5,6,7,8,9,13,14,15,16,17,18,21,24,30,33,35,38,39,40,41,42,43,44,49,50,58,59,60,61"
Comma-separated list of exported server metrics. See http://cbonte.github.io/haproxy-dconv/configuration-1.5.html#9.1
--haproxy.server-exclude-states=""
Comma-separated list of exported server states to exclude. See https://cbonte.github.io/haproxy-dconv/1.8/management.html#9.1,
field 17 statuus
--haproxy.timeout=5s Timeout for trying to get stats from HAProxy.
--haproxy.pid-file="" Path to HAProxy pid file.
If provided, the standard process metrics get exported for the HAProxy
process, prefixed with 'haproxy_process_...'. The haproxy_process exporter
needs to have read access to files owned by the HAProxy process. Depends on
the availability of /proc.
https://prometheus.io/docs/instrumenting/writing_clientlibs/#process-metrics.
--log.level=info Only log messages with the given severity or above. One of: [debug, info, warn, error]
--log.format=logfmt Output format of log messages. One of: [logfmt, json]
--version Show application version.

 

  haproxy_exporter两种监听的启动方式:

  默认监听9101端口

  1、--haproxy.scrape-uri=unix:/socket_dir,指定haproxy的socket文件监控

haproxy_exporter --haproxy.scrape-uri=unix:/run/haproxy/admin.sock

 

  2、--haproxy.scrape-uri="http://”,指定haproxy的状态页地址

haproxy_exporter --haproxy.scrape-uri="http://admin:admin@127.0.0.1:8009/admin?stats;csv" &

 

  创建haproxy_exporter service启动文件

root@mysql1:~\# vim /etc/systemd/system/haproxy_exporter.service
[Unit]
Description=Prometheus HAproxy Exporter
After=network.target
[Service]
ExecStart=/usr/local/bin/haproxy_exporter --haproxy.scrape-uri=unix:/run/haproxy/admin.sock
[Install]
WantedBy=multi-user.target

 

  浏览器访问验证

 

  prometheus-server配置haproxy_监控

- job_name: "haproxy_exporter"
static_configs:
- targets: ["192.168.100.20:9101"]

 

  prometheus-server web验证

  grafana导入模板并验证

  ID-2428、ID-367

 

 

 

 

 
posted @   PunchLinux  阅读(1167)  评论(0编辑  收藏  举报
相关博文:
阅读排行:
· DeepSeek 开源周回顾「GitHub 热点速览」
· 记一次.NET内存居高不下排查解决与启示
· 物流快递公司核心技术能力-地址解析分单基础技术分享
· .NET 10首个预览版发布:重大改进与新特性概览!
· .NET10 - 预览版1新功能体验(一)
  1. 1 Get Lucky Daft Punnk
  2. 2 烦恼歌 张学友
  3. 3 Way Back Home Conor Maynard
  4. 4 Put A Little Love In Your Heart Jackie DeShannon
  5. 5 菊次郎的夏天 久石譲
  6. 6 一丝不挂 陈奕迅
  7. 7 日落大道 梁博
  8. 8 Rage your dream m.o.v.e
  9. 9 Don't Be So Shy Imany
  10. 10 芒种 腾格尔
菊次郎的夏天 - 久石譲
00:00 / 00:00
An audio error has occurred, player will skip forward in 2 seconds.

Not available

点击右上角即可分享
微信分享提示