prometheus 监控案例-Tomcat、Redis、Mysql、Haproxy、Nginx
服务监控
可以根据prometheus官网找到对应服务的exporter
https://prometheus.io/docs/instrumenting/exporters/
监控tomcat容器
基于第三方 exporter 实现对目的服务的监控
github项目地址:https://github.com/nlighten/tomcat_exporter
所需要的Jar包获取
选择对应版本采集服务程序包,点击下载按钮,下载则是最新版本
如果想下载以往版本解决兼容某些tomcat问题,则点击版本后进入详细版本页面进行下载
根据测试,simpleclient程序包使用0.12.0开始以后版本不对8.5.x 的tomcat兼容,无法显示metrics页面,但tomcat-exporter程序包可以使用最新版本
Jar监控程序包列表:
simpleclient-0.8.0.jar
simpleclient_common-0.8.0.jar
simpleclient_hotspot-0.8.0.jar
simpleclient_servlet-0.8.0.jar
tomcat_exporter_client-0.0.17.jar
tomcat_exporter_servlet-0.0.17.war
并将tomcat_exporter_servlet-0.0.17.war修改为metrics进行tomcat页面发布
# TYPE tomcat_connections_active_total gauge tomcat_connections_active_total{name="http-nio-8080",} 2.0 # TYPE jvm_memory_bytes_used gauge jvm_memory_bytes_used{area="heap",} 2.4451216E7
制作tomcat镜像,将包含metrics监控指标的jar包导入tomcat镜像内
root@deploy:~/dockerfile# cat Dockerfile FROM harbor.cncf.net/web/tomcat:8.5.43 MAINTAINER LXH LABEL Description="tomcat-8.5.43-app1" ADD metrics.war /usr/local/tomcat/webapps/ ADD simpleclient-0.8.0.jar /usr/local/tomcat/lib ADD simpleclient_common-0.8.0.jar /usr/local/tomcat/lib ADD simpleclient_hotspot-0.8.0.jar /usr/local/tomcat/lib ADD simpleclient_servlet-0.8.0.jar /usr/local/tomcat/lib ADD tomcat_exporter_client-0.0.17.jar /usr/local/tomcat/lib EXPOSE 8080 8443
容器镜像构建脚本
root@deploy:~/dockerfile# cat build.sh #!/bin/bash version=$1 docker build -t tomcatapp:${version} ./ docker tag tomcatapp:${version} harbor.cncf.net/web/tomcatapp:${version} docker push harbor.cncf.net/web/tomcatapp:${version}
编写tomcat k8s yaml文件
k8s部署监控相关的组件的时候需要初始化很多linux的内核参数。 但是文件系统挂载到pod容器中就会变成read-only,难以进行操作实现需求。所以需要给POD privileged权限,然后在容器的初始化脚本或代码中去修改sysctl参数。给容器的spec指定`securityContext.privileged=true`参数
prometheus-server如果使用k8s内部部署并配置服务发现,则需要添加prometheus的scrape注解`annotations.prometheus.io/scrape='true'`
root@master1:~/yaml/tomcat-exporter# cat tomcat-deploy.yaml apiVersion: apps/v1 kind: Deployment metadata: name: tomcat-deployment namespace: default spec: selector: matchLabels: app: tomcat replicas: 1 template: metadata: labels: app: tomcat annotations: prometheus.io/scrape: 'true' #添加prometheus服务发现的抓取注解 spec: containers: - name: tomcat image: harbor.cncf.net/web/tomcatapp:1.0 imagePullPolicy: Always ports: - containerPort: 8080 securityContext: privileged: true #给POD privileged权限
部署tomcat server,如果svc也想被prometheus-server服务发现采集,也许要添加prometheus-server服务注解
root@master1:~/yaml/tomcat-exporter# cat tomcat-svc.yaml kind: Service apiVersion: v1 metadata: annotations: prometheus.io/scrape: 'true' #添加prometheus服务发现的抓取注解 name: tomcat-service spec: selector: app: tomcat ports: - nodePort: 31080 port: 80 protocol: TCP targetPort: 8080 type: NodePort
root@master1:~/yaml/tomcat-exporter# kubectl get pods -n default NAME READY STATUS RESTARTS AGE tomcat-deployment-b8c8457cf-fgztx 1/1 Running 0 14s root@master1:~/yaml/tomcat-exporter# kubectl get svc -n default NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE kubernetes ClusterIP 10.100.0.1 <none> 443/TCP 17d
访问tomcat页面和指标
配置prometheus-server配置文件,添加tomcat-exporter
root@prometheus:/usr/local/prometheus# vim prometheus.yml - job_name: "tomcat" static_configs: - targets: ["192.168.100.3:31080"] root@prometheus:/usr/local/prometheus# systemctl restart prometheus
部署导入grafana官方模板
https://github.com/nlighten/tomcat_exporter/tree/master/dashboard
复制example.json导入grafana
监控redis
通过 redis_exporter 监控 redis 服务状态。
prometheus官网 https://prometheus.io/docs/instrumenting/exporters/ ,找到redis_exporter
redis_exporter的GIthub项目地址 https://github.com/oliver006/redis_exporter
下载redis-exporter
https://github.com/oliver006/redis_exporter/releases
[root@redis1 local]# tar xf redis_exporter-v1.43.1.linux-amd64.tar.gz
redis-exporter会默认监听本机localhost:6379端口,来返回指标数据/metrics。启动默认监听9121端口
命令帮助信息
root@redis1 redis_exporter]# ./redis_exporter --help Usage of ./redis_exporter: -check-key-groups string Comma separated list of lua regex for grouping keys -check-keys string Comma separated list of key-patterns to export value and length/size, searched for with SCAN -check-keys-batch-size int Approximate number of keys to process in each execution, larger value speeds up scanning. WARNING: Still Redis is a single-threaded app, huge COUNT can affect production environment. (default 1000) -check-single-keys string Comma separated list of single keys to export value and length/size -check-single-streams string Comma separated list of single streams to export info about streams, groups and consumers -check-streams string Comma separated list of stream-patterns to export info about streams, groups and consumers, searched for with SCAN -config-command string What to use for the CONFIG command (default "CONFIG") -connection-timeout string Timeout for connection to Redis instance (default "15s") -count-keys string Comma separated list of patterns to count (eg: 'db0=production_*,db3=sessions:*'), searched for with SCAN -debug Output verbose debug information -export-client-list Whether to scrape Client List specific metrics -export-client-port Whether to include the client's port when exporting the client list. Warning: including the port increases the number of metrics generated and will make your Prometheus server take up more memory -include-config-metrics Whether to include all config settings as metrics -include-system-metrics Whether to include system metrics like e.g. redis_total_system_memory_bytes -is-cluster Whether this is a redis cluster (Enable this if you need to fetch key level data on a Redis Cluster). -is-tile38 Whether to scrape Tile38 specific metrics -log-format string Log format, valid options are txt and json (default "txt") -max-distinct-key-groups int The maximum number of distinct key groups with the most memory utilization to present as distinct metrics per database, the leftover key groups will be aggregated in the 'overflow' bucket (default 100) -namespace string Namespace for metrics (default "redis") -ping-on-connect Whether to ping the redis instance after connecting -redact-config-metrics Whether to redact config settings that include potentially sensitive information like passwords (default true) -redis-only-metrics Whether to also export go runtime metrics -redis.addr string Address of the Redis instance to scrape (default "redis://localhost:6379") -redis.password string Password of the Redis instance to scrape -redis.password-file string Password file of the Redis instance to scrape -redis.user string User name to use for authentication (Redis ACL for Redis 6.0 and newer) -script string Path to Lua Redis script for collecting extra metrics -set-client-name Whether to set client name to redis_exporter (default true) -skip-tls-verification Whether to to skip TLS verification -tls-ca-cert-file string Name of the CA certificate file (including full path) if the server requires TLS client authentication -tls-client-cert-file string Name of the client certificate file (including full path) if the server requires TLS client authentication -tls-client-key-file string Name of the client key file (including full path) if the server requires TLS client authentication -tls-server-ca-cert-file string Name of the CA certificate file (including full path) if the web interface and telemetry should require TLS client authentication -tls-server-cert-file string Name of the server certificate file (including full path) if the web interface and telemetry should use TLS -tls-server-key-file string Name of the server key file (including full path) if the web interface and telemetry should use TLS -tls-server-min-version string Minimum TLS version that is acceptable by the web interface and telemetry when using TLS (default "TLS1.2") -version Show version information and exit -web.listen-address string Address to listen on for web interface and telemetry. (default ":9121") -web.telemetry-path string Path under which to expose metrics. (default "/metrics")
常用选项:
-redis.addr: 连接redis地址:端口
-redis.password: 连接redis的密码
-web.listen-address: 指定redis_exporter的服务端口,格式为-web.listen-address ":9121
启动脚本
使redis_exporter监听在自定义端口9333,默认端口监听为9121
[root@redis1 redis_exporter]# cat start.sh #!/bin/bash cmd=/usr/local/redis_exporter ip=192.168.100.104 #reidis地址 password=123456 #redis密码 nohup $cmd/redis_exporter -redis.addr $ip:6379 -redis.password $password -web.listen-address ":9333" > /dev/null 2>&1 & [ $? -eq 0 ] && echo "start success!" || echo "start failed"
root@redis1 redis_exporter]# cat stop.sh #!/bin/bash PID=$(ps -ef|grep redis_exporter|grep -v grep|awk '{print $2}') if [ -z ${PID} ];then echo "redis_expoter isn't running" else kill $PID sleep 2s [ ! -z ${PID} ] echo "redis_expoter is stopped" fi
监控redis单实例
编辑启动脚本后,运行启动脚本
[root@redis1 redis_exporter]# cat start.sh #!/bin/bash cmd=/usr/local/redis_exporter ip=192.168.100.104 #reidis地址 password=123456 #redis密码 nohup $cmd/redis_exporter -redis.addr $ip:6379 -redis.password $password -web.listen-address ":9333" > /dev/null 2>&1 & [ $? -eq 0 ] && echo "start success!" || echo "start failed" [root@redis1 redis_exporter]# ./start.sh start success!
prometheus-server配置
scrape_configs: - job_name: redis_exporter static_configs: - targets: ['192.168.100.104:9333']
监控redis多实例集群
查看集群信息
分别为192.168.100.104上的6379-6384端口的6个redis为集群
官网示例:
https://github.com/oliver006/redis_exporter
Using below config instead of the /metric endpoint the /scrape endpoint will be used by prometheus. As an example the first target will be queried with this web request:
http://exporterhost:9121/scrape?target=first-redis-host:6379
解释
1、创建一个名称为redis_exporter_targets job,来获取每个redis节点的实例。
将采集redis_expoter Http URI修改为/scrape,原因是默认prometheus会收集目标target的/metrics URI,所以需要在prometheus-server配置文件中指定metrics_path 的路径,并进行redis_exporter的标签重写,将\__address\__ 标签内容为`redis实例ip:6379、6380、6381`节点实例,替换为替换为instance标签,并将value修改为`redis_exporter http`的访问地址。最终prometheus-server就能获得每个redis的target:`http://exporterhost:9121/scrape?target=first-redis-host:6379`
2、还需要单独添加一个名称为redis_exporter job来抓取redis_exporter的地址
scrape_configs: ## config for the multiple Redis targets that the exporter will scrape - job_name: 'redis_exporter_targets' static_configs: - targets: - redis://first-redis-host:6379 - redis://second-redis-host:6379 - redis://second-redis-host:6380 - redis://second-redis-host:6381 metrics_path: /scrape relabel_configs: - source_labels: [__address__] target_label: __param_target - source_labels: [__param_target] target_label: instance - target_label: __address__ replacement: <<REDIS-EXPORTER-HOSTNAME>>:9121 ## config for scraping the exporter itself - job_name: 'redis_exporter' static_configs: - targets: - <<REDIS-EXPORTER-HOSTNAME>>:9121
redis_expoter配置
redis_expoter只要能连接到一个集群的一个节点,因此只需要配置连接集群中一个redis节点就能查询其他节点的指标
[root@redis1 redis_exporter]\# cat start.sh #!/bin/bash cmd=/usr/local/redis_exporter ip=192.168.100.104 password=123456 nohup $cmd/redis_exporter -redis.addr $ip:6379 -redis.password $password -web.listen-address ":9333" > /dev/null 2>&1 & [ $? -eq 0 ] && echo "start success!" || echo "start failed"
prometheus-server配置
root@prometheus:~\# cat /usr/local/prometheus/prometheus.yml # my global config global: scrape_interval: 15s # Set the scrape interval to every 15 seconds. Default is every 1 minute. evaluation_interval: 15s # Evaluate rules every 15 seconds. The default is every 1 minute. # scrape_timeout is set to the global default (10s). alerting: alertmanagers: - static_configs: - targets: # - alertmanager:9093 rule_files: # - "first_rules.yml" # - "second_rules.yml" scrape_configs: - job_name: 'redis_exporter_targets' static_configs: - targets: - redis://192.168.100.104:6379 - redis://192.168.100.104:6380 - redis://192.168.100.104:6381 - redis://192.168.100.104:6382 - redis://192.168.100.104:6383 - redis://192.168.100.104:6384 metrics_path: /scrape relabel_configs: - source_labels: [__address__] target_label: __param_target - source_labels: [__param_target] target_label: instance - target_label: __address__ replacement: 192.168.100.104:9333 ## config for scraping the exporter itself - job_name: 'redis_exporter' static_configs: - targets: - 192.168.100.104:9333
将标签重写后,URI修改为/scripe?target=redis节点:6379
配置grafana导入redis-exporter模板
ID:11835
监控 k8s redis集群
这里每个redis节点使用statefulset部署,并暴露集群端口供redis_exporter使用。redis_expoter单独使用deployment部署
部署redis集群参考kubernetes 案例redis集群配置
验证redis集群
查看redis集群pod
查看集群状态
redis_exporter镜像构建
root@deploy:~/dockerfile/redis-exporter\# cat Dockerfile FROM harbor.cncf.net/os/ubuntu:20.04 ADD redis_exporter.tar.gz /usr/local/ ENTRYPOINT ["/usr/local/redis_exporter/start.sh"]
root@deploy:~/dockerfile/redis-exporter\# ll redis_exporter/ total 8892 drwxr-xr-x 2 root root 91 Sep 24 03:29 ./ drwxr-xr-x 3 root root 91 Sep 24 03:30 ../ -rw-r--r-- 1 root root 1063 Sep 24 03:10 LICENSE -rw-r--r-- 1 root root 43382 Sep 24 03:10 README.md -rwxr-xr-x 1 root root 9048064 Sep 24 03:10 redis_exporter* -rwx------ 1 root root 202 Sep 24 03:29 start.sh* -rwx------ 1 root root 224 Sep 24 03:10 stop.sh*
准备redis_exporter启动脚本
redis的ip地址修改为k8s中redis集群中的master节点的pod服务名称,redis-0.redis.redis-test.svc.cluster.local
root@deploy:~/dockerfile/redis-exporter\# cat redis_exporter/start.sh #!/bin/bash cmd=/usr/local/redis_exporter ip="redis-0.redis.redis-test.svc.cluster.local" password=123456 $cmd/redis_exporter -redis.addr $ip:6379 -redis.password $password -web.listen-address ":9121"
准备镜像构建脚本
root@deploy:~/dockerfile/redis-exporter\# cat build.sh #!/bin/bash version=$1 docker build -t harbor.cncf.net/project/redis_exporter:$version ./ docker push harbor.cncf.net/project/redis_exporter:$version
部署redis_exporter
创建deployment,与redis集群在同一个ns下
root@master1:~/yaml/redis-exporter# cat redis-exporter.yaml apiVersion: apps/v1 kind: Deployment metadata: name: redis-exporter namespace: redis-test spec: replicas: 1 selector: matchLabels: app: redis-exporter template: metadata: labels: app: redis-exporter spec: containers: - name: redis-exporter image: harbor.cncf.net/project/redis_exporter:1.7 imagePullPolicy: Always resources: requests: cpu: 100m memory: 100Mi ports: - containerPort: 9121
创建svc
root@master1:~/yaml/redis-exporter# cat redis-exporter-svc.yaml kind: Service #service 类型 apiVersion: v1 metadata: name: redis-redis-service namespace: redis-test spec: selector: app: redis-exporter ports: - nodePort: 39121 name: redis-exporter port: 9121 protocol: TCP targetPort: 9121 type: NodePort
验证部署
root@master1:~/yaml\# kubectl get svc -n redis-test redis-redis-service NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE redis-redis-service NodePort 10.100.135.230 <none> 9121:39121/TCP 99m root@master1:~/yaml\# kubectl get pods -n redis-test NAME READY STATUS RESTARTS AGE redis-0 1/1 Running 0 134m redis-1 1/1 Running 0 142m redis-2 1/1 Running 0 142m redis-3 1/1 Running 0 142m redis-4 1/1 Running 0 142m redis-5 1/1 Running 0 126m redis-exporter-bf6bf8576-k9p4x 1/1 Running 0 101m
验证查看redis_exporter web界面
配置k8s中的prometheus-server,添加redis-exporter服务发现
target为redis-exporter的svc名称
- job_name: 'redis_exporter' static_configs: - targets: - redis-redis-service.redis-test:9121
配置后重新应用promentheus configmap并删除prometheus-server pod重启
root@master1:~/yaml\# kubectl apply -f prometheus-cfg.yaml configmap/prometheus-config configured root@master1:~/yaml\# kubectl delete -f prometheus-deployment.yaml deployment.apps "prometheus-server" deleted root@master1:~/yaml\# kubectl apply -f prometheus-deployment.yaml
查看prometheus target界面验证target
查看grafana导入dashboard
监控mysql
通过 mysqld_exporter 监控 MySQL 服务的运行状态
官方exporter:
prometheus官网: https://prometheus.io/docs/instrumenting/exporters/,找到mysql_exporer
Github项目地址:https://github.com/prometheus/mysqld_exporter
监控mysql单示例
1、确保监控的mysql实例正常运行,添加授权监控账号权限。
已经安装mysql服务,端口监听在3307
CREATE USER 'exporter'@'localhost' IDENTIFIED BY '123456' WITH MAX_USER_CONNECTIONS 3; GRANT PROCESS, REPLICATION CLIENT, SELECT ON *.* TO 'exporter'@'localhost';
测试账号权限
root@mysql1:~\# mysql -u exporter -p #输入密码 mysql> select user(); +--------------------+ | user() | +--------------------+ | exporter@localhost | +--------------------+ 1 row in set (0.00 sec)
下载地址:https://github.com/prometheus/mysqld_exporter/releases
解压安装
root@mysql1:/usr/local\# tar xf mysqld_exporter-0.14.0.linux-amd64.tar.gz root@mysql1:/usr/local\# ln -s mysqld_exporter-0.14.0.linux-amd64 mysql_exporter
root@mysql1:~\# vim /root/.my.cnf [client] host=localhost port=3307 user = exporter password = 123456
测试权限,直接本地执行mysql命令登录mysql
root@mysql1:~\# mysql Welcome to the MySQL monitor. Commands end with ; or \g. Your MySQL connection id is 5 Server version: 5.7.36 MySQL Community Server (GPL) Copyright (c) 2000, 2021, Oracle and/or its affiliates. Oracle is a registered trademark of Oracle Corporation and/or its affiliates. Other names may be trademarks of their respective owners. Type 'help;' or '\h' for help. Type '\c' to clear the current input statement. mysql>
3、创建mysql_exporter服务启动文件
设置端口监听在本机出口网卡的ip:9105,并指定mysql的客户端配置文件。默认端口为9104
root@mysql1:~\# vim /etc/systemd/system/mysqld_exporter.service [Unit] Description=Prometheus Node Exporter After=network.target [Service] ExecStart=/usr/local/bin/mysqld_exporter --config.my-cnf=/root/.my.cnf --web.listen-address=192.168.0.29:9105 [Install] WantedBy=multi-user.target
mysqld_exporter的主要参数介绍
Name | Description |
---|---|
mysqld.address | Hostname and port used for connecting to MySQL server, format: host:port . (default: locahost:3306 ) |
mysqld.username | Username to be used for connecting to MySQL Server |
config.my-cnf | Path to .my.cnf file to read MySQL credentials from. (default: ~/.my.cnf ) |
log.level | Logging verbosity (default: info) |
exporter.lock_wait_timeout | Set a lock_wait_timeout (in seconds) on the connection to avoid long metadata locking. (default: 2) |
exporter.log_slow_filter | Add a log_slow_filter to avoid slow query logging of scrapes. NOTE: Not supported by Oracle MySQL. |
tls.insecure-skip-verify | Ignore tls verification errors. |
web.config.file | Path to a |
web.listen-address | Address to listen on for web interface and telemetry. |
web.telemetry-path | Path under which to expose metrics. |
version | Print the version information. |
命令帮助信息
root@mysql1:/usr/local\# mysqld_exporter --help usage: mysqld_exporter [<flags>] Flags: -h, --help Show context-sensitive help (also try --help-long and --help-man). --exporter.lock_wait_timeout=2 Set a lock_wait_timeout (in seconds) on the connection to avoid long metadata locking. --exporter.log_slow_filter Add a log_slow_filter to avoid slow query logging of scrapes. NOTE: Not supported by Oracle MySQL. --collect.heartbeat.database="heartbeat" Database from where to collect heartbeat data --collect.heartbeat.table="heartbeat" Table from where to collect heartbeat data --collect.heartbeat.utc Use UTC for timestamps of the current server (`pt-heartbeat` is called with `--utc`) --collect.info_schema.processlist.min_time=0 Minimum time a thread must be in each state to be counted --collect.info_schema.processlist.processes_by_user Enable collecting the number of processes by user --collect.info_schema.processlist.processes_by_host Enable collecting the number of processes by host --collect.info_schema.tables.databases="*" The list of databases to collect table stats for, or '*' for all --collect.mysql.user.privileges Enable collecting user privileges from mysql.user --collect.perf_schema.eventsstatements.limit=250 Limit the number of events statements digests by response time --collect.perf_schema.eventsstatements.timelimit=86400 Limit how old the 'last_seen' events statements can be, in seconds --collect.perf_schema.eventsstatements.digest_text_limit=120 Maximum length of the normalized statement text --collect.perf_schema.file_instances.filter=".*" RegEx file_name filter for performance_schema.file_summary_by_instance --collect.perf_schema.file_instances.remove_prefix="/var/lib/mysql/" Remove path prefix in performance_schema.file_summary_by_instance --collect.perf_schema.memory_events.remove_prefix="memory/" Remove instrument prefix in performance_schema.memory_summary_global_by_event_name --web.config.file="" [EXPERIMENTAL] Path to configuration file that can enable TLS or authentication. --web.listen-address=":9104" Address to listen on for web interface and telemetry. --web.telemetry-path="/metrics" Path under which to expose metrics. --timeout-offset=0.25 Offset to subtract from timeout in seconds. --config.my-cnf="/root/.my.cnf" Path to .my.cnf file to read MySQL credentials from. --tls.insecure-skip-verify Ignore certificate and server verification when using a tls connection. --collect.mysql.user Collect data from mysql.user --collect.info_schema.tables Collect metrics from information_schema.tables --collect.info_schema.innodb_tablespaces Collect metrics from information_schema.innodb_sys_tablespaces --collect.info_schema.innodb_metrics Collect metrics from information_schema.innodb_metrics --collect.global_status Collect from SHOW GLOBAL STATUS --collect.global_variables Collect from SHOW GLOBAL VARIABLES --collect.slave_status Collect from SHOW SLAVE STATUS --collect.info_schema.processlist Collect current thread state counts from the information_schema.processlist --collect.perf_schema.tablelocks Collect metrics from performance_schema.table_lock_waits_summary_by_table --collect.perf_schema.eventsstatements Collect metrics from performance_schema.events_statements_summary_by_digest --collect.perf_schema.eventsstatementssum Collect metrics of grand sums from performance_schema.events_statements_summary_by_digest --collect.perf_schema.eventswaits Collect metrics from performance_schema.events_waits_summary_global_by_event_name --collect.auto_increment.columns Collect auto_increment columns and max values from information_schema --collect.binlog_size Collect the current size of all registered binlog files --collect.perf_schema.tableiowaits Collect metrics from performance_schema.table_io_waits_summary_by_table --collect.perf_schema.indexiowaits Collect metrics from performance_schema.table_io_waits_summary_by_index_usage --collect.perf_schema.replication_group_member_stats Collect metrics from performance_schema.replication_group_member_stats --collect.perf_schema.replication_applier_status_by_worker Collect metrics from performance_schema.replication_applier_status_by_worker --collect.info_schema.userstats If running with userstat=1, set to true to collect user statistics --collect.info_schema.clientstats If running with userstat=1, set to true to collect client statistics --collect.perf_schema.file_events Collect metrics from performance_schema.file_summary_by_event_name --collect.perf_schema.file_instances Collect metrics from performance_schema.file_summary_by_instance --collect.perf_schema.memory_events Collect metrics from performance_schema.memory_summary_global_by_event_name --collect.perf_schema.replication_group_members Collect metrics from performance_schema.replication_group_members --collect.info_schema.query_response_time Collect query response time distribution if query_response_time_stats is ON. --collect.engine_tokudb_status Collect from SHOW ENGINE TOKUDB STATUS --collect.engine_innodb_status Collect from SHOW ENGINE INNODB STATUS --collect.heartbeat Collect from heartbeat --collect.info_schema.tablestats If running with userstat=1, set to true to collect table statistics --collect.info_schema.schemastats If running with userstat=1, set to true to collect schema statistics --collect.info_schema.innodb_cmp Collect metrics from information_schema.innodb_cmp --collect.info_schema.innodb_cmpmem Collect metrics from information_schema.innodb_cmpmem --collect.slave_hosts Scrape information from 'SHOW SLAVE HOSTS' --collect.info_schema.replica_host Collect metrics from information_schema.replica_host_status --log.level=info Only log messages with the given severity or above. One of: [debug, info, warn, error] --log.format=logfmt Output format of log messages. One of: [logfmt, json] --version Show application version.
systemctl daemon-reload && systemctl enable mysqld_exporter && systemctl start mysqld_exporter
访问URI /metrics
4、配置prometheus-server配置文件,添加mysqld_exporter服务采集
scrape_configs: - job_name: "mysqld_exporter" static_configs: - targets: ["192.168.0.29:9105"]
浏览器访问prometheus-server/targets验证
5、配置grafana添加myqld_exporter的dashboard
ID-13106、ID-11323
监控Haproxy
通过 haproxy_exporter 监控 haproxy
prometheus官网:https://prometheus.io/docs/instrumenting/exporters/,找到haproxy_exporer
Github项目地址:https://github.com/prometheus/haproxy_exporter
haproxy_exporter的监控方式
1、通过haproxy 的状态页URL进行监控
Github示例:
2、通过haproxy socket文件进行监控
配置haproxy 添加状态页与添加socket文件
global stats socket /run/haproxy/admin.sock mode 660 level admin expose-fd listeners listen statistics #自定义监听端口 bind *:8009 #启用基于程序编译时默认设置的统计报告 stats enable #统计页面用户名和密码设置 stats auth admin:admin #自定义统计页面的URL,默认为/haproxy?stats stats uri /admin?stats #隐藏统计页面上HAProxy的版本信息 stats hide-version #统计页面自动刷新时间 stats refresh 30s #如果认证通过就做管理功能,可以管理后端的服务器 stats admin if TRUE #统计页面密码框上提示文本,默认为Haproxy\ Statistics stats realm Hapadmin
浏览器访问验证:
haproxy 状态页URI:/admin?stats
验证haproxy服务器本地socket文件
root@haproxyA:~\# ll /run/haproxy/admin.sock srw-rw---- 1 root haproxy 0 Sep 26 08:44 /run/haproxy/admin.sock=
部署 haproxy_exporter
Github下载页面:
https://github.com/prometheus/haproxy_exporter/releases
解压部署
root@haproxyA:/usr/local\# tar xf haproxy_exporter-0.13.0.linux-amd64.tar.gz root@haproxyA:/usr/local\# cd haproxy_exporter-0.13.0.linux-amd64/ root@haproxyA:/usr/local/haproxy_exporter-0.13.0.linux-amd64\# cp haproxy_exporter /usr/local/bin/
root@haproxyA:~\# haproxy_exporter --help usage: haproxy_exporter [<flags>] Flags: -h, --help Show context-sensitive help (also try --help-long and --help-man). --web.config.file="" [EXPERIMENTAL] Path to configuration file that can enable TLS or authentication. --web.listen-address=":9101" Address to listen on for web interface and telemetry. --web.telemetry-path="/metrics" Path under which to expose metrics. --haproxy.scrape-uri="http://localhost/;csv" URI on which to scrape HAProxy. --haproxy.ssl-verify Flag that enables SSL certificate verification for the scrape URI --haproxy.server-metric-fields="2,3,4,5,6,7,8,9,13,14,15,16,17,18,21,24,30,33,35,38,39,40,41,42,43,44,49,50,58,59,60,61" Comma-separated list of exported server metrics. See http://cbonte.github.io/haproxy-dconv/configuration-1.5.html#9.1 --haproxy.server-exclude-states="" Comma-separated list of exported server states to exclude. See https://cbonte.github.io/haproxy-dconv/1.8/management.html#9.1, field 17 statuus --haproxy.timeout=5s Timeout for trying to get stats from HAProxy. --haproxy.pid-file="" Path to HAProxy pid file. If provided, the standard process metrics get exported for the HAProxy process, prefixed with 'haproxy_process_...'. The haproxy_process exporter needs to have read access to files owned by the HAProxy process. Depends on the availability of /proc. https://prometheus.io/docs/instrumenting/writing_clientlibs/#process-metrics. --log.level=info Only log messages with the given severity or above. One of: [debug, info, warn, error] --log.format=logfmt Output format of log messages. One of: [logfmt, json] --version Show application version.
haproxy_exporter两种监听的启动方式:
默认监听9101端口
1、--haproxy.scrape-uri=unix:/socket_dir,指定haproxy的socket文件监控
haproxy_exporter --haproxy.scrape-uri=unix:/run/haproxy/admin.sock
2、--haproxy.scrape-uri="http://”,指定haproxy的状态页地址
haproxy_exporter --haproxy.scrape-uri="http://admin:admin@127.0.0.1:8009/admin?stats;csv" &
创建haproxy_exporter service启动文件
root@mysql1:~\# vim /etc/systemd/system/haproxy_exporter.service [Unit] Description=Prometheus HAproxy Exporter After=network.target [Service] ExecStart=/usr/local/bin/haproxy_exporter --haproxy.scrape-uri=unix:/run/haproxy/admin.sock [Install] WantedBy=multi-user.target
浏览器访问验证
prometheus-server配置haproxy_监控
- job_name: "haproxy_exporter" static_configs: - targets: ["192.168.100.20:9101"]
prometheus-server web验证
grafana导入模板并验证
ID-2428、ID-367
本文来自博客园,作者:PunchLinux,转载请注明原文链接:https://www.cnblogs.com/punchlinux/p/16856793.html
【推荐】国内首个AI IDE,深度理解中文开发场景,立即下载体验Trae
【推荐】编程新体验,更懂你的AI,立即体验豆包MarsCode编程助手
【推荐】抖音旗下AI助手豆包,你的智能百科全书,全免费不限次数
【推荐】轻量又高性能的 SSH 工具 IShell:AI 加持,快人一步
· DeepSeek 开源周回顾「GitHub 热点速览」
· 记一次.NET内存居高不下排查解决与启示
· 物流快递公司核心技术能力-地址解析分单基础技术分享
· .NET 10首个预览版发布:重大改进与新特性概览!
· .NET10 - 预览版1新功能体验(一)