欢迎您访问程序员文章站本站旨在为大家提供分享程序员计算机编程知识!
您现在的位置是: 首页  >  php教程

安装MySQL慢查询日志工具Anemometer

程序员文章站 2022-03-24 11:47:58
...

首先安装LNMP环境,要求PHP-5.3以上版本. 参考:http://isadba.com/?p=82 或者参考http://isadba.com/?p=572 然后下载Anemometer git clonehttps://github.com/box

首先安装LNMP环境,要求PHP-5.3以上版本.
参考:?p=82 或者参考 ?p=572

然后下载Anemometer
git clone https://github.com/box/Anemometer.git anemometer

配置LNMP将下载下来的anemometer部署到LAMP上面,可以通过web打开页面,会提示你没有配置网站.
mysql -uroot -p mysql -uroot -p

安装percona的toolkit工具
yum install
yum install percona-toolkit -y

修改配置文件
cd /var/www/anemometer/conf/
cp sample.config.inc.php config.inc.php
修改35-40行的数据库配置,连接到anemometer的数据库,当然,你数据库也需要授权.

#使用pt-query-digest工具,将一些慢查询数据导入数据库
pt-query-digest –user=anemometer –password=anemometerpass –review h=192.168.11.28,D=slow_query_log,t=global_query_review \
–history h=192.168.11.28,D=slow_query_log,t=global_query_review_history \
–no-report –limit=0% –filter=” \$event->{Bytes} = length(\$event->{arg}) and \$event->{hostname}=\”$HOSTNAME\”" \
/usr/local/mariamysql/data/localhost-slow.log

执行命令以后,h=192.168.11.28,D=slow_query_log,t=global_query_review_history和global_query_review表已经已经有一些数据了,你可以尝试手工查询.

现在可以再通过web访问anemometer,就能看到相关信息了.

到这一步,anemometer算是跑通了,下一步就是真正的部署使用.

现在的环境是,有多台MySQL服务器的slow log需要监控.
我们有三种方案.
1、手工导入的install.sql里面的global_query_review_history表包含了hostname_max和db_max,通过hostname_max区分把多个数据源存放到一个表里.
2、每台mysql把处理后的binlog数据放到当前服务器,然后anemometer连接对应的服务器获取数据.
3、所有的mysql把处理后的binlog数据统一存放到anemometer所在的数据库上,然后通过表名或者数据库名字区分.

如果选择第二种或者第三种方案,需要定义多个数据源,并且anemometer_collect.sh.里面对应的history_db_name要对应.
在conf/config.inc.php总修改数据源的配置,注意可以不同的数据源指向的是不同的db,一般可以按照项目区分.
在conf/config.inc.php总修改数据源的配置,注意不同的数据源指向的是不同的db.

$conf['datasources']['192.168.11.28'] = array( 'host' => '192.168.11.28', 'port' => 3306, 'db' => 'slow_query_log', 'user' => 'anemometer', 'password' => 'anemometerpass', 'tables' => array( 'global_query_review' => 'fact', 'global_query_review_history' => 'dimension' ), 'source_type' => 'slow_query_log' ); $conf['datasources']['192.168.11.17'] = array( 'host' => '192.168.11.28', 'port' => 3306, 'db' => 'slow_query_log_192_168_11_17', 'user' => 'anemometer', 'password' => 'anemometerpass', 'tables' => array( 'global_query_review' => 'fact', 'global_query_review_history' => 'dimension' ), 'source_type' => 'slow_query_log' );

在所有MySQL上部署scripts/anemometer_collect.sh收集慢查询日志.
anemometer_collect.sh脚本有几个注意点
1、PATH里面是否有mysql的bin路径
2、如果history_db_name需要根据不同机器做修改,和config.inc.php定义的对应
3、注意原始慢查询日志路径是否正确
4、通过设置long_query_time设置慢查询的时长
5、所有MySQL服务器上需要安装percona-toolkit的工具箱.

并且在anemometer_collect.sh当前目录下增加下面两个配置文件.
连接到本地MySQL服务器的账号,需要有super权限.

[root@localhost scripts]# cat anemometer-localhost.cnf [client] user=root password= host=localhost socket=/tmp/mysql.sock

连接到anemometer服务器数据库的账号,需要对anemometer_collect.sh脚本中history_db_name对应的数据库有all privileges权限.

[root@localhost scripts]# cat anemometer-server.cnf [client] user=anemometer password=anemometerpass

使用下面命令调试脚本.

sh -x ./anemometer_collect.sh --interval 30 --history-db-host 192.168.11.17 --defaults-file ./anemometer-localhost.cnf --history-defaults-file ./anemometer-server.cnf

调试通过以后,在crontab添加如下命令,实现定期采集慢查询日志到数据库存储.间隔可以根据自己的需求设置,我的设置基本上是采集所有的慢查询日志.
*/1 * * * * /opt/anemometer_collect.sh –interval 59 –history-db-host 192.168.11.28 –defaults-file /opt/anemometer-localhost.cnf –history-defaults-file /opt/anemometer-server.cnf

由于anemometer不能跟踪SQL是哪个账号,哪个主机的请求,所以我们有必要保存原始日志,来做例外分析.
下面是我修改后的anemometer_collect.sh,增加了可定制的慢查询问题,mysql path,慢查询时长变量设置,保存历史的慢查询日志.

[root@localhost scripts]# cat anemometer_collect.sh #/usr/bin/env bash # anemometer collection script to gather and digest slow query logs # this is a quick draft script so please give feedback! # # basic usage would be to add this to cron like this: # */5 * * * * anemometer_collect.sh --interval 15 --history-db-host anemometer-db.example.com # # This will have to run as a user which has write privileges to the mysql slow log # # Additionally there are two sets of permissions to worry about: The local mysql instance, and the remote digest storage instance # These are handled through defaults files, just create a file in the: my.cnf format such as: # [client] # user= # password= # # use --defaults-file for permissions to the local mysql instance # and use --history-defaults-file for permissions to the remote digest storage instance # # PATH=/usr/local/mysql-6/bin/:$PATH socket= defaults_file= rate_limit= mysqlopts= interval=30 digest='/usr/bin/pt-query-digest' #set log prefix LOG_PREFIX='/usr/local/mariamysql/data/' #set slow log history file LOG_HISTORY_FILE='/var/log/slow_query_log_history_3306' long_query_time=0.1 #SET IP HOSTNAME=`/sbin/ifconfig | grep 'inet addr' | egrep '172.|192.' | awk '{print $2}' | awk -F ":" '{print $2}'` PORT=3306 HOSTNAME="$HOSTNAME\:$PORT" history_db_host= history_db_port=3306 history_db_name='slow_query_log' history_defaults_file= help () { cat &2 "Invalid argument: $1" ;; esac shift done if [ ! -e "${digest}" ]; then echo "Error: cannot find digest script at: ${digest}" exit 1 fi if [ ! -z "${defaults_file}" ]; then mysqlopts="--defaults-file=${defaults_file}" fi # path to the slow query log LOG=$( mysql $mysqlopts -e " show global variables like 'slow_query_log_file'" -B | tail -n1 | awk '{ print $2 }' ) LOG="$LOG_PREFIX$LOG" if [ $? -ne 0 ]; then echo "Error getting slow log file location" exit 1 fi echo "Collecting from slow query log file: ${LOG}" # simple 30 second collection if [ ! -z "${rate}" ]; then mysql $mysqlopts -e "SET GLOBAL log_slow_rate_limit=${rate}" fi mysql $mysqlopts -e "SET GLOBAL long_query_time=$long_query_time" mysql $mysqlopts -e "SET GLOBAL slow_query_log=1" if [ $? -ne 0 ]; then echo "Error: cannot enable slow log. Aborting" exit 1 fi echo "Slow log enabled; sleeping for ${interval} seconds" sleep "${interval}" mysql $mysqlopts -e "SET GLOBAL slow_query_log=0" echo "Done. Processing log and saving to ${history_db_host}:${history_db_port}/${history_db_name}" # process the log if [[ ! -e "$LOG" ]] then echo "No slow log to process"; exit fi mv "$LOG" /tmp/tmp_slow_log if [ ! -z "${history_defaults_file}" ]; then pass_opt="--defaults-file=${history_defaults_file}" fi "${digest}" $pass_opt \ --review h="${history_db_host}",D="$history_db_name",t=global_query_review \ --history h="${history_db_host}",D="$history_db_name",t=global_query_review_history \ --no-report --limit=0\% \ --filter="\$event->{Bytes} = length(\$event->{arg}) and \$event->{hostname}=\"$HOSTNAME\" " \ "/tmp/tmp_slow_log" #store history slow_log cat /tmp/tmp_slow_log >> $LOG_HISTORY_FILE ,