【MySQL】一条SQL使磁盘暴涨并导致MySQL Crash

收到一个MySQL实例的磁盘告警,看到监控图,磁盘使用率增长速度非常快,在两个多小时的时间里,已经涨了170多G。
检查到binlog目录并不大,而datadir已经180多G了,发现ibtmp1非常大,并且持续增长

赶紧上去看processlist,线程不多,检查到有一个SELECT case ... when ... 的线程,状态是sending data,并且已经持续几千秒了。
有点懵,没有第一时间kill掉这个线程,再一次show processlist的时候……发现大概已经挂了………………


  1. mysql> show processlist;
  2. ERROR 2006 (HY000): MySQL server has gone away
  3. No connection. Trying to reconnect...
  4. ERROR 2002 (HY000): Can't connect to local MySQL server through socket '$datadir/mysqld.sock' (2)
  5. ERROR: Can't connect to the server


那好,既然已经挂了,那么开始找原因吧…

看一下datadir挂载的磁盘,已经释放空了,果然已经被重启了。
ext4    197G  4.6G  191G   3% $datadir


检查一下mysqld进程,发现已经被mysqld_safe拉起来了。

检查一下error log:
红字是关于mysqld crash的信息,很显然,datadir挂载的磁盘满了。
没有多余的空间写binlog和ibtmp1临时表空间文件(5.7新增)。
而且ibtmp1文件最后达到了201876045824bytes,将近190G,而挂载的磁盘总大小才不到200G。

蓝字是关于sql thread的问题,在mysqld起来之后,sql thread也出现了问题。
这个之后再修复

  1. [ERROR] InnoDB: posix_fallocate(): Failed to preallocate data for file $datadir/ibtmp1, desired size 67108864 bytes. Operating system error number 28. Check that the disk is not full or a disk quota exceeded. Make sure the file system supports this function. Some operating system error numbers are described at http://dev.mysql.com/doc/refman/5.7/en/ operating-system-error-codes.html
  2. [ERROR] Disk is full writing '$datadir/mysql-bin.000004' (Errcode: 15870576 - No space left on device). Waiting for someone to free space...
  3. [ERROR] Retry in 60 secs. Message reprinted in 600 secs
  4. [Warning] InnoDB: 1048576 bytes should have been written. Only 647168 bytes written. Retrying for the remaining bytes.
  5. [Warning] InnoDB: Retry attempts for writing partial data failed.
  6. [ERROR] InnoDB: Write to file $datadir/ibtmp1 failed at offset 201911697408, 1048576 bytes should have been written, only 647168 were written. Operating system error number 28. Check that your OS and file system support files of this size. Check also that the disk is not full or a disk quota exceeded.
  7. [ERROR] InnoDB: Error number 28 means 'No space left on device'
  8. [Note] InnoDB: Some operating system error numbers are described at http://dev.mysql.com/doc/refman/5.7/en/operating-system-error-codes.html
  9. [Warning] InnoDB: Error while writing 67108864 zeroes to $datadir/ibtmp1 starting at offset 201876045824
  10. [ERROR] $basedir/bin/mysqld: The table '$tmpdir/#sql_37c5_0' is full
  11. [ERROR] InnoDB: posix_fallocate(): Failed to preallocate data for file ./thread_quartz/QRTZ_FIRED_TRIGGERS.ibd, desired size 32768 bytes. Operating system error number 28. Check that the disk is not full or a disk quota exceeded. Make sure the file system supports this function. Some operating system error numbers are described at http://dev.mysql.com/doc/refman/5.7/en/ operating-system-error-codes.html
  12. [Warning] InnoDB: Retry attempts for writing partial data failed.
  13. [Warning] InnoDB: Error while writing 32768 zeroes to ./thread_quartz/QRTZ_FIRED_TRIGGERS.ibd starting at offset 442362017-07-06T11:49:21.893377Z mysqld_safe Number of processes running now: 0
  14. mysqld_safe mysqld restarted
  15. ……………………………………………………………………………………………………………………
  16. [Note] InnoDB: Last MySQL binlog file position 0 690908428, file name mysql-bin.000004
  17. [Note] InnoDB: Starting in background the rollback of uncommitted transactions
  18. [Note] InnoDB: Rollback of non-prepared transactions completed
  19. [Note] InnoDB: Removed temporary tablespace data file: "ibtmp1"
  20. [Note] InnoDB: Creating shared tablespace for temporary tables
  21. [Note] InnoDB: Setting file '$datadir/ibtmp1' size to 12 MB. Physically writing the file full; Please wait ...
  22. [Note] InnoDB: File '$datadir/ibtmp1' size is now 12 MB.
  23. [Note] InnoDB: 96 redo rollback segment(s) found. 96 redo rollback segment(s) are active.
  24. [Note] InnoDB: 32 non-redo rollback segment(s) are active.
  25. [Note] InnoDB: Waiting for purge to start
  26. [Note] InnoDB: 5.7.12 started; log sequence number 4828513952
  27. [Note] InnoDB: page_cleaner: 1000ms intended loop took 7748ms. The settings might not be optimal. (flushed=0 and evicted=0, during the time.)
  28. [Note] InnoDB: Loading buffer pool(s) from $datadir/ib_buffer_pool
  29. [Note] Plugin 'FEDERATED' is disabled.
  30. [Note] InnoDB: Buffer pool(s) load completed at 170706 19:49:30
  31. [Note] Recovering after a crash using $basedir/mysql-bin
  32. [ERROR] Error in Log_event::read_log_event(): 'read error', data_len: 579, event_type: 2
  33. [Note] Starting crash recovery...
  34. [Note] InnoDB: Starting recovery for XA transactions...
  35. [Note] InnoDB: Transaction 6729603 in prepared state after recovery
  36. [Note] InnoDB: Transaction contains changes to 1 rows
  37. [Note] InnoDB: 1 transactions in prepared state after recovery
  38. [Note] Found 1 prepared transaction(s) in InnoDB
  39. [Note] Crash recovery finished.
  40. [Note] Crashed binlog file $basedir/mysql-bin.000004 size is 690909184, but recovered up to 690908428. Binlog trimmed to 690908428 bytes.
  41. [Warning] Failed to set up SSL because of the following SSL library error: SSL context is not usable without certificate and private key
  42. ……………………………………………………………………………………………………………………
  43. [ERROR] Error in Log_event::read_log_event(): 'read error', data_len: 835, event_type: 2
  44. [Warning] Error reading GTIDs from relaylog: -1
  45. [Note] Slave I/O thread: Start asynchronous replication to master '*****' in log 'mysql-bin.000014' at position 286486095
  46. [Warning] Storing MySQL user name or password information in the master info repository is not secure and is therefore not recommended. Please consider using the USER and PASSWORD connection options for START SLAVE; see the 'START SLAVE Syntax' in the MySQL Manual for more information.
  47. [Note] Slave I/O thread for channel '': connected to master '*****',replication started in log 'mysql-bin.000014' at position 286486095
  48. [Warning] Slave SQL for channel '': If a crash happens this configuration does not guarantee that the relay log info will be consistent, Error_code: 0
  49. [Note] Slave SQL thread for channel '' initialized, starting replication in log 'mysql-bin.000014' at position 286485153, relay log '$datadir/mysql-relay.000013' position: 286485326
  50. [ERROR] Error in Log_event::read_log_event(): 'read error', data_len: 835, event_type: 2
  51. [ERROR] Error reading relay log event for channel '': slave SQL thread aborted because of I/O error
  52. [ERROR] Slave SQL for channel '': Relay log read failure: Could not parse relay log event entry. The possible reasons are: the master's binary log is corrupted (you can check this by running 'mysqlbinlog' on the binary log), the slave's relay log is corrupted (you can check this by running 'mysqlbinlog' on the relay log), a network problem, or a bug in the master's or slave's MySQL code. If you want to check the master's binary log or slave's relay log, you will be able to know their names by issuing 'SHOW SLAVE STATUS' on this slave. Error_code: 1594
  53. [ERROR] Error running query, slave SQL thread aborted. Fix the problem, and restart the slave SQL thread with "SLAVE START". We stopped at log 'mysql-bin.000014' position 286485153
  54. [Note] Event Scheduler: Loaded 0 events
  55. [Note] $basedir/mysqld: ready for connections.
  56. Version: '5.7.12' socket: '$datadir/mysqld.sock' port: 3306 Source distribution


因为已经挂了,而且没有来得及将当时执行的sql完整记录下来,只能看到show processlist前面的一小部分。

联系到业务,业务说这一台是slave,无读的业务,但 偶尔 会手动执行 报表相关的SELECT操作。

那么检查一下crash前的两个小时的 slow log:
发现反复有这么几条SQL:
  1. ……………………
  2. 00:00:11.392881 select * from tb ……………………
  3. 00:00:04.779748 select (case when ……………………
  4. 00:00:04.779748 select ( case when t2.col1 ……………………
  5. 00:00:03.328248 select ( case when t2.col1 ……………………
  6. 00:00:04.276773 select count(t1.id) from tb1 
  7. 00:00:05.039027 select (case when t2.col ……………………
  8. 00:00:10.263063 select (case when t2.col ……………………
  9. 00:00:03.131713 select t2.* from tb1 t1 ……………………
  10. 00:00:15.909456 select t2.* from tb1 t1 ……………………
  11. 00:00:14.367047 select * from tb ……………………
  12. ……………………

虽然
没有一条超过20秒的,但不要忘了,在磁盘被打爆之前的一秒,有一条执行了几千秒的SQL。

并且case when 只有一个order by。
出于取出其中一条,检查一下
执行计划:
  1. +--------+---------------+---------+---------+---------------------+--------+----------+----------------------------------------------+
  2. | type   | possible_keys | key     | key_len | ref                 | rows   | filtered | Extra                                        |
  3. +--------+---------------+---------+---------+---------------------+--------+----------+----------------------------------------------+
  4. | ALL    | NULL          | NULL    | NULL    | NULL                | 472765 | 20.99    | Using where; Using temporary; Using filesort |
  5. | eq_ref | PRIMARY       | PRIMARY | 8       | $tb1.col1           | 1      | 10.00    | Using where                                  |
  6. +--------+---------------+---------+---------+---------------------+--------+----------+----------------------------------------------+

看起来
不是很好,但是也不至于用到190G临时表。
虽然临时表是长期积累下来的,但看监控,磁盘使用率确实是在这一个多小时迅速积攒起来的。

首先 这个从库是没有业务的,而是通过客户 那边可能手动连接 过来做一些查询服务的。
知道这个信息之后, 继续对照slow log,发现slow log关于case when的这一条sql并不完全一样……
这可能意味着客户仍然 在不停地调试这条SQL…
而那条执行了几千秒的case when可能写得十分烂。

可惜的是,在mysqld crash的最后一瞬间,该sql仍然没有执行完成,导致没有被记录到slow log。
那条sql估计也永远无法再复现。

那么究竟是多烂的SQL,或者多大的表,可以用到这么多基于磁盘的临时表或文件排序呢??
赶紧检查一下库表的情况:

    表1:100w,表2:40w,也没有blob,text等字段。
    ② 重启后的datadir只有3、4G的数据量

所以我有个大但的想法:
可能客户在调试的过程中,产生了一条类似这样的SQL:
我对照了已经记录在slow log的几条sql,脱敏,格式化之后还原如下:
  1. SELECT ( CASE
  2.     WHEN t2.col='$info_000' THEN '$a'
  3.     WHEN t2.col='$info_001' THEN '$b'
  4.     WHEN t2.col='$info_002' THEN '$c'
  5.     WHEN t2.col='$info_003' THEN '$d'
  6.     WHEN t2.col='$info_004' THEN '$e'
  7.     WHEN t2.col='$info_005' THEN '$f'
  8.     WHEN t2.col='$info_006' THEN '$g'
  9.     ELSE t2.col
  10.     END ) AS 来源XX,
  11.     ( CASE
  12.     WHEN t2.clo2='$info_000' THEN '$a'
  13.     WHEN t2.clo2='$info_001' THEN '$b'
  14.     WHEN t2.clo2='$info_002' THEN '$c'
  15.     WHEN t2.clo2='$info_003' THEN '$d'
  16.     WHEN t2.clo2='$info_004' THEN '$e'
  17.     WHEN t2.clo2='$info_005' THEN '$f'
  18.     WHEN t2.clo2='$info_006' THEN '$g'
  19.     ELSE t2.col 
  20.     END ) AS 目标XX,
  21.     t2.col4 AS XX时间,
  22.     t2.col5 AS 金额
  23. FROM $tb1 t1 JOIN $tb2 t2
  24. WHERE t1.col3 = 4 AND (t2.col LIKE '$info%' OR t2.clo2 LIKE '$info%')
  25. ORDER BY t1.col4 DESC;

对,这两张表做了一个JOIN,但是没有条件。
产生了一个很大很大很大的 笛卡尔积

那么算一下:
列:13 + 19 = 32 
行: 1088181*440650 =  479,506,957,650  rows

再测试一下:
将两张表dump成sql文件,看了一下大小:330MB,并不大。
导入自己的实验环境,再将上述产生笛卡尔积的SQL跑一下。

经过一段时间等待:
【MySQL】一条SQL使磁盘暴涨并导致MySQL Crash_第1张图片

已经执行了半个小时多了。并且该SQL仍然在执行:
说出来你们可能不信,此时的ibtmp1大小:
  1. # du -sh *
  2. 4.0K auto.cnf
  3. 4.0K ib_buffer_pool
  4. 2.9G ibdata1
  5. 48M ib_logfile0
  6. 48M ib_logfile1
  7. 54G ibtmp1
  8. 12M mysql
  9. 1.1M performance_schema
  10. 676K sys
  11. 633M test
  12. 47M test_2

大概也好解释,为什么两张并不是很大的表,在执行某些查询之后会产生这么大的临时表了…

当然,在做实验的等待时间里,已经将复制的ERROR给修复了。
当时复制的报错:

  1. ……
  2. Slave_IO_Running: Yes
  3. Slave_SQL_Running: No
  4. ……
  5. Seconds_Behind_Master: NULL
  6. Master_SSL_Verify_Server_Cert: No
  7. Last_IO_Errno: 0
  8. Last_IO_Error:
  9. Last_SQL_Errno: 1594
  10. Last_SQL_Error: Relay log read failure: Could not parse relay log event entry. The possible reasons are: the master's binary log is corrupted (you can check this by running 'mysqlbinlog' on the binary log), the slave's relay log is corrupted (you can check this by running 'mysqlbinlog' on the relay log), a network problem, or a bug in the master's or slave's MySQL code. If you want to check the master's binary log or slave's relay log, you will be able to know their names by issuing 'SHOW SLAVE STATUS' on this slave.
结合之前贴的error log可以看到更多信息,推断是relay log有损坏。


那么就好解决了:
  1. STOP SLAVE;

  2. -- master_log_file和master_log_pos为报错时所记录的位置。
  3. CHANGE MASTER master_log_file='xxx', master_log_pos=xxx;

  4. START SLAVE;
此处不再赘述。


到目前为止,我实验环境的datadir已经满了,但是和此处正式环境有区别的是。
在磁盘写满之时,并没有马上导致mysql crash,而是输出了:
  1. [ERROR] InnoDB: Write to file ./ibtmp1failed at offset 83579895808, 1048576 bytes should have been written, only 958464 were written. Operating system error number 28. Check that your OS and file system support files of this size. Check also that the disk is not full or a disk quota exceeded.
  2. [ERROR] InnoDB: Error number 28 means 'No space left on device'
  3. [Note] InnoDB: Some operating system error numbers are described at http://dev.mysql.com/doc/refman/5.7/en/operating-system-error-codes.html
  4. [Warning] InnoDB: Error while writing 67108864 zeroes to ./ibtmp1 starting at offset 83563118592
  5. [ERROR] /data/mysql-base/mysql57/bin/mysqld: The table '#sql_b26_0' is full
  6. [ERROR] Disk is full writing '/data/mysql-data/mysql57-3357/binlog/mysql-bin.000015' (Errcode: 15868064 - No space left on device). Waiting for someone to free space...
  7. [ERROR] Retry in 60 secs. Message reprinted in 600 secs
实际上正常情况下,mysql在磁盘满了的情况下, 会每隔1分钟做一次检查,每隔10分钟将检查信息输出到错误日志。

既然mysqld没有被干掉,我再次执行了一下那条可怕的sql呢?……

当然……error log如愿输出被kill的消息……
并且mysqld进程被杀掉,此时:


  1. [ERROR] /data/mysql-base/mysql57/bin/mysqld: Binary logging not possible. Message: An error occurred during flush stage of the commit. 'binlog_error_action' is set to 'ABORT_SERVER'. Hence aborting the server.
  2. 14:42:48 UTC - mysqld got signal 6 ;
  3. This could be because you hit a bug. It is also possible that this binary
  4. or one of the libraries it was linked against is corrupt, improperly built,
  5. or misconfigured. This error can also be caused by malfunctioning hardware.
  6. Attempting to collect some information that could help diagnose the problem.
  7. As this is a crash and something is definitely wrong, the informati
有趣的是,“information”这个单词都没有输出完,大概是真的一个字符都写不下了吧。


试图用vim在error log里 添加几个字符并保存,也报错。


那么总结一下,供参考:
在线上环境执行的sql,建议进行审核,报表等OLAP需求也需要正式一些。
考虑是否可以限制ibtmp1的大小,也就是设定innodb_temp_data_file_path的最大值。
可能需要定期重启mysqld来收缩临时表空间。
做好监控和及时响应,这次就是响应时间过长,达到阈值,然后到机子上检查时,已经晚了,眼睁睁地看到mysqld在我眼前crash。







来自 “ ITPUB博客 ” ,链接:http://blog.itpub.net/29773961/viewspace-2142197/,如需转载,请注明出处,否则将追究法律责任。

转载于:http://blog.itpub.net/29773961/viewspace-2142197/

你可能感兴趣的:(【MySQL】一条SQL使磁盘暴涨并导致MySQL Crash)