Just noticed that one of my server is showing unusually high IO wait time, average over a week is 6.5% with peaks of 65%. Looking at Sysmon, it is obvious that the IO wait time has increased significantly over the past 3-4 weeks. High IO percentages alternate with low values every hours or so. Messages log shows:
Mar 16 19:32:57 mx1 squid[4375]: WARNING: Disk space over limit: 176200 KB > 102400 KB
Mar 16 19:32:58 mx1 kernel: ata1.00: exception Emask 0x0 SAct 0x0 SErr 0x0 action 0x0
Mar 16 19:32:58 mx1 kernel: ata1.00: (irq_stat 0x40000001)
Mar 16 19:32:58 mx1 kernel: ata1.00: cmd 25/00:6c:e1:44:1c/00:00:1d:00:00/e0 tag 0 cdb 0x0 data 55296 in
Mar 16 19:32:58 mx1 kernel: res 51/40:00:2c:45:1c/00:00:1d:00:00/e0 Emask 0x9 (media error)
Mar 16 19:32:58 mx1 kernel: ata1.00: configured for UDMA/133
Mar 16 19:32:58 mx1 kernel: SCSI error : <0 0 0 0> return code = 0x8000002
Mar 16 19:32:58 mx1 kernel: Info fld=0x4000000 (nonstd), Invalid sda: sense = 72 11
Mar 16 19:32:58 mx1 kernel: end_request: I/O error, dev sda, sector 488391905
Mar 16 19:32:58 mx1 kernel: Buffer I/O error on device sda2, logical block 122045765
Mar 16 19:32:58 mx1 kernel: ata1: EH complete
This system is currently at 7.3 level, Raid1, 2 x 250GB SATA. I was about to upgrade to 7.4 when I noticed this issue, response is very slow at times over ssh. The server is at a remote site, hard to access. I suspect a hardware issue ( hda?) but am not sure about the Disk space over limit warning.
Any assistance would be very appreciated, thanks.