On 04/04/2014 01:17 PM, CS_DBA wrote:
> Hi All;
>
> We have a server running Scientific Linux 6.4 with a large 4.1TB RAID 10
> volume
>
> We keep all of our KVM's on this volume
>
> Today we found that we are unable to access the VM's
>
> After some checking we found errors such as the ones below in
> /var/log/message
>
> We unmounted the RAID volume (umount /dev/md0) thinking we could run
> some diagnostics and maybe fix it, however I'm less than up to speed on
> managing RAID volumes...
>
> Can someone help me debug this? How do I run a "check" on a RAID volume?
cat /proc/mdstat
will tell you the state of the array(s). Run that and report back.
The errors in the log imply the drive or controller ata5.01 has problems.
The EXT4-fs error implies the underlying device (md0) cannot be read.
Remember the "0" in RAID10 means that both of the underlying RAID1
arrays need to be clean or you won't be able to use the RAID0.
Unfortunately this also implies that one of your RAID1 arrays is unreadable.
Jeff