![]() It is currently (2016) being updated to read the status directly from /sys, and should hopefully soon be able to display raid status correctly. It is capable of displaying the state of raid arrays, but unfortunately currently the code is broken - it reads mdstat, and doesn't understand the current output. Xosview is a venerable utility, and one of the author's favourites. This will tell you the state of your arrays, and very importantly it will tell you whether any drives have failed, and whether any arrays are degraded. The exact command depends upon your Linux distribution: chkconfig -add mdadm (Red Hat, Fedora and SUSE)Īs always, be certain that you use your own e-mail address and the names of the actual hard drives and arrays in your system.You should get to know /proc/mdstat, looking at it often. To ensure the service starts at boot, you’ll need to add it to the boot sequence. Some distributions use the name mdadm and others use mdmonitor: /etc/init.d/mdadm start Once you have written the configuration file, you need to start the service. If this is the case, run mdadm -detail /dev/md1 for each RAID. MAILADDR this example, any changes to the listed md devices will be immediately e-mailed to that some newer versions of mdadm require that devices be identified by UUID (e.g. To monitor Linux software RAIDs, you’ll need at least the following lines in /etc/nf: DEVICE /dev/sd1 /dev/sd5 /dev/sd6 /dev/sd7 /dev/sd8 /dev/sd10ĪRRAY /dev/md1 devices=/dev/sda1,/dev/sdb1ĪRRAY /dev/md5 devices=/dev/sda5,/dev/sdb5ĪRRAY /dev/md6 devices=/dev/sda6,/dev/sdb6ĪRRAY /dev/md7 devices=/dev/sda7,/dev/sdb7ĪRRAY /dev/md8 devices=/dev/sda8,/dev/sdb8ĪRRAY /dev/md10 devices=/dev/sda10,/dev/sdb10 Update-rc.d mdadm defaults (Debian) mdadm The exact command depends upon your Linux distribution: chkconfig -add smartd (Red Hat, Fedora and SUSE) Once you have written the configuration file, you need to start the service: /etc/init.d/smartd start Run a long self test on each drive once a month (4am for sda 5am for sdb).Run a short self test on each drive once a week (3am for sda 4am for sdb).Enables Automatic Offline Testing of the drives.Monitor the results of hard drive self tests.Watch for any critical failure warnings in the SMART data.E-mailing all alerts to Sending one test e-mail upon startup.In this example, they monitor drives /dev/sda and /dev/sdb by performing the following tasks: Here are the standard entries I use in /etc/nf: /dev/sda -a -d ata -m -H -l error -l selftest -M test -o on -S on -s (S/././3/03|L/./15/./04) Additionally, you may be able to monitor hard drive health data if the controller supports it (3ware and ARECA cards are known to work) – see the smartd man page. However, you must be sure to properly configure the automated alerts within the controller’s management interface – check the manual for full instructions. If you are using a hardware RAID controller, then it manages some of these tasks. ![]() Similarly, mdadm watches the health of your Linux software RAIDs for any problems. health data directly off the drives and sends alerts of any changes. The smartd daemon reads hard drive S.M.A.R.T. If you do not have a dedicated hardware RAID controller, there are two utilities to be configured and started: smartd and mdadm. ![]() However, that doesn’t mean that you can’t monitor the situation and receive an alert as soon as the first problem develops. I have seen RAIDs run in degraded mode for months or years until additional drive failures ruined any chance of data recovery.ĭrives and operating systems are designed to work around issues as best they can until absolute failure. Even if you are using a software or hardware RAID, it will only continue to function if you replace failed drives. By default, you won’t find out that one of your hard drives has failed until the data is gone. ![]()
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |