View Single Post
Unread 02-13-2008, 09:08 AM   #18
netwraith
Cooling Neophyte
 
Join Date: Dec 2007
Location: New Hampshire USA
Posts: 24
Default Re: Using SATA adapter to mimic PATA/IDE drive.

Quote:
Originally Posted by netwraith
Also, BTW... the SMART logic on the newer ATA/EIDE drives gives me the most problems with ATA drives... (They often schedule self tests at odd times, and most LINUX based systems use the error data returned by the SMART logic and will fail a drive if SMART says it failed certain tests... ) I am not sure if GOS uses the SMART logic and/or has the smartctl command that most LINUX systems have. You might want to try the INFO switch and see what kind of stuff the SMART logic monitors and return to the system.... If not in a SnapServer, you might want to check it out on another LINUX box....
Follow up comment: Just for information, I logged into two different Guardian units, one 4400, and one 4500 and both had different SMART configurations. The 4400 was delivered with SMART turned off. The 4500 was delivered with SMART turned on. Both have 120GB Hitachi drives and both work without issues otherwise.

Another unit I have was inconsistent. i.e. some drives off and some drives on. That unit had Quantum Fireball drives in it. (I think they were non original drives)..

While this proves nothing at all, it may be good information for us to make a practice to check our SMART status... just for consistency reasons and perhaps better reliability. I am not going to state whether it should be on or off, since, both the 4400 and 4500 with the Hitachi drives are 100% reliable to this point. I was not having any problems with the other 4400 either, but, turned them all on just to be consistent.

On Guardian units:

The command to check the status of the SMART is:

smartctl -a /dev/hda
(Substitute with /dev/hdc , /dev/hde, /dev/hdg to check all of your internal drives).
Below are two samples using the -i form of the command... (Brief output .. still shows what's needed)
Notice how the first sample is disabled and the second is enabled (No cracks on the MAXTOR in the example... this set is pretty reliable)..

one:
Code:
# smartctl -i /dev/hda
smartctl version 5.36 [i686-pc-linux-gnu] Copyright (C) 2002-6 Bruce Allen
Home page is http://smartmontools.sourceforge.net/

=== START OF INFORMATION SECTION ===
Model Family:     IBM/Hitachi Deskstar 120GXP family
Device Model:     IC35L120AVVA07-0
Serial Number:    VNC607A6CBVTGJ
Firmware Version: VA6OA52A
User Capacity:    123,522,416,640 bytes
Device is:        In smartctl database [for details use: -P show]
ATA Version is:   5
ATA Standard is:  ATA/ATAPI-5 T13 1321D revision 1
Local Time is:    Wed Feb 13 14:56:59 2008 EST
SMART support is: Available - device has SMART capability.
SMART support is: Disabled

SMART Disabled. Use option -s with argument 'on' to enable it.
#
two:
Code:
# smartctl -i /dev/hda
smartctl version 5.33 [i686-pc-linux-gnu] Copyright (C) 2002-4 Bruce Allen
Home page is http://smartmontools.sourceforge.net/
=== START OF INFORMATION SECTION ===
Device Model:     Maxtor 7Y250P0
Serial Number:    Y65A9CEE
Firmware Version: YAR41BW0
User Capacity:    251,000,193,024 bytes
Device is:        In smartctl database [for details use: -P show]
ATA Version is:   7
ATA Standard is:  ATA/ATAPI-7 T13 1532D revision 0
Local Time is:    Wed Feb 13 14:58:08 2008 EST
SMART support is: Available - device has SMART capability.
SMART support is: Enabled

#
to change the state of a drive:

To turn SMART :ON: for /dev/hda
smartctl -s on /dev/hda

To turn SMART :OFF: for /dev/hda
smartctl -s off /dev/hda

Whichever you decide, repeat the command for all drives until they are consistent.

Finally, if your unit is reliable and they are all OFF, I would not change what is working. If they are inconsistent or your unit is having reliability issues, I would turn these ON... One final note... IF your unit is having issues and you turn this on, it is possible for the unit to *COMPLETELY* fail one or more drives. So, try to get a good backup before you do...

You can check the status of your ARRAY's with the command:

cat /proc/mdstat

Here is what it could look like....

Code:
# cat /proc/mdstat
Personalities : [linear] [raid0] [raid1] [raid5] [spare] 
read_ahead 1024 sectors
md0 : active raid5 hda4[0] hdg4[3] hde4[2] hdc4[1]      346594944 blocks level 5, 64k chunk, algorithm 2 [4/4] [UUUU]      
md101 : active raid1 hda5[0] hdc5[3] hde5[2] hdg5[1]      273024 blocks [4/4] [UUUU]      
md100 : active raid1 hda2[0] hdc2[3] hde2[2] hdg2[1]      546112 blocks [4/4] [UUUU]      
unused devices: <none>
#
you are looking for the :

[4/4] [UUUU] lines.... I you have ARRAYS of more or less than 4 drives, your numbers will be different, but you are still looking for all drives online. If you see something like:

Code:
<snip>
md0 : active raid5 hda4[0] hdg4[3] hde4[2] hdc4[1] (F)      
      346594944 blocks level 5, 64k chunk, algorithm 2 [4/3] [UUU_]
<snip>
It means you have a failed partition and are in DEGRADED mode. You really need a backup before you do anything else... Be careful of one more thing.... It is possible for a drive to show up as *FAILED* in one array, but, not the other ... i.e. it's OK in the md100 ARRAY, but, shows dead in the md0 ARRAY... This is because the LINUX MDADM software fails partitions and not drives...

Last edited by netwraith; 02-13-2008 at 11:33 AM.
netwraith is offline   Reply With Quote