C
Curious George
On Tue, 30 Nov 2004 17:42:03 +0100, "Joris Dobbelsteen"
One of the main points of raid is to be proactive about failures &
uptime. So I'm talking about things like being able see SMART/PFA
status and have the array move to a hot spare when SMART fails.
Providing notification about such an event (like SNMP traps, email,
pages, popup). Also initiating a process to validate the data against
the ecc data & check the media surface. Being able to reconfigure or
upgrade or provide info about the array without taking the whole
machine down. Being able to automate & schedule these upgrades or
validation checks for a more convenient time (so there is no
discernable performance hit). All this may seem like overkill, but it
really isn't if you want to get the full benefits of non-zero raid. I
don't see RAID0 as a viable choice for most scenarios. I also assume
that if you are looking to firmware raid (even ROMB, etc) you have
higher expectations than the quick and dirty OS software striped set.
AND model revisions
That's all more of an issue with ATA raid for a number or reasons esp
not being as configurable as scsi drives as well as firmware
limitations.
Well that's fine for the ideal and simplest of scenarios - its not the
only one.
Yes but AFAIK not always automatically or at least not always on a
strictly hardware/low level. Also it seems to me the ideal is to have
a raid controller initiate & manage a continuous background scan of
the media and restore data from bad sectors using the redundant data
residing on good sectors rather than trying to recover/read from weak
sectors. (better ata & scsi controllers do this)
I don't know much about memory stick architecture. It seems to me
this is more a result of file system overhead and perhaps conflicting
measurements of raw capacity (like it usually is w' storage) rather
than say "reserved space" for some low-level harware-implemented
automatic recovery process.
Please educate me if wrong.
Could you direct us to this resource? AFAIK SATA has the scsi
protocol on top of ata - so it has greater overhead. If there is 1
disk there isn't really any contention for that resource in either
case. Because SATA is point to point there is never drive arbitration
per bus, but in multi-disk sata the overall efficiency depends more on
details of the controller(s) than the point to point design per se.
Educate me if I'm wrong.
Then I would expect you to favor the larger 7200 rpm sata drives
marketed as "Personal Storage" devices (some are quite good). The
"enterprise" raptors are similar to current 10k scsi in multiple
regards including price, capacity, & performance (& theoretical MTBF).
AFAIK the price "advantage" of raptors has more to do with a
comparison with the "need" for a "deluxe" retail boxed Adaptec
controller - not really capacity per disk or $/MB (from what I've
seen).
Throwing ATA raid0 into the mix (as you first suggested it) the trade
off variables have to be expanded to include complexity, cost & to
some extent reliability.
True. But often things seem to be luxury simply because of sticker
shock. In some cases when time is very valuable products that bring
even small increases in productivity or assurance of quality can bring
real value despite this initial sticker shock. This is all really
relative though & must be taken case by case.
Outright and complete failure is one thing. Erratic behavior due to
poor design, overheating, damage & imminent failure is another. Yes
mechanical devices don't last as long as IC's but I think there is a
bit more to reliability than time to total failure.
Frankly I'm not that worried about an ATA raid controller dying
prematurely or before a scsi hba. I'm more concerned about a low-end
controller having limitations which interfere with the ability for the
raid to reliability deliver on its core features/promises or conflicts
or poorly written code which waste the user's/administrator's time &
eats away at the assumed cost savings.
Yes but the array MTBF calculation characterizes arrays as more
complex than a single drive with more potential points of failure.
Yes when one disk drops off the other keeps going, but there is more
to it than that.
1. if both disks are the same age with the same wear they might die at
similar times so you might not have as much time as you think to get
the replacement. Failure rates occur in a "U" pattern are not linear
across time.
2. not all failures are neat and tidy
Either not so uncommon cases are real potential time/productivity
wasters which can invalidate the expected benefits of non-zero raid.
I don't understand. You mean rebuilding the data to a new disk? Than
not "probably" -"definitely" because the process is supposed to be
seamless as opposed to backup file recovery & bare metal restore &
redoing the work since the last backup. The minor performance hit and
short time it takes to rebuild in the background is hardly worth even
trying to compare or consider (unless you are really concerned about
power failure and UPS runtime).
But that's a pretty crude comparison of reliability. That assertion
also depends on a lot of things.
Also you can't really say:
1. both are equally reliable because both are reliable enough to
usually work during a normal service life.
and at the same time say:
2. "two RAID1 raptors have equal costs to a single Cheetah 15K.3 and
a much better MTBF (theoretically)"
without the two statements either being contradictorily or virtually
valueless. Certainly I don't yet see the second statement as being
proven, explained, or correct.
The MTBF calculation I cited highlights the added complexity and
potential points of failure that raid brings and that is normally
interpreted as an array being "theoretically less reliable" than a
single disk.
That being said a properly implemented non-zero RAID "should" yield
"more reliable storage" but that has more to do with doing the work
and making the investment to "dot your i's" and "cross your t's" than
any theoretical calculation or characterization of _all_ or _any_
non-zero raid. It's very easy to botch a raid implementation and end
up with storage that is more expensive, more work, and less reliable
over normal disks. Storage & technical discussion groups regularly
have "help I didn't do a backup and I can't get my ultra-cheap raid
back online" posts. Once in a while you also see "Boy! I just found
out the hard way that most raid 5's are susceptible to transient write
errors" posts. These users were not 100% protected just because they
got disks to work together and generate ecc data.
Yes. However it is not so uncommon for "minor" hardware error with
"working" devices. When you don't invest enough time to really
scientifically dissect & troubleshoot these issues they appear as
software problems when they are not or are simply unsolved & forgotten
about or ignored because you were lucky that it didn't affect anything
that important.
Often yes. My preoccupation with robust data integrity features in
raid (here and elsewhere) has to do with transient error and failing
but still spinning media or power failure which just shouldn't ever
get the chance to crap on anything. If you have non-zero raid
something is very wrong if you _ever_ have to go to a backup or
"reinstall" or "rollback" or troubleshoot due to any kind of storage
HW issue. Without significant gains it's hard to justify the
additional expense, effort, or system complexity.
Well if you are going to exceed the service life so dramatically you
are not exactly "wearing a belt & suspenders" no matter how much the
$$$ investment. All that redundancy is good to ensure uptime for a
"normal period" but is not necessarily a great tool for trying to
drain the last drop of blood out of antiquated and worn out HW because
of the overhead and expense.
<acerbic comment>
A 5-year-old machine is hardly a "luxury."
</acerbic comment>
I have a few machines like that and nearly twice as old still running
and in service with mostly original parts (now doing very limited
tasks of course). It's exactly those machines that impressed upon me
some time ago that just because its "up" and "seems OK" doesn't
necessarily mean you can really depend on it 100%. Also timesavings
and confidence in HW really go a long way and greatly offset some
"sticker shock" expenses or at least regular upgrades/decommissions.
I've been finding it MUCH cheaper to replace these "working,"
"capable" machines than to try to continue to plug-along with "old,"
or "cheap" HW for a variety or reasons including reliability.
Of course I may just be too picky and fortunate.
What management, just install the array and you are done. It works just like
a normal disk (except for setting up the array once).
One of the main points of raid is to be proactive about failures &
uptime. So I'm talking about things like being able see SMART/PFA
status and have the array move to a hot spare when SMART fails.
Providing notification about such an event (like SNMP traps, email,
pages, popup). Also initiating a process to validate the data against
the ecc data & check the media surface. Being able to reconfigure or
upgrade or provide info about the array without taking the whole
machine down. Being able to automate & schedule these upgrades or
validation checks for a more convenient time (so there is no
discernable performance hit). All this may seem like overkill, but it
really isn't if you want to get the full benefits of non-zero raid. I
don't see RAID0 as a viable choice for most scenarios. I also assume
that if you are looking to firmware raid (even ROMB, etc) you have
higher expectations than the quick and dirty OS software striped set.
With some controllers you might get in trouble when you use different disks,
so use the same brand AND model.
AND model revisions
That's all more of an issue with ATA raid for a number or reasons esp
not being as configurable as scsi drives as well as firmware
limitations.
Recovery: RAID1: turn of the system, remove the defective drive and replace
it. Turn on, repair the array, wait a few seconds for the disk copy and
done.
Well that's fine for the ideal and simplest of scenarios - its not the
only one.
Todays disks are capable of relocating damaged sectors,
Yes but AFAIK not always automatically or at least not always on a
strictly hardware/low level. Also it seems to me the ideal is to have
a raid controller initiate & manage a continuous background scan of
the media and restore data from bad sectors using the redundant data
residing on good sectors rather than trying to recover/read from weak
sectors. (better ata & scsi controllers do this)
they do it all (same
reason your 128MB USB drive/memory stick only has 120 MB storage capacity).
I don't know much about memory stick architecture. It seems to me
this is more a result of file system overhead and perhaps conflicting
measurements of raw capacity (like it usually is w' storage) rather
than say "reserved space" for some low-level harware-implemented
automatic recovery process.
Please educate me if wrong.
Simply call it resource contention. For a single-user system the Raptor will
handle the resource contention better than the SCSI system.
Of course this is subject to the opinion expressed by a third party, who may
resonabily be expected to have sufficient knowledge of the system to provide
such an 'opinion'.
Could you direct us to this resource? AFAIK SATA has the scsi
protocol on top of ata - so it has greater overhead. If there is 1
disk there isn't really any contention for that resource in either
case. Because SATA is point to point there is never drive arbitration
per bus, but in multi-disk sata the overall efficiency depends more on
details of the controller(s) than the point to point design per se.
Educate me if I'm wrong.
Usually response times, throughput and storage capacity requires a
trade-off.
My trade-off would favor storage capacity over throughput over response
times.
Then I would expect you to favor the larger 7200 rpm sata drives
marketed as "Personal Storage" devices (some are quite good). The
"enterprise" raptors are similar to current 10k scsi in multiple
regards including price, capacity, & performance (& theoretical MTBF).
AFAIK the price "advantage" of raptors has more to do with a
comparison with the "need" for a "deluxe" retail boxed Adaptec
controller - not really capacity per disk or $/MB (from what I've
seen).
Throwing ATA raid0 into the mix (as you first suggested it) the trade
off variables have to be expanded to include complexity, cost & to
some extent reliability.
Indeed, but if you want luxery, you are (or someone else is if you are
lucky) going to pay for it anyway. Its just considering how much you are
willing to spend for you luxery.
However for the same luxery (or even the same essential product that you
simply need) there is a large variation of prices that you can pay.
True. But often things seem to be luxury simply because of sticker
shock. In some cases when time is very valuable products that bring
even small increases in productivity or assurance of quality can bring
real value despite this initial sticker shock. This is all really
relative though & must be taken case by case.
Lets asume most chip manufacturers (NOT designers, there are only a few
manufacturers) are equally capable of making the same quality product.
Besides the mechanical parts are more likely to fail than electrical.
A very hot CPU would last for 10 years (its designed for it anyway). I
expect the same for chipsets.
I only saw electronics fail because of ESD, lightning storms and some
chemicals (e.g. from batteries).
I wouldn't consider the controller to be a major problem with disk
subsystems.
Outright and complete failure is one thing. Erratic behavior due to
poor design, overheating, damage & imminent failure is another. Yes
mechanical devices don't last as long as IC's but I think there is a
bit more to reliability than time to total failure.
Frankly I'm not that worried about an ATA raid controller dying
prematurely or before a scsi hba. I'm more concerned about a low-end
controller having limitations which interfere with the ability for the
raid to reliability deliver on its core features/promises or conflicts
or poorly written code which waste the user's/administrator's time &
eats away at the assumed cost savings.
When using 2-disk RAID 1 (NOT RAID 0): when 1 disks fails the system
continues to operate correctly, leaving you time to replace the defective
material with no loss of continuity.
One disk systems will stop working when the disk failes.
Yes but the array MTBF calculation characterizes arrays as more
complex than a single drive with more potential points of failure.
Yes when one disk drops off the other keeps going, but there is more
to it than that.
1. if both disks are the same age with the same wear they might die at
similar times so you might not have as much time as you think to get
the replacement. Failure rates occur in a "U" pattern are not linear
across time.
2. not all failures are neat and tidy
Either not so uncommon cases are real potential time/productivity
wasters which can invalidate the expected benefits of non-zero raid.
Besides recovery times for RAID1 are probably lower than for one-disk
systems.
I don't understand. You mean rebuilding the data to a new disk? Than
not "probably" -"definitely" because the process is supposed to be
seamless as opposed to backup file recovery & bare metal restore &
redoing the work since the last backup. The minor performance hit and
short time it takes to rebuild in the background is hardly worth even
trying to compare or consider (unless you are really concerned about
power failure and UPS runtime).
Basically under normal operations the system will continue to work without
failing once.
But that's a pretty crude comparison of reliability. That assertion
also depends on a lot of things.
Also you can't really say:
1. both are equally reliable because both are reliable enough to
usually work during a normal service life.
and at the same time say:
2. "two RAID1 raptors have equal costs to a single Cheetah 15K.3 and
a much better MTBF (theoretically)"
without the two statements either being contradictorily or virtually
valueless. Certainly I don't yet see the second statement as being
proven, explained, or correct.
The MTBF calculation I cited highlights the added complexity and
potential points of failure that raid brings and that is normally
interpreted as an array being "theoretically less reliable" than a
single disk.
That being said a properly implemented non-zero RAID "should" yield
"more reliable storage" but that has more to do with doing the work
and making the investment to "dot your i's" and "cross your t's" than
any theoretical calculation or characterization of _all_ or _any_
non-zero raid. It's very easy to botch a raid implementation and end
up with storage that is more expensive, more work, and less reliable
over normal disks. Storage & technical discussion groups regularly
have "help I didn't do a backup and I can't get my ultra-cheap raid
back online" posts. Once in a while you also see "Boy! I just found
out the hard way that most raid 5's are susceptible to transient write
errors" posts. These users were not 100% protected just because they
got disks to work together and generate ecc data.
You are probably have more problems with software than you
will have with hardware. Most down-time is either human or software related,
not hardware.
Yes. However it is not so uncommon for "minor" hardware error with
"working" devices. When you don't invest enough time to really
scientifically dissect & troubleshoot these issues they appear as
software problems when they are not or are simply unsolved & forgotten
about or ignored because you were lucky that it didn't affect anything
that important.
The issue is that when its hardware related, recovery costs
much more time and you have a bigger risk of losing valuable data.
When the disk will start to fail, it will probably be obsolute anyways,
Often yes. My preoccupation with robust data integrity features in
raid (here and elsewhere) has to do with transient error and failing
but still spinning media or power failure which just shouldn't ever
get the chance to crap on anything. If you have non-zero raid
something is very wrong if you _ever_ have to go to a backup or
"reinstall" or "rollback" or troubleshoot due to any kind of storage
HW issue. Without significant gains it's hard to justify the
additional expense, effort, or system complexity.
unless your system lasts for more than 6 years. Of course, when you want
this, you should rather prepare for the worst and have a 4-computer
clustered installed with fail-over capability.
Well if you are going to exceed the service life so dramatically you
are not exactly "wearing a belt & suspenders" no matter how much the
$$$ investment. All that redundancy is good to ensure uptime for a
"normal period" but is not necessarily a great tool for trying to
drain the last drop of blood out of antiquated and worn out HW because
of the overhead and expense.
Asuming its for luxery and I have here a system that is in operation for
already 5 years and is subject to frequent transports and some very
disk-intensive work at times, it never left me alone due to a hardware
failure (the normal minor stuff because I forgot some cables or didn't
attach them too well provided). All the products I used where the cheapest
compared to competitors, however some trades where made between brands when
I think for only a very small difference I could get something I expect to
be more reliable or better.
<acerbic comment>
A 5-year-old machine is hardly a "luxury."
</acerbic comment>
I have a few machines like that and nearly twice as old still running
and in service with mostly original parts (now doing very limited
tasks of course). It's exactly those machines that impressed upon me
some time ago that just because its "up" and "seems OK" doesn't
necessarily mean you can really depend on it 100%. Also timesavings
and confidence in HW really go a long way and greatly offset some
"sticker shock" expenses or at least regular upgrades/decommissions.
I've been finding it MUCH cheaper to replace these "working,"
"capable" machines than to try to continue to plug-along with "old,"
or "cheap" HW for a variety or reasons including reliability.
Of course I may just be too picky and fortunate.
