If this is your first visit, be sure to check out the FAQ by clicking the link above. You may have to register before you can post: click the register link above to proceed. To start viewing messages, select the forum that you want to visit from the selection below. |
|
|
Thread Tools | Display Modes |
#21
|
|||
|
|||
Enterprise versus "consumer" grade drives
On 14/10/12 01:44, miso wrote:
On 10/12/2012 9:58 PM, Arno wrote: miso wrote: [...] So is there any advantage to buying a Seagate Constellation versus a Baracuda? Or Ultrastar versus Deskstar? No. Unless you have a hard limit to only use one drive. "Enterprise" drives are a bit better in reliability, but RAID1 is so massively better that enterprise drives are laughable. The only place these pay of (somewhat) is if drive replacement is expensive, e.g. because somebody has dro drive to the datacenter. There, even a small imporivement in reliability can justify a larger price increase. [...] Given that the OS with be on SSD and the magnetic media is on RAID 0, would it still make sense to go with enterprise grade drives, presuming they are more reliable that the consumer grade? If you use RAID0, your data is basically doomed anyways, unless you have good backup. If you have good backup, no need to go for enterprise drives. RAID0 is basically always a bad choice except for cache and buffer applications that need high throughput, such as video-capture and editing. But RAID0 should never be used as actual longer-term storage. Arno I meant mirror, i.e. raid 1, for the hard drives. I'll consider RAID5. I've done that before. Not with the greatest results though. When the mobo died, I was able to put the RAID 10 array in another PC and all the data was still there. The RAID5 array couldn't be recovered. I had a pretty good backup. Raid 5 at least buys you something in the way of more storage with just as good security, well provided the mobo doesn't get Chinese cap disease. That's /always/ a risk with hardware raid solutions. It's normally not too bad if you can stick to the same supplier and replace the motherboard / hardware raid card with another using the same basic chipsets - but it is always a risk. RAID5 is good for some things, but can have very poor performance for some usage - small writes are particularly bad. |
#22
|
|||
|
|||
Enterprise versus "consumer" grade drives
miso wrote:
On 10/12/2012 9:58 PM, Arno wrote: miso wrote: [...] So is there any advantage to buying a Seagate Constellation versus a Baracuda? Or Ultrastar versus Deskstar? No. Unless you have a hard limit to only use one drive. "Enterprise" drives are a bit better in reliability, but RAID1 is so massively better that enterprise drives are laughable. The only place these pay of (somewhat) is if drive replacement is expensive, e.g. because somebody has dro drive to the datacenter. There, even a small imporivement in reliability can justify a larger price increase. [...] Given that the OS with be on SSD and the magnetic media is on RAID 0, would it still make sense to go with enterprise grade drives, presuming they are more reliable that the consumer grade? If you use RAID0, your data is basically doomed anyways, unless you have good backup. If you have good backup, no need to go for enterprise drives. RAID0 is basically always a bad choice except for cache and buffer applications that need high throughput, such as video-capture and editing. But RAID0 should never be used as actual longer-term storage. Arno I meant mirror, i.e. raid 1, for the hard drives. Ah, that is something different. I'll consider RAID5. I've done that before. Not with the greatest results though. When the mobo died, I was able to put the RAID 10 array in another PC and all the data was still there. The RAID5 array couldn't be recovered. I had a pretty good backup. You should test RAID recovery when you design that array, i.e. before you put data on it. Software RAID is better than hardware RAID is better than FAKE RAID. Raid 5 at least buys you something in the way of more storage with just as good security, well provided the mobo doesn't get Chinese cap disease. That should be long over now. This was an isolated incident of industrial espionage and basically all caps from that time should now be dead. Arno -- Arno Wagner, Dr. sc. techn., Dipl. Inform., CISSP -- Email: GnuPG: ID: 1E25338F FP: 0C30 5782 9D93 F785 E79C 0296 797F 6B50 1E25 338F ---- Cuddly UI's are the manifestation of wishful thinking. -- Dylan Evans |
#23
|
|||
|
|||
Enterprise versus "consumer" grade drives
"miso" wrote:
I haven't built a PC with hard drives in a few years and I have to say I am amazed at the hard drive market. [Note I have bought a few USB drives, so I am referring to internal drives here.] First of all, it seems everyone bought everyone else. Samsung went to Seagate. Hitachi went to WD. Fujitsu went to Toshiba, which is presume is waiting to go elsewhere. Well now I can see why the hard drive market never fell back to the pre-Thailand flood prices. There are three, no make that 2.5 suppliers. I always had the best luck with Seagate. So I do the usual market survey (translation: read Newegg reviews) and it seems Seagate now sucks. Also the 5 year warranty is 3 years, and that is on a good day. Seagate has some drive with 1 year warranties. OK, so check out WD. Hmmh, they seem to suck now too. So is there any advantage to buying a Seagate Constellation versus a Baracuda? Or Ultrastar versus Deskstar? Have you noticed some vendors selling new drives without warranties? When did that start happening? FWIW, the system I plan on building will use intel SSD for the OS. I've done two systems with intel SSD and no headaches, well other than having to pay top dollar for the SSD. [I had a Corsair SSD arrive DOA. That is my only non-intel experience.] I plan on getting two large hard drives (normal, not SSD) and running RAID0. [Raid can be a pain if the controller dies. Raid 0 may be inefficient, but at least the drives are readable without RAID. I had a mobo fail that had a RAID 10 and a Raid 5 array on it. I got the RAID 10 going on another PC, but the RAID 5 just refused to load. I had to go to the backup.] Given that the OS with be on SSD and the magnetic media is on RAID 0, would it still make sense to go with enterprise grade drives, presuming they are more reliable that the consumer grade? Incidentally, I noticed WD now has a 4Tbyte drive whose description is similar to the Hitachi 4Tbyte. I'm leading towards using 3Tbyte since they are substantially cheaper, though Fry's occasionally discount the Hitachi 4Tbyte drives. This article in Tom's Hardware argues that Serial-Attached SCSI (SAS) HHDs are better better than SATA in the argument that Enterprise class quality is worthwhile in Enterprise environments for several reasons: http://www.tomshardware.com/us/spons...Hard-Drive-158 *TimDaniels* |
#24
|
|||
|
|||
Enterprise versus "consumer" grade drives
On 10/14/2012 8:34 AM, Arno wrote:
miso wrote: On 10/12/2012 9:58 PM, Arno wrote: miso wrote: [...] So is there any advantage to buying a Seagate Constellation versus a Baracuda? Or Ultrastar versus Deskstar? No. Unless you have a hard limit to only use one drive. "Enterprise" drives are a bit better in reliability, but RAID1 is so massively better that enterprise drives are laughable. The only place these pay of (somewhat) is if drive replacement is expensive, e.g. because somebody has dro drive to the datacenter. There, even a small imporivement in reliability can justify a larger price increase. [...] Given that the OS with be on SSD and the magnetic media is on RAID 0, would it still make sense to go with enterprise grade drives, presuming they are more reliable that the consumer grade? If you use RAID0, your data is basically doomed anyways, unless you have good backup. If you have good backup, no need to go for enterprise drives. RAID0 is basically always a bad choice except for cache and buffer applications that need high throughput, such as video-capture and editing. But RAID0 should never be used as actual longer-term storage. Arno I meant mirror, i.e. raid 1, for the hard drives. Ah, that is something different. I'll consider RAID5. I've done that before. Not with the greatest results though. When the mobo died, I was able to put the RAID 10 array in another PC and all the data was still there. The RAID5 array couldn't be recovered. I had a pretty good backup. You should test RAID recovery when you design that array, i.e. before you put data on it. Software RAID is better than hardware RAID is better than FAKE RAID. Raid 5 at least buys you something in the way of more storage with just as good security, well provided the mobo doesn't get Chinese cap disease. That should be long over now. This was an isolated incident of industrial espionage and basically all caps from that time should now be dead. Arno Tell the Samsung TV owners that Chinese bad caps are a thing of the past. http://www.capacitorindustry.com/how...death%E2%80%9D I rebuilt a raid array once. Very nerve racking. I don't know if things have changed, but back then you "delete" the bad drive and then the controller on the mobo somehow rebuilds the data on that drive. Delete is a nasty work, so I got a usb drive and backed up the unhealthy RAID array just in case. I think ultimately if I am going to be a cyber pack rat, I need to go NAS. Maybe a Drobo. |
#25
|
|||
|
|||
Enterprise versus "consumer" grade drives
Arno wrote:
You should test RAID recovery when you design that array, i.e. before you put data on it. Software RAID is better than hardware RAID is better than FAKE RAID. What's fake RAID? Raid 5 at least buys you something in the way of more storage with just as good security, well provided the mobo doesn't get Chinese cap disease. That should be long over now. This was an isolated incident of industrial espionage and basically all caps from that time should now be dead. I had a mobo get cap disease a couple of months ago, and it was used constantly. Another was late last year. -- Reply in group, but if emailing add one more zero, and remove the last word. |
#26
|
|||
|
|||
Enterprise versus "consumer" grade drives
On 10/14/2012 6:34 PM, Tom Del Rosso wrote:
Arno wrote: You should test RAID recovery when you design that array, i.e. before you put data on it. Software RAID is better than hardware RAID is better than FAKE RAID. What's fake RAID? Raid 5 at least buys you something in the way of more storage with just as good security, well provided the mobo doesn't get Chinese cap disease. That should be long over now. This was an isolated incident of industrial espionage and basically all caps from that time should now be dead. I had a mobo get cap disease a couple of months ago, and it was used constantly. Another was late last year. I assume the "reality" of RAID goes like this: 1) Real raid uses a rain controller than plugs into a card slot (the faster the better). The controller does all the hard work, removing the processing burden from the host PC. 2) Fake raid uses some chip on the mobo, in addition to software that the CPU needs to run all the time. 3)software raid is self explanatory. I've only done fake raid. The raid cards cost more than the mobo, and it is tough to justify the expense with the fake raid on the mobo already. However, in theory, when you get into these situations where the mobo fails and you have a real raid controller card, you can plug that card into another Pc and it will be able to read all your drives. [As interface slots have migrated over the years, this might not be possible in all cases. That is, the old controller needs to work in the new PC.] When you installed a fake raid, you might of had to take a step where you inserted 3rd party drivers during the installation phase. I know I did this in win2kpro, but not in win7 pro. Opensuse does a good job of being equipped with the fake raid drivers. |
#27
|
|||
|
|||
Enterprise versus "consumer" grade drives
On 10/14/2012 9:56 AM, Timothy Daniels wrote:
"miso" wrote: I haven't built a PC with hard drives in a few years and I have to say I am amazed at the hard drive market. [Note I have bought a few USB drives, so I am referring to internal drives here.] First of all, it seems everyone bought everyone else. Samsung went to Seagate. Hitachi went to WD. Fujitsu went to Toshiba, which is presume is waiting to go elsewhere. Well now I can see why the hard drive market never fell back to the pre-Thailand flood prices. There are three, no make that 2.5 suppliers. I always had the best luck with Seagate. So I do the usual market survey (translation: read Newegg reviews) and it seems Seagate now sucks. Also the 5 year warranty is 3 years, and that is on a good day. Seagate has some drive with 1 year warranties. OK, so check out WD. Hmmh, they seem to suck now too. So is there any advantage to buying a Seagate Constellation versus a Baracuda? Or Ultrastar versus Deskstar? Have you noticed some vendors selling new drives without warranties? When did that start happening? FWIW, the system I plan on building will use intel SSD for the OS. I've done two systems with intel SSD and no headaches, well other than having to pay top dollar for the SSD. [I had a Corsair SSD arrive DOA. That is my only non-intel experience.] I plan on getting two large hard drives (normal, not SSD) and running RAID0. [Raid can be a pain if the controller dies. Raid 0 may be inefficient, but at least the drives are readable without RAID. I had a mobo fail that had a RAID 10 and a Raid 5 array on it. I got the RAID 10 going on another PC, but the RAID 5 just refused to load. I had to go to the backup.] Given that the OS with be on SSD and the magnetic media is on RAID 0, would it still make sense to go with enterprise grade drives, presuming they are more reliable that the consumer grade? Incidentally, I noticed WD now has a 4Tbyte drive whose description is similar to the Hitachi 4Tbyte. I'm leading towards using 3Tbyte since they are substantially cheaper, though Fry's occasionally discount the Hitachi 4Tbyte drives. This article in Tom's Hardware argues that Serial-Attached SCSI (SAS) HHDs are better better than SATA in the argument that Enterprise class quality is worthwhile in Enterprise environments for several reasons: http://www.tomshardware.com/us/spons...Hard-Drive-158 *TimDaniels* That was good reading, though it says it is a sponsored paper, hence the heavy Seagate emphasis. A couple of sections bothered me. Under Reliability: "These added safeguards are sufficient to give enterprise drives an order of magnitude greater data protection. Whereas a Barracuda desktop drive will experience an unrecoverable read error once in every 10E14 bits read, a Constellation ES nearline drive will experience one such error in every 10E15 bits. In a three-drive RAID, this improvement would drop the chance of an unrecoverable read error from 12% down to under 2%. More volumes in the RAID and/or the use of a more fault-tolerant RAID type, minimize the risk further." I don't follow how they get 12% and 2%, and over what time period? This section notes that the MTBF for consumer drives is done at 40deg C, while enterprise drives use 60 deg C. I'm not so sure Seagate should hint that enterprise firmware is less buggy than consumer firmware. It might be true, but that sends a bad message. In any event, all software has bugs. |
#28
|
|||
|
|||
Enterprise versus "consumer" grade drives
"miso" wrote:
Timothy Daniels wrote: This article in Tom's Hardware argues that Serial-Attached SCSI (SAS) HHDs are better better than SATA in the argument that Enterprise class quality is worthwhile in Enterprise environments for several reasons: http://www.tomshardware.com/us/spons...Hard-Drive-158 That was good reading, though it says it is a sponsored paper, hence the heavy Seagate emphasis. A couple of sections bothered me. Under Reliability: "These added safeguards are sufficient to give enterprise drives an order of magnitude greater data protection. Whereas a Barracuda desktop drive will experience an unrecoverable read error once in every 10E14 bits read, a Constellation ES nearline drive will experience one such error in every 10E15 bits. In a three-drive RAID, this improvement would drop the chance of an unrecoverable read error from 12% down to under 2%. More volumes in the RAID and/or the use of a more fault-tolerant RAID type, minimize the risk further." I don't follow how they get 12% and 2%, and over what time period? Beats me! Not being an engineering paper, the math logic may have been edited out for brevity. This section notes that the MTBF for consumer drives is done at 40deg C, while enterprise drives use 60 deg C. I'm not so sure Seagate should hint that enterprise firmware is less buggy than consumer firmware. It might be true, but that sends a bad message. In any event, all software has bugs. I think that the only safe conclusion may be that Enterprise HHDs are designed *and priced* for environments that are different from Consumer environments, and that the manufacturers want to maintain that differentiation in the minds of buyers. If you don't run your desktop HHD in an Enterprise environment (i.e. up 24/7 with lots of random non-serial reads/writes alongside many other similarly-loaded HHDs), why pay for it - especially when you can protect yourself from data loss via redundancy (RAID or frequent backups)? *TimDaniels* |
#29
|
|||
|
|||
Enterprise versus "consumer" grade drives
On 10/14/2012 9:47 PM, Timothy Daniels wrote:
"miso" wrote: Timothy Daniels wrote: This article in Tom's Hardware argues that Serial-Attached SCSI (SAS) HHDs are better better than SATA in the argument that Enterprise class quality is worthwhile in Enterprise environments for several reasons: http://www.tomshardware.com/us/spons...Hard-Drive-158 That was good reading, though it says it is a sponsored paper, hence the heavy Seagate emphasis. A couple of sections bothered me. Under Reliability: "These added safeguards are sufficient to give enterprise drives an order of magnitude greater data protection. Whereas a Barracuda desktop drive will experience an unrecoverable read error once in every 10E14 bits read, a Constellation ES nearline drive will experience one such error in every 10E15 bits. In a three-drive RAID, this improvement would drop the chance of an unrecoverable read error from 12% down to under 2%. More volumes in the RAID and/or the use of a more fault-tolerant RAID type, minimize the risk further." I don't follow how they get 12% and 2%, and over what time period? Beats me! Not being an engineering paper, the math logic may have been edited out for brevity. This section notes that the MTBF for consumer drives is done at 40deg C, while enterprise drives use 60 deg C. I'm not so sure Seagate should hint that enterprise firmware is less buggy than consumer firmware. It might be true, but that sends a bad message. In any event, all software has bugs. I think that the only safe conclusion may be that Enterprise HHDs are designed *and priced* for environments that are different from Consumer environments, and that the manufacturers want to maintain that differentiation in the minds of buyers. If you don't run your desktop HHD in an Enterprise environment (i.e. up 24/7 with lots of random non-serial reads/writes alongside many other similarly-loaded HHDs), why pay for it - especially when you can protect yourself from data loss via redundancy (RAID or frequent backups)? *TimDaniels* Seagate does claim the enterprise grade drive will last longer, which is something to consider. However, mirror raid is really simple. Not efficient, but simple. You still need backups in the event the computer pees all over your data. Keeping the OSs on a SSD and data on the hard drive does make it easier to migrate to new drives. When the ghosting programs couldn't handle mirror raid, I just synced the mirrors then copied one drive to an external drive. That seems to work without lose of data, so I don't think a RAID mirror drive is too different than a non-RAID drive. Now that the ghosting programs understand fake RAID, this is less critical. |
#30
|
|||
|
|||
Enterprise versus "consumer" grade drives
On 15/10/2012 05:33, miso wrote:
On 10/14/2012 6:34 PM, Tom Del Rosso wrote: Arno wrote: You should test RAID recovery when you design that array, i.e. before you put data on it. Software RAID is better than hardware RAID is better than FAKE RAID. What's fake RAID? Raid 5 at least buys you something in the way of more storage with just as good security, well provided the mobo doesn't get Chinese cap disease. That should be long over now. This was an isolated incident of industrial espionage and basically all caps from that time should now be dead. I had a mobo get cap disease a couple of months ago, and it was used constantly. Another was late last year. I assume the "reality" of RAID goes like this: 1) Real raid uses a rain controller than plugs into a card slot (the faster the better). The controller does all the hard work, removing the processing burden from the host PC. It's not "real" raid, it's "hardware raid". A hardware raid card handles the raid on the controller card. This has some advantages, such as more efficient use of the system's IO bandwidth and the possibility of using a battery backup. It also has some disadvantages, such as the lack of flexibility (limited raid modes, no mixing of modes with the same disks, no usb disks for extra safety during replacements, etc.), the disks are tied to the particular hardware (if the board dies, you need an identical replacement to be sure of recovery of your array), and management is often poor (such as having to reboot to the raid bios, or using hideous management software). Typically, an OS can see the hardware raid as a single big drive without extra software or drivers, which is nice. But you normally need extra software for management of the raid. And the hardware raid may or may not be faster than software, depending on the type of system and the type of usage. The main cpu in a modern system is far faster at calculating parities than the cpu in a raid controller card, for example, and software solutions can often give faster raid layouts (Linux raid10,f2 on two harddisks will outperform anything a hardware raid can do with two disks in either raid0 or raid1). And of course, hardware raid cards cost a lot of money - especially with a battery backup (and they are pretty pointless without a battery). You also often have to pay extra for "advanced" features such as raid6. 2) Fake raid uses some chip on the mobo, in addition to software that the CPU needs to run all the time. Fake raid don't actually use any motherboard chips - they are basically a very limited software raid implemented in software in the bios so that you can configure it or do recovery from the bios setup screens, and the OS can boot from it. Beyond that, it requires drivers in the OS to support it (as the OS does not use the bios), and it's all run in software. Fake raid has all the disadvantages of software raid, and all the disadvantages of a really cheapo hardware raid (inflexible, tied to the one system, etc.). But if you are using an OS that doesn't have proper support for software raid (i.e., Windows, whose software raid is very limited), then it can be a convenient and easy-to-use system. 3)software raid is self explanatory. Software raid means the OS handles it. This means that the raid is only as safe as the rest of the system - you can't get the advantage of battery backed caches. And unclean shutdowns (crashes, power cuts, etc.) can lead to time-consuming checks and re-syncs, depending on your setup. You also use more IO bandwidth - if you are writing two copies of everything to raid1, your cpu has to write everything twice, rather than letting a hardware raid card do the duplication. And of course the main cpu has to do all the calculations, but that is usually a small burden on modern cpus. In return, you get an array that will work on any copy of the same OS on any hardware, and that can be hugely more flexible than hardware solutions (assuming you are using an OS with good software raid support). You can mix and match raid types to suit particular requirements, you can change things easily from within the system. With Linux (and probably other *nixs, but I haven't tried with them) you can freely mix different disks of different types and sizes within arrays, and you can re-shape and re-arrange your arrays while running. As an example, this means you can temporarily add an external USB disk to a raid5 array and re-sync it to a lopsided raid6 with all parity on the USB disk. Then you can re-arrange the raid5 disks (perhaps swapping them out for bigger disks or replacing old ones before they fail) step by step, without ever losing your redundancy. Once everything is finished, you can remove the USB disk and go back to raid5. And of course all your tools are integrated into the OS. I've only done fake raid. The raid cards cost more than the mobo, and it is tough to justify the expense with the fake raid on the mobo already. However, in theory, when you get into these situations where the mobo fails and you have a real raid controller card, you can plug that card into another Pc and it will be able to read all your drives. [As interface slots have migrated over the years, this might not be possible in all cases. That is, the old controller needs to work in the new PC.] When you installed a fake raid, you might of had to take a step where you inserted 3rd party drivers during the installation phase. I know I did this in win2kpro, but not in win7 pro. Opensuse does a good job of being equipped with the fake raid drivers. Don't use fakeraid with Linux - mdadm software raid is better in every way (except perhaps ease of setup if the distro you are using does not support it in its installer). |
Thread Tools | |
Display Modes | |
|
|
Similar Threads | ||||
Thread | Thread Starter | Forum | Replies | Last Post |
DABS "Warranty Repaired" B-Grade return. | js.b1 | UK Computer Vendors | 2 | July 12th 09 12:56 PM |
"true life" vs. "anti-glare" of Vostro 1500: What are the brightness & contrast ratios??? | Thomas G. Marshall | Dell Computers | 1 | April 11th 08 10:47 PM |
"desktop quality" (or better) drives in 2.5" laptop form factor? | phreak | Storage (alternative) | 0 | December 26th 06 09:07 PM |
Downside of changing "Max frames to render ahead"/"Prerender Limit" to 1/0? | Jeremy Reaban | Nvidia Videocards | 2 | March 31st 06 04:24 AM |
ASUS A8V & ATI AIW 9600 "inf" "thunk.exe" error message? | ByTor | AMD x86-64 Processors | 5 | January 13th 06 06:50 PM |