Jump to content

BackBlaze Storage Pod 3.0 [March 2014 Update: Now it's at 4.0]


neilt0

Recommended Posts

Good info!

 

Seagate – ST3000DM001?9YN166 (slightly higher failure rate)

 

Unfortunately for me I just added one of these to my array as parity. Oh well that's why we have redundancy.

At least it's parity and not data. Parity is the most important speed wise (all writes to the protected array are effected) and the least important failure wise (no data is directly in jeopardy if it fails). Any two drive simultaneous failure will lose data, but if one of those drives is the parity, you only lost 1 drive worth of data. If two data drives fail, you lost them both, and parity is worthless as a recovery.
Link to comment

Good info!

 

Seagate – ST3000DM001?9YN166 (slightly higher failure rate)

 

Unfortunately for me I just added one of these to my array as parity. Oh well that's why we have redundancy.

At least it's parity and not data. Parity is the most important speed wise (all writes to the protected array are effected) and the least important failure wise (no data is directly in jeopardy if it fails). Any two drive simultaneous failure will lose data, but if one of those drives is the parity, you only lost 1 drive worth of data. If two data drives fail, you lost them both, and parity is worthless as a recovery.

 

Very true, I kind of knew that going into it with that drive (had read that it might be a little more prone to failure) but it was the cheapest at the time and 7200 and I knew I was going to be using it for parity. So not all that bad. Curious which drives they prefer. Probably which ever is the cheapest at the time due to the amount they purchase.

Link to comment
Curious which drives they prefer. Probably which ever is the cheapest at the time due to the amount they purchase.

No, they are very much in to reliability as well as cost-effectiveness. Earlier, they posted stats on drive failures and the Hitachi drives had a much lower failure rate than any other.

Link to comment

I found this note on vibration rather interesting:

 

The key advantage of the drive bay assemblies is to reduce vibration. These assemblies not only keep the drives still, they also keep them firmly seated in the backplanes. Over the past several months we have tested different models of drives in the new drive bay assemblies and we have seen a dramatic improvement in overall system performance along with lower drive failure rates.

 

I can see how vibration could cause connectors to work loose causing an apparent drive failure, but they appear to be implying that the vibrations might be causing a reduction in drive performance as well.

 

Regards,

 

Stephen

Link to comment

I found this note on vibration rather interesting:

 

I can see how vibration could cause connectors to work loose causing an apparent drive failure, but they appear to be implying that the vibrations might be causing a reduction in drive performance as well.

 

Regards,

 

Stephen

 

Link to comment

I found this note on vibration rather interesting:

 

I can see how vibration could cause connectors to work loose causing an apparent drive failure, but they appear to be implying that the vibrations might be causing a reduction in drive performance as well.

 

Regards,

 

Stephen

 

 

Alright now that was very interesting. I wonder what effect a sustained vibration would cause.

Link to comment

I found this note on vibration rather interesting:

 

I can see how vibration could cause connectors to work loose causing an apparent drive failure, but they appear to be implying that the vibrations might be causing a reduction in drive performance as well.

 

Regards,

 

Stephen

 

Wow....that's mind blowing.

 

 

Link to comment
... they appear to be implying that the vibrations might be causing a reduction in drive performance as well.

 

What the Sun Video is showing is that it's the latency, or seek time, which suffers.  I guess that this is understandable - the head is moving across the disk surface, trying to position accurately onto one track.  If the whole device is shaking, then the head may well over or undershoot, requiring a read to establish actual track position, followed by a step to the correct track.

Link to comment

Vibration has a multitude of negative impact, reliability, performance, power, even noise (a positive feedback loop). In a chassis like backblaze 1.0, traditional RAID would fail out drives for response timeout. The vibration effects heads, forcing them off track, requiring reseek, since the entire stripe is delayed, the drive is dropped. All the reseeking drives up the power and adds vibrations.

 

This area has a lengthy patent list...

 

Link to comment

Vibration has a multitude of negative impact, reliability, performance, power, even noise (a positive feedback loop). In a chassis like backblaze 1.0, traditional RAID would fail out drives for response timeout. The vibration effects heads, forcing them off track, requiring reseek, since the entire stripe is delayed, the drive is dropped. All the reseeking drives up the power and adds vibrations.

 

This area has a lengthy patent list...

 

So I guess we don't want to locate a rack of disks in a tap dance studio :-)

 

Stephen

 

Link to comment
  • 2 weeks later...
  • 1 year later...

Will unraid ever support 45 drives?

 

Myk

Tom has said it would require an almost top down rewrite, because of the 4 character drive designations required. (/dev/sdaa etc.) So... probably not. It seems like the current drive limit matches the physical characteristics of relatively cheap server hardware, moving beyond 24 drives is pretty expensive. Plus, with drive sizes continuing to increase regularly, I think the main concern is adding protection with dual parity. (insert large drive size here) is a crapload of data to lose. Increasing the spindle count increases the risk of dual failure more than I'm comfortable with.
Link to comment

.... I think the main concern is adding protection with dual parity. (insert large drive size here) is a crapload of data to lose. Increasing the spindle count increases the risk of dual failure more than I'm comfortable with.

 

I totally agree. I decided a while ago I won't add more than 10 disks while we still only have a single parity disk. 4TB, and now 5TBs, is a shitload of data to lose if two disks fail. If my storage needs ever outpace how much can be stored on a 10 disk array of the largest size disks I'll just use my currently unused second license and setup a second array.

Link to comment

Same here I reduced my array to 8 data disks because I did have a two disks go bad after a short in a shredder tripped the mains fuse.  Luckily I had backups of all photos and documents (and managed to copy most things off of one of the drives and rebuild sort of the other),  so only lost a few movies and corrupted a few more.

 

Double parity is definitely high on my wishlist for feature enhancements for unraid.

 

Sent from my Nexus 4 using Tapatalk

 

 

Link to comment

Right.  And I've said before, and others I think too, that the fastest way to sort of get what we want is to add the ability to create two arrays in a single instance.  For now, it would just be possible to use Xen to create another instance of Unraid but that means using pass thru of SATA and as we've been told that is not ideal.

 

But dual array in a single box would be the fastest way to >24 drives and to afford more parity protection without the efforts to create p+q or some other dual parity method.

 

Sent from my XT1060 using Tapatalk

 

 

Link to comment

Join the conversation

You can post now and register later. If you have an account, sign in now to post with your account.
Note: Your post will require moderator approval before it will be visible.

Guest
Reply to this topic...

×   Pasted as rich text.   Restore formatting

  Only 75 emoji are allowed.

×   Your link has been automatically embedded.   Display as a link instead

×   Your previous content has been restored.   Clear editor

×   You cannot paste images directly. Upload or insert images from URL.

×
×
  • Create New...