Jump to content


  • Content Count

  • Joined

  • Last visited

Community Reputation

8 Neutral

About bman

  • Rank
    Advanced Member


  • Gender

Recent Profile Visitors

The recent visitors block is disabled and is not being shown to other users.

  1. Okay thanks for the explanation. This is a useful plugin, thanks very much for your efforts!
  2. I'd like to request this goes one further in having the ability to hide individual slots, or mark them in some way as N/A (not available) with their own colour. Reasoning is because sometimes through poor quality control or over-use/abuse, individual slots on a backplane may no longer be functional. I have two Supermicro 24-bay chassis where this is the case and I need to remember not to try to use those slots, so I physically mark them with tape, but it would be nice to have the plugin able to also mark individual slots as defunct, so when I am several miles away from the physical server, I can easily know which slot(s) are faulty for making upgrade or drive swap plans before visiting the site.
  3. tbh I haven't given Intel SSDs a fair shake since premature failure on 6 of 8 purchased 520 series products left students without their days' worth of video footage upon returning. Swore off them because even though the warranty was good, the reliability wasn't. Seeing the price of the D3-4510 960G versus the competition has me rethinking things, albeit for different use cases. I think you're on the right track with that one.
  4. $0.02 on write endurance - Samsung 850 or 860 Pro series wins. I've only seen one dud out of a couple dozen I've used over the past several years. Write endurance seems to be a Sammy strongpoint: https://techreport.com/review/27909/the-ssd-endurance-experiment-theyre-all-dead
  5. Agreed: return to sender. Had lots of drives dead on arrival in my time. Had a drive not quite failed but prone to future problems based upon SMART information and inconsistent data transfer rates after the first cycle was complete, so I cycled it through preclear for several days until it barfed so I had no hassle at the returns desk.
  6. preclear is a script you can run on your unRAID server to read from and write to every sector on a hard drive, then read again to verify the writes happened properly. It is an optional step in preparing a drive for use with a server (I preclear all new drives whether for unRAID or not). The original purpose of the script is now irrelevant, but it was also very useful for weeding out weak hard drives before you entrusted them with your data, which presumably you do not wish to lose. By forcing a drive to read and write every sector multiple times (I use a cycle count of 3 - for 12TB drives that would likely take a week) you are forcing the drive to find any sectors that are perhaps bad from the factory. It's a burn-in test. For me it has found a few drives that I was able to return within the 15 days my local computer shop allows me to walk in and replace a failed drive with them. After that time period I'd have to deal with the manufacturer's RMA process which takes much longer. It's also useful for assessing the relative health of a drive that has been used before. It erases everything on the drive, so it's something you only do before or in between "uses". preclear is unrelated to SMART data and testing except to say that SMART counters will definitely be altered after preclear has been run. SMART test results after the completion of preclear can be decent indicators of whether you can rely on your drive to keep your data safe, but nothing is foolproof. SMART data will never reveal the full story: Almost regardless of the SMART data, your drive may live a very long life, or it may fail a week later without any hint of trouble beforehand. The majority of the time (75%, give or take, if you believe Backblaze data on the subject) SMART results are good enough to determine the relative health of a hard drive.
  7. The total capacity of your data array is irrelevant to parity, so long as you follow the rule of parity being at least as large as your largest data drive. You can have twenty-eight 8TB drives (for a capacity of 224TB) all protected by your single 8TB parity drive, if you so choose. 28 data disks is the current unRAID limit.
  8. There are exceptions to every rule. I have a 320GB cache drive that's running great and is 11 years old. Most of my IDE drives from 2004 still worked fine when I tossed them out in 2015. I have SCSI drives from the 1980's that are running fine today. Today I want drives that last 7 years or more and I am hoping the better 'ratings' of the higher-priced drives will prove to be of value in terms of longevity and total cost of ownership. I get the insurance thing, law of averages, conspiracy theory and share some of the same thoughts. Taking WD's lineup for example, purple for... surveillance? Give me a break. That's different than NAS how? I used to have dozens of WD green drives spinning, but I can count on one hand how many lasted longer than 5 years, and over 70% of them failed before 4 years. I have (so far) had much better luck with WD Red, Black and Gold drives, though few of them are older than 4 years at this point.
  9. Put better, 5-year warranty means you won't have to shell out more money to replace the drive any time soon Drives fail whenever they want to based upon lots of factors, but in general those warranted for five years have passed more stringent tests at the factory and can be expected to last longer, else the manufacturer wouldn't put a longer warranty on them in the first place.
  10. In general, what you're after is a different server OS than unRAID. You can use unRAID as a spot to host virtual machines, and in that way get the "different" OS that is required for your performance needs. In short, though, if you're not using unRAID for its storage abilities, you may be better served installing another OS on your bare metal build. The author of the video is already running unRAID for other reasons, but unRAID is not what is enabling the speed between his demo systems. Rather, the speeds demonstrated come from the Windows, Linux or OS X machines (whichever he has booted into at the time) in concert with RAM used as disks and 10GbE network interface cards. Proper network setup is necessary too, naturally.
  11. While I do not recommend you run without parity, technically this is also possible. If I hadn't used parity, I can recall at least 6 times I would have lost multiple TB of data in the past, if that helps you make the decision on parity use. You may be surprised how short a time period your initial array size will last (given your intended use), so I'd say save some cash for a couple 5-year-warranted 8TB drives as your next upgrade, then you can just add such "large" drives to your array as needed. Going from 2TB to 4, then 6, then 8 for your parity drive as an example, means you've purchased at least two too many hard drives in a short (maybe 2 year) time period. By buying 5-year-warranted drives (WD Gold or Black or Red Pro as examples) you can rest assured the drives won't fail anytime soon, and by buying large it takes much longer to fill your chassis with hard drives. (I experienced some regret as had to remove perfectly functional, not-very-old 2TB drives in favour of larger ones just because I hadn't any more physical space for more drives.)
  12. Quite right; It was not obvious where to go from my point of not assigning drives to the array, but the idea is that once you've got them outside the control of unRAID you can build a virtual machine to handle them in any speedy fashion you choose.
  13. Hello Ettiene, There are upgrade procedures to go to v6. Did you try to implement the method here? https://lime-technology.com/wiki/UnRAID_6/Upgrade_Instructions For your case you'd need to perform a "clean install" s outlined from Tom's post here: https://lime-technology.com/forums/topic/39381-upgrading-unraid-5-to-unraid-6/ You have not included any details of your server. Can you indicate which components you have working with 4.7? Version 6 will not work with all the same hardware as version 4.7 does. (I for example had to buy a new motherboard, CPU and RAM since my 2005-era server was too old for v6)
  14. As nuhll said, you can run VMs of any Linux flavour you like (or other operating systems as desired) and get all kinds of performance. But the storage subsystem unRAID uses is not suited to the speeds you're looking for. Using drives outside the array (by simply not assigning physical drives to unRAID's storage pool) with any VMs you might like to run will get you the speeds you're looking for. You can have a hybrid system in this way, using unRAID for archive type storage (write-once, read many concept) and also using it as a spot to host virtual machines from. Or you can ignore the core functionality of unRAID and just use it to host VMs and fast drives + network access, similar to what Linus did in his 7-gamers-on-one-PC video as mentioned above.
  15. Dual parity may not be 100% successful in cases where you lose a drive or two simultaneously, because failing hard drives can cause controller lock-ups and other inconsistencies which may in turn negatively affect your otherwise healthy drives. Once any healthy drive gets bad bits written to it in the throes of disk failure, parity is no longer of any use to you in rebuilding when you replace your failed drive(s). I'd suggest you get a functional backup running for the data you consider most important (the Kodi/database stuff, at least) then just follow the upgrade procedures to get to v6. My own transition was painless. MySQL runs fine on 6, but I do not know how to transfer an existing database to a new server so cannot comment on what troubles may show up there. Your cache drive ostensibly could (should?) be empty so whether its format gets changed or not during the upgrade is irrelevant with respect to being able to reformat it to work again after a downgrade. If important data reside on your cache drive, back them up too before you start.