Jump to content

mathomas3

Members
  • Posts

    359
  • Joined

  • Last visited

Everything posted by mathomas3

  1. You might also check the setting for the share that you are copying too... and ensure that the 'Use cache pool' is set to yes... sounds like you currently have it set to only
  2. From my experience, disable docker and envoke the mover which should move everything that is not currently being accessed... in short... disable everything that would have access to the cache drive to include vms if that's where you have them running from and then envoke mover I learned this the hard way when I moved somethings around in my system... but that's the way I would go about it
  3. Keep this in mind... files are written as a whole on a single drive(while also having the parity drive) so if you lost 2 disks... you would still have access to the files on the remaining disks... A suggestion that I tell everyone. When looking at your shares and what write method to use... I suggest using the 'fill-up' choice... Files will be written on disk 1 till it's full then it will go to the next... decreases wear on the other disks and you dont have to wait for 3 disks to spin up...
  4. You might consider it... It would give peace of mind... next time it happens plug up a monitor to the machine and see if the console is responsive For me. The machine was still up and running but the console was blank... a restart later with a UPS helped...
  5. The last time this happened to me was when we had a power hiccup... Is your server behind a UPS?
  6. Also something to consider... unRAID is hard on hdds... Imo fill it up and leave it like that... The excess read/writes ever time kill a hdd... Write to the drive and be done with it...
  7. I wouldn't say it's time to pass on it... Just from my experience these drive have failed regularly... Of the 20 drives in my array I only have two 5tb drive remaining... I will be replacing them soon
  8. I have never received a used hdd with wiped smart history before... is that a thing?!? my two golden rules for used hdds are... that they have to be SAS disks and also enterprise disks... I have a good stack of 3,5,6tb reds and greens that all died a young death... 8tb sas drives have done me well... that's why I have 26 of them... 8 currently waiting to be added to the array
  9. A question... I have 3 SSDs that I have in my 24 bay DAS... when I tried using these three drives in the old server chassis they came up with a weird ID, thus UR wasnt able to identify them as my cache/app drives... once I added them into the DAS everything was normal again... I would like to reclaim these three slots for data disks moving forward... Given that this is a new server with a new RAID controller... what are the odds that I will have this issue again... What caused that issue to begin with? Should I try to move these drives into the R430 chassis and I run into this same issue... what can I do to correct the issue? Thank you
  10. that or looking up the smart history on the drive...I have only been buying used HDDs of recent... havent had issues yet 14 HDDs into it... if your happy with the price and the drive doesnt give you many red flags... roll with it... I recently spent 45$ for each of my 8tb drives... and I bought 10 of them... yea... I should have only bought 5 of them given my current restrictions on adding addl HDDs... but meh... it was a good price
  11. I have around x16 8tb drives in my setup... also have a stack of x8 more should I have to expand... found a good deal on them for $45 each... I have not had a singe one of them fail...
  12. It was only after I had the system open that I noticed that the existing memory had matching specs with the memory that I was installing... happy little 16gb surprise
  13. Just got everything together and it's up and running... so happy... and now I think I have enough cores
  14. can you manually set the IP in the console and access the server after this?
  15. While I am not an expert here... This is what I would try... replace the CMOS battery... validate that time/date in the bios is close to the real time... restart the machine...
  16. I question the PSU then... I would connect everything up again and boot into the GUI... and while your there run a parity check
  17. Hmmm so if you boot to gui mode it's fine but not a normal boot... when you booted into the GUI did you have everything still plugged in and running when you do this? have you added any disks into the array recently? Have you made any changes to the system recently... updates?
  18. I would suggest that you take this to this docker's support thread... You really havent given much details here that would allow us to help you..
  19. IE... you play for shipping plus a Starbucks GC? I would be willing to provide proof that the card works... I would need to use some of my older hdds+and under powered PSU to do so... the choice is yours
  20. Heck... I have a sas controller that I was using 8 months ago(with breakout cables) that I would be willing to part with should you and I come to an agreeable arrangement...
  21. another good reason to have a sas controller... should you decide to upgrade the mobo/CPU combo... the sas controller gives you greater flexibility on the combo you decide on...
  22. Dont think about using WD green/red drives... If you are... you and I will have words
  23. The reason I like SAS controllers is that they are NOT consumer grade products... SAS is enterprise... I have had too many sata devices fail(over 10 disks) on me... I refuse to add them to my array due to failures... each sas contoller port can handle 8 disks each... normally they have two ports... thus they can handle 16hdd... this board was a quick search here... just searching google for sas+unraid should provide you with some good options... I would like to counter Trurl's post... "fewer larger disks instead of more smaller disks" I have 2x5tb disks and 20x8tb disks... the 8tb disks have never failed on me... and should I want to upgrade to something like a 16tb or 18tb... I would have to buy 3hdd just to increase the space by 8tb or 10tb... for me it makes more sense to add another 8tb disk then buying something newer... My point being is... dont be afraid to use renewed enterprise SAS HDD, they are built to last... The prices on these drives tend to be lower because SAS is enterprise thus many are not interested/know about these drives yet there are many on the market the sas controller should be a good fit for you due to your 15hdd requirement(breakout cables will be needed) and will provide you with better motherboard options(not needing 6sata ports on the MOBO) ... I would suggest this... find a good deal on a stack of hdds(used or not that are NOT consumer grade) 12tb or more and the build system around that... Im stuck on the 8tb drives as they are around 50 bucks a piece... IMO in the long run... you will find SAS drives are cheaper and they will last longer... (if you willing to use renewed drives and or find a good deal) this was a long post... im drunk, sorry... but in the ten years of using unraid... I strongly recommend using SAS drives... if you would like me to count the number of my sata vs sas drive failures I can... but it would be something like this... SASx0 SATAx12+
  24. I think you have a pretty good idea as to what you need/have... I would still suggest a sas controller to make things easier for you down the road should you need more space... I would think that your cache drives could be stored inside the case vs using a tray for it and allowing you more room for hdds... Having a spare is always nice and you can avoid the stress of preclearing/testing a drive having a cache drive comes in handy when there are high r/rw operations. I think you could do without... Though having a app pool with two drives ssd would be a good idea One other suggestion I would make... when making the share permission and what not... I would suggest the allocation setting be set to 'fill-up'... reason being is that when you start transfer data to the pool... only 1 disk + parity drives are spinning and I think it's less wear and tear on the drives, because once the drive is full it only has to read from then on... it's a theory
×
×
  • Create New...