Jump to content

tiwing

Members
  • Posts

    98
  • Joined

Everything posted by tiwing

  1. Figured another reply since I see this morning nothing has changed in terms of errors and offline drives. But this picture ... There are a lot of reads and writes for a drive that is supposedly offline (parity), and both the drives are now showing up in UD ... that is 16 hours since reboot. Edit: also, ran extended test as above, then for giggles started extended test on my other 10TB drive. It's still running 8 hours later. So the above extended test on drive 7 I assume failed but didn't show a failed error.
  2. Ran extended test, but unsure where to get results. All I see is this, but I still have read errors showing on drive 7 (sdf) and parity is still disabled (sdg). Now drive 2 (sdi) is showing one read error as well. indicates a controller failure perhaps???
  3. LOL I attached above first. not so good at following directions apparently (9 year old?? LOL I'm actually 47...) thanks! kscs-fvm2-diagnostics-20201111-1620.zip
  4. Hi, I have found my parity disk is offline, and another disk has read errors. Both are 10TB Reds and both were purchased about 18 months ago and have been flawless. this is a lightly used home server. All disks are connected to an LSI SAS controller (LSI00301 SAS 9207-8i) which has been perfect for a year. Both failed and read error drives test fine on a quick smart test. I followed other advice on the forums after a successful SMART test and shut down the server, re-seated the card, powered back up, unassigned the parity, started the array, stopped the array, assigned the parity, started the array .... and it's still disabled. I currently have the webGUI working, my VMs have started (running on cache drive), and all my dockers have started fine. Unassigned Devices is working fine. I'm a rookie, please treat me like I'm 9 years old What do I do next? What other info can I provide for the kind members here to help out? thank you! Tiwing edit: attaching diagnostic zip kscs-fvm2-diagnostics-20201111-1620.zip
  5. Hi, I have the same issue - link above no longer works and I suck at searching - tried, no luck. Any updated link, or revised instructions for this scenario? thanks tiwing
  6. hi all, Fix common problems recently identified that I have two shares with different cases. Media, and media. the lower case share is the one I have always used for 5 years and is the share referenced in ALL my dockers. However, the capital M share is the share that's showing up on my shares page within Unraid. So even if I delete the empty Media folder using Krusader, whenever I restart the box the Media share reappears, and I get the Fix Common Problems warning again. CAN I delete the current Media share (says it's empty) and reboot the server, and will it magically work? Magically.... Is there a way to rename the actual share in Unraid to be lower case? Doing through the Shares UI does not work. I've looked at (not changed) the .cfg files on the USB drive in /boot/config/shares and the cfg file is lower case. I can't see anywhere else to look, and don't want to mess stuff up... Help, please? thanks
  7. kscs-fvm2-diagnostics-20200109-0741.zip Hi! Thankfully I've never had to post diagnostics here before. But twice now I've had what appears to be a random restart. Uptime is showing right now as 4 hours 50 mins which is exactly how long parity check has been running. (I just woke up). This is not a business machine or mission critical and all important data is backed up locally and offsite, so I'm not super concerned. Only recent changes are installation of tdarr docker and removal of two NSF shares to another box in unassigned devices - both restarts have been after tdarr. I have noticed once memory usage going to 100% and all cores maxed. Restarting tdarr solved that. Thoughts? Thanks! tiwing
  8. I was planing to get one of those fans that goes in a pci slot and exhausts out the back to go in the slot right below the controller, I heard the cards run hot. Appreciate the info. Like this bit not this. Need different power. https://www.amazon.ca/dp/B0000510SS/ref=as_li_ss_tl?slotNum=7&ie=UTF8&linkCode=g12&linkId=a9e6b57b44fb0be28f2049cf4005d38e&imprToken=6-8g74E5x8UusuJUkoEMQQ&tag=healprodrevca-20
  9. That's frigging awesome info, thank you! I didn't realize fan was pushing into the card. Makes sense. Actually having the graphics above the controller might help a little to draw hot air away at least to circulate it. Maybe I'll just put it back. Cheers
  10. Hi, sort of an unraid question in that I'm using unraid. more of a general computer question. My mobo has 2 16x slots. One is currently for graphics, the other will be for the new LSI 9207-8i card that's coming. While I'm preparing for it, I want to move graphics to the lower 16x slot so it's not blowing hot air straight on the LSI card. Question is does it matter which 16x slot is used for which purpose? Graphics was installed in the top 16x slot and I've always just left it there. Lenovo Thinkstation S20 now in a different case. specs https://support.lenovo.com/ca/en/solutions/pd005642 not my site but some good pics of the internals https://thinkorama.wordpress.com/thinkstation/thinkstation-s20/thinkstation-s20-part-4/ Thanks.
  11. Attached is the smart report. Please let me know if this isn't the right thing to attach? thanks t kscs-fvm2-smart-20191020-1557.zip kscs-fvm2-diagnostics-20191020-2003.zip
  12. Got it. Thanks! Very glad it works this way! Will post once I get back home.
  13. Hi all. My server emailed me at 1:30 this morning. Had drive 6 go out on me, again. First was an old 4TB red and I just assumed it had gone bad. So replaced with a 10TB Red. less than 85 hours later it has also failed after a full rebuild (24 hours) plus a couple days usage. I swapped cables between two drives and looked still not good (would is self-correct during power on in unraid or once it's got a red X it requires manual intervention?). Considering raw_read_error_rate is zero I've unassigned the drive and reassigned and it's currently rebuilding. (All critical data is backed up twice in my house and offsite. Non critical data is backed up only in my house, and Plex stuff I just don't care about isn't backed up, but is on the array.) By swapping cables and the drive didn't come back to life, I think I've eliminated the expansion card and the cable. (? right ?) I've had 2 unraid boxes running for 4+ years 24/7. Had 10 drives running in one for over a year. Nothing else changed in the last 6 months.. and this is the first failure ever. Basic server info - Thinkstation s20 in a new tower with Xeon 3550 CPU, 24GB EEC, plugged into UPS. All drive are WD reds of various ages and sizes. Question: based on below SMART report (do you need anything else to help analyze?) - should I be in touch with Western Digital, go back the store for a replacement, or is this just a *shrug* "weird s*&% happens" kind of thing and see how it goes for another week? Thanks for your help. tiwing. SMART Attributes Data Structure revision number: 16 Vendor Specific SMART Attributes with Thresholds: ID# ATTRIBUTE_NAME FLAGS VALUE WORST THRESH FAIL RAW_VALUE 1 Raw_Read_Error_Rate PO-R-- 100 100 016 - 0 2 Throughput_Performance --S--- 130 130 054 - 108 3 Spin_Up_Time POS--- 100 100 024 - 0 4 Start_Stop_Count -O--C- 100 100 000 - 3 5 Reallocated_Sector_Ct PO--CK 100 100 005 - 0 7 Seek_Error_Rate -O-R-- 100 100 067 - 0 8 Seek_Time_Performance --S--- 128 128 020 - 18 9 Power_On_Hours -O--C- 100 100 000 - 85 10 Spin_Retry_Count -O--C- 100 100 060 - 0 12 Power_Cycle_Count -O--CK 100 100 000 - 3 22 Helium_Level PO---K 100 100 025 - 100 192 Power-Off_Retract_Count -O--CK 100 100 000 - 140 193 Load_Cycle_Count -O--C- 100 100 000 - 140 194 Temperature_Celsius -O---- 191 191 000 - 34 (Min/Max 23/41) 196 Reallocated_Event_Count -O--CK 100 100 000 - 0 197 Current_Pending_Sector -O---K 100 100 000 - 0 198 Offline_Uncorrectable ---R-- 100 100 000 - 0 199 UDMA_CRC_Error_Count -O-R-- 200 200 000 - 0 Device Statistics (GP Log 0x04) Page Offset Size Value Flags Description 0x01 ===== = = === == General Statistics (rev 1) == 0x01 0x008 4 3 --- Lifetime Power-On Resets 0x01 0x010 4 85 --- Power-on Hours 0x01 0x018 6 20935128920 --- Logical Sectors Written 0x01 0x020 6 20521877 --- Number of Write Commands 0x01 0x028 6 1958493232 --- Logical Sectors Read 0x01 0x030 6 4981555 --- Number of Read Commands 0x01 0x038 6 307138550 --- Date and Time TimeStamp 0x03 ===== = = === == Rotating Media Statistics (rev 1) == 0x03 0x008 4 51 --- Spindle Motor Power-on Hours 0x03 0x010 4 51 --- Head Flying Hours 0x03 0x018 4 140 --- Head Load Events 0x03 0x020 4 0 --- Number of Reallocated Logical Sectors 0x03 0x028 4 0 --- Read Recovery Attempts 0x03 0x030 4 0 --- Number of Mechanical Start Failures 0x04 ===== = = === == General Errors Statistics (rev 1) == 0x04 0x008 4 0 --- Number of Reported Uncorrectable Errors 0x04 0x010 4 0 --- Resets Between Cmd Acceptance and Completion 0x05 ===== = = === == Temperature Statistics (rev 1) == 0x05 0x008 1 34 --- Current Temperature 0x05 0x010 1 35 N-- Average Short Term Temperature 0x05 0x018 1 - N-- Average Long Term Temperature 0x05 0x020 1 41 --- Highest Temperature 0x05 0x028 1 23 --- Lowest Temperature 0x05 0x030 1 40 N-- Highest Average Short Term Temperature 0x05 0x038 1 25 N-- Lowest Average Short Term Temperature 0x05 0x040 1 - N-- Highest Average Long Term Temperature 0x05 0x048 1 - N-- Lowest Average Long Term Temperature 0x05 0x050 4 0 --- Time in Over-Temperature 0x05 0x058 1 65 --- Specified Maximum Operating Temperature 0x05 0x060 4 0 --- Time in Under-Temperature 0x05 0x068 1 0 --- Specified Minimum Operating Temperature 0x06 ===== = = === == Transport Statistics (rev 1) == 0x06 0x008 4 5 --- Number of Hardware Resets 0x06 0x010 4 0 --- Number of ASR Events 0x06 0x018 4 0 --- Number of Interface CRC Errors 0xff ===== = = === == Vendor Specific Statistics (rev 1) == |||_ C monitored condition met ||__ D supports DSN |___ N normalized value Pending Defects log (GP Log 0x0c) No Defects Logged SATA Phy Event Counters (GP Log 0x11) ID Size Value Description 0x0001 2 0 Command failed due to ICRC error 0x0002 2 0 R_ERR response for data FIS 0x0003 2 0 R_ERR response for device-to-host data FIS 0x0004 2 0 R_ERR response for host-to-device data FIS 0x0005 2 0 R_ERR response for non-data FIS 0x0006 2 0 R_ERR response for device-to-host non-data FIS 0x0007 2 0 R_ERR response for host-to-device non-data FIS 0x0008 2 0 Device-to-host non-data FIS retries 0x0009 2 3 Transition from drive PhyRdy to drive PhyNRdy 0x000a 2 2 Device-to-host register FISes sent due to a COMRESET 0x000b 2 0 CRC errors within host-to-device FIS 0x000d 2 0 Non-CRC errors within host-to-device FIS
  14. write cached is OK. I suspect it's all related to the 1x slots. Saving my pennies for the LSI SAS card.. it's not critical to get this "fixed" so I'll prob wait for a year or so. Thank you all for your help. Mods, feel free to close the thread...
  15. thanks for the suggestions - will dig into write cache in the next few days and will post up if I find anything out. Thanks!
  16. Hi, I've recently expanded the number of drives in my system and now see parity rebuilds happening at 1/2 the speed it did before. Hoping for some help on where to physically connect each drive please! Old parity rebuild was at 100/120 MB/s, hoping to achieve that again. Base Computer came from a Lenovo thinkstation s20 with quad core xeon. 24GB EEC Ram installed. https://support.lenovo.com/ca/en/solutions/pd005642 5 onboard SATA 2 connectors. (Based on speed even of SATA2 I don't think this would be my bottleneck as sata 2 is theoretically maxed at 300MB/s) 2x 4 port SATA cards - 1: IOCrest SI-PEX40064, 2: QNINE PCIe SATA Card . Both report Marvell chipsets but don't know which ones. Both cards are supposed to be sata 3, and are installed in 1x slots. I'm wide open to buying new cards if that is my issue. Suggestions please. Drives: All Spinning disks are WD Reds of various ages, but all are SATA 3. Cache Drives: 2x 256 GB SSD BTRFS (Samsung Evo). My large writes (mp4/mkv mostly) write directly to the array so speed is not critical here. Plex and all my dockers are here, and one Windows VM, none of which require huge speed. Parity: 1 x 10TB Spinning disks in array: 1 x 10TB, 4 x 4TB, 2 x 3TB, 1 x 2TB All upgrades in future will be to 10TB Reds as I need more space. not planning to expand to more drives. Outside array (unassigned devices): 1x 256 GB SSD, 1x 2TB. I don't care how fast these drives are - they're temp drives for video editing and one Windows VM on SSD. Can you please suggest where to connect each drive for maximum array read performance? I'll be unable to read or respond for a few days... I look forward to your suggestions! thanks. Tiwing.
  17. Thanks to those who helped! After a couple more hours of testing and remembering to turn off advanced docker settings to see the logs, I realized that couldn't get certs because by default all my traffic goes out through VPN. Once I created a WAN rule for the letsencrypt docker it all worked great. Thanks again!
  18. removed the docker, deleted appdata, and repulled. there are no logs..!! (in the /log folder). port 80 is forwarded to the right IP:80... "This site can’t be reached". There has to be something so basic that I'm missing.
  19. Please please ..... - letsencrypt running on static ip on my lan 192.168.13.11 . - all settings are default - I put my dynamic DNS name in the domain name field. - router port forwarded 9999 to 80 at 192.168.13.11 - confirmed dynamic dns service forwards to my IP because I have other ports forwarded no problem. But want to replace all my port forwarding with letsencrypt - sonarr, for instance running on 192.168.13.45:8989 - I have set base URL in sonarr to /sonarr and restarted - I have not touched proxy.conf on one machine - win10/chrome returns a 404 in top left corner of the screen on another machine - win10/chrome returns "site cannot be reached" I'm so confused. My nginx.conf file looks like this but of course it's wrong... I have another 8 dockers I want to set up the same - some don't have a base URL setting such as deluge - is it possible to help with that as well? server { server_name _; location ^~ /sonarr { # enable the next two lines for http auth #auth_basic "Restricted"; #auth_basic_user_file /config/nginx/.htpasswd; # enable the next two lines for ldap auth, also customize and enable ldap.conf in the default conf #auth_request /auth; #error_page 401 =200 /login; #include /config/nginx/proxy.conf; #resolver 192.168.13.45 valid=30s; #set $upstream_sonarr sonarr; proxy_pass http://192.168.13.45:8989; } location ^~ /sonarr/api { #include /config/nginx/proxy.conf; #resolver 192.168.13.45 valid=30s; #set $upstream_sonarr sonarr; proxy_pass http://192.168.13.45:8989; } }
  20. ahhh. ok that makes sense. Do you know if WOL magic packets will work over a bridged wifi connection (I've read they won't but that might be old news)? I haven't bought my second router yet for bridge so haven't been able to test. thanks.
  21. I'm about 3 years into unraid. Never had 5.x installed, always had 6.x server. So far experience has been awesome. Flawless except for when I screwed up things myself. Looking for suggestions and things I might have missed. Primary server consists of 5 misc drives of 4, 3, and 2tb for a 10tb array + single parity. 256GB dual cache drives, no unassigned devices. 24GB EEC ram. 1 VM running on cache, Plex and handbrake in dockers, and about 4TB of media. Drives are a mix of 1 and 3 years old. Server is lightly used in a home environment. Parity set to calculate monthly. Shares - appdata, cachedisk, system, docker, and vm cache=prefer. media (contains all movies, music) cache=no. All other shares (personal files) cache=yes. Mover set to run every 6 hours. local user logins are matched on the server for user ID and PW, and all share are explicitly set to secure or private depending on need. Sensitive shares are set to hidden export. only one VM exists and is on cache disk and manages all backups and is my torrent downloader. VM runs 24/7. Plex uses /tmp for transcoding installed plugins: trash bin, krusader, community applications, activestreams, ssd trim, proftpd, dynamix schedules Offsite Backup server is 3 drives for 6tb + single parity and will grow as the primary space fills up. 12GB ram. No cache drives since this is a backup server. But Plex, handbrake, and a copy of the VM are on the array (not running) in case they're needed in a pinch until my primary server can be saved - I can handle slow writes and parity drive going nuts for a few days. Backup is a combination of daily and weekly managed by the windows 10 VM of all files. Server is currently accessible through a point to point vpn, but is being moved to a bridged wifi connection in my neighbors basement. Local backup 4TB USB drive attached to my router - nightly backups managed by windows 10 VM of all NON-media Things I messed up: - installed original unraid server on a PARTITIONED usb thumb drive. duh. still haven't figured that one out but that server is stuck at v 6.5.0 until I clear space on the drive. That's actually this afternoon's project. - when I took the cache drives out of my backup server I lost all my dockers in the GUI, but configs were still there on the array so a quick install and they came back. - I messed around with unassigned devices, but never got it working right and figured everything should be protected anyhow. But my experience has been awesome so far. Very glad I made the switch from freenas to unraid - seems like it was the right time, and have never looked back. So - what can I do better? what have I missed? cheers Tiwing -
  22. I've been trying to find something to wake up the server at a specific time, but all I found so far was WOL and BIOS level stuff, both which won't work in my case. Any plugins or hidden system settings to accomplish this? Why: My headless backup server is located in my neighbor's basement connected to my network via wifi bridge (so WOL won't work right?) And I have a habit of messing around with backup settings and timing and don't want to big him or pull the server out every time I want to change/test something. Ideally would be able to just log in to the interface, change a setting, and log out... Thanks Tiwing
  23. Hi All, I'm using some older hardware in my unraid box and trying to figure out the best way to physically connect drives. Motherboard uses SATA1. An expansion card uses SATA2. Currently: Moherboard has 4TB Parity, two 3TB and 1 2TB data drives connected Expansion card has both 256GB SSD cache drives and final 2TB data drive. System is an 8 core Xeon with 24GB Ram. Upgrading hardware is not an option except possibly the expansion card if it's worth it in the slower/older slots. Cache drives are allowed for all shares except my "media" share - movies/tv/music with my mover set to 12 hour intervals (it's not a heavy use server) Cache drives also house 1 Windows 10 VM and docker for Plex and Handbrake. I'm only asking because sometimes when copying large files or multiple files to the box it copies at ~300 MB/s for the first Gig or so, then drops to zero for a while, then resumes at 50-80MB/s for a while, then drops to 5MB/s for a while, etc. Copies never fail and files are never corrupt. What's the "right" way? Cheers.
  24. I want to provide an update to this thread. The problem i was having seems to have been solved! I have had both unraid boxes running fine for over 90 days now with no problems whatsoever. I found a possible solution in another thread (can't find it again now to link to it), It appears the issue is caused by running a windows 10 VM connected to a wireless printer (on my servers I have unraid 6.1.9). Hard wiring the printer to the network seems to have solved the problem. No idea why it worked that way, but for anyone else who has the same kind of intermittent failure, try running a network cable to your printer instead of using it via wifi. This also explains why one server went down weekly, and the other less frequently - it was due to the frequency of printing at different locations. Hopefully this post is useful to someone else who might have the same kind of problem. Cheers, Tiwing.
×
×
  • Create New...