PSteward

Members
  • Posts

    31
  • Joined

  • Last visited

Everything posted by PSteward

  1. Updated to add a check for the liquidctl initialize, and fixed it so it doesn't set speed if already correct, plus additional checks for zero RPM. Just update your docker.
  2. Hey Greg, Never noticed! I put that in because I noticed sometimes when I rebooted or started the server not all fans would work (showing 0 RPM) - honestly it doesn't really fix that issue though. It is REALLY important to check the logs at reboot to ensure that all fans are reporting more than 0 RPM and if not shutdown - then start again (reboot doesn't always resolve). I never had the issue once all fans are running but only on the occasional startup or reboot (I may do that every few months, so definitely important to verify). I'll review the coding and see if I can make some improvements based on how its been working of late - in addition at one point I'd like to extract the data into a plugin and show it on the interface (% + Speed).
  3. Recently I have been having on-going issues where Unraid will hang/freeze. This causes all services to freeze (GUI direct on server, SSH, web interface, dockers) and unfortunately even with a syslog server there is nothing posting to the log when the freeze happens. I am assuming it's a hardware issue (CPU temps get up around 85c when I have a specific docker running and a VM at the same time), but not sure about that 100% (I am going to change my cooler / thermal paste this week however). My question is - if Unraid is hung up on a specific docker or VM, is it possible to have it force close that (is there a script that can still run?) and recover from the freeze up? I can still ping my server, and ssh TRIES to get in but just sits there, same with the GUI - I can type the username! I really wonder if there is a script that recover this type of issue. Thanks, --Phil
  4. Yep that was it, Q24-27 are on point. I used localhost to access the qbitorrent hosting the VPN for this docker and then added the ports for my external qbittorrent (on a different machine). Thanks!
  5. Next time I'm on that server I will! FYI it's not that it doesn't Auto Mount (can't mount what's not online...) - it shows "Remote Server is OFFLINE" until I push it to search for servers, then they all reconnect.
  6. Hello! I tried using this container as a network for another however it appears the it blocks the LAN network (can ping items but not get to them). For example I put Lidarr to use the container as it's network (which works) but now Lidarr cannot access the local clients. I tried "curl http://LAN_address:port" and it times out - is this normal? Thanks!
  7. Quick question: SMB Shares never reconnect after a reboot, but if I go to "Add Remote SMB" then click on the "windows logo" then "Next" then press "Search For Servers" - they all reconnect! What is it doing when it searches? I would like to add a user script to run after boot to reconnect the SMB drives since UD doesn't appear to be doing it at the moment. Thanks!
  8. Hello! Server just completely froze up, couldn't even SSH or control it on my KVM (though it did have video output showing the login screen). It did respond to pings but all dockers, shares, ssh, emhttp were gone and I was forced to restart. Diag attached. Any ideas of the cause? Version: 6.12.3 Thanks, --Phil homeserver-diagnostics-20230824-1248.zip
  9. Hello! I set this up today but have an issue when adding files going through the Cloudflare Tunnel URL: When I access internally it works fine for uploading (so http://localip:port vs https://external.domain.name). EDIT: There is a solution on Github: https://github.com/filebrowser/filebrowser/issues/2588#issuecomment-1664335304
  10. Hey Jorge, I don't have any sessions open typically but power failures don't wait for you to close your ssh session!!! Besides I have a the plugin to kill them anyways. At this point it looks like an issue with the version 6.12.2 as per that other post. Before updating I didn't have these issues when I tested. Thanks, --Phil
  11. Hey Jorge, It appears to be a bug: I cannot even stop my array...: Jul 10 09:00:37 HOMESERVER emhttpd: Unmounting disks... Jul 10 09:00:37 HOMESERVER emhttpd: shcmd (432479): umount /mnt/cache Jul 10 09:00:37 HOMESERVER root: umount: /mnt/cache: target is busy. Jul 10 09:00:37 HOMESERVER emhttpd: shcmd (432479): exit status: 32 Jul 10 09:00:37 HOMESERVER emhttpd: Retry unmounting disk share(s)... lsof shows nothing open for /mnt/cache Thanks, --Phil
  12. Hey Jorge, There is no kill scripts to ensure a clean shutdown? Power outages are an unknown variable, you never know when or what is happening with a server at that time, what's the point of UPS if the server can't even kill open sessions to shut down cleanly? Thanks, --Phil
  13. Hey Jorge, What if there is an open ssh session preventing shutdown/reboot or something else? Is there anything that the shutdown script will not be able to overcome? As a note I had previously set up a timeout that worked but it appears that somewhere between 6.11 and 6.12.2 my settings were reset. Thanks, --Phil
  14. Hey Jorge, Wouldn't that not fix the underlying issue? Why did the server shutdown unclean - as in the test it didn't lose power? Thanks, --Phil
  15. Recently lost power and had an unclean shutdown, I just assumed the server timed out trying to shut down before the UPS ran out of power... Today I tested that theory and the server shut itself down while the UPS was still providing power and it still had an unclean shutdown! Diagnostics attached, please help! Thanks, --Phil homeserver-diagnostics-20230709-1606.zip
  16. Hello! I've been trying to get nextcloud-aio-mastercontainer going without success. I'm stuck now at two issues, one is with nextcloud talk: "Message: Server error: `POST http://localhost/v1.41/containers/nextcloud-aio-talk/start` resulted in a `500 Internal Server Error` " And if I remove that add-on, nextcloud still won't start and throws a: error: connection to server at "nextcloud-aio-database" (172.19.0.5), port 5432 failed: FATAL: password authentication failed for user "oc_nextcloud" I've scrubbed the install a handful of times, and verified the timezone. I've been using Swag for reverse proxy (NPM gave issues), all around tried so much without success to get this thing going. Any thoughts on obvious things to check would be great. Thanks, --Phil
  17. Hey Ginhead, Do you have NZXT RGB & Fan Controller or equiv that works with the liquidctl docker? If you do: Go to the docker tab, click Add Container. Name: liquidtemp Repository: pstewardyul/liquidtemp:latest Privileged: ON Click on Add another Path, Port, Variable, Label or Device: Config Type: Variable Name: Thresholds Key: THRESHOLDS_ENV Value: 20,25,30,35,40,45 Save. Click on Add another Path, Port, Variable, Label or Device: Config Type: Variable Name: Fan Speeds Key: FAN_SPEEDS_ENV Value: 20,30,40,50,60,100 Save. Click on Add another Path, Port, Variable, Label or Device: Config Type: Variable Name: Excluded Drives from temp calc Key: EXCLUDED_DRIVES_ENV Value: {ENTER WHAT EVER DRIVES YOU DON'T WANT TO CHECK TEMP,. example "/dev/sda" } Save. Click on Add another Path, Port, Variable, Label or Device: Config Type: Variable Name: Fans Key: FAN_QUANTITY_ENV Value: {ENTER QUANTITY OF FANS UNDER CONTROL, example "3" } Save. Click Done, after it loads, check the logs for its run status.
  18. FYI: I'm not using his container. Why would this be considered a hddtemp issue and not an unRAID issue? I assume reading drive information should not cause spun down drives to break your unRAID array.
  19. I have been using hddtemp in a docker to read drive temperatures and have noticed that when I spin down drives and then let the hddtemp try to read it's temperature it knocks a drive out of the array (it will show up in assigned devices). I have replicated it this a few times, and to stop it from occuring now I check if the drive is on standby (with smartmontools) before using hddtemp to read the temp. Drive is a 10TB Seagate, but note it won't happen with all Seagate drives, just appears to be this one (so far). Steps to cause issue: Spin down drive via unRAID interface Use 'hddtemp -n /dev/sdX' on drive -- Drive shows still spun down on interface and appears with new sdX assignment in Unassigned Devices -- Steps to fix/restore: Stop array (if possible, if not reboot) New config Start array homeserver-diagnostics-20230523-1318.zip
  20. Hello! I've been trying to find a solution to the issue of linux and reading/controlling fans, I stumbled upon liquidctl and the existing docker got me on the path to build something myself that reads hard drive temps, averages them and sets fan speed based on that. I am using the NZXT RGB & Fan Controller - AC-CRFR0-B1-6 and focusing only on controlling the fan speed (fan1, fan2, fan3). Here is a docker: pstewardyul/liquidtemp:latest Variables needed to set are: THRESHOLDS_ENV - Set the temperature thresholds that will change the fan speeds - DEFAULT "20,25,30,35,40,45" FAN_SPEEDS_ENV - Set the speeds that will change with the threshold temperatures - DEFAULT "20,30,40,50,60,100" EXCLUDED_DRIVES_ENV - If you need to exclude drives from the average calculation - NO DEFAULT - EXAMPLE "/dev/sda,/dev/sdb" FAN_QUANTITY_ENV - Set to 1, 2, 3 (how many you are connecting to your controller) - DEFAULT "3" The container requires "privileged" to read/control related hardware. Also recommend to set "--log-opt max-size=50m" in extra parameters to limit the log size as every minute the script will run to check/update fan speed, these changes are viewable in the log. If you don't set any variables - no worries - it will load defaults (except the log limit). This an AMD64 docker image, not sure how to work with the other platforms, and just started with making scripts and building docker this afternoon so even though it works on my system, not sure of others! FYI - The script just works with /dev/sdX not NVME, also some /dev/sdX don't produce the smart output compatible (like my kingston MSATA). The script can be found at "/usr/local/bin/fan_control.sh". EDIT: Changed from hddtemp (causing issues with spun down drives) to smartctl and added more error checking. --Phil
  21. So to conclude to this, I haven't had any more read errors. I upgraded my power supply and assume all previous READ errors were related to running all drives at once since all other times everything worked fine but obviously didn't pull that amount of wattage.
  22. I stopped what I was doing and did new config and then selected the option to re-use the existing parity before starting the array. Now I am running the parity check again (error correcting) and here I am: Total size:16 TB Elapsed time:3 hours, 32 minutes Current position:2.41 TB (15.1 %) Estimated speed:179.5 MB/sec Estimated finish:21 hours, 2 minutes Sync errors corrected:3 As a note, I haven't physically changed anything (not in the same city as my server anyways!), BUT I did spin down ALL off my un-assigned drives. So I am leaning towards not getting enough power and already ordered a 1200w power supply for absolute overkill just in case.
  23. Thanks for the response. I will investigate power issues, but to be clear there is more usage during a parity check/rebuild than during a parity build? My parity build had all drives running full for 1.5 days without issue.
  24. Ok, more info. During the rebuild it gave me read errors on disk1 (one of tons): May 12 19:16:53 HOMESERVER kernel: md: disk1 read error, sector=221271952 SO I plugged this into badblocks (badblocks -v /dev/md1/ RANGE OF SECTORS) AND there are zero 0 bad blocks found and no read errors. Data is good, Smart test is good, drive is good, badblock read is good only conclusion I have is that is a system bug with Unraid parity?