dtempleton

Members
  • Posts

    167
  • Joined

  • Last visited

Everything posted by dtempleton

  1. I'll try rebooting safe mode but right now the server is running and in use; just can't access the GUI. But an update; This URL seems to access it OK this morning (I thought I tried this yesterday): http://192.168.1.220:8080/Main While the one setup with myUNRAID.net doesn't: https://192-168-1-220.eb4a424017fa08cd2cc8d2bbd9cd528f6c202fc6.myunraid.net:8443/Main (I altered this specific key to post) or it didn't until just a minute ago and now the HTML loads but no info from the server loads. 10 minutes now and no array or pool data. at the same time http://lime-technology.com or http://lime-technology.com.index flash a loading page then send a blank page without error. I think something is going on with our friends' server. Anyway; I'll wait to see if it fixes itself, but it makes me a bit leery of being dependent on the myunraid.net service yet.
  2. My GUI won't load the main page, but instead gives a 503 nginx errror, I rebooted to the same problem. Im pasting the diagnostics zip file below, but the syslog lists hundreds of these: Jan 22 14:49:26 Tower nginx: 2023/01/22 14:49:26 [error] 6094#6094: *6 limiting requests, excess: 20.456 by zone "authlimit", client: 192.168.1.101, server: hash.myunraid.net, request: "GET /login HTTP/2.0", host: "hash.myunraid.net:8443", referrer: "https://hash.myunraid.net:8443/Settings/Userscripts" Jan 22 14:49:27 Tower nginx: 2023/01/22 14:49:27 [error] 6094#6094: *25 limiting requests, excess: 20.602 by zone "authlimit", client: 192.168.1.101, server: hash.myunraid.net, request: "GET /login HTTP/2.0", host: "hash.myunraid.net:8443", referrer: "https://hash.myunraid.net:8443/Main" Jan 22 14:49:28 Tower nginx: 2023/01/22 14:49:28 [error] 6094#6094: *6 limiting requests, excess: 20.473 by zone "authlimit", client: 192.168.1.101, server: hash.myunraid.net, request: "GET /login HTTP/2.0", host: "hash.myunraid.net:8443", referrer: "https://hash.myunraid.net:8443/Main" Jan 22 14:49:30 Tower nginx: 2023/01/22 14:49:30 [error] 6094#6094: *6 limiting requests, excess: 20.463 by zone "authlimit", client: 192.168.1.101, server: hash.myunraid.net, request: "GET /login HTTP/2.0", host: "hash.myunraid.net:8443", referrer: "https://hash.myunraid.net:8443/Settings/Userscripts" Jan 22 14:49:30 Tower nginx: 2023/01/22 14:49:30 [error] 6094#6094: *6 limiting requests, excess: 20.110 by zone "authlimit", client: 192.168.1.101, server: hash.myunraid.net, request: "GET /login HTTP/2.0", host: "hash.myunraid.net:8443", referrer: "https://hash.myunraid.net:8443/Main" Jan 22 14:49:30 Tower nginx: 2023/01/22 14:49:30 [error] 6094#6094: *25 limiting requests, excess: 20.102 by zone "authlimit", client: 192.168.1.101, server: hash.myunraid.net, request: "GET /login HTTP/2.0", host: "hash.myunraid.net:8443", referrer: "https://hash.myunraid.net:8443/Main" I also found that this link : http://lime-technology.com/wiki/index.php?title=Troubleshooting#Capturing_your_syslog gives a Cloudfare error from Limetech. Could this be a server problem from myunraid.net? Thanks! Dennis tower-diagnostics-20230122-1504.zip I see the reference to myunraid.net which I installed a few days ago but it seemed to work OK
  3. OK, I figured this one out myself. The symptoms of crashing were not related to the parity rebuild, but coincidental. The syslog held the key- this several-times repeating error: ntpd[1135]: kernel reports TIME_ERROR: 0x41: Clock Unsynchronized I found a post on Reddit that suggested turning off ntpd auto update, but that did nothing. However one post there said "check your motherboard battery", so I did that. Replacing the battery on the motherboard prevented any subsequent UNRAID OS crashes. Hope this helps someone down the line. Dennis
  4. Hi; I'm running 6.11 I've had three unraid OS crashes while trying to do a data rebuild from a failed drive. After about 3 hours of an 8 hour expected rebuild the system just goes offline. The server seems to work ok for these hours until the crash. I can't access it via ssh or the GUI, and can't get a diagnostic report. I have to restart with the reboot button. The last time it was in safe mode. Edit: I learned how to turn on "Mirror syslog to flash" and I'll report back... I finally for a syslog from the crash, it's below. I don't see many errors except a failed time synch. The crash seems to come at about the same percent data rebuild, so it could be related to a disk access error but I don't see evidence of that. I'll try booting the server but not mounting the array Thanks for any advice in advance syslog20221215.zip
  5. My server is working now, thanks for your input JorgeB and others. I bought this 5 port card: https://www.amazon.com/dp/B08BHZQVP7?psc=1&ref=ppx_yo2ov_dt_b_product_details and all seems to be ok except several of my drives report UMDA CRC errors (that seem to be permanent but unimportant). Regarding the previously purchased card from Ziyituod, it's a mess. It was listed at Amazon as originally listed as an ASMedia controller, then with the same part number a second version was listed on Amazon that was clearly a Marvell controller. When I pulled the controller the board actually showed neither of the Ziyituod model numbers and no identifiers at all. In future if I get a card that doesn't look like that advertised I'll send it right back. Thanks for helping put this back together. Dennis
  6. sorry I posted the one after reboot. The earlier one had blown up with read error warnings, 128mb! I zipped it and add it herebelow I'll report after Wednesday when the new JMicro controller comes in. syslog.1.txt.zip
  7. It looks like that Marvell 9215 controller is not on the approved list now, but the forum is full of questions. I just ordered a different one: https://www.amazon.com/gp/product/B08BHZQVP7/ref=ppx_yo_dt_b_asin_title_o00_s00?ie=UTF8&psc=1 PCIe SATA Card, Electop SATA III 6 Gbps Expansion Controller, JMB585/SATA 3.0 Non-Raid ,Support 5 Ports with 5 SATA Cables, Standard & Low Profile Bracket for Desktop PC that has the JMicro controller that is recommended here: https://forums.unraid.net/topic/102010-recommended-controllers-for-unraid/ won't know for a few days if it fixes things ChatNoir: Thanks, our messages passed each other simultaneously.
  8. Thanks; reboot looks the same Here is the new syslog syslog.zip and diagnostics file tower-diagnostics-20221127-1441.zip I realize that all of the drive errors I'm seeing are controlled by one controller; this one: https://www.amazon.com/gp/product/B07SZDK6CZ/ref=ppx_yo_dt_b_search_asin_title?ie=UTF8&psc=1 Ziyituod PCIe SATA Card, 4 Port with 4 SATA Cable, SATA Controller Expansion Card with Low Profile Bracket, Marvell 9215 Non-Raid, Boot as System Disk its a Marvell 9215 device, about 7 months old; I thought that was a usable one. I'll go look at the list of usable controllers
  9. My server has been working great for the past 7-8 months, no problems at all, mainly as a media server and TimeMachine. One morning the server was not accessible and I had to do a hard reboot with the power button. The array reported a bad Parity2 drive, which I removed from the array and then did a parity rebuild that completed. I reformatted the failed Parity drive (XFS) and the Attributes seemed OK. It's my newest drive, an 8 GB WD drive about a year old, maybe a bit more. I rebooted at least once and now have a failed array drive which shows that the contents are emulated (the data seems to be there). Thinking the reformatted 8G drive was good, I substituted it for the failed disk 3 and rebooted, which initiated a rebuild that never finished. Now the syslog is full of disk0 read errors, the tail is below I have some new drives coming, but its unclear that the issue is actually drive fail. I tried the extended SMART test on the former P2 drive, but it stopped before completion. The short SMART test reports "no such device". The drive attributes "could not be read". Today I tried to get a diagnostics file but the script starts and never completes. I attach a diagnostics file from the time of the parity drive fail though and can try to get a current one if anyone has suggestions on how to get it. I'll format the new drives as soon as I can, but it will be days before they're ready. At this point its not clear how the array drive can be emulated, if it cant read from the Parity drive (!?) Thanks for any help Dennis Here's the hardware profile:HW profile.xml.zip tower-smart-20221125-2237.zip tower-diagnostics-20221125-1150.zip
  10. I get a similar warning using nzbget. I have mappings to container directories /data ; /media ; and /config. /media and /config are self explanatory but I'm not sure what goes in /data I put it inside /appdata but maybe it needs to map to my /nzbget download directory??
  11. I'm having problems getting SSH setup under 6.9.2 It was working until I upgraded to 6.9.2 from 6.3.5 when I lost the ability to log in automagically with authorized keys. I figured that I would sort that out later and logged in successfully with the root password. I tried several things to get the authorized keys to work, and eventually renamed the /boot/config/ssh/ directory so that a new one would be created and I could set that one up from scratch. More on that later. In the meantime I lost the ability to log into the server as root, even though the root password allowed login to the webGUI. in the /boot/config/sshd_config file I see the following lines (about 27) AllowUsers root #LoginGraceTime 2m PermitRootLogin prohibit-password PermitRootLogin prohibit-password #StrictModes yes MaxAuthTries 6 #MaxSessions 10 I didn't know what 'prohibit-password' meant but from https://ubuntuforums.org/showthread.php?t=2359172 I learned that this is the new default for ssh_config to force automatic login, i.e. prohibiting the use of root passwords. OK, but for now I need the password option, so I changed the lines using nano to: AllowUsers root #LoginGraceTime 2m #PermitRootLogin prohibit-password PermitRootLogin yes #StrictModes yes then stopped and restarted SSH with the plugin. I still can't use a root password, and find that the sshd_config has been changed back to what it was before: # Authentication: AllowUsers root #LoginGraceTime 2m PermitRootLogin prohibit-password PermitRootLogin prohibit-password #StrictModes yes MaxAuthTries 6 #MaxSessions 10 What is happening? Is something writing a default sshd_config file here on every SSH restart, perhaps the SSH plugin? (Edit: sure enough deleting the plugin allows 'PermitRootLogin yes' to persist and login works) If so this is an important enough setting that it might should be accessible by the plugin GUI as an option Secondly, I note that in UnRAID /root/.ssh redirects to /boot/config/ssh/root but, the ssh keys that are generated are placed in /boot/config/ssh not in the root subdirectory, Is that as planned? I did try copying the keys to .../root and then to my client authorized keys but no joy so far. I'm more flummoxed by not being able to use a root password. Thanks for all you do! Dennis
  12. Sorry for what seems like a lame question: Does the ssh plugin have an interface? Nothing happens when I click on the green icon on the plugins page. Does the plugin do all of its work 'behind the scenes' and we set selections by editing the sshd_config file in /etc/ssh ? thanks
  13. Re h265ize First, thanks for making this docker image. It should be very useful, but hasn't been working for me. The input directory is monitored and the file queue loaded, and the script is loaded into the temp directory just fine. The log files say that the files are processed, but gives an error that "ffmpeg exited with code 1". There is an output version with size 0b made in the temp folder. I tried reinstalling ffmpeg within the container, but found that ffmpeg version is already the latest version. Second, I really support the comment above about passing options through during the build of the docker image. For example, the bit depth of 8 is default, but the h265ize page says that this can be changed with the variable --bitdepth (https://github.com/FallingSnow/h265ize) thanks for your advice! [log] [h265ize]: Processing ABCD-1080p Remux.mkv... [h265ize]: [verbose] Encoding started at Wed, 12:02 PM [h265ize]: [verbose] Getting metadata... [h265ize]: [verbose] Resolving video streams... [h265ize]: [verbose] Upconverting... [h265ize]: [verbose] Setting video bit depth... [h265ize]: [verbose] Mapping streams... [h265ize]: English [h265ize]: Spanish; Castilian [h265ize]: [verbose] Detecting auto crop... [h265ize]: Crop Detection: 1/12 [1000D[1A[2K[h265ize]: Crop Detection: 2/12 [1000D[1A[2K[h265ize]: Crop Detection: 3/12 [1000D[1A[2K[h265ize]: Crop Detection: 4/12 [1000D[1A[2K[h265ize]: Crop Detection: 5/12 [1000D[1A[2K[h265ize]: Crop Detection: 6/12 [1000D[1A[2K[h265ize]: Crop Detection: 7/12 [1000D[1A[2K[h265ize]: Crop Detection: 8/12 [1000D[1A[2K[h265ize]: Crop Detection: 9/12 [1000D[1A[2K[h265ize]: Crop Detection: 10/12 [1000D[1A[2K[h265ize]: Crop Detection: 11/12 [1000D[1A[2K[h265ize]: Crop Detection: 12/12 [h265ize]: [alert] Output will be cropped to 1920x1072. Originally 1920x1080 [h265ize]: [verbose] Mapping HE Audio streams... [h265ize]: [verbose] Setting ffmpeg settings... [h265ize]: [verbose] Encoding video... [h265ize]: ffmpeg exited with code 1: [h265ize]: [verbose] Finished encoding at Wed, 12:03 PM (a minute)
  14. I came to posses three small Toshiba SSDs (256g) when I dismantled my hackintosh in favor of an upgraded board and CPU for unraid (that went smoothly!) I decided to put them into new pools that would separate vms and appdata, and use two of them for a pool for the traditional cache buffer. One of these drives, the old UNRAID cache, had begun throwing CRC errors (about one a week, it seems, and now 50) so I thought that being in a pool with a sounder identical SSD would be a safe thing. I mean, you have two redundant drives and you lose one you're still OK, right? So I started having problems with rsync for files on this pool ('read only file system' errors) and the log was full of btrfs errors. Uh Oh. On reboot, both the SSDs in the pool are unmountable with no file system. All the data (not much, mostly testing) was just gone. Reformatting the SSDs using default btrfs remounted them and re-formed the pool. SMART reports show one is still fine and the other has a few more CRC errors. I can't see what happened. If one SSD poops out, how did that cause the other one to lose its FS? I'm thinking that with an iffy SSD I'm better removing it that trying to use it to create some redundancy. Any thoughts? Thanks!
  15. Thank you; it makes sense now. For others stumbling across this here is what I learned. /mnt/cache is data actually on the cache drive(s)/pool /mnt/user0 (now deprecated, but I think valuable) is data actually on the array drives /mnt/user is a logical summation of data on both cache pool and data array The recommended way of moving data between cache and array is to specify preferences in shares and then to invoke the Mover. I of course understood that regular data files written to /mnt/user behaved like that when the cache pool is used for writing new data, I just thought that writing to /mnt/cache was different. I actually found a 0 byte file in /mnt/user0/appdata that is not in /mnt/cache/appdata 🙃 so managed to clean appdata up after all. Thanks again
  16. I have a data array of 10 drives (xfs) and a small ssd for a half-empty cache (btrfs), in a Pro system running 6.9.2 as a recent upgrade from 6.3.5. For now I have all data except appdata and docker.image written only to the data arrays. My docker apps all point to /mnt/cache/appdata (though some now unused dockers had pointed to /mnt/user/appdata in error). For some reason I can't understand the contents of /mnt/cache/ show up also in /mnt/user/, though of course /mnt/data/ has all of the other array data. Regular data written to /mnt/user/Media don't get copied to cache, just appdata and docker.img. I don't think I specified this mirror. I checked and they aren't soft links to each other, and at least for the similarly mirrored docker.img the inum of the files in the two locations are different, so they're not hardlinked. If I create a junk directory in one it shows up instantly and if disappears if I delete its twin. It feels like data is written to both cache and array at the same time, but I *think* I have not set the cache behavior this way. On the other hand I can't see right now how to tell the cache drive in the webGUI how to behave. The shares are all 'array only' at present or 'prefer cache'. I have mover set to move monthly but since data isn't currently being written to the cache it isn't moving much. Can anyone see where I might have specified this behavior? I don't want it since it likely slows down the cache to have to write to the array as well, and it consumes 100GB+ of array needlessly. Thanks in advance
  17. OK, I followed a DL link in a post about 6.5.2 but it instead installed LATEST (6.9) There were a few hiccups but less than an hour it was working better than before. I was upgrading to access SMB timemachine for my new Mac, but overall everything is snappier and more reliable. For others worried about upgrades, my experience is that it is worth it. Thanks for your encouragement!
  18. Yes, I've been following that very helpful post, to 6.5.x stable Wasn't certain it was sufficient for 6.9.x
  19. cuz so many things get broken with each upgrade. I''m following the thread of upgradeto 6.5.3 and there are probably 12 things that have to be attended to. How many more are there to 6.9? Even if I choose to upgrade to the current stable I need to know how to do this manually.
  20. I'm trying to upgrade incrementally from unRAID 6.3.5 by moving to 6.5.3 first Attempting to upgrade the OS via the Plugin updater gives an error that the zip file is not available anymore from AWS. (downloading: https://s3.amazonaws.com/dnld.lime-technology.com/stable/unRAIDServer-6.5.3-x86_64.zip ... failed (Invalid URL / Server error response) OK, I'll try upgrading manually. But I can't find a downloadable copy of the 6.5.3 OS. Can you advise a URL that will lead to this file? Also, where does this file go? Do I put it on the literal boot/flash drive or can I put it in the Boot directory of the running server? Sorry this is likely very obvious, but I'm not seeing it. Thanks! Dennis
  21. In trying to make a new docker instance using docker-compose I encountered this error that prevented the image download: ERROR: Get https://lscr.io/v1/_ping: x509: certificate has expired or is not yet valid Googling around this seems to be related to an incorrect clock setting on the local machine, or a certificate that has actually expired. The fixes I cound (one on CentOS) reference a directory not present in UnRAID. How can I reset my clock on UnRAID or obtain a refreshed certificate? TIA Dennis
  22. I'm resurrecting an ignored server and preparing for an update. HOWEVER, as it is now the WebUI doesn't load in my normal environment. I checked if this might be due to a plugin conflict by rebooting in Safe Mode, and sure enough, the WebUI loads in Safe Mode. Now I want to remove (or disable) suspect plugins one or a few at a time to find the conflict. How can I do this? I looked in /usr/local/emhttp/plugins and (in safe mode) only Dynamix apps and UnRAID are listed. The go script auto_installs those so marked in /boot/packages/ Can I just relocate these to say /packages_inactive to get them not to load, or am I missing a simpler way.? TIA