Cull2ArcaHeresy

Members
  • Posts

    98
  • Joined

  • Last visited

Everything posted by Cull2ArcaHeresy

  1. same issue of size option missing (so a screenshot is provided), but mine might be a full/almost full cache/array problem. If adding new disk does not clear the problem, ill create a new thread about it (datacenter feburb drive so doing preclear...not sure if 1 or 2 or 3 passes yet tho).
  2. i've had scatters going from multiple drives to multiple drives (and multiple to one) do the same thing saying that will not be transferred (idr if had space comment), so i reselected what was left and it just took multiple passes to finish. But mine was more than 1 working transfers per drive. I'll dig into history to be more specific if needed to help hunt down the bug or cause (assuming not intentional)
  3. no but there was talk about multiple pools previously (1 or 2 years ago i think), but boiled down to it being hard to identify pools that aren't "cache"
  4. updated from 6.9.2 to 6.11.5 and now in a screen session i have to run "export TERM=linux" for many commands to be able run. Without the export command i get "Error opening terminal: vt100." that some commands throw (nload for example). Also screen -r partial_session_name tab complete completes a file name from current dir instead of session number or name. Is this an issue with version/config of screen in nerdtools or 6.11 bug to report? Nerdpack had screen 4.7.0 (according to my backup at least), and the version installed with nerdtools is running as 4.9.0
  5. updating to 6.11.5 seems to have fixed this, but not marking as solved till server is up for a month without it happening again
  6. updated from 6.9.2 to 6.11.5 and now in a screen session i have to run "export TERM=linux" for all commands to run instead of "Error opening terminal: vt100." that some commands throw (nload for example). Also screen -r partial_session_name tab complete completes a file name from current dir instead of session number or name.
  7. forgot to come update this when tried moving 9202-16e to the 16x slot, in multiple reboots neither server hardware scanning thing that runs on boot nor unraid saw those drives so ended up just moving it back to the riser that has the 2 9207-8i cards that connect to the 14 server drive bays. Due to a different server issue (web ui problem a little while after booting up) i can't look up parity history, but iirc most checks have been just under 2 days this year. In the past the GPU that was in that slot was allocated to a VM, or at a different point a network card was in that slot allocated to a different VM, but that dealt with blocking off a PCIe device id (at least on the NIC) not the PCIe slot. One of those things on the "possibly investigate in the future" list...or when it is needed for some other reason. notification of your comment reminded me to reply above, but if there was an OS change, would be curious what since my speed has gone from ~120-140 down to ~80-100 (more like 80). Or if the subset of us with the issue are just having confirmation bias 😜
  8. Rebooted and generated a copy of diagnostics from web ui. At some point between 1 and 2 days the issue happened again (does seem to have happened faster this time). Generated new copy of diagnostics via cli shortly before making this post. After reboot i saw there were updates to some plugins and dockers, but for consistency did not update them yet. When back home after holiday trip i can reboot, update plugins and dockers, reboot, and wait for issue to happen again. Also having the ui back (for a short bit) can look back into what prevented me from updating the os and after resolving that will update os (unless you say not to until this issue is resolved). Also a non-optimized version of my dupe finder is running on a selection of files to "quickly" free up some space to be able to redistribute space between the drives to make sure that isn't somehow causing a problem (will probably be done tomorrow and have 3 to 4 tb to free up). rightafterreboot_raza-diagnostics-20221221-0834.zip issuehappened_raza-diagnostics-20221223-1354.zip
  9. on the plugins i have meant to look into that for a while (grab history stuff before removing), also i think i am a couple updates behind on base os. Total free space (if you include cache) is ~1.5tb, and it complains about different disks over time as stuff gets moved around. The tv_pool is full, but acts more as a level 3 cache (when there is room i change share setting to move from array to that cache and then after mover runs change it back to move from ssd cache to array). I'm hoping to add at least 16tb in the next month and then can rebalance to make it happy. I do have a script that am gonna finish writing in next week to check for duplicate files, but optimizing it to where it will have a thread hashing per drive/pool to speed it up (rough estimate of 1 to 5 tb to clear so can redistribute some and make it happy then too). This problem has been happening for a while now and a reboot fixes it for some time. Just to double check, 1) reboot, 2) grab diagnostics, 3) when issue happens again grab diagnostics and post both here? After that i can look into updating os.
  10. i've had this issue for months now (think it started early 2022 but might have been at some point in 2021). When server boots everything is fine, and for a while after it is fine. But after a few days to a couple weeks it will start having this issue. The "frame" of the web ui loads, but the content does not. Sometimes stuff will load...a few times a day if i'm lucky. Behaves the same on dedicated 10g nic (172...) and local ip (192...), but used the 172 screenshot because dark reader messes with 192 differently. Also works the same on chrome, ff, ie, edge, browser in unraid gui, and i feel like i might have tried another browser too (plus mobile chrome and ff do same). The error below screenshot is what it gives after the page sits with the frame but not content loaded for a few minutes. Dec 17 08:40:30 Raza nginx: 2022/12/17 08:40:30 [error] 35647#35647: *7626036 upstream timed out (110: Connection timed out) while reading upstream, client: 172.16.16.2, server: , request: "GET /Dashboard HTTP/1.1", upstream: "fastcgi://unix:/var/run/php5-fpm.sock:", host: "172.16.16.1", referrer: "http://172.16.16.1/Dashboard" SSH and vms (via other vnc access) and dockers and file serving all work fine, just the content of page fails to load. All the other answers i've seen have to do with nothing loading, unless i missed this issue being answered. raza-diagnostics-20221217-0850.zip
  11. with this last boot (well multiple when troubleshooting) pulseway stopped working and so then multiple reboots trying to fix/update it. Currently just saying screw it will fix pulseway later. I would assume no relation just coincidence in timing. I assume these errors would not be a "drive tray needs to be reseated into DAS" type of issue? How can i move the data off without repairing it first, or is it like it just reads from the "good" copy and ignores the errors if any (then deletes file from move) and so no prob to move? I see unraid 6.10 brings kernel up to 5.15, guessing 5.20 is a ways off then. i know server is a mess, the 3 non-cache pools are just 24tb overflows (for specific uses) to keep the main array cleaner (and over drive limit). I know i'm at the size where i need a second unraid or other server to handle all the categorical archive stuff, but that is not feasible right now so its a down the line thing (still looking at you 45 drives).
  12. Cache pool finished scrub correcting 2 errors, 2 other pools are still running with ~60 hours left and found no errors yet...but 1 pool just instantly aborts scrub. Tried it from web interface with reapir checked and nothing in log but seeing the command run, same thing when btrfs scrub status /mnt/archive_one is run from command. But if i run it from web interface with repair unchecked, i get the following in the log and it still goes right to aborted.
  13. On Reddit like 5 years ago I was reading a post that was a few years old, but can't find it again. It had to do with rtorrent (or some other p2p client) and so are on unraid using some combination of soft links and hard links. Iirc it was something like both sides used a hard link that pointed to a softlink that pointed to the actual files. Is there something like this? I currently have a share for each of p2p, tv, and movie. A seedbox does most of the actual downloading and then imports it to my server. I do have binhex's rtorrentvpn docker, but I use the seedbox mainly for speed (and currently my VPN subscription expired). The automation is that seedbox downloads it and then seedbox sonarr imports and rsync grabbed the completed torrent to local to continue seeding. I know this makes linking more complicated too, I assume I would need to write a post import/sync linking script. If I need to create a share that has root folders for pep, tv, movie I can, but would really rather not. [Edit] plex runs on local server for media
  14. Update, file integrity was running during parity check. Having fixed that setting, back to ~90 mb/s. When moving server this weekend will rearrange hba cards to test davids theory...but hit was from file integrity plugin.
  15. The same way that generally you assume a gigabit switch is not a bottleneck, i didn't even think about the LSI HBA cards being one (gigabit switch comment is about common "normal" network activity that is not mass file transfers). Looks like the pcie riser card has each of the pcie slots at 8x based on online images of the r720xd riser card. Plenty when that riser just had the 2 9207-8i cards that connect to the 14 r720xd bays. I would assume the riser gets 24 pcie lanes for the 3 8x slots, but even with that, my 2 ds4246 shelfs are connected to a 9202-16e that is in that 3rd slot with only 8 lanes. The ds4246s have 36 drives in them now, but 24 of those are pool drives now instead of array and the other 12 are part of main array (moved them to pools for multiple reasons). Since netapp designed it to use 1 cable, i'm assuming it is not restricting at all, but now thinking i should map it all out and calculate each step to see if there are other places of accidental bottleneck like putting a 16x in an 8x slot. When i reorganized the layout of the drives i did make sure the SSDs and parity drives are in the r720xd bays, as well as the 16tb disks and as many of 12tb as could still fit...really glad i did that (especially if bottleneck was big). Parity check takes 2 to 3.5 days now (or 5.625 when a bunch of random automatic rsyncs happen during it pulling hundreds of gigs from remote server for hours). But next time i shut the server down, since gpu x16 slot is open again, i'll try to remember to move the 9202-16e HBA over to it (well if it came with a full height plate). Guess i've gone from near the limit of my server to pretty much out grown it...but that upgrade has to come later down the line..and 2 to 4 more 16tb drives is needed sooner than the upgrade. Thanks for pointing that out, which could help future peoples.
  16. sort by type="Cache" then by ["poolname*"] to group ["tv_pool6"] in with ["tv_pool"] and ["cache2"] with ["cache"]? Unless you can name 2 pools the same in which case edit-oh i see the name="cache3" also now
  17. notification of reply got lost, thus why it took so long. Used the tool and grabbed the sanitized disks.ini out of it. If you need the other stuff i can check it before attaching. Btw my streamlabs share and tv share use the main array. First i filled the 2 cache pools as much as could and then set them to main array. For using unbalance that might make things weird, but i assume i am not a normal case with this. Btw monthly parity check is still going for another estimated 15 hours. I assume that it would not produce different data, but pointing it out if unraid hides something during check. disks.ini
  18. your vpn login is saved in plain text there too
  19. yes. Like the speed limits need to be set in config file to be persistent. Unless that is an old issue that no longer exists in which case this is outdated info.
  20. i have 3 pools along with cache and main array if you need it. Would have to find disks.ini and see if it contains any data that would need to be redacted before sharing.
  21. Either that was a fast update (thanks) or i've never looked at the "TOTAL" line (oops hidden in plain sight).
  22. With the new pool options, i'm trying to have different shares for different things. Currently trying to have downloads be in a share that is on a pool (pool only), and done/seeding to be on main array (mover pool->array). Initially tried having p2p share (for incoming and everything else) mount as /data and also shares mounting as /data/downloading and /data/DONE but that caused conflicts where some stuff went to proper share and some went to p2p/[downloading or DONE]. Now do not have a base /data mount, but mounting everything needed...which seems to work ok but save to cannot escape whatever share it is currently in, and auto tools does not always succeeded. Is there a "right" way to do this? If can get this working, the idea is to start splitting off different categories into their own pools (like all defcon/infocon torrents to be in a separate pool). Considered adding links within p2p share to the other shares, but that just seemed like a really bad idea.
  23. Unless there is a longer log elsewhere, the one in the ui doesn't go back far enough. But those were the last 3 lines to be transferred to disk 7, who currently has 49.2 kb free. Source disk still has 36.8 gb on it, belonging to those 3 lines. So i'm assuming that was the issue. Mover was disabled, but guess something else was writing to that disk during transfers. All lines before were green (disk 7), and after those 3 lines (disk 12) were also green checks. One feedback (been meaning to give for a while). A select/deselect all checkbox would be great. Currently working on clearing off 8 4tb drives to move into a pool instead of main array, and each time it requires select source + shares, then unselecting over 20 destination disks. The auto-select-all is great when all/most of your disks are <70% full, but otherwise a quick unselect all would be nice to have. Either way thanks for the tool
  24. do the colors of the icons indicate a problem or just ui bug?