schreibman

Members
  • Posts

    30
  • Joined

  • Last visited

Everything posted by schreibman

  1. have 6.12.8 w/ WARN: Multi mapping detected! attached "tower-diagnostics" tower-diagnostics-20240303-1657.zip
  2. Any ideas on what causes this error? "[radarr] tar verification failed!" [21.01.2024 09:48:56][❌][radarr] tar verification failed! Tar said: tar: from member names; /*stdin*\ : Read error (39) : premature end; tar: Unexpected EOF in archive; tar: Child returned status 1; tar: Error is not recoverable: exiting now Line 1114: [21.01.2024 09:45:35][debug][radarr] Container got excludes! Line 1115: /mnt/user/appdata/radarr/MediaCovers/ Line 1116: /mnt/user/appdata/radarr/Backups/ Line 1117: /mnt/user/appdata/radarr/logs/ Line 1118: [21.01.2024 09:45:35][ℹ️][radarr] Calculated volumes to back up: /mnt/user/appdata/radarr Line 1119: [21.01.2024 09:45:35][debug][radarr] Target archive: /mnt/remotes/w/backup/ab_20240121_094432/radarr.tar.zst Line 1120: [21.01.2024 09:45:35][debug][radarr] Generated tar command: --exclude '/mnt/user/appdata/radarr/MediaCovers' --exclude '/mnt/user/appdata/radarr/Backups' --exclude '/mnt/user/appdata/radarr/logs' -c -P -I zstdmt -f '/mnt/remotes/w/backup/ab_20240121_094432/radarr.tar.zst' '/mnt/user/appdata/radarr' Line 1121: [21.01.2024 09:45:35][ℹ️][radarr] Backing up radarr... Line 1122: [21.01.2024 09:48:27][debug][radarr] Tar out: Line 1123: [21.01.2024 09:48:27][ℹ️][radarr] Backup created without issues Line 1124: [21.01.2024 09:48:27][ℹ️][radarr] Verifying backup... Line 1125: [21.01.2024 09:48:27][debug][radarr] Final verify command: --exclude '/mnt/user/appdata/radarr/MediaCovers' --exclude '/mnt/user/appdata/radarr/Backups' --exclude '/mnt/user/appdata/radarr/logs' --diff -f '/mnt/remotes/w/backup/ab_20240121_094432/radarr.tar.zst' '/mnt/user/appdata/radarr' Line 1126: [21.01.2024 09:48:56][debug][radarr] Tar out: tar: Removing leading `/' from member names; /*stdin*\ : Read error (39) : premature end; tar: Unexpected EOF in archive; tar: Child returned status 1; tar: Error is not recoverable: exiting now Line 1127: [21.01.2024 09:48:56][❌][radarr] tar verification failed! Tar said: tar: Removing leading `/' from member names; /*stdin*\ : Read error (39) : premature end; tar: Unexpected EOF in archive; tar: Child returned status 1; tar: Error is not recoverable: exiting now Line 1128: [21.01.2024 09:49:04][debug][radarr] lsof(/mnt/user/appdata/radarr) Line 1133: [21.01.2024 09:49:04][debug][radarr] AFTER verify: Array debug.7z
  3. As indicated, after reboot no pool devices recognized. Causes Docker fail, img/apps on a pool device. Reboot/ Poweroff tried. no luck. The devices show unmounted in unassigned devices, but fail if I try to mount through the gui: Jan 14 12:36:12 Tower unassigned.devices: Error: Device '/dev/sdw1' mount point 'hd36pool' - name is reserved, used in the array or a pool, or by an unassigned device. Jan 14 12:36:12 Tower unassigned.devices: Disk with serial 'ST18000NE000-3G6101_ZVT0JX4D', mountpoint 'hd36pool' cannot be mounted. Jan 14 12:36:18 Tower unassigned.devices: Error: Device '/dev/sdac1' mount point 'hd36pool' - name is reserved, used in the array or a pool, or by an unassigned device. Jan 14 12:36:18 Tower unassigned.devices: Disk with serial 'ST18000NM000J-2TV103_ZR59ZJJV', mountpoint 'hd36pool' cannot be mounted. Jan 14 12:36:21 Tower unassigned.devices: Error: Device '/dev/nvme1n1p1' mount point 'nvme' - name is reserved, used in the array or a pool, or by an unassigned device. Jan 14 12:36:21 Tower unassigned.devices: Disk with serial 'Sabrent_SB-RKT4P-2TB_48790469800627', mountpoint 'nvme' cannot be mounted. Jan 14 12:36:26 Tower unassigned.devices: Error: Device '/dev/nvme0n1p1' mount point 'nvme' - name is reserved, used in the array or a pool, or by an unassigned device. Jan 14 12:36:26 Tower unassigned.devices: Disk with serial 'Sabrent_SB-R Also, can only access GUI 'terminal, SSH fails 'target machine refuses.' any help appreciated / diagnostics of course attached 🙂 tower-diagnostics-20240114-1349.zip
  4. I'm sure this is user error, but I select from: Disk 1 select 14 TB folder to: Disk 1-4 all with 18 TB free When I run it 'scatters' all to the first disk selected, even though multiple disks selected. I'm sure this is user error.
  5. there is a (worse?) way to do this and have it in syslog: but I can only image this is an even more horrible solution as your unraid syslog now will have all file activity: logger -n ${SYSLOG_SERVER} -P 514 -t unraid-file-activity < "/var/log/file.activity.log"
  6. Agreed, Basically what I did is more of a kludge, and I'm sure this is a bad way to address this, but here's what I did: create shell to cp "/var/log/file.activity.log" to my log dest "/mnt/user/logs/fileactivity$(date +%Y%m%d%H%M%S).log" then rm "/var/log/file.activity.log" then - created job in user scripts and scheduled its not going into the syslog itself, just using the same location. [Yeah, its a mess, lots of issues with my approach, not recommending it, was a stop gap for me]
  7. would like to send file activity log to syslog server for diagnostics. where are they stored?
  8. Looks like that was it! Found it: Line 123075: Oct 17 08:37:03 Tower apcupsd[13439]: Power failure. Line 123076: Oct 17 08:37:03 Tower apcupsd[13439]: Battery power exhausted. Line 123093: Oct 17 08:37:13 Tower apcupsd[13439]: Power is back. UPS running on mains.
  9. Hello, I've had my Unraid server poweroff unexpectedly a couple of times. Can you point me to the correct logs that I can inspect to help identify the root cause? For example, is there a log I can check to see if the UPS sent a poweroff signal? Are there other logs I can look at to help resolve this issue? I've had to manually power on the server to resolve. tower-diagnostics-20231017-1406.zip
  10. @Swarles+1 this would be awesome. Would resolve several related 'asks' 🙂
  11. Any idea how to find files with filenames containing [DV] in the filename? I tried [DV] (returns all files ) and \[DV]. Example: "Filename (2018) {imdb-tt7044444} [WEBDL-2160p][DV][DTS-HD MA 5.1][x265]"
  12. is there a way to log wireguard stats, specifically peer info? this info on a given frequency: peer: xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx preshared key: (hidden) endpoint: 172.58.111.98:47136 allowed ips: 10.253.0.2/32 latest handshake: 1 minute, 45 seconds ago transfer: 851.32 KiB received, 8.84 MiB sent *and by chance is exposing the wg interface, like wg0 to the GUI dashboard network INTERFACE widget?* edit: Dhoope its already there ! scroll down ! VPN
  13. Maybe a bug? Maybe user stupidity? Backup was created with no issues then hanging and then failing in verification when backing up docker Plex-Media-Server: [26.04.2023 21:47:31][Plex-Media-Server][info] Backup created without issues [26.04.2023 21:47:31][Plex-Media-Server][info] Verifying backup... [26.04.2023 21:47:31][Plex-Media-Server][debug] Final verify command: --diff -f '/mnt/user/Backup/CommunityApp/ab_20230426_214539/Plex-Media-Server.tar.zst' '/mnt/user/appdata/Plex-Media-Server' '/mnt/user/appdata/Plex-Media-Server' [26.04.2023 21:50:57][Plex-Media-Server][debug] Tar out: tar: Removing leading `/' from member names; tar: Removing leading `/' from hard link targets; tar: /mnt/user/appdata/Plex-Media-Server: Not found in archive; tar: Exiting with failure status due to previous errors [26.04.2023 21:50:57][Plex-Media-Server][error] tar verification failed! More output available inside debuglog, maybe. [26.04.2023 21:52:34][Plex-Media-Server][debug] lsof(/mnt/user/appdata/Plex-Media-Server) I noticed this info message, like it was trying to back up the same path twice: [26.04.2023 21:45:53][Plex-Media-Server][info] Calculated volumes to back up: /mnt/user/appdata/Plex-Media-Server, /mnt/user/appdata/Plex-Media-Server And saw the Docker had set AppData Config Path: [2] = Host Path 2: [3] = "/mnt/user/appdata/Plex-Media-Server": [26.04.2023 21:45:53][Plex-Media-Server][debug] usorted volumes: Array ( [0] => /mnt/user/data [1] => /mnt/remotes/10.1.1.55_O [2] => /mnt/user/appdata/Plex-Media-Server [3] => /mnt/user/appdata/Plex-Media-Server ) So I changed the docer paths to be unique: [2] => /mnt/user/appdata/Plex-Media-Server/Config [3] => /mnt/user/appdata/Plex-Media-Server/Transcode And all was perfect! Again, maybe is user stupidity? but if a bug, attatched logs: ab.debug (1).log - Debug when not working ab.debug (2).log - Debug now when working ab.debug (1).logab.debug (2).log
  14. Would love to see similar color added to directories which exist more than once.
  15. Love this plugin and appreciate your work. Would love the ability to manually add a disk. As silly as this may seem, my use case is that Unraid contains 90% of my 'infrastructure' and it would be useful to be able to manually add disks, and then I would assign them to an 'offline' disk tray layout. Just a simple way of keeping 100% of my disk inventory visable. Also would use for any 'cold-swap' disks I may have.
  16. So - bumping my topic with a bit of new info. Basically, Pool 'hdd' is mounted as read only - I'm sure I filled it and caused btfrs to puke. Wanting to 'move' the data into the array, but obviously mover fails as pool is read-only: I'm thinking this should do the trick, using cp or cpg (think cpg won't show progress if I tee output), creating also an output file to check : screen cp -vR /mnt/hdd/data/* /mnt/user0/data/ | tee hddfilelist Is there a better method to move a large (60 TB) number of files off a failing read-only pool?
  17. all good ! thanks!!! fmi, is there a list of reserved words?
  18. Read through this thread, and wanted to see if there was a solution to: Enable plugged-in hot spare USB on-board the server. From a business continuity perspective, I'm responsible for minimizing my response time objectives[RTO], and defense in layers (hot/offsite/cloud) so my solution needs are: Layers: offsite and cloud [yes] hot [no] For "Hot": "Hot" Thumb-drive. Selectable in EFI like 'USB2" whatever Selectable from gui reboot screen Restarts / and can* start array as from a "clean shutdown" (*if configured): I agree, if "opps I dwerped off my USB" - about 30 minutes recovery time. My concern is my drive fails when I am not at the server. I need to facter in the average time it will take for me to get there, or to talk someont through it. Again, all cool, just factoring in the time. Hot-Spare to keep recovery time down as close to time to do regular restart. Agreed, as part of the depth of the solution. I'd state it as "2nd [3rd,..] drive aside in a safe location tl;dr, looking for HOT PLUGGED IN USB, aka: a 2 keys 1 cron
  19. I agree on both points, @pwm's explaination about Unraid and how it works was very nice. However, my tl;dr on this is: If I'm accountable to manage [1+ TB of data] OR [0+ b of critical data] I would deploy a hot spare as routine part of my business continuity plan. Do any docker/apps/user scripts provide Hot Spare functionality at array or cache pool ?
  20. Fix Common Problems Identifies this issue: Share log references non existent pool cache. If you have renamed a pool this will have to be adjusted in the share's settings Clicking Share Settings results in: ! Share 'log' has been deleted. I can see directories : /media/user[0]/log both exist. SyslogSettings set to share named "logs", working as expected. Attempts to resolve: SyslogSettings set to share "log", does not appear as an option Add share named "log" results (as expected): Invalid share name. Do not use reserved names Idea I thought to try, seems like an awful idea: rm -f /media/user?/log Thinking "maybe" caused somewhere around here: SyslogSettings reference "log" "log" pointed to pool cache named " " deleted pool "cache" created new pool "nvme" created new share "logs" repointed SyslogSettings to share:"logs" pool:"nvme" And if I had to guess it was in step #2. I recall a page listing some reserved words, like UNRAID, but can't locate the url. Resolution Currently, I ignore the finding. Any other way to resolve?
  21. Good call out. Assume that: Is this correct? : [Share Minimum Free Space setting] = [Share Minimum Free Space setting] = [2 * Largest File Size]
  22. Hi, I have a btfrs pool device (hdd) that is failing, showing as rw when array starts then as completely full and read-only within a couple minutes. To manually move the files to the array, would this be the correct process/or would there be a better method? It has about 60 TB of data needing to be copied if that makes a difference. Understanding user0 is without items on the cache drive pool drive = 'hdd' folder to copy = 'data' Think opening terminal and: cp -R /media/hdd/data/.* /media/user0/data/ Thanks in advance tower-diagnostics-20230331-2253.zip
  23. THANKS! (... aaand it was right there on the screeen as well " Choose a value which is equal or greater than the biggest single file size you intend to copy to the share. Include units KB, MB, GB and TB as appropriate, e.g. 10MB. "