dalben

Members
  • Posts

    1453
  • Joined

  • Last visited

About dalben

  • Birthday 09/30/1966

Retained

  • Member Title
    sleepy

Converted

  • Gender
    Male
  • Location
    Singapore
  • Personal Text
    Too old to know everything

Recent Profile Visitors

3651 profile views

dalben's Achievements

Community Regular

Community Regular (8/14)

45

Reputation

  1. From what I can work out, with Docker setup to Host Access to custom networks, you get a second network appearing on the same Server IP address. I have 2 NICS, one for the docker vlan one for the server management, both show two networks/MAC addresses at the same IP. Hasn't degregaded performance. What would be nice, and I'm not sure if possible, is to fix/lock a MAC address in for br0 so that my network controller can label it properly. It seems to change at every reboot. It's not a performance thing or show stopper, just helps my OCD cope with my network controller better.
  2. A quick that must probably has been answered before but my search fu isn't working today. Due to a catastrophic mobo error, I have to quickly replace my board and get my server up again quickly. The board that's arriving is an ASUS TUF B365M-Plus Gaming. After recently upgrading to 18Tb disks I can use the 6 onboard SATA + 2xM.2 with ports to spare. I also have an LSI 9211-i8 controller that's in the current dead server and a recently purchased 2nd hand Dell HBA, which is an LSI 9207-i8. What's going to give me the best performance, all on the onboard SATA, or on an LSI running on the x4 PCIe slot? Would using the x16 slot make a difference? If we're talking miniscule differences I might go the LSA simply because the cabling will be neater. Any advice would be most welcome.
  3. @JorgeB All reformatted and finished playing shuffle files. No more errors. Thanks
  4. Thanks. No errors prior to NUT losing the USB connection.
  5. Still getting this error with the latest update on my APC MODEL Back-UPS BX750MI. Once it loses connection I need to pull the USB cable and plug it in again for it to work. Tried switching to the native Unraid UPS monitor while in this states and it complains of connection error. After resetting the USB cable the Unraid UPS monitor works fine without any issues.
  6. Thanks @JorgeB. Appreciate the troubleshooting and testing. Yeah, but 18Gb drives for me. What fun
  7. I've started seeing parity tuning errors in syslog. Mar 21 06:40:20 tdm root: RTNETLINK answers: File exists Mar 21 06:40:20 tdm Parity Check Tuning: PHP error_reporting() set to E_ERROR|E_WARNING|E_PARSE|E_CORE_ERROR|E_CORE_WARNING|E_COMPILE_ERROR|E_COMPILE_WARNING|E_USER_ERROR|E_USER_WARNING|E_USER_NOTICE|E_STRICT|E_RECOVERABLE_ERROR|E_USER_DEPRECATED Mar 21 06:40:20 tdm emhttpd: nothing to sync Mar 21 06:40:20 tdm sudo: root : PWD=/ ; USER=root ; COMMAND=/bin/bash -c /usr/local/emhttp/plugins/controlr/controlr -port 2378 -certdir '/boot/config/ssl/certs' -showups Mar 21 06:40:20 tdm sudo: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=0) Mar 21 06:40:20 tdm Parity Check Tuning: PHP error_reporting() set to E_ERROR|E_WARNING|E_PARSE|E_CORE_ERROR|E_CORE_WARNING|E_COMPILE_ERROR|E_COMPILE_WARNING|E_USER_ERROR|E_USER_WARNING|E_USER_NOTICE|E_STRICT|E_RECOVERABLE_ERROR|E_USER_DEPRECATED Mar 21 06:40:21 tdm kernel: eth0: renamed from veth505b9c4 They are fairly regular throughout syslog. Noticed them today while running RC3, still there after RC4
  8. Not sure if it was addressed in RC4 and if so there was any expectation of the error going away, but it's still there Mar 21 06:40:06 tdm kernel: XFS (md1): Mounting V5 Filesystem Mar 21 06:40:06 tdm kernel: XFS (md1): Ending clean mount Mar 21 06:40:06 tdm emhttpd: shcmd (34): xfs_growfs /mnt/disk1 Mar 21 06:40:06 tdm kernel: xfs filesystem being mounted at /mnt/disk1 supports timestamps until 2038 (0x7fffffff) Mar 21 06:40:06 tdm root: xfs_growfs: XFS_IOC_FSGROWFSDATA xfsctl failed: No space left on device Mar 21 06:40:06 tdm root: meta-data=/dev/md1 isize=512 agcount=32, agsize=137330687 blks Mar 21 06:40:06 tdm root: = sectsz=512 attr=2, projid32bit=1 Mar 21 06:40:06 tdm root: = crc=1 finobt=1, sparse=1, rmapbt=0 Mar 21 06:40:06 tdm root: = reflink=1 bigtime=0 inobtcount=0 Mar 21 06:40:06 tdm root: data = bsize=4096 blocks=4394581984, imaxpct=5 Mar 21 06:40:06 tdm root: = sunit=1 swidth=32 blks Mar 21 06:40:06 tdm root: naming =version 2 bsize=4096 ascii-ci=0, ftype=1 Mar 21 06:40:06 tdm root: log =internal log bsize=4096 blocks=521728, version=2 Mar 21 06:40:06 tdm root: = sectsz=512 sunit=1 blks, lazy-count=1 Mar 21 06:40:06 tdm root: realtime =none extsz=4096 blocks=0, rtextents=0 Mar 21 06:40:06 tdm kernel: XFS (md1): EXPERIMENTAL online shrink feature in use. Use at your own risk! Mar 21 06:40:06 tdm emhttpd: shcmd (34): exit status: 1 Mar 21 06:40:06 tdm emhttpd: shcmd (35): mkdir -p /mnt/disk2 Mar 21 06:40:06 tdm emhttpd: shcmd (36): mount -t xfs -o noatime,nouuid /dev/md2 /mnt/disk2 Mar 21 06:40:06 tdm kernel: XFS (md2): Mounting V5 Filesystem Mar 21 06:40:07 tdm kernel: XFS (md2): Ending clean mount Mar 21 06:40:07 tdm kernel: xfs filesystem being mounted at /mnt/disk2 supports timestamps until 2038 (0x7fffffff) Mar 21 06:40:07 tdm emhttpd: shcmd (37): xfs_growfs /mnt/disk2 Mar 21 06:40:07 tdm root: xfs_growfs: XFS_IOC_FSGROWFSDATA xfsctl failed: No space left on device Mar 21 06:40:07 tdm root: meta-data=/dev/md2 isize=512 agcount=32, agsize=137330687 blks Mar 21 06:40:07 tdm root: = sectsz=512 attr=2, projid32bit=1 Mar 21 06:40:07 tdm root: = crc=1 finobt=1, sparse=1, rmapbt=0 Mar 21 06:40:07 tdm root: = reflink=1 bigtime=0 inobtcount=0 Mar 21 06:40:07 tdm root: data = bsize=4096 blocks=4394581984, imaxpct=5 Mar 21 06:40:07 tdm root: = sunit=1 swidth=32 blks Mar 21 06:40:07 tdm root: naming =version 2 bsize=4096 ascii-ci=0, ftype=1 Mar 21 06:40:07 tdm root: log =internal log bsize=4096 blocks=521728, version=2 Mar 21 06:40:07 tdm root: = sectsz=512 sunit=1 blks, lazy-count=1 Mar 21 06:40:07 tdm root: realtime =none extsz=4096 blocks=0, rtextents=0 Mar 21 06:40:07 tdm emhttpd: shcmd (37): exit status: 1 Mar 21 06:40:07 tdm emhttpd: shcmd (38): mkdir -p /mnt/disk3 Mar 21 06:40:07 tdm emhttpd: shcmd (39): mount -t xfs -o noatime,nouuid /dev/md3 /mnt/disk3 Mar 21 06:40:07 tdm kernel: XFS (md3): Mounting V5 Filesystem Mar 21 06:40:07 tdm kernel: XFS (md3): Ending clean mount Mar 21 06:40:07 tdm kernel: xfs filesystem being mounted at /mnt/disk3 supports timestamps until 2038 (0x7fffffff) Mar 21 06:40:07 tdm emhttpd: shcmd (40): xfs_growfs /mnt/disk3 Mar 21 06:40:07 tdm root: meta-data=/dev/md3 isize=512 agcount=32, agsize=30523583 blks Mar 21 06:40:07 tdm root: = sectsz=512 attr=2, projid32bit=1 Mar 21 06:40:07 tdm root: = crc=1 finobt=1, sparse=1, rmapbt=0 Mar 21 06:40:07 tdm root: = reflink=1 bigtime=0 inobtcount=0 Mar 21 06:40:07 tdm root: data = bsize=4096 blocks=976754633, imaxpct=5 Mar 21 06:40:07 tdm root: = sunit=1 swidth=32 blks Mar 21 06:40:07 tdm root: naming =version 2 bsize=4096 ascii-ci=0, ftype=1 Mar 21 06:40:07 tdm root: log =internal log bsize=4096 blocks=476930, version=2 Mar 21 06:40:07 tdm root: = sectsz=512 sunit=1 blks, lazy-count=1 Mar 21 06:40:07 tdm root: realtime =none extsz=4096 blocks=0, rtextents=0 Mar 21 06:40:07 tdm emhttpd: shcmd (41): mkdir -p /mnt/cache
  9. On the pool disks tab. Second pool shows this error: Balance Status btrfs filesystem df: Data, single: total=30.00GiB, used=21.21GiB System, RAID1: total=32.00MiB, used=16.00KiB Metadata, RAID1: total=2.00GiB, used=29.98MiB GlobalReserve, single: total=261.73MiB, used=0.00B btrfs balance status: No balance found on '/mnt/rad' Current usage ratio: Warning: A non-numeric value encountered in /usr/local/emhttp/plugins/dynamix/include/DefaultPageLayout.php(515) : eval()'d code on line 542 0 % --- Warning: A non-numeric value encountered in /usr/local/emhttp/plugins/dynamix/include/DefaultPageLayout.php(515) : eval()'d code on line 542 Full Balance recommended
  10. Who would know then? And would that person know if all large drives formatted with 6.10 need to be reformatted or just the one throwing the error?
  11. Installed 6.10.0-RC3 and I still get the same error/warning Mar 12 09:50:42 tdm kernel: XFS (md1): Mounting V5 Filesystem Mar 12 09:50:42 tdm kernel: XFS (md1): Ending clean mount Mar 12 09:50:42 tdm kernel: xfs filesystem being mounted at /mnt/disk1 supports timestamps until 2038 (0x7fffffff) Mar 12 09:50:42 tdm emhttpd: shcmd (34): xfs_growfs /mnt/disk1 Mar 12 09:50:42 tdm kernel: XFS (md1): EXPERIMENTAL online shrink feature in use. Use at your own risk! Mar 12 09:50:42 tdm root: xfs_growfs: XFS_IOC_FSGROWFSDATA xfsctl failed: No space left on device Mar 12 09:50:42 tdm root: meta-data=/dev/md1 isize=512 agcount=32, agsize=137330687 blks Mar 12 09:50:42 tdm root: = sectsz=512 attr=2, projid32bit=1 Mar 12 09:50:42 tdm root: = crc=1 finobt=1, sparse=1, rmapbt=0 Mar 12 09:50:42 tdm root: = reflink=1 bigtime=0 inobtcount=0 Mar 12 09:50:42 tdm root: data = bsize=4096 blocks=4394581984, imaxpct=5 Mar 12 09:50:42 tdm root: = sunit=1 swidth=32 blks Mar 12 09:50:42 tdm root: naming =version 2 bsize=4096 ascii-ci=0, ftype=1 Mar 12 09:50:42 tdm root: log =internal log bsize=4096 blocks=521728, version=2 Mar 12 09:50:42 tdm root: = sectsz=512 sunit=1 blks, lazy-count=1 Mar 12 09:50:42 tdm root: realtime =none extsz=4096 blocks=0, rtextents=0 Mar 12 09:50:42 tdm emhttpd: shcmd (34): exit status: 1 Mar 12 09:50:42 tdm emhttpd: shcmd (35): mkdir -p /mnt/disk2 Mar 12 09:50:42 tdm emhttpd: shcmd (36): mount -t xfs -o noatime,nouuid /dev/md2 /mnt/disk2 Mar 12 09:50:42 tdm kernel: XFS (md2): Mounting V5 Filesystem Mar 12 09:50:42 tdm kernel: XFS (md2): Ending clean mount Mar 12 09:50:42 tdm emhttpd: shcmd (37): xfs_growfs /mnt/disk2 Mar 12 09:50:42 tdm kernel: xfs filesystem being mounted at /mnt/disk2 supports timestamps until 2038 (0x7fffffff) Mar 12 09:50:43 tdm root: xfs_growfs: XFS_IOC_FSGROWFSDATA xfsctl failed: No space left on device Mar 12 09:50:43 tdm root: meta-data=/dev/md2 isize=512 agcount=32, agsize=137330687 blks Mar 12 09:50:43 tdm root: = sectsz=512 attr=2, projid32bit=1 Mar 12 09:50:43 tdm root: = crc=1 finobt=1, sparse=1, rmapbt=0 Mar 12 09:50:43 tdm root: = reflink=1 bigtime=0 inobtcount=0 Mar 12 09:50:43 tdm root: data = bsize=4096 blocks=4394581984, imaxpct=5 Mar 12 09:50:43 tdm root: = sunit=1 swidth=32 blks Mar 12 09:50:43 tdm root: naming =version 2 bsize=4096 ascii-ci=0, ftype=1 Mar 12 09:50:43 tdm root: log =internal log bsize=4096 blocks=521728, version=2 Mar 12 09:50:43 tdm root: = sectsz=512 sunit=1 blks, lazy-count=1 Mar 12 09:50:43 tdm root: realtime =none extsz=4096 blocks=0, rtextents=0 Mar 12 09:50:43 tdm emhttpd: shcmd (37): exit status: 1 Mar 12 09:50:43 tdm emhttpd: shcmd (38): mkdir -p /mnt/disk3 Mar 12 09:50:43 tdm emhttpd: shcmd (39): mount -t xfs -o noatime,nouuid /dev/md3 /mnt/disk3 Mar 12 09:50:43 tdm kernel: XFS (md3): Mounting V5 Filesystem Mar 12 09:50:43 tdm kernel: XFS (md3): Ending clean mount Mar 12 09:50:43 tdm emhttpd: shcmd (40): xfs_growfs /mnt/disk3 Mar 12 09:50:43 tdm kernel: xfs filesystem being mounted at /mnt/disk3 supports timestamps until 2038 (0x7fffffff) Mar 12 09:50:43 tdm root: meta-data=/dev/md3 isize=512 agcount=32, agsize=30523583 blks Mar 12 09:50:43 tdm root: = sectsz=512 attr=2, projid32bit=1 Mar 12 09:50:43 tdm root: = crc=1 finobt=1, sparse=1, rmapbt=0 Mar 12 09:50:43 tdm root: = reflink=1 bigtime=0 inobtcount=0 Mar 12 09:50:43 tdm root: data = bsize=4096 blocks=976754633, imaxpct=5 Mar 12 09:50:43 tdm root: = sunit=1 swidth=32 blks Mar 12 09:50:43 tdm root: naming =version 2 bsize=4096 ascii-ci=0, ftype=1 Mar 12 09:50:43 tdm root: log =internal log bsize=4096 blocks=476930, version=2 Mar 12 09:50:43 tdm root: = sectsz=512 sunit=1 blks, lazy-count=1 Mar 12 09:50:43 tdm root: realtime =none extsz=4096 blocks=0, rtextents=0 Mar 12 09:50:43 tdm emhttpd: shcmd (41): mkdir -p /mnt/cache
  12. Thanks, that makes sense. Didn't think of that, but yes, I'll try the Unraid way first. Makes life a lot easier
  13. Only just started to really use this plugin and dropped adding scripts to crontab. Kicking myself for not migrating to this earlier. Quick question. I have a need to add an entry to the hosts file of about 3 dockers after they start up (a server i am trying to reach has had it's domain name snatched so a hosts file entry is the only way to get to it). Is there a way I can trigger CAUS to to run a script upon docker start? Or maybe 5/10 mins after array start? While I am at it, any tips on writing a script that can be run from the console to had hosts entries in docker containers? Or is there a simpler way to do it?
  14. Did this last weekend. Being slighty OCD about my server, there's no way I could leave it with gaps in the disk numbers. 20odd hours for the parity rebuild but server is up and running while it was doing it so there was no impact really.