sota

Members
  • Posts

    691
  • Joined

Recent Profile Visitors

The recent visitors block is disabled and is not being shown to other users.

sota's Achievements

Enthusiast

Enthusiast (6/14)

32

Reputation

  1. starts at 218, drops to 103 at the end. my last parity check of 5+1 drives... Duration: 18 hours, 9 minutes, 19 seconds. Average speed: 153.0 MB/sec
  2. I have no idea, as the only machine I run unRAID on has ECC
  3. we talking commodity memory, or ECC? if the latter, run the following... grep "[0-9]" /sys/devices/system/edac/mc/mc*/csrow*/ch*_ce_count it'll show you if any chips have been having correctable faults.
  4. (was going to post this as a question, but since it's now resolved, i'm posting as a documentation of my journey.) This has been an on going problem for a long time for me, but i'm just now able to deal with it. /var/log/syslog no longer populates. as such http://server/logging.htm never populates. if I delete the /var/log/syslog file, I get... /usr/bin/tail: cannot open '/var/log/syslog' for reading: No such file or directory /usr/bin/tail: no files remaining and it never recreates the files. tmpfs 128M 18M 111M 14% /var/log so I have plenty of space. /etc/rc.d/rc.rsyslogd restart does no good. restarting the machine does no good. finally tried this... deleting the file AND rebooting appears to have solved it for me. deleting the file AND ONLY restarting the service does not work. not sure what background process(es) are also affected, but I figured I'd post this as how it worked for me.
  5. Since the still working data disks were mounted on another UNRAID machine under UD and as Ready Only, I'm hoping things are pretty decent, but I'm willing to accept some potential data loss. I'll give that a shot in the next couple days, so thank for that @JorgeB
  6. UNRAID 6.9.2 long story short, had 2 drives in a 6+1 array die (one of which was parity). Since I needed to get the box working RIGHT NOW, I yanked all 7 disks, build a new system, copied everything I could from the working drives, and set these disks to the side. Now, I'm trying to see if I can't recover something extra. Parity "failed", but with some effort on my part (HDD REGEN tool) I got it back and working, for now. that leaves 6 data disks. 5 of them work. I want to try and recreate the array, and somehow get it to rebuild the data from the 6th drive. What steps do i need to take to convince a trial version of UNRAID that this array is a valid, but 1 disk is listed as "missing", so I can force the rebuild to a replacement when I add it in? I know which drives are which (thank god for good note taking!) I stupidly did NOT save a copy of the CONFIG before blowing the original box up, so that's on me. I'm sure there's details i'm missing that needed providing, so just let me know and I'll do my best to gather that info. thanks!
  7. ok I don't know what the he double hockey sticks was going on there, as i've been trying to fix notifications on this server for 3 days now, and been getting nothing but 534 and 535 errors, no matter what I tried. NOW, it works!
  8. (v6.9.2) So, Gmail seems to no longer work for sending notifications, due to security settings on their end. What alternate and free email service providers are people using I can sign up for to send out system notifications? Thanks!
  9. that's exactly how I'd do it if I needed to upgrade to several bigger disks all at once.
  10. Honestly... leave the array intact, copy the data from disk to disk as files, do that 7 times, swap 'em all out, and rebuild the parity drives. that'll put the least amount of stress on the rest of the disks.
  11. Given my backup methodology relies on a Windows VM to execute the backup, I went with NTFS. See sig for link to what i do.
  12. yea. that was the free space. the problem is the disk only was using 77GB until recently, when I moved the VMs off of it. it clearly hasn't been TRIMming on a regular basis.
  13. fstrim -v /mnt/cache /mnt/cache: 444 GiB (476746616832 bytes) trimmed that's on a 500GB SSD with 6.9.2 running. we've had performance issues on the system recently, and when I copied the VMs off the drive, it was barely moving data. then I ran the command above. something tells me it's not TRIMming the drive like its supposed to.
  14. Apr 21 23:14:44 Tigger kernel: e1000e 0000:00:19.0 eth0: Detected Hardware Unit Hang: Apr 21 23:14:44 Tigger kernel: TDH <2d> Apr 21 23:14:44 Tigger kernel: TDT <44> Apr 21 23:14:44 Tigger kernel: next_to_use <44> Apr 21 23:14:44 Tigger kernel: next_to_clean <2c> Apr 21 23:14:44 Tigger kernel: buffer_info[next_to_clean]: Apr 21 23:14:44 Tigger kernel: time_stamp <13a8c3651> Apr 21 23:14:44 Tigger kernel: next_to_watch <2d> Apr 21 23:14:44 Tigger kernel: jiffies <13a8c3f00> Apr 21 23:14:44 Tigger kernel: next_to_watch.status <0> Apr 21 23:14:44 Tigger kernel: MAC Status <80083> Apr 21 23:14:44 Tigger kernel: PHY Status <796d> Apr 21 23:14:44 Tigger kernel: PHY 1000BASE-T Status <3800> Apr 21 23:14:44 Tigger kernel: PHY Extended Status <3000> Apr 21 23:14:44 Tigger kernel: PCI Status <10> Apr 21 23:14:46 Tigger kernel: e1000e 0000:00:19.0 eth0: Detected Hardware Unit Hang: Apr 21 23:14:46 Tigger kernel: TDH <2d> Apr 21 23:14:46 Tigger kernel: TDT <44> Apr 21 23:14:46 Tigger kernel: next_to_use <44> Apr 21 23:14:46 Tigger kernel: next_to_clean <2c> Apr 21 23:14:46 Tigger kernel: buffer_info[next_to_clean]: Apr 21 23:14:46 Tigger kernel: time_stamp <13a8c3651> Apr 21 23:14:46 Tigger kernel: next_to_watch <2d> Apr 21 23:14:46 Tigger kernel: jiffies <13a8c46c0> Apr 21 23:14:46 Tigger kernel: next_to_watch.status <0> Apr 21 23:14:46 Tigger kernel: MAC Status <80083> Apr 21 23:14:46 Tigger kernel: PHY Status <796d> Apr 21 23:14:46 Tigger kernel: PHY 1000BASE-T Status <3800> Apr 21 23:14:46 Tigger kernel: PHY Extended Status <3000> Apr 21 23:14:46 Tigger kernel: PCI Status <10> Apr 21 23:14:48 Tigger kernel: e1000e 0000:00:19.0 eth0: Detected Hardware Unit Hang: Apr 21 23:14:48 Tigger kernel: TDH <2d> Apr 21 23:14:48 Tigger kernel: TDT <44> Apr 21 23:14:48 Tigger kernel: next_to_use <44> Apr 21 23:14:48 Tigger kernel: next_to_clean <2c> Apr 21 23:14:48 Tigger kernel: buffer_info[next_to_clean]: Apr 21 23:14:48 Tigger kernel: time_stamp <13a8c3651> Apr 21 23:14:48 Tigger kernel: next_to_watch <2d> Apr 21 23:14:48 Tigger kernel: jiffies <13a8c4ec0> Apr 21 23:14:48 Tigger kernel: next_to_watch.status <0> Apr 21 23:14:48 Tigger kernel: MAC Status <80083> Apr 21 23:14:48 Tigger kernel: PHY Status <796d> Apr 21 23:14:48 Tigger kernel: PHY 1000BASE-T Status <3800> Apr 21 23:14:48 Tigger kernel: PHY Extended Status <3000> Apr 21 23:14:48 Tigger kernel: PCI Status <10> Apr 21 23:14:49 Tigger kernel: e1000e 0000:00:19.0 eth0: Reset adapter unexpectedly Apr 21 23:14:50 Tigger kernel: bond0: (slave eth0): link status definitely down, disabling slave Apr 21 23:14:50 Tigger kernel: device eth0 left promiscuous mode Apr 21 23:14:50 Tigger kernel: bond0: now running without any active interface! Apr 21 23:14:50 Tigger kernel: br0: port 1(bond0) entered disabled state Apr 21 23:14:53 Tigger kernel: e1000e 0000:00:19.0 eth0: NIC Link is Up 1000 Mbps Full Duplex, Flow Control: None Apr 21 23:14:53 Tigger kernel: bond0: (slave eth0): link status definitely up, 1000 Mbps full duplex Apr 21 23:14:53 Tigger kernel: bond0: (slave eth0): making interface the new active one Apr 21 23:14:53 Tigger kernel: device eth0 entered promiscuous mode Apr 21 23:14:53 Tigger kernel: bond0: active interface up! Apr 21 23:14:53 Tigger kernel: br0: port 1(bond0) entered blocking state Apr 21 23:14:53 Tigger kernel: br0: port 1(bond0) entered forwarding state Started having this problem recently. Not physically at the machine right now, but does this look like a faulty card, bad port on the switch, or a bad cable? Or, something else software related. Machine is basically a glorified file cabinet, running a single Windows 7 x64 VM for SageTV (haven't gotten the docker to work to my liking, and i'm under a time crunch with a failing physical SageTV server and a 4/25 "hard" cut over date for Cablevision switching to encrypted.) Everything seemed to be working fine until a couple days ago, when while I was remoted into the VM (anydesk) I kept and keep getting disconnected. Finally tried to watch the log, only to discover /var/log was full. I caught the above after the most recent disconnect. diagnostics and syslog.1 are attached. syslog.zip tigger-diagnostics-20220421-2322.zip