geekypenguin

Members
  • Posts

    49
  • Joined

  • Last visited

Recent Profile Visitors

The recent visitors block is disabled and is not being shown to other users.

geekypenguin's Achievements

Rookie

Rookie (2/14)

5

Reputation

  1. Would trying a bios update be worthwhile before I go spending money?
  2. Sorry it's taken a few days to respond, there's been a lot to work through and while I still don't know the cause, I reached a point where I had to stop and revert to known good. Firstly I had the macvlan kernel issue that's known in 6.12 which frustrated things. With both sticks of ram on their own, I was getting data corruption errors, always in the same disk. I was also getting corruption of my docker.img which was causing docker's to crash without the cache going read only. As the nvme drives were new, I got a warranty replacement on the nvme drive with all the errors and attempted to rebuild the cache pool onto the second drive, but was flooded with "nvme frozen state error detected, reset controller" etc messages for the replacement drive. I read in a few bug reports to add ```nvme_core.default_ps_max_latency_us=0 pcie_aspm=off``` to the boot config but this didn't help either. This is unfortunately where I had to stop. I've removed the second cache drive and reverted to single drive mode for my cache which has been working fine for a few days now with all the ram re-installed. Not sure where else to go from this to be honest. I can stay like this with no redundancy on my cache I suppose but would like to get to the bottom of it.
  3. Thanks for you help. Removed one ram stick and reset the stats and configured the userscript as suggested. I'll let you know how it gets on
  4. Ok First two passes of memtest have returned zero errors. I'll keep it running a bit longer to be sure none materialise
  5. Thanks, I'll run a memtest now. I saw those messages about nvme1 which is what made me suspect a bad ssd? I also see messages about multiple uncorrected fatal error received, frozen state error detected, and device recovery successful.
  6. A second diagnostics download immediately after rebooting if that's of any use: lisa-diagnostics-20231124-1015.zip
  7. I replaced my cache drive about a week ago and also added a second drive as a raid1 pool. Since then, approximately every 2 days the dockers and VMs lock up and trying to write to the cache drive returns a message that the file system is read only. I've tried running a balance and a scrub and the scrub returns no errors, yet the problem keeps recurring. The only way to bring it back to life is to reboot, but it soon happens again. What have I missed? Or could the new SSDs just be faulty? Diagnostics attached. lisa-diagnostics-20231124-0959.zip
  8. Good point, I completely missed that people were using the LTS tag and not the latest tag
  9. There's your problem When I said export and import I didn't mean backing up your app data folder. If you restore the old appears it will push you back to the old version. From unifi itself there is an export config option. Use this. Still take a backup of your appdata incase everything brakes as a fallback, but that's not the process for upgrading
  10. Have a look on unifi's website for the update path from the version you're at to v5.13. you can't jump from 5.6 to 5.13 in one go. If memory serves there is at least one to go to before changing to the latest .13 tag. Alternatively, backup your config, bin the docker and fire up a new one with latest, before importing your old config
  11. Ok so another update, now been running for over 3 days with the most current docker container, no ram cap and DPI turned off, and my ram is sitting just below 800MB used.
  12. I haven't set a limit yet on mine but turning off DPI has worked well for me. There was still a memory leak somewhere but it's not as bad, I got to 950MB after 5 days. However, I noticed another update yesterday and it's at 705MB after 12 hours running (it started at 709mb) so curious to know what's changed as unifinis still reporting the same controller version (5.13.29) Edit: another hour later and the ram has fallen further to 682MB!?!
  13. I haven't limited mine to 4GB yet, I'll set a cap and turn on DPI later today
  14. Currently up to 765MB so it has grown but would usually be over 1.5GB by now with 31 hours up-time. I'm keeping the DPI off on mine as it's clearly helping but I think it's too soon to call the issue fixed for version 5.13.29 And I'm just using a user script: docker restart unifi