Jump to content

Domotics

Members
  • Posts

    21
  • Joined

Posts posted by Domotics

  1. Hello friends,

    Am running the latest stable version on UnRaid 6.10.3.


    I have been observing an issue with my Unraid server lately. At times the WebUI goes offline and I cannot login anymore. after several tries, it comes back but very very slow to load the UI. then, I tried to stop a few containers (namely PiHole, UnRaid-API, Grafana-Unraid-Stack etc) and then the server slowly comes back. Last time it happened, I permanently stopped `Pi-Hole`, thinking that was the culprit. But it happened again this morning, and I stopped `UnRaid-API` and the server started responding again, it took a while to stabilize, though. I also checked the “htop” during this time, and found some processes under “/usr/local/emhttp/WebGui/nchan/*” running very high on CPU.

     

    Anybody faced this issue before ? 

     

    I took a diagnostics (after the fact), not sure if it helps..

    fusion-diagnostics-20220809-0928.zip

  2. On 7/6/2022 at 3:06 PM, ich777 said:

    This should be fine, since the SHUTDOWN temperature is set to ~105 °C I think...

     

    The shutdown is only triggered to my knowledge when the shutdown temperature is reached and it is recommended to reboot after it was triggered, not to reset the device.

     

    Anyways, keep me updated...

     

    BTW You can of course set the Shutdown temperature to something insanely high but keep in mind that these Coral devices can be a fire hazard (even the official documentation mentions this) when used wrong. Keep in mind I won't recommend to do this because if it really runs sometimes too hot it can simply catch on fire.

    I don't know if it really helped, but since i moved the frigate.db from share to cache , it is running fine so far. haven't crashed since last few days. of course i don't consider it as stable yet. i will monitor for another week or so... 

  3. 2 minutes ago, ich777 said:

    Have you yet tried the User Script that I've linked multiple times in the GitHub Issue that you've linked?

    You mean the temperature monitoring ? yes i have that installed. for now it is running fine and the temperature is shown below 50 most times.

     

    As part of the troubleshooting, I have moved the frigate.db to nvme (it was on the share before and the share was setup with cache enabled). Now the DB should always stay on the nvme, hopefully it will make some differance. 

  4. On 6/26/2022 at 8:42 PM, ich777 said:

    I only can imagine that it got completely overloaded and actually run into the temperature limit, but I can't tell for sure if this was the case here.

    This is pretty much the case. I opend this issue a while ago and closed it thinking my issue was fixed.

    https://github.com/blakeblackshear/frigate/issues/3283

     

    Appears like it is not. now it is pretty much clear the the sudden hike in the coral module temperature is causing shutdown of TPU and this is leading frigate to crash. i am almost sure it is not caused by an unassigned device because i don't have one and it still fails. am scratching my head to figure out what triggers the sudden increase of TPU temperature.. In my case, am using a mini PCIE module attched on a miniPCIE to PCI express card (specifically this one). This is a cheap device so it could be the reason (possibily).. waiting for some more people facing this issue to come up with their theory..

    My system is well ventilated and have 5 FANS running to keep things under low temp.. my CPU/MoBo, Disks etc, run below the threshoulds.. it is just this TPU going crazy at times. 

  5. 10 minutes ago, JonathanM said:

    Perhaps you should add the rest of the instructions to this post.

    That is a fair point. posted couple of screenshots above. more more clarity i will mentioned the details again below.

     

     

    1. setup a new share for storing log files (preferably set to cache only so it will not harm the spinning disks) - example screenshot below

     

    ext_logs_share.thumb.PNG.5f86baeb6a09d3f52610d868b6bd48bd.PNG

    2. Edit nextcloud container settings and add an additional path and point it to the share created in the previous step.

     

    nextcloud_ext_logs.PNG.8a5745e5a610abbc07b3e94835d928a5.PNG

     

    3. edit the config.php (normally located on /mnt/user/appdata/nextcloud/www/nextcloud/config/config.php) and add the below lines somewhere in between the other configs there.

    'logfile' => '/ext_logs/nextcloud_logs/nextcloud.log',
    'logfile_audit' => '/ext_logs/nextcloud_logs/audit.log',

     

    4. restart the nextcloud container.

     

    5. verify the disk with nextcloud data is spinning down after the timeout you set.


    Hope this helps.

  6. 6 minutes ago, Domotics said:

    ext_logs is share with cache (only) and this share is mounted as an external path on the nextcloud container. so this is not on the docker itself. 

     

    this is how is it setup on nextcloud

     

    nextcloud_ext_logs.PNG.07e00994627aba136e65e5d75eccc9e8.PNG

     

    and this is the share settings.

    ext_logs_share.thumb.PNG.13ce0dd952988ad2133a2187a5e9f583.PNG

     

  7. 11 hours ago, JonathanM said:

    Unless you also implement something to clean them out this will likely balloon your docker image file.

    ext_logs is share with cache (only) and this share is mounted as an external path on the nextcloud container. so this is not on the docker itself. 

  8. Hi There,

     

    Hope everyone is doing well.

     

    This is not really an issue but i am trying to do some optimisation to my unraid setup by spinning down the disks when there is no activity. i have found couple of applications prevent this happening. I have tried with `File activity` & `open files ` plugins but i couldn't find anything that is directly accessing disk3 (in the case of cctv).

     

    1. Nextcloud

     

    For nextcloud, i have share setup with no cache enabled. it seems there is a log file `nextcloud.log` having some activities very frequently and this is preventing the disk spin down. if i keep the nextcloud container stopped for a while this disk spins down after the time out i set. I have seen another post with this discussion but unable to fix this with the methods mentioned on that post.

    2. Windows VM with Blue Iris

    I have a Windows 10 VM running with Blue Iris and it has the recordings saved to a share named `cctv`. This share has the cache enabled and this share has only one disk included. I checked the files on the disk and i can't find anything that was modified after the last mover run. if i keep this VM shut for a while, the disks would spin down after the time out. I can see the write and read counter increases on the disk but not sure what is causing it.

     

    below logs you will see frigate in the file names thats because i have frigate as well. but i am still using Blue Iris as my frigate is not stable yet and I am trying fix it so i can completly remove the need of Blue Iris. And i stopped Frigate container as well to ensure that is not the culprit.

     

    the last file modification was around 4.35 AM which is just before the mover finsihed running.

     

    root@Fusion:/mnt/disk3# find . -type f -printf "%t - %p\n" | sort -n | tail -10
    Wed Jun  8 04:35:09.0360880340 2022 - ./cctv/frigate/recordings/2022-06/08/04/parking/34.57.mp4
    Wed Jun  8 04:35:13.8880906960 2022 - ./cctv/frigate/recordings/2022-06/08/04/front/34.59.mp4
    Wed Jun  8 04:35:13.9640907370 2022 - ./cctv/frigate/recordings/2022-06/08/04/patio/34.59.mp4
    Wed Jun  8 04:35:19.0260935130 2022 - ./cctv/frigate/recordings/2022-06/08/04/parking/35.07.mp4
    Wed Jun  8 04:35:23.8880961790 2022 - ./cctv/frigate/recordings/2022-06/08/04/front/35.09.mp4
    Wed Jun  8 04:35:24.0620962740 2022 - ./cctv/frigate/recordings/2022-06/08/04/patio/35.09.mp4
    Wed Jun  8 04:35:33.8891016600 2022 - ./cctv/frigate/recordings/2022-06/08/04/front/35.19.mp4
    Wed Jun  8 04:35:34.0731017600 2022 - ./cctv/frigate/recordings/2022-06/08/04/patio/35.19.mp4
    Wed Jun  8 04:35:44.8341076530 2022 - ./cctv/frigate/recordings/2022-06/08/04/patio/35.29.mp4
    Wed Jun  8 04:35:54.0301126860 2022 - ./cctv/frigate/recordings/2022-06/08/04/patio/35.39.mp4


     

     

    347385051_Screenshot2022-06-08at16_16_54.thumb.png.a39d5f7b72b359473785c1c1adbcb7d0.png

    1096343202_Screenshot2022-06-08at17_50_00.thumb.png.99c55da3ec653f0ae9e3d68eb83a1c7e.png

     

    1393131896_Screenshot2022-06-08at17_51_44.thumb.png.5f4d3cbce61d039e5ae2e0eb16e4d8e4.png

     

     

    Anyone else had simialr issues ? can you point to how you fixed it ?

     

    Also attached the diagnostics, not sure if that will be helpful in anyway.

     

    Thanks

     


     

    fusion-diagnostics-20220608-1829.zip

  9. Anyone else having this issue where this plugin is causing some error at startup ? this erros is seen at Unraid startup. it goes away when the system is started/rebooted without the vm backup plugin installed. 

     

    
    error: failed to connect to the hypervisor 
    error: Operation not supported: Cannot use direct socket mode if no URI is set

     

    as i understand, it's because the VM service isn't started when the plugin is being installed during boot. is there a setting in the plugin that can be set so the plugin only kicks off after the VM service is started or it wont start at all if the vm/array auto start is disabled. 

     

    Note: i am running unraid 6.10.2

  10. I have a freshly installed unraid running on 6.10.2 which I setup couple of days ago. I then installed all plugins, dockers and VMs i want. at some point today the server went unresponsive and I could not reach the device over network nor via the monitor attached to the server. So i decided to hard-shutdown the system (press and hold the power button), once the server booted, i took a diagnostics (attached) and tried to start the array. after couple of seconds, the browser hung with a popup (saying i need to refresh to go back to previous window) which i followed. for a moment i read a notification on the screen which said something about stale configuration. so i googled and found the below post and deleted the super.dat file and now the array is starting after I reassigned all drives. hopefully no data is lost and it will stay up.

     

    I am curious to find any evidence of the crash and what lead to it. Previously I had the 6.10.0 rc-8 and 6.10.0/1 running for several days without any issues. Only change I made is adding new cache (samsung 870 QVO ) and a PCI card with Coral accelerator (mini PCI version), not sure any of this caused crash. I don't have much data to lose at the moment (and I have a backup) but it is a couple of days work getting everything installed from scratch if it decides to crash again.

     

    my main HW is

     

    Intel i7 12700K (not overclocked)

    32GB DDR4 2666MHz

    MSI z590 A-PRO MB

    Seagate Exos 14TB Parity

    2xSeagate Exos 14TB Data

    1 x Segate barracuda 6TB

    1 x WD RED 6TB

    1 x WD Blue M.2 (Cache for appdata/vm)

    1x Samsung 870 SSD for normal cache)

     

     

    Any hint is appreciated.

     

     

    fusion-diagnostics-20220530-1613.zip

×
×
  • Create New...