alturismo

Moderators
  • Posts

    6101
  • Joined

  • Last visited

  • Days Won

    47

Report Comments posted by alturismo

  1. 36 minutes ago, limetech said:

    Please capture the system log:

    
    cat /var/log/syslog /boot/syslog.txt

    That will put syslog.txt in root of usb flash boot device.

    cat didnt work out, always "no such file .." on target

     

    cat /var/log/syslog /boot/syslog.txt

    ....

    Oct 15 16:50:05 AlsServer sshd[8312]: Starting session: shell on pts/0 for root from 88.73.177.160 port 47984 id 0
    cat: /mnt/user/appdata/syslog.txt: No such file or directory
    root@AlsServer:~# 

     

    i copied the syslog file now to be sure ;)

  2. Hi, issue here on 6.9 30 Beta, webui not accessible anymore since somewhen today ...

    root@AlsServer:~# uptime
     15:56:46 up 8 days,  9:04,  1 user,  load average: 16.96, 16.27, 15.04

    pretty much since the release of the .30 beta, worked like a charm since today ;)

     

    before i try to restart nginx (would be nice to have a proper way todo so on unraid) i try to fetch some diagnostics,

    bus seems like i have no luck with that ...

    root@AlsServer:~# diagnostics
    Starting diagnostics collection... 

    Resting there forever ...

     

    VM's, Dockers, Shares, ... all seem to work fine, only webui is not working anymore

     

    may some hints howto "safe" restart nginx on unraid ? thanks ahead

  3. Hi, seems i have a issue now when trying to move my VM's from UAD to a 2nd cache pool.

     

    my libvrt service fails to start after shutting down the VM system and trying to restart it.

     

    made a reboot now with the same result.

     

    any chance to "repair" or only solution, wipe the img file and redo VM's with existing vdisks and hope i match the settings ;)

     

    ### EDIT, sorry, my fault, just saw i made a wrong path for default VM's ...

     

    logs and diags attached

     

  4. g morning, after updating to .25 following in the buttom line

     

     

     Array Started•Warning: file_get_contents(/sys/block/ROOTFS/queue/rotational): failed to open stream: No such file or directory in /usr/local/emhttp/plugins/unassigned.devices/include/lib.php on line 635

     

    all looking good (also UAD), just this error line came up

     

    logs attached

     

    alsserver-syslog-20200713-0429.zipalsserver-diagnostics-20200713-0631.zip

  5. ok, seems i was too fast, i guess its another issue why this happens

     

    when i read changelog again ;)

     

    webgui: VMs: change default network model to virtio-net

     

    this virtio-net does not exist here, i only have those

     

    image.thumb.png.b7a2a8181e228cd44ad392ed3701229b.png

     

    so i have chosen the virbr0 which is 192.168.122.0/24, whatever this is about cause i never added it.

     

    so, how can i add this virtio-net to unraid ?

     

    as note, when i add a new vm i also dont have the option, its still default to br0 and i can choose from br0, br1 or virbr0 only.

  6. may another question about VM and virtio network, i see this correct that they have to be on their own subnet ?

     

    its weird here now due my vm's are on 192.168.122.x while my regular net is 192.168.1.x

     

    i never added or changed anything on virtio, so its default i guess.

     

    is this a must have ? cause i guess i ll run into issues when i try to rdp to my VM from non unraid mashines.

     

    ### Update, confirmed, i cant reach any win10 vm from either guac or laptop in LAN Network 192.168.1.x .... so, this is still open when i read correctly that the kernel errors appear when using the same network as dockers do ?

  7. may some questions before trying the beta

     

    i currently use

    1 cache drive (nvme 1tb) xfs formatted

    1 UD device where my VM's sit, nvme 500gb xfs formatted

    4 drives in array all xfs formatted

     

    when i now upgrade, does this collide with the current setup ? do cache drives have to be btrfs or am i still good to go ? reason why xfs, my cache drive was btrfs before and crashed 2x completely with complete data loss ... since xfs im good. when i read correctly this should still be fine as single drive pool.

     

    Next would be, change my current UD drive to cache2 (sep single drive cache pool for my VM's), also possible as its a seperated pool with single formatted xfs drive ?

    or rather keep it UD

     

    VM's need network change so the kernel error would be gone, when changing this (webui) also probably all manual changes still gone or are they persistent now ? (sample cpu pinning)

     

    VFiO PCIe config should be obsolete then when using the new setting feature, uninstall plugin before update or doesnt matter ?

     

    last for now, new kernel fixes nested virt ? ... win 10 wsl2 subsystem activation lets VM crash badly, any changes done there in webui or still same procedure to activate ?

     

    for some notes thanks ahead

  8. installed over 683, so far looking ok for unraid.

     

    my main win10 VM aint booting anymore (blue screen, diagnostic, start issues ...) but i guess its just bad luck cause the 2nd win10 vm boot just fine ... so i guess its not 69 beta related ;) also the ubuntu mashine spins up fine.

     

    so i look to check if i can repair the win10 vm somehow

  9. 11 hours ago, ljm42 said:

    Thanks for the fix @bluemonster !

     

    Here is a bash file that will automatically implement the fix in 6.7.2 (and probably earlier, although I'm not sure how much earlier):
      https://gist.github.com/ljm42/74800562e59639f0fe1b8d9c317e07ab

     

    It is meant to be run using the User Scripts plugin, although that isn't required.

    Note that you need to re-run the script after every reboot.

    Remember to uninstall the script after you upgrade to Unraid 6.8

    More details in the script comments. 

    also thanks from here to @bluemonster for the fix and @ljm42 for the script

  10. ok, happened again today

     

    i was testing some http api commands for tvheadend and one command stalled ... after that there was no more repsond, also could not quit the command at all, even a tvheadend restart (to definately kill the http connection) also didnt help, just as note.

     

    /etc/rc.d/rc.nginx restart <- worked then well here

  11. 8 hours ago, limetech said:

    Finally had this happen on a test server.  I could not find any reason why the ttyd daemon quit responding to connection requests.  Probably this is a bug in ttyd or libwebsockets it depends on.  We have updated the latter in the next release and let's see if this happens again.

     

    To recover from this you can use telnet/ssh/console and type this command:

    
    /etc/rc.d/rc.nginx restart

     

    thanks for the Tipp

  12. g morning, just as quick test today morning i see it happened again, as note ...

    command was ps x

     

    the last entry from system log

     

    Mar 1 07:01:42 AlsServer login[12318]: ROOT LOGIN on '/dev/pts/2'

     

    is there no option to restart the service ? exept server reboot ?

    image.thumb.png.1f4f738459e2d8d420ba7cb62de76974.png

  13. ok, happened again now, i have the webterminal windows open where it "stalled", i opened a new window to see what happens and its the same behavior, about:blank ....

     

    happened in the middle of typing, not even executed anything ...

     

    heres a screenshot and diags attached

    image.thumb.png.8b03814e357ea1dde32223995aa6578c.png

     

    now, what to look for or what diags could be helpful, i keep all open for now ...

    alsserver-diagnostics-20190225-0609.zip