Jump to content

civic95man

Members
  • Posts

    224
  • Joined

  • Last visited

Posts posted by civic95man

  1. 3 hours ago, erniej said:

    Any advice or thoughts on this?

    You are using a Marvell sata controller for a majority of your drives.  Marvell controllers are not recommended as they can drop drives and cause problems. Best bet would be to find a LSI controller.

    3 hours ago, erniej said:

    I'm just going to pull this drive and take it out of service than risk this type of thing happening again.   Alternatively, if I pull this drive now - do you think unRaid would show it as a failed drive and them emulate the previous contents?   

    If it's not currently disabled then the emulated drive will have the same contents as the physical drive.  Parity is always in sync with the drives.  You may need to read up on how parity works, it's not a backup but just a way to recover from one (or two with dual parity) missing/failed drives.  Parity cannot recover from data loss resulting from writes to disk (ie. formatting, deleting files, ransomware, etc).  You would need a full backup solution in that case.

     

    Edit:  It also looks like you're using a port multiplier on that Marvell controller.  That is also not recommended.

    • Like 1
  2. Because you have CA appData Backup running every morning at 5:00.  When it backs up all of the containers, it first shuts them down.

     

    Sep 30 05:00:01 Tower CA Backup/Restore: #######################################
    Sep 30 05:00:01 Tower CA Backup/Restore: Community Applications appData Backup
    Sep 30 05:00:01 Tower CA Backup/Restore: Applications will be unavailable during
    Sep 30 05:00:01 Tower CA Backup/Restore: this process.  They will automatically
    Sep 30 05:00:01 Tower CA Backup/Restore: be restarted upon completion.
    Sep 30 05:00:01 Tower CA Backup/Restore: #######################################

     

    Maybe set the backup schedule to something like 3 days or more if you can handle running that long without a valid backup.

    • Like 1
  3. 51 minutes ago, JukeBug said:

    I have now gone down the VM rabbit hole with my research.

    Thats an approach that many people seem to take.  Just keep in mind that you *may* run into stability issues since windows is in a VM (not saying you will and probably wont), but just keep that in mind that it may crash/lockup.  I personally run 2 windows VMs 24/7 with absolutely no issues (~150 days uptime) - but the potential is always there.

    51 minutes ago, JukeBug said:

    1. After researching, I think this would be sufficient to convert my windows machine to UNRAID + windows VM. I would also use a Plex docker. Would I need to beef up any of the hardware?

    That would be a good approach and the cpu has a good amount of cores/threads to share between the VM and unraid/plex.  It also depends on what you want to use the windows VM for. With plex, if you transcode, you may run into issues depending on how many cores you assign to the VM vs leaving free for unraid and the content to transcode.

    51 minutes ago, JukeBug said:

    2. I am fine with tinkering to get it to work, but would there be any limitations to using VM for my everyday Windows machine that I may be overlooking?

    Depends on what you want to use it for.  Gaming? Surfing the web? You just want to make sure that you keep enough cores available to unraid (maybe 4 threads max for VM and 8 for unraid).  Too many people assign all of their cores to the VM and leave nothing for unraid to manage the VM with.

     

    If you want to test, just pull your windows HD out and try booting that system with your flash.  If you feel comfortable, you could try creating a VM and using the HD with the windows install as your disk.  I believe their are tutorials out there for converting a baremetal w10 to a VM.  I would recommend a flash drive backup before doing this just in case.

     

    EDIT:  I forgot the most important thing - you need to make sure that your IOMMU groups are split ideally for VM use.  It would probably be a good idea to have a few USB ports on its own group.  That would allow you to pass them through and allow them to natively function within windows (so you can easily plug in flash drive, etc)

    • Thanks 1
  4. 6 hours ago, Urbanpixels said:

    I had separate dockers on static IP's using VLAN's and still had these issues

    Your unraid system is on vlan 130; are your docker containters using a different vlan?  I believe ALL of your containers that have a static IP must be on a different vlan than that of your unraid system.  My understanding for this is because most routers will not route broadcast packets between networks - with the broadcast packets being the cause of the macvlan call traces.

  5. 1 minute ago, JukeBug said:

    So unfortunately the computer still froze while using the Lubuntu OS from a different USB drive. I am guessing this means the motherboard is toast.

    Yes, that would mean that its at the hardware level

    2 minutes ago, JukeBug said:

    what is my best bet for finding a motherboard with an AM2 socket, 6 sata connections, and a couple of PCIE slots

    if you are comfortable with used, then i would check ebay.  I acquired most of my build through parts from ebay.  Just test everything really well before you are comfortable with it and any return policy expires.

     

    otherwise you would need to look into NOS, but at that point it may be cheaper to use lower end modern equipment and buy everything new.

    • Thanks 1
  6. 7 hours ago, trurl said:

    Connection issues on 2 disks, disk4 and another that might have been disk6, serial ending EZJ3.

    Looks like you are using a Marvell controller for four of your drives, two of which are ones that you were having issues with in this post.  I don't believe Marvell controllers are recommended and all of your connections issues may be caused by this card.  It could have also been the cause of the corrupted file system on disk6.  

     

    As for formatting disk6, judging by your logs, it looks like disk6 (sdh) drops in and out because of connection issues (presumably). I can only guess that it looks like it drops out when it should have been assigned the "disk 6" designation but still maintains a md6 presence.  so the format is performed against md6.  Just my educated guess.. in that case, maybe this is a bug

  7. The /config of your flash drive backup has all of the settings. If you want to restore everything then just copy it all over to your current flash drive (overwriting everything).  Optionally, if you just care about your disk assignments then just copy over the super.dat file.

    • Thanks 1
  8. 11 minutes ago, JukeBug said:

    can I simply replace my Unraid USB with a LUBUNTU bootable USB (different USB drives)

    Yes, remove the unraid USB and insert the bootable flavor of our choise.

     

    12 minutes ago, JukeBug said:

    without creating any problems for my HDD data

    as long as you do not try to mount any of your disks outside of unraid, then you will be fine and parity will stay valid.  If you don't trust yourself then maybe disconnect all of your drives. I don't know if the live versions will try to auto mount disks now - its been a while since I've used one, but my assumption is that they won't mount until you try to access them (via GUI if you prefer) 

  9. Maybe try booting your system with a live version of another flavor of linux to see if you have similar issues (lockup/boot problems).

     

    Could also try clearing the bios to defaults and checking the battery as a last resort.

     

    I didn't see anything in the manual about not using the pcie x16 slot for anything but graphics.

    • Thanks 1
  10. Your syslog shows that your network adapter for eth0 might be having problems.  This may or may not be related to this option:

    pcie_acs_override=downstream

    have you tried running your VM without this option?  Is that option necessary for your IOMMU groups to properly split?

     

    Failing that, you could try the beta version of unraid as it has a newer kernel with a newer version of qemu.

  11. 13 minutes ago, animeking said:

    if i want it to go back to transcoding for plex/emby all i would have to do is delete the vm correct?

    no, you would need to use the nvidia build, and remove the "stub" for the card.  The stub is reserving the card for the VM and is preventing unraid from seeing it and thus prevents the driver from loading.  By removing the stub, I mean remove it and all reference to it's related iommu in the vfio-pci plugin (assuming you are using the plug-in).

     

    You do not need to delete the VM.  However, if you try starting the VM when the card is set for transcoding, bad things could happen - best case, the VM fails to load and just dies; worst case, the VM rips the card from an active transcode and your system crashes.

  12. 20 minutes ago, animeking said:

    would this be better:

     

    rclone mount "google_drive_cache": /mnt/user/Media/GoogleDrive --allow-other --dir-cache-time 1000h --log-level INFO --timeout=1h --umask 002 --user-agent shedman214 --cache-dir=/mnt/user/Media/GoogleDrive_cache --vfs-cache-mode writes --vfs-cache-max-size 1G --vfs-cache-max-age 10s &

    I'm sorry, but I am the wrong person to ask.  I was just looking around with google and found a lot of people having issues similar to yours. The uninitialized csrf_token seems related to the the rootfs filling up. And a mis-configured rclone seems to eat up a lot of memory when many files are open.  I guess you could try that setting and see if a media scan completes without messing everything up.  I wonder if you could kill the rclone process if this happens again and reclaim the memory - at least then you might be able to generate diagnostics after it happens.

     

    20 minutes ago, animeking said:

    and on the nvidia i was trying to passthrough my VM so i can use my gpu for gaming

    yes, I figured; in that case the nvidia build is pointless since the VM (windows) is handling the gpu and drivers.  The nvidia build is primarily used if you want to allow the gpu to transcode in plex/emby, in which case it provides the necessary drivers for unraid/linux to see and talk to it.  Unfortunately, it's one or the other: either keep it in unraid for transcoding OR pass through to the VM

  13. 12 minutes ago, animeking said:

    is there a way to fix that

    Like I said, I'm no expert on that front.  From what i was able to understand as a possible fix would be to reduce the --vfs-cache-max-age to seconds (10s).  In my experience with plex, it will try to scan all of my media and keep multiple files open.  And from my understanding of rclone and your configuration, it will try to keep those open files in memory for 336h.  There are a few posts out there that describe optimizations for rclone and streaming content in plex/emby.  That might help you. You could also try the rclone support thread.

     

    And in regard to the 'nvidia build' of unraid, for troubleshooting, its best to simplify the process and that build is not "officially" supported (third party).  Also, since you stubbed that video card anyway to pass it to a VM, unraid itself can't see it so the driver will never load and becomes pointless; and spams the log.

  14. I am by no means an expert here but my best guess is that rclone is using up all of your memory to cache your media files when plex is performing a scan.  I tried to read through the different cache and buffer options and meanings but now my head just hurts.

  15. Judging by looking around on other posts, you might be filling up your rootfs (remember this resides in ram) you already assigned half of your memory to a VM.  I assume you have plex - are you transcoding in there to RAM as well (/tmp)?

  16. Sorry, I was reviewing the motherboard manual and no, you can't select the primary display.  Its usually the card in the first slot so you could try to physically swap your 2080ti and 1650 in their slots and see if the 1650 comes up as primary.  If you do this, it will most likely swap around their location within the IOMMU group and move them to different addresses, so you will need to change your binding and the VM gpu assignment.  

     

    But since you have it working, you can decide if you really need to admin it locally or remotely via the web interface. 

     

    In either case I'm glad you have it working and glad I could help you some.

    • Thanks 1
  17. That is what I expect, I'm assuming that is the output from the 2080ti.  So basically, at that point the kernel binds that video card at the specified hardware location (based on the vfio-cfg.cfg file) to the vfio-pci driver. Thats basically just a placeholder for when the VM starts and it gets passed through.  But once that card binds with that driver, unraid can no longer see or talk to it.

     

    So I expect that and your system is still functioning, not hanging.

     

    3 minutes ago, ne10g said:

    Interestingly, if I unplug the DP cable from the 2080ti then I can boot, certainly headless. Will try GUI now.

    The 2080ti must be checking if the monitor is attached, if not then it excludes itself from being available. Does the other 1650 takes over?

     

    So you have three options it sounds like, see if you can specify in the bios which card is the primary, rearrange your gpus on the motherboard so the 1650 appears as the primary, or find a kernel boot parameter to specify which card to treat as the primary (haven't checked if possible but I'm sure it it)

  18. 6 minutes ago, ne10g said:

    I hit the hurdle that I had with only one gpu, I go into device settings, then select to bind the 2080ti (it selects all 4 devices, including sound etc..) then as soon as I reboot it hangs.

    What do you mean by it hangs?  Do you get to the boot menu to choose the GUI/non-GUI/safemode and then nothing appears on the screen?

     

    Try switching to the other gpu (1650) when that happens.  When you isolate the 2080ti, unraid essentially doesn't see it so it can't use it.  Therefore it will fall back to its other gpu (1650).  Failing that, it will go into "headless" mode and not use any video card, in which case you should still see it on the network.

  19. 18 minutes ago, ne10g said:

    So much clearer now! I've just popped back to update.

     

    1) The trial is no good as i now have too many disks...lol! i will purchase a license anyway, it's gonna be worth it!

    2) I ran out of SATA data connectors, typical! Managed to scrape 3 together, so for now just put the 3x 14Tb drives in. Assuming I can put the 4th in and set as the parity disk at a later date?

     

    The journey begins properly now, thanks again!

    1.  Trial should be good for more than 3 disks. Do you have internet connection to your server so that it can "phone home?"  It only contacts unraid's servers while under the trial license.  You If you aren't already on the beta, then you may need to for the network drivers.  Else, maybe connect to a standard 1gb port instead of a 2.5gbe if your board has those - this is all assuming that your server isn't visible on the network, if it is then disregard.

     

    2.  Yes that is perfectly fine.  You can easily add new drive as you please, including parity.  Granted, it will take a good part of a day to create the parity so be aware of that when the time comes.  Not so much critical now since you don't have parity, but you want to ensure all of your disk connections are high quality (power and data).  Don't want to use questionable sata cables as they can cause problems, same for power. 

  20. 12 minutes ago, ne10g said:

    Am I right in setting up like:

     

    1 x 14TB Parity disk

    3 x 14TB Disks

    yeah, if you want parity (i recommend it, but remember that parity is not a backup) then that looks good.  With parity, you will want to make sure turbo write is on (don't know if that is default by now but most likely is).  If you are initially transferring data from another source then some people don't bother with parity until everything has been transferred to speed things up. if you don't have anything to transfer or very little then it doesn't hurt to setup parity now.

     

    15 minutes ago, ne10g said:

    Then what would be the best way to utilise the 2 x 2TB NVME's and 1x512GB SSD please?

     

    I'd like Plex and the dockers to have use of most of the fast space on the NVME's. 

    You can setup the NVMe(s) as your cache and then run docker/VM from there.  It's really your preference.  If you decide to use both NVMes as cache then you can choose raid0 or raid1 (raid0 will get you 4tb storage, whereas raid1 will give you a 2tb mirror which can protect you if lose one of the devices, raid0 cannot).  I would recommend the backup plugin on CA which can backup your docker containers to your array so if things do go south then you have a backup of your containers.

     

    As for the single SSD, its your call. You could utilize that to run your VM, or with the beta version of unraid, you can set it up as another pool

    • Thanks 1
  21. 8 minutes ago, ne10g said:

    Then (the bit I was missing!) when I start W10 with the 2080ti dedicated to it, I can use the main monitor, just use the other input?

    That's what I do.  Granted, I just use mine headless - I still have the monitor attached to the igpu for diagnostics and to mess with the BIOS.  I don't even boot up into the GUI mode - all of my admin work is done on another computer (or even my phone in a pinch).  Most of the time I have win10 running within the VM and output going to the monitor with it's dedicated mouse/keyboard.

     

    11 minutes ago, ne10g said:

    I've been dabbling with the server for a few days, then realised I needed another GPU, so just got that today and will try again. I was using the nvidia build funnily enough, stumbled upon that in the CA, seems very good!

    I'm assuming you're running the beta version - 6.9? with that new of hardware, I'm sure you must be.

    12 minutes ago, ne10g said:

    The penny has dropped now, I should just run it headless, then I can manage it either from inside the W10 VM, or from an ipad or something on the same network.  Then (the bit I was missing!) when I start W10 with the 2080ti dedicated to it, I can use the main monitor, just use the other input?

    Funny enough, i don't remember if there is an option to make the VM autostart on boot, otherwise you will need a way to start the VM - but your ipad will work.  I personally don't have anything on mine set to autostart - simply because if the server restarted, it wasn't because of me and I want to track down what caused it before potentially making things worse.

     

    14 minutes ago, ne10g said:

    I knew I was being thick haha!

    It's always better to ask questions when in doubt rather than [potentially] making reckless mistakes, such as when it comes to your data

    • Thanks 1
×
×
  • Create New...