Jump to content


Popular Content

Showing content with the highest reputation since 10/11/19 in all areas

  1. 12 points
    Note: this community guide is offered in the hope that it is helpful, but comes with no warranty/guarantee/etc. Follow at your own risk. What can you do with WireGuard? Let's walk through each of the connection types: Remote access to server: Use your phone or computer to remotely access your Unraid server, including: Unraid administration via the webgui Access dockers, VMs, and network shares as though you were physically connected to the network Remote access to LAN: Builds on "Remote access to server", allowing you to access your entire LAN as well. Server to server access: Allows two Unraid servers to connect to each other. LAN to LAN access: Builds on "Server to server access", allowing two entire networks to communicate. May require additional settings, TBD. Server hub & spoke access: Builds on "Remote access to server", except that all of the VPN clients can connect to each other as well. Note that all traffic passes through the server. LAN hub & spoke access: Builds on "Server hub & spoke access", allowing you to access your entire LAN as well. VPN tunneled access: Route traffic for specific Dockers and VMs through a commercial WireGuard VPN provider (see this guide) Remote tunneled access: Securely access the Internet from untrusted networks by routing all of your traffic through the VPN and out Unraid's Internet connection In this guide we will walk through how to setup WireGuard so that your trusted devices can VPN into your home network to access Unraid and the other systems on your network. Prerequisites You must be running Unraid 6.8 with the Dynamix WireGuard plugin from Community Apps Be aware that WireGuard is is technically classified as experimental. It has not gone through a full security audit yet and has not reached 1.0 status. But it is the first open source VPN solution that is extremely simple to install, fast, and designed from the ground up to be secure. Understand that giving someone VPN access to your LAN is just like giving them physical access to your LAN, except they have it 24x7 when you aren't around to supervise. Only give access to people and devices that you trust, and make certain that the configuration details (particularly the private keys) are not passed around insecurely. Regardless of the "connection type" you choose, assume that anyone who gets access to this configuration information will be able to get full access to your network. This guide works great for simple networks. But if you have Dockers with custom IPs or VMs with strict networking requirements, please see the "Complex Networks" section below. Unraid will automatically configure your WireGuard clients to connect to Unraid using your current public IP address, which will work until that IP address changes. To future-proof the setup, you can use Dynamic DNS instead. There are many ways to do this, probably the easiest is described in this 2 minute video from SpaceInvaderOne If your router has UPnP enabled, Unraid will be able to automatically forward the port for you. If not, you will need to know how to configure your router to forward a port. You will need to install WireGuard on a client system. It is available for many operating systems: https://www.wireguard.com/install/ Android or iOS make good first systems, because you can get all the details via QR code. Setting up the Unraid side of the VPN tunnel First, go to Settings -> Network Settings -> Interface eth0. If "Enable bridging" is "Yes", then WireGuard will work as described below. If bridging is disabled, then none of the "Peer type of connections" that involve the local LAN will work properly. As a general rule, bridging should be enabled in Unraid. If UPnP is enabled on your router and you want to use it in Unraid, go to Settings -> Management Access and confirm "Use UPnP" is set to Yes On Unraid 6.8, go to Settings -> VPN Manager Give the VPN Tunnel a name, such as "MyHome VPN" Press "Generate Keypair". This will generate a set of public and private keys for Unraid. Take care not to inadvertently share the private key with anyone (such as in a screenshot like this) By default the local endpoint will be configured with your current public IP address. If you chose to setup DDNS earlier, change the IP address to the DDNS address. Unraid will recommend a port to use. You typically won't need to change this unless you already have WireGuard running elsewhere on your network. Hit Apply If Unraid detects that your router supports UPnP, it will automatically setup port forwarding for you: If you see a note that says "configure your router for port forwarding..." you will need to login to your router and setup the port forward as directed by the note: Some tips for setting up the port forward in your router: Both the external (source) and internal (target/local) ports should be the set to the value Unraid provides. If your router interface asks you to put in a range, use the same port for both the starting and ending values. Be sure to specify that it is a UDP port and not a TCP port. For the internal (target/local) address, use the IP address of your Unraid system shown in the note. Google can help you find instructions for your specific router, i.e. "how to port forward Asus RT-AC68U" Note that after hitting Apply, the public and private keys are removed from view. If you ever need to access them, click the "key" icon on the right hand side. Similarly, you can access other advanced setting by pressing the "down chevron" on the right hand side. They are beyond the scope of this guide, but you can turn on help to see what they do. In the upper right corner of the page, change the Inactive slider to Active to start WireGuard. You can optionally set the tunnel to Autostart when Unraid boots. Defining a Peer (client) Click "Add Peer" Give it a name, such as "MyAndroid" For the initial connection type, choose "Remote access to LAN". This will give your device access to Unraid and other items on your network. Click "Generate Keypair" to generate public and private keys for the client. The private key will be given to the client / peer, but take care not to share it with anyone else (such as in a screenshot like this) For an additional layer of security, click "Generate Key" to generate a preshared key. Again, this should only be shared with this client / peer. Click Apply. Note: Technically, the peer should generate these keys and not give the private key to Unraid. You are welcome to do that, but it is less convenient as the config files Unraid generates will not be complete and you will have to finish configuring the client manually. Configuring a Peer (client) Click the "eye" icon to view the peer configuration. If the button is not clickable, you need to apply or reset your unsaved changes first. If you are setting up a mobile device, choose the "Create from QR code" option in the mobile app and take a picture of the QR code. Give it a name and make the connection. The VPN tunnel starts almost instantaneously, once it is up you can open a browser and connect to Unraid or another system on your network. Be careful not to share screenshots of the QR code with anyone, or they will be able to use it to access your VPN. If you are setting up another type of device, download the file and transfer it to the remote computer via trusted email or dropbox, etc. Then unzip it and load the configuration into the client. Protect this file, anyone who has access to it will be able to access your VPN. About DNS The 2019.10.20 release of the Dynamix Wireguard plugin includes a "Peer DNS Server" option (thanks @bonienl!) If you are having trouble with DNS resolution on the WireGuard client, return to the VPN Manager page in Unraid and switch from Basic to Advanced mode, add the IP address of your desired DNS server into the "Peer DNS Server" field, then install the updated config file on the client. You may want to use the IP address of the router on the LAN you are connecting to, or you could use a globally available IP like This is required for "Remote tunneled access" mode, if the client's original DNS server is no longer accessible after all traffic is routed through the tunnel. If you are using any of the split tunneling modes, adding a DNS server may provide name resolution on the remote network, although you will lose name resolution on the client's local network in the process. The simplest solution is to add a hosts file on the client that provides name resolution for both networks. Complex Networks (added Oct 24) The instructions above should work out of the box for simple networks. With "Use NAT" defaulted to Yes, all network traffic on Unraid uses Unraid's IP, and that works fine if you have a simple setup. However, if you have Dockers with custom IPs or VMs with strict networking requirements, things may not work right (I know, kind of vague, but feel free to read the two WireGuard threads for examples) A partial solution is: In the WireGuard config, set "Use NAT" to No In your router, add a static route that lets your network access the WireGuard "Local tunnel network pool" through the IP address of your Unraid system. For instance, for the default pool of you should add this static route: Network: (aka with subnet Gateway: <IP address of your Unraid system> (Note that this covers the entire class B 10.253.x.x network, so you can add other WireGuard tunnels without having to modify your router setup again.) With these changes, your network should work normally. However, your WireGuard clients still may not be able to access Dockers on custom IPs or VMs. If you find a solution to this, please comment!
  2. 12 points
    Overview of what Macinabox does. This is a container that is designed to help make installing a macOS KVM Virtual Machine very easy. The VM doesnt run in a docker container but runs as a full fat Unraid KVM VM selectable in the VM tab of the webUI. However your server's hardware must be 'fairly' modern to run a macOS VM. You are going to need a CPU that supports SSE 4.2 & AVX2 for macOS Mojave and above to work. Both Intel and AMD processors are fine to use. To use just select the OS type, vdisk type and size. ( I suggest you use raw disk type) Then let Macinabox make a vdisk for the install, download the recovery media, clover boot-loader and create a vm xml file that is preconfigured to work. (The xml files created will have unique uuids and network mac addresses) Sit back and let the container do its stuff - note - To see the progress of the container, you do this by looking at the log whilst it runs. You will know when it has finished as you will see a message saying to stop then start the array. This container doesn't have a webUI (but clicking on the webUI button of this container will just take you too a video of how to use this container) - - - - - - - - - - - So after the container has done its stuff. Stop the array then start it again and the VM will become visible in the Unraid VM manger. (you will not see it if you dont do this) Click start to start the VM and you will boot into a clover boot-loader. Then press enter to continue to load the recovery media. Goto disk utility and format the vdisk. Close disk utility. Select re-install macOS then sit back and wait until done.. Please be patient when installing as the install speed will depend on your internet connection and how busy the Apple servers are. After installing the VM don't run the container again or else it will overwrite the vdisk with the install on. (I will change this so it cant happen soon) Probably best after installing to remove the container for now just to be safe. edit - I have now added checks to stop the container re downloading install media if run again. It will also check for an existing vdisk and if found not create another and therefore not overwrite it. Same goes for the xml file. However if the container is run again it will download another clover and ovmf files. I have done this so people can easily update clover and ovmf files if needed.
  3. 8 points
  4. 4 points
    tldr: Starting with 6.8.0-rc2 please visit Settings/Disk Settings and change the 'Tunable (scheduler)' to 'none'. Then run with SQLite DB files located on array disk shares and report whether your databases still become corrupted. When we first started looking into this issue one of the first things I ran across was this monster topic: https://bugzilla.kernel.org/show_bug.cgi?id=201685 and related patch discussion: https://patchwork.kernel.org/patch/10712695/ This bug is very very similar to what we're seeing. In addition Unraid 6.6.7 is on the last of the 4.18 kernels (4.18.20). Unraid 6.7 is on 4.19 kernel and of course 6.8 is on 5.3 currently. The SQLite DB Corruption bug also only started happening with 4.19 and so I don't think this is coincidence. In looking at the 5.3 code the patch above is not in the code; however, I ran across a later commit that reverted that patch and solved the bug a different way: https://www.spinics.net/lists/linux-block/msg34445.html That set of changes is in 5.3 code. I'm thinking perhaps their "fix" is not properly handling some I/O pattern that SQLite via md/unraid is generating. Before I go off and revert the kernel to 4.18.20, please test if setting the scheduler to 'none' makes any difference in whether databases become corrupted.
  5. 3 points
    LibreELEC, TBS-OS, Digital Devices v6.8.0rc1 all done. TBS-Crazy-Cat broken at the moment.
  6. 2 points
    You can help a lot by describing in more detail the I/O patterns that lead to this. For example, suppose you wanted to describe to someone how to reproduce the error - how would you describe it? Saying, "bad performance still" does not help.
  7. 2 points
    ***Update*** : Apologies, it seems like there was an update to the Unraid forums which removed the carriage returns in my code blocks. This was causing people to get errors when typing commands verbatim. I've fixed the code blocks below and all should be Plexing perfectly now Y =========== Granted this has been covered in a few other posts but I just wanted to have it with a little bit of layout and structure. Special thanks to [mention=9167]Hoopster[/mention] whose post(s) I took this from. What is Plex Hardware Acceleration? When streaming media from Plex, a few things are happening. Plex will check against the device trying to play the media: Media is stored in a compatible file container Media is encoded in a compatible bitrate Media is encoded with compatible codecs Media is a compatible resolution Bandwith is sufficient If all of the above is met, Plex will Direct Play or send the media directly to the client without being changed. This is great in most cases as there will be very little if any overhead on your CPU. This should be okay in most cases, but you may be accessing Plex remotely or on a device that is having difficulty with the source media. You could either manually convert each file or get Plex to transcode the file on the fly into another format to be played. A simple example: Your source file is stored in 1080p. You're away from home and you have a crappy internet connection. Playing the file in 1080p is taking up too much bandwith so to get a better experience you can watch your media in glorious 240p without stuttering / buffering on your little mobile device by getting Plex to transcode the file first. This is because a 240p file will require considerably less bandwith compared to a 1080p file. The issue is that depending on which format your transcoding from and to, this can absolutely pin all your CPU cores at 100% which means you're gonna have a bad time. Fortunately Intel CPUs have a little thing called Quick Sync which is their native hardware encoding and decoding core. This can dramatically reduce the CPU overhead required for transcoding and Plex can leverage this using their Hardware Acceleration feature. How Do I Know If I'm Transcoding? You're able to see how media is being served by playing a first something on a device. Log into Plex and go to Settings > Status > Now Playing As you can see this file is being direct played, so there's no transcoding happening. If you see (throttled) it's a good sign. It just means is that your Plex Media Server is able to perform the transcode faster than is necessary. To initiate some transcoding, go to where your media is playing. Click on Settings > Quality > Show All > Choose a Quality that isn't the Default one If you head back to the Now Playing section in Plex you will see that the stream is now being Transcoded. I have Quick Sync enabled hence the "(hw)" which stands for, you guessed it, Hardware. "(hw)" will not be shown if Quick Sync isn't being used in transcoding. PreRequisites 1. A Plex Pass - If you require Plex Hardware Acceleration Test to see if your system is capable before buying a Plex Pass. 2. Intel CPU that has Quick Sync Capability - Search for your CPU using Intel ARK 3. Compatible Motherboard You will need to enable iGPU on your motherboard BIOS In some cases this may require you to have the HDMI output plugged in and connected to a monitor in order for it to be active. If you find that this is the case on your setup you can buy a dummy HDMI doo-dad that tricks your unRAID box into thinking that something is plugged in. Some machines like the HP MicroServer Gen8 have iLO / IPMI which allows the server to be monitored / managed remotely. Unfortunately this means that the server has 2 GPUs and ALL GPU output from the server passed through the ancient Matrox GPU. So as far as any OS is concerned even though the Intel CPU supports Quick Sync, the Matrox one doesn't. =/ you'd have better luck using the new unRAID Nvidia Plugin. Check Your Setup If your config meets all of the above requirements, give these commands a shot, you should know straight away if you can use Hardware Acceleration. Login to your unRAID box using the GUI and open a terminal window. Or SSH into your box if that's your thing. Type: cd /dev/dri ls If you see an output like the one above your unRAID box has its Quick Sync enabled. The two items were interested in specifically are card0 and renderD128. If you can't see it not to worry type this: modprobe i915 There should be no return or errors in the output. Now again run: cd /dev/dri ls You should see the expected items ie. card0 and renderD128 Give your Container Access Lastly we need to give our container access to the Quick Sync device. I am going to passively aggressively mention that they are indeed called containers and not dockers. Dockers are manufacturers of boots and pants company and have nothing to do with virtualization or software development, yet. Okay rant over. We need to do this because the Docker host and its underlying containers don't have access to anything on unRAID unless you give it to them. This is done via Paths, Ports, Variables, Labels or in this case Devices. We want to provide our Plex container with access to one of the devices on our unRAID box. We need to change the relevant permissions on our Quick Sync Device which we do by typing into the terminal window: chmod -R 777 /dev/dri Once that's done Head over to the Docker Tab, click on the your Plex container. Scroll to the bottom click on Add another Path, Port, Variable Select Device from the drop down Enter the following: Name: /dev/dri Value: /dev/dri Click Save followed by Apply. Log Back into Plex and navigate to Settings > Transcoder. Click on the button to SHOW ADVANCED Enable "Use hardware acceleration where available". You can now do the same test we did above by playing a stream, changing it's Quality to something that isn't its original format and Checking the Now Playing section to see if Hardware Acceleration is enabled. If you see "(hw)" congrats! You're using Quick Sync and Hardware acceleration [emoji4] Persist your config On Reboot unRAID will not run those commands again unless we put it in our go file. So when ready type into terminal: nano /boot/config/go Add the following lines to the bottom of the go file modprobe i915 chmod -R 777 /dev/dri Press Ctrl X, followed by Y to save your go file. And you should be golden!
  8. 2 points
    For anyone running into the e1000-82545em bridging-to-br0 weirdness under Catalina, I have a workaround that's working fine for me: Install AppleIntelE1000e.kext (I'm using the latest build from the fork at https://github.com/chris1111/AppleIntelE1000e) either to /Library/Extensions (the advantage being simplicity; you can install it manually or with the simple KextBeast utility) or by injecting it with Clover (the advantage being that it will likely work while installing macOS or when booted into Recovery Mode). Change the Interface definition in your XML to use the 'e1000e' virtual NIC: <model type='e1000e'/> Having done this, I can bridge to br0 under Catalina without issue, and even access the App Store and use iCloud services. I'm hoping to be able to make 'virtio-net-pci' work one of these days, but no luck so far.
  9. 2 points
    Yeah I screwed up the release, -rc3 on the way....
  10. 2 points
    The problem was that in 6.8.0-rc1 I could not access the the Flash Drive ( boot drive) from Krusader. The reason being that Limetech has changed the permission to the flash drive when it is mounted in the version. The only way in this version (and all future versions) to access the drive is through SMB or as user 'root'. (Understand that they are insistent that they are not going to change.) Currently, Krusader is being run as user 'nobody'. For anyone else looking for an answer, the answer is "YES". All you have to do is edit the Krusader Docker and change the PUID and PGID to that for the 'root' user. It is currently set for the 'nobody' user. On my system, the 'root' PUID is 0 and PGID is 0. I did run some tests and I did not find any problems after making the changes.
  11. 2 points
    Good evening! I'd like to not only request this as a feature, but explain how one could implement this on their own! Basically, the idea is that all user customization done via configuration files located in "/root" are lost on each boot. I know this is intentional, but there's an "easy" way to implement this with clever failsafe mechanics. I also know that one can work around this by adding a couple of lines to /boot/config/go, and storing the configuration files on the flash drive. This isn't as desirable as Fat32 doesn't properly handle Linux permissions, and can require other manual edits to the go file down the road. Enter OverlayFS (a feature built into the Linux kernel for eons) First we create the container for our data. I use the truncate command as it is safe and "quick" (note: we are writing over USB so this step will take time no matter which option we use) truncate -s 4000M /boot/config/root.persist I chose to go with 4000M as it is close to the Fat32 ceiling of "4gb" (note: if you specify 4G you will receive an error) Next we format that image, and set up some important directories within it: mkfs.ext4 /boot/config/root.persist mkdir /tmp/overlay mount /boot/config/root.persist /tmp/overlay mkdir /tmp/overlay/upper mkdir /tmp/overlay/workdir Finally the special sauce that overlays the image we created on top of the normal unraid /root/ directory: mount -t overlay -o lowerdir=/root,upperdir=/tmp/overlay/upper,workdir=/tmp/overlay/workdir none /root Anything written to /root/ after this command is run will actually be writting to /tmp/overlay/upperdir, and permanently stored there. The lowerdir will never be modified in this situation as it isn't addressable since we are placing the overlay on top of lowerdir. And to make it persistent, we add this block to /boot/config/go: if [ -f /boot/config/root.persist ]; then mkdir /tmp/overlay mount /boot/config/root.persist /tmp/overlay mount -t overlay -o lowerdir=/root,upperdir=/tmp/overlay/upper,workdir=/tmp/overlay/workdir none /root fi A couple of notes: The if statement above makes sure that we don't try doing anything if there isn't a persistent image for the root folder. It's kind of redundant (the first and second mount commands will just fail and regurgitate errors if the file isn't there) but I prefer a clean console log. If the image becomes corrupt, or unusable you can safely discard it this way. Safe mode shouldn't use /boot/config/go so if anything goes wrong safe mode will undo any of the changes contained in the image. Meaning you can boot into safe mode, manually mount the image, and undo whatever you did in upperdir and be back up and running. I'm not sure what you could do to cause those sorts of things. This also allows for: Persistent bash history (forget that command you ran before you rebooted? No more.) Persistent config file storage (tmux preferences, terminal colors, and htop profiles? Oh my.) Anything you would normally want a home directory to be useful for in LinuxLand.
  12. 1 point
    There was an update for the container midnight and my server auto updated at 6am without problems, and looks like my dbengine database files are persistent now. So looks like the "delete obsolete charts files = no" setting solved the problem.
  13. 1 point
    I'm wondering the same thing. I'm using /mnt/user0 and /mnt/cache to ad hoc moving of data between cache and user shares without messing with share use cache configuration. Maybe it can be preserved for experienced user via system tunable?
  14. 1 point
    You can't use a subdomain of the main domain in the extra domain. That is why you get that error.
  15. 1 point
    Are they both on the same controller? Some controllers (old / ancient) only supported a max of 2.2TB
  16. 1 point
    1 thing I forgot to add. mount_unionfs will also protect your cloud storage since any change will be downloaded first and then any change is done (including delete - which unionfs just hides the file).
  17. 1 point
    Noted, and thankfully noted is all that's needed since you saved me from falling into a hole of disappointment if that's what I tried. Occasional gaming etc is enough for me in any case In this regard, Unraid gaming in general my lifestyle perfectly at the moment (kid, condo, job, minimal free time).
  18. 1 point
    remove the space after the coma for the last device id and reboot. you would have to check your scheduler and the logs to see what is happening at that time.... could be trim, could be anything.
  19. 1 point
    Side Note: It can't be included due to it using a graphing library that is incompatible with unRaid's licence (or more to the point, it can't be included without significantly raising the cost of your licence)
  20. 1 point
    I'm not using Macinabox, as I prefer to install with the "old method", so I'm not sure if you leave the default settings/paths in the container configuration if the vdisks will be overwritten or if there's some sort of folder structure depending on the chosen installed os, preventing the overwriting, or if the container will warn about that vdisks are already in place and it will not continue. Looking at the screenshot it seems the container is smart enough to install things on different vdisks, named accordingly to the chosen os? Maybe the clover image will be the same (but I repeat I'm not sure): this can be the same with no problems for basic booting, but if you need different clover images for different OSes you will need to change the xml of the vms to point to the right (and different) clover images.
  21. 1 point
    A little over a month ago I was having daily SQL errors with my Plex / Sonarr setup. I had to completely stop adding any new media and have been following this and many other threads looking for a fix. Last night I updated to 6.8.0-rc5 and have been really stressing the database today with TV and Movie updates over the past few months. So far no errors, thank you!!!
  22. 1 point
    The first thing you need to do is post diagnotics (Tools --> Diagnostics in the GUI) so your configuration and settings can be examined by the experts.
  23. 1 point
    On the post that helped you hover your mouse over the heart in the bottom right, then click the appropriate reaction.
  24. 1 point
    Btrfs is the only filesystem that supports that, and you'd need to be taking manual or scripted snapshots, if you don't know what that is then you wouldn't be doing it, so no way to recover.
  25. 1 point
    Absolutely not. Here's how it works. There are two ways to specify an encryption key: Using a passphrase. In this case what you type is exactly what will be used for the key, without any newline at the end (because did you type a newline? No). The max length for a passphrase is 512 bytes (characters). As specified in the Help, "It is highly advisable to only use the 95 printable characters from the first 128 characters of the ASCII table." A passphrase is meant to be some kind of string that you commit to memory and type in following server reboot. Use a keyfile. Whatever contents are in this file are what is used for the key. The max length used from a keyfile is 8192 bytes. Since contents are in a file, each byte can take on any value. Often one would use a file with random binary data. If the specified file is greater than 8192 in length, only the first 8192 bytes comprise the key. If you use a passphrase, whatever you typed is not stored in a file, though it is kept in memory. If you use a keyfile, then file uploaded data is written to /root/keyfile and it stays there for as long as sever has power. There is a button in webGUI that lets you delete the file if you want. But then each time you Start array you will have to upload file again, and then presumably delete again. Using a keyfile lets you specify a far stronger key, but onus is on you to delete the file.
  26. 1 point
    Let's flip the question back. Why would you want to update to the RC version if it sounds like you cannot afford to have issue? The general consensus is to always run stable on production servers and RC on test servers (or if you have a specific issue for which a fix is not yet available in stable).
  27. 1 point
    Managed to sort it out. Opened console window of Nextcloud docker and entered this command: sudo -u abc php /config/www/nextcloud/occ db:add-missing-indices After that I still had one more error, sorted it out with: sudo -u abc php /config/www/nextcloud/occ db:add-missing-indices
  28. 1 point
    It's under the settings section. Then theming. You can change the favor icon, login logo, login text and background picture. There is also a dark mode that can be turned on also.
  29. 1 point
    Not technically. 32GB is max for FAT32 with 16K cluster size, which is the max that Windows let's you set. However third party tools allow you to format FAT32 with larger sector size, and that is exactly what our USB Flash Creator tool does.
  30. 1 point
  31. 1 point
    A couple minor md/unraid driver changes made in 6.8.0-rc4 but please retest before trying below. If DB corruption persists, then there are two tweak we can try that influence how I/O is generated. Both of these make use of mdcmd set md_restrict <value> The default <value> is 3, that is both bits 0 and 1 set. Bit 0 determines how incoming I/O requests are broken up. If set, then they are broken into 128K requests. If not set they are processed as-is (meaning a really large request will be entirely scheduled by md/unraid driver. Bit 1 determines how to handle read-ahead requests when pool resources are limited. If set, then if a stripe cannot be allocated the read-ahead is terminated (this is valid). If not set, then read-ahead will compete with all other I/O until completion. To make this effective you must Stop array, type the command in terminal window, then Start array (with no browser refresh in-between). I suggest trying first: mdcmd set md_restrict 2 If still corruption, let's try: mdcmd set md_restrict 0
  32. 1 point
    No. You will not need to purchase a new license. 🙂
  33. 1 point
    Can you do a traceroute (windows = tracert) from the peer to an IP address of the docker container >tracert -d Tracing route to over a maximum of 30 hops 1 3 ms 2 ms 2 ms 2 2 ms 2 ms 2 ms Trace complete. Above is a traceroute from my laptop to one of my containers (pi-hole) with a custom IP address ( It goes over the tunnel and the container respond as well.
  34. 1 point
    Here is a link to a tutorial reported by @bland328 for iCloud/AppStore/iMessage with general and detailed information, also valid for our VMs. It seems my problem was related to something messed up in my System Preference panel, as I noticed that the error was related to it and moreover I was not able to add any account under "Internet Account". So, I deleted my administrator account and created a new one and after following all the advices in the tutorial I was able to login to iCloud; I didn't try iMessage as I'm not using it. I also switched from MacPro7,1 to iMacPro1,1 since it appears that the serial is an important parameter, and serials for MacPro7,1 may be not recognized yet since it's not on the market.
  35. 1 point
    Excellent workaround! Do you happen to know if it persists through updates to the nordvpn container?
  36. 1 point
    The procedure is correct if you WANT to reorder disks, but it is not clear from anything else you have said that this is the case. Most people simply add a new data disk in the next available slot in the Unraid GUI so that no re-ordering is required. it might be worth pointing out that re-ordering disks in the GUI has nothing to do with how they are physically connected. Unraid recognises disks by their serial number, not by where they are connected.
  37. 1 point
    v6.8.0rc3 uploaded. Still no TBS Crazy Cat....
  38. 1 point
    That's not possible. That's the easiest way to solve the issue. You should configure the 2nd NIC without IP address under network settings, and assign it to VMs
  39. 1 point
  40. 1 point
    You can't clear those errors. I had same happen to me and the drive eventually failed. If you get anymore I would suggest to replace the drive.
  41. 1 point
    You can also get that type of indication on the Dashboard if the CRC error count has increased. CRC errors do not show up when you run a SMART test as they indicate connection issues rather than disk failures. In such a case clicking the Orange icon and selecting the Acknowledge option will make the icon turn green until another error occurs. The CRC count never resets to 0 so Unraid only notifies you again if it increases.
  42. 1 point
    Yes, it's finally there
  43. 1 point
    @Marshalleq once again thank you. The issue was simpler than expected. I just had to select USB controller 3.0 and it worked. @John_M off course. Asus ROG STRIX X470-F GAMING motherboard. And the working bios is Version 4207 2018/12/17
  44. 1 point
    It should be tried by anyone having the issue, it could be fixed by the md driver changes introduced to deal with the bad v6.7 array performance, there's also a much newer kernel and many other changes.
  45. 1 point
    SAS disks cannot be spun down, and will result in that error being logged.
  46. 1 point
    @Josh.5 FYI, Someone at Datahoarding has created a template for unManic and published this on CA. I believe earlier you had stated that you did not wish unManic to be published as of yet. Is this still the case? Should the template be removed from CA? Ideally though if you would like it in CA it would be better for it to be published within your own repository.
  47. 1 point
    Suggestion for v3 -- I had a file get corrupted in my Radarr docker appdata, and had to retrieve it, and extracting the 2mb .xml file took forever because my .tar.gz backup file is huge. Can you add a setting to back up each docker container to individual .tar.gz files, rather than one enormous file?
  48. 1 point
    I have a 2GB USB2 Sandisk Cruzer that has been running in my Build since day 1. I'm at 700Meg used with everything on it.
  49. 1 point
    I also wanted to get the ZFS Event Daemon (zed) working on my unRAID setup. Most of the files needed are already built into steini84's plugin (thanks!) but zed.rc needs to be copied into the file system at each boot. I created a folder /boot/config/zfs-zed/ and placed zed.rc in there - you can get the default from /usr/etc/zfs/zed.d/zed.rc. Add the following lines to your go file: #Start ZFS Event Daemon cp /boot/config/zfs-zed/zed.rc /usr/etc/zfs/zed.d/ /usr/sbin/zed & To use built in notifications in unRAID, and to avoid having to set up a mail server or relay, set your zed.rc with the following options: ZED_EMAIL_PROG="/usr/local/emhttp/webGui/scripts/notify" ZED_EMAIL_OPTS="-i warning -s '@SUBJECT@' -d '@SUBJECT@' -m \"\`cat $pathname\`\"" $pathname contains the verbose output from ZED, which will be sent in the body of an email alert from unRAID. I have this set to alert level of 'warning' as I have unRAID configured to always email me for warnings. You'll also want to adjust your email address, verbosity level, and set up a debug log if desired. Either place the files and manually start zed, or reboot the system for this to take effect. Pro tip, if you want to test the notifications, zed will alert on a scrub finish event. If you're like me and only have a large pools that takes hours/days to scrub, you can set up a quick test pool like this: truncate -s 64M /root/test.img zpool create test /root/test.img zpool scrub test When you've finished testing, just destroy the pool.