Jump to content

Leaderboard


Popular Content

Showing content with the highest reputation since 02/21/17 in all areas

  1. 17 points
    I was wanting to do GPU Hardware Acceleration with a Plex Docker but unRAID doesn't appear to have the drivers for the GPUs loaded. would be nice to have the option to install the drivers so the dockers could use them.
  2. 13 points
    Hi guys, this is a simple plugin that allows users to clear their disks before add them to the array. The main characteristics of this plugin are: Modularity: can be used standalone or in conjunction with Joe L. or bjp999 scripts; Ease of use: with a few clicks you can start a clear session on your disk; Integration: you can always access the plugin under Tools > Preclear Disk menu. If you have Unassigned Devices installed, you can start/stop/view preclear sessions directly from Main > Unassigned Devices. All dependencies included: you don't need SCREEN to run a preclear session in the background; all jobs are executed in the background by default, so you can close your browser while the preclear runs. You can install it directly or via Community Apps. Q & A: Q) Why Joe L. or bjp999 scripts are not included? A) I'm not authorized by Joe L. to redistribute his script, so you need to download a copy from the topic above and put it under /boot/config/plugins/preclear.disk/ renaming it to preclear_disk.sh if necessary. bjp999 modifications are unofficial, so I decided not to include it by default. Q) By default, I see a "gfjardim" script always available. Why? A) Since I'm not authorized to redistribute Joe L. script and the recent slow support by the author, I decided a major code rewrite in the script was needed. The new script is being actively supported, compatible with unRAID notifications, is faster than bjp999 script and has a cleaner output so users can easily visualizes what's going on with their disks. Q) I want to use one of the older scripts(Joe L. or bjp999) in conjunction with notifications. Is that possible? A) Yes. I've made some adjustments on both scripts so they become compatible with unRAID notifications; Joe L. version can be found here and bjp999 can be found here. Q) Is there any howtos available? A) gridrunner made a awesome video explaining why preclearig a hard disk is a good idea, and how you can accomplish that: Q) The plugin asked me to send some statistics information. How does the statistics report system work? Is it safe? Is it anonymous? A) To better track the usage of the plugin, a statistics report system was put in place. The main goals I intend to archive are: know number of disks that gets precleared; fix any silent bugs that gets reported on the logs; know average size of disks, their model, their average speed and elapsed time we should expect from that model; success rate; rate of disks with SMART problems; This system is totally optional and users will get prompted if they want to send each report. It is also safe and totally anonymous, since all data is sent to Google Forms and no identifying data is exported, like disks serial numbers. Detailed info can be found here. The statistics are public and can be found here. Q) How can I download a copy of the plugin log? A) Please go to Tools, then Pleclear Disk, and click on the Download icon: Q) Which are the differences between Erase and Clear? A) The Clear option uses zeroes to fill the drive; at the end, the drive can be added to the array the array immediately. The Erase All the Disk option uses random data to wipe out the drive; the resulting drive can't be quickly added to the array. If you want to add if after erase, you must select Erase and Clear the Disk.
  3. 13 points
    Hi, Guys. This is the first part of a two-part video about setting up a Windows 10 KVM VM in unRAID. (second part in a day or 2 if work lets me !) The first part deals with setting up the VM correctly to be able to use as a 'daily driver'. Then the second part passing through hardware to turn it into a gaming VM. The first part consists of Download a windows 10 iso. Where to Buy a license for windows 10 pro for $20 How to assign resources and correctly pin you CPUs. How to install the virtio drivers including the qxl graphics driver. How to remove or block the windows 10 data mining - phone home - etc with anti beacon. How to install multiple useful programmes with ninite Using Splashtop desktop for good quality remote viewing How to install a virtual sound card to have sound in Splashtop/RDP etc. Using mapped drives and symlinks to get the most out of the array. Windows tweaks for VM compatibility. general tips Hope you find it useful The best way to install and setup a windows 10 vm as a daily driver or a Gaming VM Below is the T second part of a two-part video about setting up a Windows 10 KVM VM in unRAID. The second part deals with passing through hardware and potential problems and solutions showing you how to turn it into a gaming VM. Hope you find it useful.
  4. 12 points
    Currently unRAID uses basic auth to enter credentials for the web gui, but many password managers don't support this. Would be great if we could get a proper login page. Examples This kind of login page always works with password managers. This does not
  5. 12 points
    I would love to be able to view the cpu thread pairs from the vm templates like this
  6. 10 points
    Community Applications This thread is rather long (and is mostly all off-topic), and it is NOT necessary to read it in order to utilize Community Applications (CA) Just install the plugin, go to the apps tab and enjoy the freedom. If you find an issue with CA, then don't bother searching for answers in this thread as all issues (when they have surfaced) are fixed generally the same day that they are found... (But at least read the preceding post or two on the last page of the thread) - This is without question, the best supported plugin / addon in the universe - on any platform. Simple interface and easy to use, you will be able to find and install any of the unRaid docker or plugin applications, and also optionally gain access to the entire library of applications available on dockerHub (~1.8 million) INSTALLATION To install this plugin, paste the following URL into the Plugins / Install PlugIn section: https://raw.githubusercontent.com/Squidly271/community.applications/master/plugins/community.applications.plg After installation, a new tab called "Apps" will appear on your unRaid webGUI. To see what the various icons do, simply press Help or the (?) on unRaid's Tab Bar. Note All screenshots in this post are subject to change as Community Applications continues to evolve Easily search or browse applications Get full details on the application Easily reinstall previously installed applications Able to optionally search for any container from dockerHub And much, much more USING CA CA also has a dedicated Settings section (click CA Settings) which will let you fine tune certain aspects of its operation. Buy Andrew A Beer!
  7. 10 points
    How to setup Dockers to have own IP address without sharing the host IP address: This is only valid in unRAID 6.3 series going forward. 6.4.0 has this built into the GUI but currently have a limitation of needing extra IP addresses for each of your interfaces and needs to delete all manually created docker networks. If you don't like that, you can opt to create the network manually like here and disable the docker network auto cleanup. 6.4.1 is now more intelligent about this and presents a relatively powerful UI that covers most simple cases. You don't need to assign unnecessary IP address to additional interfaces anymore. refer to https://lime-technology.com/forums/topic/62107-network-isolation-in-unraid-64/ for more details Single NIC only: Some assumptions: We'll be using a shared interface br0 (This allows us to use the same nic with virtual machines, otherwise its alright to use eth0) The IP address details are: unRAID = 192.168.1.2 Gateway/router = 192.168.1.1 Subnet = 192.168.1.0/24 Docker IP pool = 192.168.1.128/25 (192.168.1.128-254) A new docker network will be established called homenet Login via SSH and execute this: # docker network create \ -o parent=br0 \ --driver macvlan \ --subnet 192.168.1.0/24 \ --ip-range 192.168.1.128/25 \ --gateway 192.168.1.1 \ homenet Modify any Docker via the WebUI in Advanced mode Set Network to None Remove any port mappings Fill in the Extra Parameters with: --network homenet Apply and start the docker The docker is assigned an IP from the pool 192.168.1.128 - 192.168.1.254; typically the first docker gets the first IP address # docker inspect container | grep IPAddress "SecondaryIPAddresses": null, "IPAddress": "", "IPAddress": "192.168.1.128", # docker exec container ping www.google.com PING www.google.com (122.2.129.167): 56 data bytes 64 bytes from 122.2.129.167: seq=0 ttl=57 time=36.842 ms 64 bytes from 122.2.129.167: seq=1 ttl=57 time=36.496 ms ^C # docker exec container ping 192.168.1.2 PING 192.168.1.2 (192.168.1.2): 56 data bytes ^C # At this point, your gateway/router will have a first class network citizen with the specified IP address An additional Extra Parameter can be specified to fix the IP address: --ip 192.168.1.128 The container will not be allowed to talk to unRAID host due to the underlying security implementation with the macvlan driver used by Docker. This is by design That's it. Secondary NIC is available: Some assumptions: We'll be using a dedicated interface br1 (the native eth1 interface can used here too) There is no IP address assigned to the interface The IP address details are: Gateway/router = 10.0.3.1 Subnet = 10.0.3.0/24 Docker IP pool = 10.0.3.128/25 (10.0.3.128-254) A new docker network will be established called docker1 unRAID has an ip of 10.0.3.2 Login via SSH and execute this: # docker network create \ -o parent=br1 \ --driver macvlan \ --subnet 10.0.3.0/24 \ --ip-range 10.0.3.128/25 \ --gateway 10.0.3.1 \ docker1 Modify any Docker via the WebUI in Advanced mode Set Network to None Remove any port mappings Fill in the Extra Parameters with: --network docker1 Apply and start the docker The docker is assigned an IP from the pool 10.0.3.128 - 10.0.3.254; typically the first docker gets the first IP address # docker inspect container | grep IPAddress "SecondaryIPAddresses": null, "IPAddress": "", "IPAddress": "10.0.3.128", # docker exec container ping www.google.com PING www.google.com (122.2.129.167): 56 data bytes 64 bytes from 122.2.129.167: seq=0 ttl=57 time=36.842 ms 64 bytes from 122.2.129.167: seq=1 ttl=57 time=36.496 ms ^C # docker exec container ping 10.0.3.2 PING 10.0.3.2 (10.0.3.2): 56 data bytes 64 bytes from 10.0.3.2: seq=0 ttl=64 time=0.102 ms 64 bytes from 10.0.3.2: seq=1 ttl=64 time=0.075 ms 64 bytes from 10.0.3.2: seq=2 ttl=64 time=0.065 ms 64 bytes from 10.0.3.2: seq=3 ttl=64 time=0.069 ms ^C --- 10.0.3.2 ping statistics --- 4 packets transmitted, 4 packets received, 0% packet loss round-trip min/avg/max = 0.065/0.077/0.102 ms At this point, your gateway/router will have a first class network citizen with the specified IP address An additional Extra Parameter can be specified to fix the IP address: --ip 10.0.3.128 The container can happily talk to unRAID as the packets go out via br1 and talk to the host on br0 That's it. Some caveats: With only a single NIC, and no VLAN support on your network, it is impossible for the host unRAID to talk to the containers and vice versa; the macvlan driver specifically prohibits this. This situation prevents a reverse proxy docker from proxying unRAID, but will work with all other containers on the new docker network. We can only have one network defined per gateway. So if you already have docker network br0 on your main LAN (gateway 192.168.1.1), docker will not allow you to create another network referencing the same gateway. I cannot confirm yet what happens in the case of two or more NICs bridged/bonded together (but it should be the same as a single NIC) unRAID 6.4.0 We need to disable the docker network auto generation and cleanup if we want these settings to remain (until the auto cleanup is made more intelligent). The code below should be inserted into the go file before /usr/local/sbin/emhttp is started. The code below will disable docker network auto creation too, so it will behave like 6.3.5. It seems that the docker page will happily show any docker network you have defined, so you can just use them as normal and no longer need to set Extra Parameters. Again not needed with 6.4.1 going forward. # stop docker network creation and pruning sed -i -e '/# cleanup/,+3s/^/## /;s/^ custom_networks/## custom_networks/' /etc/rc.d/rc.docker
  8. 9 points
    Please report only issues/bugs which are new in the prerelease. We've been holding back on this release a few days waiting for updated Intel ixgbe (10Gbit) OOT driver which builds correctly with 4.18 kernel. But no update yet so this release uses 4.17.19 kernel. Every major subsystem has been updated. In addition @bonienl has added some really cool features and update the default themes to match our new branding. Release notes: Version 6.6.0-rc1 2018-08-31 Base distro: aaa_base: version 14.2 (rev 5) aaa_elflibs: version 15.0 (rev 2) acl: version 2.2.53 acpid: version 2.0.29 adwaita-icon-theme: version 3.28.0 appres: version 1.0.5 (rev 2) at-spi2-atk: version 2.26.2 at-spi2-core: version 2.28.0 at: version 3.1.20 (rev 5) atk: version 2.28.1 attr: version 2.4.48 avahi: version 0.7 (rev2) bash: version 4.4.023 bin: version 11.1 (rev 2) bridge-utils: version 1.5 (rev 2) btrfs-progs: version v4.17 bzip2: version 1.0.6 (rev 3) ca-certificates: version 20180409 cairo: version 1.15.12 celt051: version 0.5.1.3 (rev 2) cifs-utils: version 6.7 (rev 2) coreutils: version 8.29 (rev 2) cpio: version 2.12 (rev 2) cpufrequtils: version 08 (rev 2) cryptsetup: version 1.7.5 (rev 2) curl: version 7.61.0 (CVE-2018-1000300, CVE-2018-1000301, CVE-2018-0500) cyrus-sasl: version 2.1.27_rc8 db48: version 4.8.30 (rev 4) dbus: version 1.12.8 dcron: version 4.5 dejavu-fonts-ttf: version 2.37 (rev 4) dhcpcd: version 7.0.6 diffutils: version 3.6 (rev 2) dmidecode: version 3.1 (rev 2) dnsmasq: version 2.79 (rev 2) docker: version 18.06.1-ce e2fsprogs: version 1.44.2 ebtables: version 2.0.10 (rev 2) editres: version 1.0.7 (rev 2) elvis: version 2.2_0 (rev 4) encodings: version 1.0.4 (rev 2) etc: version 15.0 (rev 6) ethtool: version 4.17 eudev: version 3.2.5 (rev2) file: version 5.33 findutils: version 4.4.2 (rev 2) flex: version 2.6.4 (rev 3) floppy: version 5.5 (rev 2) fontconfig: version 2.12.6 (rev 2) freetype: version 2.9 (rev 2) fribidi: version 1.0.4 fuse: version 2.9.7 (rev3) gawk: version 4.2.1 (rev 2) gd: version 2.2.5 (rev2) gdbm: version 1.15 gdk-pixbuf2: version 2.36.12 genpower: version 1.0.5 (rev 3) getty-ps: version 2.1.0b (rev 4) glew: version 2.1.0 (rev 2) glib2: version 2.56.1 glibc-solibs: version 2.27 (rev 4) glibc-zoneinfo: version 2018e (rev 3) glibc: version 2.27 (rev 4) glu: version 9.0.0 (rev 2) gmp: version 6.1.2 (rev 2) gnome-themes-standard: version 3.22.3 (rev 2) gnupg: version 1.4.23 (CVE-2018-12020) gnutls: version 3.6.2 (rev 2) gptfdisk: version 1.0.3 (rev 2) grep: version 3.1 (rev 2) gtk+3: version 3.22.30 gzip: version 1.9 (rev 2) harfbuzz: version 1.8.1 haveged: version 1.9.2 hdparm: version 9.56 hicolor-icon-theme: version 0.17 (rev 2) hostname: version 3.18 (rev 2) htop: version 2.2.0 icu4c: version 61.1 imlib2: version 1.5.1 inetd: version 1.79s (rev 11) infozip: version 6.0 (rev 4) inotify-tools: version 3.14 (rev 2) intel-microcode: version 20180807 iproute2: version 4.17.0 iptables: version 1.6.2 (rev 2) iputils: version s20140519 (rev 2) kernel-firmware: version 20180814_f1b95fe keyutils: version 1.5.10 (rev 2) kmod: version 25 (rev 2) lbzip2: version 2.5 less: version 530 (rev 3) libaio: version 0.3.111 libarchive: version 3.3.2 (rev 3) libcap-ng: version 0.7.8 (rev 3) libcgroup: version 0.41 (rev 5) libcroco: version 0.6.12 (rev 2) libdaemon: version 0.14 (rev2) libdmx: version 1.1.4 libdrm: version 2.4.92 libedit: version 20180525_3.1 libee: version 0.4.1 (rev 2) libepoxy: version 1.4.3 (rev 2) libestr: version 0.1.10 (rev 2) libevdev: version 1.5.9 (rev 2) libevent: version 2.1.8 (rev 3) libfastjson: version 0.99.8 (rev2) libffi: version 3.2.1 (rev 2) libfontenc: version 1.1.3 (rev 2) libgcrypt: version 1.8.3 (CVE-2018-0495) libgpg-error: version 1.31 libgudev: version 232 (rev 2) libICE: version 1.0.9 (rev 3) libidn: version 1.35 libjpeg-turbo: version 1.5.3 (rev 2) liblogging: version 1.0.6 (rev2) libmnl: version 1.0.4 (rev 3) libnetfilter_conntrack: version 1.0.7 libnfnetlink: version 1.0.1 (rev 2) libnftnl: version 1.1.0 libnl3: version 3.4.0 (rev 2) libpcap: version 1.8.1 (rev 2) libpciaccess: version 0.14 (rev 2) libpng: version 1.6.34 (rev 2) libpthread-stubs: version 0.4 (rev 2) librsvg: version 2.42.5 libseccomp: version 2.3.3 (rev2) libSM: version 1.2.2 (rev 3) libssh2: version 1.8.0 (rev 3) libtasn1: version 4.13 (rev 2) libtirpc: version 1.0.3 (rev 2) libunistring: version 0.9.10 libunwind: version 1.2.1 libusb-compat: version 0.1.5 (rev2) libusb: version 1.0.22 libvirt-php: version 0.5.4 (rev3) libvirt: version 4.6.0 libwebsockets: version 2.4.2 libX11: version 1.6.5 (rev 2) libx86: version 1.1 (rev 3) libXau: version 1.0.8 (rev 3) libXaw: version 1.0.13 (rev 2) libxcb: version 1.13 (rev 2) libXcomposite: version 0.4.4 (rev 3) libXcursor: version 1.1.15 (rev 2) libXdamage: version 1.1.4 (rev 3) libXdmcp: version 1.1.2 (rev 3) libXevie: version 1.0.3 (rev 3) libXext: version 1.3.3 (rev 3) libXfixes: version 5.0.3 (rev 2) libXfont2: version 2.0.3 (rev 2) libXfontcache: version 1.0.5 (rev 3) libXft: version 2.3.2 (rev 4) libXi: version 1.7.9 (rev 2) libXinerama: version 1.1.3 (rev 3) libxkbfile: version 1.0.9 (rev 2) libxml2: version 2.9.8 (rev 2) libXmu: version 1.1.2 (rev 3) libXpm: version 3.5.12 (rev 2) libXrandr: version 1.5.1 (rev 2) libXrender: version 0.9.10 (rev 2) libXres: version 1.2.0 (rev 2) libxshmfence: version 1.3 (rev 2) libxslt: version 1.1.32 (rev 2) libXt: version 1.1.5 (rev 2) libXtst: version 1.2.3 (rev 2) libXxf86dga: version 1.1.4 (rev 3) libXxf86misc: version 1.0.3 (rev 3) libXxf86vm: version 1.1.4 (rev 3) listres: version 1.0.4 (rev 2) lm_sensors: version 3.4.0 (rev 2) logrotate: version 3.14.0 (rev 2) lsof: version 4.91 lsscsi: version 0.29 lvm2: version 2.02.177 lz4: version 1.8.2 mc: version 4.8.21 mesa: version 18.1.2 mkfontdir: version 1.0.7 (rev 2) mkfontscale: version 1.1.3 (rev 2) mozilla-firefox: version 61.0.2 (CVE-2018-12359, CVE-2018-12360, CVE-2018-12361, CVE-2018-12358, CVE-2018-12362, CVE-2018-5156, CVE-2018-12363, CVE-2018-12364, CVE-2018-12365, CVE-2018-12371, CVE-2018-12366, CVE-2018-12367, CVE-2018-12368, CVE-2018-12369, CVE-2018-12370, CVE-2018-5186, CVE-2018-5187, CVE-2018-5188) mpfr: version 4.0.1 (rev 2) mtdev: version 1.1.5 (rev 2) nano: version 2.9.8 ncompress: version 4.2.4.4 (rev 2) ncurses: version 6.1_20180616 net-tools: version 20170208_479bb4a (rev 2) netatalk: version 3.1.11 (rev2) nettle: version 3.4 (rev 2) network-scripts: version 15.0 nghttp2: version 1.32.0 (CVE-2018-1000168) nginx: version 1.14.0 ntfs-3g: version 2017.3.23 (rev 2) ntp: version 4.2.8p12 (CVE-2016-1549, CVE-2018-12327) numactl: version 2.0.11 (rev2) openldap-client: version 2.4.46 openssh: version 7.7p1 openssl-solibs: version 1.1.0i openssl10-solibs: version 1.0.2o openssl: version 1.1.0i (CVE-2018-0732, CVE-2018-0737) p11-kit: version 0.23.12 pango: version 1.42.1 patch: version 2.7.6 (rev 3) (CVE-2018-1000156) pciutils: version 3.5.6 (rev 2) pcre: version 8.42 php: version 7.2.8 pixman: version 0.34.0 (rev 2) pkgtools: version 15.0 (rev 20) pm-utils: version 1.4.1 (rev 6) procps-ng: version 3.3.15 (CVE-2018-1124, CVE-2018-1126, CVE-2018-1125, CVE-2018-1123, CVE-2018-1122) qemu: version 3.0.0 reiserfsprogs: version 3.6.27 (rev2) rpcbind: version 0.2.4 (rev 4) rsync: version 3.1.3 (rev 2) rsyslog: version 8.36.0 samba: version 4.8.4 (CVE-2018-1139, CVE-2018-1140, CVE-2018-10858, CVE-2018-10918, CVE-2018-10919) sed: version 4.5 sessreg: version 1.1.1 (rev 2) setxkbmap: version 1.3.1 (rev 2) shadow: version 4.2.1 (rev 4) shared-mime-info: version 1.9 (rev 2) spice: version 0.14.0 (rev2) sqlite: version 3.24.0 ssmtp: version 2.64 (rev5) startup-notification: version 0.12 (rev 3) sudo: version 1.8.23 sysfsutils: version 2.1.0 (rev 2) sysvinit-scripts: version 2.1 (rev 12) sysvinit: version 2.90 talloc: version 2.1.13 tar: version 1.30 (rev 2) tcp_wrappers: version 7.6 (rev 2) tdb: version 1.3.15 (rev 2) telnet: version 0.17 (rev 4) tevent: version 0.9.36 (rev 2) traceroute: version 2.1.0 (rev 2) transset: version 1.0.2 (rev 2) tree: version 1.7.0 (rev 2) ttyd: version 1.4.0 (rev2) usbredir: version 0.7.1 (rev2) usbutils: version 010 utempter: version 1.1.6 (rev 3) util-linux: version 2.32 vala: version 0.28.1 (rev2) vbetool: version 1.2.2 (rev 2) vsftpd: version 3.0.3 (rev 5) vte3: version 0.44.3 (rev2) wget: version 1.19.5 (CVE-2018-0494) which: version 2.21 (rev 2) xauth: version 1.0.10 (rev 2) xcb-util: version 0.4.0 (rev 3) xclock: version 1.0.7 (rev 3) xdpyinfo: version 1.3.2 (rev 2) xdriinfo: version 1.0.6 (rev 2) xev: version 1.2.2 (rev 2) xf86-input-evdev: version 2.10.6 xf86-input-keyboard: version 1.9.0 (rev 3) xf86-input-mouse: version 1.9.3 xf86-input-synaptics: version 1.9.1 xf86-video-ast: version 1.1.5 (rev 5) xf86-video-mga: version 1.6.5 (rev 3) xf86-video-vesa: version 2.4.0 (rev 3) xfsprogs: version 4.16.1 xhost: version 1.0.7 (rev 2) xinit: version 1.4.0 (rev 2) xkbcomp: version 1.4.2 xkbevd: version 1.1.4 (rev 2) xkbutils: version 1.0.4 (rev 3) xkeyboard-config: version 2.22 (rev 2) xkill: version 1.0.5 (rev 2) xload: version 1.1.3 (rev 2) xlsatoms: version 1.1.2 (rev 2) xlsclients: version 1.1.4 (rev 2) xmessage: version 1.0.5 (rev 2) xmodmap: version 1.0.9 (rev 2) xorg-server: version 1.20.0 (rev 2) xprop: version 1.2.3 (rev 2) xrandr: version 1.5.0 (rev 2) xrdb: version 1.1.1 (rev 2) xrefresh: version 1.0.6 (rev 2) xset: version 1.2.4 (rev 2) xsetroot: version 1.1.2 (rev 2) xsm: version 1.0.4 (rev 2) xterm: version 333 xtrans: version 1.3.5 (rev 2) xwd: version 1.0.7 (rev 2) xwininfo: version 1.1.4 (rev 2) xwud: version 1.0.5 (rev 2) xz: version 5.2.4 zlib: version 1.2.11 (rev 2) Linux kernel: version 4.17.19 intel ixgbe: version 5.3.7 intel ixgbevf: version 4.3.5 highpoint r750: version 1.2.11 highpoint rr3740a: version 1.17.0 added per customer request: CONFIG_BLK_DEV_SKD: STEC S1120 Block Driver CONFIG_BNXT: Broadcom NetXtreme-C/E support CONFIG_CHR_DEV_ST: SCSI tape support changed from built-in to module: CONFIG_BLK_DEV_SR: SCSI CDROM support removed CONFIG_SENSORS_IT87: ITE IT87xx and compatibles it87: version 20180709 groeck Linux Driver for ITE LPC chips (https://github.com/groeck/it87) set CRYPTO_DEV_SP_PSP=N otherwise udev deadlocks on threadripper Management: Docker: add optional per-container wait before starting Enable IPv6 for NTP rc.cpufreq: For CPUs using intel_pstate, always use the performance governor. This also provides power savings on Intel processors while avoiding the ramp-up lag present when using the powersave governor (which is the default if ondemand is requested on these machines). restore docker custom networks upon docker start update smartmontools drivedb and hwdata/{pci.ids,usb.ids,oui.txt,manuf.txt} webgui: docker: Correct docker container icon name generation webgui: docker: added cpu load and memory load display webgui: docker: added shell/bash selection for console webgui: docker: Make cpu-load and mem-load only visible in advanced mode webgui: Show accumulated encryption status webgui: Prevent openbox from clipping webgui: Plugins: Show support thread if present webgui: Remove control codes, and extended ascii characters from plugin urls on install webgui: Update link for privileged help text webgui: fix regex matching for unraid.net domain name validation when using new openssl1.1 webgui: docker: Prevent arbitrary bash execution or redirection on docker create/run commands webgui: Plugins: Preserve support link when updating plugin webgui: diagnostics: Replace consecutive repeated lines in syslog webgui: Remove empty entries when storing individual disk settings webgui: docker: fix connect container console to new window webgui: Use timeout in command execution of diagnostics webgui: Fixed 'Update All Containers' button sometimes hidden in docker list webgui: Added version control in docker API calls webgui: Show docker allocations in column format webgui: fix css help button in themes azure and gray webgui: Added docker autostart wait period webgui: Added CPU selection to docker container edit section webgui: Verify internet access using NCSI method NCSI = network connection status indicator. This method tries to access a specific Microsoft site to test internet access. The same method is used in Windows. webgui: Added docker autostart wait period webgui: New CPU pinning functionality webgui: Change wording on dashboard page "memory total usable" webgui: Include meminfo in diagnostics webgui: Docker containers page: show container starting up message Starting docker service and auto-starting containers has been decoupled. This lets the docker page return as soon as the service is ready. The container overview page shows a new message, and will automatically reload to update the status until all containers are auto-started webgui: VMs: preserve XML custom settings in VM config updates webgui: dockerMan: Avoid filename collisions on FAT32 webgui: Extract disk log from multiple rotated logs webgui: theme-match marketing site
  9. 9 points
    I just wanted to say a big thanks to all the hard work that you guys put into unRAID. Your commitment to constantly improving the product is truely excellent. We get constant updates and new features and we only buy the license once. It really is amazing how you guys do it. Please anyone reading this who feels the same post beneath and lets let the guys know how much we appreciate what they do. I started my journey with computers with an Atari 800xl and absolutely loved it and wasted many hours. Then later had a career in IT and it became a bit of a boring thing I had to do. Having found unRAID I feel like I was when I was a boy and it's just as exciting as my old 800XL was. So unRAID has put the fun back into computers for me. Well, that or I am just in my second childhood already!
  10. 9 points
    With all due respect man, this is unwarranted. We take security very seriously. Case in point: totally dropped development last month to incorporate CSRF protection as fast as possible and it was a hell of a lot of work. We are team of 2 developers in unRAID OS core, and one of us can only spend half time at it because of other hats that must be worn. Reality is 99% of CVE's do not affect unRAID directly. Many are in components we don't use. Many apply only to internet-facing servers. We have always advised, unRAID OS should not be directly attached to the internet. The day is coming when we can lift that caveat but for now VM's can certainly serve that role if you know what you are doing. If you find a truly egregious security vulnerability in unRAID we would certainly appreciate an email. We see every one of those, whereas we don't read every single forum post. Send to tomm@lime-technology.com
  11. 9 points
    I had the opportunity to test the “real word” bandwidth of some commonly used controllers in the community, so I’m posting my results in the hopes that it may help some users choose a controller and others understand what may be limiting their parity check/sync speed. Note that these tests are only relevant for those operations, normal read/writes to the array are usually limited by hard disk or network speed. Next to each controller is its maximum theoretical throughput and my results depending on the number of disks connected, result is observed parity check speed using a fast SSD only array with Unraid V6.1.2 (SASLP and SAS2LP tested with V6.1.4 due to performance gains compared with earlier releases) Values in green are the measured controller power consumption with all ports in use. 2 Port Controllers SIL 3132 PCIe gen1 x1 (250MB/s) 1 x 125MB/s 2 x 80MB/s Asmedia ASM1061 PCIe gen2 x1 (500MB/s) - e.g., SYBA SY-PEX40039 and other similar cards 1 x 375MB/s 2 x 206MB/s 4 Port Controllers SIL 3114 PCI (133MB/s) 1 x 105MB/s 2 x 63.5MB/s 3 x 42.5MB/s 4 x 32MB/s Adaptec AAR-1430SA PCIe gen1 x4 (1000MB/s) 4 x 210MB/s Marvell 9215 PCIe gen2 x1 (500MB/s) - 2w - e.g., SYBA SI-PEX40064 and other similar cards (possible issues with virtualization) 2 x 200MB/s 3 x 140MB/s 4 x 100MB/s Marvell 9230 PCIe gen2 x2 (1000MB/s) - 2w - e.g., SYBA SI-PEX40057 and other similar cards (possible issues with virtualization) 2 x 375MB/s 3 x 255MB/s 4 x 204MB/s 8 Port Controllers Supermicro AOC-SAT2-MV8 PCI-X (1067MB/s) 4 x 220MB/s (167MB/s*) 5 x 177.5MB/s (135MB/s*) 6 x 147.5MB/s (115MB/s*) 7 x 127MB/s (97MB/s*) 8 x 112MB/s (84MB/s*) *on PCI-X 100Mhz slot (800MB/S) Supermicro AOC-SASLP-MV8 PCIe gen1 x4 (1000MB/s) - 6w 4 x 140MB/s 5 x 117MB/s 6 x 105MB/s 7 x 90MB/s 8 x 80MB/s Supermicro AOC-SAS2LP-MV8 PCIe gen2 x8 (4000MB/s) - 6w 4 x 340MB/s 6 x 345MB/s 8 x 320MB/s (205MB/s*, 200MB/s**) *on PCIe gen2 x4 (2000MB/s) **on PCIe gen1 x8 (2000MB/s) Dell H310 PCIe gen2 x8 (4000MB/s) - 6w – LSI 2008 chipset, results should be the same as IBM M1015 and other similar cards 4 x 455MB/s 6 x 377.5MB/s 8 x 320MB/s (190MB/s*, 185MB/s**) *on PCIe gen2 x4 (2000MB/s) **on PCIe gen1 x8 (2000MB/s) LSI 9207-8i PCIe gen3 x8 (4800MB/s) - 9w - LSI 2308 chipset 8 x 440MB/s+ (*) * SSDs used max speed SAS Expanders HP 6Gb (3Gb SATA) SAS Expander - 11w Single Link on Dell H310 (1200MB/s*) 8 x 137.5MB/s 12 x 92.5MB/s 16 x 70MB/s 20 x 55MB/s 24 x 47.5MB/s Dual Link on Dell H310 (2400MB/s*) 12 x 182.5MB/s 16 x 140MB/s 20 x 110MB/s 24 x 95MB/s * Half 6GB bandwidth because it only links @ 3Gb with SATA disks Intel® RAID SAS Expander RES2SV240 - 10w Single Link on Dell H310 (2400MB/s) 8 x 275MB/s 12 x 185MB/s 16 x 140MB/s (112MB/s*) 20 x 110MB/s (92MB/s*) Dual Link on Dell H310 (4000MB/s) 12 x 205MB/s 16 x 155MB/s (185MB/s**) Dual Link on LSI 9207-8i (4800MB/s) 16 x 275MB/s * Avoid using slower linking speed disks with expanders, as it will bring total speed down, in this example 4 of the SSDs were SATA2, instead of all SATA3. ** Two different boards have consistent different results, will need to test a third one to see what's normal, 155MB/s is the max on a Supermicro X9SCM-F, 185MB/s on Asrock B150M-Pro4S. Sata 2 vs Sata 3 I see many times on the forum users asking if changing to Sata 3 controllers or disks would improve their speed, Sata 2 has enough bandwidth (between 265 and 275MB/s according to my tests) for the fastest disks currently on the market, if buying a new board or controller you should buy sata 3 for the future, but except for SSD use there’s no gain in changing your Sata 2 setup to Sata 3. Single vs. Dual Channel RAM In arrays with many disks, and especially with low “horsepower” CPUs, memory bandwidth can also have a big effect on parity check speed, obviously this will only make a difference if you’re not hitting a controller bottleneck, two examples with 24 drive arrays: Asus A88X-M PLUS with AMD A4-6300 dual core @ 3.7Ghz Single Channel – 99.1MB/s Dual Channel - 132.9MB/s Supermicro X9SCL-F with Intel G1620 dual core @ 2.7Ghz Single Channel – 131.8MB/s Dual Channel – 184.0MB/s DMI There is another bus that can be a bottleneck for Intel based boards, much more so than Sata 2, the DMI that connects the south bridge or PCH to the CPU. Socket 775, 1156 and 1366 use DMI 1.0, socket 1155, 1150 and 2011 use DMI 2.0, socket 1151 uses DMI 3.0 DMI 1.0 (1000MB/s) 4 x 180MB/s 5 x 140MB/s 6 x 120MB/s 8 x 100MB/s 10 x 85MB/s DMI 2.0 (2000MB/s) 4 x 270MB/s (Sata2 limit) 6 x 240MB/s 8 x 195MB/s 9 x 170MB/s 10 x 145MB/s 12 x 115MB/s 14 x 110MB/s DMI 3.0 (3940MB/s) 6 x 330MB/s (Onboard SATA only*) 10 X 297.5MB/s 12 x 250MB/s 16 X 185MB/s *Despite being DMI 3.0, Skylake, Kaby Lake and Coffee Lake chipsets have a max combined bandwidth of approximately 2GB/s for the onboard SATA ports. DMI 1.0 can be a bottleneck using only the onboard Sata ports, DMI 2.0 can limit users with all onboard ports used plus an additional controller onboard or on a PCIe slot that shares the DMI bus, in most home market boards only the graphics slot connects directly to CPU, all other slots go through the DMI (more top of the line boards, usually with SLI support, have at least 2 slots), server boards usually have 2 or 3 slots connected directly to the CPU, you should always use these slots first. You can see below the diagram for my X9SCL-F test server board, for the DMI 2.0 tests I used the 6 onboard ports plus one Adaptec 1430SA on PCIe slot 4. UMI (2000MB/s) - Used on most AMD APUs, equivalent to intel DMI 2.0 6 x 203MB/s 7 x 173MB/s 8 x 152MB/s Ryzen link - PCIe 3.0 x4 (3940MB/s) 6 x 467MB/s (Onboard SATA only) I think there are no big surprises and most results make sense and are in line with what I expected, exception maybe for the SASLP that should have the same bandwidth of the Adaptec 1430SA and is clearly slower, can limit a parity check with only 4 disks. I expect some variations in the results from other users due to different hardware and/or tunnable settings, but would be surprised if there are big differences, reply here if you can get a significant better speed with a specific controller. How to check and improve your parity check speed System Stats from Dynamix V6 Plugins is usually an easy way to find out if a parity check is bus limited, after the check finishes look at the storage graph, on an unlimited system it should start at a higher speed and gradually slow down as it goes to the disks slower inner tracks, on a limited system the graph will be flat at the beginning or totally flat for a worst-case scenario. See screenshots below for examples (arrays with mixed disk sizes will have speed jumps at the end of each one, but principle is the same). If you are not bus limited but still find your speed low, there’s a couple things worth trying: Diskspeed - your parity check speed can’t be faster than your slowest disk, a big advantage of Unraid is the possibility to mix different size disks, but this can lead to have an assortment of disk models and sizes, use this to find your slowest disks and when it’s time to upgrade replace these first. Tunables Tester - on some systems can increase the average speed 10 to 20Mb/s or more, on others makes little or no difference. That’s all I can think of, all suggestions welcome.
  12. 9 points
    Can I manually create and use multiple btrfs pools? Work in progress. Multiple cache pools are not supported yet, though they are planned for the future, until then you can still use multiple btrfs pools with the help of the Unassigned Devices plugin. There are some limitations and most operations creating and maintaining the pool will need to be made using the command line, so if you're not comfortable with that wait for LT to add the feature. If you want to use the now, here's how: -If you don't have yet install the Unassigned Devices plugin. -Better to start with clean/wiped devices, so wipe them or delete any existing partitions. -Using UD format the 1st device using btrfs, choose the mount point name and optionally activate auto mount and share -Using UD mount the 1st device, for this example it will be mounted at /mnt/disks/yourpoolpath -Using UD format the 2nd device using btrfs, no need to change the mount point name, and leave auto mount and share disable. -Now on the console/SSH add the device to the pool by typing: btrfs dev add -f /dev/sdX1 /mnt/disks/yourpoolpath Replace X with correct identifier, note the 1 in the end. -Device will be added and you will see the extra space on the 1st disk free space graph, whole pool will be accessible in the original mount point, in this example: /mnt/disks/yourpoolpath -By default the disk is added in single profile mode, i.e., it will extend the existing volume, you can change that to other profiles, like raid0, raid1, etc, e.g., to change to raid1 type: btrfs balance start -dconvert=raid1 -mconvert=raid1 /mnt/disks/yourpoolpath See here for the other available modes. -If you want to add more devices to that pool just repeat the process above Notes -Only mount the first device with UD, all other members will mount together despite nothing being shown on UD's GUI, same to unmount, just unmount the 1st device to unmount the pool. -It appears that if you mount the pool using the 1st device used/free space are correctly reported by UD, unlike if you mount using e.g. the 2nd device, still for some configurations the space might be incorrectly reported, you can always check it using the command line: btrfs fi usage /mnt/disks/yourpoolpath -You can have as many unassigned pools as you want, example how it looks on UD: sdb+sdc+sdd+sde are part of a raid5 pool, sdf+sdg are part of raid1 pool, sdh+sdi+sdn+sdo+sdp are another raid5 pool, note that UD sorts the devices by identifier (sdX), so if sdp was part of the first pool it would still appear last, UD doesn't reorder the devices based on if they are part of a specific pool. You can also see some of the limitations, i.e., no temperature is shown for the secondary pools members, though you can see temps for all devices on the dashboard page, still it allows to easily use multiple pools until LT adds multiple cache pools to Unraid. Remove a device: -to remove a device from a pool type (assuming there's enough free space): btrfs dev del /dev/sdX1 /mnt/disks/yourpoolpath Replace X with correct identifier, note the 1 in the end Note that you can't go below the used profile minimum number of devices, i.e., you can't remove a device from a 2 device raid1 pool, you can convert it to single profile first and then remove the device, to convert to single use: btrfs balance start -f -dconvert=single -mconvert=single /mnt/disks/yourpoolpath Then remove the device normally like above. Replace a device: To replace a device from a pool (if you have enough ports to have both old and new devices connected simultaneously): You need to partition the new device, to do that format it using the UD plugin, you can use any filesystem, then type: btrfs replace start -f /dev/sdX1 /dev/sdY1 /mnt/disks/yourpoolpath Replace X with source, Y with target, note the 1 in the end of both, you can check replacement progress with: btrfs replace status /mnt/disks/yourpoolpath If the new device is larger you need to resize it to use all available capacity, you can do that with: btrfs fi resize X:max /mnt/disks/yourpoolpath Replace X with the correct devid, you can find that with: btrfs fi show /mnt/disks/yourpoolpath
  13. 9 points
    After successfully bricking the Fujitsu D2607 by downflashing it I'm proud to be able to contribute to this thread and hereby report: LSI MegaRAID with SAS2008 chipsets 3) DELL Perc H310 as well as H200 Flashed successfully to LSI9211-8i IT (P20) 3TB Drive Support with this card: YES (UPDATE: 5.0Beta7 added 3TB Drive support) Drive Spin Down support: YES (UPDATE: Added as of 5.0Beta7) Drive Temp Readings: YES Toolset_PercH310 to LSIMegaraid.zip (DOS, via bootable usb key) http://www45.zippyshare.com/v/51016808/file.html (for some reason I can't embed the link...) MD5:80174075959fb7d1ff8c6362f7241bfe Update on 06.08.2014 Included the P19 firmware http://www21.zippyshare.com/v/9541812/file.html Update on 01.12.2014 Possible issues with P20 firmware! See this post and this. Update on 23.10.2015 There is an new version of Avago (former LSI) P20 (20.00.04.00) which seems to be OK with unRAID. See this post. Update on 15.09.2015 User opentoe found out that the DELL IT firmware is also working with unRAID. It's your decision what to flash. Flashing DELL firmware is easier and supported by DELL! opentoe's verdict on DELL IT or Avago (former LSI). Update on 07.06.2016 There is a new firmware from Avago. P20.00.07.00 The toolset has been updated accordingly. First impressions. http://www3.zippyshare.com/v/xZKIOHaz/file.html https://www.mediafire.com/?8f82hx4c032a929 MD5: 24f7d428292e00f9dee6f715d8564202 Update on 30.12.2016 Firmware is still P20.00.07.00 Switch to RUFUS for bootdisk creation. Added alternative ways to extract controller info if MegaCli is not working. https://www.mediafire.com/?9cbklh4i1002n23 MD5: 7d90f84c831e8b939c5536d9eb03ba81 Update on 23.02.2017 Firmware is still P20.00.07.00 Uses sas2flsh through the whole process. Tested on a "backflashed" H200, to be confirmed on a stock H200 card and on H310's. Card backup is now dumping the full flash. This can be used to restore the initial condition of the card. Added script for automatic SAS address extraction. No reboot necessary any more. https://www.mediafire.com/?0op114fpim9xwwf MD5: 2fbe3d562846e493714a9e8ac3f15923 Due to missing UEFI environment, no changes nor testing with UEFI shell. Update on 30.03.2017, v2 Firmware is still P20.00.07.00 Spiced up the routines with some checks to automatically select the right tool if one is not working. Tested on a stock H310 as well as a H200 - works for me. Post your experience in the forum. https://www.mediafire.com/?6b77v4s7czluvs2 MD5: 6cb92336ff537aeb838085a028aa6601 Update on 11.04.2017, v3 Firmware is still P20.00.07.00 Added files for use in an EFI environment. Untested due to missing hardware. Post your experience in the forum. https://www.mediafire.com/?9ovj2rxuaf43wv4 MD5: t.b.d. Update on 17.04.2017, v4 <--- this is the latest, use this one! Firmware is still P20.00.07.00 Corrections for EFI environment. Untested due to missing hardware. Post your experience in the forum. https://www.mediafire.com/?py9c1w5u56xytw2 MD5: t.b.d. If you experience the "failed to initialize PAL" error somewhere in step 5, you have to boot from UEFI shell and try again or use another mainboard. See here how to use UEFI shell (Kudos 2 Maglin). Make sure you read and understand the __READMEFIRST.txt before starting! If you experience troubles or something is not clear, don't hesitate to ask for help. You can help improving the howto by doing so. Chances are small but you can brick the controller!
  14. 8 points
    This plugin is designed to find and offer up suggestions to a multitude of configuration errors, outright problems, etc for numerous aspects of your unRaid server. To install this plugin, just head over to the Apps tab and search for Fix Common Problems. After installation, you will see a lifesaver within Settings / User Utilities which will launch a manual scan (and give you the option to set the background scan settings) For every error or warning that this plugin finds, a suggested course of action will also be displayed. Additionally, should you seek additional help for any error / warning, the errors are all logged into your syslog so that persons helping can easily find the issue when you post your diagnostics. Scans can be scheduled to run automatically in the background (you have the option of hourly, daily, weekly, and monthly). Additionally, if the background scans find an issue they will send out a notification (depending upon your notification settings in this plugin) The current list of tested items will be maintained in the second post of this thread. Any support for problems this plugin finds, should be posted in the General v6 section of these forums. Problems relating to false positives, suggestions for more checks, why I made the decisions I did, wording mistakes in suggestions, etc. should be posted here. As usual for anything written by me, updates are frequent as new ideas pop into my head. Highly recommended to turn on auto-updates for this plugin. Additionally, a special "Troubleshooting Mode" is available to assist with problems involving random crashes / shutdowns / lockups / etc A video with a basic run through of FCP can be found here: (at about 18:25)
  15. 8 points
    I would love a progress bar/percentage thingy for mover
  16. 8 points
    Hi Guys, This Video is a tutorial on how to setup Rclone on unRAID. You will see how to install it then connect to 3 different cloud storage providers. Amazon, dropbox and google drive. You will see how to encrypt and decrypt files in the cloud. You will see how to connect a docker container to rclone and even stream and encrypted media file to emby or plex. You will then see how to make the rclone mount into a network share. Finally, you will see how to easily sync a folder to the cloud. How to setup and use rclone. Copy sync and encrypt files to the cloud. Even stream media
  17. 8 points
    Hi, Guys. This is a series of 3 videos about tuning the unRAID server. It is a guide that is for the server as a whole but has a lot of information for VMs so I thought this forum section the best place to post this. Some of the topics are:- Cpu governor and enabling turbo boost About vCPUs and hyperthreading. How VMs and Docker containers and affect each other performance. Pinning cores to Docker containers. Using the same container with different profiles Allocating resources to Docker containers. Decreasing latency in VMs Using emuatorpin Isolating CPU cores Setting extra profiles in syslinux for isolated cores Checking wether cores have been correctly isolated Disabling hyperthreading Having unRAID manage vCPUs as opposed to vCPU pinning. Hope this video is interesting Part 1 Part 2 Part 3
  18. 8 points
    I took the GUI style efforts of Kode and Drakknar together and blended that in a new theme GRAY. Much obliged to them The new color theme has the 'header' buttons at the left side which expand when hovering over them. The top header is fixed, meaning status information stays on screen when scrolling down the page. All-in-all it gives a complete new fresh view to the GUI, hope you guys like this direction. Next, see if this can be integrated with the upcoming release of unRAID. Below some screen pictures to give you an idea.
  19. 8 points
    Well gents, just wanted to give you a update on Ryzen w/ c-states enabled locking up unRAID... or rather how we've fixed it! On my Ryzen test machine, with array stopped (to keep it idle) with c-states enabled in bios, I'm approaching 7 days of uptime. Before the changes we made to the kernel in the upcoming RC7, It would only make it a few hours and lockup. I think you guys will find RC7 just epyc! (sorry, couldn't resist) FYI, these are the two kernel changes we had to add to make it stable for Ryzen: - CONFIG_RCU_NOCB_CPU: Offload RCU callback processing from boot-selected CPUs - CONFIG_RCU_NOCB_CPU_ALL: All CPUs are build_forced no-CBs CPUs Thanks to everyone here for testing and helping narrow it down to c-states. That made it easier for us to test and find a solution.
  20. 7 points
    Hi, guys. I was frustrated as to why we couldn't use the techpowerup vbios files in our VMs. So I have discovered it is due to a header put there for later Nvidia cards which isn't compatible with KVM. (the header is for the NVflash programme) So in this video, you will see how to pass through a Nvidia GPU as primary in a KVM virtual machine on unRAID by removing the header from a vbios dumped by GPU-z/ Techpowerup making it compatible with a KVM virtual machine. This is very useful as techpowerup.com has an extensive database of GPU vbioses. Also, should you want to dump your own previously you had to have 2 GPUs and 2 PCIe slots to do this. Well now you can do it using GPU-Z easily then convert it for KVM. How to easily passthough a Nvidia GPU as primary without dumping your own vbios!
  21. 7 points
    I just realised this morning that it’s been just over a year since I built my unRAID server. I meant to write this post last year, but never got around to it. I’ve taken some very quick photos – apologies for the quality (and some of the dust on my PC – hasn’t been cleaned for a while…) as I only had my Nexus 6P! Overall I’m amazed by how much unRAID has allowed me to do with this machine, and a year in I still haven’t scratched the surface of what it can do. I’m convinced I’ll be able to, with a few select upgrades along the way, keep this machine going for another 10 years like my last machine. ------------------------------------------------------------ This build has been both exciting and sad at the same time. Sad, because my 1st Build 'Cylon', which was lovingly created as a xmas present to myself way back in 2005, was finally deactivated. Cylon evolved over the years with all of the parts changing except for my trusty Gigabyte 3DAurora chassis , including the CPU from a AMD Athlon 64 X2 4200+ (2005) --> Intel Core 2 Quad Pro Q6600 (2008), i7 920 (2009) to its final i5-3450 (2012). I've built several HTPCs (and a WHS server at one point) along the way that used to be connected to each of my TVs, but they've all gone over the years with only my trusty Cylon staying the course. What drove the new build? Cylon was starting to fall apart and was unable to keep cool and was too noisy. Some of my HDDs weren't even in bays anymore...... I needed more power - my setup couldn't keep up with the Plex streaming needs of my family, with multiple concurrent transcodes locally and remotely I needed more workstations - MY PC seemingly overnight was being shared by my wife and 2 kids, so I was looking for a way to cost effectively support concurrent users I miss building and configuring PCs and I finally had a good excuse for a major new challenge! It's been exciting getting up to speed with the last 10 years of PC development While I was trying to find the best solution, two things sent me down the path that I choose: reading about amazing dual Xeon E5-2670 systems being built, thanks to cheap CPUs flooding eBay stumbling onto LinusTechTips '2 Gaming Rigs, 1 Tower' video and realising that unRAID would let me build one powerful PC capable of running multiple concurrent workstations and media centers, just struck a chord. Adding parity protection to my HDDs was also attractive, as it’s been a real pain in the ass replacing failed drives over the years in terms of working out what files had been lost and replacing them. Part Selection CPU: My original intention was to build a system around dual E5-2670s I had picked up on eBay. But, after struggling to find reasonably priced motherboards due to lack of availability and being forced to buy SSI-EEB cases I didn't like, I decided I needed a more modern system with as many current gen products as possible, so that my next system has a chance of surviving as long as Cylon with future upgrades. I was lucky enough to find a E5-2683V3 on eBay for 1/10th of the cost of a new one - 14 cores Vs the 16 of the 2670s, with lower power consumption for a 24/7 ATX not SSI-EEB machine and support for DDR4, USB3.1, M.2 etc Motherboard: Future-proofing was a major consideration (32Mbps M.2, 32Mbps USB 3.1, Type C, ) and 3+ PCIe x16 slots allowing me to run multiple virtual machines in unRAID Case: Although it's been out 2 years, the R5 is still a great case. It's not ostentatious and just has a quality look about it. It's also very quiet, which was a vital element of the overall build. I didn't want a massive window that showed my drives. I've bought this case to hopefully last as long as the Aurora3D, and it should be flexible enough to handle any scenario in the future GPU: I haven't played PC games since AOE/AOC, so I don't need a powerful card at the moment so the focus again has been on 'silent'. However, I am tempted to get a gaming card for one of my VMs as a later upgrade CPU Cooler: overkill for the Xeon which isn't overclocked, but capable of cooling hotter CPUs in the future. Plus it looks so damn sexy. Oh, and it's quiet PSU: efficiency important for a 24/7 machine and I wanted enough wattage headroom to support one mid-range GPU if necessary in the future. Quiet Lighting: I resisted buying a Hue+ as this isn't a gaming rig, but I couldn't resist adding some cheap lighting Original Build I couldn't be more pleased with the outcome. Everything runs soooo fast compared to Cylon, RIP. Temps are fantastic and the noise levels are acceptable (more later). Working inside the R5 was a dream, with plenty of room to manoeuvre and no need to ask my wife to be an extra pair of hands. Cost wise, I’m very happy with the results as it cost me £1,234 to build THREE (unRAID + 2VMs) powerful machines: £952 old components (monitor, HDD drives, keyboard) £1,234 for new components Grand Total: £2,186 - All powered by one chassis, so electricity and space savings are huge Changes over the year - '2 more PCs', TV Server, Bigger Array Storage, M.2 storage, bigger cache pool - £925 I purchased a 3rd monitor on eBay for £75 and added a GT730 and cheap Logitech keyboard to create a 3rd VM. Now, 3 of my family can use the PC concurrently I added another 32GB of RAM as at times I was getting too close to 100% utilisation I initially bought a 500GB SSD to use as an unassigned drive for my VMs and used the 2nd 250GB SSD I was using previously to create a cache pool. I then gave in and bought a 256GB M.2. SM961 and threw the 500GB SSD into the cache pool, doubling the size Added an extra 6TB X300 to my array I purchased a WinTV Quad and then replaced with a TBS6205 TV card I created a 4th VM, pfSense by buying a AOC-SGP-I2 NIC on eBay for £60 Overall - £3,110 It’s been a fantastic year. I’ve more than achieved my goals of delivering more concurrent users and an efficient Plex server. What’s been amazing is the number of additional tasks my server has taken on-board over the year, without breaking a sweat: Kodi mysql database. I now use kodi throughout my home rather than Plex. This has been a very happy development as my media center journey started 15 years ago when I modded orignial Xbox consoles to run XBMC, and Cylon was originally built to deliver content to those XBMC machines TV server – I have ditched all my STBs over the year and now use a TVheadend docker delivering TV to Android and Fire TVs Unifi Controller – I’ve added 2 AC Pro APs over the year controlled by a docker and after 14 years of trying, I finally have a wi-fi network I am happy with pfSense router – excellent piece of software Home Automation through a Home Assistant Docker Automated media management dockers (sonar, radar, nzbget etc). I even use the excellent calibre-web docker to provide easy access to my book collection MineOS server docker for the kids Letsencrypt reverse proxy, openvpn-as and resilio-sync for secure remote access to server and files It really has become my HOME server – controlling my tvs, lighting, heating, media and internet usage! Future Plans Add a gaming GPU as the GT730 can’t even manage the most basic graphics settings Add more storage, probably moving to 8TB drives at the next purchase. This will probably mean I will have to build a dedicated pfSense box as I only have one SATA slot left on my motherboard, so I will need the PCIe slot to add a SATA card. I’m not keen on spending approx. £300-400 on a SATA card and a 2nd pfsense machine that will need maintaining, and I’m toying with maybe spending a bit more to get a new SSI-EEB motherboard (unfortunately this would mean a new case as well) with more PCIe and SATA ports for better future expansion options Think about the location of my box as it’s starting to get noisy with the number of drives spinning, which will only get worse as I add more. I think I might have a plan to tackle this in the future by using my cache drive for downloads (I don’t currently) and my latest media files before they are moved to the array; spending some time using the unbalance plugin and optimising my split levels to reduce the number of drives spinning. This also might involve using HDMI/USB over ethernet or something similar like Raspberry Pi thin clients, to put the server in a different room Here's to another 10 years!
  22. 7 points
    diskspeed.sh, version 2.6.5 UNRAID 5 Users: Download version 2.5 A new, enhanced version is being created for UNRAID 6 and later. Version 2.6.5 contains an edit by bonienl to allow it to work in version 6.4 Note: Disks under 25 GB will be skipped. Should work in virtualized environments. This utility will perform read tests at different points on each drive attached to the system (even those not assigned to the UNRAID array) to compute the average speed & generate a graph. Useful if you want to see if you have a drive that's slower than the others and negatively affecting your parity drive speeds. Even drives of the same make & model can perform marginally different - example with the cache drive and disk 9 in the below graph utilizing my backup server with drives retired from my main server. To execute, ensure no other processes are running on your UNRAID server that are accessing your hard drives (which will result in slower results) and run the diskspeed.sh script. Execution time will be approx. 90 seconds multiplied by the number of drives attached to the system with the default sample & test iterations. If you suspect a disk is failing, run the following command and replace "sdx" with the drive you want to test. It will test just that drive at every 2% of the capacity. diskspeed.sh -s 51 -n sdx Syntax diskspeed.sh -i # -s # diskspeed.sh --iterations # --samples # diskspeed.sh -f Examples diskspeed.sh Test 11 sample points one time (0%, 10%, 20%, 30%, 40%, 50%, 60%, 70%, 80%, 90%, 100%) diskspeed.sh -s 11 -i 1 Same as above as these are the default values diskspeed.sh -s 5 Test five sample points (0%, 25%, 50%, 75%, 100%) diskspeed.sh -s 3 Test three sample points (0%, 50%, 100%) diskspeed.sh -i 3 Test each sample point three times and take the average diskspeed.sh -s 21 Test the drive every 5% diskspeed.sh -f Perform a fast test where 200 MB is read at each location vs 1 GB. Not as accurate. diskspeed.sh -x sda,sdb Exclude drives sda & sdb diskspeed.sh -n sdc,sdd Only test drives sdc & sdd Expanding info on sample points: The script will check the start of the hard drive (0%) and the end of the hard drive (100%). The rest of the sample points are divided evenly between the start and end. So a sample request of 3 would test the start, end, and middle of the drive. If your graph seems "spiky", try running the script with the "-i 3" option to test each location three times and take the average. To view the graph, navigate to the location you executed the script file via your preferred file explorer on your UNRAID share (ex: \\tower\flash\scripts) and open the diskspeed.html file. You can toggle each drive off & on by clicking on its designation in the legend. Example console output The drives are tested in the order they were assigned by the OS diskspeed.sh for UNRAID, version 2.3 By John Bartlett. Support board @ limetech: http://goo.gl/ysJeYV /dev/sdb (Disk 4): 107 MB/sec avg /dev/sdc (Disk 2): 158 MB/sec avg /dev/sdd (Disk 5): 98 MB/sec avg /dev/sde (Disk 10): 98 MB/sec avg /dev/sdf (Disk 6): 100 MB/sec avg /dev/sdg (Disk 7): 99 MB/sec avg /dev/sdh (Disk : 97 MB/sec avg /dev/sdi (Disk 9): 97 MB/sec avg /dev/sdj (Disk 3): 123 MB/sec avg /dev/sdk (Disk 11): 104 MB/sec avg /dev/sdl (Disk 1): 112 MB/sec avg To see a graph of the drive's speeds, please browse to the current directory and open the file diskspeed.html in your Internet Browser application. Example graph How it works The script utilizes the dd utility to do a direct read at various offsets Change Log Version 2.6.4 Added support for UNRAID 6.3.0-RC9 Version 2.6.3 Changed memory check to ignore cache memory Version 2.6.2 Added a check to ensure there is enough free memory available to execute the dd command Added -n | --include option to specify which drives to test, comma delimited Ignore floppy drives Added support for nvme drives Version 2.6.1 Fixed issue identifying drives assigned sdxx (more than 26 drives attached) Fixed issue with data drives over 9 having the last digit truncated Version 2.6 Removed checks for invalid drives, redundent Altered drive inventory to exclude md? drives/identify drive/cache/parity assignments Modified to support UNRAID 6.2 running under OS 4.4.x and higher Version 2.5 Fixed computation for percentages less than 10% Reverted to 1 GB scans for better results but slower Added -f --fast to scan 200 MB instead of 1 GB, same as version 2.3 & 2.4 Version 2.4 If the drive model is not able to be determined via fdisk, extract it from mdmcd Add -l --log option to create the debug log file diskspeed.log Modified to not display the MB sec in drive inventory report for excluded drives Modified to compute the drive capacity from the number of bytes UNRAID reports for external drive cards. Added -g --graph option to display the drive by percentage comparison graph Added warning if files on the array are open which could mean drives are active Added spin up drive support by reading a random spot on the drive Version 2.3 Changed to use the "dd" command for speed testing, eliminates risk of hitting the end of the drive. The app will read 200MB of data at each testing location. Before scanning each spot, uses the "dd" command to place the drive head at the start of the test location. Added -o --output option for saving the file to a given location/name (credit pkn) Added report generation date & server name to the end of the report (credit pkn) Added a Y axis floor of zero to keep the graph from display negative ranges Hid graph that compared each drive by percentage. If you wish to re-enable it, change the line "ShowGraph1=0" to "ShowGraph1=1" Added average speed to the drive inventory list below the graph Added -x --exclude option to ignore drives, comma seperated. Ex: -x sda,sdb,sdc Added -o --output option to specify report HTML file name Version 2.2 Changed method of identifying the UNRAID boot drive and/or USB by looking for the file /bzimage or /config/ident.cfg if the device is mounted Skip drives < 25 GB Route fdisk errors to the bit bucket Removed the max size on the 2nd graph to allow smaller drives to scale if larger drives are hidden Version 2.1 Fixed GB Size determination to minimize hdparm hitting the end of the drive while performing a read test at the end of the drive (credit doron) Fixed division error in averaging sample sizes (credit doron) Updated graphs to size to 1000 px wide Added 2nd graph which shows drive speeds in relation to the largest drive size; this is a better indication of how your parity speeds may run Added drive identification details below the graphs Added support for scanning all hard drives attached to the system Version 2.0 Added ability to specify the number of tests performed at each sample spot Added ability to specify the number of samples to take, min of 3 samples. first sample will be at the start of the drive, last sample at the end, and the rest spread out evenly on the drive Added help screen Formatted the graph tool tip to display the information in a easy to read format Do not run if the parity sync is in process Added support for gaps in drive assignments Added support for arrays with no parity drive Version 1.1 Fix bug for >= 10 drives in array (credit bonienl) Fix graph bug so graph displays in MB Version 1.0 Initial Release UNRAID 5 Users: Use version 2.5 diskspeed.v2.5.zip diskspeed_2.6.4.zip diskspeed_2.6.5.zip
  23. 7 points
    Hi, guys, this is a three-part video series on building a 10 core server from start to finish. The server is based off an x-99 platform with a Xeon CPU. These videos follow what I did to make a complete server running unRAID. This includes building the hardware. Setting the bios. Creating the unRAID USB. Preclearing the disks, setting up the array, installing various plugins and tools. Installing media containers such as emby, sonarr radarr and some download clients then a couple of VMs. Hope some of you new guys to unRAID may find it interesting. Part 1 is hardware setup and build. Part 2 is unRAID instalation and configuration. Part 3 is Docker container and VM configuration and testing.
  24. 7 points
    I was asked by another user to put this update in this thread as well, but it appears the issue relating to Nested Page Tables with AMD platforms has been resolved via a patch created by Paolo Bonzini: https://marc.info/?l=kvm&m=150891016802546&w=2 Another user actually found the troublesome code segment, supplied a hack of a patch, then Paolo saw the real problem and implemented a proper solution. That said, it hasn't made its way into a full release of the kernel yet, so we are going to patch it in manually with the release of the next rc. Needless to say, we are incredibly excited if this patch truly does resolve all these nasty GPU pass through performance problems. Ryzen offers a strong use-case with unRAID and VMs with GPU pass through so long as it can keep on par with Intel for performance and price. I'm excited to hear what you guys achieve after the next update. All the best, Jon
  25. 7 points
    Hi Guys. I have made a video about setting up and configuring Duplicati on unRAID for cloud and network backups.