Jump to content

Leaderboard


Popular Content

Showing content with the highest reputation since 12/17/18 in Posts

  1. 7 points
    Hey there Unraid community! Just wanted to take a moment between all the present opening, food eating, and carrol singing to wish everyone here happy holidays! Thank you all for being so awesome!! All the best, Team LT (Tom, Jon, and Eric) Sent from my Pixel 3 XL using Tapatalk
  2. 4 points
    Support for Nginx Proxy Manager docker container Application Name: Nginx Proxy Manager Application Site: https://nginxproxymanager.jc21.com Docker Hub: https://hub.docker.com/r/jlesage/nginx-proxy-manager/ Github: https://github.com/jlesage/docker-nginx-proxy-manager Make sure to look at the complete documentation, available on Github ! Post any questions or issues relating to this docker in this thread.
  3. 4 points
    Added to Unraid 6.7
  4. 4 points
    Scheduled for Unraid OS 6.7 release.
  5. 3 points
    I was wanting to do GPU Hardware Acceleration with a Plex Docker but unRAID doesn't appear to have the drivers for the GPUs loaded. would be nice to have the option to install the drivers so the dockers could use them.
  6. 3 points
    MAN I KNOW RIGHT! I always want to create new accounts on boards i've never posted on and ask questions that are crazily discussed at holiday parties! It's like the other day, for the second night of Hanukah, me and Ishmael were having some friends over, and some guy was like "if I was going to do 10gbe in my house, I would go fiber, i don't care about the cable expense!" But this other guy was all "MAN, you gotta go copper. It's a little pricey now for the cards, but it's the future!" So after that, I went and created an account on the synology forums and posted: Do you use 10gbe? I have tried several 10gbe hardware providers like Intel, Quanta and as of lately, Mellanox. Would rank Mellanox the highest because it has the best quality for a very low price, really impressed with the provider. What do you use? Totally legit. Welcome to the forums.
  7. 2 points
    Notice: You must be running unRAID version 6.1 or later to use these plugins The easiest way of installing plugins is thru Community Applications. This is an apps installation manager developed by Squid and needs to be installed separately. The alternative way of installing an optional plugin is from the Plugin page in the WebGui and use the tab Install Pugin. The URLs of the optional plugins as mentioned below can be copied and pasted in the install box. Available Dynamix plugins Active Streams shows in real-time any open SMB and AFP network streams. This allows instant view of who is accessing the server - either by IP address or name - and see what content is opened. Optionally streams can be stopped from the GUI. Cache Dirs keeps folder information in memory to prevent unnecessary disk spin up. Dynamix builds a GUI front-end to allow entering of parameters for the cache_dirs script which is running in the background. S3 Sleep defines the conditions under which the system will go to S3 sleep mode. It also adds an unconditional 'sleep' button on the Array Operation page. System Info shows various details of your system hardware and BIOS. This includes processor, memory and sub-system components. System Stats shows in real-time the disk utilizations and critical system recources, such CPU usage, memory usage, interface bandwidth and disk I/O bandwidth. System Temp shows in real-time the temperature of the system CPU and motherboard. Temperatures can be displayed in Celsius or Fahrenheit. Your hardware must support the necessary probes, and additional software drivers may be required too. This plugin requires PERL, this package needs to be installed separately. System AutoFan allows automatic fan control based on the system temperature. High and low thresholds are used to speed up or speed down the fan. This is a new plugin and still under development. Schedules is a front-end utility for the built-in hourly, daily, weekly and monthly schedules. It allows the user to alter the schedule execution times using the GUI. See Settings -> Scheduler -> Fixed Schedules. System Buttons adds an one-click button to the header which allows for instant sleep, reboot, shutdown of the system or array start/stop. Local Master supports detection of the local master browser in an SMB network. It will display an icon in the header at the top-right when unRAID is elected as local master browser. Under SMB Workgroup settings more information about the current elected local master browser is given. SSD TRIM allows the creation of a cronjob to do regular SSD TRIM operations on the cache device(s). The command 'fstrim -v /mnt/cache' is executed at the given interval. File Integrity Real-time hashing and verification of files stored on the data disks of the array. This plugin reports on failed file content integrity and detects silent file corruption (aka bit-rot). WARNING: USING THIS PLUGIN ON DISKS FORMATTED IN REISERFS MAY LEAD TO SYSTEM INSTABILITY. IT IS RECOMMENDED TO USE XFS. SCSI Devices (unRAID 6.2 or later) updates the udev persistent storage devices rules file (courtesy of bubbaQ), which allows proper naming of SCSI attached disks. Please by aware that after installation of this plugin, it might be necessary to re-assign disks due to their changed names! Date Time (unRAID 6.2 or later) adds an interactive world map to the date and time settings. This allows the user to simply click on his/her country and select the corresponding time zone. In addition the world map highlights the countries in the currently selected time zone. Stop Shell (unRAID 6.4 or later) adds a script which gets invoked when the array is stopped. This script looks for any open shells in /mnt/... and terminate them. This ensures the array can be stopped. Be aware that automatic terminating of open shells may lead to data loss if an active process is writing to the array. Day Night (unRAID 6.5 or later) automatically toggles between a day theme and a night theme. Based on the sunrise and sunset times of your location. Installation URLs (copy & paste) Active Streams - https://raw.github.com/bergware/dynamix/master/unRAIDv6/dynamix.active.streams.plg Cache Dirs - https://raw.github.com/bergware/dynamix/master/unRAIDv6/dynamix.cache.dirs.plg S3 Sleep - https://raw.github.com/bergware/dynamix/master/unRAIDv6/dynamix.s3.sleep.plg System Info - https://raw.github.com/bergware/dynamix/master/unRAIDv6/dynamix.system.info.plg System Stats - https://raw.github.com/bergware/dynamix/master/unRAIDv6/dynamix.system.stats.plg System Temp - https://raw.github.com/bergware/dynamix/master/unRAIDv6/dynamix.system.temp.plg System AutoFan - https://raw.github.com/bergware/dynamix/master/unRAIDv6/dynamix.system.autofan.plg Schedules - https://raw.github.com/bergware/dynamix/master/unRAIDv6/dynamix.schedules.plg System Buttons - https://raw.github.com/bergware/dynamix/master/unRAIDv6/dynamix.system.buttons.plg Local Master - https://raw.github.com/bergware/dynamix/master/unRAIDv6/dynamix.local.master.plg SSD TRIM - https://raw.github.com/bergware/dynamix/master/unRAIDv6/dynamix.ssd.trim.plg File Integrity - https://raw.github.com/bergware/dynamix/master/unRAIDv6/dynamix.file.integrity.plg SCSI Devices - https://raw.github.com/bergware/dynamix/master/unRAIDv6/dynamix.scsi.devices.plg Date Time - https://raw.github.com/bergware/dynamix/master/unRAIDv6/dynamix.date.time.plg Stop Shell - https://raw.github.com/bergware/dynamix/master/unRAIDv6/dynamix.stop.shell.plg Day Night - https://raw.github.com/bergware/dynamix/master/unRAIDv6/dynamix.day.night.plg gridrunner aka Spaceinvader One has made a very nice video tutorial about several Dynamix plugins. A recommended watch when you like to learn more. You like my Dynamix plugins?
  8. 2 points
    The last time I checked, most of the SCSI changes were implemented in 4.19. I haven't done a full 4.19 vs 4.20 breakdown in the SCSI and FS areas/modules to see what additional changes were implemented. If 6.7 drops with 4.19, we "should" be good. If 6.7 comes with a Slackware re-baseline, even better, as there are several updated packages that would compliment the improvements. The other aspect I have been becoming familiar with is UNMAP. Similar to FSTRIM, it provides instructions to the PCI bus to perform certain actions. Again, learning as time permits. Nevertheless, it seems the SCSI community acknowledged the collapse of several modules and programming language/library optimization has effected several functionalities in the HBA world. I'm really hoping it all comes to bed at 4.19 or 4.20. Again, fingers crossed.
  9. 2 points
    It is included in the upcoming Unraid version 6.7
  10. 2 points
    better storage setup ability to add more nifty things via docker ability to run more vm's including a firewall your coolness factor increases by a factor of 5 virtual 10gbe connection to server from windows vm vs buying 10gbe hardware. ability to divide resources and not waste them if someone watches plex while you're gaming on a standalone computer, your game can suffer vs on unraid you can isolate them from each other did i mention your coolness factor increases by a factor of 5?
  11. 2 points
    Only real change on this release is that there is now a backup application feed server running (thank you @limetech @eschultz @jonp) In the event that the primary server cannot be reached, the feed will download instead from the secondary server, and failing that a USB stored feed (if present - ask for details) will be utilized instead. Oh yeah, happy ho-ho.
  12. 2 points
    Although I have a lot of sympathy for this view but in my mind the biggest objection I have is that it is not easily ‘discoverable’. I was thinking that it should appear on the Shares tab, but under the Disk Shares section (regardless of whether disk shares are enabled or not), not the User Shares section.
  13. 2 points
    Well I don't know what changed, but I tried it again today and it worked fine! Thanks for this patch!
  14. 2 points
    Looks like you have quite the variety of drives, so that complicates things. Here's how I would proceed. 1) Build a new array, ultimately your goal will be a solid, reliable array. Don't reuse any of your old drives since we are going to try an extract data. Note that with this method of recovery, I don't think you can rely on any drive giving you back 100%, so if you have to do a rebuild on any given drive (assuming you fully recover that many drives), I don't know how reliable your rebuilt drive would be, either. You're welcome to try it. If not, maybe start with 1 Parity and 1 Data and work your way up from there. I'm assuming you know which of the old drives are data and which were parity. This method recovers the data treating the old array drives as JBOD. 2) Take say the STBV5000100, buy another exact model drive. Last time, I bought a used working drive off E-Bay. Test the newer drive, make sure it works and is reliable. Replace the bad drive's board with the new drive's board. Plug the bad drive into the server and use something like Unassigned Devices to mount it, then see how much data you can copy off of it. Once you have extracted as much data as you can, unmount and remove your bad drive. Swap the controller boards back. Bad drive goes on the shelf in case you need it for further recovery. The newer drive can be pre-cleared and added to the array. Repeat this step for all drives. Something I heard was that reallocated sectors are recorded somewhere on the controller board. I heard it quite a long time ago, so I don't know if it is/was true. If it is, your recovery may involve accessing some incorrect sectors, which is why I think the data isn't guaranteed to be 100%, but again anything is better than 0%. This should also be non-destructive, so you could still use other methods to recover your data if you like. I have not heard of the diode fix, nor have I ever attempted to alter a controller board in any way. All I have done is a straight board swap, and hope that any data losses are livable. Thankfully, this isn't something that I have had to do regularly, but it has worked once or twice. PS: Dunno about the warranty, but I'd skip the soldering iron if you intend to go this route.
  15. 2 points
    kizer is spot on--the community here is incredible--I've had my butt saved by incredible members here. I've been using the same license (and same USB key now that I think about it hmmm) since buying a 1TB drive was considered HUGE and cost prohibitive. If you look at the cost of my license over the last decade plus, it's amounted to absolute peanuts.
  16. 2 points
    I am sorry I am going to have to skip this update. I just can't... Any ETA on 6.6.7?
  17. 2 points
  18. 1 point
    In a recent reply to a post by another forum member (archedraft), I provided this guide to help him assign one of his NIC devices to a virtual machine, leaving the other for host networking (unRAID OS). I didn't see much point in this because with KVM and VirtIO, we can create virtual network interfaces that offer little to no overhead over a physical NIC, but after testing with pfSense, archedraft confirmed for me that he saw a dramatic performance increase. The reason? In this particular instance, pfSense was acting as a firewall and is based on FreeBSD. The FreeBSD kernel used by pfSense, while having support for VirtIO, appears to be out of date and was not allowing full 1gbps LAN throughput as it does with Linux or Windows VMs. Passing through a physical Ethernet controller to his pfSense VM in this instance resolved his issue. So we have found at least one use case thus far to consider such a method, but in the future, we may find more. And since the question comes up from time to time, I thought it prudent to post this here as an advanced guide for those that want to try it. WARNING: If you do not have multiple NICs in your system, doing this will result in your server losing all network connectivity. IMPORTANT: Regarding VM to Host Networking Performance When VMs utilize VirtIO, their is another distinct advantage in that networking between the host and guest can take place without traversing the copper wire. This allows for much faster throughput than the physical NIC hardware even supports at the port level. As an example, in mounting an SMB share to my SSD-based cache pool from inside my Windows VM, I was able to see IO throughput to the share exceed 250MB/s (that's megabytes, not bits). When a VM is assigned a physical network controller, this advantage disappears as the VM will communicate with the host as if it was a separate physical machine, going out the one NIC, down to your router/switching infrastructure, and then back in. This will limit your network throughput to that of the physical hardware. In my previous Windows VM / SMB example, I would be limited to 1gbps or 125MB/s. Guide 1 - Login to your server via ssh. 2 - Type the following command: lspci You will get a list like this: 00:00.0 Host bridge: Intel Corporation 4th Gen Core Processor DRAM Controller (rev 06) 00:01.0 PCI bridge: Intel Corporation Xeon E3-1200 v3/4th Gen Core Processor PCI Express x16 Controller (rev 06) 00:01.1 PCI bridge: Intel Corporation Xeon E3-1200 v3/4th Gen Core Processor PCI Express x8 Controller (rev 06) 00:02.0 VGA compatible controller: Intel Corporation Xeon E3-1200 v3/4th Gen Core Processor Integrated Graphics Controller (rev 06) 00:03.0 Audio device: Intel Corporation Xeon E3-1200 v3/4th Gen Core Processor HD Audio Controller (rev 06) 00:14.0 USB controller: Intel Corporation 8 Series/C220 Series Chipset Family USB xHCI (rev 04) 00:16.0 Communication controller: Intel Corporation 8 Series/C220 Series Chipset Family MEI Controller #1 (rev 04) 00:19.0 Ethernet controller: Intel Corporation Ethernet Connection I217-V (rev 04) 00:1b.0 Audio device: Intel Corporation 8 Series/C220 Series Chipset High Definition Audio Controller (rev 04) 00:1c.0 PCI bridge: Intel Corporation 8 Series/C220 Series Chipset Family PCI Express Root Port #1 (rev d4) 00:1c.3 PCI bridge: Intel Corporation 8 Series/C220 Series Chipset Family PCI Express Root Port #4 (rev d4) 00:1f.0 ISA bridge: Intel Corporation Z87 Express LPC Controller (rev 04) 00:1f.2 SATA controller: Intel Corporation 8 Series/C220 Series Chipset Family 6-port SATA Controller 1 [AHCI mode] (rev 04) 00:1f.3 SMBus: Intel Corporation 8 Series/C220 Series Chipset Family SMBus Controller (rev 04) 01:00.0 VGA compatible controller: Advanced Micro Devices, Inc. [AMD/ATI] Cedar [Radeon HD 5000/6000/7350/8350 Series] 01:00.1 Audio device: Advanced Micro Devices, Inc. [AMD/ATI] Cedar HDMI Audio [Radeon HD 5400/6300 Series] 02:00.0 VGA compatible controller: NVIDIA Corporation GK110 [GeForce GTX 780] (rev a1) 02:00.1 Audio device: NVIDIA Corporation GK110 HDMI Audio (rev a1) 04:00.0 Multimedia video controller: Device 1a0a:6202 (rev 01) Identify the Ethernet controller you wish to assign. Note the PCI address for the device (from my list, it would be 00:19.0). From my list, I only have one network card, so I shouldn't do this, but if you have multiple, either one SHOULD be fine to select. 3 - Type the following command: lspci -n 00:00.0 0600: 8086:0c00 (rev 06) 00:01.0 0604: 8086:0c01 (rev 06) 00:01.1 0604: 8086:0c05 (rev 06) 00:02.0 0300: 8086:0412 (rev 06) 00:03.0 0403: 8086:0c0c (rev 06) 00:14.0 0c03: 8086:8c31 (rev 04) 00:16.0 0780: 8086:8c3a (rev 04) 00:19.0 0200: 8086:153b (rev 04) 00:1b.0 0403: 8086:8c20 (rev 04) 00:1c.0 0604: 8086:8c10 (rev d4) 00:1c.3 0604: 8086:8c16 (rev d4) 00:1f.0 0601: 8086:8c44 (rev 04) 00:1f.2 0106: 8086:8c02 (rev 04) 00:1f.3 0c05: 8086:8c22 (rev 04) 01:00.0 0300: 1002:68f9 01:00.1 0403: 1002:aa68 02:00.0 0300: 10de:1004 (rev a1) 02:00.1 0403: 10de:0e1a (rev a1) 04:00.0 0400: 1a0a:6202 (rev 01) 4 - Identify your network card by PCI address (first column of results). 5 - Obtain the vendor/product ID for that device from the last column. 00:19.0 from my example is 8086:153b. 6 - Edit your syslinux.cfg file and add the following after the append but before initrd=/bzroot. pci-stub.ids=8086:153b REPLACE THE VENDOR/PRODUCT ID FROM MY EXAMPLE ABOVE WITH THE ONE YOU OBTAINED IN STEP 5. 7 - Reboot your system. 8 - Edit your VM using the XML editor mode. 9 - Add the following between the <devices> and </devices> tags. <hostdev mode='subsystem' type='pci' managed='yes'> <driver name='vfio'/> <source> <address domain='0x0000' bus='0x00' slot='0x19' function='0x0'/> </source> </hostdev> Modify the address line entering in the two digit bus, slot, and function from your ID. So 00:19.0 translates to what I have above. Save the XML and start your VM. All should be right as rain! NOTE: If you get an error, it could be because your NIC is in an IOMMU group with another in-use PCI device (either assigned to the host or to another VM). In this instance, you can attempt to use the PCIE ACS Override option under the VM Manager settings page, but use of this toggle is considered experimental.
  19. 1 point
    You can use SAS or SATA on any of the shelfs, you can even have mixed SAS and SATA on the same shelf, only ting to be aware is speed will be limited if using slower linking devices, like SAS1/SATA2 devices in any of them.
  20. 1 point
    Hi so I just follow the steps. The file is on google drive encrypted and nothing in folder upload. Did also the temps command. What should i do next ? thx Edit: Now i see a mountcheck in unionfs and in rclone mount for you information
  21. 1 point
    If all hdd connect with power plug haven't shutdown immediately during power on, then more serious issue was happen. BTW a simple check (by mulitmeter) you can do, pls found 2 diode near power plug, 5v and 12v, a negative mark was connect to power +, if you got 5v or 12v on diode negative mark (N means negative, P is positive), then you can ignore problem on the diode. You need further troubleshoot. If not success, then you need donor board, swap board usually won't work https://www.donordrives.com/pcb-replacement-guide (this is ref. only and not up to date)
  22. 1 point
    yes - you should change to this new repository. Unfortunately when I updated the template, this does not roll down to existing users
  23. 1 point
    OP from the linked thread here, I happened to stumble across this thread. I used the guide from Intel here, which worked like a charm. You need to load the software onto a thumb drive and run it from the internal EFI shell. You can then save the output as a .sel file back onto the thumb drive and open from Notepad++ in Windows or whatever. Hope that helps. EDIT: words.
  24. 1 point
    docker restart jackett Sent from my SM-T560NU using Tapatalk
  25. 1 point
    I've had my server for a few years now.