Jump to content


  • Content Count

  • Joined

  • Last visited

  • Days Won


brainbone last won the day on January 31

brainbone had the most liked content!

Community Reputation

8 Neutral

About brainbone

  • Rank
    Advanced Member


  • Gender

Recent Profile Visitors

The recent visitors block is disabled and is not being shown to other users.

  1. Crap. I see that now for devices under the array. Unfortunately, some of my NVMe devices are Unassigned Devices. Doesn't look it's supported there. Guess this is a question for the Unassigned Devices plugin thread. Thanks.
  2. Is there a way to have certain devices, like NVMe drives, use a higher threshold for temperature warnings? NVMe can and do run at much hotter temperatures than HDDs. When I set the threshold to warn me when HDD temps are getting too high, I end up getting constant warnings when my NVMe devices get heavy write activity.
  3. But I currently have no use for multiple array support. I'd much rather have HA and replication as core features. But above that, I'd like GPU drivers baked in and officially supported. May I suggest cheering in the multiple array thread instead of jeering in the GPU one?
  4. Yeah, let's just get rid of docker, VMs, the whole plugin system ... all that stuff that's not "genuine NAS", whatever that means.
  5. That could be said about any feature request. You may not find it useful, but many others would. For this specific feature, I think some of you may be conflating the burden for Linuxserver.io to do each Unraid Nvidia release with what it would be for Limetech. The burden on Limetech would be many orders of magnitude less, which is why this feature would be beneficial.
  6. Please add GPU drivers to unraid builds. Doesn't Limetech already do this for some NICs, SAS controllers, etc? If so, I see no reason why GPU drivers (specifically for Nvidia in my case) shouldn't be added as well. There's no need to keep up to date with the latest driver, unless there are serious bugs/exploits that need to be patched, just like with any other driver Unraid uses. I really don't get all the "sky is falling" negativity surrounding this request. The work involved for Limetech to include these drivers is far less than for the Linuxserver.io team (and a huge thanks to each of them for that effort!) to add them after the fact.
  7. Until there's ubiquitous, unmetered, gigabit internet just about everywhere, there's going to be a need for transcoding. Transcoding to lower bit rates for streaming, and syncing, while on the road is one of my main uses of Unraid Nvidia. I don't see that going away any time soon.
  8. Having an asrock board and multiple nvidia GPUs, I was a little gun-shy updating to 6.8.1 based on a few past comments. Updated to Nvidia 6.8.1 without issue. @CHBMB Thank you!
  9. Letting the GPU handle decoding/encoding leaves your CPU open for other work. Generally, the encodes from a GPU (hardware encoder) will be lower quality than the encodes from the CPU (when using x264), though newer generation hardware encoders are starting to close this gap. (the K10 does not have a newer generation hardware encoder.) If you have a powerful enough CPU to keep up with the transcodes being requested and everything else being requested of your unraid server, you'll likely see no real benefit with a hardware encoder. "Unraid Nvidia" lets you use the GPU with Plex/Emby/etc. dockers. See this list for what you can expect in Plex. The K10 isn't included in that list, but if it works at all, I'd expect it to perform something like the other GK104's No experience with any of the Tesla cards, but how useful it will be depends on your specific use case. I use a Quadro P400 for nvenc/nvdec in my Plex Docker, leaving my CPU (and GTX 1080) open for use with a gaming VM. For me, it was worth the time involved passing the 1080 GTX through to my gaming VM, and using Unraid Nvidia to offload Plex transcoding to the P400. Allows me to dedicate more CPU cores to the Gaming VM.
  10. I just use a second NVMe for transcoding and other stuff that makes more sense on a scratch disk. This saves some endurance and bandwidth on the main cache SSD, where I'd rather not have its appdata and domains mounts go down, and leaves ram open for more useful endeavors like VMs, dockers, read cache, etc. If you're confident you don't need a huge amount of space for transcodes / scratch disk, a 64GB (or even 32GB, if you don't have that many plex clients) intel Optane will have much higher endurance than a typical NVMe. However, I just use a pair of 500GB 970 EVOs -- one for a standard cache drive + docker/vm storage, the other for transcode/scratch. I'll likely need to replace them every 3 to 5 years.
  11. Yikes! I'm behind the curve here. I updated to 6.7.2 from 6.6.6 a few days ago, before seeing this information. I've not yet had any corruption, but I'm a little worried. Glancing over this and related threads, so far it seems like only those storing SQLite DBs directly on disk rather than on SSD/Cache are seeing this issue. Is that correct, or are there reports of people storing appdata on cache with this issue? Having all my dockers in appdata, and having appdata set to cache only, am I immune to this issue, or should I roll back to 6.6.7?
  12. Please edit or delete that last post. The first post in this thread forbids talking about that!
  13. Make sure you pay attention to what's highlighted in green/yes. The 1030, for example, is a GPU you should NOT get. This list has more detailed information on GPUs with Plex.
  14. I'm not exactly sure how much extra power my GTX 1050 is chewing sitting idle without being in a power saving mode, but my guess (based on readings from a "Kill A Watt" with the GTX installed vs. not.) is around 15 to 20 watts. 15 to 20w x $0.15 per kwh = ~ $20 to $25 per year. I'd certainly be willing to donate at least that much towards getting this done.
  15. I think I'm seeing this same issue. When mover is running, Plex is for the most part frozen. Happened both before and after installing an old Geforce 1050 (using Unraid-Nvidia) for transcoding, so it seems unrelated to Unraid Nvidia. I'm not certain if this happened before I upgraded to 6.7.0 as I seldom manually ran the mover when I was on 6.5.x -- though I don't recall it happening before the upgrade to 6.7.0.