• Posts

  • Joined

  • Last visited

  • Days Won


Everything posted by 1812

  1. works fine for me either way on both machines I have running 6.12.3. For reference, I'm on safari Version 16.5.1 (18615.
  2. upgraded 2 similar machines from .2 to .3 no issues except this is new: Jul 17 14:54:50 Tower1 smbd[24128]: [2023/07/17 14:54:50.767379, 0] ../../source3/lib/adouble.c:2363(ad_read_rsrc_adouble) Jul 17 14:54:50 Tower1 smbd[24128]: ad_read_rsrc_adouble: invalid AppleDouble resource .DS_Store but I'm not that concerned about it at the moment because of the file type/what that does. I haven't had any file transfer issues (yet) will come back if it becomes a thing.
  3. helped me out a few times (both in my own threads and when searching for previously solved threads), much appreciated and congrats!
  4. regular user shares are a little bottlenecked on unraid. to increase smb performance look at https://forums.unraid.net/topic/100855-workaround-how-to-get-macos-smb-transfers-at-1gbs-on-10gbe/ and the original source https://forums.unraid.net/topic/97165-smb-performance-tuning/#comment-895985
  5. Happy New Year! I'm new to 3d printing so I thought I'd give this docker a try my an ender 3 s1 pro. I ran into the same issue with serial/by-id issue and had the same result using "udevadm info --name=/dev/bus/usb/001/002" showing "device node not found". This may not be the "right" way but this is what I did to get my printer recognized: 1. open a terminal window on unraid server web page 2. enter "cd /dev/serial/by-id" 3. enter "dir" this showed me what I was after, which in my case was: usb-1a86_USB_Serial-if00-port0 in the docker template I changed it to "/dev/serial/by-id/usb-1a86_USB_Serial-if00-port0" I then followed the directions in the second post of this topic (even though I don't have multiple printers) which were 1) go to settings/serial connections and add /dev/serial/by-id/* to "Additional serial ports" 2) save 3) go to settings/serial connections and change AUTO in the serial port list to your device 4) save Just completed my first test and it all seems to be doing what it is suppose to. Hope that helps!
  6. updated 2 machines from 6.11.x to this. no issues.
  7. upgraded 2 boxes: one from 6.11.3, the other from 6.10.3. no issues currently observed. SPIFFY!
  8. 1. you can assign as many cores as you have to your vm or a number of vm's. 2. Any level of unRaid does this. The difference in the levels is the amount of disks you can attach. you can assign whatever and however many cores to any vm, and even stack vm's on the same cores (though you'll get an understandable performance penalty.) You can even isolate the cores away from unRaid running as host and reserve the cores to only be used by the vm's you choose. 3. open a question in the appropriate place on the forum and someone will probably be able to help you with that.
  9. and I broke the open files plugin by running it out of memory
  10. Then I'm not sure why this is currently working? I mean, I know what value I entered on purpose for max open files (just to see what would happen). But it seems like I'm chugging along beyond it. Maybe it'll all come to a screeching halt soon? ¯\_(ツ)_/¯
  11. just a heads up. A Mac photo library that is of any substantial size (like 175GB for example) blows past the 40964 open file limit when transferring to the server and the file limit has to be increased way beyond that. [learned from experience and several failures today before finding this thread and increasing to a ludicrous number to try and get this moved over]. Hopefully it will be more easily user adjustable in the future release.
  12. Over 4 years later and now I'm changing my +1 to a +10
  13. side note: you can get around the halted boot process for missing devices by disabling them in the bios.
  14. why does it keep installing the newest Nvidia driver after every update? I use a gt for Plex and it always rolls to the latest vs keeping me on the 470.129.06 which is the one that works for this card.
  15. FWIW I have 2 HP ML30 Gen 9 servers updated to 6.10.2, both with dual: 02:00.0 Ethernet controller: Broadcom Inc. and subsidiaries NetXtreme BCM5720 Gigabit Ethernet PCIe 02:00.1 Ethernet controller: Broadcom Inc. and subsidiaries NetXtreme BCM5720 Gigabit Ethernet PCIe I did the "un-blacklist" procedure and have experienced no errors. I will also add that the Broadcom controllers are not eth0 in both systems either, and as far as I know there were no reports of this model of HP servers with issues.
  16. soke too soon, having a problem with my gpu, which went from functional to not. Log shows: May 19 13:46:23 Tower kernel: nvidia-nvlink: Unregistered the Nvlink Core, major device number 245 May 19 13:46:23 Tower kernel: nvidia-nvlink: Nvlink Core is being initialized, major device number 245 May 19 13:46:23 Tower kernel: NVRM: The NVIDIA GeForce GT 730 GPU installed in this system is May 19 13:46:23 Tower kernel: NVRM: supported through the NVIDIA 470.xx Legacy drivers. Please May 19 13:46:23 Tower kernel: NVRM: visit http://www.nvidia.com/object/unix.html for more May 19 13:46:23 Tower kernel: NVRM: information. The 510.73.05 NVIDIA driver will ignore May 19 13:46:23 Tower kernel: NVRM: this GPU. Continuing probe... May 19 13:46:23 Tower kernel: NVRM: No NVIDIA GPU found. It appears that my GPU driver was automatically updated and now I have to roll back to the 470.xx option in the Nvidia plugin. Ok I guess... a bit of an annoyance to have to fix something like this that was working just fine.
  17. updated 2 machines. first one no problems. Second one had all disks in the 2 pools "missing" as the device names had changed (those using an H240). I noted the disk positions/assignments, used new config and preserved the array disks, went back to the main tab and re-assigned the pool disks, marked parity as correct, and started the array. Normal operation as expected with no loss of data (also expected).
  18. I already listed my hardware specs. So moving on: Firstly, don't write to the array. If you have to, make sure you're on reconstruct write. For cache use ssd/nvme cache disks if you want fastest performance. second, see here: Try that and see if your experience changes. *Note* I don't have this setting on anything but a fast network share. The speeds I posted yesterday are just a a basic share that uses a cache drive.
  19. M1 MacBook Pro > OWC Thunderbolt Pro Doc > 10gbe ethernet cable > mikrotik 4 port 10gbe switch > direct attach copper cable > mellanox connectx-2 card > hp ml30 gen9 unraid server. file copy from the server, off a cache pool of 6 data ssd's in raid10, writing is about 100MB/s slower. Reading from my spinning array gets me about 170MB/s give or take using exos drives. Writing is a little slower. So, it works for me and my Mac. YMMV depending on server/client hardware specifics and tuning.
  20. the very first post says "Installation (procedure used for unRaid 6.9.0 RC 2 and up to current stable versions, previous versions not supported). This thread will be updated/deprecated when 6.10 goes stable to reflect the changes with that version." The patch is a part of the OS currently in the RC versions, so technically it should work if you follow step 4 and 6 in the first post, rebooting after step 4. I sold my last proliant a year ago, and don't always run RC versions so I don't have the ability to test at the moment, hence why RC is not supported (plus things can unexpectedly change from version to version.) If you do try steps 4 and 5, kindly report back the results.
  21. Current does not equate to RC/beta versions. Current pertains to stable versions. But I have updated that text to reflect that since there seems to be confusion.
  22. the second will be listed as de[recated once the stable version comes out, thanks to you!
  23. She's selling a service to set it up for you.