1812

Members
  • Posts

    2625
  • Joined

  • Last visited

  • Days Won

    16

Everything posted by 1812

  1. I don't think you understand, I am not contradicting your statement regarding Ram vs HDD speeds. I am saying that it is not as relevant to the topic as you attest. The topic is not Ram vs HDD speeds in unRaid via SMB. It is about working around the bottleneck of user shares to achieve greater SMB performance. As a side note, Disk Shares don't work for pools as of 6.9.0 rc 2, although the same effect of bypassing the user shares is achieved.
  2. As I wrote above, I did use a large file for file transfer testing. This did exceed my dirty ram percentage but I did not include that bit of information. And retesting using the ram clearing command shows the same resulting speeds as before, which is expected because I exceed the dirty ram percentage. But I will disagree with you on using Black Magic. Those results demonstrate and reinforce the point of this thread, that SMB transfers are bottlenecked using regular user shares on Unraid on macOS, and that high rate SMB transfers are possible using this workaround. Your implied point that transferring to RAM is faster is valid, and your point about SSD's possibly not being as fast or fast enough is also valid. But not so much as to dismiss the results of the Black Magic test. Each person using the workaround it will have different results based on their setup, which is why I detailed mine along with the results achieved.
  3. Firstly, credit to @testdasi for sharing this info in a similar thread by @mgutt for windows. (edited to fix attribution) I looked around the board a bit to try and figure out why mac's get such horrible smb performance on unRaid, the os we all know and love. I wanted to setup a faster network scratch drive, but seemed to hit a wall at 400-600ish MB/s. Setup: M1 MacBook Pro running Big Sur 11.1, connected to an OWC thunderbolt pro dock with 10GBE <--Cat6a--> MikroTik CRS305-1G-4S+IN <--DAC--> ML350p Gen 8 with a cheap mellanox connectx-2, running unRaid 6.9.0 RC2, with the share hosted on a pool of 4 Samsung EVO 860 1tb drives in RAID 10, mtu 1500 on both network adapters. With this setup, perf3 yields: This equates to 1.17GB/s, which is great for zero tuning. But I've never achieved anywhere near that on smb transfers on Unraid, until today. Original directions here: Following those as a guide almost worked, but it required manual entry via finder>go>connect to server every time I wanted to connect. Yes, I know you can have it auto connect at launch, but for security reasons, I don't store passwords to servers on my laptop. And I hate getting nagged when I'm away from the house about not being able to find/connect to the server. So my smb-custom.conf looked like this: [Scratch-Fast] path = /mnt/netdrive/Scratch comment = browseable = yes Force User = nobody valid users = banannajoe writeable = yes write list = admin vfs objects = fruit For me, my pool wasn't named cache, it was named netdrive. I made the share browsable so it was accessible via the finder and added "fruit" to the vfs objects for better macOS compatibility (or so we're told.) I also needed to add "writeable = yes" otherwise the share was readonly for me. Per the instructions, I made the share name different than the original share which was titled "Scratch". So when I use the finder to view the folders on the server, I use the one titled "Scratch-Fast". The resulting smb single large file transfer speeds: My receive speed is essentially the same now as iperf3, with the send being about 15% slower (which also matches an unposted iperf3 result.) Again, this is on mtu 1500. Since this for me intended as a fast, "cache" only share, I don't know the ramifications towards using this method of accessing a share and having the mover do it's thing later. It probably shouldn't matter, but I haven't tested it. So you assume the risk if you do this workaround. Since I still have access to the original folder via the normal user share (shfs) way, here is another comparison Normal user share: Workaround: There may be room for improvement on the write side, but this is good enough for me, for now.... maybe...
  4. hp ml350p kernel panic after boot. booting into safe mode with no plugins allows boot. Not sure which was causing the issue as I manually nuked them all and only reinstalled a few that I know I use regularly. I know this isn't that helpful, but reporting nonetheless.
  5. Hello not yet. I'm going on holiday soon but should be able to start on it sometime around the 27th or so
  6. h220 works fine. I have used them for years.
  7. moderately sized, about 70TB in disk space, which turns into 24TB array, 26TB pool for backup, 1tb mirrored cache, 2 tb mirrored pool w/4 disks
  8. just the bz that you replaced only. Or you can upgrade to a newer Unraid version via the normal method and it will also work without the patch.
  9. have you tried setting up syslog server or writing a copy to usb? it'll preserve everything up to the moment of lockup
  10. see if you can capture what the call trace is for. that could help.
  11. I would wait and see if you can get stability first. I don’t think it’s an issue but it’s one other thing that could be eliminated in a step by step process. If you are stable without it, you can always add/replace cache with it after.
  12. I see lots of network errors in the last syslog you uploaded, along with some BRRFS warnings. But one thing at a time. Sometimes HP Proliant onboard network controllers are stupid. So let's try this: Make a copy of your network.cfg file for safe keeping. Then delete the one you have off the usb drive. This will reset network configuration. Also disable the openvpn plugin. Boot normally and see if the problem fixes itself. If so, then re-enamble openvpn and reboot. If it crashes, then you have a configuration issue with openvpn OR the onboard controllers are being finicky with correct settings and there is nothing more to be done except to try the latest beta with newer linux kernel or go without openvpn. If it crashes regardless of the above, then if you're only using one of your ports, disable the others in Unraid networking settings. If it crashes, then go into the bios and manually disable the ports you aren't using, delete the network.cfg file again, and reboot.
  13. EDAC MC0: 1 CE error on CPU#0Channel#1_DIMM#0 EDAC MC0: 1 CE error on CPU#0Channel#2_DIMM#0 looking it up on the interwebs shows ram issue. So, probable ram error on Dim 0. Reseat the dim and see if it goes away. If not, then swap it to a different slot with another stick and see if the error follows it to the new dim number. If so, bad ram stick. If the error stays on dim 0, then there may be a problem with the slot or cpu. But most likely ram stick issue. that's my guess/troubleshooting.
  14. I think I win... or lose... overnight "idle" low is about 217 watts. But it's never really at idle.That includes: 2 8 core processors (32 threads) 128 GB ram 5 disk main array in various states of spun up and down 7 disk backup pool spun down 2 SSD cache up 4 SSD pool up 1 vm running with 2 rx580's and some various dockers contributing to a few watts here and there.
  15. status is still pending. it looked like it was working, but was not. it's a little tricky to explain but compiling showed no errors and passthrough was intermittent for some undetermined reason. once 6.9 goes into RC phase we will then attempt to try again. this is because it may take a combination of 6.9 but using a more recent linux kernel. or not. we'll know more once we get there. if anyone wants to playground, then I recommend downloading the kernel compiling plugin as a start. in some ways it has expedited the process of working on it.
  16. But if it was later, how could we really know?
  17. brothers and sisters, I'm here to testify! thanks to this thread and specifically to @ghost82 I am a a redeemed clover convert and a believer in opencore. not only is the vm booting faster, but I'm benchmarking 10% higher than the same os image on clover. And free performance is worth a lot. THANKS!
  18. xml (had to copy from the running vm) <?xml version='1.0' encoding='UTF-8'?> <domain type='kvm' id='34' xmlns:qemu='http://libvirt.org/schemas/domain/qemu/1.0'> <name>Desktop</name> <uuid>9e8e043c-3cfd-8a45-aeac-d80f5cbd234f</uuid> <metadata> <vmtemplate xmlns="unraid" name="MacOS" icon="/mnt/user/domains/MacinaboxCatalina/icon/catalina.png" os="Catalina"/> </metadata> <memory unit='KiB'>101187584</memory> <currentMemory unit='KiB'>101187584</currentMemory> <memoryBacking> <nosharepages/> </memoryBacking> <vcpu placement='static'>24</vcpu> <cputune> <vcpupin vcpu='0' cpuset='4'/> <vcpupin vcpu='1' cpuset='20'/> <vcpupin vcpu='2' cpuset='5'/> <vcpupin vcpu='3' cpuset='21'/> <vcpupin vcpu='4' cpuset='6'/> <vcpupin vcpu='5' cpuset='22'/> <vcpupin vcpu='6' cpuset='7'/> <vcpupin vcpu='7' cpuset='23'/> <vcpupin vcpu='8' cpuset='8'/> <vcpupin vcpu='9' cpuset='24'/> <vcpupin vcpu='10' cpuset='9'/> <vcpupin vcpu='11' cpuset='25'/> <vcpupin vcpu='12' cpuset='10'/> <vcpupin vcpu='13' cpuset='26'/> <vcpupin vcpu='14' cpuset='11'/> <vcpupin vcpu='15' cpuset='27'/> <vcpupin vcpu='16' cpuset='12'/> <vcpupin vcpu='17' cpuset='28'/> <vcpupin vcpu='18' cpuset='13'/> <vcpupin vcpu='19' cpuset='29'/> <vcpupin vcpu='20' cpuset='14'/> <vcpupin vcpu='21' cpuset='30'/> <vcpupin vcpu='22' cpuset='15'/> <vcpupin vcpu='23' cpuset='31'/> </cputune> <resource> <partition>/machine</partition> </resource> <os> <type arch='x86_64' machine='pc-q35-3.1'>hvm</type> <loader readonly='yes' type='pflash'>/mnt/scratch/vm/Desktop/ovmf/OVMF_CODE.fd</loader> <nvram>/mnt/scratch/vm/Desktop/ovmf/OVMF_VARS.fd</nvram> </os> <features> <acpi/> <apic/> <ioapic driver='qemu'/> </features> <cpu mode='host-passthrough' check='none'> <cache mode='passthrough'/> </cpu> <clock offset='utc'> <timer name='rtc' tickpolicy='catchup'/> <timer name='pit' tickpolicy='delay'/> <timer name='hpet' present='no'/> </clock> <on_poweroff>destroy</on_poweroff> <on_reboot>restart</on_reboot> <on_crash>restart</on_crash> <devices> <emulator>/usr/local/sbin/qemu</emulator> <disk type='file' device='disk'> <driver name='qemu' type='qcow2' cache='none' io='native' discard='unmap'/> <source file='/mnt/user/vm/Desktop/Clover.qcow2' index='4'/> <backingStore/> <target dev='hdc' bus='sata'/> <boot order='1'/> <alias name='sata0-0-2'/> <address type='drive' controller='0' bus='0' target='0' unit='2'/> </disk> <disk type='file' device='disk'> <driver name='qemu' type='qcow2' cache='none' io='native' discard='unmap'/> <source file='/mnt/scratch/vm/Desktop/macos_disk.qcow2' index='3'/> <backingStore/> <target dev='hdd' bus='sata'/> <alias name='sata0-0-3'/> <address type='drive' controller='0' bus='0' target='0' unit='3'/> </disk> <disk type='file' device='disk'> <driver name='qemu' type='qcow2' cache='none' io='native' discard='unmap'/> <source file='/mnt/scratch/vm/Scratch and Apps Disk/Scratch.qcow2' index='2'/> <backingStore/> <target dev='hde' bus='sata'/> <alias name='sata0-0-4'/> <address type='drive' controller='0' bus='0' target='0' unit='4'/> </disk> <disk type='file' device='disk'> <driver name='qemu' type='qcow2' cache='writeback'/> <source file='/mnt/disks/MacBackup/Backup/vdisk1.qcow2' index='1'/> <backingStore/> <target dev='hdf' bus='sata'/> <alias name='sata0-0-5'/> <address type='drive' controller='0' bus='0' target='0' unit='5'/> </disk> <controller type='usb' index='0' model='ich9-ehci1'> <alias name='usb'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x07' function='0x7'/> </controller> <controller type='usb' index='0' model='ich9-uhci1'> <alias name='usb'/> <master startport='0'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x07' function='0x0' multifunction='on'/> </controller> <controller type='usb' index='0' model='ich9-uhci2'> <alias name='usb'/> <master startport='2'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x07' function='0x1'/> </controller> <controller type='usb' index='0' model='ich9-uhci3'> <alias name='usb'/> <master startport='4'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x07' function='0x2'/> </controller> <controller type='sata' index='0'> <alias name='ide'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x1f' function='0x2'/> </controller> <controller type='pci' index='0' model='pcie-root'> <alias name='pcie.0'/> </controller> <controller type='pci' index='1' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='1' port='0x8'/> <alias name='pci.1'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x0' multifunction='on'/> </controller> <controller type='pci' index='2' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='2' port='0x9'/> <alias name='pci.2'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x1'/> </controller> <controller type='pci' index='3' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='3' port='0xa'/> <alias name='pci.3'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x2'/> </controller> <controller type='pci' index='4' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='4' port='0xb'/> <alias name='pci.4'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x3'/> </controller> <controller type='pci' index='5' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='5' port='0xc'/> <alias name='pci.5'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x4'/> </controller> <controller type='pci' index='6' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='6' port='0xd'/> <alias name='pci.6'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x5'/> </controller> <controller type='pci' index='7' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='7' port='0xe'/> <alias name='pci.7'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x6'/> </controller> <controller type='pci' index='8' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='8' port='0xf'/> <alias name='pci.8'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x7'/> </controller> <controller type='pci' index='9' model='pcie-to-pci-bridge'> <model name='pcie-pci-bridge'/> <alias name='pci.9'/> <address type='pci' domain='0x0000' bus='0x07' slot='0x00' function='0x0'/> </controller> <controller type='pci' index='10' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='10' port='0x10'/> <alias name='pci.10'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x02' function='0x0' multifunction='on'/> </controller> <controller type='pci' index='11' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='11' port='0x11'/> <alias name='pci.11'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x02' function='0x1'/> </controller> <controller type='virtio-serial' index='0'> <alias name='virtio-serial0'/> <address type='pci' domain='0x0000' bus='0x02' slot='0x00' function='0x0'/> </controller> <interface type='bridge'> <mac address='52:54:00:0a:24:44'/> <source bridge='br0'/> <target dev='vnet0'/> <model type='vmxnet3'/> <alias name='net0'/> <address type='pci' domain='0x0000' bus='0x01' slot='0x00' function='0x0'/> </interface> <serial type='pty'> <source path='/dev/pts/0'/> <target type='isa-serial' port='0'> <model name='isa-serial'/> </target> <alias name='serial0'/> </serial> <console type='pty' tty='/dev/pts/0'> <source path='/dev/pts/0'/> <target type='serial' port='0'/> <alias name='serial0'/> </console> <channel type='unix'> <source mode='bind' path='/var/lib/libvirt/qemu/channel/target/domain-34-Desktop/org.qemu.guest_agent.0'/> <target type='virtio' name='org.qemu.guest_agent.0' state='connected'/> <alias name='channel0'/> <address type='virtio-serial' controller='0' bus='0' port='1'/> </channel> <input type='mouse' bus='ps2'> <alias name='input0'/> </input> <input type='keyboard' bus='ps2'> <alias name='input1'/> </input> <input type='keyboard' bus='ps2'> <alias name='input2'/> </input> <input type='keyboard' bus='ps2'> <alias name='input3'/> </input> <hostdev mode='subsystem' type='pci' managed='yes'> <driver name='vfio'/> <source> <address domain='0x0000' bus='0x47' slot='0x00' function='0x0'/> </source> <alias name='hostdev0'/> <rom file='/mnt/user/domains/vbios/580.dump'/> <address type='pci' domain='0x0000' bus='0x03' slot='0x00' function='0x0'/> </hostdev> <hostdev mode='subsystem' type='pci' managed='yes'> <driver name='vfio'/> <source> <address domain='0x0000' bus='0x41' slot='0x00' function='0x0'/> </source> <alias name='hostdev1'/> <rom file='/mnt/user/domains/vbios/580a.dump'/> <address type='pci' domain='0x0000' bus='0x04' slot='0x00' function='0x0'/> </hostdev> <hostdev mode='subsystem' type='pci' managed='yes'> <driver name='vfio'/> <source> <address domain='0x0000' bus='0x09' slot='0x00' function='0x0'/> </source> <alias name='hostdev2'/> <address type='pci' domain='0x0000' bus='0x05' slot='0x00' function='0x0'/> </hostdev> <hostdev mode='subsystem' type='pci' managed='yes'> <driver name='vfio'/> <source> <address domain='0x0000' bus='0x0a' slot='0x00' function='0x0'/> </source> <alias name='hostdev3'/> <address type='pci' domain='0x0000' bus='0x06' slot='0x00' function='0x0'/> </hostdev> <hostdev mode='subsystem' type='pci' managed='yes'> <driver name='vfio'/> <source> <address domain='0x0000' bus='0x0b' slot='0x00' function='0x0'/> </source> <alias name='hostdev4'/> <address type='pci' domain='0x0000' bus='0x08' slot='0x00' function='0x0'/> </hostdev> <hostdev mode='subsystem' type='pci' managed='yes'> <driver name='vfio'/> <source> <address domain='0x0000' bus='0x0c' slot='0x00' function='0x0'/> </source> <alias name='hostdev5'/> <address type='pci' domain='0x0000' bus='0x0a' slot='0x00' function='0x0'/> </hostdev> <memballoon model='none'/> </devices> <seclabel type='dynamic' model='dac' relabel='yes'> <label>+0:+100</label> <imagelabel>+0:+100</imagelabel> </seclabel> <qemu:commandline> <qemu:arg value='-usb'/> <qemu:arg value='-device'/> <qemu:arg value='usb-kbd,bus=usb-bus.0'/> <qemu:arg value='-device'/> <qemu:arg value='isa-applesmc,osk=theysayonceyougomacyounevergoback'/> <qemu:arg value='-smbios'/> <qemu:arg value='type=2'/> <qemu:arg value='-cpu'/> <qemu:arg value='IvyBridge,kvm=on,vendor=GenuineIntel,+invtsc,vmware-cpuid-freq=on,+pcid,+ssse3,+sse4.2,+popcnt,+avx,+aes,+xsave,+xsaveopt,check'/> <qemu:arg value='-set'/> <qemu:arg value='device.sata0-0-2.rotation_rate=1'/> <qemu:arg value='-set'/> <qemu:arg value='device.sata0-0-3.rotation_rate=1'/> <qemu:arg value='-set'/> <qemu:arg value='device.sata0-0-4.rotation_rate=1'/> </qemu:commandline> </domain> running Catalina 10.15.4 sending you the config.plist now much appreciated!
  19. I only have what was pulled of of Leoyzen's GitHub page. Just now I attempted to locate it on the web but apparently my google-fu is a bit weak today...
  20. OK so I decided to take the plunge after reading all 21 pages. My assumption was that I could take a perfectly working Catalina vm made with macinabox, mount the clover qcow2 image, trash the clover efi folder, replace with the one from @Leoyzen's github and stand a decent chance of it booting, even with no alternations to the xml. Well that didn't happen. I am able to get to the oc boot selector, but regardless of what is selected, the result is the same. This happens almost instantaneously after the boot selector. I also tried another OC.qcow2 image that was uploaded in this thread and the result was the same. Not sure where to go from here. Maybe I'm missing something obvious? Anything I can provide to help in troubleshooting?
  21. 128gb ecc. 96gb for a vm, the rest for dockers and things