sonic6

Members
  • Posts

    610
  • Joined

  • Last visited

  • Days Won

    2

Everything posted by sonic6

  1. Great to hear. My Swag is older than this variables, i never used them, so i don't know if those works normally without problems. Maybe wrong Editor or Encoding?
  2. my subfolder sample looks a bit different: ## Version 2021/05/18 # make sure that your dns has a cname set for tautulli and that your tautulli container is not using a base url server { listen 443 ssl; listen [::]:443 ssl; server_name tautulli.*; include /config/nginx/ssl.conf; client_max_body_size 0; # enable for ldap auth, fill in ldap details in ldap.conf #include /config/nginx/ldap.conf; # enable for Authelia #include /config/nginx/authelia-server.conf; location / { # enable the next two lines for http auth #auth_basic "Restricted"; #auth_basic_user_file /config/nginx/.htpasswd; # enable the next two lines for ldap auth #auth_request /auth; #error_page 401 =200 /ldaplogin; # enable for Authelia #include /config/nginx/authelia-location.conf; include /config/nginx/proxy.conf; include /config/nginx/resolver.conf; set $upstream_app tautulli; set $upstream_port 8181; set $upstream_proto http; proxy_pass $upstream_proto://$upstream_app:$upstream_port; } location ~ (/tautulli)?/api { include /config/nginx/proxy.conf; include /config/nginx/resolver.conf; set $upstream_app tautulli; set $upstream_port 8181; set $upstream_proto http; proxy_pass $upstream_proto://$upstream_app:$upstream_port; } location ~ (/tautulli)?/newsletter { include /config/nginx/proxy.conf; include /config/nginx/resolver.conf; set $upstream_app tautulli; set $upstream_port 8181; set $upstream_proto http; proxy_pass $upstream_proto://$upstream_app:$upstream_port; } location ~ (/tautulli)?/image { include /config/nginx/proxy.conf; include /config/nginx/resolver.conf; set $upstream_app tautulli; set $upstream_port 8181; set $upstream_proto http; proxy_pass $upstream_proto://$upstream_app:$upstream_port; } }
  3. Hier für eignet sich der CF-Tunnel (Argo) nur bedingt, da hier abgesehen von http/https Datentransfers über 100MB unterbunden werden. Solang man NC nur über das Webinterface und die Apps nutzt, ist das kein Problem. WebDAV klappt hier aber leider nicht. Hier würde mir in erste Linie auch nur der Weg über den SSH Tunnel einfallen. Habe ich aber selbst noch nicht eingerichtet. Ich glaube damit sieht es schon ganz gut aus, dass sein Tunnel funktioniert. Der Rest müsste dann ja ablaufen wie in den restlichen NPM Guides mit den Subdomain. Ich nutze leider kein NPM. Ich meine mich zu erinnern, dass wir das Thema schon einmal wegen DS-Lite hier hatten. Vielleicht kann @flinx01 hier ein wenig helfen?
  4. Bitte im Hinterkopf behalten, dass C-States bei Ryzen CPUs Probleme machen kann:
  5. root@Unraid-1:~# lspci -vv | awk '/ASPM/{print $0}' RS= | grep --color -P '(^[a-z0-9:.]+|ASPM )' 00:01.0 PCI bridge: Intel Corporation 6th-10th Gen Core Processor PCIe Controller (x16) (rev 05) (prog-if 00 [Normal decode]) LnkCap: Port #2, Speed 8GT/s, Width x16, ASPM L0s L1, Exit Latency L0s <256ns, L1 <8us LnkCtl: ASPM Disabled; RCB 64 bytes, Disabled- CommClk+ 00:1b.0 PCI bridge: Intel Corporation Comet Lake PCI Express Root Port #21 (rev f0) (prog-if 00 [Normal decode]) LnkCap: Port #21, Speed 8GT/s, Width x4, ASPM L0s L1, Exit Latency L0s <1us, L1 <16us LnkCtl: ASPM L0s L1 Enabled; RCB 64 bytes, Disabled- CommClk+ 00:1d.0 PCI bridge: Intel Corporation Comet Lake PCI Express Root Port #9 (rev f0) (prog-if 00 [Normal decode]) LnkCap: Port #9, Speed 8GT/s, Width x4, ASPM L0s L1, Exit Latency L0s unlimited, L1 <16us LnkCtl: ASPM L0s L1 Enabled; RCB 64 bytes, Disabled- CommClk- pcilib: sysfs_read_vpd: read failed: No such device 01:00.0 Serial Attached SCSI controller: Broadcom / LSI SAS2008 PCI-Express Fusion-MPT SAS-2 [Falcon] (rev 03) LnkCap: Port #0, Speed 5GT/s, Width x8, ASPM L0s, Exit Latency L0s <64ns LnkCtl: ASPM Disabled; RCB 64 bytes, Disabled- CommClk+ 02:00.0 PCI bridge: ASMedia Technology Inc. ASM1184e 4-Port PCIe x1 Gen2 Packet Switch (prog-if 00 [Normal decode]) LnkCap: Port #0, Speed 5GT/s, Width x1, ASPM L0s L1, Exit Latency L0s <2us, L1 unlimited LnkCtl: ASPM L0s L1 Enabled; Disabled- CommClk+ 03:01.0 PCI bridge: ASMedia Technology Inc. ASM1184e 4-Port PCIe x1 Gen2 Packet Switch (prog-if 00 [Normal decode]) LnkCap: Port #1, Speed 5GT/s, Width x1, ASPM L0s L1, Exit Latency L0s <2us, L1 unlimited LnkCtl: ASPM L1 Enabled; Disabled- CommClk+ 03:03.0 PCI bridge: ASMedia Technology Inc. ASM1184e 4-Port PCIe x1 Gen2 Packet Switch (prog-if 00 [Normal decode]) LnkCap: Port #3, Speed 5GT/s, Width x1, ASPM L0s L1, Exit Latency L0s unlimited, L1 unlimited LnkCtl: ASPM L0s L1 Enabled; Disabled- CommClk- 03:05.0 PCI bridge: ASMedia Technology Inc. ASM1184e 4-Port PCIe x1 Gen2 Packet Switch (prog-if 00 [Normal decode]) LnkCap: Port #5, Speed 5GT/s, Width x1, ASPM L0s L1, Exit Latency L0s <2us, L1 unlimited LnkCtl: ASPM L1 Enabled; Disabled- CommClk+ 03:07.0 PCI bridge: ASMedia Technology Inc. ASM1184e 4-Port PCIe x1 Gen2 Packet Switch (prog-if 00 [Normal decode]) LnkCap: Port #7, Speed 5GT/s, Width x1, ASPM L0s L1, Exit Latency L0s <2us, L1 unlimited LnkCtl: ASPM L1 Enabled; Disabled- CommClk+ 04:00.0 PCI bridge: Advanced Micro Devices, Inc. [AMD/ATI] Navi 10 XL Upstream Port of PCI Express Switch (rev c1) (prog-if 00 [Normal decode]) LnkCap: Port #1, Speed 16GT/s, Width x8, ASPM L1, Exit Latency L1 <64us LnkCtl: ASPM L1 Enabled; Disabled- CommClk+ 05:00.0 PCI bridge: Advanced Micro Devices, Inc. [AMD/ATI] Navi 10 XL Downstream Port of PCI Express Switch (prog-if 00 [Normal decode]) LnkCap: Port #0, Speed 16GT/s, Width x16, ASPM L0s L1, Exit Latency L0s <64ns, L1 <1us LnkCtl: ASPM L0s L1 Enabled; Disabled- CommClk+ 06:00.0 VGA compatible controller: Advanced Micro Devices, Inc. [AMD/ATI] Navi 23 [Radeon RX 6600/6600 XT/6600M] (rev c1) (prog-if 00 [VGA controller]) LnkCap: Port #0, Speed 16GT/s, Width x16, ASPM L0s L1, Exit Latency L0s <64ns, L1 <1us LnkCtl: ASPM L0s L1 Enabled; RCB 64 bytes, Disabled- CommClk+ 06:00.1 Audio device: Advanced Micro Devices, Inc. [AMD/ATI] Navi 21 HDMI Audio [Radeon RX 6800/6800 XT / 6900 XT] LnkCap: Port #0, Speed 16GT/s, Width x16, ASPM L0s L1, Exit Latency L0s <64ns, L1 <1us LnkCtl: ASPM L0s L1 Enabled; RCB 64 bytes, Disabled- CommClk+ 08:00.0 PCI bridge: Advanced Micro Devices, Inc. [AMD/ATI] Navi 10 XL Upstream Port of PCI Express Switch (rev c1) (prog-if 00 [Normal decode]) LnkCap: Port #5, Speed 16GT/s, Width x8, ASPM L1, Exit Latency L1 <64us LnkCtl: ASPM L1 Enabled; Disabled- CommClk+ 09:00.0 PCI bridge: Advanced Micro Devices, Inc. [AMD/ATI] Navi 10 XL Downstream Port of PCI Express Switch (prog-if 00 [Normal decode]) LnkCap: Port #0, Speed 16GT/s, Width x16, ASPM L0s L1, Exit Latency L0s <64ns, L1 <1us LnkCtl: ASPM L0s L1 Enabled; Disabled- CommClk+ 0a:00.0 VGA compatible controller: Advanced Micro Devices, Inc. [AMD/ATI] Navi 23 [Radeon RX 6600/6600 XT/6600M] (rev c1) (prog-if 00 [VGA controller]) LnkCap: Port #0, Speed 16GT/s, Width x16, ASPM L0s L1, Exit Latency L0s <64ns, L1 <1us LnkCtl: ASPM L0s L1 Enabled; RCB 64 bytes, Disabled- CommClk+ 0a:00.1 Audio device: Advanced Micro Devices, Inc. [AMD/ATI] Navi 21 HDMI Audio [Radeon RX 6800/6800 XT / 6900 XT] LnkCap: Port #0, Speed 16GT/s, Width x16, ASPM L0s L1, Exit Latency L0s <64ns, L1 <1us LnkCtl: ASPM L0s L1 Enabled; RCB 64 bytes, Disabled- CommClk+ 0b:00.0 PCI bridge: Advanced Micro Devices, Inc. [AMD/ATI] Navi 10 XL Upstream Port of PCI Express Switch (rev c1) (prog-if 00 [Normal decode]) LnkCap: Port #7, Speed 16GT/s, Width x8, ASPM L1, Exit Latency L1 <64us LnkCtl: ASPM L1 Enabled; Disabled- CommClk+ 0c:00.0 PCI bridge: Advanced Micro Devices, Inc. [AMD/ATI] Navi 10 XL Downstream Port of PCI Express Switch (prog-if 00 [Normal decode]) LnkCap: Port #0, Speed 16GT/s, Width x16, ASPM L0s L1, Exit Latency L0s <64ns, L1 <1us LnkCtl: ASPM L0s L1 Enabled; Disabled- CommClk+ 0d:00.0 VGA compatible controller: Advanced Micro Devices, Inc. [AMD/ATI] Navi 23 [Radeon RX 6600/6600 XT/6600M] (rev c1) (prog-if 00 [VGA controller]) LnkCap: Port #0, Speed 16GT/s, Width x16, ASPM L0s L1, Exit Latency L0s <64ns, L1 <1us LnkCtl: ASPM L0s L1 Enabled; RCB 64 bytes, Disabled- CommClk+ 0d:00.1 Audio device: Advanced Micro Devices, Inc. [AMD/ATI] Navi 21 HDMI Audio [Radeon RX 6800/6800 XT / 6900 XT] LnkCap: Port #0, Speed 16GT/s, Width x16, ASPM L0s L1, Exit Latency L0s <64ns, L1 <1us LnkCtl: ASPM L0s L1 Enabled; RCB 64 bytes, Disabled- CommClk+
  6. this could be the reason. this forum is about Unraid, not Oracle VirtualBox. Sorry
  7. Ja das ist tatsächlich so. Was mich nach wie vor irritiert ist, dass sogar eine UD, welche nicht gemounted ist, von diesen Spinups betroffen ist. Jedoch bin ich nach wie vor nicht dahinter gestiegen und hatte auch wenig Zeit in den letzten Woche mich darum weiter zu kümmern.
  8. sorry hatte mich verschrieben. Ich habe "/data" auf den Cache verschoben.
  9. Zumindest für den "Unraid mit Array, Cache und Spindown"-Usecase nicht optimal. Da ich jedoch nach wie vor recht häufige Spinups meines Array habe, bin ich wieder zum dauerhaft aktiven Array zurück gekehrt. Jedoch habe ich den kompletten /data/ Pfad auf dem redundanten Cache verschoben. Tut ja letztendlich nicht weh. Für den Fall-der-Fälle wird wöchentlich davon ein Backup aufs Array gemacht. Dennoch danke für das aufklamüsern.
  10. Einfach selbständig gewisse Wartungsintervalle einhalten: Sichtkontrolle der Akkus in gewissen Zyklen. Ist der Akku aufgebläht, treten Substanzen aus... Natürlich sollte der Akku auch ab und an ausgetauscht, gegen mechanische Einflüsse und gegen Kurzschluss geschützt werden. Dann sollte nichts passieren.
  11. i think the most of unraid users are running Unraid headless. also HiveOS will be running headless in the most cases. so this is the easiest and most common way.
  12. Wir sollten das Thema einfach belassen. Man konnte die Geschichte verfolgen und sich seinen Teil denken. Letztendlich hat Unraid un er einfach nicht zusammen gepasst und jegliche weitere Urteile die wir hier öffentlich fällen, werden nur ein schlechtes Licht auf Unraid und unsere sehr hilfsbereite Community. Im schlimmsten Fall steigert es für neue und/oder unerfahrene User Ihre Fragen zu stellen.
  13. Abgesehen, vom Guide, aber wäre "Unbalance" nicht genau jenes?
  14. einfach den Server neu starten ohne den "powertop --auto-tune" Befehl aus dem /boot/config/go File. Ist das nicht ein Multiplier und sind diese nicht inkompatible mit Unraid?
  15. I used this just for troubleshooting. Thank you! you helped me alot about understanding how the xml works. It "works" at this point with the "bus"-solution and if a found another way, i will report!
  16. This runs fine: If i change the GPUs to "bus 0x01"... then i got the windows from this post okay, now i will be complicated: i am using this m.2 to PCIe riser then a 1-to-4 PCIe Card (picture) and three PCIe x16 riser i know this is a very weird and messi setup. but when i delete just one card from the xml (doesn't matter which one) the values are fine and the vm starts up. so in my eyes, the hardware setup isn't the problem.
  17. So, i try adding noVNC, but it looks like the VM won't really start up: if i add all GPUs under pci-root (0x00) i can follow the start up on VNC. My bindigs are ok? Maybe there are limitations with 3 GPUs on Unraid, QEMU, KVM? Are there some other places, where i can look at Logs or something else?
  18. Sorry, i wasn't at home. So, if i start the "working" xml i can see in the unraid dashbaord that Thread "1" goes up to 100% for a few second. Then the load from Thread "1" goes down and Thread "7" goes up to 100%. After that both are beetween 20%-30%. With "your" xml only Thread "1" goes up to 100% for a few seconds and then nothing. Both Threads goes down to 5% and i can't connect to the VM with ssh. I don't have any error or messeages in the logs, so i don't know where or for what i can take a look.
  19. Does not work... I don't understand this... maybe it is the image?
  20. I never change this, and it works without problems for two month. Thank you very much, I will try your recommend xml and report later.
  21. I don't know why, but i copied the wrong XML. Hi used multifunction on all GPUs but the main difference between the Q35 and i440 template is, that the CPU uses diffent "bus" on Q35 and "slot" on i440. If i change it to bus='0x00' on all GPUs and use slot='0x01', '0x02', '0x03' then the VM starts. But i don't understad why this will not startup: <?xml version='1.0' encoding='UTF-8'?> <domain type='kvm' id='2'> <name>hiveos</name> <uuid>76c45fbe-5d5b-13bd-b6bf-5360db4c5437</uuid> <metadata> <vmtemplate xmlns="unraid" name="Ubuntu" icon="ubuntu.png" os="ubuntu"/> </metadata> <memory unit='KiB'>4194304</memory> <currentMemory unit='KiB'>4194304</currentMemory> <memoryBacking> <nosharepages/> </memoryBacking> <vcpu placement='static'>2</vcpu> <cputune> <vcpupin vcpu='0' cpuset='1'/> <vcpupin vcpu='1' cpuset='7'/> </cputune> <resource> <partition>/machine</partition> </resource> <os> <type arch='x86_64' machine='pc-q35-6.1'>hvm</type> <loader readonly='yes' type='pflash'>/usr/share/qemu/ovmf-x64/OVMF_CODE-pure-efi.fd</loader> <nvram>/etc/libvirt/qemu/nvram/76c45fbe-5d5b-13bd-b6bf-5360db4c5437_VARS-pure-efi.fd</nvram> </os> <features> <acpi/> <apic/> </features> <cpu mode='host-passthrough' check='none' migratable='on'> <topology sockets='1' dies='1' cores='1' threads='2'/> <cache mode='passthrough'/> </cpu> <clock offset='utc'> <timer name='rtc' tickpolicy='catchup'/> <timer name='pit' tickpolicy='delay'/> <timer name='hpet' present='no'/> </clock> <on_poweroff>destroy</on_poweroff> <on_reboot>restart</on_reboot> <on_crash>restart</on_crash> <devices> <emulator>/usr/local/sbin/qemu</emulator> <disk type='file' device='disk'> <driver name='qemu' type='raw' cache='writeback'/> <source file='/mnt/cache_vm/domains/hiveos/vdisk1.img' index='1'/> <backingStore/> <target dev='hdc' bus='sata'/> <boot order='1'/> <alias name='sata0-0-2'/> <address type='drive' controller='0' bus='0' target='0' unit='2'/> </disk> <controller type='pci' index='0' model='pcie-root'> <alias name='pcie.0'/> </controller> <controller type='pci' index='1' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='1' port='0x8'/> <alias name='pci.1'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x0' multifunction='on'/> </controller> <controller type='pci' index='2' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='2' port='0x9'/> <alias name='pci.2'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x1'/> </controller> <controller type='pci' index='3' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='3' port='0xa'/> <alias name='pci.3'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x2'/> </controller> <controller type='pci' index='4' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='4' port='0xb'/> <alias name='pci.4'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x3'/> </controller> <controller type='pci' index='5' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='5' port='0xc'/> <alias name='pci.5'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x4'/> </controller> <controller type='pci' index='6' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='6' port='0xd'/> <alias name='pci.6'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x5'/> </controller> <controller type='pci' index='7' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='7' port='0xe'/> <alias name='pci.7'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x6'/> </controller> <controller type='pci' index='8' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='8' port='0xf'/> <alias name='pci.8'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x7'/> </controller> <controller type='pci' index='9' model='pcie-to-pci-bridge'> <model name='pcie-pci-bridge'/> <alias name='pci.9'/> <address type='pci' domain='0x0000' bus='0x01' slot='0x00' function='0x0'/> </controller> <controller type='virtio-serial' index='0'> <alias name='virtio-serial0'/> <address type='pci' domain='0x0000' bus='0x02' slot='0x00' function='0x0'/> </controller> <controller type='sata' index='0'> <alias name='ide'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x1f' function='0x2'/> </controller> <controller type='usb' index='0' model='ich9-ehci1'> <alias name='usb'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x07' function='0x7'/> </controller> <controller type='usb' index='0' model='ich9-uhci1'> <alias name='usb'/> <master startport='0'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x07' function='0x0' multifunction='on'/> </controller> <controller type='usb' index='0' model='ich9-uhci2'> <alias name='usb'/> <master startport='2'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x07' function='0x1'/> </controller> <controller type='usb' index='0' model='ich9-uhci3'> <alias name='usb'/> <master startport='4'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x07' function='0x2'/> </controller> <interface type='bridge'> <mac address='52:54:00:da:1a:0e'/> <source bridge='br0'/> <target dev='vnet1'/> <model type='e1000'/> <alias name='net0'/> <address type='pci' domain='0x0000' bus='0x09' slot='0x01' function='0x0'/> </interface> <serial type='pty'> <source path='/dev/pts/0'/> <target type='isa-serial' port='0'> <model name='isa-serial'/> </target> <alias name='serial0'/> </serial> <console type='pty' tty='/dev/pts/0'> <source path='/dev/pts/0'/> <target type='serial' port='0'/> <alias name='serial0'/> </console> <channel type='unix'> <source mode='bind' path='/var/lib/libvirt/qemu/channel/target/domain-2-hiveos/org.qemu.guest_agent.0'/> <target type='virtio' name='org.qemu.guest_agent.0' state='disconnected'/> <alias name='channel0'/> <address type='virtio-serial' controller='0' bus='0' port='1'/> </channel> <input type='tablet' bus='usb'> <alias name='input0'/> <address type='usb' bus='0' port='1'/> </input> <input type='mouse' bus='ps2'> <alias name='input1'/> </input> <input type='keyboard' bus='ps2'> <alias name='input2'/> </input> <audio id='1' type='none'/> <hostdev mode='subsystem' type='pci' managed='yes'> <driver name='vfio'/> <source> <address domain='0x0000' bus='0x0a' slot='0x00' function='0x0'/> </source> <alias name='hostdev0'/> <address type='pci' domain='0x0000' bus='0x04' slot='0x00' function='0x0' multifunction='on'/> </hostdev> <hostdev mode='subsystem' type='pci' managed='yes'> <driver name='vfio'/> <source> <address domain='0x0000' bus='0x06' slot='0x00' function='0x0'/> </source> <alias name='hostdev1'/> <address type='pci' domain='0x0000' bus='0x05' slot='0x00' function='0x0' multifunction='on'/> </hostdev> <hostdev mode='subsystem' type='pci' managed='yes'> <driver name='vfio'/> <source> <address domain='0x0000' bus='0x0d' slot='0x00' function='0x0'/> </source> <alias name='hostdev2'/> <address type='pci' domain='0x0000' bus='0x06' slot='0x00' function='0x0' multifunction='on'/> </hostdev> <hostdev mode='subsystem' type='pci' managed='yes'> <driver name='vfio'/> <source> <address domain='0x0000' bus='0x0a' slot='0x00' function='0x1'/> </source> <alias name='hostdev3'/> <address type='pci' domain='0x0000' bus='0x04' slot='0x00' function='0x1'/> </hostdev> <hostdev mode='subsystem' type='pci' managed='yes'> <driver name='vfio'/> <source> <address domain='0x0000' bus='0x06' slot='0x00' function='0x1'/> </source> <alias name='hostdev4'/> <address type='pci' domain='0x0000' bus='0x05' slot='0x00' function='0x1'/> </hostdev> <hostdev mode='subsystem' type='pci' managed='yes'> <driver name='vfio'/> <source> <address domain='0x0000' bus='0x0d' slot='0x00' function='0x1'/> </source> <alias name='hostdev5'/> <address type='pci' domain='0x0000' bus='0x06' slot='0x00' function='0x1'/> </hostdev> <memballoon model='none'/> </devices> <seclabel type='dynamic' model='dac' relabel='yes'> <label>+0:+100</label> <imagelabel>+0:+100</imagelabel> </seclabel> </domain> If i delete one of the Cards(+multifunction) the VM will start. Doen't matter which one. (I also don't need the multifunction as well, for a VM startup) Or isn't it possible to start a VM with 3 GPU on Bus? (I need the Card on "bus", because the card values on "slot" are all the same (temperatures, hashrate, etc) *edit* and thank you @ghost82 ! i saw you post post and try to understand how the adressing fpr passthroug works. but i dont get it.
  22. ca 40W im Spinup ca 25W im Spindown