cpu

Members
  • Posts

    36
  • Joined

  • Last visited

Everything posted by cpu

  1. You need to attach terminal and first get two packages that are missing - wget and either vi/vim/nano whatever you preffer. Then add repositories as per Installation — Proxmox Backup 3.0.3-1 documentation and remove enterprise Once you've got those added you can simply apt update && apt dist-upgrade and restart container
  2. @mguttdo you know by any chance if PCIe NVME adapter can cause lack of >c2 state? Those PCIe cards can differ from each other as I have 2 x 970 Evo and I don't want to ditch mirror for my cache pool yet my system had no problems in past to go as low as C8 -> i5-7500 with Fujitsu D3401 motherboard.
  3. @biggiesize looks like something is broken in latest gluetunvpn as on same config it does not start / nothing in logs either. Went back to pr-1268 so previous tag and it's all working again - can you have a look? As for my config - regular PIA via OpenVPN nothing special added.
  4. Anone has tips how to run go2rtc to use webrtc along frigate 0.12b1 ? I'm struggling with that.
  5. According to docs privileged is only needed when passing through Edge TPU. Have you tired without this option enabled as Frigate works just fine if you pass usb device. I will be able to test it after xmas. Also port is missing and icon - any chance for fixes @roflcoopter? You're maintainer of the project right? 🙂
  6. Why there are no default permissions set on postgres container? As it would simplyfi life to have 99:100 rather than non existing group - @Flight777can you have a look on that?
  7. @JorgeByou may check my post -> and consider updating power draw for different cards.
  8. I bought "Die-Wu ASM1166 6 port SATA PCIEx2 adapter from aliexpress" it has red heatsink and red SATA ports on back of card. In mean time ordered ECS06 just in case and to be honest it's not always an easy task to upgrade this ASM1166. Even with SilverStone I had couple failes when software didn't recognize my card properly, found out that I already had latest firmware on ECS06 but flashed in anyway - if there are any problems just start flashing software once again until it'll say "It can take some time to detect devices - please be patient". So let's dive into Die-Wu -> it was not recognized (used same software that came with ECS06) - and firmware was outdated from 2019, but I managed to flash it without any major issues - now it has same firmware (you can check this on POST screen). Currently I need to figure out idle power consumption with NVME disk installed - once I have that data I'll check both cards power consumption and C8 power states. EDIT#1 With Die-Wu power consumption goes up from ~ 8 W to ~ 13 W on average measured on wall plug. Way more than mentioned 2 W. I used powertop tunning at beggining then run powertop --auto-tune just to make sure - we're loosing C8. EDIT#2 With SilverStone ECS06 installed I'm reading same consumption as on Die-Wu - ~ 13 W maybe a 0.1-0.2 W less - but this proves that actual power consumption is around 4 W not 2 W. EDIT#3 So I've got 9211-8i flashed in IT mode and well this thing itself without HDD connected draws 8.5 W !! So power consumption goes up from 8 W without any card to 16.5 W in idle - same Unraid, same plugins, same setup. This is insane to be honest. I've got same measuring device - exactly same setup. And system takes 8 W and HBA card more than that...
  9. New build to replace my power hungry Ryzen 1600 AF + B450 mATX -> I used Fujitsu D3402-B2 + Core i5-7500 + 16 GB RAM and Corsair RM550x PSU currently idling at 7.2 W average after running powertop tune. Will report back once ASM1166 will arrive and 970 Pro Evo NVME. I've seen that ASM1166 requires firmware upgrade to properly support sleep states. Summary I've reduced usage from 70 W to 7.2 W so far but realistically I'm aiming at 11-12 W with fully deployed system.
  10. Once it'll arrive I'll post info here for sake of anyone going that way to reduce power consumption -> HBA to ASM1166.
  11. I'm planning to swap from HP H220 (LSI 9205i IT mode) and move to ASM1166 6xSATA card (https://pl.aliexpress.com/i/4001034338141.html) I assume moving array from old controller to new isn't going to be straight forward?
  12. Last resort help is to swap HBA card to PICE-SATA with ASM1166 you get 6xSATA without multiplexer so really good performance and power usage compared to HBA card is 4 times less. It's also way cheeper - 30 EUR on aliexpress.
  13. But once you create share with cache set to only games will sit on SSD which have btrfs
  14. Trying to utilize this with 1660 Super & Ryzen 1600AF - it's super slow... AC Origins 25 mins and it's still launching - games are on SSD share, shaders are done, using 5 GHz wifi. Should I pin some cores for this container exclusively? Memory is not an issue as I have 30 GB free...
  15. Same here - everything was working fine for long time and crash broke everything.
  16. OK so job's done I beleive - 3rd check finished without errors and RAM is set to 2400 MHz - thanks for your help @JorgeB 2022-01-24, 05:16:34 14 hr, 34 min, 50 sec 152.4 MB/s OK 0
  17. Thanks for reply, so second parity check finished with exactly same results, I just starter 3rd check so expect this to end tomorrow. 2022-01-23, 09:28:31 14 hr, 35 min, 32 sec 152.3 MB/s OK 475953 2022-01-21, 05:21:24 14 hr, 33 min, 47 sec 152.6 MB/s OK 475953 About RAM - I've got 2 x Micron MTA18ASF4G72AZ-3G2 32 GB UDIMM ECC - 3200 MHz so according to your post I should set 2400 MHz as it's DUAL RANK with 2nd gen Ryzen (Zen+) CPU - is it correct?
  18. Hello, I've recently had a chance to shutdown my Unraid server (due to some electrical maintanance) and it was a time to do parity check - it ended with: Event: Unraid Parity check Subject: Notice [SIGMA] - Parity check finished (475953 errors) Description: Duration: 14 hours, 33 minutes, 47 seconds. Average speed: 152.6 MB/s Importance: warning So far everything is running stable, no problems at all but I decided to run another parity check to see if errors are gone and so far I've got: Total size: 8 TB Elapsed time: 3 hours, 43 minutes Current position: 2.09 TB (26.2 %) Estimated speed: 143.7 MB/sec Estimated finish: 11 hours, 25 minutes Sync errors corrected: 475915 I'm attaching diag report, my only cue is LSI HBA controller - shouldn't be overheating especially now when it's winter time and this room is around 18 degC ambient + 2 front fans are pushing air toward rear of case. Anyone has any idea what to do next - should I run 3rd parity check? I know that my HDDs are 8TB IronWolf which is problematic with unraid >6.9 but I've disabled low current spinup and EPC and since that time I had 0 erros. Also all previous checks were 0 with same drive setup. I have UPS, shutdowns are clean, I suspected that might be some errors during this first parity check as there was unclean shutdown in past but I had hope that during this last check there'll be no errors. sigma-diagnostics-20220122-2238.zip
  19. I'm struggling to passthrough on my Quadro P1000 (primary GPU) to Win 10 machine. Any help appreciated. I've dumped vbios with Space Invader video but even with vbios I've got code 43. Does Nvidia block GPU passthrough for Quadro family? I've used at strt of array: echo 0 > /sys/class/vtconsole/vtcon0/bind echo 0 > /sys/class/vtconsole/vtcon1/bind echo efi-framebuffer.0 > /sys/bus/platform/drivers/efi-framebuffer/unbind I've got PCIe ACS override set to 'Both' IOMMU groups: VM template: <?xml version='1.0' encoding='UTF-8'?> <domain type='kvm'> <name>Win10</name> <uuid>1d260a9d-6f57-3777-730e-0ac7ba49f075</uuid> <metadata> <vmtemplate xmlns="unraid" name="Windows 10" icon="windows.png" os="windows10"/> </metadata> <memory unit='KiB'>8388608</memory> <currentMemory unit='KiB'>8388608</currentMemory> <memoryBacking> <nosharepages/> </memoryBacking> <vcpu placement='static'>4</vcpu> <cputune> <vcpupin vcpu='0' cpuset='4'/> <vcpupin vcpu='1' cpuset='10'/> <vcpupin vcpu='2' cpuset='5'/> <vcpupin vcpu='3' cpuset='11'/> </cputune> <os> <type arch='x86_64' machine='pc-q35-4.1'>hvm</type> <loader readonly='yes' type='pflash'>/usr/share/qemu/ovmf-x64/OVMF_CODE-pure-efi.fd</loader> <nvram>/etc/libvirt/qemu/nvram/1d260a9d-6f57-3777-730e-0ac7ba49f075_VARS-pure-efi.fd</nvram> </os> <features> <acpi/> <apic/> <hyperv> <relaxed state='on'/> <vapic state='on'/> <spinlocks state='on' retries='8191'/> <vendor_id state='on' value='none'/> </hyperv> </features> <cpu mode='host-passthrough' check='none' migratable='on'> <topology sockets='1' dies='1' cores='2' threads='2'/> <cache mode='passthrough'/> <feature policy='require' name='topoext'/> </cpu> <clock offset='localtime'> <timer name='hypervclock' present='yes'/> <timer name='hpet' present='no'/> </clock> <on_poweroff>destroy</on_poweroff> <on_reboot>restart</on_reboot> <on_crash>restart</on_crash> <devices> <emulator>/usr/local/sbin/qemu</emulator> <disk type='file' device='disk'> <driver name='qemu' type='raw' cache='writeback'/> <source file='/mnt/user/domains/Windows 10/vdisk1.img'/> <target dev='hdc' bus='virtio'/> <boot order='1'/> <address type='pci' domain='0x0000' bus='0x04' slot='0x00' function='0x0'/> </disk> <disk type='file' device='cdrom'> <driver name='qemu' type='raw'/> <source file='/mnt/user/isos/virtio-win-0.1.208-1.iso'/> <target dev='hdb' bus='sata'/> <readonly/> <address type='drive' controller='0' bus='0' target='0' unit='1'/> </disk> <controller type='usb' index='0' model='qemu-xhci' ports='15'> <address type='pci' domain='0x0000' bus='0x00' slot='0x07' function='0x0'/> </controller> <controller type='pci' index='0' model='pcie-root'/> <controller type='pci' index='1' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='1' port='0x8'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x0' multifunction='on'/> </controller> <controller type='pci' index='2' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='2' port='0x9'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x1'/> </controller> <controller type='pci' index='3' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='3' port='0xa'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x2'/> </controller> <controller type='pci' index='4' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='4' port='0xb'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x3'/> </controller> <controller type='pci' index='5' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='5' port='0xc'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x4'/> </controller> <controller type='pci' index='6' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='6' port='0xd'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x5'/> </controller> <controller type='pci' index='7' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='7' port='0xe'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x6'/> </controller> <controller type='pci' index='8' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='8' port='0xf'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x01' function='0x7'/> </controller> <controller type='pci' index='9' model='pcie-root-port'> <model name='pcie-root-port'/> <target chassis='9' port='0x10'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x02' function='0x0'/> </controller> <controller type='pci' index='10' model='pcie-to-pci-bridge'> <model name='pcie-pci-bridge'/> <address type='pci' domain='0x0000' bus='0x01' slot='0x00' function='0x0'/> </controller> <controller type='virtio-serial' index='0'> <address type='pci' domain='0x0000' bus='0x02' slot='0x00' function='0x0'/> </controller> <controller type='sata' index='0'> <address type='pci' domain='0x0000' bus='0x00' slot='0x1f' function='0x2'/> </controller> <interface type='bridge'> <mac address='52:54:00:56:b0:c0'/> <source bridge='br0'/> <model type='virtio-net'/> <address type='pci' domain='0x0000' bus='0x03' slot='0x00' function='0x0'/> </interface> <serial type='pty'> <target type='isa-serial' port='0'> <model name='isa-serial'/> </target> </serial> <console type='pty'> <target type='serial' port='0'/> </console> <channel type='unix'> <target type='virtio' name='org.qemu.guest_agent.0'/> <address type='virtio-serial' controller='0' bus='0' port='1'/> </channel> <input type='tablet' bus='usb'> <address type='usb' bus='0' port='1'/> </input> <input type='mouse' bus='ps2'/> <input type='keyboard' bus='ps2'/> <audio id='1' type='none'/> <hostdev mode='subsystem' type='pci' managed='yes'> <driver name='vfio'/> <source> <address domain='0x0000' bus='0x08' slot='0x00' function='0x0'/> </source> <address type='pci' domain='0x0000' bus='0x05' slot='0x00' function='0x0' multifunction='on'/> </hostdev> <hostdev mode='subsystem' type='pci' managed='yes'> <driver name='vfio'/> <source> <address domain='0x0000' bus='0x08' slot='0x00' function='0x1'/> </source> <address type='pci' domain='0x0000' bus='0x05' slot='0x00' function='0x1'/> </hostdev> <memballoon model='none'/> </devices> </domain> Diagnostics: sigma-diagnostics-20211227-1844.zip
  20. What about ownership? Have you tried exporting PUID and PGID ?
  21. Just set another PATH -> /mnt/disks/[backup] <- I've got unassigned HDD dedicated for backups and it's location is inside /mnt/disks
  22. Cheer for that container! Small sugestion - in your template source directory should be set to Read Only as there's no point to have this as Read/Write with full access (especially that someone may use /mnt/user instead of separate RO mount points)
  23. At that time I created new share and as I remember I changed appdata settings to export as samba share with r/w access for my super user. Server was unresponsive all time during file transfer - basically during copy operation when I try to do anything else ie. delete folder, access folder with lots of files I've got timeout. Even if folder I want to delete or access is on different hdd (part of array), I tought that it might be smb issue as this is 6.10rc2 but well when I went to terminal and tried to delete folder with mc I had timeout too, then I tried winscp and same story.
  24. diagnostics attached sigma-diagnostics-20211121-1557.zip
  25. Is it normal that when I start moving a lot of data from one share to another I can't access anything via samba or scp or even in terminal? I started moving media library to /data share for atomic moves and hardlinks and wanted to check few things in mean time - anyway I try to access data I just wait, wait, wait and eventually I've got timeout. I know parity and other two drive are involved but with Ryzen 5 and 64 GB ECC I would assume at least some reasonable access time to data should be possible. // No issues with cables etc as I checked SMART for those errors, all drives are attached to LSI controller in HBA mode.