crcobb Posted March 20, 2016 Share Posted March 20, 2016 I'm upgrading a server that has been running unRaid since version 5 and upgrading from 6.1.9. Upgrading to beta19 the server takes 2 hours to boot. I have attached a diagnostics file and a copy of the dmesg. The syslog in the diagnostics doesn't have correct timestamps for all the boot related messages. Looking at the boot messages the delay appears to be when identifying the USB ports. In testing I have unplugged the only USB items, mouse and keyboard, so the the only item still plugged in is the boot flash drive. I seem to have basic functionality but still looking at docker and my VM's. nas-diagnostics-20160320-0712.zip dmesg.zip Link to comment
bigjme Posted March 20, 2016 Share Posted March 20, 2016 Can't see my PCI device to passthrough in VM manager under 'Other PCI Devices' even though it is detected fine if i run 'lspci' I tried with and without PCI ACS Override on, no luck. 09:00.0 SATA controller: ASMedia Technology Inc. ASM1062 Serial ATA Controller (rev 01) 0a:00.0 USB controller: Etron Technology, Inc. EJ168 USB 3.0 Host Controller (rev 01) You have stubbed them in the boot right? If not, please read the first post in the beta 18 topic showing how to assigned those devices Link to comment
saskwatch Posted March 20, 2016 Share Posted March 20, 2016 My server has crashed 5 times since to this beta. It was working fine on stable and beta 18. I have posted diagnostics a few posts ago. I do not know how helpful that is, since I have to run diagnostics after reboot. Very frustrating. I am running: Asus Z170 Deluxe Motherboard (Skylake) Core i7 (not overclocked) 32GB Corsair Vengance 4 WD 3TB Reds 2 WD 4TB Reds 2 Sandisk 1TB SSDs Link to comment
Squid Posted March 20, 2016 Share Posted March 20, 2016 Question: Why don't we set the default mode for all docker paths to be RW:Slave instead of RW. Quick and dirty tests show that there isn't any issues (when I switched my local paths to be RW Slave), and this switch only would solve problems in the long term (if there are no ill effects) due to quite a number of users running their appdata and/or media shares outside of unRaid's control (Unassigned Devices). Or put another way, what downsides are there to running local (unRaid controlled shares) with RW Slave. And a related feature request, I would also like to see a Read-Only Slave option available Link to comment
jonp Posted March 20, 2016 Share Posted March 20, 2016 Question: Why don't we set the default mode for all docker paths to be RW:Slave instead of RW. Quick and dirty tests show that there isn't any issues (when I switched my local paths to be RW Slave), and this switch only would solve problems in the long term (if there are no ill effects) due to quite a number of users running their appdata and/or media shares outside of unRaid's control (Unassigned Devices). Or put another way, what downsides are there to running local (unRaid controlled shares) with RW Slave. And a related feature request, I would also like to see a Read-Only Slave option available I tried this for the appdata folder for Plex and it doesn't work. Shared mode seems to have issues with that where regular mode does not. Link to comment
Squid Posted March 20, 2016 Share Posted March 20, 2016 Question: Why don't we set the default mode for all docker paths to be RW:Slave instead of RW. Quick and dirty tests show that there isn't any issues (when I switched my local paths to be RW Slave), and this switch only would solve problems in the long term (if there are no ill effects) due to quite a number of users running their appdata and/or media shares outside of unRaid's control (Unassigned Devices). Or put another way, what downsides are there to running local (unRaid controlled shares) with RW Slave. And a related feature request, I would also like to see a Read-Only Slave option available I tried this for the appdata folder for Plex and it doesn't work. Shared mode seems to have issues with that where regular mode does not. Fair enough. I'd still like to see RO:Slave, as I (any many other users) do not give RW access to paths where the app doesn't require it. Link to comment
piotrasd Posted March 20, 2016 Share Posted March 20, 2016 Hi After login over FTP im always getting [ 369.831674] vsftpd[18787]: segfault at 0 ip 00002aeeb9980e2a sp 00007ffd221ff098 error 4 in libc-2.23.so[2aeeb98e2000+1c0000] [ 374.350005] vsftpd[18833]: segfault at 0 ip 00002ad7ea42ee2a sp 00007ffeafd81418 error 4 in libc-2.23.so[2ad7ea390000+1c0000] [ 380.910259] vsftpd[18904]: segfault at 0 ip 00002b7b6d8eee2a sp 00007ffd4bdaaa98 error 4 in libc-2.23.so[2b7b6d850000+1c0000] [ 396.746897] vsftpd[19151]: segfault at 0 ip 00002b7e89b2ee2a sp 00007fffb6490ca8 error 4 in libc-2.23.so[2b7e89a90000+1c0000] [ 399.897735] vsftpd[19181]: segfault at 0 ip 00002b58c40e4e2a sp 00007ffe83003b78 error 4 in libc-2.23.so[2b58c4046000+1c0000] This same was in b18 @limetech @jnop please try yourself connect and download/upload some bigger file. (thx) Link to comment
eschultz Posted March 20, 2016 Share Posted March 20, 2016 Hi After login over FTP im always getting [ 369.831674] vsftpd[18787]: segfault at 0 ip 00002aeeb9980e2a sp 00007ffd221ff098 error 4 in libc-2.23.so[2aeeb98e2000+1c0000] [ 374.350005] vsftpd[18833]: segfault at 0 ip 00002ad7ea42ee2a sp 00007ffeafd81418 error 4 in libc-2.23.so[2ad7ea390000+1c0000] [ 380.910259] vsftpd[18904]: segfault at 0 ip 00002b7b6d8eee2a sp 00007ffd4bdaaa98 error 4 in libc-2.23.so[2b7b6d850000+1c0000] [ 396.746897] vsftpd[19151]: segfault at 0 ip 00002b7e89b2ee2a sp 00007fffb6490ca8 error 4 in libc-2.23.so[2b7e89a90000+1c0000] [ 399.897735] vsftpd[19181]: segfault at 0 ip 00002b58c40e4e2a sp 00007ffe83003b78 error 4 in libc-2.23.so[2b58c4046000+1c0000] This same was in b18 @limetech @jnop please try yourself connect and download/upload some bigger file. (thx) I haven't been able to get it to crash yet. Transferred a 3GB file in both directions to the cache folder. And then again to user0, both directions. No errors or segfaults. All I can think of is some installed plugin might have downgraded a package (e.g. openssl, libcap) that vsftpd was dependent on? Please submit diagnostics here (or via the webgui's feedback system). Link to comment
gfjardim Posted March 20, 2016 Share Posted March 20, 2016 Question: Why don't we set the default mode for all docker paths to be RW:Slave instead of RW. Quick and dirty tests show that there isn't any issues (when I switched my local paths to be RW Slave), and this switch only would solve problems in the long term (if there are no ill effects) due to quite a number of users running their appdata and/or media shares outside of unRaid's control (Unassigned Devices). Or put another way, what downsides are there to running local (unRaid controlled shares) with RW Slave. And a related feature request, I would also like to see a Read-Only Slave option available I tried this for the appdata folder for Plex and it doesn't work. Shared mode seems to have issues with that where regular mode does not. Fair enough. I'd still like to see RO:Slave, as I (any many other users) do not give RW access to paths where the app doesn't require it. By the way, RW/Slave only works with paths under /mnt, because that is the only path unRAID shares by default. It's a workaround for a specific problem, not a solution for all volume mounts. Link to comment
spylex Posted March 20, 2016 Share Posted March 20, 2016 You have stubbed them in the boot right? If not, please read the first post in the beta 18 topic showing how to assigned those devices I feel silly now, thank you, that got it. Comma separated works for multiple devices. Does this only work with OVMF or SeaBIOS too? Link to comment
Bjonness406 Posted March 20, 2016 Share Posted March 20, 2016 This bug still exist in 6.2 b19, any chance you can take a look at it and fix it before the next beta? http://lime-technology.com/forum/index.php?topic=40826.0 Link to comment
Squid Posted March 21, 2016 Share Posted March 21, 2016 Getting varying amounts of these HTML warnings on every tab in dynamix. Always on line #1 (<!DOCTYPE HTML>. Page reference at the right will change depending upon which tab you're on (obviously I took the screen shot while I was in the Apps tab) (using Chrome 49.0.2623.87 m) unRaid 6.1.x was always clear of any / all errors / warnings Link to comment
bubbaQ Posted March 21, 2016 Share Posted March 21, 2016 Don't know if this is new to 6.2Beta, but when exploring a disk device, and selecting the "Attributes" tab, it is not checking for valid SMART data being returned... it just parses and throws up whatever it gets back: Vendor: Seagate Product: ST4000DM000-1F21 Revision: R001 User Capacity: 4,000,787,030,016 Bytes [4.00 TB] Logical Block Size: 512 Bytes Rotation Rate: 10000 Rpm Logical Unit Id: 0x001b4d2000000000 Serial Number: Z3052DR8 Device Type: Disk Transport Protocol: Fibre Channel (FCP-2) Local Time Is: Sun Mar 20 21:59:02 2016 EDT SMART Support Is: Available - Device Has SMART Capability. SMART Support Is: Enabled Temperature Warning: Disabled Or Not Supported === START OF READ SMART DATA SECTION === Current Drive Temperature: 30 C Drive Trip Temperature: 25 C Manufactured In Week 30 Of Year 2002 Specified Cycle Count Over Device Lifetime: 4278190080 Accumulated Start-stop Cycles: 256 Elements In Grown Defect List: 0 The "Identify" tab is correct, and the "Capabilities" tab correctly indicates it did not get capabilities data from the drive. Link to comment
Frank1940 Posted March 21, 2016 Share Posted March 21, 2016 Don't know if this is new to 6.2Beta, but when exploring a disk device, and selecting the "Attributes" tab, it is not checking for valid SMART data being returned... it just parses and throws up whatever it gets back: Vendor: Seagate Product: ST4000DM000-1F21 Revision: R001 User Capacity: 4,000,787,030,016 Bytes [4.00 TB] Logical Block Size: 512 Bytes Rotation Rate: 10000 Rpm Logical Unit Id: 0x001b4d2000000000 Serial Number: Z3052DR8 Device Type: Disk Transport Protocol: Fibre Channel (FCP-2) Local Time Is: Sun Mar 20 21:59:02 2016 EDT SMART Support Is: Available - Device Has SMART Capability. SMART Support Is: Enabled Temperature Warning: Disabled Or Not Supported === START OF READ SMART DATA SECTION === Current Drive Temperature: 30 C Drive Trip Temperature: 25 C Manufactured In Week 30 Of Year 2002 Specified Cycle Count Over Device Lifetime: 4278190080 Accumulated Start-stop Cycles: 256 Elements In Grown Defect List: 0 The "Identify" tab is correct, and the "Capabilities" tab correctly indicates it did not get capabilities data from the drive. It seems to be working properly on my Test Bed Server. You might trying clearing your browser cache. Link to comment
bubbaQ Posted March 21, 2016 Share Posted March 21, 2016 It seems to be working properly on my Test Bed Server. You might trying clearing your browser cache. No, I correctly reported it as a bug. I'm sure it does work for a lot of people, but if a system (as does mine) returns SMART data differently than what is expected, the UI should detect it is not what was expected, rather than parse whatever text comes back and show garbage in the UI. That's basic input validation. Link to comment
Squid Posted March 21, 2016 Share Posted March 21, 2016 Posted a defect report over here http://lime-technology.com/forum/index.php?topic=47696.0, where its going to look like you've completely lost all of your appdata (but you haven't) for docker apps in most circumstances (for those who've upgraded from 6.1.x) due to how dockerMan operates if you happen to reinstall an application Link to comment
henriquelb Posted March 21, 2016 Share Posted March 21, 2016 I got several lines of the same error: Mar 21 00:16:55 Inefavel kernel: pcieport 0000:00:07.0: AER: Multiple Uncorrected (Non-Fatal) error received: id=0000 Mar 21 00:16:55 Inefavel kernel: pcieport 0000:00:07.0: can't find device of ID0000 00:07.0 is a graphics card GTX 960, however I have another GTX 960 exactly same model, running in a exactly the same VM configuration giving me no errors. As far I'm aware, I never saw it in past releases. syslog.zip Link to comment
dAigo Posted March 21, 2016 Share Posted March 21, 2016 Known Issues 6.2 beta18 GPU Pass Through CPU Overhead Some users are reporting an overly high amount of CPU usage when running Windows guest VMs with GPU pass through and opening applications that engage 3D graphics, video, or audio. The usage doesn't appear to actually affect the performance of the guest, but the utilization reported on the host definitely doesn't appear to match the guest. We are actively investigating this issue, but know that it shouldn't affect functionality of your system or usability, just how much your CPU reports to utilize on the host. Still an active issue on my end. The attached diagnostic was collected, while a video was running and the host showed 99% on all cores, that are attached to the vm. Guest load was around 4-10% on any core... Your wording makes it almost sound, as if you may suspecrt an issue with the "reported load". According to my cpu-temps (see chart) and powerusage measured on the wall, the Host-Load is accurate. During the 10 minute playback, cpu-temp was at 50°C. At around 5 minutes into the test, I startet a cpu stresstest. Guest CPU went to 90%+ but temp stayed at ~50°C. 50°C is around max, because I removed all overclocking and the cpu is watercooled. And powerusage goes from ~100watts to ~150watts. I'll stay on 6.2 for this week, if you need additional info or if I should try anything, feel free to ask. Rolling back to 6.1 is a pita, because I am testing the nvme support for cache drives... Rolling back means a lot of files beeing moved. Just a quick update: I can reproduce that problem on another VM, that is also Win10 (1511), also SeaBIOS but has NO GPU passthrough... However, I can't reproduce it on VMs that run OVMF and Win10 (1507)/Server 2012 R2, regardless of passthrough or not. At least for me, it seams SeaBIOS and/or Win10 (1511) has something to do with it. I could add another daig. with the ovfm vms running, if the first one didn't help. Link to comment
bonienl Posted March 21, 2016 Share Posted March 21, 2016 Getting varying amounts of these HTML warnings on every tab in dynamix. Always on line #1 (<!DOCTYPE HTML>. Page reference at the right will change depending upon which tab you're on (obviously I took the screen shot while I was in the Apps tab) (using Chrome 49.0.2623.87 m) unRaid 6.1.x was always clear of any / all errors / warnings Chrome somehow doesn't fully like the newer version 4.5.0 of font-awesome. unRAID v6.1.9 is using version 4.3.0... Link to comment
bonienl Posted March 21, 2016 Share Posted March 21, 2016 It seems to be working properly on my Test Bed Server. You might trying clearing your browser cache. No, I correctly reported it as a bug. I'm sure it does work for a lot of people, but if a system (as does mine) returns SMART data differently than what is expected, the UI should detect it is not what was expected, rather than parse whatever text comes back and show garbage in the UI. That's basic input validation. New smartctl version 6.4 is used, looks like it behaves differently. Any chance of posting a raw output of your findings? Link to comment
sparklyballs Posted March 21, 2016 Share Posted March 21, 2016 This started in beta18, happened a couple of times with that and it just now happened with beta19. after a period of time opensuse VM reports this in its log and is dead. Domain id=1 is tainted: high-privileges Domain id=1 is tainted: host-cpu ATTACHED DIAGNOSTICS unraid-nas-diagnostics-20160321-0852.zip Link to comment
sparklyballs Posted March 21, 2016 Share Posted March 21, 2016 This started in beta18, happened a couple of times with that and it just now happened with beta19. after a period of time opensuse VM reports this in its log and is dead. Domain id=1 is tainted: high-privileges Domain id=1 is tainted: host-cpu ATTACHED DIAGNOSTICS Not sure if it's related or not, but my sysload seems to be permanently stuck at 2.0 - ish where it should normally be around 0.1 ish at idle. Link to comment
piotrasd Posted March 21, 2016 Share Posted March 21, 2016 Hi After login over FTP im always getting [ 369.831674] vsftpd[18787]: segfault at 0 ip 00002aeeb9980e2a sp 00007ffd221ff098 error 4 in libc-2.23.so[2aeeb98e2000+1c0000] [ 374.350005] vsftpd[18833]: segfault at 0 ip 00002ad7ea42ee2a sp 00007ffeafd81418 error 4 in libc-2.23.so[2ad7ea390000+1c0000] [ 380.910259] vsftpd[18904]: segfault at 0 ip 00002b7b6d8eee2a sp 00007ffd4bdaaa98 error 4 in libc-2.23.so[2b7b6d850000+1c0000] [ 396.746897] vsftpd[19151]: segfault at 0 ip 00002b7e89b2ee2a sp 00007fffb6490ca8 error 4 in libc-2.23.so[2b7e89a90000+1c0000] [ 399.897735] vsftpd[19181]: segfault at 0 ip 00002b58c40e4e2a sp 00007ffe83003b78 error 4 in libc-2.23.so[2b58c4046000+1c0000] This same was in b18 @limetech @jnop please try yourself connect and download/upload some bigger file. (thx) I haven't been able to get it to crash yet. Transferred a 3GB file in both directions to the cache folder. And then again to user0, both directions. No errors or segfaults. All I can think of is some installed plugin might have downgraded a package (e.g. openssl, libcap) that vsftpd was dependent on? Please submit diagnostics here (or via the webgui's feedback system). Ok Thanks ! Link to comment
spylex Posted March 21, 2016 Share Posted March 21, 2016 I get this all the time in my VMs, i thought it was normal? Domain id=1 is tainted: high-privileges Domain id=1 is tainted: host-cpu Link to comment
ktbken Posted March 21, 2016 Share Posted March 21, 2016 I get this all the time in my VMs, i thought it was normal? Domain id=1 is tainted: high-privileges Domain id=1 is tainted: host-cpu Same here. It has happened with with all 6.1 and 6.2 versions I have used Link to comment
Recommended Posts