unRAID Server Release 6.2.0-beta19 Available


Recommended Posts

I'm upgrading a server that has been running unRaid since version 5 and upgrading from 6.1.9.  Upgrading to beta19 the server takes 2 hours to boot.  I have attached a diagnostics file and a copy of the dmesg.  The syslog in the diagnostics doesn't have correct timestamps for all the boot related messages.  Looking at the boot messages the delay appears to be when identifying the USB ports.  In testing I have unplugged the only USB items, mouse and keyboard, so the the only item still plugged in is the boot flash drive. 

I seem to have basic functionality but still looking at docker and my VM's.

nas-diagnostics-20160320-0712.zip

dmesg.zip

Link to comment
  • Replies 194
  • Created
  • Last Reply

Top Posters In This Topic

Can't see my PCI device to passthrough in VM manager under 'Other PCI Devices' even though it is detected fine if i run 'lspci'

 

I tried with and without PCI ACS Override on, no luck.

 

09:00.0 SATA controller: ASMedia Technology Inc. ASM1062 Serial ATA Controller (rev 01)

0a:00.0 USB controller: Etron Technology, Inc. EJ168 USB 3.0 Host Controller (rev 01)

 

You have stubbed them in the boot right? If not, please read the first post in the beta 18 topic showing how to assigned those devices

Link to comment

My server has crashed 5 times since to this beta.  It was working fine on stable and beta 18.  I have posted diagnostics a few posts ago.  I do not know how helpful that is, since I have to run diagnostics after reboot.  Very frustrating.  I am running:

 

  • Asus Z170 Deluxe Motherboard (Skylake)
  • Core i7 (not overclocked)
  • 32GB Corsair Vengance
  • 4 WD 3TB Reds
  • 2 WD 4TB Reds
  • 2 Sandisk 1TB SSDs

Link to comment

Question:

 

Why don't we set the default mode for all docker paths to be RW:Slave instead of RW.  Quick and dirty tests show that there isn't any issues (when I switched my local paths to be RW Slave), and this switch only would solve problems in the long term (if there are no ill effects) due to quite a number of users running their appdata and/or media shares outside of unRaid's control (Unassigned Devices).

 

Or put another way, what downsides are there to running local (unRaid controlled shares) with RW Slave.

 

 

And a related feature request, I would also like to see a Read-Only Slave option available

Link to comment

Question:

 

Why don't we set the default mode for all docker paths to be RW:Slave instead of RW.  Quick and dirty tests show that there isn't any issues (when I switched my local paths to be RW Slave), and this switch only would solve problems in the long term (if there are no ill effects) due to quite a number of users running their appdata and/or media shares outside of unRaid's control (Unassigned Devices).

 

Or put another way, what downsides are there to running local (unRaid controlled shares) with RW Slave.

 

 

And a related feature request, I would also like to see a Read-Only Slave option available

I tried this for the appdata folder for Plex and it doesn't work. Shared mode seems to have issues with that where regular mode does not.

Link to comment

Question:

 

Why don't we set the default mode for all docker paths to be RW:Slave instead of RW.  Quick and dirty tests show that there isn't any issues (when I switched my local paths to be RW Slave), and this switch only would solve problems in the long term (if there are no ill effects) due to quite a number of users running their appdata and/or media shares outside of unRaid's control (Unassigned Devices).

 

Or put another way, what downsides are there to running local (unRaid controlled shares) with RW Slave.

 

 

And a related feature request, I would also like to see a Read-Only Slave option available

I tried this for the appdata folder for Plex and it doesn't work. Shared mode seems to have issues with that where regular mode does not.

Fair enough.  I'd still like to see RO:Slave, as I (any many other users) do not give RW access to paths where the app doesn't require it.
Link to comment

Hi

After login over FTP im always getting

[  369.831674] vsftpd[18787]: segfault at 0 ip 00002aeeb9980e2a sp 00007ffd221ff098 error 4 in libc-2.23.so[2aeeb98e2000+1c0000]

[  374.350005] vsftpd[18833]: segfault at 0 ip 00002ad7ea42ee2a sp 00007ffeafd81418 error 4 in libc-2.23.so[2ad7ea390000+1c0000]

[  380.910259] vsftpd[18904]: segfault at 0 ip 00002b7b6d8eee2a sp 00007ffd4bdaaa98 error 4 in libc-2.23.so[2b7b6d850000+1c0000]

[  396.746897] vsftpd[19151]: segfault at 0 ip 00002b7e89b2ee2a sp 00007fffb6490ca8 error 4 in libc-2.23.so[2b7e89a90000+1c0000]

[  399.897735] vsftpd[19181]: segfault at 0 ip 00002b58c40e4e2a sp 00007ffe83003b78 error 4 in libc-2.23.so[2b58c4046000+1c0000]

 

This same was in b18 @limetech @jnop please try yourself connect and download/upload some bigger file. (thx)

Link to comment

Hi

After login over FTP im always getting

[  369.831674] vsftpd[18787]: segfault at 0 ip 00002aeeb9980e2a sp 00007ffd221ff098 error 4 in libc-2.23.so[2aeeb98e2000+1c0000]

[  374.350005] vsftpd[18833]: segfault at 0 ip 00002ad7ea42ee2a sp 00007ffeafd81418 error 4 in libc-2.23.so[2ad7ea390000+1c0000]

[  380.910259] vsftpd[18904]: segfault at 0 ip 00002b7b6d8eee2a sp 00007ffd4bdaaa98 error 4 in libc-2.23.so[2b7b6d850000+1c0000]

[  396.746897] vsftpd[19151]: segfault at 0 ip 00002b7e89b2ee2a sp 00007fffb6490ca8 error 4 in libc-2.23.so[2b7e89a90000+1c0000]

[  399.897735] vsftpd[19181]: segfault at 0 ip 00002b58c40e4e2a sp 00007ffe83003b78 error 4 in libc-2.23.so[2b58c4046000+1c0000]

 

This same was in b18 @limetech @jnop please try yourself connect and download/upload some bigger file. (thx)

 

I haven't been able to get it to crash yet.  Transferred a 3GB file in both directions to the cache folder.  And then again to user0, both directions.  No errors or segfaults. 

 

All I can think of is some installed plugin might have downgraded a package (e.g. openssl, libcap) that vsftpd was dependent on?  Please submit diagnostics here (or via the webgui's feedback system).

Link to comment

Question:

 

Why don't we set the default mode for all docker paths to be RW:Slave instead of RW.  Quick and dirty tests show that there isn't any issues (when I switched my local paths to be RW Slave), and this switch only would solve problems in the long term (if there are no ill effects) due to quite a number of users running their appdata and/or media shares outside of unRaid's control (Unassigned Devices).

 

Or put another way, what downsides are there to running local (unRaid controlled shares) with RW Slave.

 

 

And a related feature request, I would also like to see a Read-Only Slave option available

I tried this for the appdata folder for Plex and it doesn't work. Shared mode seems to have issues with that where regular mode does not.

Fair enough.  I'd still like to see RO:Slave, as I (any many other users) do not give RW access to paths where the app doesn't require it.

 

By the way, RW/Slave only works with paths under /mnt, because that is the only path unRAID shares by default. It's a workaround for a specific problem, not a solution for all volume mounts.

Link to comment

You have stubbed them in the boot right? If not, please read the first post in the beta 18 topic showing how to assigned those devices

 

I feel silly now, thank you, that got it. Comma separated works for multiple devices.

 

Does this only work with OVMF or SeaBIOS too?

Link to comment

Getting varying amounts of these HTML warnings on every tab in dynamix.  Always on line #1 (<!DOCTYPE HTML>.  Page reference at the right will change depending upon which tab you're on (obviously I took the screen shot while I was in the Apps tab) (using Chrome 49.0.2623.87 m)

 

unRaid 6.1.x was always clear of any / all errors / warnings

 

Untitled_zpsmrdgahvy.png

Link to comment

Don't know if this is new to 6.2Beta, but when exploring a disk device, and selecting the "Attributes" tab, it is not checking for valid SMART data being returned... it just parses and throws up whatever it gets back:

 

Vendor:	Seagate
Product:	ST4000DM000-1F21
Revision:	R001
User	Capacity:	4,000,787,030,016	Bytes	[4.00	TB]
Logical	Block	Size:	512	Bytes
Rotation	Rate:	10000	Rpm
Logical	Unit	Id:	0x001b4d2000000000
Serial	Number:	Z3052DR8
Device	Type:	Disk
Transport	Protocol:	Fibre	Channel	(FCP-2)
Local	Time	Is:	Sun	Mar	20	21:59:02	2016	EDT
SMART	Support	Is:	Available	-	Device	Has	SMART	Capability.
SMART	Support	Is:	Enabled
Temperature	Warning:	Disabled	Or	Not	Supported
===	START	OF	READ	SMART	DATA	SECTION	===
Current	Drive	Temperature:	30	C
Drive	Trip	Temperature:	25	C
Manufactured	In	Week	30	Of	Year	2002
Specified	Cycle	Count	Over	Device	Lifetime:	4278190080
Accumulated	Start-stop	Cycles:	256
Elements	In	Grown	Defect	List:	0

 

The "Identify" tab is correct, and the "Capabilities" tab correctly indicates it did not get capabilities data from the drive.

Link to comment

Don't know if this is new to 6.2Beta, but when exploring a disk device, and selecting the "Attributes" tab, it is not checking for valid SMART data being returned... it just parses and throws up whatever it gets back:

 

Vendor:	Seagate
Product:	ST4000DM000-1F21
Revision:	R001
User	Capacity:	4,000,787,030,016	Bytes	[4.00	TB]
Logical	Block	Size:	512	Bytes
Rotation	Rate:	10000	Rpm
Logical	Unit	Id:	0x001b4d2000000000
Serial	Number:	Z3052DR8
Device	Type:	Disk
Transport	Protocol:	Fibre	Channel	(FCP-2)
Local	Time	Is:	Sun	Mar	20	21:59:02	2016	EDT
SMART	Support	Is:	Available	-	Device	Has	SMART	Capability.
SMART	Support	Is:	Enabled
Temperature	Warning:	Disabled	Or	Not	Supported
===	START	OF	READ	SMART	DATA	SECTION	===
Current	Drive	Temperature:	30	C
Drive	Trip	Temperature:	25	C
Manufactured	In	Week	30	Of	Year	2002
Specified	Cycle	Count	Over	Device	Lifetime:	4278190080
Accumulated	Start-stop	Cycles:	256
Elements	In	Grown	Defect	List:	0

 

The "Identify" tab is correct, and the "Capabilities" tab correctly indicates it did not get capabilities data from the drive.

 

It seems to be working properly on my Test Bed Server.  You might trying clearing your browser cache.

Link to comment

It seems to be working properly on my Test Bed Server.  You might trying clearing your browser cache.

 

No, I correctly reported it as a bug.  I'm sure it does work for a lot of people, but if a system (as does mine) returns SMART data differently than what is expected, the UI should detect it is not what was expected, rather than parse whatever text comes back and show garbage in the UI.  That's basic input validation.

Link to comment

I got several lines of the same error:

Mar 21 00:16:55 Inefavel kernel: pcieport 0000:00:07.0: AER: Multiple Uncorrected (Non-Fatal) error received: id=0000

Mar 21 00:16:55 Inefavel kernel: pcieport 0000:00:07.0: can't find device of ID0000

 

00:07.0 is a graphics card GTX 960, however I have another GTX 960 exactly same model, running in a exactly the same VM configuration giving me no errors.

 

As far I'm aware, I never saw it in past releases.

syslog.zip

Link to comment

Known Issues 6.2 beta18

 

GPU Pass Through CPU Overhead

 

Some users are reporting an overly high amount of CPU usage when running Windows guest VMs with GPU pass through and opening applications that engage 3D graphics, video, or audio.  The usage doesn't appear to actually affect the performance of the guest, but the utilization reported on the host definitely doesn't appear to match the guest.

 

We are actively investigating this issue, but know that it shouldn't affect functionality of your system or usability, just how much your CPU reports to utilize on the host.

 

Still an active issue on my end.

The attached diagnostic was collected, while a video was running and the host showed 99% on all cores, that are attached to the vm.

Guest load was around 4-10% on any core...

 

Your wording makes it almost sound, as if you may suspecrt an issue with the "reported load".

According to my cpu-temps  (see chart) and powerusage measured on the wall, the Host-Load is accurate.

 

During the 10 minute playback, cpu-temp was at 50°C.

At around 5 minutes into the test, I startet a cpu stresstest. Guest CPU went to 90%+ but temp stayed at ~50°C.

50°C is around max, because I removed all overclocking and the cpu is watercooled.

 

And powerusage goes from ~100watts to ~150watts.

 

I'll stay on 6.2 for this week, if you need additional info or if I should try anything, feel free to ask.

Rolling back to 6.1 is a pita, because I am testing the nvme support for cache drives... Rolling back means a lot of files beeing moved.

 

Just a quick update:

I can reproduce that problem on another VM, that is also Win10 (1511), also SeaBIOS but has NO GPU passthrough...

However, I can't reproduce it on VMs that run OVMF and Win10 (1507)/Server 2012 R2, regardless of passthrough or not.

 

At least for me, it seams SeaBIOS and/or Win10 (1511) has something to do with it.

I could add another daig. with the ovfm vms running, if the first one didn't help.

Link to comment

Getting varying amounts of these HTML warnings on every tab in dynamix.  Always on line #1 (<!DOCTYPE HTML>.  Page reference at the right will change depending upon which tab you're on (obviously I took the screen shot while I was in the Apps tab) (using Chrome 49.0.2623.87 m)

 

unRaid 6.1.x was always clear of any / all errors / warnings

 

Untitled_zpsmrdgahvy.png

 

Chrome somehow doesn't fully like the newer version 4.5.0 of font-awesome. unRAID v6.1.9 is using version 4.3.0...

Link to comment

It seems to be working properly on my Test Bed Server.  You might trying clearing your browser cache.

 

No, I correctly reported it as a bug.  I'm sure it does work for a lot of people, but if a system (as does mine) returns SMART data differently than what is expected, the UI should detect it is not what was expected, rather than parse whatever text comes back and show garbage in the UI.  That's basic input validation.

 

New smartctl version 6.4 is used, looks like it behaves differently. Any chance of posting a raw output of your findings?

Link to comment

This started in beta18, happened a couple of times with that and it just now happened with beta19.

 

after a period of time opensuse VM reports this in its log and is dead.

 

Domain id=1 is tainted: high-privileges
Domain id=1 is tainted: host-cpu

 

 

 

ATTACHED DIAGNOSTICS

 

Not sure if it's related or not, but my sysload seems to be permanently stuck at 2.0 - ish where it should normally be around 0.1 ish at idle.

Link to comment

Hi

After login over FTP im always getting

[  369.831674] vsftpd[18787]: segfault at 0 ip 00002aeeb9980e2a sp 00007ffd221ff098 error 4 in libc-2.23.so[2aeeb98e2000+1c0000]

[  374.350005] vsftpd[18833]: segfault at 0 ip 00002ad7ea42ee2a sp 00007ffeafd81418 error 4 in libc-2.23.so[2ad7ea390000+1c0000]

[  380.910259] vsftpd[18904]: segfault at 0 ip 00002b7b6d8eee2a sp 00007ffd4bdaaa98 error 4 in libc-2.23.so[2b7b6d850000+1c0000]

[  396.746897] vsftpd[19151]: segfault at 0 ip 00002b7e89b2ee2a sp 00007fffb6490ca8 error 4 in libc-2.23.so[2b7e89a90000+1c0000]

[  399.897735] vsftpd[19181]: segfault at 0 ip 00002b58c40e4e2a sp 00007ffe83003b78 error 4 in libc-2.23.so[2b58c4046000+1c0000]

 

This same was in b18 @limetech @jnop please try yourself connect and download/upload some bigger file. (thx)

 

I haven't been able to get it to crash yet.  Transferred a 3GB file in both directions to the cache folder.  And then again to user0, both directions.  No errors or segfaults. 

 

All I can think of is some installed plugin might have downgraded a package (e.g. openssl, libcap) that vsftpd was dependent on?  Please submit diagnostics here (or via the webgui's feedback system).

 

Ok Thanks !

Link to comment
Guest
This topic is now closed to further replies.