jonp Posted August 8, 2015 Share Posted August 8, 2015 Here’s a nice btrfs overview which includes an interesting comparison with ZFS. He also mentions two reasons why we also don’t include ZFS in unRaid. First is that ZFS is real memory hog and would likely blow up typical small unRAID NAS servers. Second is our questionable ability to distribute ZFS along with the linux kernel. Beyond that, we think for our purposes btrfs is better anyway. We plan to leverage many btrfs features in future unRaid OS releases, especially snapshots! http://marc.merlins.org/linux/talks/Btrfs-LinuxCon2014/Btrfs.pdf Link to comment
prostuff1 Posted August 31, 2015 Share Posted August 31, 2015 I wants me a RAID0 cache pool setup ability in the unRAID web GUI... please and thank you Link to comment
limetech Posted August 31, 2015 Share Posted August 31, 2015 I wants me a RAID0 cache pool setup ability in the unRAID web GUI... please and thank you Posted this elsewhere: You can set up a raid-0 cache pool. When you click on the 'cache' device after array Started and cache is mounted, scroll down to 'Balance' section and you will see default Balance options are: -dconvert=raid1 -mconvert=raid1 You can change the 'raid level' by editing these options and click Balance. For example, to have data raid0 but leave metadata raid1 you could change to: -dconvert=raid0 -mconvert=raid1 You can monitor syslog to watch balance operation proceed. Note that, at present, the s/w will "revert" the default balance options back to: -dconvert=raid1 -mconvert=raid1 That is, we don't store your selection. Similarly if you add a device to an existing pool it will automatically kick off a balance with the default options. We plan to address this in a future release, but for now you can experiment with it. Link to comment
Lake-end Posted July 17, 2016 Share Posted July 17, 2016 I´m really sorry for resurrecting old thread. I´m planning on getting a second SSD for my cache pool. Is the default still raid1 in 6.1.9? If I go and change it, will it still revert back to it? If so, how can you prevent this? Link to comment
JorgeB Posted July 17, 2016 Share Posted July 17, 2016 I´m really sorry for resurrecting old thread. I´m planning on getting a second SSD for my cache pool. Is the default still raid1 in 6.1.9? If I go and change it, will it still revert back to it? If so, how can you prevent this? Default is still RAID1, different config survives a reboot, it only reverts if you change the pool, e.g., add another disk. Link to comment
kode54 Posted September 29, 2016 Share Posted September 29, 2016 Resurrecting this old thread again, to point out how you can easily create a RAID0 cache from the SSH console, in the event that you have a weird setup, like mine. I have a 256GB SSD and an older 128GB SSD, and the web interface was simply showing that there was no balance in the arrangement at all. It was creating an unbalanced RAID1 of a 192GB capacity, with 2/3 of the 256GB drive mirrored by the other 1/3 of the 256GB and the 128GB drive. First you need to take your array offline, and configure the cache to the number of drives you wish to merge. Then you need to find the /dev/sdx nodes for the two or more SSDs you wish to add. Then you need to run the following from a bash prompt logged into unRAID: mkfs.btrfs -d raid0 -m raid0 -f /dev/sdx /dev/sdx <etc> Where /dev/sdx items listed above are the cache drives, in the order you want them in. These steps are only necessary if you want to format the cache manually. A rebalance may work now, but there's one addition you need, as of some version of the btrfs tool, is the -f switch, to force conversion. This step is necessary now, as it will otherwise error out at the attempt to convert down to an insecure RAID0 array. Link to comment
wedge22 Posted December 23, 2016 Share Posted December 23, 2016 So kode54 is this the only way to convert the cache drives over to RAID0, I have the same configuration as yourself a 240GB and 128GB SSD and the current RAID1 setup means I am losing valuable disk space. I do not mind losing out on RAID1 as I can create backups of appdata, Dockers and Vdisks. This is my current config. Data, RAID0: total=221.52GiB, used=197.20GiB System, RAID1: total=32.00MiB, used=16.00KiB Metadata, RAID1: total=1.00GiB, used=218.52MiB GlobalReserve, single: total=80.00MiB, used=0.00B Total devices 2 FS bytes used 197.41GiB devid 1 size 111.79GiB used 111.79GiB path /dev/sdg1 devid 2 size 223.57GiB used 111.79GiB path /dev/sdf1 Link to comment
saarg Posted December 23, 2016 Share Posted December 23, 2016 So kode54 is this the only way to convert the cache drives over to RAID0, I have the same configuration as yourself a 240GB and 128GB SSD and the current RAID1 setup means I am losing valuable disk space. I do not mind losing out on RAID1 as I can create backups of appdata, Dockers and Vdisks. This is my current config. Data, RAID0: total=221.52GiB, used=197.20GiB System, RAID1: total=32.00MiB, used=16.00KiB Metadata, RAID1: total=1.00GiB, used=218.52MiB GlobalReserve, single: total=80.00MiB, used=0.00B Total devices 2 FS bytes used 197.41GiB devid 1 size 111.79GiB used 111.79GiB path /dev/sdg1 devid 2 size 223.57GiB used 111.79GiB path /dev/sdf1 No its not the only way. There is the easy way here. Link to comment
Recommended Posts
Archived
This topic is now archived and is closed to further replies.