• Posts

  • Joined

  • Last visited


  • Gender
  • URL
  • Location

Recent Profile Visitors

The recent visitors block is disabled and is not being shown to other users.

glennv's Achievements


Contributor (5/14)



  1. I have lots of foreign movies (old dvd converts etc) spread all over my library with vobsubs that do not have any online available srt's etc, so i have to keep the vobsub track. Is there a way to tell tdarr to skip converting the file if it detects a vobsub track ? I can not risk loosing any of these while converting my entire library using tdarr, and manually checking thousands of videos which would be affected is a nogo. Any suggestions ?
  2. Your welcome. PLay with it and i thing with all the info sofar you will figure it out (and learn more about zfs in the process). I also started that way, step by step . But zfs and the commands (only 2 commands to do everything . zpool <blabla> and zfs <blabla>) are super simple and very well documented on the internet. Have fun learning. I have to do some other stuff so wont be repying today anymore. Have a great evening.
  3. Then there is still something wrong with the mountpoint or you manualy mounted it using mount commands. You can see that if you check : zpool get altroot zfs There it should only show the root folder (zo without the name of the zfs) altroot "/mnt/disks" It will then always use the name of the zpool (in your case "zfs") to create the full mount path as /mnt/disks/zfs In there you will see the content of the pool. If it does not show that, export again and reimport using the commands given before. The import mounts it for you at the right place so no need to manualy mount it. I must admit i have had it in the past that it remembered the old mountpoint in the altroot value but a few export/imports solved it in the end i think Its best to already specify the right mountpoint during pool creation to prevent this , but it should be possible to adjust bu using export/import. If not and it gets messed up at every boot, you can add the 2 export/import commands to your go file (/boot/config/go) so it will do an export and import to the right mountpoint you want at boot. If you want you can create a dataset named finalfolder in the zpool, which will then be mounted under /mnt/disks/zfs/finalfolder. datasets are sub devices in zpools with their own characteristics and they inherit the characteristics of the main pool unless set otherwise. You can simply create one (after you got the mountpoint sorted out to be /mnt/disks/zfs) by using the command. zfs create zfs/finalfolder
  4. Yes indeed . Your ghost share is [Backup] Between brackets you define the name you want the share to be visible by in windows. So you could reuse that and just change the "path" variable underneath to the actual mountpoint of your zfs (eg /mnt/disks/zfs) The only thing extra i have (which you may or may not need) is write list = yourusername valid users = yourusername With yourusername to be replaced with your unraid share user you want it to have access to. Last thing you need to do is open up the permissions of the zfs mountpoint on unraid. chmod 777 /mnt/disks/zfs chown nobody:users /mnt/disks/zfs Then restart samba (or the whole array )
  5. check my previous screenshot for the content you can add to smb extra to share your zfs. Maybe the ghost share you see is also in there So in your case you could add something like this: [My-AMAZING-ZFS-SHARE] path = /mnt/disks/zfs comment = browseable = yes # Public writeable = yes read list = write list = yourusername valid users = yourusername vfs objects =
  6. Think about it, everything in unraid os is already dynamic as build up in memory. During boot zfs is installed every time from scratch . Then (if set as such) it will import all available pools . In the pool parameters the mountpoint is set (altroot) and used/created during import. With "zpool get all <poolname>" you will see all current parameters for the pool.
  7. Yes That is the nice thing with zfs. Its very portable.
  8. p.s. When rebooting set array temporarily to not autostart . That also helps as then you only have to deal with zfs and not with the array, untill you are ready with the zfs stuff
  9. Ok so when you reboot witout drives conected, you check with zpool status , should be no pool (if pool, run export command) . Then plug in drives. Wait untill recognised. (should not auto import the pool ). Then run the import command i gave you. and check again afterwars with zpool status.
  10. If zpool status shows no pools , you can use the import command i gave you directly and skip the export.
  11. what is the output of "zpool status" ?
  12. That can have several causes. For example the pool is beeing used/bussy. If you dont know how to fix that , you can do what you did before eg shutdown, disconnect the 2 zfs drives, start the system . Then , if you start the array as you mentioned you will have all you normal shares back and the zfs pool is not imported (as missing all drives). Run the export command just to be sure the pool is gone. Then you can plug in the drives , wait untill they are detected and just run the second command "zpool import -R /mnt/disks zfs" . You can also do this export/import "before" the array is started as zfs is already active immediately after boot. That way you can check / make sure the zfs is mounted correctly before you start the array.
  13. yes, after these 2 command you can type "df" and you should see your zfs pool mounted under the new mountpoint. The /mnt/user/zfs should then be empty (doublecheck !!) and you can then remove that directory/share. Then i would restart the array and check if all your other shares will come back now zfs is not blocking/accessing it anymore. So step by step.
  14. If your zfs pool is named "zfs" then you can do that via an export and import of the pool as you can only set the "altroot" parameter during creation. For example setting the mountpoint of the pool named "zfs" to "/mnt/disks/zfs" (so the "altroot" is /mnt/disks) you would use : zpool export zfs zpool import -R /mnt/disks zfs Any datasets (if any) created underneath will inherit this mountpoint.