Jump to content

giantkingsquid

Members
  • Posts

    52
  • Joined

  • Last visited

Everything posted by giantkingsquid

  1. Hi everyone, I did some searching the forums and reading the help files on the UI and can't find a reason for this behaviour on my array. Basically, Disk 1 is going to fill up while there is a lot of room left on the other two disks. See below for how my array is set up: Now I have had this array going since 2017, so I haven't been in an fiddled share settings for five or six years, but from my reading of the share settings the data should be getting written to disk 3 in preference to disk 1 and 2 (Disk 3 is the only one with space above the high water mark). All of my shares are set up along these lines: I have double checked and no shares are forcing the use of Disk 1. Thanks very much, Tom
  2. Thanks for the suggestions. I ended up hosting it directly rather than in a VM, and it seems happy there. Cheers.
  3. This looks to just be some scripts Oh ok, I thought it was neatly packaged up as a stand-alone as you can get it as a port or pkg for FreeBSD. What I really want is plain old mysqldump but I’m not sure if it’s available without the entire database.
  4. Hi can I please request mysqldump-secure? I’d like to use it to pull database dumps from a remote server back to the array. https://github.com/cytopia/mysqldump-secure/blob/master/README.md thanks, tom
  5. Has anyone managed to get this working through a Nginx reverse proxy? I've tried two methods based on setups I've used successfully for other containers with no luck. #Lychee proxy attempt #1 location /lychee/ { proxy_pass http://192.168.178.25:8082; add_header X-Frame-Options SAMEORIGIN; proxy_set_header Host $host; proxy_set_header X-Forwarded-Host $server_name; proxy_set_header X-Real-IP $remote_addr; proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for; proxy_set_header X-Forwarded-Ssl on; proxy_set_header X-Forwarded-Proto $scheme; proxy_read_timeout 90; proxy_redirect http://192.168.178.25:8082 https://$host; } This first method results in the formatting looking all wrong. I'm guessing the CSS is not making it through. See Capture.PNG. The second method: #Lychee proxy attempt #2 location /lychee { proxy_pass http://192.168.178.25:8082/; proxy_set_header X-Lychee-Base "/lychee/"; add_header X-Frame-Options SAMEORIGIN; } This method results in what looks like the right format but no site. See Capture2.PNG. Any ideas on this please? Thanks, Tom
  6. Thanks very much, got around to making the changes and all is working well. Thanks again.
  7. Hi All, I've had this docker working fine for a few years now and it seems to have stopped working, presumably since upgrading to 6.7. I can access the web ui, but cannot start the server. I get the following error in the web ui: Has anyone got any ideas for that? I've read the last few pages of this thread but couldn't see anything relevent. Cheers, Tom
  8. Yes that's right, and the files that fail to back up seem to vary each time. There are some logs, but not great. Here are some of them. I believe that the 2nd and 3rd have been solved. The illegal symbol was some OS X metadata that I have since excluded, the permissions issue was fixed by changing the backup destination to root:root. The top one still has them stumped. Are they on the right track with their blaming of the fuse.shfs? Or is it something simple I wonder? Thanks, Tom 2018-06-14 18:03:40,919441 [ERROR]: [ CBB ] [ 4 ] Backup error: /storage/appdata/couchpotato/data/cache/python/bb056859be5dab3d6d2620d739c6fea0 message: Error on create destination path code: 2 2018-06-12 18:56:28,940418 [ERROR]: [ CBB ] [ 4 ] Ignoring file: " /storage/documents/Wedding/Icon " file contain illegal symbol! 2018-06-09 05:51:03,762566 [ERROR]: [ CBC ] [ 8 ] Error open destination file data, to append data path: /localbackup/CBB_Apollo/storage/windows_backups/thoma/DIONYSUS/Data/C/Users/thoma/Downloads/signal-desktop-win-1.9.0 (2018_05_10 07_41_41 UTC).exe:/20180510072718/signal-desktop-win-1.9.0 (2018_05_10 07_41_41 UTC).exe message: Permission denied
  9. Hi, I've been persisting with this for a while now and I still cannot get a full backup to complete. I've been dealing with Cloudberry support but they seem fairly confused as well. This is my latest response from them. What do you think? The test backup they're referring to was when they asked me to backup one file only that had failed when the full plan was run. The host directory for storage is a USB3 drive through Unassigned devices, I have it available to this docker with the rw slave option. Things I've tried so far: Changed path to external drive from rw to rw slave. This seemed to increase the speed of the backup drastically. Changed permissions of storage directory to root:root. The Cloudberry processes seem to run as root. Any ideas would be appreciated. Tom
  10. I followed the steps on the plex site to repair a database and had no luck. The integrity check said it wasn't a database! I did a .dbinfo on it and it said it had zero tables... Anyhow I rolled back to a backup db from 2 days ago and everything is up and running again! Thanks very much for your help
  11. Thanks for the tip, it does appear to be corrupt Any ideas on how to repair a corrupt Plex DB? May 04, 2018 06:08:45.823 [0x150267066700] INFO - Plex Media Server v1.12.3.4973-215c28d86 - ubuntu docker x86_64 - build: linux-ubuntu-x86_64 ubuntu - GMT 10:00 May 04, 2018 06:08:45.823 [0x150267066700] INFO - Linux version: 4.14.16-unRAID (#1 SMP PREEMPT Wed Jan 31 08:51:39 PST 2018), language: en-US May 04, 2018 06:08:45.824 [0x150267066700] INFO - Processor Intel(R) Core(TM) i7-6700 CPU @ 3.40GHz May 04, 2018 06:08:45.824 [0x150267066700] INFO - /usr/lib/plexmediaserver/Plex Media Server May 04, 2018 06:08:45.822 [0x150272a8b800] DEBUG - BPQ: [Idle] -> [Starting] May 04, 2018 06:08:45.824 [0x150272a8b800] DEBUG - Opening 20 database sessions to library (com.plexapp.plugins.library), SQLite 3.13.0, threadsafe=1 May 04, 2018 06:08:45.824 [0x150272a8b800] ERROR - SQLITE3:0x10, 26, file is encrypted or is not a database May 04, 2018 06:08:45.825 [0x150272a8b800] ERROR - Database corruption: sqlite3_statement_backend::prepare: file is encrypted or is not a database for SQL: PRAGMA cache_size=2000
  12. 15% of 25 GB. I downloaded and installed the official Plex container and it works fine, so it must be something specific to the LinuxServer container. Hopefully I can get it going again as my database is quite extensive!
  13. Here you go: root@localhost:# /usr/local/emhttp/plugins/dynamix.docker.manager/scripts/docker run -d --name='plex' --net='host' -e TZ="Australia/Sydney" -e HOST_OS="unRAID" -e 'PUID'='99' -e 'PGID'='100' -e 'VERSION'='latest' -v '/mnt/user/movies/':'/media':'rw' -v '/tmp':'/transcode':'rw' -v '/mnt/user/appdata/plex':'/config':'rw' 'linuxserver/plex' bcfff21d0457b684aeda8c996f4c7d3cc3fadd582b5471d5c29b04c27c94ad1a Thanks in advance for any suggestions
  14. Hi guys, I've been happily running this container for years now without any problems. I went to use it today and got nothing, so checked the log and it's doing some weird: It all looks good until it actually gets to the line "Starting Plex Media Server", where it just repeats and never starts the server. I'm not sure whether it's something to do with last weekend's build? Though a quick google revealed no other problems out there. Any ideas? [s6-init] making user provided files available at /var/run/s6/etc...exited 0. [s6-init] ensuring user provided files have correct perms...exited 0. [fix-attrs.d] applying ownership & permissions fixes... [fix-attrs.d] done. [cont-init.d] executing container initialization scripts... [cont-init.d] 10-adduser: executing... usermod: no changes ------------------------------------- _ () | | ___ _ __ | | / __| | | / \ | | \__ \ | | | () | |_| |___/ |_| \__/ Brought to you by linuxserver.io We gratefully accept donations at: https://www.linuxserver.io/donations/ ------------------------------------- GID/UID ------------------------------------- User uid: 99 User gid: 100 ------------------------------------- [cont-init.d] 10-adduser: exited 0. [cont-init.d] 30-dbus: executing... [cont-init.d] 30-dbus: exited 0. [cont-init.d] 40-chown-files: executing... [cont-init.d] 40-chown-files: exited 0. [cont-init.d] 50-plex-update: executing... No update required [cont-init.d] 50-plex-update: exited 0. [cont-init.d] done. [services.d] starting services Starting Plex Media Server. [services.d] done. Starting dbus-daemon dbus[276]: [system] org.freedesktop.DBus.Error.AccessDenied: Failed to set fd limit to 65536: Operation not permitted Starting Avahi daemon Starting Plex Media Server. Found user 'avahi' (UID 106) and group 'avahi' (GID 107). Successfully dropped root privileges. avahi-daemon 0.6.32-rc starting up. No service file found in /etc/avahi/services. *** WARNING: Detected another IPv4 mDNS stack running on this host. This makes mDNS unreliable and is thus not recommended. *** *** WARNING: Detected another IPv6 mDNS stack running on this host. This makes mDNS unreliable and is thus not recommended. *** Joining mDNS multicast group on interface veth288010d.IPv6 with address fe80::473:83ff:fee3:4a39. New relevant interface veth288010d.IPv6 for mDNS. Joining mDNS multicast group on interface virbr0.IPv4 with address 192.168.122.1. New relevant interface virbr0.IPv4 for mDNS. Joining mDNS multicast group on interface vethde15982.IPv6 with address fe80::8c7c:13ff:fe06:4d73. New relevant interface vethde15982.IPv6 for mDNS. Joining mDNS multicast group on interface vethe43364f.IPv6 with address fe80::a080:87ff:fe0d:1d26. New relevant interface vethe43364f.IPv6 for mDNS. Joining mDNS multicast group on interface veth6cfbdfa.IPv6 with address fe80::4a8:42ff:fecf:47e1. New relevant interface veth6cfbdfa.IPv6 for mDNS. Joining mDNS multicast group on interface docker0.IPv6 with address fe80::42:9fff:fe44:581d. New relevant interface docker0.IPv6 for mDNS. Joining mDNS multicast group on interface docker0.IPv4 with address 172.17.0.1. New relevant interface docker0.IPv4 for mDNS. Joining mDNS multicast group on interface br0.IPv6 with address fe80::9c30:56ff:fe9b:fc4e. New relevant interface br0.IPv6 for mDNS. Joining mDNS multicast group on interface br0.IPv4 with address 192.168.178.25. New relevant interface br0.IPv4 for mDNS. Joining mDNS multicast group on interface eth0.IPv6 with address fe80::7285:c2ff:fe21:72b3. New relevant interface eth0.IPv6 for mDNS. Network interface enumeration completed. Registering new address record for fe80::473:83ff:fee3:4a39 on veth288010d.*. Registering new address record for 192.168.122.1 on virbr0.IPv4. Registering new address record for fe80::8c7c:13ff:fe06:4d73 on vethde15982.*. Registering new address record for fe80::a080:87ff:fe0d:1d26 on vethe43364f.*. Registering new address record for fe80::4a8:42ff:fecf:47e1 on veth6cfbdfa.*. Registering new address record for fe80::42:9fff:fe44:581d on docker0.*. Registering new address record for 172.17.0.1 on docker0.IPv4. Registering new address record for fe80::9c30:56ff:fe9b:fc4e on br0.*. Registering new address record for 192.168.178.25 on br0.IPv4. Registering new address record for fe80::7285:c2ff:fe21:72b3 on eth0.*. Server startup complete. Host name is Apollo.local. Local service cookie is 77526132. Starting Plex Media Server. Starting Plex Media Server. Starting Plex Media Server. Starting Plex Media Server. Starting Plex Media Server. Starting Plex Media Server. Starting Plex Media Server. Starting Plex Media Server. Starting Plex Media Server. Starting Plex Media Server. Starting Plex Media Server. Starting Plex Media Server. Starting Plex Media Server.
  15. Hi, I got an email from Crashplan today stating that my backup hadn't run for three days. My backup destination is a Lacie external RAID thing with 2x4TB drives in RAID 1. It is connected through USB3 and UD. Well, it was connected through UD. Upon inspection it was not mounted. I tried to mount it and got this: Aug 04 20:21:03 Mount drive command: /sbin/mount -t xfs -o rw,noatime,nodiratime '/dev/sdk2' '/mnt/disks/Hebe' Aug 04 20:21:03 Mount of '/dev/sdk2' failed. Error message: mount: wrong fs type, bad option, bad superblock on /dev/sdk2, missing codepage or helper program, or other error I ran xfs_repair without and with the -L flag but that doesn't appear to have helped. It still won't mount So, given that I'm currently 2000 km away from the machine I can't physically look at the enclosure to see if it's having some hardware RAID melt down or disk failure, have I got any other options? Assuming that I haven't corrupted the data on the good disk with the -L flag, that is if I've had a single drive failure. Cheers, Tom Update: I got someone to have a look at the enclosure and all was well, no LEDs indicating single or double drive failure. They also rebooted the enclosure but the problem still persists... Oh, there's another small EFI partition on the disk, not sure what it's for, but it's been there since I set it up and it mounts and unmounts fine, so I'm guessing the problem lies with the XFS filesystem on the sdk2 partition. What's the best way to reformat if it's required?
  16. I'll give it a go! Thanks for your help to date
  17. Yes that is the case. So if it was called nfljhslfh9w87520455h25g8ghnwto8.mkv it would not be recognized and would be left alone until the transcode finished and it was renamed to the correct name. However if it was was called NCIS_S11E01.mkv and Sickrage had already snatched (downloaded a torrent and sent it to the torrent program) that episode it would try to move it before Handbrake was finished encoding it because it knows it has downloaded that episode and is expecting it to turn up at some point in the future. Yes that is true if the movie file was the only file in the directory. If there is a .nfo, .idx, .jpg or anything else the directory is not removed. Just hit it with a rm -rf
  18. Yes I'm sure there are many ways to skin this cat, but if the Djoss' script could prevent this happening with just a few lines of code that would help many users I'm sure.
  19. Hi Djoss, I finally got around to testing your new version. Everything you've done works as expected, however it's still not working for me. The problem I am having is that the Handbrake docker writes the files to Output slowly (obviously), and that both Sickrage and CouchPotato are moving the partially written file to their final destination before the encode is complete. Obviously this freaks out all programs involved! So what I was thinking was that if you initially store the input file name as a variable, then wrote the output file to a randomly named file, for instance you could call it the MD5 hash that you calculate near the top of your script. Then when the conversion in complete and Handbrake is finished, rename the file back to the stored real file name. That way, both Sickrage and CP would not recognize the file and try to move it until it is complete. Second, minor problem (again for my work flow), is that the docker does not remove the directory when it completes the encode. Eg After /watch/movie/movie.mp4 is finished the whole /watch/movie directory should be removed. Thanks mate.
  20. In my experience, with CouchPotato and SickRage, they can both watch the same output folder and work out which movie files belong to which program. I've been using them both like this for six months now without a single misstep. I assume that they know that they downloaded x.mp4 so ignore all else. DJoss, I haven't had a chance to test the new version yet, will endeavor to today and let you know how it goes. I'll test for this situation as well.
  21. I think it would be better (certainly for my setup using CouchPotato) if the output is put in the output directory itself. Eg. /watch/movie/movie.mp4 --> /output/movie.mp4 But it's your docker so it's up to you really One thing that I noticed is that some torrents include a short sample video. I think this would be ok to get transcoded and outputted as well because most of the auto management tools (like CouchPotato) know to delete the sample movie when it appears in its watch folder. Cheers, Tom
  22. Great container, it works great Just one suggestion: Any chance that the auto conversion script could be modified to look in directories for the movie file? /watch/movie.mp4 works fine. /watch/movie/movie.mp4 does not. It's just that many of my movies downloaded via couch potato and deluge end up like that. edit: And obviously delete the whole directory, not just the movie file itself after the conversion is done. Thanks, Tom
  23. Just a few quick pointers for anyone wanting to use this plugin. I discovered this the hard way after a reboot... Copy your rsnapshot.conf to the flash drive and refer to it using the -v flag when you run the command. Similar for your SSH config, particularly the strict host checking. Make sure you turn it off in the ssh_config on your flash drive or else any cron jobs won't run if you're using rsync over ssh. Make sure you tell your remote (in the authorized_keys) to look for one of the keys (I used RSA) that are copied from the flash drive to /etc/ssh/ on boot.
×
×
  • Create New...