danimal86

Members
  • Posts

    247
  • Joined

  • Last visited

About danimal86

  • Birthday 03/30/1986

Recent Profile Visitors

The recent visitors block is disabled and is not being shown to other users.

danimal86's Achievements

Contributor

Contributor (5/14)

19

Reputation

1

Community Answers

  1. Just following up....it does in fact delete all the files that weren't on the source at the end of the transfer. Thanks for your help.
  2. Ok. I'll let it run and then see what it does. I just didn't want to have a bunch of duplicates
  3. strange, i enabled that "delete files on the destination" but it appears that it just copied but didn't remove the old one. does rsync go through and delete at the very end after copying?
  4. Dang....i looked right past that! Thank you
  5. LuckyBackups: I recently went through and changed my file structure for how i was housing media. I went through Luckybackups and updated the source/target paths and everything seems to be working now, but i've been going through and upgrading older low quality versions of movies to new high quality ones and i'm wondering if its possible for Luckybackups to remove the old copies that had previously been backed up. The top is my main server, and bottom is my backup. My backup is onsite, so its not the end of the world if i have to just trash and re-backup, but i'd rather not have to write about 40tb of data if i dont have to.
  6. @alturismo Wondering if i can pick your brain about setting up xteve_vpn. I dumped the .ovpn file in /appdata/xteve_vpn folder but not sure what else i need to do. I cant get any streams to run so i'm sure i haven't configured it correctly? Any requirements for the openvpn config (protical/port....)? Edit: So i dumped my 13.ovpn file created from torguard in /conf and pointed towards it in the docker template. Is that correct? and then as far as the logindata.conf, dumb question but what goes there? login info for torguard?
  7. First of all, sorry for placing this in the General Support area. I couldn't find a page for the Plex Inc Docker and my search for "Plex Remote Access" didn't come up with anything plex related. I had remote access working fine for years but all of a sudden i was noticing that my remote streams were limited to 2mbps and it looked like it was going through the Plex Relay and my server wasn't reachable remotely. I have upnp enabled in my asus router but it still doesn't seem to work. I tried setting up a port forwarding (external & internal ports set to 32400 and internal ip 192.168.1.99) and selected the Manually Specified Public Port in plex, but it still doesn't work. Anyone have any ideas? I haven't changed anything network related for years, so i dont think its a double NAT issue, but i dont completely understand it. Edit: i had a remote stream going and it was playing indirect The only think i have changed recently is installed jellyfin, so i deleted the docker and when i checked back on Remote Access in plex it was all green, but now its back to red. Edit2: This is bizarre. I have a vpn setup through built in wireguard. I have Sabnzbd setup to go through the vpn (i use ssl but cant be too safe) but plex is on Host mode. I went and dissabled the vpn on wireguard and BAM! All green and running a remote test it all working fine. How the hell did that affect my plex?
  8. Thank you! Cant believe i didn't put 2 & 2 together with that setting. Seems like it could be helpful to have that more towards the top with the other settings. Thats really what i want to do, kinda fine tune the days to keep on cache so it keeps cache around 80% full, ( seems like i ingest 1tb/month, so that would leave some extra room in there) and it will keep drives from spinning up and have fast play times.
  9. Could use some advice: I run two 2tb m.2's in raid0 for my cache. I would like to keep files on the cache drive for as long as possible before the mover takes over and puts them on the array. Seems to work fine, it will fill up to around 80ish % then all of a sudden, bam....moves everything to the array. This has been happening for a while, and its really hard to diagnose, because it take a bit of time to fill up the cache Here's my mover settings Here's my tuning settings: Is it something i have setup wrong? Ultimately i'd like to have it keep 60 days worth of media on the cache, then move it to the array. I'm a little confused on what to have "Only move at this threshold of used cache space" set to if i only want the 60 day old media kept
  10. Got it figured out. Go into CA Backup Appdata v2.5 and uder the Backlup/Settings tab, go down to the bottom to Show Advanced Settings, and uncheck Firefox to stop when backing up appdata.
  11. i tried using the stopwatch and it had reset itself, so it wasn't the appdata backup I'll give it a try again and see if it will not reset overnight. Edit: it reset overnight
  12. Dang, i didn't even think about this. Yes i do, i remember running into a problem with Lucky Backups that it would disconnect when my appdata backups, nightly. I'll set it to Exclude the Firefox appdata folder during my daily backups. @Djoss I bet this is the problem. Going to try the stopwatch again.
  13. I just checked on it, and the stopwatch had reset itself to the Stopwatch or Countdown homepage. So it seems like it is something with firefox resetting.
  14. Thats a great idea! I've got it running now.
  15. I didn't think about that. Any way to track that? I think for the most part i'll just have to set it up the day of....Not the biggest issue in the world. Thanks for your help