• Posts

  • Joined

  • Last visited

Everything posted by BloodBlight

  1. I freed up some space, and DDing out in 100MB chunks. It let me write to the cache drive all the way down to 113MBs... But I also noticed this: This got my thinking that this could be a simple math error. So I adjusted the minimum cache size to 1.6 GBs (came to that with some math I can't remember) and Shazam! No more writes to cache drive even for 1k files. I ended up with 376 MB free when I SHOULD have no less than 1.6GBs. But I re-did my tests and was able to zero it out again... So IDK at this point...
  2. Interesting... So, SSHFS, no change. If I rsync files to the drive and cause the error, then do a DD to the share, it goes directly to the array. If I delete a single 400MB file, taking the free space up from 220MBs to 620MBs, the DD will go to the cache drive even though it has less than the 1GB minimum and errors out!
  3. Indeed! I will do a dd with oflag=direct. I THINK that works through NFS. Oh... Ya know, I might try an SSHFS mount and see if it does the same thing. Testing now...
  4. I don't think that it does. I am working under the assumption that at least one file will cause a policy violation before triggering the rule to store elsewhere. I picked these files and sizes intentionally. The 1GB minimum free size should be able to accommodate the largest file two times over (400MB). So one file takes it below the 1GB mark (by even 1 byte), the OS can start writing a second file (assuming the file that would break the rule is still flushing), but by the time it gets to the third file, the first should by flushed and trigger the rule to move it down to the next tear of storage. Does that line of thinking sound right? Sorry, I am very good at breaking things.
  5. For my process, I am deleting the folder, that frees up about 2GBs on the cache drive, then re-starting the copy. Is there some sort of delay on the free space calculation? This is not a multi threaded copy so..
  6. Understood. The largest file being copied is only about 400MBs, so that should not be an issue here. And right now, even with that set, it is sitting at 220MBs free: Even if you subtract the largest file in the copy (400MBs) from the 1GB, it should never go below 600MBs free... So it seems like something isn't working right.
  7. I think so, yes. Is this what you mean:
  8. Off lined array can grabbed a screenshot of the cache drive:
  9. I am getting my feet wet with unRAID in a VM with some pass-through devices, so I have nothing in this config that can't be tossed and re-built. This is all in prep before the “big move”. I am using version: 6.9.2 Share setting has cache set to “Yes” Screen shots: (Main & Share Settings) Log dump attached. I have a small cache disk that is VM disk sitting on an SSD (only 5GBs for now, I intentionally made this small). I have three spinals passed through as LUNs (3TBs each, double parity). So far, everything is running great, until I enable the cache drive and start coping files. As soon as the cache disk fills I get this from rsync: rsync: close failed on "/mnt/###CENSORED FILE NAME###": No space left on device (28) rsync error: error in file IO (code 11) at receiver.c(853) [receiver=3.1.3 I have tried stopping the array and setting a minimum free space on the cache disk to 1GB (more than double the size of any single file I am trying to copy), but it just fills the cache drive to 100% anyway. It seems to free up space from the failed copy, but never tries moving data to the array (see 220 Mbs free in screenshot even with the 1GB minimum free). Waiting a while and restarting the copy seems to work, but it seems that the minimum free space on the cache disk is just not working…. It was my understanding that once a cache disk is full (or in this case minimum free is reached), any future writes will go to the array directly. Is that wrong? Is this a config issue, or have I hit a bug? Also, I am assuming this is happening because once a write starts to a cache disk, it can’t be moved to the array mid-copy... So a minimum free space on the cache disk should be a default config I would think (but seems to be blank). As I am writing this, and after I took the log capture, I just NOW am seeing this in the logs: Jan 16 10:55:29 Tower shfs: share cache full ### [PREVIOUS LINE REPEATED 22209 TIMES] ### Could this be related to having copy on write enabled (BTRFS not letting space go fast enough)? Please advise. Thanks!!! Side note, and “Anonymize diagnostics” function doesn’t seem to remove user names from the logs, so I did manually.
  10. I think I miss read something there the first time around (prepping to switch to unRAID from LizardFS). What I was hoping for was a way to remove a disk while maintaining parity. I think the clean drive and remove thing would do this. Sorry for burning time.
  11. If your array has dual parity bits, and you follow the guide to remove a single data disk, will the second parity disk continue to provide protection during the rebuild operation? If so, are you any more vulnerable to power outages during this window? Thanks.