• Posts

  • Joined

  • Last visited

Recent Profile Visitors

The recent visitors block is disabled and is not being shown to other users.

KrisMin's Achievements


Newbie (1/14)



  1. Fresh install. Somehow I cant access the webui on the https port. The default http port works fine. Bug or something with my setup, i cant figure out.
  2. Hello and thanks for posting the guide. I managed to get it working with a default OVMF bios just fine. I am running a pair of GTX 1060's on economic settings and it runs clean. hiveOS is a nice convenient tool for managing miners. I got around 1-2% more hash on windows, but managing mining windows VM on Unraid is crappy.
  3. OK, that was an issue with subnet availability. For some reason the NC could not connect to the database when pointed the database IP and port to it. When I joined the same bridge as the mariadb, the issue disappeared. As far as i know this should not be happening, but it somehow did.
  4. I had a server crash and now my NC seems to not get connected to database. ","app":"remote","method":"GET","url":"/status.php","message":{"Exception":"Doctrine\\DBAL\\DBALException","Message":"Failed to connect to the database: An exception occurred in driver: SQLSTATE[HY000] [2002] Host is unreachable","Code":0, Anyone knows how to fix this? The config and data volumes look to be fine. I have no idea why can't it reach the mariadb...
  5. FYI for anyone running a node here. The official Storj node GUI is crappy, but luckily for us, there's one awesome Grafana dashboard available for nodes statistics. Give it a go.
  6. Thanks! I think that should be written in OP - to use as Cache = yes.
  7. Update about running this thing on "cache preferred" mode. Looks like storj has some silly rule built in which returns an error and switched the node offline when the space runs below ~450GB. This rule makes no sense to me, why...? 2021-01-27T19:29:00.600Z ERROR piecestore:cache error getting current used space: {"error": "context canceled; context canceled; context canceled; context canceled; context canceled; context canceled", "errorVerbose": "group:\n--- context canceled\n--- context canceled\n--- context canceled\n--- context canceled\n--- context canceled\n--- context canceled"} Error: piecestore monitor: disk space requirement not met So I had to switch the cache off for my storj share. As soon as I did that, the node started working again. Maybe there's some workaround on this? Would more exprerienced Unraid users would even prefer to run such thing on cache preferred? I suspect it would put less stress on HDD's when done so.
  8. The issue was with windows explorer, which does not accept a self signed certificate. I got it working with FileZilla. However, it looks like webdav is considerably slower in moving files than a mapped external drive copy inside the nextcloud. The latter has other issues, like not being able to copy several files or directories at once and gives an error if multiple directories are put to copy/move. So I stick to webdav and FileZilla combo, which seems to be working without any issues. Just takes a bit more time, which is fine by me. Besides, FileZilla has useful built in rules in case a version of a same file already exists in destination - like "overwrite if a version is newer".
  9. Hey Everyone! I ask some help with getting this thing working via http. I did a -p8080:80 extra parameter, but It always redirects to https. When i want to map the nextcloud data dir with webdav, I get asked for certs and i want to avoid that.
  10. OK, that was my issue too! Thanks! What an annoying bug, wasted a couple of hours digging in google and docs. Dno why I didn't double check if ports were mapped right in docker.
  11. Sorry if i didn't explain it too well. Error in the web browser when trying to reach the app by the domain name. Both, the NPM and theAPP were reachable by their LAN IP's
  12. Can anyone explain me why NPM does not work if both, the NPM and the APP sit on the LAN ("br0" in mine) and not in the unraid default bridge. I gave a local static IP to the NPM container and also to the APP container I wanted to proxy to. When done so, APP can not be reached with its domain name (Both can be reached with their IP's). In detail: - my router redirects all 80 to NPM IP:PORT and all 443 to NPM IP:PORT - my NPM has a proxy host: to APP IP:PORT I get connection refused error. may I ask why please?
  13. Sure, you can turn autoupdate on-off for every container individually via Unraid Auto-update Appliations plugin. Just go in to this app (under settings) and you'll find docker tab and auto-update toggle buttons there.
  14. Here: Yeah, probably the best option for us would be to run 8TB or 10TB nodes and only start a next one after the first one gets vetted. I do not know If the second one will still get vetted twice slower than the first one or will it get it as fast as the first one. Need to test or ask from the official forum about that.
  15. Of what I have read, there is no hard cap on the storage size per node and the number is just a strong suggestion from the dev team. Someone in the forum did some math with hes years of statistics data and came up with a conlcusion that a theoretical limit would be aound 40TB of per IP subnet if IP filters work like we think they do. That number is based on hes statistics that a ~5% of all data gets trashed and therefor a pool of >40TB would never fill up, because 40TB would be the line in which point ingress = trashed. Of course if in the future an average Ingress rate increases, then the max pool size would also be larger. On our case (running nodes on docker), I would keep a single node size less or equal to 24TB, because it's less risky. If one node gets a bad reputation for some reason, then you still have other good ones to compensate and you could easily increase their size if needed. I started two identical nodes running on the same machine and the same network. Somehow one of them got uptime warning (from one us. sattelite) and lowered its uptime score, even tho they are identical in terms of availability and uptime. How did that happen, I have no idea. Must be some kind of bug. If I see this happening again, then I'll investigate it some more. Hopefully it was just once.