Aluavin
-
Posts
30 -
Joined
-
Last visited
Content Type
Profiles
Forums
Downloads
Store
Gallery
Bug Reports
Documentation
Landing
Posts posted by Aluavin
-
-
Thank you for your reply, I checked and as far as I understand my config should do the trick, however the Docker Image keeps filling. My current config looks something like this:
docker run -d --name='TimeMachine' --net='br0' --ip='192.168.178.9' -e TZ="Europe/Berlin" -e HOST_OS="Unraid" -e HOST_HOSTNAME="Aluavin" -e HOST_CONTAINERNAME="TimeMachine" -e 'VOLUME_SIZE_LIMIT'='1 T' -e 'TM_USERNAME'='Aluavin' -e 'PASSWORD'='<removed>' -e 'ADVERTISED_HOSTNAME'='timemachine' -e 'CUSTOM_SMB_CONF'='false' -e 'CUSTOM_USER'='false' -e 'DEBUG_LEVEL'='1' -e 'MIMIC_MODEL'='TimeCapsule8,119' -e 'EXTERNAL_CONF'='' -e 'HIDE_SHARES'='no' -e 'TM_GROUPNAME'='timemachine' -e 'TM_UID'='1000' -e 'SET_PERMISSIONS'='false' -e 'SMB_INHERIT_PERMISSIONS'='no' -e 'SMB_NFS_ACES'='yes' -e 'SMB_METADATA'='stream' -e 'SMB_PORT'='445' -e 'SMB_VFS_OBJECTS'='acl_xattr fruit streams_xattr' -e 'WORKGROUP'='WORKGROUP' -e 'TM_GID'='1000' -e 'SHARE_NAME'='TimeMachine' -l net.unraid.docker.managed=dockerman -l net.unraid.docker.icon='https://upload.wikimedia.org/wikipedia/de/f/f4/Time_Machine_%28Apple%29_Logo.png' -v '/mnt/user/timemachine/fewald':'/opt/timemachine':'rw' --hostname timemachine 'mbentley/timemachine' ba827ece46dd7bde741e3240fa345c1e8564b463a1732b54be840f20b1f482c5
-
So I got the TM running and the Backup also starts on my Macbook. So this is progress
However, I noticed that my Docker Image fills up once started. In order to fix it I had to Restart and than turn the docker off. somehow it recovered than. before I want to give it another shot I thought to maybe ask here if someone encountered the same issue or has an idea how i can fix that.
-
Hi,
my Nextcloud was working previously. However the current update breaks it in a way where I cant access my Files right now - which is an issue. Nextcloud runs on the Domain https://nextcloud.domain.tld/updater/
I'm running the mentioned docker and tried to update nextcloud withing the UI. However It won't do the job. I'm stuck at
Step 6 is currently in process. Please reload this page later.
So I I tried to searh a bit on how to fix it. It's also mentioned in this thread here, however the only suggestion was "give it a little time" - which doesnt work for me because this message appeard for like a week. Generally it seems like deleting the File within the docker /data/updater-xxxxxx/.step will bring you back to the Updater - which is a step backwards, but it is working.
Than set in in the appdater/nextcloud/www/nextcloud/config/config.php the parameter
'maintenance' => false,
However, If I make a request to to nextcloud dashboard I still see the message Update in progress.So all I can do is to head over https://nextcloud.domain.tld/updater/ and start the process again - which I did multiple times. I even ssh into the docker and renamed the /data/updater-xxxxxx to a different name. The Folder was than created successfull when I entered the /updater URL and ran the process.
One suggestion which I also tried was to run occ - which doesnt run at all due to
Could not open input file: /config/www/nextcloud/occ
So I thought hey maybe a permission issue and googled. However, this file seems to only exsist within the Updater Folder, so when I run inside the folder
root@dockerid:/data/updater-oci1k05iuicv/downloads/nextcloud
the command
sudo -u abc php occ maintenance:mode
it seem to work, so naturally I tried
sudo -u abc php occ maintenance:mode Maintenance mode is currently disabled
Which is the same which is noted in the unraid/appdata - File mentioned above.
However, when I try to access the UI I still get the message
Update in process.
I am completly lost and a bit mad right now. Can someone please tell me on how I might get my instance back and running? I will check in later in here because it is 2:35 in my timezone and i have to start working in a couple of hours but I'll check in later.
-
1 hour ago, horphi said:
Hello, can i run PM also with CPU instead GPU?
BR
PhilPM is a Miner for GPUs -see https://phoenixminer.org/
I would recomend to read about cryptos first.
But basically it is possible to mine ETH with a CPU but very much not recomended. Have a look at monero for that purpose.
came to this thread to point out that you have to install and start the docker first w/o the extra parameter for nvidia. than the docker will start but give you an error. Edit the docker and add the nvidia parameter => starts up fine. my 1070 does around 28MH/s which can be tweaked by some overclocking & undervolting (to around 30-32) which I have no idea how to do in unRaid.
- 2
-
Hi, just that this infomation is also in this Thread.
There is currently an issue with the latest version (.77). The deconz Team is debugging with the comunity right now.
Everyone affected should run version .75 since this version was running just fine.
Issue: Mesh doesn't build and some devices get dropped. (out of 50 nodes only 10 were connecting in my place)
To have a constructive thing for this thread. maybe provide an additional parameter which determines the version. is that a thing?
edit: the docker for HomeAssistant has recently rolled back to version .75
edit2: provided github links as source
-
If someone is curious: You don't have to setup everything again, it takes the same config.
Plus you have a webgui - so you dont need to run the manager on your other maschines (which leads to noice...)
-
+1
-
Currently I have a Script which starts letsencrypt docker (which just gets a cert from LE via DNS Challenge) then copies the cert to my usb device so my unraid has a green lock.
But it feels kinda rocky, anyone has a Idea how to improve this script?
#!/bin/bash echo "Starting Docker" docker start letsencrypt-server echo "Wait for 60s" sleep 10 echo "Wait for 50s" sleep 10 echo "Wait for 40s" sleep 10 echo "Wait for 30s" sleep 10 echo "Wait for 20s" sleep 10 echo "Wait for 10s" sleep 10 echo "Copy Files" cp /mnt/user/appdata/letsencrypt/keys/letsencrypt/priv-fullchain-bundle.pem /boot/config/ssl/certs/Columbus_unraid_bundle.pem echo "Stopping docker letsencrypt-server" docker stop letsencrypt-server echo "Wait for 10s" sleep 10 echo "Done"
-
HI,
Finally had enough of my manual creation of certs for my server. So I bought a Throwaway Domain and set the NS Records to Cloudflare.(For reasons I don't want to use my personal domains with CF). Cert gets created and 👍
My unraid WebUI is only available in my local network but resolves with server.domain.tld
One thing to do now is to copy the priv-fullchain-bundle.pem over to \flash\config\ssl\certs so I can access the web UI of unraid. How would I do this?
-
sorry for gravedigging!
could you manage to store those files to appdata?
- 1
-
21 hours ago, Ryonez said:
it's a hidden share called flash. So goto: `\\<unraidip>\flash`
I know how to find the flash drive I want to minimize the read access to this one. that's why I want to move those .ovpn files to my SSD.
-
On 12/18/2018 at 9:36 PM, Ryonez said:
Thank you, that worked.
Would you be able to tell me what the "Extended Routing:" is for? Will I be able to have dockers/vms use the vpn, but still be able to access their web interfaces?yes. you can test it: just get the firefox from the CA. Then go to a page which lists your IP: whatismyipaddress.com
so all traffic will go through this VPN. locally you can still access everything.
On 12/18/2018 at 9:00 PM, Squid said:Make a folder called openvpn in the root of your flash drive.
Toss your .ovpn file in there. Reload the settings page and you'll be able to select it and go from there.
can I somehow read those files from my appdata/ssd somehow?
-
Can I use the custom scripts to stop & start the docker Service?
-
Hi,
So for a while now I have this issue.
Whenever I start my Unraid System (sometimes you need to reboot) the Dockers can't start automatically.
So I went into edit Mode and I saw that the `Network Type` is set to none, instead of `br0`. Since this Interface is only missing for Docker (my VM templates still offers me to select it) I went to Settings/Docker and disabled the service and enabled it again.
I went back to edit the container and tada: network interface is already selected properly and container boots just fine.
Having a look in the Syslog prompts me:
Oct 26 09:31:42 Columbus emhttpd: Autostart enabled Oct 26 09:31:42 Columbus root: Starting diskload Oct 26 09:31:42 Columbus kernel: br0: port 1(eth0) entered disabled state Oct 26 09:31:43 Columbus tips.and.tweaks: Tweaks Applied Oct 26 09:31:43 Columbus emhttpd: Mounting disks... Oct 26 09:31:43 Columbus emhttpd: shcmd (53): /sbin/btrfs device scan Oct 26 09:31:43 Columbus root: Scanning for Btrfs filesystems Oct 26 09:31:43 Columbus emhttpd: shcmd (54): mkdir -p /mnt/disk1 Oct 26 09:31:43 Columbus emhttpd: shcmd (55): mount -t xfs -o noatime,nodiratime /dev/mapper/md1 /mnt/disk1 Oct 26 09:31:43 Columbus kernel: SGI XFS with ACLs, security attributes, no debug enabled Oct 26 09:31:43 Columbus kernel: XFS (dm-0): Mounting V5 Filesystem Oct 26 09:31:43 Columbus dhcpcd[1735]: br0: carrier lost Oct 26 09:31:43 Columbus avahi-daemon[5309]: Withdrawing address record for 192.168.178.26 on br0. Oct 26 09:31:43 Columbus avahi-daemon[5309]: Leaving mDNS multicast group on interface br0.IPv4 with address 192.168.178.26. Oct 26 09:31:43 Columbus avahi-daemon[5309]: Interface br0.IPv4 no longer relevant for mDNS. Oct 26 09:31:43 Columbus dhcpcd[1735]: br0: deleting route to 192.168.178.0/24 Oct 26 09:31:43 Columbus dhcpcd[1735]: br0: deleting default route via 192.168.178.1
Does this means the br0 will be disabled for some time (maybe in this time also means, that the docker service is starting)?
What can I do to solve this issue?
-
Hi,
so I have a Windows 10 VM which starts from a nvme SSD. I followed the Guide of Spaceinvader One.
Of course, I had to make adjustments, so I stubbed all USB Controllers except of 1 (Keyboard for Unraid & USB Stick). This was done because my Mainboard is weird and didn't let me split up those IOMMU Groups to a degree I needed to. However, stubbing all USB Controllers is no Issue since only 1 VM is running at a Time.
When I start Ubuntu as a VM everything is fine and it feels like bare metal.
When I start Windows(VM) the Story is differently. I have an Audio USB Interface connected, every now and then this Interface loses the connection for about 1 second. THe USB light flickers also and of course the sound skips for this 1 second. My Mouse and Keyboard (connected to the same USB Controler) still maintain the connection and I didn't notice any stutter with those.
Did someone else encounter this problem? Also, an Idea on how to fix it would be great, since I relay like to listen to music on my Windows VM.
-
This looks awsome. Do you mind writing up a bit? Id like to fill in the 5,25 bay on my Maschine with this VM Controller :)
-
ordered new ddr3 ram. Unfortunately, the CPU doesn't support ECC RAM, so I have to play the lottery again. when I do an upgrade next year I will also take a focus on ECC RAM.
-
-
1 hour ago, johnnie.black said:
If not using ECC RAM run memtest.
will do as soon as the advanced smart test is done
-
21 minutes ago, testdasi said:
I have had jD failing to extract perfectly good archives all the time. Have you tried extracting manually in a Windows envi with the latest version of WinRAR?
Yes, tried this also but I got an error like "corrupted file" or cannot find foo-bar.
-
Hi,
QuoteM/B: MSI - Z97 GAMING 7 (MS-7916)CPU: Intel® Core™ i5-4690K CPU @ 3.50GHzHVM: EnabledIOMMU: EnabledCache: 256 kB, 1024 kB, 6144 kBMemory: 16 GB (max. installable capacity 32 GB)OS: Unraid 6.5.3HDDs: (5+1)*Iron Wolf 4 TB (Storage)SSD: 950 PRO 512GB NVMe (Cache Drive)I'm currently debugging my system. So I use the Jdownloader Docker to Download Stuff. A while ago I noticed that 40-50% of files I download fail at crc checks. Most (>99%) of them are split into multiple parts - like 5-10 .rar archives.
To Debug I did the follwoing:
- Disabling the CRC Check to try to force a unzip => resulting in a corrupt file while extracting.
- tuned on "download again" if-CRC-check-fails-option to workaround this issue => resulting in infinite loops
- delete and setup docker again w/ deleting all concerning files in appdata to determine if it is something i messed up in configs => no difference
- switched to a different docker (there are only 2 on the Com. Apps Store) because maybe some weired update hickup => no difference
- asked a friend: does this happen to you with <same-files>? to determine if files are corrupt => Answer no
- started the Windows App, download <same-files> to determine if it is my ISP => no issue
- Checked SMART of SSD => seems fine
- Extended Check of HDDs => still running
So what I think out of the questions are currently
- ISP
- broken Docker(s)
- broken original files on the hoster
So right now I'm figuring out how to debug my Hardware. All Downloads go to Cache Drive (nvme SSD/ 950 PRO 512GB) first and will be moved at 11 pm to HDD. So naturally, I removed the Cache Drive from the Download Share. Also no change. So right now I'm figuring out if it is a coincidence that HDDs (6x Iron Wolfs (4TB) that all fail. The SMART report of my SSD seems to be fine. The extended Test is still running on all HDDs.
Is there something else I can check? I want to narrow down the issue and possibly solve it.
-
+1
also this guy:
-
On 6/6/2018 at 2:01 AM, aptalca said:
My question is, what is your purpose for getting the cert? If you want to use the cert for openvpn, you can use letsencrypt to get it and manage it, and then you can mount the folder in the openvpn container and point to the cert.
3The main purpose is that I have a clean URL while maintaining a TLS Connection so my Drives can be encrypted. I know that I could mock the URL with a custom DNS, but for now, an A-Level record to my local IP does the deal. same goes for nextcloud.domain.tld pointing to nextclouds docker IP address.
However, I think I have to accept that I have to use Cloudflare (or a different supported provider, preferably route53), open up my server so I can use http validation (won't do that) or to obtain the cert manually. In this case, option #3 seems to be I have to take until I've got some spare time to evaluate a domain movement to route53, Cloudflare is to sketchy for me.
-
Can I set somehow a spindown timer for this docker? The drives containing Nextcloud Data are spun up all the time. Did I miss a setting?
edit: found that nextcloud runs a cob every 15minutes which access the HDD is there a way to disable or even delay this one?
Timemachine Application Support Thread
in Docker Containers
Posted
This did the trick. so the template itself has to be modified. Thank you very much!