• Posts

  • Joined

  • Last visited


  • Gender
  • URL
  • Location

Recent Profile Visitors

The recent visitors block is disabled and is not being shown to other users.

Kacper's Achievements


Rookie (2/14)



  1. Hi nice to know u like it. I used docker called: unraid-kernel-helper Then I set to manual compile or sth. like that. Then I did my own make menuconfig - like it is described in the manual for this docker. Then I compiled with make commands. At the end i have copied over ssh battery modules.
  2. I am trying to find disk temperature with snmpwalk, but I cannot. What oid it suppose to be or what string should i search for to find disk temperature?
  3. Hi guys, I want to announce my simple ntp server, that can provide docker host localtime over ntp protocol. What does it do? This simple NTP server delivers time over ntp protocol to dummy devices like ip cameras, cheap routers that cannot handle time zone changes and day light saving correctly. With this container You can set your dummy device to use UTC time with no daylight saving and then point it to sychronize time ovet ntp protocol. This container will handle timezone correctly and deliver local time LTC to dummy device instead of UTC time (like regular ntp server does). Install under unraid. Go to docker tab in your unraid panel. At the bottom click add container. Then configure like at the image bellow. Replace ip with your own network ip address. Make sure selected ip is not taken by other device. After successful you will see: Point your dummy device to sync time over ntp. This server localtime can be different, than docker host. Local time zone is provided by TZ variable. Link to dockerhub: https://hub.docker.com/repository/docker/k4cus/ntpserverltc This is my first docker, so maybe It is not very clean implementation.
  4. Hi, I have written an ntp server which can serve local time LTC instead of UTC. It is usefull for dummy devices, that can not handle time zone, day light and so on. I am using it with ip cameras, that always messed their clocks even when synchronized with ntp server. I this case you can set utc zone with no day light saving and connect to my ntp server. My ntp server takes local time for OS so it is always handled correctly by the os. I would like to share this docker image, I have it running on my unraid. I was reading about making docker and I have docker hub account, but it is unclear for me howto make proper unraid image. It is even more unclear howto make it easy to install on unraid via community plugin, so paths and configuration option appear in unraid automatically? Can You provide some good tutorial? There are many tutorials, but very few that are reliable
  5. This docker has a little bug. /root/.ssh/ dir schould be mounted and persistent, as after container reinstall known_hosts and private keys are lost. As workaround I am using: ssh -i /config/id_rsa -o UserKnownHostsFile=/config/known_hosts -p 222 root@your.server.ip and my settings in rsnaphost.conf: ssh_args -i /config/id_rsa -o UserKnownHostsFile=/config/known_hosts -p 222 rsync_short_args -az # My mysql dump on my debian machine: backup_exec ssh -i /config/id_rsa -o UserKnownHostsFile=/config/known_hosts -p 222 root@your.server.ip "mysqldump --defaults-file=/etc/mysql/debian.cnf --all-databases| bzip2 > /root/mysqldump.sql.bz2" # Then backup whole /root folder backup root@your.server.ip:/root your.server.ip/ # and etc www vmain folders - it is ISPCONFIG machine backup root@your.server.ip:/etc/ your.server.ip/ backup root@your.server.ip:/var/www/clients your.server.ip/ backup root@your.server.ip:/var/vmail your.server.ip/ Of course to install private key ssh login option do: # Only once to generate key ssh-keygen # do this for each server where you want to install your private key ssh-copy-id -i /config/id_rsa -p 222 root@your.server.ip To make this post complete: it is required to modify crontabs: root@nas:/mnt/user/appdata/rsnapshot-backup/crontabs# cat root # do daily/weekly/monthly maintenance # min hour day month weekday command */15 * * * * run-parts /etc/periodic/15min 0 * * * * run-parts /etc/periodic/hourly 0 2 * * * run-parts /etc/periodic/daily 0 3 * * 6 run-parts /etc/periodic/weekly 0 5 1 * * run-parts /etc/periodic/monthly # rsnapshot examples 0 0 * * * rsnapshot daily 0 1 * * 1 rsnapshot weekly 0 2 1 * * rsnapshot monthly and in rsnapshot.conf: ######################################### # BACKUP LEVELS / INTERVALS # # Must be unique and in ascending order # # e.g. alpha, beta, gamma, etc. # ######################################### retain daily 7 retain weekly 4 retain monthly 3
  6. Hi guys, after 5 days of struggling with shinobi I must say that this docker is the best one I have found if You want to use nvidia gpu. I can explain quickly. This docker is based on Ubuntu 18.04 - which has cuda-10.2, which is the last version of cuda supporting cuda capable cards 3.0. Here is a list, check your gpu: https://developer.nvidia.com/cuda-gpus This docker has gpu working out of a box - ffmpeg + yolo ! My recommendation to setup camera are: Advantage is the performance. I can see ffmpeg working on gpu. There is no cpu usage. Sorry I made mistake when measuring cpu usage. Even if ffmpeg show it is running on gpu it still uses a lot of cpu. Be aware that when you decode using gpu, then in recording section you cant use gpu, or when you set recording to use gpu codec, then in input section you have to set to use cpu. In conlusion, when I have added all 5 cameras 2 of them keep disconnecting. Also cpu load is not less than when using zoneminder. Only Yolo plugin seams to benefit from gpu usage a lot. For these reason I will keep using zoneminder as it is more reliable for me, all cameras work fine. To asses update matters. Run console and type sh UPDATE.sh Then your shinobi will be updated. Of course if you change anything in container settings it will reverse back to original version from image. That is how dockers are superior over lxc or openvz One of the use case for my todays code:
  7. And I fully agree with this guy. And I fully disagree, I am even disgusted :D:D:D
  8. You right, maybe it is only my that find things 10 times harder to make them work in docker than in regular linux machine Zoneminder is not the only application. Any complicated application with heavy dependence on libraries particular version or data version will fail on docker or will be not worth the effort of porting to docker. One hint I can give is - use ubuntu 18.04 as base. Cuda version 10.2 supports cuda compatibility 3.0, which is important for users to use older, but still efficient gpus. Nvidia dropped compatibility in cuda 11.2 - it is cc 5.0 or 6.0 now, so useless for many people. We don't want to buy 1000 dollars gpu just to run ffmpeg on it and maybe tensorflow Running ffmpeg on my quadro K2000M takes 1% for one camera, on cpu 15% (intel i7) for decoding, so it is worth to struggle.
  9. Hi, I would like to share with community my new modification. Maybe limetech will include this option in unraid. To the point. When complicated application, with many system dependencies has to be deployed or developed it is convenient to use virtual machine. However virtual machine has extra overtread. Also dedicated RAM or GPU has to be attached to this VM only. With docker one GPU can be used in several containers, also ram is shared dynamically. It is nothing new as openvz can do that very well. Unfortunately unraid is lacking openvz. On the other hand in microservices approach present in docker environment it is assumed that only data is persistent and whole operating system libraries, application libraries are included in image and can be erased and restored from the image at any time in any version. It works well only when application designer is aware of this. However there are other use cases, when docker approach will fail: In development, when using application that depends heavily on libraries to be in one particular version – e.g. shinobi with yolo, tensorflow + cuda. Very often an effort or cost of adjusting application to work in docker philosophy is to high. When migrating physical machine or virtual machine. When one gpu should accelerate several applications. For this reason I have created small feature for unraid to make container persistent. Unraid is working in agreement with docker philosophy, therefore at any parameter change in unraid’s gui or when update is released it will erase whole container and recreate it. My code makes container detached from main repo and persistent across parameter edit. To make it work user.script plugin is required. Then my script should be copied to folder “/boot/config/plugins/user.scripts/scripts/persistentDockerContainers” and scheduled to be run once after array is started. How to use my code? You need to edit repository field, by adding “repository-“ prefix. Each container that is used as persistent should be having unique repository field !!! It is also advised that name of the container should be changed to unique, because unraid has undesirable behavior that, when another container is installed via community apps and container with the same name exists it will be deleted without any warning. In short, to make container persistent: 1. Change name to something unique. 2. Set unique repository name starting with prefix “persistent-“ That is all folks. Give me some feedback if You find it useful please. I don’t know if I should continue posting my solutions, as it takes a lot of time. I will do it only if it is useful for community. Best Regards, Kacper p.s. For people who are reasonable not to run someones code without checking it: root@nas:~# diff CreateDocker.php CreateDocker.php.org 108,126d107 < < // check what is name of the container to image, important when changing container name < if ($existing && $DockerClient->doesContainerExist($existing)) {$ContainerNameToImage = $existing; } < else if ($DockerClient->doesContainerExist($Name)){ $ContainerNameToImage = $Name; } < else{ $ContainerNameToImage = ""; < } < < // do the image < if ($DockerClient->doesContainerExist($ContainerNameToImage) && (strpos($Repository, "persistent-") !== false)){ < // repository name must by unique for presistent containers - this not guaranteed if user creates several containers from one repository and makes all containers persistent < $oldContainerInfo = $DockerClient->getContainerDetails($ContainerNameToImage); < if (!empty($oldContainerInfo) && !empty($oldContainerInfo['State']) && !empty($oldContainerInfo['State']['Running'])) { < // attempt graceful stop of container first < $startContainer = true; < stopContainer($ContainerNameToImage); < } < $DockerClient->commitContainer($ContainerNameToImage, $Repository); // commit persistent container < } < root@nas:~# diff DockerClient.php DockerClient.php.org 772,777d771 < public function commitContainer($id, $repo) { < $this->getDockerJSON("/commit?container=$id&repo=$repo", 'POST', $code); < $this->flushCache($this::$containersCache); < return $code; < } < persistentDockerContainers_unraid_6_9_0.zip
  10. I agree it would be great to have system containers avaliable. I find docker extremely complicated and taking a lot of effort to do really simple things. I don't understand the idea of dockers and when I see how zoneminder integration with ES is painfull in dockers I am supprised people force other people to use dockers. Openvz was really cool, now lxc does the same as openvz. Would be greate to have it in unraid. It is much easier to install everything from source code in system container than find or create docker.
  11. Hi guys, how to configure shinobi in such a way, that when I download video date and time will be overlayed on the video? I wan't it to behave like regular DVR. Video is quite useless for the police if there is no timestamp on it
  12. Hi, I would like to ask maybe off topic question. Unfortunatelly I can't register to zoneminder forum, I am marked as spamer regardles It was my first visit there in my life My question is: When I download video via: export or export video option, the mp4 image has no date and time on it. How to make zoneminder add date and time on the video? Thanks, Kacper
  13. 1. Firstly make a backup of your flash drive! 2. If you have backup of your previous version maybe try to restore it. or 2. If you don't, then I would do fresh 6.9.0 install and then make sure it boots. 3. When it boots successfully then I would copy over manually my configs. Maybe not all at once, but first unraid configs, if it boots then plugin configs. Maybe make sure that your usb flash drive is not faulty. Run checkdisk... In my case update was painless, so I would gues it is your system specyfic problem, which has to be investigated by You Good luck.
  14. I also remeber that during my testing performance of uploading files to owncloud when it was mounted /mnt/user/owncloud was degradeted so much that I was unable to use it. But it was long time ago since I have installed my server and I don't remember everything
  15. One drawback of my workaround, by mounting /mnt/disk1/owncloud is that I can have only one drive in use. Second drawback is that i can't use cache. I don't know what would happen if i used cache, but I think data corruption will occure if I mount /mnt/disk1/owncloud to my docker and cache owncloud share.