tazire

Members
  • Posts

    375
  • Joined

  • Last visited

Everything posted by tazire

  1. Having a little trouble with permissions atm... I know its something i've done but looking for a bit of a hand to get it back working. I recently moved all my docker appdata and docker.img to an unassigned device nvme. I had a few permissions issues with a couple of dockers not being able to edit or access their config files and i think thats whats happening here. I had to redo my sabz and rutorrent dockers. both back up and running as they should. but as a result ive had to go and change the app api in my sonarr radarr etc. radarr and lidarr worked as they should. however when i change the api the test works fine and connects but the save button does nothing. so when i leave that part of the settings and go back in the api has not been updated. Anything i can do to fix this? im reluctant to redo sonarr and have to readd all my media to it again. cheers. EDIT Just to add to this. Getting the following errors in logs. 2020-07-17 14:37:04,424 DEBG 'sonarr' stdout output: [Error] TaskExtensions: Task Error [v2.0.0.5344] System.Data.SQLite.SQLiteException (0x80004005): attempt to write a readonly database attempt to write a readonly database at System.Data.SQLite.SQLite3.Reset (System.Data.SQLite.SQLiteStatement stmt) [0x00083] in <61a20cde294d4a3eb43b9d9f6284613b>:0 at System.Data.SQLite.SQLite3.Step (System.Data.SQLite.SQLiteStatement stmt) [0x0003c] in <61a20cde294d4a3eb43b9d9f6284613b>:0 at System.Data.SQLite.SQLiteDataReader.NextResult () [0x0016b] in <61a20cde294d4a3eb43b9d9f6284613b>:0 at System.Data.SQLite.SQLiteDataReader..ctor (System.Data.SQLite.SQLiteCommand cmd, System.Data.CommandBehavior behave) [0x00090] in <61a20cde294d4a3eb43b9d9f6284613b>:0 at (wrapper remoting-invoke-with-check) System.Data.SQLite.SQLiteDataReader..ctor(System.Data.SQLite.SQLiteCommand,System.Data.CommandBehavior) at System.Data.SQLite.SQLiteCommand.ExecuteReader (System.Data.CommandBehavior behavior) [0x0000c] in <61a20cde294d4a3eb43b9d9f6284613b>:0 at System.Data.SQLite.SQLiteCommand.ExecuteScalar (System.Data.CommandBehavior behavior) [0x00006] in <61a20cde294d4a3eb43b9d9f6284613b>:0 at System.Data.SQLite.SQLiteCommand.ExecuteScalar () [0x00006] in <61a20cde294d4a3eb43b9d9f6284613b>:0 at Marr.Data.QGen.InsertQueryBuilder`1[T].Execute () [0x00046] in C:\BuildAgent\work\5d7581516c0ee5b3\src\Marr.Data\QGen\InsertQueryBuilder.cs:140 at Marr.Data.DataMapper.Insert[T] (T entity) [0x0005d] in C:\BuildAgent\work\5d7581516c0ee5b3\src\Marr.Data\DataMapper.cs:728 at NzbDrone.Core.Datastore.BasicRepository`1[TModel].Insert (TModel model) [0x0002d] in C:\BuildAgent\work\5d7581516c0ee5b3\src\NzbDrone.Core\Datastore\BasicRepository.cs:111 at NzbDrone.Core.Messaging.Commands.CommandQueueManager.Push[TCommand] (TCommand command, NzbDrone.Core.Messaging.Commands.CommandPriority priority, NzbDrone.Core.Messaging.Commands.CommandTrigger trigger) [0x0013d] in C:\BuildAgent\work\5d7581516c0ee5b3\src\NzbDrone.Core\Messaging\Commands\CommandQueueManager.cs:82 at NzbDrone.Core.Messaging.Commands.CommandQueueManager.Push (System.String commandName, System.Nullable`1[T] lastExecutionTime, NzbDrone.Core.Messaging.Commands.CommandPriority priority, NzbDrone.Core.Messaging.Commands.CommandTrigger trigger) [0x000b7] in C:\BuildAgent\work\5d7581516c0ee5b3\src\NzbDrone.Core\Messaging\Commands\CommandQueueManager.cs:95 at NzbDrone.Core.Jobs.Scheduler.ExecuteCommands () [0x00043] in C:\BuildAgent\work\5d7581516c0ee5b3\src\NzbDrone.Core\Jobs\Scheduler.cs:42 at System.Threading.Tasks.Task.InnerInvoke () [0x0000f] in /build/mono/src/mono/external/corert/src/System.Private.CoreLib/src/System/Threading/Tasks/Task.cs:2476 at System.Threading.Tasks.Task.Execute () [0x00000] in /build/mono/src/mono/external/corert/src/System.Private.CoreLib/src/System/Threading/Tasks/Task.cs:2319
  2. im an idiot... XD i have usually got adv set on everything so never even dawned on me to check this... thanks for pointing it out.
  3. Just a question in relation to docker spacifically... i have offloaded my appdata and .img to a UD. I keep getting the error relating to the config path not being set as R/W Slave. When i try to edit the docker container to fix this the config is the one path that i dont have the edit option next to in order to set this as required. Is this something i should redefine myself by adding a new path and leaving the existing config path empty? obviously as the edit and delete button arent available on the config as standard this isnt an option.
  4. i just F5 on the site and it loaded up just fine after getting the above error.
  5. Im curious about using rsync for local backups. I essentially want a cache only share to sync to an array only share. Would rsync be the right thing to use or is there something else i should be looking at?
  6. Ok i tried to google this and just cant find anything to help. essentially the title says it all really. I am making a few changes. I am moving my appdata off the cache drives onto an nvme drive im expecting. I then want to change my cache drives to RAID 0 and run nextcloud off this only. I currently have duplicati backing up my nextcloud to backblaze but id like to have another share on my array as a direct copy of the nextcloud (cache only) share. for immediate availablity should i have cache drive issues. What is my best option to achive this? rsync the whole share to the other share? and whats the best way to have this occure automatically every day or once a week etc?
  7. Ok just in case anyone else finds similar issues. I checked my config and for some reason the version was down as 17.x.x so i changed that to 19.0.0 and all good now
  8. Yea that's the steps I followed. But when I get to the update command I get that error in the terminal. I then tried getting into my nextcloud instance and obv maintenance mode was on causing issues... Turned that off and it gave me the update button which I tried which again gave me a similar error to above only on the nextcloud ui EDIT Ok i tried a new install and just copying the old config from the nextcloud back up and it just goes back to the same issue... unfortunately. Also tried to rollback to linuxserver/nextcloud:18.0.4-ls81 but it just gives me the same result.
  9. hmmmm just trying to go from version 18.0.6 to 19 I got the following error... Nextcloud or one of the apps require upgrade - only a limited number of commands are available You may use your browser or the occ upgrade command to do the upgrade Set log level to debug Exception: Updates between multiple major versions and downgrades are unsupported. Update failed Maintenance mode is kept active Reset log level im aware this is an issue when you jump a major update but im fairly sure 18.0.6 was the most recent prior to 19? Not really sure where to go with this. Cant update and kinda stuck now. I read this thread https://help.nextcloud.com/t/updates-between-multiple-major-versions-are-unsupported/7094 but this seems to be if you skip a major release. At this point will i have to reinstall?
  10. i keep getting the following errors and then ombi becomes very unresponsive... [41m[30mfail[39m[22m[49m: Ombi.Api.Api[1000] StatusCode: InternalServerError, Reason: Internal Server Error, RequestUri: http://x.x.x.x:8989/sonarr/api/Episode?seriesId=46 [41m[30mfail[39m[22m[49m: Ombi.Schedule.Jobs.Sonarr.SonarrSync[2006] Exception when trying to cache Sonarr Newtonsoft.Json.JsonSerializationException: Cannot deserialize the current JSON object (e.g. {"name":"value"}) into type 'System.Collections.Generic.List`1[Ombi.Api.Sonarr.Models.Episode]' because the type requires a JSON array (e.g. [1,2,3]) to deserialize correctly. To fix this error either change the JSON to a JSON array (e.g. [1,2,3]) or change the deserialized type so that it is a normal .NET type (e.g. not a primitive type like integer, not a collection type like an array or List<T>) that can be deserialized from a JSON object. JsonObjectAttribute can also be added to the type to force it to deserialize from a JSON object. a restart of the docker brings it back for a short period until it happens again. Any ideas. I might just do a full fresh install soon.
  11. Ok just for anyone who happens to have the same problem. I ended up just reformatting my cache drives, reseating them to make sure it wasnt a connection issue and then just restored all my appdata using the plugin and readded my dockers. Slightly time consuming but worth it in the end. No more errors
  12. Exact error is as follows Apr 27 22:32:53 SERVER kernel: BTRFS critical (device sdaa1): corrupt node: root=5 block=1759166464 slot=120, bad key order, current (81626601 84 706685372) next (81592320 12 81626604) I googled it a bit and was wondering if I simply reformat my cache drives and start from an appdata backup will this resolve the issue? Server is runninng and I can use everything but the error is repeating and filling up the logs. any help greatly appreciated.
  13. hmmm got this installed there but having issues... unraid sees the card but when i plug the eithernet cable in i get nothing at all (light doesnt even appear on the switch to indicate connection)... plugged back into the motherboard 1gbe nic and all fine... EDIT Nevermind. This solved my issue.
  14. Perfect thanks for the definitive answer. I might try dual link with my current setup as I believe I have unoccupied backplanes I can just disconnect just to see what if any difference I feel In other uses. I have a few other things I'll be doing separately to try improve my streaming experience in the mean time. Thanks again
  15. I only have a very odd 4k file thats maybe more than that. But the other thing is this was never an issue for me prior to the lockdown. And the increase in people using my server. Its just the usage on the server seems to be slowing down my experience. So i was just trying to figure out if there is anything i can do to get everything back working seemlessly. Without going to much into the problem my main question is would a change in expansion card and change to dual link to the HBA improve the throughput and possibly my experience? This is very much a hobby for me too so i enjoy all the upgrading and learning about this stuff too. I mean i really have no real reason to upgrade to 10Gb networking but again i just enjoy all that kinda stuff. Anyway thanks for the input either way.
  16. How do you do this? Sorry never heard of this. Unless you mean transcode to RAM which I already have setup.
  17. Its not something i had noticed until this lockdown though... with more and more people accessing my plex server it seems the throughput of the HBA or expansion is possibly slowing down my experience. Just that with 2 or 3 people possibly reading data from 2 or 3 different drives at once it leaves less throughput for me to watch what i want. Im just wondering if the upgrade in expansion card and going to a dual link to the HBA will improve my experience?
  18. Its not so much about the speed.... as improving my personal experience within my home. Atm i am having slowdown viewing 4k content from my plex server while 2 to 3 others are watching from my server. This is direct play 4k content so i know its not a transcoding issue. Im essentially just trying to eliminate the bottlenecks which may be causing this. It may simply be the read speed of the drives are causing the issue but I just want to elimate the throughput of the drive access as a possible cause.
  19. Title says it all really. I currently have a 24 bay setup with 19 bays populated. my current setup is r7 1700 64gb RAM at default speed nvidia p4000 hba lsi 9240 8i Expander RES2SV240 1GB NIC soon to be 10GB Basically im looking to maximise my drive speeds as best i can. From reading other threads I was kinda leaning towards an expander upgrade to the RES2CV360 and use a dual link to the existing HBA, currently single link to the RES2SV240. Im just curious will the HBA be the bottleneck in this instance and should i upgrade it for best performance? Cheers in advance.
  20. Sorry to reopen an old thread but its exactly my question. I have the ASUS card on the way. Just curious if its as simple as plug and play or is there some level of configuration involved?
  21. No i never did after. Just couldnt get it working right. I didnt give it an awful lot of thought after the initial effort though.
  22. In relation to getting this to work with nextcloud.... If i remove the scan path set will that stop the auto scanning? Then im setting the path to clamav in nextcloud. Then I plan on setting this path in the nextcloud settings. Hope is that clamav instance sits idle until nextcloud activates it to scan uploads?? Just setting this up now so ive no idea if it will work.
  23. Same issue here. i just went with the docker and its all working as expected. just left the app alone while its early days.
  24. The app and community server app within nextcloud did get an update very recently. this is exactly the issue im having. i tried getting the community server app within nextcloud to work when it was updated but still no joy on that. EDIT- Found this on github https://github.com/ONLYOFFICE/onlyoffice-nextcloud/issues/228