Connor Moloney

Members
  • Posts

    48
  • Joined

  • Last visited

Posts posted by Connor Moloney

  1. In my unraid server I have started receiving errors on one of my WD drives however no reallocated sectors as of yet. I have only had the drive about 9 months and it is already showing 'old-age' and 'pre-fail' in smart reports being run on the server. How should I proceed now? Should I request an RMA as it is still in warranty or leave it until it starts spitting out sector errors? I can provide latest smart reports if necessary. 

  2. 2 minutes ago, CHBMB said:

     

    I'm not sure you could accomplish that with a click.......

    Generally takes the command line to do something like that.  You always used our container?  Or migrated from somebody elses version?

    Yeah I have always used your sonarr version have recently transferred to your plex media server and tautulli builds as well, in the process of trying to configure them exactly as I had them previously but then I noticed that hundreds of my films had disappeared :( so using radarr to get them back and then will finish configuring the server

  3. 10 minutes ago, CHBMB said:

    I dunno what you've done, but you've changed permissions on some pretty important files there.

     

    run this then restart the container.

     

    
    chown -R abc:users /config

     

    You need to work out what the hell changed those perms to www-data and make sure it doesn't happen again.

    Okay so I ran this, restarted the docker then afterwards re-ran the commands you gave me earlier to get this output now:

    image.png.54b0f727c5bb63fe3d0b08f74a826b1f.png

    is this normal?

  4. 4 minutes ago, CHBMB said:

    I'm gonna guess you've followed some tutorial on a webserver blindly as www-data is commonly used in that context.  Either that or migrated data from somewhere else.

    I don't think I have done either, only thing I have done recently which could remotely be connected to this is set up a dns for my openvpn and that's about it? So how do you think it's happened then?

  5. 7 minutes ago, CHBMB said:

     

    Have you run "New Perms" or changed permissions on your Unraid box at all?  Just a hunch.

     

    Try running these two commands and posting the output of the second one.

     

    
    docker exec -it sonarr bash
    
    ls -la /config

     

    I ran that and got this output:

    image.png.c3395f97efaf1e0cacccf51f83e91158.png

  6. 6 minutes ago, CHBMB said:

    My gut feeling looking at those logs is some sort of database corruption tbh.  SQLite is readonly which is a problem.

     

    You may want to grab some debug logs and post over at the Sonarr forums, I'm not as convinced as @saarg that the remote mount is the issue, although it's a possibility.  Be interesting to see @hayleraid's logs etc because until we see them, to assume it's the same problem is a little too speculative for my liking.

    I just tried turning sonarr log mode to debug and it said "failed to save general settings"

    I think you are right with the database corruption as all it says in the log table is stuff about the database and errors, how can I fix that?

    image.thumb.png.84626f87056e8637a5c29d3737010617.png

  7. Okay so after trying to reconfigure sonarr to work with deluge (container i used to use for downloads) I would set it all up correctly click test it would work got the notification in the corner popping up saying it had succeeded. So i press save and wait for the spinny circle to disappear then press the 'x' in the top right and deluge isn't there in the download clients. Like I said earlier it's almost as if sonarr just isn't responding, I have tried restarting the container, turning docker on and off but don't know what else to try?

  8. 6 minutes ago, saarg said:

     

    Have you tried to use unraid for completed storage? 

    When I first started using this docker I had the download client on my unraid server so yes I have done that in the past but now I download using my desktop (vpn) so the location is on that computer now

  9. 7 hours ago, CHBMB said:

    Docker run commands please, you're the only two reports of this that we've seen, nothing in Discord or Github about it.

    Stopping container: sonarr

    Successfully stopped container 'sonarr'

     

    Removing container: sonarr

    Successfully removed container 'sonarr'

     

    Command:root@localhost:# /usr/local/emhttp/plugins/dynamix.docker.manager/scripts/docker run -d --name='sonarr' --net='bridge' -e TZ="Europe/London" -e HOST_OS="unRAID" -e 'PUID'='99' -e 'PGID'='100' -p '8989:8989/tcp' -v '/dev/rtc':'/dev/rtc':'ro' -v '/mnt/user/Media/TV Shows/':'/tv':'rw' -v '/mnt/disks/192.168.1.72_My_Pc_-_Spare_Storage/Movies/completed/':'/downloads':'rw,slave' -v '/mnt/cache/appdata/sonarr':'/config':'rw' 'linuxserver/sonarr' 

    dfa8aadc778577e64943c3ac6fc3a814925d814b2eda5fd9be4a90ed98e52abe

    The command finished successfully!

  10. Having a few issues with this container it has been working for ages then 7 days ago it stopped working, its still running but none of the automated tasks are running, it doesn't send requests to look for new episodes, it isn't recognising when new episodes are added to the array and also the log file hasn't been updated in 7 days. Any thoughts?

     

    So in the log shown in the docker window this is repeated god knows how many times:

    [v2.0.0.5163] System.Data.SQLite.SQLiteException (0x80004005): attempt to write a readonly database
    attempt to write a readonly database
    at System.Data.SQLite.SQLite3.Reset (System.Data.SQLite.SQLiteStatement stmt) [0x00083] in <61a20cde294d4a3eb43b9d9f6284613b>:0
    at System.Data.SQLite.SQLite3.Step (System.Data.SQLite.SQLiteStatement stmt) [0x0003c] in <61a20cde294d4a3eb43b9d9f6284613b>:0
    at System.Data.SQLite.SQLiteDataReader.NextResult () [0x0016b] in <61a20cde294d4a3eb43b9d9f6284613b>:0
    at System.Data.SQLite.SQLiteDataReader..ctor (System.Data.SQLite.SQLiteCommand cmd, System.Data.CommandBehavior behave) [0x00090] in <61a20cde294d4a3eb43b9d9f6284613b>:0
    at (wrapper remoting-invoke-with-check) System.Data.SQLite.SQLiteDataReader..ctor(System.Data.SQLite.SQLiteCommand,System.Data.CommandBehavior)
    at System.Data.SQLite.SQLiteCommand.ExecuteReader (System.Data.CommandBehavior behavior) [0x0000c] in <61a20cde294d4a3eb43b9d9f6284613b>:0
    at System.Data.SQLite.SQLiteCommand.ExecuteScalar (System.Data.CommandBehavior behavior) [0x00006] in <61a20cde294d4a3eb43b9d9f6284613b>:0
    at System.Data.SQLite.SQLiteCommand.ExecuteScalar () [0x00006] in <61a20cde294d4a3eb43b9d9f6284613b>:0
    at Marr.Data.QGen.InsertQueryBuilder`1[T].Execute () [0x00046] in M:\BuildAgent\work\b69c1fe19bfc2c38\src\Marr.Data\QGen\InsertQueryBuilder.cs:140
    at Marr.Data.DataMapper.Insert[T] (T entity) [0x0005d] in M:\BuildAgent\work\b69c1fe19bfc2c38\src\Marr.Data\DataMapper.cs:728
    at NzbDrone.Core.Datastore.BasicRepository`1[TModel].Insert (TModel model) [0x0002d] in M:\BuildAgent\work\b69c1fe19bfc2c38\src\NzbDrone.Core\Datastore\BasicRepository.cs:111
    at NzbDrone.Core.Messaging.Commands.CommandQueueManager.Push[TCommand] (TCommand command, NzbDrone.Core.Messaging.Commands.CommandPriority priority, NzbDrone.Core.Messaging.Commands.CommandTrigger trigger) [0x0013d] in M:\BuildAgent\work\b69c1fe19bfc2c38\src\NzbDrone.Core\Messaging\Commands\CommandQueueManager.cs:82
    at (wrapper dynamic-method) System.Object.CallSite.Target(System.Runtime.CompilerServices.Closure,System.Runtime.CompilerServices.CallSite,NzbDrone.Core.Messaging.Commands.CommandQueueManager,object,NzbDrone.Core.Messaging.Commands.CommandPriority,NzbDrone.Core.Messaging.Commands.CommandTrigger)
    at NzbDrone.Core.Messaging.Commands.CommandQueueManager.Push (System.String commandName, System.Nullable`1[T] lastExecutionTime, NzbDrone.Core.Messaging.Commands.CommandPriority priority, NzbDrone.Core.Messaging.Commands.CommandTrigger trigger) [0x000b7] in M:\BuildAgent\work\b69c1fe19bfc2c38\src\NzbDrone.Core\Messaging\Commands\CommandQueueManager.cs:95
    at NzbDrone.Core.Jobs.Scheduler.ExecuteCommands () [0x00043] in M:\BuildAgent\work\b69c1fe19bfc2c38\src\NzbDrone.Core\Jobs\Scheduler.cs:42
    at System.Threading.Tasks.Task.InnerInvoke () [0x0000f] in <71d8ad678db34313b7f718a414dfcb25>:0
    at System.Threading.Tasks.Task.Execute () [0x00010] in <71d8ad678db34313b7f718a414dfcb25>:0

    [Error] TaskExtensions: Task Error

     

    But in the log section in the actual docker it shows it not having been updated for 7 days?

  11. 13 hours ago, NGMK said:

    Well I build my server. The initial learning curb can be overwhelming but am so glad I decided to go this route rather than the Synology way. I was able to find cheap (169.99) WD Red Nas drives from Bestbuy 8TB each. Question I was initially planing on having a 16GB of RAm but one of may ram stick failed, the server is running fine now with 12GB DD3, at the moment i'm only running ruTorrent, Plex and PiHole, I do plan to play around with couple more dockers and even possibly a VM. Should I be lucking at a 32GB ram kit or not. Ram is expensive right but If I need it then it is what it is . 

    I haven't played around with vm's yet but I guess it depends on what you want to be running on it? If its just a small one for browsing the internet then you won't need much ram dedicated to it but if you want to do anything more powerful on it you should probably put more in :)

    • Upvote 1
  12. 32 minutes ago, Tward2 said:

    Thanks, that did it. I disabled the Plex docker and used mover before but I didn't disable the docker service. After I did that and pressed mover they are all spinning down now but the cache disc. What is the brand and size of a SSD cache drive that is recommended?

    The cache drive wont spin down if you have your dockers installed on it no matter what size or brand that it i. However it seems that majority of users use samsung 850 or 860 evo's as they are good for the price and very reliable. I would probably go for a 250gb but it depends on what you want to use it for

  13. On 5/4/2018 at 4:17 PM, Tward2 said:

    The heat is one of my main concerns I think that was because the drives weren't spinning down. I installed a cache drive yesterday and now all of them are spinning down except the parity and drive 1. How do I figure out what is keeping them spun up?

    Turn off all of your dockers to see whether the drives spin down or not, then start the dockers up again one by one to see which is writing to the array and keeping it spun up

  14. On 4/29/2018 at 2:29 PM, L0rdRaiden said:

    Sonnar have been adding epsiodes to ruTorrent (those are added in stop mode but I can live with this) when the download finished sonnar says "eppisode missing from disk" but the files are there.

    This was working in the past with other series, so now I don't understand why it doesn't work.

     

    2.thumb.PNG.32f848ef75cd3b50fed3c5a47e115745.PNG3.thumb.PNG.656120d4a7304b599a2bdd4b2655555f.PNG1.thumb.PNG.f68f4bd7620782dd34c86ef287b8a6c7.PNG

     

    
    18-4-29 15:24:54.2|Info|SceneMappingService|Updating Scene mappings
    18-4-29 15:24:55.3|Info|RefreshSeriesService|Updating Info for XXXXX
    18-4-29 15:24:55.3|Info|RefreshEpisodeService|Starting episode info refresh for: [296762][XXXXX]
    18-4-29 15:24:55.3|Info|RefreshEpisodeService|Finished episode refresh for series: [296762][XXXXX].
    18-4-29 15:24:55.3|Info|DiskScanService|Scanning disk for XXXXX
    18-4-29 15:24:55.3|Info|DiskScanService|Completed scanning disk for XXXXX
    18-4-29 15:24:55.5|Info|MetadataService|Series folder does not exist, skipping metadata creation

    XXXXX= hide on purpose

    7.PNG

    I may be wrong but I had similar issues when the files weren't named with their extension in the name e.g. .mkv or whatever the media type is

  15. On 5/5/2018 at 9:42 AM, bally12345 said:

    Will be sticking with the WD Reds but at £167.99 each I really wish I had another spare bay instead of taking out the perfectly functioning 4tb drive.

    Sent from my SM-G930F using Tapatalk
     

    Keep hold of it, you may be able to find a use for it in a different build or something. You could also sell it on the forums here to make some money back on it towards the new drive

  16. On 5/5/2018 at 12:15 AM, John_M said:

     

    Any differences there are mostly irrelevant for unRAID.

    Yes true the ironwolf coming out on top by just a few MB/s and if you really need extra MB/s you would probably be looking at different drives anyway