Jump to content

papnikol

Members
  • Posts

    341
  • Joined

  • Last visited

Posts posted by papnikol

  1. First thank you very much for your willingness to help.

     

    I also learned about the concept of Rebuilding a drive onto itself. I would usually do a new config and accept the array. I hope to exploit it, hopefully, for one last time :)

    So, I removed drive 8 and its emulated counterpart seems to be working. I also mounted drive 8 as unassigned and it seems to work but, still, although I changed cables, the SMART test seems to get stuck at 90%, while it works for other drives. This is a bit strange, so I am thinking of using another drive for a rebuild.

     

    Does that make sense? Also, would it hurt if I simultaneously allowed the new 2nd parity to rebuild? That would allow me to be sooner protected from another failure...

  2.   

    Thanks a lot.

     

    3 hours ago, JorgeB said:

    Also see if disks 1, 2 and 3 share anything in common, like a miniSAS cable or power splitter.

    Disks 1, 2, 3 (along with 4) share both the same SATA splitter and miniSAS. I will have to investigate.

     

    Just a few clarifications:

    3 hours ago, JorgeB said:

    if the emulated disk is mounting

    what do you mean by that? since disk 8 is redballed, wouldn't the only way to mount it be to take it out of the array? Should I take it out of the array and try mounting with UD? (btw, I changed SATA cable and it still gets stuck at 90% of the short SMART test)

     

    3 hours ago, JorgeB said:

    you can rebuild on top.

    -What exactly do you mean by "rebuild on top"? Accept disk 8 as good and rebuild the new parity drive?
     

  3. Hi everyone,,

    I posted this question to r/unraid but I am posting it here too because it is somehow urgent:

     

    I have an array with dual parity. A few days ago I decided to replace both parity drives with larger ones (16TB). The first parity drive replacement went smoothly.

     

    The second parity sync stuck at some point and I had to reboot. When I came back one data drive was red balled. I run a short SMART test but it gets stuck at 90% (I tried other drives and they completed their short SMART test). This indicates that the drive might be problematic.

     

    So, my situation is: One of the 2 parity drives is un-synced and one data drive might need to be replaced. I also have a feeling that another disk might be on the verge of failing. What is the optimal solution:?

    1. replace the redballed disk with my previous parity drive (which was obviously removed from the array) and have the array simultaneously rebuild the new parity drive and the failed drive.
    2. 1st rebuild the new parity drive and then the failed data drive (so, 2 runs)
    3. Remove the new parity drive, accept the array as correct and check if the drive works. Then, if it works, add again the second parity drive for parity sync
    4. Some other solution I have not though of...

    Any input is welcome

    towerp-diagnostics-20240314-1206.zip

  4. 19 hours ago, JorgeB said:

    I guess I wasn't clear, having a PSU cable with 4 SATA plugs (or more) is perfectly fine, splitting a single SATA plug to 4 SATA plugs is not.

    No, you were, I get it now. So, a viable solution would be to use a splitter with only 2 SATA connectors? Again, do you think I could use 1 splitter on every connector of the PSU cable? like that (where every "[" is a 1->2 splitter):

     

    PSU
       |
       |    
       |----[
       |
       |----[
       |
       |----[
       |
       |----[





     

  5. 34 minutes ago, ChatNoir said:

    I tend to prefer Molex to Sata adapters.

    Generally the cables are bigger.

    And they are easier to find, but, generally, from all the comments in various forums, the common wisdom is that they should be avoided (although that is what I have been using up to now)

  6. There are a few but they are relatively expensive:
    https://de.pcpartpicker.com/products/power-supply/#A=550000000000,2050000000000&D=20&sort=price&page=1&E=5,14

     

    So, going back to the original question(s):
     

    Quote

     

    1. Is it safe to connect a 4x SATA expander to a 4x SATA cable (thus having 7 drives per cable in total)?

    2. Is it better to connect the expander at the 1st or the last connector of the SATA cable. Or is it of no importance?

     

     

  7. 13 minutes ago, whipdancer said:

    Are you talking about power splitters?

    951471377_71I7YgzS-NL._AC_SX425_PIbundle-2TopRight00_SH20_.jpg.d121d084e23231d7f38b6d25d764baba.jpg
     

     

    No, I am talking about this:

     

    pyo4sata_main.jpg.627ed1e977f0724d748939989750d9d9.jpg

     

    13 minutes ago, whipdancer said:

    Regardless, you will need enough power to deal with the spin-up power requirements. A relatively simple calculation to estimate.

    One reason you may want a more expensive PSU is to have separate rails. When powering that many drives, you want to make sure you don't starve something else because you have momentarily heavy demand.
     

     

    Thanks for the info, I am aware, but even many good quality 750W PSUs usually do not have enough connectors to accommodate 20 drives. Whenever I built a PC, the last thing I use to cut costs is the PSU, but, obviously, I am trying to avoid buying something that I dont need, hence the power extenders (I also happen to have them at hand).

     

  8. Hi everyone,

    I am looking for a PSU that will power many HDDs/SDDs (up to 20). I can buy a PSU with 6 cables, with 4x and 2x SATA connectors but these are higher wattage PSUs that are unnecessarily expensive for my purpose.

     

    Trying to avoid IDE -> 4x SATA cables, I wanted to try for my first time SATA expanders. So, I have 2 questions:
    1. Is it safe to connect a 4x SATA expander to a 4x SATA cable (thus having 7 drives per cable in total)?

    2. Is it better to connect the expander at the 1st or the last connector of the SATA cable. Or is it of no importance?

    Thanks in advance for your answers....

  9. So I tried various proposed stuff (and was away for a bit, sorry for taking too many days to get back to you after your prompt inputs).

     

    What I found out is by disabling docker, the problem was fixed. Additionally, I noticed in the dashboard that docker was using quite a lot of memory, when enabled. More particularly 67% of my 6GBs, which is almost 4GBs. I noticed that in my other unraid machine docker was also using ~4GB. I do not understand why that would be since all my containers were stopped. Maybe the memory is reserved by docker?

  10. 31 minutes ago, itimpi said:

    The ‘shfs’ process is what supports User Shares so that getting killed is why the User Shares disappear.  The question is why?
     

    You are likely to get better informed feedback if you attach your system’s diagnostics zip file to your next post in this thread taken after the problem occurs.

    Thanks a lot for the input.

    I will boot normally and when the problem occurs again (I am guessing soon), I will post the diagnostics.

     

    31 minutes ago, itimpi said:

    Have you tried booting in Safe Mode with Docker disabled to see if the problem still occurs?

    I might try that too, afterwards. A plugin might be the problem but I seriously doubt it has something to do with Docker since I am nor running automatically any of the installed plugins.

  11.   

    Hi everyone,

    I am using unraid v6.11.3. Suddenly (without me having made any significant changes), the user shares have started disappearing and they do not reappear until I reboot. Initially it happened once in a while but it keeps happening more and more often (3 times today). Here is what I tried (having also searched the forum for suggestions):

    1. Tried stopping and restarting the array - didn't work

    2. Changed Settings > Global Share Settings > Tunable (support Hard Links) to "No" - didn't work

    3. I have some Docker Containers but they are not running so I dont think they could be the problem

    4. Updated all apps - didn't work

    5. Checked the XFS filesystem in all disks

     

    EDIT: I just noticed an error in the logs:

    Quote

    kernel: Out of memory: Killed process 8995 (shfs) total-vm:569000kB, anon-rss:269704kB, file-rss:0kB, shmem-rss:524kB, UID:0 pgtables:588kB oom_score_adj:0

     

    Could this be the problem? My server has 6GB and I am running cache_dirs (folder caching). I also have a LOT of files in the server. I always had but they have been slowly increasing. Could I have reached a limit where the memory is somehow filled?

     

    Any help would be very welcome as my server has been rendered unusable.

  12. I just stumbled upon this this Silverstone "ECS07 - 5-Port SATA Gen3 6Gbps Non-RAID M.2 PCIe storage expansion card".

    I just wanted to mention it because it seems like a very interesting alternative. If a Mobo has 2 M.2 slots one could easily add 10 SATA ports. If they work as advertised, I would certainly prefer them to HBAs.

     

    One big caveat is that the price has not been released yet so I dont know if they cost 50 or 500 euros....

  13. On 6/13/2022 at 7:26 PM, jbartlett said:

    Click on the DiskSpeed Docker icon and open a console window.

     

    Copy-n-paste the following command after verifying your parity drive is still sdp and change it here if it is different.

    dd if=/dev/sdp of=/dev/null bs=1310720 skip=0 iflag=direct status=progress conv=noerror

     

    You should see it starting to copy data and updating every second with the progress. The MB/s value should quickly settle down with a typical 2-3 MB variance from one second to the next. When you press CTRL-C to abort (or let it read the entire drive), it should report something like the following:

    2081423360 bytes (2.1 GB, 1.9 GiB) copied, 9.00526 s, 231 MB/s^C
    1655+0 records in
    1654+0 records out
    2167930880 bytes (2.2 GB, 2.0 GiB) copied, 9.40043 s, 231 MB/s

     

     

     

    On 6/13/2022 at 8:33 PM, jbartlett said:

     

    Go to the tools page in UNRAID and click on the System Devices icon. Find your storage card and take note of the domain ID. It'll look something like "07:00.0" and be listed after the two hex numbers in brackets. Open a console window and enter replacing the domain id with your value.

    lspci -vv -s 07:00.0

     

    Look for lines starting with "LnkCap" and "LnkSta" that report a speed & width. Please copy-n-paste the results here.

     

    On 6/13/2022 at 8:42 PM, JorgeB said:

    It's reporting the controller link speed, PCIe 1.0 x4 is correct for a SASLP

     

    I hate it when people try to help me and I seemingly ignore them. I read your answers but I had to be away for >1 month. When I came back, 1 HDD was dead so I first had to replace it. Now the problem has disappeared for some reason.

     

    Thanks for your input and know that i did not ignore your input (at least intentionally)...

    • Like 1
    • Thanks 1
  14.  

     

    15 hours ago, John_M said:

     

    Are you using the Azure or Gray theme (I use the original White)? I notice that Unraid 6.10.3 will have some changes to help with compatibility with Firefox. They are already present in the -rc1:

     

     

    EDIT: Ah, I scrolled up and see that you seem to be using the Black theme. Are there any related nginx errors in the syslog?

     

     

    I did not notice any such errors in the systems log.

     

    Thanks for pointing me to the upcoming changes. I think I will wait for 6.10.3, so as not to bother anyone for no reason, and come back if the problem persists.

  15. 44 minutes ago, John_M said:

     

    When you get different results with different browsers the first thing to do is clear the cache, whitelist your server in any ad-blocking extension and make sure the browser itself is up to date. Does it work in private/incogneto mode?

     

    I had already done all the above but I did not think to try the private mode. unfortunately, it did not change anything.

     

    PS: Just to be sure, I tried a fresh Firefox portable but I get the same results...

  16. Hi, I upgraded from 6.9.2 straight to 6.10.2. I noticed that, after the upgrade, for some reason, the main window content takes a few seconds to load (and meanwhile it shows as in the attached file). This happens in Firefox but not in e.g. chrome.

     

    It is a bit annoying but minor, I am just mentioning because I dont think it has been already mentioned.

    main_unraid.jpg

  17. Hi,

    I have encountered an error that (from what I managed to find) I have not seen mentioned somewhere else. I was trying to benchmark this controller:

    
    MV64460/64461/64462 System Controller, Revision B
    
    Marvell Technology Group Ltd.
    RAID bus controller
    
    Type: Add-on Card in PCIe Slot 2 (x1 PCI Express)
    Current & Maximum Link Speed: 2.5GT/s width x4 (1 GB/s max throughput)
    Capabilities: storage pm msi pciexpress bus_master cap_list rom
    
    Port 1: sdj 8TB Seagate ST8000AS0002 Rev AR17  (Disk 13)
    Port 2: sdk 2TB Western Digital WD20EARX Rev 51.0AB51 (Disk 7)
    Port 3: sdl 4TB Western Digital WD40EFRX Rev 82.00A82 (Disk 9)
    Port 4: sdm 2TB Western Digital WD20EARS Rev 80.00A80 (Disk 3)
    Port 5: sdn 8TB Seagate ST8000DM004 Rev 0001 (Disk 12)
    Port 6: sdo 8TB Western Digital WD80EZAZ Rev 83.H0A83
    Port 7: sdp 8TB Western Digital WD80EZAZ Rev 83.H0A83(Parity)

    I should add that according to the manual for the mobo, this slot is suposed to be PCIe v.2  (PCI Express 2.0 x16 Slot (PCIE5 @ x4 mode) which seems different with what Diskspeed is reporting.

     

    and I get the following error (the error appears when the "Benchmarking Parity (sdp)" bar reaches 100%).. What could be causing it?

    Quote

    Benchmarking Parity (sdp):
    Lucee 5.2.9.31 Error (expression)
    Message     can't cast [dd:] string to a number value
    Stacktrace     The Error Occurred in
    /var/www/CustomTags.cfm: line 145

        143: <CFIF ListLen(CurrLine," ") GT 1 AND Find("record",CurrLine) EQ 0>
        144: <CFSET BytesRead=ListFirst(CurrLine," ")>
        145: <CFSET BytesDiff=BytesRead - LastBytesRead>
        146: <CFSET LastBytesRead=BytesRead>
        147: <CFIF Skip GT 0>

    called from /var/www/TestControllerBandwidth.cfm: line 171

        169: <CFFILE action="write" file="#BWDir#/#DriveID#_benchmark.sh" mode="766" output="#cmd#" addnewline="NO">
        170: <CFEXECUTE name="#BWDir#/#DriveID#_benchmark.sh" timeout="3060" />
        171: <CFSET Result=GetReadAvg(ResultsFN,Drive.Attrib.Configuration.RPM,2,MaxGap)>
        172: <CFSET Avg=ListFirst(Result,"|")>
        173: <CFSET Max=ListLast(Result,"|")>

    called from /var/www/DispController.cfm: line 97

        95: </CFIF>
        96: <CFELSE>
        97: <CFINCLUDE template="TestControllerBandwidth.cfm">
        98: <CFOUTPUT><br></CFOUTPUT>
        99: </CFIF>

    Java Stacktrace     lucee.runtime.exp.CasterException: can't cast [dd:] string to a number value
         at lucee.runtime.op.Caster.toDoubleValueViaDate(Caster.java:564)
         at lucee.runtime.op.Caster.toDoubleValue(Caster.java:540)
         at lucee.runtime.op.Caster.toDoubleValue(Caster.java:438)
         at lucee.runtime.op.Operator.minusRef(Operator.java:1004)
         at customtags_cfm$cf.udfCall1(/CustomTags.cfm:145)
         at customtags_cfm$cf.udfCall(/CustomTags.cfm)
         at lucee.runtime.type.UDFImpl.implementation(UDFImpl.java:107)
         at lucee.runtime.type.UDFImpl._call(UDFImpl.java:357)
         at lucee.runtime.type.UDFImpl.call(UDFImpl.java:226)
         at lucee.runtime.type.scope.UndefinedImpl.call(UndefinedImpl.java:803)
         at lucee.runtime.util.VariableUtilImpl.callFunctionWithoutNamedValues(VariableUtilImpl.java:756)
         at lucee.runtime.PageContextImpl.getFunction(PageContextImpl.java:1718)
         at testcontrollerbandwidth_cfm$cf.call(/TestControllerBandwidth.cfm:171)
         at lucee.runtime.PageContextImpl._doInclude(PageContextImpl.java:933)
         at lucee.runtime.PageContextImpl._doInclude(PageContextImpl.java:823)
         at lucee.runtime.PageContextImpl.doInclude(PageContextImpl.java:805)
         at dispcontroller_cfm$cf.call(/DispController.cfm:97)
         at lucee.runtime.PageContextImpl._doInclude(PageContextImpl.java:933)
         at lucee.runtime.PageContextImpl._doInclude(PageContextImpl.java:823)
         at lucee.runtime.listener.ClassicAppListener._onRequest(ClassicAppListener.java:66)
         at lucee.runtime.listener.MixedAppListener.onRequest(MixedAppListener.java:45)
         at lucee.runtime.PageContextImpl.execute(PageContextImpl.java:2464)
         at lucee.runtime.PageContextImpl._execute(PageContextImpl.java:2454)
         at lucee.runtime.PageContextImpl.executeCFML(PageContextImpl.java:2427)
         at lucee.runtime.engine.Request.exe(Request.java:44)
         at lucee.runtime.engine.CFMLEngineImpl._service(CFMLEngineImpl.java:1090)
         at lucee.runtime.engine.CFMLEngineImpl.serviceCFML(CFMLEngineImpl.java:1038)
         at lucee.loader.engine.CFMLEngineWrapper.serviceCFML(CFMLEngineWrapper.java:102)
         at lucee.loader.servlet.CFMLServlet.service(CFMLServlet.java:51)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:729)
         at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:292)
         at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:207)
         at org.apache.tomcat.websocket.server.WsFilter.doFilter(WsFilter.java:52)
         at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:240)
         at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:207)
         at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:212)
         at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:94)
         at org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:492)
         at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:141)
         at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:80)
         at org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:620)
         at org.apache.catalina.valves.RemoteIpValve.invoke(RemoteIpValve.java:684)
         at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:88)
         at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:502)
         at org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1152)
         at org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:684)
         at org.apache.tomcat.util.net.AprEndpoint$SocketProcessor.doRun(AprEndpoint.java:2527)
         at org.apache.tomcat.util.net.AprEndpoint$SocketProcessor.run(AprEndpoint.java:2516)
         at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
         at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
         at org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61)
         at java.lang.Thread.run(Thread.java:748)
     

     

    container run:

    Quote

    root@localhost:# /usr/local/emhttp/plugins/dynamix.docker.manager/scripts/docker run -d --name='DiskSpeed' --net='bridge' --privileged=true -e TZ="Europe/Athens" -e HOST_OS="Unraid" -e HOST_HOSTNAME="towerS" -e HOST_CONTAINERNAME="DiskSpeed" -l net.unraid.docker.managed=dockerman -l net.unraid.docker.webui='http://[IP]:[PORT:8888]/' -l net.unraid.docker.icon='https://raw.githubusercontent.com/jbartlett777/DiskSpeed/master/DiskSpeed.png' -p '18888:8888/tcp' -v '/mnt/user/appdata/DiskSpeed/':'/tmp/DiskSpeed':'rw' -v '/var/local/emhttp':'/var/local/emhttp':'ro' 'jbartlett777/diskspeed'

     

    I should also mention that i have this controller in the same mobo and it is benchmarked with no problems  (although it might also be misreported) :

    MV64460/64461/64462 System Controller, Revision B
    
    Super Micro Computer Inc (Marvell Technology Group Ltd.)
    SCSI storage controller
    
    Type: Onboard Controller
    Current & Maximum Link Speed: 2.5GT/s width x4 (1 GB/s max throughput)
    Capabilities: scsi pm msi pciexpress bus_master cap_list rom

     

  18. On 2/20/2021 at 10:53 AM, Alex R. Berg said:

    Hi yo all,

     


    I appreciate the thanks, you are welcome.

    I've done some more testing myself, and I have not seen any spin-ups. So also think it is working now, without scanning of user. It's really good to get rid of that as it was CPU intensive.

    Its relatively easy to test, by checking disk status in unRaids' Main dashbard. I can click the dots to spin disks down, and see if they are spun up. I then find folder on eg. disk3 and access it though user-share and see if it spins up. So far I have not reproduced any spin-ups on unRaid 6.8.3. So I think we are good, since it also sounds like it works for you out there.

    Regarding maintenance, life moves on for all of us. I'm not really maintaining it anymore, but fortunately it does not need any maintenance. I might fix it though if a problem occurs that I can fix.

    ==== Custom Mover Script ===
    Regarding @papnikol Q1, what I do if I have many disks? Well actually regardless of cache_dirs, I use my own custom mover before the official mover script. My custom mover moves certain folders to certain disks. That way I can keep all my most used files on disk1. I have 'ton' of files that consumes little space by todays standard, whereas mostly movies and backups consume huge chunks. So I just keep all those important things on disk1 and configured unRaid to never spin down disk1. It works far better for me than UnRaids default strategy. If my script hasn't chosen the location for a folder, it will be moved by the official mover. I only move to my allocated disks by custom mover when there's at least 50 GB available, and fall back to default strategy, moving to other disks. When space becomes available, my script moves it to correct disk later. All without me having to do anything besides the initial allocation of which folder goes where. If I want to move a share from one disk to another, I typically also just add a rule, and let it sort it out in the background automatically 

    My custom mover also reduces risk of loss, if I lose two disks in unRaid. Because if a git repository/program is spread across many disks, I'll probably lose it all if one of the disks data is lost. UnRaid GUI cannot do that, while still falling back to using any disks, when I go full.



     

     

    custom_mover 8.11 kB · 4 downloads move_if_idle 743 B · 1 download mover_all 113 B · 1 download

     

    Sorry @Alex R. Berg, somehow I missed your comment (and, again, thank you for your contribution).

     

    Your strategy is useful for cases like yours, but for cases like mine, where I cannot foresee which file I will need to use, It is not very helpful. Additionally - and this might be less important - in my windows machine I use Everything for indexing/searching, and cache_dirs works well with it... Still, thanks for sharing your scripts, one additional value for me is that they can help improve my shell scripting....
     

  19. 11 hours ago, Frank1940 said:

     

    The 'Joe' is actually  @Joe L..  He was very active in supporting Unraid for many years but he has been largely inactive since his last post in about 2016.  In that post, he stated that he had been using 'sed' for forty years.  I suspect that he is now retired and hopefully enjoying life!  He was a real craftsman in the writing of shell scripts!!!

     

     

    I actually remember @Joe L., I had one or two very pleasurable interactions with him.

     

    I both hope that he is having a good time and also that @Alex R. Berg keeps maintaining cache_dirs

  20. 11 hours ago, Alex R. Berg said:

    I'm the last active maintainer of the plugin cache-dir script ...

     

     

    @Alex R. Bergyou have a sincere "thank you" from me.

     

    just 2 questions for anyone invloved:

    1. The thing is, I think cache_dirs is very important (at least to me). If you have many disks you have to wait a long time for a listing. What do people who do not use it do? Is there an alternative?
    2. Why does Joe not still maintain it? Is there an "official" reason i have missed?
    • Like 1
  21. Hi, I am having a similar problem, Mobo is booting, 2 green LEDs turn on but I have no image from the GPU.

     

    There might be an incompatibility but shouldn't the Dr Debug LED show something? It never turns on.

    Does anyone know if the Dr Debug LED always shows something or only when there is an error?

×
×
  • Create New...