peppingc

Members
  • Posts

    19
  • Joined

  • Last visited

Everything posted by peppingc

  1. For those following I seem to have solved my issue using the guide here :- https://kodi.wiki/view/SMB Note I used the steps outlined for Device Name method rather than IP address (skip step 4 and go straight to step 5). It's not as simple as when SMB v1 was enabled and I could just browse to the relevant file path but is apparently a more secure way of sharing files.
  2. Thanks so much, just waiting for a parity check to finish and will do that. I tried Googling a solution but got lost in the confusing posts it returned. I'm assuming I can also turn off SMB 1.0/CIFS Client on the Winows 10 side when I apply this change. Edit2: I followed your suggested steps and turned off SMB 1.0/CIFS Client in Windows 10 but now when I browse for my media directories over on the UnRAID server I can not see the existing file paths, just a single entry for BORG.modem (Borg being the name of my UnRAID server. Is there anything else I need to do on the UnRAID server? Edit3: I can see the file paths if I open up the BORG.modem entry and can access the files that way however I can not see my other Windows based file sources on other PCs any more (other PCs are all still running Windows 7 though, not 10.
  3. I have recently built a HTPC running Kodi via Windows 10. In order to get the HTPC to see my SMB shares I have had to renable SMB 1.0/CIFS Client as per this article :- https://www.windowscentral.com/how-access-files-network-devices-using-smbv1-windows-10 I understand SMBv1 is no longer supported in Windows 10 by default due to security concerns but how do I make UnRAID support SMBv2?
  4. Thanks, that's what I thought from previous dabbling with UNRAID on my old N36L Microserver some time ago, just wanted to make sure it was still the same. I did check in the FAQs and online manual but couldn't see it mentioned there.
  5. With Unraid 6.8.3 once an array has been configured and the drive letters assigned (sda, sdb, etc) does it matter if the drives are moved around on the same controller? I have a HP Microserver Gen8 where the 4 HDDs are attached to a single SATA backplane attached to a single SAS cable connected directly to the motherboard. I have my 2 x 5tb HDDs located directly next door to each other in the first 2 slots and my 2 x 4TB HDDs in slot 3 and 4. I've noticed the 5TB HDDs run a little warm so wanted to shuffle them around to a 5/4/5/4 arrangement. The array is currently live and contains data on all drives, the 1st 5TB is my parity drive.
  6. Won't be an issue for me as the drive bays on my Gen8 Microserver are non-hot swap. Can't recall exactly which model it is but it only has 4 bays and is running a Pentium G2020T CPU.
  7. Thanks JorgeB, it is a Gen8 Proliant Microserver I found the issue discussed in that thread and have followed the suggested fix. I'm just awaiting a response to my question there before marking this as solved.
  8. ACPI Error Message Issues I was getting the same ACPI error messages detailed towards the start of this thread on my Gen8 Microserver which was already running the latest J06 BIOS dated 04/04/2019. I followed the advice given there namely ;- "This is essentially a misreading and does not affect anything but spamming your system log. To remove, disable "acpi_power_meter" module from the kernel by adding this line to the /boot/config/go file: rmmod acpi_power_meter" After a reboot I am now seeing this message in the system logs :- Nov 25 12:21:25 Borg kernel: ACPI: Early table checksum verification disabled Nov 25 12:21:25 Borg kernel: ACPI: RSDP 0x00000000000F4F00 000024 (v02 HP ) Nov 25 12:21:25 Borg kernel: ACPI BIOS Warning (bug): 32/64X length mismatch in FADT/Pm1aControlBlock: 16/32 (20180810/tbfadt-564) Nov 25 12:21:25 Borg kernel: ACPI BIOS Warning (bug): 32/64X length mismatch in FADT/Pm2ControlBlock: 8/32 (20180810/tbfadt-564) Nov 25 12:21:25 Borg kernel: ACPI BIOS Warning (bug): Invalid length for FADT/Pm1aControlBlock: 32, using default 16 (20180810/tbfadt-669) Nov 25 12:21:25 Borg kernel: ACPI BIOS Warning (bug): Invalid length for FADT/Pm2ControlBlock: 32, using default 8 (20180810/tbfadt-669) I have attached a copy of the post-reboot system log for your reference. Is there anything to worry about here or are we all good to go now? borg-syslog-20201125-0124.zip
  9. I have started transferring data overnight and am seeing a bunch of the following errors in the system log (BORG is my server's name). Nov 24 12:19:49 Borg kernel: ACPI Error: SMBus/IPMI/GenericSerialBus write requires Buffer of length 66, found length 32 (20180810/exfield-390) Nov 24 12:19:49 Borg kernel: ACPI Error: Method parse/execution failed \_SB.PMI0._PMM, AE_AML_BUFFER_LIMIT (20180810/psparse-514) Nov 24 12:19:49 Borg kernel: ACPI Error: AE_AML_BUFFER_LIMIT, Evaluating _PMM (20180810/power_meter-338) I am just using the default disk settings and the file system is XFS. Is there something I need to change?
  10. It was more than just a case of changing the file system. 2 of the drives came out of a previous unRAID server that had failed drives. I was just making sure these 2 drives were still good prior to putting them back into service as the old box had been sitting idle for some time. The other 2 drives were new drives and my understanding is it's good practice to stress test them at least once before putting them into service. It's all good, the drives precleared just fine and I have a functioning unRAID server again.
  11. Thanks for the explanation. I did not know it was possible to run VMs from a cache drive. At the moment this box is purely for file serving purposes but that may change as I learn more about what unRAID can do. How do you determine the most appropriate cache size for any given array? I have 2 x 4tb and 2 x 5tb drives in mine with 1 of the 5tb drives being a parity drive so I effectively have 13tb of usable space in my array.
  12. I'm assuming it will be used for buffering during large file transfers. I have a lot of data to migrate to this new server, the data is currently on various HDDs spread across multiple machines. Yes, will run with parity on a single drive.
  13. The Preclear tool would not allow me to select the drive for wiping until I unassigned it. In any event it's done now, 2 drives done and 2 to go.
  14. I'm in the process of building up a unRAID NAS box running on a HP Microserver Gen8 platform. I only have 4 3.5" HDD bays all taken up with data drives (2 x 4tb & 2 x 5tb). I can sneak in a cache drive if use a SATA SSD or 2.5" HDD, there is no room for any more 3.5" drives. I will probably only have a single cache drive and am wondering if a 1TB 5400rpm HDD will be enough or should I spend a bit more and go the SSD route? There is only a single SATA port free on the mobo and I will have to piggy back power off something else.
  15. That's okay, I'm going to clear the drives and reformat them anyway. Obviously I need to stop the array before I do this, I'm assuming I also need to unassign the drives.
  16. Thanks JorgeB! I have installed the plugin via community apps but I note it says you also need to separately install the script. Where can I find the script? Edit: Have just reread the 1st post on that thread and note the following "A) the bjp999 script is now included by default.". I can see a preclear.sh file under /boot/config/plugins/preclear.disk/ so I assume the script is included with the plugin install. Also, I note one of my unmountable drives went to XFS format when I formatted it, the other drives are all reiserfs. I'm assuming XFS is the new default filesystem. Since I don't care about the existing content of the drives I'm assuming it's okay just to zap them with a preclear and then format to XFS. Is that correct?
  17. I am in the process of building a new UnRAID server using v6.8.3. I have pulled some of the drives from my previous UnRAID server and am not concerned with retaining any existing data on them. The official documentation seems to suggest a preclear operation is not required but seeing as it's a new build with a mix of old and new drives I thought it might be prudent to clear all the drives before use, plus it would allow me to stress test all the drives before loading live data on to them. It's been a while since I have built an unRAID box so I need some help with locating the Preclear plugin and locating instructions on how to use it. Please point me in the right direction, all help appreciated.