caplam

Members
  • Posts

    333
  • Joined

  • Last visited

Posts posted by caplam

  1. thanks. It looks like it's the right diag. 

    Changed the port to the ahci controller for theses 2 disks (they were the only ones on the scu controller).

    No issue at booting and currently running diskspeed benchmark. 

    Never had a glitch in years with the scu controller. Luckily i don't need these 4 ports. 

    What drove you to scu in diags ?

    I made the same deduction but only when i saw that the 2 drives were the only one attached to the scu.

  2. Hello,

     

    Recently i repurposed my main server into a backup server with a plus license.

    The array has 7 disks with no parity. Cache pool is 2x1Tb ssd btrfs pool.

    I have a strange issue (see the picture). 2 disks are visible both as array disks and unassigned devices.

    I can't stop the array. 

    1 disk has read errors. For the other one with issue i don't know. Perhaps a connection. 

    Before repurposing the server disks had no problems. 2 disks are in a sas das (disks 6&7). The other ones are on internal sata ports.

     

     

     

     

     

    Capture d’écran 2024-04-10 à 13.58.06.png

    z620-diagnostics-20240410-1357.zip

  3. back in unraid 6.8.3 i had a similar issue. Docker was running from a btrfs pool. Writes on the pool killed my ssds in 8 months.

    I managed to lower writes by

    - using directory for docker instead of image

    - use a single xfs pool

    - add --no-healthcheck as extra parameters for most of my containers

    - and use a ram disk for dockers logs. This one is great thanks to @mguttlink to the topic

     

    Besides of this i also bought longer life ssds. Depending of my activity i have around 50 running containers.

  4. My new unraid server is a ms01 so i want to do better usage of networking interfaces (2x2,5G 2x10G). one 2,5G is not used reserved for vpro

    My old unraid server is now a backup server. (2x1G 1x10G). one 1G is not used reserved for vpro.

    I also want to improve my networking skills. I bought some chinese switchs for upgrading to 10G (definitely too expensive to keep an all unifi gear for this kind of stuff)

    The simplest way to do is to use 10G interfaces for both servers. Eventually make a 2x10G bonding on main server.

    But as i want to learn networking i plan to create vlans.

    If i understood correctly unraid doesn't like to have more than one interface in a network.

    I could setup main server as follow:

    1x10G :etho default vlan all unraid traffic

    1x10G: eth1 vlan10 traffic to backup server

     

    backup server

    1G: eth0 default vlan all traffic 

    10G: eth1 vlan10 traffic from main server

     

    I tried with direct connection between the 2 servers (no vlan, no dhcp, no gateway, static adresses). It works ok at 2Gbs with luckybackup on main server (without cache involved).

    As the 2 servers won't be in the same room i will have a backbone fiber link between  switches. 

    Network is as follow:

    switch0--------10G-------->switch3-------10G----->switch4

    there is another link:    switch0-------1G-------->unifi dream router------>ISP

    switch0 is in the basement

    switch3 is on 3rd floor

    switch4 is on 4rd floor

    Am i correct to assume i have a trunk (to allow all vlans) to create on 1 port of each switch ?

    For the 2 servers i will have 1 interface in default vlan and 1 in vlan 10. To put an interface in a vlan i can make it with the switch config.

    For what use is the option in unraid "enable vlans" ?  can you create subinterfaces in differents vlans?

    Later when i will be more familiar with vlans i plan to creates others: iot, voip, cameras.

     

    Is it a correct way to enhance my network usage and security ?

    What should i do differently?

     

     

  5. There was a time i had many devices and servers on my network. All that stuff was quite power hungry.

    I moved and dropped most of the stuff i had. I kept network devices (unifi gear ) that i'm willing to sell to have smaller and savier devices.

    I'm replacing my main unraid server which is a HP Z620 with sfp connect x3 10Gb card with a small minisforum ms01 (2xSFP+, 2 x 2.5G).

    I use one of the 2 2,5G port fort vpro management (corresponding interface is down in unraid).

    In my basement i have a switch (unifi 48 poe) with uplink to 1st and 3rd floor (1G link from switch sfp cages to transceivers). There is only 5 ports on 52 used and it consumes around 65W.

    On the basement switch there will be the minisforum MS01 and a rpi.

    The Z620 will be moved to third floor and will be a backup server. As a good geek i'd like to leverage the capacity of ms01.

    So i was thinking on using a 10G interface for dockers, vms, unraid gui and the other 10G interface for the link with backup server. 

    Is it possible ? and does it make sense?

    i would need to replace the us48 switch with a smaller one with at least 4sfp+ cages (hassivo seems to be a good option) and the 3rd floor transceiver with a  sfp+ switch.

     

  6. After updating to 6.12.8 my unraid gui is bound  on port 80. The problem is that it's set on port 3080. It's also bound on port 3080.

    As a result nginx proxy manager can't respond and my nextcloud instance can't be joined.

     

     

    godzilla-diagnostics-20240304-1636.zip

     

    edit: netstat doesn't show a socket listening to port 80. it's weird.

     

    edit2: after a reboot things are back to normal.

     

  7. My plan is to move all disks and ssds from the old server to the new. The new server is going to replace the old; so normally i have not to transfer data across the network.

    I want to keep everything (shares, dockers, vms,...). I have to deal with different hardware for docker (plex and tdarr use quadro p400 on the old server for hardware encoding and it will be quicksync on the new one). I have no hardware passed through to the vms.

    On the old server parity, disks 1,2&3 are on internal sata ports; the others in a sas enclosure. 

    On the new servers all spinners will be in a external sas enclosure.

    Then later configure the old as a backup server with other disks probably as unassigned drives.

    Of course i'm aware i can't have 2 servers with same ip (i have dhcp reservation on the router). I don't need both servers to run at the same time.

     

  8. Few answers here. I will rephrase.

     

    I have 3 vm, 50 dockers, 6 disks and 3 ssd to move from one server to another.

    Hardware is quite different. I will come from a bi xeon E5-2650V2 with quadro P400 to minisforum MS-01 with i9-12900H.

    I have 2 license keys: 1 pro from the old server and 1 plus not yet installed. I'd rather keep the ip of the old server for the new one so i have not to reconfigure everything.

    I will use the old server as backup server.

    What is the best way to do this?

  9. Hello,

    I have a server with a pro key.

    Array has 5 data drives + 2 parity

    There is also a 2 ssd btrfs pool and 1 ssd xfs pool.

    The server is a power hungry HP Z620 with a 9207-8e hba and a das attached.

    Backups are made on unassigned drives (btrfs) and on a external usb (exfat) attached to my macbook air.

    I moved and now my server is in the basement not very accessible. I have no ipmi and intel amt vpro is not very convenient (no kvm) on the Z620.

    For power savings purpose i've just bought a minisforum ms01 and a unraid plus key to run my unraid server. 

    The plus key is enough for my arraid and pools.

    I plan to repurpose the Z620 as backup server.

    The Z620 has 128Gb ram.

    The ms01 will have 64Gb ram and also a 9300-8e hba.

    What is the best way to set-up the ms01 to make it my principal server ?

     

    edit: in the Z620 i have a nvidia P400 used for transcoding in plex and tdarr

    in the ms01 i will use quicksync.

    I want my vm and dockers to be configured from the 1st server.

    So in these conditions what's best?

    -configure the ms01 as new server with new key, the later reconfigure backup server 

    -use the actual usb key for the new server

  10. Hello,

    i'm facing a problem i can't solve.

     

    My unraid server has 3 nics. 2 are integrated and i keep only one active for amt purpose.

    The 3rd is a mellanox 10G bound to eth0.

    Bridging and bonding are inactive.

    For docker i use custom network on eth0 with macvtap.

     

    My network gear is unifi. The gateway is an usg3.

    I want my unraid on 192.168.1.50 reservation is made in unifi controller.

    In unifi controller is see eth0. It takes the right ip but after some time the ip change to 192.168.1.34.

    This ip (34) is also the adresse of an interface on the mellanox card which is listed in unraid as vhost0 interface.

    What is this interface ? 

    Why eth0 keeps changing ip ?

    Can i delete the route to my lan with vhost0 as gateway as it has a smaller metric than the one with eth0 as gateway? 

    godzilla-diagnostics-20240110-1202.zip

  11. the settings are ok.

    Several reboots fixed the problem but that's strange.

    I may have a problem on my server. I had to reboot several times to have the server back online. 

    My server is on the basement and i'm on the third floor. I don't have ipmi only intel amt which only permits reboot. 

  12. This night my unraid server crashed. This morning i couldn't access it nor with webgui nor with ssh. It only answered ping. So i had to hard reboot.

    When it came back online my NPM docker refused to start because port 80 is already in use.

    typing "netstat -tulpn | grep 80" tells me that unraid is listenning to port 80  with the same pid that used for webgui which is set to port 3080.

    Whenever i kill this pid it's started again and listen to port 80 and 3080. The process is nginx worker.

    I never had this problem before.

    Could this be a bug?

     

    edit: ihave this lines in syslog at the time of reboot:

     10:43:47 godzilla nginx: 2024/01/04 10:43:47 [emerg] 15223#15223: bind() to 127.0.0.1:3080 failed (98: Address already in use)
    Jan  4 10:43:47 godzilla nginx: 2024/01/04 10:43:47 [emerg] 15223#15223: bind() to 192.168.178.50:3080 failed (98: Address already in use)
    Jan  4 10:43:47 godzilla nginx: 2024/01/04 10:43:47 [emerg] 15223#15223: bind() to 127.0.0.1:3080 failed (98: Address already in use)
    Jan  4 10:43:47 godzilla nginx: 2024/01/04 10:43:47 [emerg] 15223#15223: bind() to 192.168.178.50:3080 failed (98: Address already in use)
    Jan  4 10:43:48 godzilla nginx: 2024/01/04 10:43:47 [emerg] 15223#15223: bind() to 127.0.0.1:3080 failed (98: Address already in use)
    Jan  4 10:43:48 godzilla nginx: 2024/01/04 10:43:47 [emerg] 15223#15223: bind() to 192.168.178.50:3080 failed (98: Address already in use)
    Jan  4 10:43:48 godzilla nginx: 2024/01/04 10:43:47 [emerg] 15223#15223: bind() to 127.0.0.1:3080 failed (98: Address already in use)
    Jan  4 10:43:48 godzilla nginx: 2024/01/04 10:43:47 [emerg] 15223#15223: bind() to 192.168.178.50:3080 failed (98: Address already in use)
    Jan  4 10:43:49 godzilla nginx: 2024/01/04 10:43:47 [emerg] 15223#15223: bind() to 127.0.0.1:3080 failed (98: Address already in use)
    Jan  4 10:43:49 godzilla nginx: 2024/01/04 10:43:47 [emerg] 15223#15223: bind() to 192.168.178.50:3080 failed (98: Address already in use)

    but webgui is accessible trough port 80

     

    edit2: the webgui is also accessible trough port 80.  i don't understand what's happenning.