May 02 10:34:23 ********** iscsid[1854]: iSCSI daemon with pid=1857 started!
May 02 10:34:27 ********** iscsid[1854]: connect to 10.10.200.105:3260 failed (No route to host)
May 02 10:34:27 ********** iscsid[1854]: connect to 10.10.10.10:3260 failed (No route to host)
May 02 10:34:27 ********** iscsid[1854]: connect to 10.10.200.100:3260 failed (No route to host)
May 02 10:34:27 ********** iscsid[1854]: connect to 10.10.200.103:3260 failed (No route to host)
May 02 10:34:27 ********** iscsid[1854]: connect to 10.10.200.101:3260 failed (No route to host)
May 02 10:34:27 ********** iscsid[1854]: connect to 10.10.200.102:3260 failed (No route to host)
May 02 10:34:28 ********** kernel: ixgbe 0000:04:00.0 enp4s0f0: NIC Link is Up 10 Gbps, Flow Control: None
May 02 10:34:28 ********** kernel: IPv6: ADDRCONF(NETDEV_CHANGE): enp4s0f0: link becomes ready
May 02 10:34:28 ********** kernel: ixgbe 0000:04:00.1 enp4s0f1: NIC Link is Up 10 Gbps, Flow Control: None
May 02 10:34:28 ********** kernel: vmbr0: port 1(enp4s0f1) entered blocking state
May 02 10:34:28 ********** kernel: vmbr0: port 1(enp4s0f1) entered forwarding state
May 02 10:34:28 ********** kernel: IPv6: ADDRCONF(NETDEV_CHANGE): vmbr0: link becomes ready
May 02 10:34:34 ********** iscsid[1854]: connect to 10.10.200.105:3260 failed (No route to host)
May 02 10:34:34 ********** iscsid[1854]: connect to 10.10.10.10:3260 failed (No route to host)
May 02 10:34:34 ********** iscsid[1854]: connect to 10.10.200.100:3260 failed (No route to host)
May 02 10:34:34 ********** iscsid[1854]: connect to 10.10.200.103:3260 failed (No route to host)
May 02 10:34:34 ********** iscsid[1854]: connect to 10.10.200.101:3260 failed (No route to host)
May 02 10:34:34 ********** iscsid[1854]: connect to 10.10.200.102:3260 failed (No route to host)
May 02 10:34:41 ********** iscsid[1854]: connect to 10.10.200.105:3260 failed (No route to host)
May 02 10:34:41 ********** iscsid[1854]: connect to 10.10.10.10:3260 failed (No route to host)
May 02 10:34:41 ********** iscsid[1854]: connect to 10.10.200.100:3260 failed (No route to host)
May 02 10:34:41 ********** iscsid[1854]: connect to 10.10.200.103:3260 failed (No route to host)
May 02 10:34:41 ********** iscsid[1854]: connect to 10.10.200.101:3260 failed (No route to host)
May 02 10:34:41 ********** iscsid[1854]: connect to 10.10.200.102:3260 failed (No route to host)
May 02 10:34:48 ********** iscsid[1854]: connect to 10.10.200.105:3260 failed (No route to host)
May 02 10:34:48 ********** iscsid[1854]: connect to 10.10.10.10:3260 failed (No route to host)
May 02 10:34:48 ********** iscsid[1854]: connect to 10.10.200.100:3260 failed (No route to host)
May 02 10:34:48 ********** iscsid[1854]: connect to 10.10.200.103:3260 failed (No route to host)
May 02 10:34:48 ********** iscsid[1854]: connect to 10.10.200.101:3260 failed (No route to host)
May 02 10:34:48 ********** iscsid[1854]: connect to 10.10.200.102:3260 failed (No route to host)
May 02 10:34:55 ********** iscsid[1854]: connect to 10.10.200.105:3260 failed (No route to host)
May 02 10:34:55 ********** iscsid[1854]: connect to 10.10.10.10:3260 failed (No route to host)
May 02 10:34:55 ********** iscsid[1854]: connect to 10.10.200.100:3260 failed (No route to host)
May 02 10:34:55 ********** iscsid[1854]: connect to 10.10.200.103:3260 failed (No route to host)
May 02 10:34:55 ********** iscsid[1854]: connect to 10.10.200.101:3260 failed (No route to host)
May 02 10:34:55 ********** iscsid[1854]: connect to 10.10.200.102:3260 failed (No route to host)
May 02 10:35:00 ********** systemd[1]: Starting Proxmox VE replication runner...
May 02 10:35:00 ********** kernel: scsi host11: iSCSI Initiator over TCP/IP
May 02 10:35:00 ********** kernel: scsi host12: iSCSI Initiator over TCP/IP
May 02 10:35:00 ********** kernel: scsi host13: iSCSI Initiator over TCP/IP
May 02 10:35:00 ********** kernel: scsi host14: iSCSI Initiator over TCP/IP
May 02 10:35:00 ********** kernel: scsi host15: iSCSI Initiator over TCP/IP
Yes i see but i don't want to edit default iscsi.service adding a sleep time... it seems a bug in my opinion.
hmm - it does work in my local testsetup here without any modifications:
```
May 02 19:21:46 pve-local-03 systemd[1]: Reached target Network.
May 02 19:21:46 pve-local-03 systemd[1]: Reached target Network is Online.
May 02 19:21:46 pve-local-03 systemd[1]: Starting iSCSI initiator daemon (iscsid)...
May 02 19:21:46 pve-local-03 systemd[1]: Starting LXC network bridge setup...
May 02 19:21:46 pve-local-03 iscsid[1183]: iSCSI logger with pid=1194 started!
May 02 19:21:46 pve-local-03 systemd[1]: iscsid.service: Failed to read PID from file /run/iscsid.pid: Invalid argument
May 02 19:21:46 pve-local-03 iscsid[1194]: iSCSI daemon with pid=1195 started!
May 02 19:21:46 pve-local-03 systemd[1]: Started iSCSI initiator daemon (iscsid).
May 02 19:21:46 pve-local-03 systemd[1]: Starting Login to default iSCSI targets...
```
please post the output of:
* `cat /etc/network/interfaces`
* the complete log for the boot
* `systemctl -a |grep -i iscsi`
* ` dpkg -l |grep -i iscsi`
do you have any other specifics in your setup ? (e.g. using multipath)?
root@*********:~# cat /etc/network/interfaces
# network interface settings; autogenerated
# Please do NOT modify this file directly, unless you know what
# you're doing.
#
# If you want to manage part of the network configuration manually,
# please utilize the 'source' or 'source-directory' directives to do
# so.
# PVE will preserve these directives, but will NOT its network
# configuration from sourced files, so do not attempt to move any of
# the PVE managed interfaces into external files!
source /etc/network/interfaces.d/*
auto lo
iface lo inet loopback
auto enp4s0f0
iface enp4s0f0 inet static
address 10.10.200.154
netmask 255.255.255.0
gateway 10.10.200.1
dns-nameservers 10.10.200.1
dns-search ***********
iface enp2s0f0 inet manual
iface enp2s0f1 inet manual
iface enp4s0f1 inet manual
auto vmbr0
iface vmbr0 inet static
address 0.0.0.0
netmask 255.255.255.0
bridge-ports enp4s0f1
bridge-stp off
bridge-fd 0
bridge-vlan-aware yes
bridge-vids 2-4094
#Bridge ALL VLAN
root@********:~# systemctl -a |grep -i iscsi
dev-disk-by\x2did-lvm\x2dpv\x2duuid\x2dx0DNS8\x2dywk9\x2dpuQn\x2dUa68\x2dsfD2\x2d1L36\x2dh0Wo3f.device loaded active plugged iSCSI_Storage
dev-disk-by\x2did-scsi\x2d1SYNOLOGYiSCSI_Storage:dfd0ada2\x2d87bc\x2d41af\x2db5e9\x2d7536c904c1ba.device loaded active plugged iSCSI_Storage
dev-disk-by\x2did-scsi\x2d36001405dfd0ada2d87bcd41afdb5e9d7.device loaded active plugged iSCSI_Storage
dev-disk-by\x2did-scsi\x2dSSYNOLOGY_iSCSI_Storage_dfd0ada2\x2d87bc\x2d41af\x2db5e9\x2d7536c904c1ba.device loaded active plugged iSCSI_Storage
dev-disk-by\x2did-wwn\x2d0x6001405dfd0ada2d87bcd41afdb5e9d7.device loaded active plugged iSCSI_Storage
dev-disk-by\x2dpath-ip\x2d10.10.200.105:3260\x2discsi\x2diqn.2000\x2d01.com.synology:storage\x2d1.5ae93b244c\x2dlun\x2d0.device loaded active plugged iSCSI_Storage
dev-sdb.device loaded active plugged iSCSI_Storage
sys-devices-platform-host11-session1-target11:0:0-11:0:0:0-block-sdb.device loaded active plugged iSCSI_Storage
iscsid.service loaded active running iSCSI initiator daemon (iscsid)
open-iscsi.service loaded active exited Login to default iSCSI targets
root@*********:~# dpkg -l |grep -i iscsi
ii libiscsi7:amd64 1.17.0-1.1 amd64 iSCSI client shared library
ii open-iscsi 2.0.874-3~deb9u1 amd64 iSCSI initiator tools
that is not necessary/probably won't work - please remove these 2 lines.address 0.0.0.0 netmask 255.255.255.0
that is not necessary/probably won't work - please remove these 2 lines.
the dpkg and systemctl output look like in my setup.
please provide the output of `multipath -ll`
and the (anonymized) log would still be needed.
root@********:~# multipath -ll
3600c0ff00028252320c3655c01000000 dm-1 HP,MSA 1040 SAN
size=3.6T features='2 queue_if_no_path retain_attached_hw_handler' hwhandler='1 alua' wp=rw
|-+- policy='service-time 0' prio=50 status=active
| |- 12:0:0:3 sde 8:64 active ready running
| `- 14:0:0:3 sdk 8:160 active ready running
`-+- policy='service-time 0' prio=10 status=enabled
|- 13:0:0:3 sdh 8:112 active ready running
`- 15:0:0:3 sdn 8:208 active ready running
3600c0ff00028252304676c5901000000 dm-2 HP,MSA 1040 SAN
size=2.7T features='2 queue_if_no_path retain_attached_hw_handler' hwhandler='1 alua' wp=rw
|-+- policy='service-time 0' prio=50 status=active
| |- 14:0:0:2 sdj 8:144 active ready running
| `- 12:0:0:2 sdd 8:48 active ready running
`-+- policy='service-time 0' prio=10 status=enabled
|- 13:0:0:2 sdg 8:96 active ready running
`- 15:0:0:2 sdm 8:192 active ready running
3600c0ff000282523ff6c0d5a01000000 dm-0 HP,MSA 1040 SAN
size=2.7T features='2 queue_if_no_path retain_attached_hw_handler' hwhandler='1 alua' wp=rw
|-+- policy='service-time 0' prio=50 status=active
| |- 12:0:0:1 sdc 8:32 active ready running
| `- 14:0:0:1 sdi 8:128 active ready running
`-+- policy='service-time 0' prio=10 status=enabled
|- 13:0:0:1 sdf 8:80 active ready running
`- 15:0:0:1 sdl 8:176 active ready running
root@**********:~# journalctl | grep pvesr
May 03 07:41:01 ********** pvesr[2076]: ipcc_send_rec[1] failed: Connection refused
May 03 07:41:01 ********** pvesr[2076]: ipcc_send_rec[2] failed: Connection refused
May 03 07:41:01 ********** pvesr[2076]: ipcc_send_rec[3] failed: Connection refused
May 03 07:41:01 ********** pvesr[2076]: Unable to load access control list: Connection refused
May 03 07:41:01 ********** systemd[1]: pvesr.service: Main process exited, code=exited, status=111/n/a
May 03 07:41:01 ********** systemd[1]: pvesr.service: Unit entered failed state.
May 03 07:41:01 ********** systemd[1]: pvesr.service: Failed with result 'exit-code'.
May 03 07:42:00 ********** pvesr[2511]: ipcc_send_rec[1] failed: Connection refused
May 03 07:42:00 ********** pvesr[2511]: ipcc_send_rec[2] failed: Connection refused
May 03 07:42:00 ********** pvesr[2511]: ipcc_send_rec[3] failed: Connection refused
May 03 07:42:00 ********** pvesr[2511]: Unable to load access control list: Connection refused
May 03 07:42:00 ********** systemd[1]: pvesr.service: Main process exited, code=exited, status=111/n/a
May 03 07:42:00 ********** systemd[1]: pvesr.service: Unit entered failed state.
May 03 07:42:00 ********** systemd[1]: pvesr.service: Failed with result 'exit-code'.
hm it seems the service is lacking the dependency on pmxcfs - please open an enhancement request at https://bugzilla.proxmox.comFAILED TO START PROXMOX VE REPLICATION RUNNER
in my opinion pvesr.service start before network is ready:
maybe it's a problem with the huawei servers then (their 10G nics) - please try to upgrade all firmwares to the latest versions.only 2 huawei server have the same problem (huawei server have 10GBE ethernet card)
PVE parses /etc/network/interfaces for presenting it in the GUI.Is there a reason for not using systemd-networkd for proxmox network configuration?
hm it seems the service is lacking the dependency on pmxcfs - please open an enhancement request at https://bugzilla.proxmox.com
However these warning messages are not a problem (it would be a problem if you have storage replication configured and they occur after the system is fully booted up)
maybe it's a problem with the huawei servers then (their 10G nics) - please try to upgrade all firmwares to the latest versions.
do the other servers use 1g or 10g interfaces
PVE parses /etc/network/interfaces for presenting it in the GUI.