Hey all,
We have stood up a 3 node Proxmox cluster, each node with 1.2T local storage for gust servers. Each of the nodes is running Debian 10.
We have also stood up a 6 node Ceph cluster with ~400G of storage. Each of these nodes is running CentOS 8
We have mounted Ceph onto two nodes on the cluster for testing using RDB. (End goal is a poor-man's fail-over.) Steps used...
a) On Proxmox node, created a /etc/pve/priv/ceph
b) Copied the /etc/ceph/ceph.client.admin.keyring from one of the monitor servers to the above directory on the proxmox server.
c) On one of the Proxmox cluster nodes GUI's I went to Datacenter>Storage and selected add. The name for the storage is the same as the name of the keyring I copied over. I put all three monitors in the list.
The proxmox server seem to be happy with that.
This configuration allows us to create a guest machine using the external Ceph storage but every time we attempt to install any OS on the guest machine, the proxmox host reboots. Below is what we have been able to capture from the logs just prior to the reboot.
There is no evidence of a panic, just a reboot. I would make sense if the guest had issues but why would it reboot the host?
I am continuing to investigate what the issue might be but any direction would be greatly appreciated. If there are any additional configurations needed, please let me know and I will append them.
Thank you,
DHM
We have stood up a 3 node Proxmox cluster, each node with 1.2T local storage for gust servers. Each of the nodes is running Debian 10.
We have also stood up a 6 node Ceph cluster with ~400G of storage. Each of these nodes is running CentOS 8
We have mounted Ceph onto two nodes on the cluster for testing using RDB. (End goal is a poor-man's fail-over.) Steps used...
a) On Proxmox node, created a /etc/pve/priv/ceph
b) Copied the /etc/ceph/ceph.client.admin.keyring from one of the monitor servers to the above directory on the proxmox server.
c) On one of the Proxmox cluster nodes GUI's I went to Datacenter>Storage and selected add. The name for the storage is the same as the name of the keyring I copied over. I put all three monitors in the list.
The proxmox server seem to be happy with that.
This configuration allows us to create a guest machine using the external Ceph storage but every time we attempt to install any OS on the guest machine, the proxmox host reboots. Below is what we have been able to capture from the logs just prior to the reboot.
Jul 08 15:38:27 {HOSTNAME} systemd[1]: Started 122.scope.
Jul 08 15:38:27 {HOSTNAME} systemd-udevd[30890]: Using default interface naming scheme 'v240'.
Jul 08 15:38:27 {HOSTNAME} systemd-udevd[30890]: link_config: autonegotiation is unset or enabled, the speed and duplex are not writable.
Jul 08 15:38:27 {HOSTNAME} systemd-udevd[30890]: Could not generate persistent MAC address for tap122i0: No such file or directory
Jul 08 15:38:28 {HOSTNAME} kernel: device tap122i0 entered promiscuous mode
Jul 08 15:38:28 {HOSTNAME} systemd-udevd[30890]: link_config: autonegotiation is unset or enabled, the speed and duplex are not writable.
Jul 08 15:38:28 {HOSTNAME} systemd-udevd[30890]: Could not generate persistent MAC address for fwbr122i0: No such file or directory
Jul 08 15:38:28 {HOSTNAME} systemd-udevd[30890]: link_config: autonegotiation is unset or enabled, the speed and duplex are not writable.
Jul 08 15:38:28 {HOSTNAME} systemd-udevd[30890]: Could not generate persistent MAC address for fwpr122p0: No such file or directory
Jul 08 15:38:28 {HOSTNAME} systemd-udevd[30881]: link_config: autonegotiation is unset or enabled, the speed and duplex are not writable.
Jul 08 15:38:28 {HOSTNAME} systemd-udevd[30881]: Using default interface naming scheme 'v240'.
Jul 08 15:38:28 {HOSTNAME} systemd-udevd[30881]: Could not generate persistent MAC address for fwln122i0: No such file or directory
Jul 08 15:38:28 {HOSTNAME} kernel: fwbr122i0: port 1(fwln122i0) entered blocking state
Jul 08 15:38:28 {HOSTNAME} kernel: fwbr122i0: port 1(fwln122i0) entered disabled state
Jul 08 15:38:28 {HOSTNAME} kernel: device fwln122i0 entered promiscuous mode
Jul 08 15:38:28 {HOSTNAME} kernel: fwbr122i0: port 1(fwln122i0) entered blocking state
Jul 08 15:38:28 {HOSTNAME} kernel: fwbr122i0: port 1(fwln122i0) entered forwarding state
Jul 08 15:38:28 {HOSTNAME} kernel: vmbr0: port 5(fwpr122p0) entered blocking state
Jul 08 15:38:28 {HOSTNAME} kernel: vmbr0: port 5(fwpr122p0) entered disabled state
Jul 08 15:38:28 {HOSTNAME} kernel: device fwpr122p0 entered promiscuous mode
Jul 08 15:38:28 {HOSTNAME} kernel: vmbr0: port 5(fwpr122p0) entered blocking state
Jul 08 15:38:28 {HOSTNAME} kernel: vmbr0: port 5(fwpr122p0) entered forwarding state
Jul 08 15:38:28 {HOSTNAME} kernel: fwbr122i0: port 2(tap122i0) entered blocking state
Jul 08 15:38:28 {HOSTNAME} kernel: fwbr122i0: port 2(tap122i0) entered disabled state
Jul 08 15:38:28 {HOSTNAME} kernel: fwbr122i0: port 2(tap122i0) entered blocking state
Jul 08 15:38:28 {HOSTNAME} kernel: fwbr122i0: port 2(tap122i0) entered forwarding state
Jul 08 15:38:28 {HOSTNAME} pvedaemon[5599]: <USER@pve> end task UPID:HOSTNAME:000078A0:0001260A:5F064AE3:qmstart:122:USER@pve: OK
Jul 08 15:38:29 {HOSTNAME} pvedaemon[30954]: starting vnc proxy UPID:HOSTNAME:000078EA:000126F4:5F064AE5:vncproxy:122:USER@pve:
Jul 08 15:38:29 {HOSTNAME} pvedaemon[5599]: <USER@pve> starting task UPID:HOSTNAME:000078EA:000126F4:5F064AE5:vncproxy:122:USER@pve:
Jul 08 15:39:00 {HOSTNAME} systemd[1]: Starting Proxmox VE replication runner...
Jul 08 15:39:01 {HOSTNAME} systemd[1]: pvesr.service: Succeeded.
Jul 08 15:39:01 {HOSTNAME} systemd[1]: Started Proxmox VE replication runner.
Jul 08 15:39:38 {HOSTNAME} pvedaemon[5600]: <root@pam> successful auth for user 'USER@pve'
Jul 08 15:40:00 {HOSTNAME} systemd[1]: Starting Proxmox VE replication runner...
Jul 08 15:40:01 {HOSTNAME} systemd[1]: pvesr.service: Succeeded.
Jul 08 15:40:01 {HOSTNAME} systemd[1]: Started Proxmox VE replication runner.
There is no evidence of a panic, just a reboot. I would make sense if the guest had issues but why would it reboot the host?
I am continuing to investigate what the issue might be but any direction would be greatly appreciated. If there are any additional configurations needed, please let me know and I will append them.
Thank you,
DHM