Error message 595, when migrating manually

Martino Rabe

Renowned Member
Jul 24, 2016
14
1
68
44
Hello,

I installed proxmox 8.2.7 on two systems and added both to a cluster. I configured replication which seems to be working fine and added all lxc to the cluster. The problem is that I am not able to use or manually migrate the lxc systems when one system is going down. I only see the error message "No route to host (595)" when I click on the lxc container or "Verbindungsfehler (Connection error) 595: No route to host" when I want to migrate the lxc container to the working node.

The output (when one node is down) for pvecm status is

Code:
root@redmountain:~# pvecm status
Cluster information
-------------------
Name:             proxmox-cluster
Config Version:   2
Transport:        knet
Secure auth:      on

Quorum information
------------------
Date:             Wed Oct 30 17:42:04 2024
Quorum provider:  corosync_votequorum
Nodes:            1
Node ID:          0x00000002
Ring ID:          2.2c9
Quorate:          No

Votequorum information
----------------------
Expected votes:   2
Highest expected: 2
Total votes:      1
Quorum:           2 Activity blocked
Flags:        

Membership information
----------------------
    Nodeid      Votes Name
0x00000002          1 192.168.5.54 (local)

I even configured HA (even I know that 3 nodes are neccessary to get an automatic failover working) but I still get the same error when I click on migrate. When both nodes are up, I am able to migrate the lxc container to the other node without any problem.

Maybe I completely missunderstand the concept or I did some configuration error I am not able to find out ?
Does anyone have a hint for me ? I didn't find any other entry that describes the 595 error that describes the same environment.

If you need any logs please let me know I am happy to provide you with the information.
 

Attachments

  • Bildschirmfoto vom 2024-10-30 17-05-19.png
    Bildschirmfoto vom 2024-10-30 17-05-19.png
    207.1 KB · Views: 3
Last edited:
Maybe as an additional information, this is the log when both nodes are up and I manually migrate one lxc container

Code:
2024-10-31 18:53:53 shutdown CT 101
2024-10-31 18:54:05 starting migration of CT 101 to node 'redmountain' (192.168.5.54)
2024-10-31 18:54:05 found local volume 'zfs-disk:subvol-101-disk-0' (in current VM config)
2024-10-31 18:54:05 start replication job
2024-10-31 18:54:05 guest => CT 101, running => 0
2024-10-31 18:54:05 volumes => zfs-disk:subvol-101-disk-0
2024-10-31 18:54:06 create snapshot '__replicate_101-0_1730397245__' on zfs-disk:subvol-101-disk-0
2024-10-31 18:54:06 using secure transmission, rate limit: none
2024-10-31 18:54:06 incremental sync 'zfs-disk:subvol-101-disk-0' (__replicate_101-0_1730394003__ => __replicate_101-0_1730397245__)
2024-10-31 18:54:07 send from @__replicate_101-0_1730394003__ to zfs-disk/subvol-101-disk-0@__replicate_101-0_1730397245__ estimated size is 12.7M
2024-10-31 18:54:07 total estimated size is 12.7M
2024-10-31 18:54:07 TIME        SENT   SNAPSHOT zfs-disk/subvol-101-disk-0@__replicate_101-0_1730397245__
2024-10-31 18:54:08 18:54:08   7.76M   zfs-disk/subvol-101-disk-0@__replicate_101-0_1730397245__
2024-10-31 18:54:09 successfully imported 'zfs-disk:subvol-101-disk-0'
2024-10-31 18:54:09 delete previous replication snapshot '__replicate_101-0_1730394003__' on zfs-disk:subvol-101-disk-0
2024-10-31 18:54:10 (remote_finalize_local_job) delete stale replication snapshot '__replicate_101-0_1730394003__' on zfs-disk:subvol-101-disk-0
2024-10-31 18:54:10 end replication job
2024-10-31 18:54:10 # /usr/bin/ssh -e none -o 'BatchMode=yes' -o 'HostKeyAlias=redmountain' -o 'UserKnownHostsFile=/etc/pve/nodes/redmountain/ssh_known_hosts' -o 'GlobalKnownHostsFile=none' root@192.168.5.54 pvesr set-state 101 \''{"local/robustoak":{"fail_count":0,"last_node":"robustoak","last_iteration":1730397245,"last_try":1730397245,"storeid_list":["zfs-disk"],"duration":4.63319,"last_sync":1730397245}}'\'
2024-10-31 18:54:11 start final cleanup
2024-10-31 18:54:12 start container on target node
2024-10-31 18:54:12 # /usr/bin/ssh -e none -o 'BatchMode=yes' -o 'HostKeyAlias=redmountain' -o 'UserKnownHostsFile=/etc/pve/nodes/redmountain/ssh_known_hosts' -o 'GlobalKnownHostsFile=none' root@192.168.5.54 pct start 101
2024-10-31 18:54:15 migration finished successfully (duration 00:00:23)
TASK OK

So migrating is simply a problem when one node is down.
 

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!