Hello,
I am trying to migrate few VMs (offline) from one pve 8.2.2 host (HOST1) to another host running pve 8.2.2 too (HOST2).
I am using this syntax:
The target storage is lvm-thin.
I have successfuly migrated two VMs (one 127G size, another 64G size), but the third one (80G size) is always transfering up to 80G but it then it breaks:
(...)
Journalctl on HOST1 shows nothing more:
BUT there is an interesting journal on HOST2:
(...)
(...)
Whole log here: https://pastebin.com/289N8m0M
This is repeatable, tried this several times and always look the same.
HOST2 has 32G RAM.
During the migration the pveproxy process takes all the time 80% of memory on HOST2, but just when the transfer is finishing, the process quickly eats whole RAM in a few seconds and crashes.
There are two harddisks on the source vm:
Is seems that only 1st of them is being migrated and I can see it on the destination host when the migration crashes. But the second one is missing.
Is this a bug? Or can I do something more?
I am trying to migrate few VMs (offline) from one pve 8.2.2 host (HOST1) to another host running pve 8.2.2 too (HOST2).
I am using this syntax:
qm remote-migrate 103 103 apitoken='Authorization: PVEAPIToken=.....,host=192.168.115.10,fingerprint=.....' --target-bridge vmbr0 --target-storage thinpool-data
The target storage is lvm-thin.
I have successfuly migrated two VMs (one 127G size, another 64G size), but the third one (80G size) is always transfering up to 80G but it then it breaks:
Establishing API connection with remote at '192.168.115.10'
(...)
85899345920 bytes (86 GB, 80 GiB) copied, 918.303 s, 93.5 MB/s
tunnel: -> sending command "query-disk-import" to remote
tunnel: done handling forwarded connection from '/run/pve/103.storage'
2024-06-13 21:25:56 ERROR: no reply to command '{"cmd":"query-disk-import"}': reading from tunnel failed: got timeout
2024-06-13 21:25:56 aborting phase 1 - cleanup resources
2024-06-13 21:26:56 tunnel still running - terminating now with SIGTERM
CMD websocket tunnel died: command 'proxmox-websocket-tunnel' failed: interrupted by signal
2024-06-13 21:26:57 ERROR: no reply to command '{"cmd":"quit","cleanup":1}': reading from tunnel failed: got timeout
2024-06-13 21:26:57 ERROR: migration aborted (duration 00:16:32): no reply to command '{"cmd":"query-disk-import"}': reading from tunnel failed: got timeout
migration aborted
Journalctl on HOST1 shows nothing more:
Jun 13 21:26:57 pve qm[963664]: migration aborted
Jun 13 21:26:57 pve qm[963663]: <root@pam> end task UPID:pve:000EB450:022D9BFA:666B4421:qmigrate:103:root@pam: migration aborted
BUT there is an interesting journal on HOST2:
Jun 13 21:27:56 pve kernel: pveproxy worker invoked oom-killer: gfp_mask=0x140dca(GFP_HIGHUSER_MOVABLE|__GFP_COMP|__GFP_ZERO), order=0, oom_score_adj=0
Jun 13 21:27:56 pve kernel: CPU: 1 PID: 69493 Comm: pveproxy worker Tainted: P O 6.8.4-3-pve #1
(...)
Jun 13 21:27:56 pve kernel: 19274 total pagecache pages
Jun 13 21:27:56 pve kernel: 12330 pages in swap cache
Jun 13 21:27:56 pve kernel: Free swap = 0kB
Jun 13 21:27:56 pve kernel: Total swap = 8388604kB
Jun 13 21:27:56 pve kernel: 8372427 pages RAM
Jun 13 21:27:56 pve kernel: 0 pages HighMem/MovableOnly
Jun 13 21:27:56 pve kernel: 175305 pages reserved
Jun 13 21:27:56 pve kernel: 0 pages hwpoisoned
Jun 13 21:27:56 pve kernel: Tasks state (memory values in pages):
(...)
Jun 13 21:27:56 pve kernel: oom-kill:constraint=CONSTRAINT_NONE,nodemask=(null),cpuset=pveproxy.service,mems_allowed=0,global_oom,task_memcg=/system.slice/pveproxy.service,task=pveproxy worker,pid=69493,uid=33
Jun 13 21:27:56 pve kernel: Out of memory: Killed process 69493 (pveproxy worker) total-vm:51662568kB, anon-rss:31977508kB, file-rss:2176kB, shmem-rss:512kB, UID:33 pgtables:77576kB oom_score_adj:0
Jun 13 21:27:56 pve systemd[1]: pveproxy.service: A process of this unit has been killed by the OOM killer.
Jun 13 21:27:58 pve kernel: oom_reaper: reaped process 69493 (pveproxy worker), now anon-rss:0kB, file-rss:232kB, shmem-rss:0kB
Jun 13 21:27:59 pve pveproxy[1036]: worker 69493 finished
Jun 13 21:27:59 pve pveproxy[1036]: starting 1 worker(s)
Jun 13 21:27:59 pve pveproxy[1036]: worker 437765 started
Jun 13 21:28:00 pve pvedaemon[435389]: mtunnel exited unexpectedly
Jun 13 21:28:02 pve pvedaemon[1030]: <root@pam!token2> end task UPID:pve:0006A4BD:01268FD5:666B4421:qmtunnel:103:root@pam!token2: mtunnel exited unexpectedly
Whole log here: https://pastebin.com/289N8m0M
This is repeatable, tried this several times and always look the same.
HOST2 has 32G RAM.
During the migration the pveproxy process takes all the time 80% of memory on HOST2, but just when the transfer is finishing, the process quickly eats whole RAM in a few seconds and crashes.
There are two harddisks on the source vm:
Is seems that only 1st of them is being migrated and I can see it on the destination host when the migration crashes. But the second one is missing.
Is this a bug? Or can I do something more?
Last edited: