Hi,
after migrating a KVM machine to another node the migration fails with an zfs error. I am not able to destroy the image anymore. But I was able to raname it.
zfs destroy -f rpool/test
cannot destroy 'rpool/test': dataset is busy
Same behavior after reboot. The is nothing that can have a...
Hi,
So far the crash appears only at the nightly backup job. It looks like this:
vzdump --all 1 --mailto nilius@ckc.de --compress lzo --storage FreeNAS --quiet 1 --mailnotification failure --mode snapshot
I cannot abandon that this is a kernel issue. I can't imagine any scenario where any...
It does a reboot. But I don't think that this has anything todo with the hardware. I had Proxmox 3.x running for nealy two years without problems. After upgrading to 4.0 I see this on ALL cluster nodes (I have three of them). It is more likely that the kernel detected an urgent problem, probably...
Dietmar,
I found a few minutes to look into the code. First I found a problem which causes that the old snaphot is not force removed.
PVE/VZDump/LXC.pm:131
PVE::LXC::snapshot_delete($vmid, 'vzdump', 0);
calls snapshot_delete with force 0 instead of 1. So that in
PVE/LXC.pm from line 1827...
Now there comes another problem in. After a container backup the snapshot information was correctly removed but the wrong disk information was written:
local:202/vm-202-disk-1.raw,size=25G
instead of
zfs:subvol-202-disk-1,size=25G
So the conclusion is that the config file handler is broken...
The RAM usage is max. at ~38GB of 64GB, Swap utilization is nealy zero. So I think that cannot be the problem. But I post the requestet values when it happend again.
Those kernel messages are another problem. I opened a new thread: http://forum.proxmox.com/threads/24386-pveproxy-become-blocked-state-and-cannot-be-killed
Sorry, here is the full log:
https://www.dropbox.com/s/1lnctnzpyo8mnyg/full-log.gz?dl=0
Symptoms:
- web UI not accessible
systemctl status pveproxy
● pveproxy.service - PVE API Proxy Server
Loaded: loaded (/lib/systemd/system/pveproxy.service; enabled)
Active: failed (Result: timeout) since Mon 2015-11-09 06:49:06 CET; 3h 46min ago
Process: 7836...
Hi Thomas,
I can say that under some curcumstances 10 seconds are far away from enough. When the other cluster nodes are under heavy load it took much longer.
Yes this happens at the node statup.
Here is the log:
Nov 5 01:30:03 ckc-b-p0003 lxc-devsetup[1879]: Creating /dev/.lxc
Nov 5...
Here is the log from the whole day where the backup run:
Nov 7 00:00:04 ckc-b-p0005 vzdump[32477]: <root@pam> starting task UPID:ckc-b-p0005:00007EF8:01BF8802:563D30F4:vzdump::root@pam:
Nov 7 01:36:57 ckc-b-p0005 vzdump[32477]: <root@pam> end task...
It's a 3 node cluster.
uorum information
------------------
Date: Sat Nov 7 11:35:09 2015
Quorum provider: corosync_votequorum
Nodes: 3
Node ID: 0x00000002
Ring ID: 6500
Quorate: Yes
Votequorum information
----------------------
Expected...
When trying to make a new backup:
INFO: starting new backup job: vzdump 240 --remove 0 --node ckc-b-p0005 --compress lzo --mode snapshot --storage FreeNAS
INFO: Starting Backup of VM 240 (lxc)
INFO: status = running
INFO: found old vzdump snapshot (force removal)
INFO: backup mode: snapshot...
I have temporarily activated pve-no-subscription repo over enterprise. I have some other nasty problems after upgrading to 4.0 so I hope to get fixes a little faster.
proxmox-ve: 4.0-19 (running kernel: 4.2.3-2-pve)
pve-manager: 4.0-57 (running version: 4.0-57/cc7c2b53)
pve-kernel-4.2.2-1-pve...
It looks like there is a bug in writing the LXC .conf while while backup. Now I have a container which config file is completely empty after backup. After a backup there are also many XXX.conf.tmp.XXX files in the LXC config folder which never gets deleted.
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.