I guest my problem fall into NIC problem with Proxmox. Same this case: https://forum.proxmox.com/threads/dual-sfp-card-only-showing-1-nic.36661/
Tomorrow i will change my to newer NIC and retest again.
This is network traffic but peak time is differences from incident time: i think this is not the root cause.
And how network can caused all nodes reboot ? except Dell Server?
Can i upgrade straight to 5.1.3x? do i have to migrate VM/CT to other and perform rolling upgrade?
Probably. I'm checking log now. We have exp when we test MTU 9000 with following command and server reboot too:
From hv102:
ping hv101 -c 10 -M do -s 8972
Hi,
I have few Blade Server HP C7000 and i use them to host Proxmox Cluster. My Cluster have 8 nodes (is this reliable quorum?) and sometime it rebooted all nodes hosted on Blade. Other server same blade have no impact.
Last 2 weeks i decided to add 1 Dell PowerEdge R620 and migrate some...
Hi Fabian,
I can confirm this is impacted on LXC node and parameter adjusted on node2 and node1 (LXC node).
I try to separate LXC and KVM to separate nodes: some host KVM VPS and some host LXC VPS.
KVM nodes: not affected by this bug.
LXC nodes: yes, only LXC VPS caused this problem...
Hi Fabian,
I already try it but no luck yesterday.
I red your instruction in this thread: https://forum.proxmox.com/threads/failed-to-allocate-directory-watch-too-many-open-files.28700/#post-145947
Hi Fireon,
Yes, this is newly setup system and host under 50 VPS on 3 nodes. I have narrow down VPS caused this issues and temporary suspend them to avoid this error. Our server has run without reboot more than 24 hours already.
But this is just a temporary resolution. It's not the root cause...
Hi,
I have 3 nodes, i configure all basic network configuration -->install let's encrypt and access with green color on browser URL.
Then on node1 i create cluster:
pvecm create test-cluster --ring0_addr 10.10.30.1 --bindnet0_addr 10.10.30.0
on node2 i join to cluster of node1:
pvecm add...
I guest problem related to this thread
Yes
External CEPH storage
We use both platform: KVM and LXC. I try separate both platform to separate compute, some host KVM and some host LXC. After monitoring we detected problem just happen to LXC node only. KVM node is still working normal.
This...
Hi,
This morning we detected problem on our cluster running Proxmox 5.1.35
We try to migrate all VM and CT to new node, then try start services with ha-manager and see following errors:
root@node02:~# ha-manager set ct:119 --state started
Unable to create new inotify object: Too many open...
This morning my server shutdown and i can not start our VM back, when check pveproxy service is said:
root@node01:~# service pveproxy status
● pveproxy.service - PVE API Proxy Server
Loaded: loaded (/lib/systemd/system/pveproxy.service; enabled; vendor preset: enabled)
Active: active...
Hi,
We have an VM running on VMWare ESXi. We just migrate to Proxmox 5.1 with disk type as IDE and it boot up normal.
But the problem is really slow due to IDE type so we remove this HDD -->edit it as Virtio and reboot.
This boot is already hang up at:
I tried edit boot command to...
Hi dcsapak,
It return as below:
root@hv103:~# /usr/bin/ssh -o 'HostKeyAlias=hv101' root@10.10.30.151 /bin/true
Warning: the RSA host key for 'hv101' differs from the key for the IP address '[10.10.30.151]:4848'
Offending key for IP in /root/.ssh/known_hosts:2
Matching host key in...
Hi,
We having migration problem recently, promox does not migrate VM to other server if 1 server die.
We tried manual migrate from hv103 to hv101 and got error:
()
Task viewer: VM 150 - Migrate
Output
Status
Stop
task started by HA resource agent
2017-12-18 19:13:12 # /usr/bin/ssh -o...
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.