Proxmox Datacenter Manager - First Alpha Release

In some nodes, it allows adding them, and I can see real-time metrics on the VMs, but I do not get the usage history, neither of the node nor of the VMs. I find this error in the logs:

Jan 16 20:56:01 datacentermanager proxmox-datacenter-api[523]: failed to collect metrics for XXXXXX: api error (status = 501 Not Implemented): Method 'GET /cluster/metrics/export' not implemented

Are your PVE-nodes up-to-date?:
 
  • Like
Reactions: kennysamuerto
In some nodes, it allows adding them, and I can see real-time metrics on the VMs, but I do not get the usage history, neither of the node nor of the VMs. I find this error in the logs:

Jan 16 20:56:01 datacentermanager proxmox-datacenter-api[523]: failed to collect metrics for XXXXXX: api error (status = 501 Not Implemented): Method 'GET /cluster/metrics/export' not implemented
make sure the nodes are on v8.3+
 
We installed this but the web server appears unresponsive.

Processes are up, server rebooted but nothing. Port 8443 does show in use.

```
proxmox-d 540 www-data 14u IPv6 2946 0t0 TCP *:8443 (LISTEN)
proxmox-d 540 www-data 15u IPv6 2946 0t0 TCP *:8443 (LISTEN)
proxmox-d 540 552 tokio-run www-data 14u IPv6 2946 0t0 TCP *:8443 (LISTEN)
proxmox-d 540 552 tokio-run www-data 15u IPv6 2946 0t0 TCP *:8443 (LISTEN)
proxmox-d 540 553 tokio-run www-data 14u IPv6 2946 0t0 TCP *:8443 (LISTEN)
proxmox-d 540 553 tokio-run www-data 15u IPv6 2946 0t0 TCP *:8443 (LISTEN)
proxmox-d 540 554 tokio-run www-data 14u IPv6 2946 0t0 TCP *:8443 (LISTEN)
proxmox-d 540 554 tokio-run www-data 15u IPv6 2946 0t0 TCP *:8443 (LISTEN)
proxmox-d 540 555 tokio-run www-data 14u IPv6 2946 0t0 TCP *:8443 (LISTEN)
proxmox-d 540 555 tokio-run www-data 15u IPv6 2946 0t0 TCP *:8443 (LISTEN)
proxmox-d 540 559 proxmox-d www-data 14u IPv6 2946 0t0 TCP *:8443 (LISTEN)
proxmox-d 540 559 proxmox-d www-data 15u IPv6 2946 0t0 TCP *:8443 (LISTEN)
```

Has anyone else seen this?
 
looks good, would nice if the tiles along the top would let you click on them to navigate to the relevant area i.e. you'd click on Virtual machines which would then take you to the list of machines
 
I like the role of this, but it still seems like most functions exist on the hosts, not at the manager, and I'm wondering if the vision is to move things to PDCM as things get more robust. Cluster-wide load balancing is still something I relegate to Prox-LB, and I'd love for that to be native, as well as cluster-wide startup order for VM's (so that I don't have to go in after maintenance or an outage and reboot vm's in order).
 
Very good starting !

I try to add a second user, it works, but the new user has no permissions ... How to set permissions ?

I've see in the roadmap, that firewall will be managed later with SDN. I think a tool to compare and synchronize ipsets, alias and group can be a good thing. I've create my pesonnal tool for that, via API, it's very helpful when working with multiple separated proxmox.

Why unlock container can't be done via interfaces or API ?
 
Hi again with my Question, has maybe anyone an idea on that? (its using root with all permissions)
i added my cluster and some single pve nodes, i do not have the Network Traffic and Disk I/O, Overview, any ideas?

Thanks, BR
 
Hi again with my Question, has maybe anyone an idea on that? (its using root with all permissions)
i added my cluster and some single pve nodes, i do not have the Network Traffic and Disk I/O, Overview, any ideas?

Thanks, BR
Do you look at the Node or at the VMs?
 
Hi again with my Question, has maybe anyone an idea on that? (its using root with all permissions)
i added my cluster and some single pve nodes, i do not have the Network Traffic and Disk I/O, Overview, any ideas?

Thanks, BR
Are you saying you have added the nodes/clusters and they show up in PDM but you don't see any graphical representations for statistics? If I'm understanding your question correctly, make sure your nodes are on PVE v8.3+
 
getting an error migrating, yet the machine was properly moved and I was able to start it on the target system. locked status on the source:

2025-01-29 11:31:33 remote: started tunnel worker 'UPID:ms01-01:00002FCB:0002A55D:679A57E5:vzmtunnel:103:root@pam!pdm-admin:'
tunnel: -> sending command "version" to remote
tunnel: <- got reply
2025-01-29 11:31:33 local WS tunnel version: 2
2025-01-29 11:31:33 remote WS tunnel version: 2
2025-01-29 11:31:33 minimum required WS tunnel version: 2
2025-01-29 11:31:33 websocket tunnel started
2025-01-29 11:31:33 shutdown CT 103
2025-01-29 11:31:35 starting migration of CT 103 to node 'ms01-01' (192.168.30.10)
tunnel: -> sending command "bwlimit" to remote
tunnel: <- got reply
2025-01-29 11:31:35 found local volume 'M920Q-Pool:vm-103-disk-0' (in current VM config)
tunnel: -> sending command "disk-import" to remote
tunnel: <- got reply
tunnel: accepted new connection on '/run/pve/103.storage'
tunnel: requesting WS ticket via tunnel
tunnel: established new WS for forwarding '/run/pve/103.storage'
Exporting image: 1% complete...
Exporting image: 97% complete...
Exporting image: 98% complete...
Exporting image: 99% complete...
Exporting image: 100% complete...done.
tunnel: -> sending command "query-disk-import" to remote
tunnel: done handling forwarded connection from '/run/pve/103.storage'
tunnel: <- got reply
2025-01-29 11:47:35 disk-import: Importing image: 100% complete...done.
tunnel: -> sending command "query-disk-import" to remote
tunnel: <- got reply
2025-01-29 11:47:36 volume 'M920Q-Pool:vm-103-disk-0' is 'TWOTB:vm-103-disk-0' on the target
2025-01-29 11:47:36 mapped: net0 from vmbr0 to vmbr0
tunnel: -> sending command "config" to remote
tunnel: <- got reply
2025-01-29 11:47:36 start final cleanup
tunnel: -> sending command "unlock" to remote
tunnel: <- got reply
2025-01-29 11:47:36 start container on target node
tunnel: -> sending command "start" to remote
tunnel: <- got reply
2025-01-29 11:47:37 ERROR: error - tunnel command '{"cmd":"start"}' failed - failed to handle 'start' command - startup for container '103' failed
2025-01-29 11:47:37 ERROR: migration finished with problems (duration 00:16:04)
TASK ERROR: migration problems
 
getting an error migrating, yet the machine was properly moved and I was able to start it on the target system. locked status on the source:

2025-01-29 11:31:33 remote: started tunnel worker 'UPID:ms01-01:00002FCB:0002A55D:679A57E5:vzmtunnel:103:root@pam!pdm-admin:'
tunnel: -> sending command "version" to remote
tunnel: <- got reply
2025-01-29 11:31:33 local WS tunnel version: 2
2025-01-29 11:31:33 remote WS tunnel version: 2
2025-01-29 11:31:33 minimum required WS tunnel version: 2
2025-01-29 11:31:33 websocket tunnel started
2025-01-29 11:31:33 shutdown CT 103
2025-01-29 11:31:35 starting migration of CT 103 to node 'ms01-01' (192.168.30.10)
tunnel: -> sending command "bwlimit" to remote
tunnel: <- got reply
2025-01-29 11:31:35 found local volume 'M920Q-Pool:vm-103-disk-0' (in current VM config)
tunnel: -> sending command "disk-import" to remote
tunnel: <- got reply
tunnel: accepted new connection on '/run/pve/103.storage'
tunnel: requesting WS ticket via tunnel
tunnel: established new WS for forwarding '/run/pve/103.storage'
Exporting image: 1% complete...
Exporting image: 97% complete...
Exporting image: 98% complete...
Exporting image: 99% complete...
Exporting image: 100% complete...done.
tunnel: -> sending command "query-disk-import" to remote
tunnel: done handling forwarded connection from '/run/pve/103.storage'
tunnel: <- got reply
2025-01-29 11:47:35 disk-import: Importing image: 100% complete...done.
tunnel: -> sending command "query-disk-import" to remote
tunnel: <- got reply
2025-01-29 11:47:36 volume 'M920Q-Pool:vm-103-disk-0' is 'TWOTB:vm-103-disk-0' on the target
2025-01-29 11:47:36 mapped: net0 from vmbr0 to vmbr0
tunnel: -> sending command "config" to remote
tunnel: <- got reply
2025-01-29 11:47:36 start final cleanup
tunnel: -> sending command "unlock" to remote
tunnel: <- got reply
2025-01-29 11:47:36 start container on target node
tunnel: -> sending command "start" to remote
tunnel: <- got reply
2025-01-29 11:47:37 ERROR: error - tunnel command '{"cmd":"start"}' failed - failed to handle 'start' command - startup for container '103' failed
2025-01-29 11:47:37 ERROR: migration finished with problems (duration 00:16:04)
TASK ERROR: migration problems
could you check the system log on the target side around that timestamp?
 
So, I attempted to move an LXC container, got this error:
Code:
failed to handle 'config' command - vm 216 - unable to parse value of 'unprivileged' - type check ('boolean') failed - got ' '

Interestingly, the container wouldn't let me change its value in the GUI, but I edited the .conf file with nano on the CLI and gave the 'unprivileged' option a value, and then it moved fine.

I would like to submit one of the following solutions to this issue:

1. do a pre-check of the config file before the file move, giving errors before moving data. In my case, this was an 8GB container, but I'd hate myself if I found this out after moving something massive.

2. If there is an issue with the config file, prompt the user for the intended value when attempting to start the VM on the destination.

3. Offer to move the virtual disk only, and show the existing config in the Datacenter UI once done, enabling the end user to manually re-create the container/VM.
 
  • Like
Reactions: steelghost
could you check the system log on the target side around that timestamp?
so I ran it again just to be see if it wasn't a fluke and I have the same error, this is the error on the receiving host:

Code:
mtunnel started
received command 'version'
received command 'bwlimit'
received command 'disk-import'
ready

received command 'ticket'
received command 'query-disk-import'
disk-import: Importing image: 100% complete...done.

received command 'query-disk-import'
disk-import: successfully imported 'TWOTB:vm-103-disk-0'

received command 'config'
received command 'unlock'
received command 'start'
run_buffer: 571 Script exited with status 2
lxc_init: 845 Failed to run lxc.hook.pre-start for container "103"
__lxc_start: 2034 Failed to initialize container "103"
TASK ERROR: mtunnel exited unexpectedly
 
so I ran it again just to be see if it wasn't a fluke and I have the same error, this is the error on the receiving host:

Code:
mtunnel started
received command 'version'
received command 'bwlimit'
received command 'disk-import'
ready

received command 'ticket'
received command 'query-disk-import'
disk-import: Importing image: 100% complete...done.

received command 'query-disk-import'
disk-import: successfully imported 'TWOTB:vm-103-disk-0'

received command 'config'
received command 'unlock'
received command 'start'
run_buffer: 571 Script exited with status 2
lxc_init: 845 Failed to run lxc.hook.pre-start for container "103"
__lxc_start: 2034 Failed to initialize container "103"
TASK ERROR: mtunnel exited unexpectedly
So the first two attempts, the system was powered on starting the migration. I just ran it again with the container turned off beforehand and it moved it without issues. if that can help narrow it down.
 
Last edited:
I
Are your PVE-nodes up-to-date?:
have noticed this but then i changed my backend network from 1gb network to a 10g network and added a local recursive dns and i havent seen it sense
 
Why am I getting errors when adding remotes to the PDM:

api error (status = 400: error trying to connect: error:0A000086:SSL routines:tls_post_process_server_certificate:certificate verify failed:../ssl/statem/statem_clnt.c:1889:)

I shouldn't be getting connection errors because the user and pw is the same across all of the hosts. Could someone help me please?

Thank you in advance,
nexusguy59
 

Attachments

  • pdmpve2025.png
    pdmpve2025.png
    11.9 KB · Views: 10
Last edited:
Why am I getting errors when adding remotes to the PDM:

api error (status = 400: error trying to connect: error:0A000086:SSL routines:tls_post_process_server_certificate:certificate verify failed:../ssl/statem/statem_clnt.c:1889:)

I shouldn't be getting connection errors because the user and pw is the same across all of the hosts. Could someone help me please?

Thank you in advance,
nexusguy59
I figured this out. I had the wrong IPs assigned to the wrong certificates. When I swapped the IPs it worked fine. Thanks Guys.