Delay in creating disk including error with gluster

cbr

New Member
Oct 17, 2023
4
0
1
Hi all,

Currently we are setting up a cluster consisting out of 6 machines, all proxmox 8.0.4 where 3 of them are glusterfs servers and one of then an arbiter.
We added a glusterfs storage in the gui and assigned them to nodes, all is fine so far.

When we create a vm (or disk) no matter what type or what options we pass it with, the creation of a disk is quite slow due to the errors below i think.
Per error 1 glusterfs server i assume?

[2023-10-16 23:12:51.553665 +0000] I [io-stats.c:3701:ios_sample_buf_size_configure] 0-clusterdata: Configure ios_sample_buf size is 1024 because ios_sample_interval is 0 [2023-10-16 23:12:51.661126 +0000] E [MSGID: 108006] [afr-common.c:6123:__afr_handle_child_down_event] 0-clusterdata-replicate-0: All subvolumes are down. Going offline until at least one of them comes back up. [2023-10-16 23:12:51.661215 +0000] E [MSGID: 108006] [afr-common.c:6123:__afr_handle_child_down_event] 0-clusterdata-replicate-1: All subvolumes are down. Going offline until at least one of them comes back up. [2023-10-16 23:13:01.556623 +0000] I [io-stats.c:4033:fini] 0-clusterdata: io-stats translator unloaded [2023-10-16 23:13:02.560883 +0000] I [io-stats.c:3701:ios_sample_buf_size_configure] 0-clusterdata: Configure ios_sample_buf size is 1024 because ios_sample_interval is 0 [2023-10-16 23:13:02.674999 +0000] E [MSGID: 108006] [afr-common.c:6123:__afr_handle_child_down_event] 0-clusterdata-replicate-0: All subvolumes are down. Going offline until at least one of them comes back up. [2023-10-16 23:13:02.675086 +0000] E [MSGID: 108006] [afr-common.c:6123:__afr_handle_child_down_event] 0-clusterdata-replicate-1: All subvolumes are down. Going offline until at least one of them comes back up. [2023-10-16 23:13:12.563757 +0000] I [io-stats.c:4033:fini] 0-clusterdata: io-stats translator unloaded

The vm / disk is create fine and it is working, no perfomance issues or anything.
I did notice that this error is quite old but still not fixed as of today?

Just as a test we created a directory on all nodes, mounted them manually with mount -t glusterfs etc.... and then added a storage as a shared directory, used that to create a vm / disk, and no delay... no errors... and it even seems a tiny bit faster when testing.
What is going on here?

What is the difference when attaching the glusterfs storage directly trough proxmox vs manually and then adding a shared directory?
 

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!