Hi all,
Currently we are setting up a cluster consisting out of 6 machines, all proxmox 8.0.4 where 3 of them are glusterfs servers and one of then an arbiter.
We added a glusterfs storage in the gui and assigned them to nodes, all is fine so far.
When we create a vm (or disk) no matter what type or what options we pass it with, the creation of a disk is quite slow due to the errors below i think.
Per error 1 glusterfs server i assume?
The vm / disk is create fine and it is working, no perfomance issues or anything.
I did notice that this error is quite old but still not fixed as of today?
Just as a test we created a directory on all nodes, mounted them manually with mount -t glusterfs etc.... and then added a storage as a shared directory, used that to create a vm / disk, and no delay... no errors... and it even seems a tiny bit faster when testing.
What is going on here?
What is the difference when attaching the glusterfs storage directly trough proxmox vs manually and then adding a shared directory?
Currently we are setting up a cluster consisting out of 6 machines, all proxmox 8.0.4 where 3 of them are glusterfs servers and one of then an arbiter.
We added a glusterfs storage in the gui and assigned them to nodes, all is fine so far.
When we create a vm (or disk) no matter what type or what options we pass it with, the creation of a disk is quite slow due to the errors below i think.
Per error 1 glusterfs server i assume?
[2023-10-16 23:12:51.553665 +0000] I [io-stats.c:3701:ios_sample_buf_size_configure] 0-clusterdata: Configure ios_sample_buf size is 1024 because ios_sample_interval is 0
[2023-10-16 23:12:51.661126 +0000] E [MSGID: 108006] [afr-common.c:6123:__afr_handle_child_down_event] 0-clusterdata-replicate-0: All subvolumes are down. Going offline until at least one of them comes back up.
[2023-10-16 23:12:51.661215 +0000] E [MSGID: 108006] [afr-common.c:6123:__afr_handle_child_down_event] 0-clusterdata-replicate-1: All subvolumes are down. Going offline until at least one of them comes back up.
[2023-10-16 23:13:01.556623 +0000] I [io-stats.c:4033:fini] 0-clusterdata: io-stats translator unloaded
[2023-10-16 23:13:02.560883 +0000] I [io-stats.c:3701:ios_sample_buf_size_configure] 0-clusterdata: Configure ios_sample_buf size is 1024 because ios_sample_interval is 0
[2023-10-16 23:13:02.674999 +0000] E [MSGID: 108006] [afr-common.c:6123:__afr_handle_child_down_event] 0-clusterdata-replicate-0: All subvolumes are down. Going offline until at least one of them comes back up.
[2023-10-16 23:13:02.675086 +0000] E [MSGID: 108006] [afr-common.c:6123:__afr_handle_child_down_event] 0-clusterdata-replicate-1: All subvolumes are down. Going offline until at least one of them comes back up.
[2023-10-16 23:13:12.563757 +0000] I [io-stats.c:4033:fini] 0-clusterdata: io-stats translator unloaded
The vm / disk is create fine and it is working, no perfomance issues or anything.
I did notice that this error is quite old but still not fixed as of today?
Just as a test we created a directory on all nodes, mounted them manually with mount -t glusterfs etc.... and then added a storage as a shared directory, used that to create a vm / disk, and no delay... no errors... and it even seems a tiny bit faster when testing.
What is going on here?
What is the difference when attaching the glusterfs storage directly trough proxmox vs manually and then adding a shared directory?