Hi,
First post here and sadly with a problem.
I'm running a two node cluster with DRBD-storage + a third node for quorum and NFS storage for image backups. The backup is scheduled daily and have been working just fine for a some time now, until last night that is. In the middle of a backup the NFS server failed with a kernel panic due to hardware failure - system hard drive died.
Nothing strange here but what concerns me was the effect this had on the cluster node and the VM being backed up during the failure. The VM is running Windows Server 2008 and has got two disk images, one for OS, one for data. The data volume became unaccessible on the guest and the only way I could recover from this was to reboot the cluster node. The VM started just fine after that and all data was readable, no problem.
Is this expected behavior when the NFS share becomes inaccessible during a backup? It would have been nice if the backup would have timed out and just failed with an error but I also understand that NFS itself is tricky and sensitive when there is a network or host problem.
Would iSCSI be a better solution in this case? The chances of this happening again is probably slim but when possible I would like to avoid scenarios that forces a reboot of a physical node when the VM's can't be migrated.
Thanks.
Freddy
First post here and sadly with a problem.
I'm running a two node cluster with DRBD-storage + a third node for quorum and NFS storage for image backups. The backup is scheduled daily and have been working just fine for a some time now, until last night that is. In the middle of a backup the NFS server failed with a kernel panic due to hardware failure - system hard drive died.
Nothing strange here but what concerns me was the effect this had on the cluster node and the VM being backed up during the failure. The VM is running Windows Server 2008 and has got two disk images, one for OS, one for data. The data volume became unaccessible on the guest and the only way I could recover from this was to reboot the cluster node. The VM started just fine after that and all data was readable, no problem.
Is this expected behavior when the NFS share becomes inaccessible during a backup? It would have been nice if the backup would have timed out and just failed with an error but I also understand that NFS itself is tricky and sensitive when there is a network or host problem.
Would iSCSI be a better solution in this case? The chances of this happening again is probably slim but when possible I would like to avoid scenarios that forces a reboot of a physical node when the VM's can't be migrated.
Thanks.
Freddy