You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I'd like to know if there are any issues with GlusterFS and NVME.
This week I got two customer where I build 2 Proxmox VE with GlusterFS 11.
I had have created with:
On both nodes I do:
mkdir /data1
mkdir /data2
mkfs.xfs /dev/nvme1
mkfs.xfs /dev/nvme2
mount /dev/nvme1 /data1
mount /dev/nvme2 /data2
I had have installed gluster like this:
wget -qO - https://download.gluster.org/pub/gluster/glusterfs/11/rsa.pub | gpg --dearmor -o /etc/apt/trusted.gpg.d/gluster.gpg
To solve the split-brain issue, I applied this configurations:
gluster vol set VMS cluster.heal-timeout 5
gluster vol heal VMS enable
gluster vol set VMS cluster.quorum-reads false
gluster vol set VMS cluster.quorum-count 1
gluster vol set VMS network.ping-timeout 2
gluster vol set VMS cluster.favorite-child-policy mtime
gluster vol heal VMS granular-entry-heal enable
gluster vol set VMS cluster.data-self-heal-algorithm full
gluster vol set VMS features.shard on
gluster vol set VMS performance.write-behind off
gluster vol set VMS performance.flush-behind off
So this configuration allows me to power down the first server and the VMs restart on the secondary server, with no issues at all.
I have the very same scenario in another customer, but there we are working wih SSD DC600M Kingston.
Turns out that in the servers with NVME I got a lot of disk corruption inside the VM.
If I reboot, things go worse.
Does anybody know any cases about gluster and nvme issues like that?
Is there any fix for that?
Thanks
The text was updated successfully, but these errors were encountered:
Replica 2 volume is not good for consistency. It even warns while creating. I wouldn't use it for production workload.
Also these are the recommended options for VM workloads
Yes. I know replica 2 is not recommended.
But I have used this set up for couple of years and nothing really happens so worst than what's happen now.
And just with NVME. The VM was seating there for months and then, all of the sudden, crashes.
With SSD everything is ok.
Hi there.
I'd like to know if there are any issues with GlusterFS and NVME.
This week I got two customer where I build 2 Proxmox VE with GlusterFS 11.
I had have created with:
On both nodes I do:
mkdir /data1
mkdir /data2
mkfs.xfs /dev/nvme1
mkfs.xfs /dev/nvme2
mount /dev/nvme1 /data1
mount /dev/nvme2 /data2
I had have installed gluster like this:
wget -qO - https://download.gluster.org/pub/gluster/glusterfs/11/rsa.pub | gpg --dearmor -o /etc/apt/trusted.gpg.d/gluster.gpg
echo "deb https://download.gluster.org/pub/gluster/glusterfs/LATEST/Debian/bookworm/amd64/apt bookworm main" > /etc/apt/sources.list.d/gluster.list
After install glusterfs and do the peer probe, I do
gluster vol create VMS replica 2 gluster1:/data1/vms gluster2:/data1/vms gluster1:/data2/vms gluster2:/data/vms
To solve the split-brain issue, I applied this configurations:
gluster vol set VMS cluster.heal-timeout 5
gluster vol heal VMS enable
gluster vol set VMS cluster.quorum-reads false
gluster vol set VMS cluster.quorum-count 1
gluster vol set VMS network.ping-timeout 2
gluster vol set VMS cluster.favorite-child-policy mtime
gluster vol heal VMS granular-entry-heal enable
gluster vol set VMS cluster.data-self-heal-algorithm full
gluster vol set VMS features.shard on
gluster vol set VMS performance.write-behind off
gluster vol set VMS performance.flush-behind off
So this configuration allows me to power down the first server and the VMs restart on the secondary server, with no issues at all.
I have the very same scenario in another customer, but there we are working wih SSD DC600M Kingston.
Turns out that in the servers with NVME I got a lot of disk corruption inside the VM.
If I reboot, things go worse.
Does anybody know any cases about gluster and nvme issues like that?
Is there any fix for that?
Thanks
The text was updated successfully, but these errors were encountered: