We are utilizing ZFS on Linux and we create the pool and filesystem as follows:
zpool create -o ashift=12 -m /dbs dbs_pool /dev/nvme0n1
zfs create -o recordsize=4k -o atime=off -o mountpoint=/dbs/data -o redundant_metadata=most -o compression=lz4 -o logbias=throughput -o primarycache=metadata -o xattr=sa dbs/data
zfs create -o recordsize=128k -o atime=off -o mountpoint=/dbs/ogs -o redundant_metadata=most -o compression=off -o logbias=latency -o xattr=sa dbs/logs
We then start a Docker container to access the above data and logs. After a few minutes one or more of
the above filesystems are automatically unmounted.
if we do: zfs mount -a then get back the filesystems but by that time the processes accessing the
filesystems have already died.
Some posts indicate that we have to add a dependency to mount zfs before we start Docker. So this is what we did:
sudo systemctl edit docker.service and add the following lines to it:
But it did not make any difference. It seems that something in the Linux system that is causing the unmount.
Can anyone please help with this.
Kernel version: 5.10.186-179.751
OpenZFS version: 2.1.12
Docker version 20.10.23, build 7155243
Is there some way for us to check if the problem is caused by Docker.
We are not utilizing ZFS storage in Docker. We are simply providing bind mounts to Docker and it so happens that the bind mounts have ZFS filesystem on them. Docker only sees these as volumes but otherwise is not instructed in anyway about ZFS.