I am running a GKE cluster which has multiple pods trying to access a shared volume. Since GC persistent disks do not allow ReadWriteMany access, I set up an NFS server in the cluster (in the same way done by many samples like this) to allow it. I am running both a production and development environment on this cluster in different namespaces, but since both of these environments run the same application, they both need their own file system.
Currently, the solution to this has been to set up 2 NFS servers in the same way (one for prod and one for dev). It seems like when the pods that mount the volume using the NFS server are on the same node as the NFS server itself, they are unable to mount (the error is "Unable to attach or mount volumes [...]: timed out waiting for the condition"). However, this seems to only be occuring for the dev environment as the prod environment has not had any problems. Currently, both NFS servers have been allocated to the same node, which may also be contributing to the problem, but I'm not sure.
I've been trying to figure out whether there is a problem with having 2 NFS servers in this way, or if there is some problem with trying to connect pods to an NFS server running on the same node, but to no avail so far. Any ideas what could cause the problem?
Logs in nfs server pods (same for both dev and prod):
nfs-dev-server Oct 30, 2020, 3:57:23 PM NFS started
nfs-dev-server Oct 30, 2020, 3:57:22 PM exportfs: / does not support NFS export
nfs-dev-server Oct 30, 2020, 3:57:22 PM Starting rpcbind
nfs-dev-server Oct 30, 2020, 3:57:22 PM rpcinfo: can't contact rpcbind: : RPC: Unable to receive; errno = Connection refused
nfs-dev-server Oct 30, 2020, 3:57:21 PM Serving /
nfs-dev-server Oct 30, 2020, 3:57:21 PM Serving /exports