[Linux-HA] active active failover NFS server?
dejanmm at fastmail.fm
Thu Jul 19 10:07:13 MDT 2007
On Thu, Jul 19, 2007 at 05:12:56PM +0200, Sebastian Reitenbach wrote:
> Hi list,
> I am going to build an active active NFS server, where one exports a public
> directory, and the other the home directories. In case one fails, both
> should be exported by the remaining one server.
> I have a shared storage on a SAN, connected to both servers, I use the
> Filesystem ocf script to mount/umount the partitions (ext3, ocfs2 doesn't
> have ACL's, and I do not get GFS2 to work). Therefore I cannot run a nfs
> server clone, because I cannot umount the partition when the nfs server
> still lives on it, and the shared IP is wandering. I only see the LSB Script
> available for managing the nfsserver, but with the LSB script, only one NFS
> server can be started or stopped.
> So I have to configure two NFS resources using the LSB script, so that both
> can life on different servers. But now when I manually tell on nfs resource
> to move to another server, then both nfs resources will not be available for
> a short time. I also saw some problems when a dead node comes back into the
> cluster, also both nfs server resouces were not available for a short time.
> An other option would be to create a OCF script (I haven't found one) to
> manage the nfsserver. In the manual page of rpc.mountd I have seen that it
> is possible to specify a exports file and the port automatically. But I
> don't know what kind of other problems I might get, or whether it will be
> possible to run two nfsservers in parallel.
> anybody has an idea?
You can start here: http://linux-ha.org/HaNFS
> kind regards
> Linux-HA mailing list
> Linux-HA at lists.linux-ha.org
> See also: http://linux-ha.org/ReportingProblems
More information about the Linux-HA