Skip site navigation (1)Skip section navigation (2)
Date:      Thu, 17 Jul 2008 10:16:55 -0400
From:      Sven Willenberger <sven@dmv.com>
To:        Pete French <petefrench@ticketswitch.com>
Cc:        koitsu@FreeBSD.org, freebsd-stable@FreeBSD.org
Subject:   Re: Multi-machine mirroring choices
Message-ID:  <1216304215.14562.19.camel@lanshark.dmv.com>
In-Reply-To: <E1KImKO-000Gqk-ST@dilbert.ticketswitch.com>
References:  <E1KImKO-000Gqk-ST@dilbert.ticketswitch.com>

next in thread | previous in thread | raw e-mail | index | archive | help

--=-OUlfW357Kn9I5gRG34WM
Content-Type: text/plain
Content-Transfer-Encoding: quoted-printable

On Tue, 2008-07-15 at 16:20 +0100, Pete French wrote:
> > However, I must ask you this: why are you doing things the way you are?
> > Why are you using the equivalent of RAID 1 but for entire computers?  I=
s
> > there some reason you aren't using a filer (e.g. NetApp) for your data,
> > thus keeping it centralised?
>=20
> I am not the roiginal poster, but I am doing something very similar and
> can answer that question for you. Some people get paranoid about the
> whole "single point of failure" thing. I originally suggestted that we bu=
y
> a filer and have identical servers so if one breaks we connect the other
> to the filer, but the response I got was "what if the filer breaks?". So
> in the end I had to show we have duplicate independent machines, with the
> data kept symetrical on them at all times.
>=20
> It does actually work quite nicely actually - I have an "'active" databas=
e
> machine, and a "passive". The opassive is only used if the active fails,
> and the drives are run as a gmirror pair with the remote one being mounte=
d
> using ggated. It also means I can flip from active to passive when I want
> to do an OS upgrade on the active machine. Switching takes a few seconds,
> and this is fine for our setup.
>=20
> So the answer is that the descisiuon was taken out of my hands - but this
> is not uncommon, and as a roll-your-own cluster it works very nicely.
>=20
> -pete.
> _______________________________________________

I have for now gone with using ggate[cd] along with zpool and so far
it's not bad. I can fail the master, stop ggated on the slave at which
point geom reads the glabeled disks. From there I can zpool import to an
alternate root. When the master comes back up I can zpool export and
then, on the master, zpool import at which point zfs handles the
resilvering.

The *big* issue I have right now is dealing with the slave machine going
down. Once the master no longer has a connection to the ggated devices,
all processes trying to use the device hang in D status. I have tried
pkill'ing ggatec to no avail and ggatec destroy returns a message of
gctl being busy. Trying to ggatec destroy -f panics the machine.

Does anyone know how to successfully time out a failed ggatec connection
so that I can zpool detach or somehow have zfs removed the unavailable
drive?

Sven

--=-OUlfW357Kn9I5gRG34WM
Content-Type: application/pgp-signature; name=signature.asc
Content-Description: This is a digitally signed message part

-----BEGIN PGP SIGNATURE-----
Version: GnuPG v1.4.7 (GNU/Linux)

iD8DBQBIf1RXSnmnd8q3JGsRAlKBAKCE22yRLMrtzKkG4YyoY3rA2ZuU0wCdGegC
LV/XmxCaWiNzl08keY/U1AY=
=Nh7v
-----END PGP SIGNATURE-----

--=-OUlfW357Kn9I5gRG34WM--




Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?1216304215.14562.19.camel>