Skip site navigation (1)Skip section navigation (2)
Date:      Fri, 1 Jul 2016 16:41:59 +0200
From:      InterNetX - Juergen Gotteswinter <jg@internetx.com>
To:        Julien Cigar <julien@perdition.city>
Cc:        Joe Love <joe@getsomewhere.net>, freebsd-fs@freebsd.org
Subject:   Re: HAST + ZFS + NFS + CARP
Message-ID:  <01b8a61e-739e-c41e-45bc-a84af0a9d8ab@internetx.com>
In-Reply-To: <20160701143917.GB41276@mordor.lan>
References:  <20160630163541.GC5695@mordor.lan> <50BF1AEF-3ECC-4C30-B8E1-678E02735BB5@gmail.com> <20160701084717.GE5695@mordor.lan> <47c7e1a5-6ae8-689c-9c2d-bb92f659ea43@internetx.com> <20160701101524.GF5695@mordor.lan> <f74627e3-604e-da71-c024-7e4e71ff36cb@internetx.com> <20160701105735.GG5695@mordor.lan> <3d8c7c89-b24e-9810-f3c2-11ec1e15c948@internetx.com> <93E50E6B-8248-43B5-BE94-D94D53050E06@getsomewhere.net> <bbaf14e2-4ec6-545c-ba67-a1084100b05c@internetx.com> <20160701143917.GB41276@mordor.lan>

next in thread | previous in thread | raw e-mail | index | archive | help
Am 01.07.2016 um 16:39 schrieb Julien Cigar:
> On Fri, Jul 01, 2016 at 03:44:36PM +0200, InterNetX - Juergen Gotteswinter wrote:
>>
>>
>> Am 01.07.2016 um 15:18 schrieb Joe Love:
>>>
>>>> On Jul 1, 2016, at 6:09 AM, InterNetX - Juergen Gotteswinter <jg@internetx.com> wrote:
>>>>
>>>> Am 01.07.2016 um 12:57 schrieb Julien Cigar:
>>>>> On Fri, Jul 01, 2016 at 12:18:39PM +0200, InterNetX - Juergen Gotteswinter wrote:
>>>>>
>>>>> of course I'll test everything properly :) I don't have the hardware yet
>>>>> so ATM I'm just looking for all the possible "candidates", and I'm 
>>>>> aware that a redundant storage is not that easy to implement ...
>>>>>
>>>>> but what solutions do we have? It's either CARP + ZFS + (HAST|iSCSI), 
>>>>> either zfs send|ssh zfs receive as you suggest (but it's
>>>>> not realtime), either a distributed FS (which I avoid like the plague..)
>>>>
>>>> zfs send/receive can be nearly realtime.
>>>>
>>>> external jbods with cross cabled sas + commercial cluster solution like
>>>> rsf-1. anything else is a fragile construction which begs for desaster.
>>>
>>> This sounds similar to the CTL-HA code that went in last year, for which I haven’t seen any sort of how-to.  The RSF-1 stuff sounds like it has more scaling options, though.  Which it probably should, given its commercial operation.
>>
>> rsf is what pacemaker / heartbeat tries to be, judge me for linking
>> whitepapers but in this case its not such evil marketing blah
>>
>> http://www.high-availability.com/wp-content/uploads/2013/01/RSF-1-HA-PLUGIN-ZFS-STORAGE-CLUSTER.pdf
>>
>>
>> @ Julien
>>
>> seems like you take availability really serious, so i guess you also got
>> plans how to accomplish network problems like dead switches, flaky
>> cables and so on.
>>
>> like using multiple network cards in the boxes, cross cabling between
>> the hosts (rs232 and ethernet of course, using proved reliable network
>> switches in a stacked configuration for example cisco 3750 stacked). not
>> to forget redundant power feeds to redundant power supplies.
> 
> the only thing that is not redundant (yet?) is our switch, an HP Pro 
> Curve 2530-24G) .. it's the next step :)

Arubas, okay, a quick view in the spec sheet does not seem to list
stacking option.

what about power?

> 
>>
>> if not, i whould start again from scratch.
>>
>>>
>>> -Joe
>>>
>>> _______________________________________________
>>> freebsd-fs@freebsd.org mailing list
>>> https://lists.freebsd.org/mailman/listinfo/freebsd-fs
>>> To unsubscribe, send any mail to "freebsd-fs-unsubscribe@freebsd.org"
>>>
>> _______________________________________________
>> freebsd-fs@freebsd.org mailing list
>> https://lists.freebsd.org/mailman/listinfo/freebsd-fs
>> To unsubscribe, send any mail to "freebsd-fs-unsubscribe@freebsd.org"
> 




Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?01b8a61e-739e-c41e-45bc-a84af0a9d8ab>