Date: Tue, 20 Nov 2012 13:59:54 -0500 From: George Neville-Neil <gnn@neville-neil.com> To: Daichi GOTO <daichi@bsdconsulting.co.jp> Cc: freebsd-infiniband@FreeBSD.org Subject: Re: Infiniband experiment environment setting up Message-ID: <2A385CE3-72E8-4D40-8413-69E00841536D@neville-neil.com> In-Reply-To: <20121119115350.3456c580209197b4aea33d75@bsdconsulting.co.jp> References: <20121119115350.3456c580209197b4aea33d75@bsdconsulting.co.jp>
next in thread | previous in thread | raw e-mail | index | archive | help
On Nov 18, 2012, at 21:53 , Daichi GOTO <daichi@bsdconsulting.co.jp> wrote: > Hi, > > Currently, we are getting ready to test InfiniBands on FreeBSD > using follow devices: > > CPU: Intel Xeon E3-1240v2 > M/B: Super Micro X9SCM > RAM: Kingston KVR16E11/8 (8GBx4) > > IB Card: Mellanox Technologies ConnectX-2 VPI MHQH19B-XTR > IB Card: Mellanox Technologies ConnectX-3 VPI MCX353A-QCBT > IB Switch: Mellanox Technologies EXW-IS5022 > QSFP cable: MC22096-130-001 > > Is there some information about above InfiniBand devices? Well it's only the cards you have to worry about. I know the ConnectX-2 works but have not tested the ConnectX-3. Also, note that it depends on what you are testing. That is, Infiniband has low latency at low payloads. The numbers quoted by Mellanox about low latency depend on you sending one byte (yes, 1 byte). At real packet sizes Infiniband is no faster than 10GbE. Best, George
Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?2A385CE3-72E8-4D40-8413-69E00841536D>
