From owner-freebsd-net@FreeBSD.ORG Fri Mar 1 16:18:41 2013 Return-Path: Delivered-To: freebsd-net@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:1900:2254:206a::19:1]) by hub.freebsd.org (Postfix) with ESMTP id EEF43FA0 for ; Fri, 1 Mar 2013 16:18:41 +0000 (UTC) (envelope-from ncrogers@gmail.com) Received: from mail-ve0-f182.google.com (mail-ve0-f182.google.com [209.85.128.182]) by mx1.freebsd.org (Postfix) with ESMTP id AA53F7E5 for ; Fri, 1 Mar 2013 16:18:41 +0000 (UTC) Received: by mail-ve0-f182.google.com with SMTP id ox1so3062005veb.13 for ; Fri, 01 Mar 2013 08:18:41 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20120113; h=mime-version:x-received:in-reply-to:references:date:message-id :subject:from:to:cc:content-type; bh=7LmYd61o1y2dsHNLzuR0TXgbRFWXlS7FVGrTgYFeLmw=; b=o8b6GZE789nWXaOvNgA8k3lMiapSdDLMPSikFPrE3T5ZsY7FNkOvJ9ypQysA3zCE6J qVGoIkwFVXD7SeihuY3yGoOmgF8zxMcxGbbHzayZtkRgAM6JDGfwuCImvl4xfTK68Xbv 0BThmi/ngA6NwNp8xhQhnoXUzq1t9ifEtWV7O29eoGKY0oUTJI63ZBJMVuwkg/L3T/lZ rDlgXxdzutf4Xyv8IQ+N7ji4XO1oas2b3eSUbR8A0SUJekreGiC4bUno4lMyFkJvxLvg tpgrJt5UvwqrnCtHBxL60Dv3Pye+ptNpXUkWYrNziRUrKRi8X7RAH8CL7BBPfrWD9STz pvww== MIME-Version: 1.0 X-Received: by 10.220.151.144 with SMTP id c16mr4216077vcw.18.1362154720931; Fri, 01 Mar 2013 08:18:40 -0800 (PST) Received: by 10.52.176.131 with HTTP; Fri, 1 Mar 2013 08:18:40 -0800 (PST) In-Reply-To: References: <512BAA60.3060703@biostat.wisc.edu> <512BAF8D.7080308@biostat.wisc.edu> Date: Fri, 1 Mar 2013 08:18:40 -0800 Message-ID: Subject: Re: igb network lockups From: Nick Rogers To: Jack Vogel Content-Type: text/plain; charset=ISO-8859-1 Cc: freebsd-net@freebsd.org, "Christopher D. Harrison" X-BeenThere: freebsd-net@freebsd.org X-Mailman-Version: 2.1.14 Precedence: list List-Id: Networking and TCP/IP with FreeBSD List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Fri, 01 Mar 2013 16:18:42 -0000 On Fri, Mar 1, 2013 at 8:04 AM, Nick Rogers wrote: > FWIW I have been experiencing a similar issue on a number of systems > using the em(4) driver under 9.1-RELEASE. This is after upgrading from > a snapshot of 8.3-STABLE. My systems use PF+ALTQ as well. The symptoms > are: interface stops passing traffic until the system is rebooted. I > have not yet been able to gain access to the systems to dig around > (after they have crashed), however my kernel/network settings are > properly tuned (high mbuf limit, hw.em.rxd/txd=4096, etc). It seems to > happen about once a day on systems with around a sustained 50Mb/s of > traffic. > > I realize this is not much to go on but perhaps it helps. I am > debating trying the e1000 driver in the latest CURRENT on top of > 9.1-RELEASE. I noticed the Intel shared code was updated about a week > ago. Would this change or perhaps another change to e1000 since > 9.1-RELEASE possibly affect stability in a positive way? > > Thanks. Heres relevant pciconf output: em0@pci0:1:0:0: class=0x020000 card=0x10d315d9 chip=0x10d38086 rev=0x00 hdr=0x00 vendor = 'Intel Corporation' device = '82574L Gigabit Network Connection' class = network subclass = ethernet cap 01[c8] = powerspec 2 supports D0 D3 current D0 cap 05[d0] = MSI supports 1 message, 64 bit cap 10[e0] = PCI-Express 1 endpoint max data 128(256) link x1(x1) cap 11[a0] = MSI-X supports 5 messages in map 0x1c enabled ecap 0001[100] = AER 1 0 fatal 0 non-fatal 1 corrected em1@pci0:2:0:0: class=0x020000 card=0x10d315d9 chip=0x10d38086 rev=0x00 hdr=0x00 vendor = 'Intel Corporation' device = '82574L Gigabit Network Connection' class = network subclass = ethernet cap 01[c8] = powerspec 2 supports D0 D3 current D0 cap 05[d0] = MSI supports 1 message, 64 bit cap 10[e0] = PCI-Express 1 endpoint max data 128(256) link x1(x1) cap 11[a0] = MSI-X supports 5 messages in map 0x1c enabled ecap 0001[100] = AER 1 0 fatal 0 non-fatal 1 corrected em2@pci0:7:0:0: class=0x020000 card=0x10d315d9 chip=0x10d38086 rev=0x00 hdr=0x00 vendor = 'Intel Corporation' device = '82574L Gigabit Network Connection' class = network subclass = ethernet cap 01[c8] = powerspec 2 supports D0 D3 current D0 cap 05[d0] = MSI supports 1 message, 64 bit cap 10[e0] = PCI-Express 1 endpoint max data 128(256) link x1(x1) cap 11[a0] = MSI-X supports 5 messages in map 0x1c enabled ecap 0001[100] = AER 1 0 fatal 0 non-fatal 1 corrected em3@pci0:8:0:0: class=0x020000 card=0x10d315d9 chip=0x10d38086 rev=0x00 hdr=0x00 vendor = 'Intel Corporation' device = '82574L Gigabit Network Connection' class = network subclass = ethernet cap 01[c8] = powerspec 2 supports D0 D3 current D0 cap 05[d0] = MSI supports 1 message, 64 bit cap 10[e0] = PCI-Express 1 endpoint max data 128(256) link x1(x1) cap 11[a0] = MSI-X supports 5 messages in map 0x1c enabled ecap 0001[100] = AER 1 0 fatal 0 non-fatal 1 corrected > > On Mon, Feb 25, 2013 at 10:45 AM, Jack Vogel wrote: >> Have you done any poking around, looking at stats to determine why the >> hangs? For instance, >> might your mbuf pool be depleted? Some other network resource perhaps? >> >> Jack >> >> >> On Mon, Feb 25, 2013 at 10:38 AM, Christopher D. Harrison < >> harrison@biostat.wisc.edu> wrote: >> >>> Sure, >>> The problem appears on both systems running with ALTQ and vanilla. >>> -C >>> >>> On 02/25/13 12:29, Jack Vogel wrote: >>> >>> I've not heard of this problem, but I think most users do not use ALTQ, >>> and we (Intel) do not >>> test using it. Can it be eliminated from the equation? >>> >>> Jack >>> >>> >>> On Mon, Feb 25, 2013 at 10:16 AM, Christopher D. Harrison < >>> harrison@biostat.wisc.edu> wrote: >>> >>>> I recently have been experiencing network "freezes" and network "lockups" >>>> on our Freebsd 9.1 systems which are running zfs and nfs file servers. >>>> I upgraded from 9.0 to 9.1 about 2 months ago and we have been having >>>> issues with almost bi-monthly. The issue manifests in the system becomes >>>> unresponsive to any/all nfs clients. The system is not resource bound as >>>> our I/O is low to disk and our network is usually in the 20mbit/40mbit >>>> range. We do notice a correlation between temporary i/o spikes and >>>> network freezes but not enough to send our system in to "lockup" mode for >>>> the next 5min. Currently we have 4 igb nics in 2 aggr's with 8 queue's >>>> per nic and our dev.igb reports: >>>> >>>> dev.igb.3.%desc: Intel(R) PRO/1000 Network Connection version - 2.3.4 >>>> >>>> I am almost certain the problem is with the ibg driver as a friend is >>>> also experiencing the same problem with the same intel igb nic. He has >>>> addressed the issue by restarting the network using netif on his systems. >>>> According to my friend, once the network interfaces get cleared, everything >>>> comes back and starts working as expected. >>>> >>>> I have noticed an issue with the igb driver and I was looking for >>>> thoughts on how to help address this problem. >>>> >>>> http://freebsd.1045724.n5.nabble.com/em-igb-if-transmit-drbr-and-ALTQ-td5760338.html >>>> >>>> Thoughts/Ideas are greatly appreciated!!! >>>> >>>> -C >>>> >>>> _______________________________________________ >>>> freebsd-net@freebsd.org mailing list >>>> http://lists.freebsd.org/mailman/listinfo/freebsd-net >>>> To unsubscribe, send any mail to "freebsd-net-unsubscribe@freebsd.org" >>>> >>> >>> >>> >> _______________________________________________ >> freebsd-net@freebsd.org mailing list >> http://lists.freebsd.org/mailman/listinfo/freebsd-net >> To unsubscribe, send any mail to "freebsd-net-unsubscribe@freebsd.org"