Skip site navigation (1)Skip section navigation (2)
Date:      Tue, 6 Oct 2015 10:35:57 -0500
From:      Sean Kelly <smkelly@smkelly.org>
To:        Slawa Olhovchenkov <slw@zxy.spb.ru>
Cc:        FreeBSD-STABLE Mailing List <freebsd-stable@freebsd.org>, Jim Harris <jim.harris@gmail.com>
Subject:   Re: Dell NVMe issues
Message-ID:  <27228FE7-5FF9-4F58-9E23-42A66806C374@smkelly.org>
In-Reply-To: <20151006152955.GA16596@zxy.spb.ru>
References:  <BC5F191D-FEB2-4ADC-9D6B-240C80B2301C@smkelly.org> <20151006152955.GA16596@zxy.spb.ru>

next in thread | previous in thread | raw e-mail | index | archive | help


> On Oct 6, 2015, at 10:29 AM, Slawa Olhovchenkov <slw@zxy.spb.ru> =
wrote:
>=20
> On Tue, Oct 06, 2015 at 10:18:11AM -0500, Sean Kelly wrote:
>=20
>> Back in May, I posted about issues I was having with a Dell PE R630 =
with 4x800GB NVMe SSDs. I would get kernel panics due to the inability =
to assign all the interrupts because of =
https://bugs.freebsd.org/bugzilla/show_bug.cgi?id=3D199321 =
<https://bugs.freebsd.org/bugzilla/show_bug.cgi?id=3D199321>; =
<https://bugs.freebsd.org/bugzilla/show_bug.cgi?id=3D199321 =
<https://bugs.freebsd.org/bugzilla/show_bug.cgi?id=3D199321>>. Jim =
Harris helped fix this issue so I bought several more of these servers, =
Including ones with 4x1.6TB drives...
>>=20
>> while the new servers with 4x800GB drives still work, the ones with =
4x1.6TB drives do not. When I do a
>> 	zpool create tank mirror nvd0 nvd1 mirror nvd2 nvd3
>> the command never returns and the kernel logs:
>> 	nvme0: resetting controller
>> 	nvme0: controller ready did not become 0 within 2000 ms
>>=20
>> I've tried several different things trying to understand where the =
actual problem is.
>> WORKS: dd if=3D/dev/nvd0 of=3D/dev/null bs=3D1m
>> WORKS: dd if=3D/dev/zero of=3D/dev/nvd0 bs=3D1m
>> WORKS: newfs /dev/nvd0
>> FAILS: zpool create tank mirror nvd[01]
>> FAILS: gpart add -t freebsd-zfs nvd[01] && zpool create tank mirror =
nvd[01]p1
>> FAILS: gpart add -t freebsd-zfs -s 1400g nvd[01[ && zpool create tank =
nvd[01]p1
>> WORKS: gpart add -t freebsd-zfs -s 800g nvd[01] && zpool create tank =
nvd[01]p1
>>=20
>> NOTE: The above commands are more about getting the point across, not =
validity. I wiped the disk clean between gpart attempts and used GPT.
>=20
> Just for purity of the experiment: do you try zpool on raw disk, w/o
> GPT? I.e. zpool create tank mirror nvd0 nvd1
>=20

Yes, that was actually what I tried first. I headed down the path of GPT =
because it allowed me a way to restrict how much disk zpool touched. =
zpool on the bare NVMe disks also triggers the issue.

>> So it seems like zpool works if I don't cross past ~800GB. But other =
things like dd and newfs work.
>>=20
>> When I get the kernel messages about the controller resetting and =
then not responding, the NVMe subsystem hangs entirely. Since my boot =
disks are not NVMe, the system continues to work but no more NVMe stuff =
can be done. Further, attempting to reboot hangs and I have to do a =
power cycle.
>>=20
>> Any thoughts on what the deal may be here?
>>=20
>> 10.2-RELEASE-p5
>>=20
>> nvme0@pci0:132:0:0:     class=3D0x010802 card=3D0x1f971028 =
chip=3D0xa820144d rev=3D0x03 hdr=3D0x00
>>    vendor     =3D 'Samsung Electronics Co Ltd'
>>    class      =3D mass storage
>>    subclass   =3D NVM
>>=20
>> --=20
>> Sean Kelly
>> smkelly@smkelly.org
>> http://smkelly.org
>>=20
>> _______________________________________________
>> freebsd-stable@freebsd.org <mailto:freebsd-stable@freebsd.org> =
mailing list
>> https://lists.freebsd.org/mailman/listinfo/freebsd-stable =
<https://lists.freebsd.org/mailman/listinfo/freebsd-stable>;
>> To unsubscribe, send any mail to =
"freebsd-stable-unsubscribe@freebsd.org =
<mailto:freebsd-stable-unsubscribe@freebsd.org>"




Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?27228FE7-5FF9-4F58-9E23-42A66806C374>