From owner-freebsd-hackers@FreeBSD.ORG Tue Sep 21 07:12:10 2010 Return-Path: Delivered-To: freebsd-hackers@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id EBFCD106564A; Tue, 21 Sep 2010 07:12:10 +0000 (UTC) (envelope-from avg@freebsd.org) Received: from citadel.icyb.net.ua (citadel.icyb.net.ua [212.40.38.140]) by mx1.freebsd.org (Postfix) with ESMTP id 9AA478FC23; Tue, 21 Sep 2010 07:12:09 +0000 (UTC) Received: from porto.topspin.kiev.ua (porto-e.starpoint.kiev.ua [212.40.38.100]) by citadel.icyb.net.ua (8.8.8p3/ICyb-2.3exp) with ESMTP id KAA26990; Tue, 21 Sep 2010 10:12:07 +0300 (EEST) (envelope-from avg@freebsd.org) Received: from localhost.topspin.kiev.ua ([127.0.0.1]) by porto.topspin.kiev.ua with esmtp (Exim 4.34 (FreeBSD)) id 1Oxx1L-000NpQ-1w; Tue, 21 Sep 2010 10:12:07 +0300 Message-ID: <4C985AC6.60906@freebsd.org> Date: Tue, 21 Sep 2010 10:12:06 +0300 From: Andriy Gapon User-Agent: Mozilla/5.0 (X11; U; FreeBSD amd64; en-US; rv:1.9.2.9) Gecko/20100918 Lightning/1.0b2 Thunderbird/3.1.4 MIME-Version: 1.0 To: Jeff Roberson References: <4C93236B.4050906@freebsd.org> <4C935F56.4030903@freebsd.org> <4C98500D.5040109@freebsd.org> In-Reply-To: X-Enigmail-Version: 1.1.2 Content-Type: text/plain; charset=ISO-8859-1 Content-Transfer-Encoding: 7bit Cc: Andre Oppermann , Jeff Roberson , Robert Watson , freebsd-hackers@freebsd.org Subject: Re: zfs + uma X-BeenThere: freebsd-hackers@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Technical Discussions relating to FreeBSD List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Tue, 21 Sep 2010 07:12:11 -0000 on 21/09/2010 09:35 Jeff Roberson said the following: > On Tue, 21 Sep 2010, Andriy Gapon wrote: > >> on 19/09/2010 01:16 Jeff Roberson said the following: >>> Additionally we could make a last ditch flush mechanism that runs on each cpu in >> >> How would you qualify a "last ditch" trigger? >> Would this be called from "standard" vm_lowmem look or would there be some extra >> check for even more severe memory condition? > > If lowmem does not make enough progress to improve the condition. Do we have a good way to detect that? I see that currently vm_lowmem is always invoked with argument value of zero. -- Andriy Gapon