Skip site navigation (1)Skip section navigation (2)
Date:      Thu, 22 Mar 2018 14:52:01 -1000 (HST)
From:      Jeff Roberson <jroberson@jroberson.net>
To:        Justin Hibbits <jrh29@alumni.cwru.edu>
Cc:        Jeff Roberson <jeff@freebsd.org>,  src-committers <src-committers@freebsd.org>, svn-src-all@freebsd.org,  svn-src-head@freebsd.org
Subject:   Re: svn commit: r331369 - head/sys/vm
Message-ID:  <alpine.BSF.2.21.1803221451420.2307@desktop>
In-Reply-To: <CAHSQbTBumEe2-=5DMG8NG2c3C6Agf-eUok7=ti_XGE_Fp7E=MQ@mail.gmail.com>
References:  <201803221921.w2MJLBi7058560@repo.freebsd.org> <CAHSQbTBumEe2-=5DMG8NG2c3C6Agf-eUok7=ti_XGE_Fp7E=MQ@mail.gmail.com>

next in thread | previous in thread | raw e-mail | index | archive | help
Thank you, working on it.  I had done a make universe before getting 
review feedback.

Jeff

On Thu, 22 Mar 2018, Justin Hibbits wrote:

> This broke gcc builds.
>
> On Thu, Mar 22, 2018 at 2:21 PM, Jeff Roberson <jeff@freebsd.org> wrote:
>> Author: jeff
>> Date: Thu Mar 22 19:21:11 2018
>> New Revision: 331369
>> URL: https://svnweb.freebsd.org/changeset/base/331369
>>
>> Log:
>>   Lock reservations with a dedicated lock in each reservation.  Protect the
>>   vmd_free_count with atomics.
>>
>>   This allows us to allocate and free from reservations without the free lock
>>   except where a superpage is allocated from the physical layer, which is
>>   roughly 1/512 of the operations on amd64.
>>
>>   Use the counter api to eliminate cache conention on counters.
>>
>>   Reviewed by:  markj
>>   Tested by:    pho
>>   Sponsored by: Netflix, Dell/EMC Isilon
>>   Differential Revision:        https://reviews.freebsd.org/D14707
>>
>> Modified:
>>   head/sys/vm/vm_page.c
>>   head/sys/vm/vm_pagequeue.h
>>   head/sys/vm/vm_reserv.c
>>   head/sys/vm/vm_reserv.h
>>
>> Modified: head/sys/vm/vm_page.c
>> ==============================================================================
>> --- head/sys/vm/vm_page.c       Thu Mar 22 19:11:43 2018        (r331368)
>> +++ head/sys/vm/vm_page.c       Thu Mar 22 19:21:11 2018        (r331369)
>> @@ -177,7 +177,6 @@ static uma_zone_t fakepg_zone;
>>  static void vm_page_alloc_check(vm_page_t m);
>>  static void vm_page_clear_dirty_mask(vm_page_t m, vm_page_bits_t pagebits);
>>  static void vm_page_enqueue(uint8_t queue, vm_page_t m);
>> -static void vm_page_free_phys(struct vm_domain *vmd, vm_page_t m);
>>  static void vm_page_init(void *dummy);
>>  static int vm_page_insert_after(vm_page_t m, vm_object_t object,
>>      vm_pindex_t pindex, vm_page_t mpred);
>> @@ -1677,10 +1676,10 @@ vm_page_alloc_after(vm_object_t object, vm_pindex_t pi
>>   * for the request class and false otherwise.
>>   */
>>  int
>> -vm_domain_available(struct vm_domain *vmd, int req, int npages)
>> +vm_domain_allocate(struct vm_domain *vmd, int req, int npages)
>>  {
>> +       u_int limit, old, new;
>>
>> -       vm_domain_free_assert_locked(vmd);
>>         req = req & VM_ALLOC_CLASS_MASK;
>>
>>         /*
>> @@ -1688,15 +1687,34 @@ vm_domain_available(struct vm_domain *vmd, int req, in
>>          */
>>         if (curproc == pageproc && req != VM_ALLOC_INTERRUPT)
>>                 req = VM_ALLOC_SYSTEM;
>> +       if (req == VM_ALLOC_INTERRUPT)
>> +               limit = 0;
>> +       else if (req == VM_ALLOC_SYSTEM)
>> +               limit = vmd->vmd_interrupt_free_min;
>> +       else
>> +               limit = vmd->vmd_free_reserved;
>>
>> -       if (vmd->vmd_free_count >= npages + vmd->vmd_free_reserved ||
>> -           (req == VM_ALLOC_SYSTEM &&
>> -           vmd->vmd_free_count >= npages + vmd->vmd_interrupt_free_min) ||
>> -           (req == VM_ALLOC_INTERRUPT &&
>> -           vmd->vmd_free_count >= npages))
>> -               return (1);
>> +       /*
>> +        * Attempt to reserve the pages.  Fail if we're below the limit.
>> +        */
>> +       limit += npages;
>> +       old = vmd->vmd_free_count;
>> +       do {
>> +               if (old < limit)
>> +                       return (0);
>> +               new = old - npages;
>> +       } while (atomic_fcmpset_int(&vmd->vmd_free_count, &old, new) == 0);
>>
>> -       return (0);
>> +       /* Wake the page daemon if we've crossed the threshold. */
>> +       if (vm_paging_needed(vmd, new) && !vm_paging_needed(vmd, old))
>> +               pagedaemon_wakeup(vmd->vmd_domain);
>> +
>> +       /* Only update bitsets on transitions. */
>> +       if ((old >= vmd->vmd_free_min && new < vmd->vmd_free_min) ||
>> +           (old >= vmd->vmd_free_severe && new < vmd->vmd_free_severe))
>> +               vm_domain_set(vmd);
>> +
>> +       return (1);
>>  }
>>
>>  vm_page_t
>> @@ -1723,44 +1741,34 @@ vm_page_alloc_domain_after(vm_object_t object, vm_pind
>>  again:
>>         m = NULL;
>>  #if VM_NRESERVLEVEL > 0
>> +       /*
>> +        * Can we allocate the page from a reservation?
>> +        */
>>         if (vm_object_reserv(object) &&
>> -           (m = vm_reserv_extend(req, object, pindex, domain, mpred))
>> -           != NULL) {
>> +           ((m = vm_reserv_extend(req, object, pindex, domain, mpred)) != NULL ||
>> +           (m = vm_reserv_alloc_page(req, object, pindex, domain, mpred)) != NULL)) {
>>                 domain = vm_phys_domain(m);
>>                 vmd = VM_DOMAIN(domain);
>>                 goto found;
>>         }
>>  #endif
>>         vmd = VM_DOMAIN(domain);
>> -       vm_domain_free_lock(vmd);
>> -       if (vm_domain_available(vmd, req, 1)) {
>> +       if (vm_domain_allocate(vmd, req, 1)) {
>>                 /*
>> -                * Can we allocate the page from a reservation?
>> +                * If not, allocate it from the free page queues.
>>                  */
>> +               vm_domain_free_lock(vmd);
>> +               m = vm_phys_alloc_pages(domain, object != NULL ?
>> +                   VM_FREEPOOL_DEFAULT : VM_FREEPOOL_DIRECT, 0);
>> +               vm_domain_free_unlock(vmd);
>> +               if (m == NULL) {
>> +                       vm_domain_freecnt_inc(vmd, 1);
>>  #if VM_NRESERVLEVEL > 0
>> -               if (!vm_object_reserv(object) ||
>> -                   (m = vm_reserv_alloc_page(object, pindex,
>> -                   domain, mpred)) == NULL)
>> +                       if (vm_reserv_reclaim_inactive(domain))
>> +                               goto again;
>>  #endif
>> -               {
>> -                       /*
>> -                        * If not, allocate it from the free page queues.
>> -                        */
>> -                       m = vm_phys_alloc_pages(domain, object != NULL ?
>> -                           VM_FREEPOOL_DEFAULT : VM_FREEPOOL_DIRECT, 0);
>> -#if VM_NRESERVLEVEL > 0
>> -                       if (m == NULL && vm_reserv_reclaim_inactive(domain)) {
>> -                               m = vm_phys_alloc_pages(domain,
>> -                                   object != NULL ?
>> -                                   VM_FREEPOOL_DEFAULT : VM_FREEPOOL_DIRECT,
>> -                                   0);
>> -                       }
>> -#endif
>>                 }
>>         }
>> -       if (m != NULL)
>> -               vm_domain_freecnt_dec(vmd, 1);
>> -       vm_domain_free_unlock(vmd);
>>         if (m == NULL) {
>>                 /*
>>                  * Not allocatable, give up.
>> @@ -1775,9 +1783,7 @@ again:
>>          */
>>         KASSERT(m != NULL, ("missing page"));
>>
>> -#if VM_NRESERVLEVEL > 0
>>  found:
>> -#endif
>
> 'found' is now declared, but unused on powerpc64.
>
> - Justin
>



Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?alpine.BSF.2.21.1803221451420.2307>