Skip site navigation (1)Skip section navigation (2)
Date:      Sat, 31 Mar 2012 06:50:15 GMT
From:      dfilter@FreeBSD.ORG (dfilter service)
To:        freebsd-bugs@FreeBSD.org
Subject:   Re: kern/165927: commit references a PR
Message-ID:  <201203310650.q2V6oFDn034907@freefall.freebsd.org>

next in thread | raw e-mail | index | archive | help
The following reply was made to PR kern/165927; it has been noted by GNATS.

From: dfilter@FreeBSD.ORG (dfilter service)
To: bug-followup@FreeBSD.org
Cc:  
Subject: Re: kern/165927: commit references a PR
Date: Sat, 31 Mar 2012 06:45:20 +0000 (UTC)

 Author: kib
 Date: Sat Mar 31 06:44:48 2012
 New Revision: 233728
 URL: http://svn.freebsd.org/changeset/base/233728
 
 Log:
   MFC r233100:
   In vm_object_page_clean(), do not clean OBJ_MIGHTBEDIRTY object flag
   if the filesystem performed short write and we are skipping the page
   due to this.
   
   Propogate write error from the pager back to the callers of
   vm_pageout_flush().  Report the failure to write a page from the
   requested range as the FALSE return value from vm_object_page_clean(),
   and propagate it back to msync(2) to return EIO to usermode.
   
   While there, convert the clearobjflags variable in the
   vm_object_page_clean() and arguments of the helper functions to
   boolean.
   
   PR:	kern/165927
 
 Modified:
   stable/9/sys/vm/vm_contig.c
   stable/9/sys/vm/vm_map.c
   stable/9/sys/vm/vm_mmap.c
   stable/9/sys/vm/vm_object.c
   stable/9/sys/vm/vm_object.h
   stable/9/sys/vm/vm_pageout.c
   stable/9/sys/vm/vm_pageout.h
 Directory Properties:
   stable/9/sys/   (props changed)
 
 Modified: stable/9/sys/vm/vm_contig.c
 ==============================================================================
 --- stable/9/sys/vm/vm_contig.c	Sat Mar 31 01:21:54 2012	(r233727)
 +++ stable/9/sys/vm/vm_contig.c	Sat Mar 31 06:44:48 2012	(r233728)
 @@ -139,7 +139,8 @@ vm_contig_launder_page(vm_page_t m, vm_p
  			   object->type == OBJT_DEFAULT) {
  			vm_page_unlock_queues();
  			m_tmp = m;
 -			vm_pageout_flush(&m_tmp, 1, VM_PAGER_PUT_SYNC, 0, NULL);
 +			vm_pageout_flush(&m_tmp, 1, VM_PAGER_PUT_SYNC, 0,
 +			    NULL, NULL);
  			VM_OBJECT_UNLOCK(object);
  			vm_page_lock_queues();
  			return (0);
 
 Modified: stable/9/sys/vm/vm_map.c
 ==============================================================================
 --- stable/9/sys/vm/vm_map.c	Sat Mar 31 01:21:54 2012	(r233727)
 +++ stable/9/sys/vm/vm_map.c	Sat Mar 31 06:44:48 2012	(r233728)
 @@ -2591,6 +2591,7 @@ vm_map_sync(
  	vm_object_t object;
  	vm_ooffset_t offset;
  	unsigned int last_timestamp;
 +	boolean_t failed;
  
  	vm_map_lock_read(map);
  	VM_MAP_RANGE_CHECK(map, start, end);
 @@ -2620,6 +2621,7 @@ vm_map_sync(
  
  	if (invalidate)
  		pmap_remove(map->pmap, start, end);
 +	failed = FALSE;
  
  	/*
  	 * Make a second pass, cleaning/uncaching pages from the indicated
 @@ -2648,7 +2650,8 @@ vm_map_sync(
  		vm_object_reference(object);
  		last_timestamp = map->timestamp;
  		vm_map_unlock_read(map);
 -		vm_object_sync(object, offset, size, syncio, invalidate);
 +		if (!vm_object_sync(object, offset, size, syncio, invalidate))
 +			failed = TRUE;
  		start += size;
  		vm_object_deallocate(object);
  		vm_map_lock_read(map);
 @@ -2658,7 +2661,7 @@ vm_map_sync(
  	}
  
  	vm_map_unlock_read(map);
 -	return (KERN_SUCCESS);
 +	return (failed ? KERN_FAILURE : KERN_SUCCESS);
  }
  
  /*
 
 Modified: stable/9/sys/vm/vm_mmap.c
 ==============================================================================
 --- stable/9/sys/vm/vm_mmap.c	Sat Mar 31 01:21:54 2012	(r233727)
 +++ stable/9/sys/vm/vm_mmap.c	Sat Mar 31 06:44:48 2012	(r233728)
 @@ -509,6 +509,8 @@ sys_msync(td, uap)
  		return (EINVAL);	/* Sun returns ENOMEM? */
  	case KERN_INVALID_ARGUMENT:
  		return (EBUSY);
 +	case KERN_FAILURE:
 +		return (EIO);
  	default:
  		return (EINVAL);
  	}
 
 Modified: stable/9/sys/vm/vm_object.c
 ==============================================================================
 --- stable/9/sys/vm/vm_object.c	Sat Mar 31 01:21:54 2012	(r233727)
 +++ stable/9/sys/vm/vm_object.c	Sat Mar 31 06:44:48 2012	(r233728)
 @@ -101,9 +101,10 @@ SYSCTL_INT(_vm, OID_AUTO, old_msync, CTL
      "Use old (insecure) msync behavior");
  
  static int	vm_object_page_collect_flush(vm_object_t object, vm_page_t p,
 -		    int pagerflags, int flags, int *clearobjflags);
 +		    int pagerflags, int flags, boolean_t *clearobjflags,
 +		    boolean_t *eio);
  static boolean_t vm_object_page_remove_write(vm_page_t p, int flags,
 -		    int *clearobjflags);
 +		    boolean_t *clearobjflags);
  static void	vm_object_qcollapse(vm_object_t object);
  static void	vm_object_vndeallocate(vm_object_t object);
  
 @@ -774,7 +775,7 @@ vm_object_terminate(vm_object_t object)
   * page should be flushed, and FALSE otherwise.
   */
  static boolean_t
 -vm_object_page_remove_write(vm_page_t p, int flags, int *clearobjflags)
 +vm_object_page_remove_write(vm_page_t p, int flags, boolean_t *clearobjflags)
  {
  
  	/*
 @@ -783,7 +784,7 @@ vm_object_page_remove_write(vm_page_t p,
  	 * cleared in this case so we do not have to set them.
  	 */
  	if ((flags & OBJPC_NOSYNC) != 0 && (p->oflags & VPO_NOSYNC) != 0) {
 -		*clearobjflags = 0;
 +		*clearobjflags = FALSE;
  		return (FALSE);
  	} else {
  		pmap_remove_write(p);
 @@ -805,21 +806,25 @@ vm_object_page_remove_write(vm_page_t p,
   *	Odd semantics: if start == end, we clean everything.
   *
   *	The object must be locked.
 + *
 + *	Returns FALSE if some page from the range was not written, as
 + *	reported by the pager, and TRUE otherwise.
   */
 -void
 +boolean_t
  vm_object_page_clean(vm_object_t object, vm_ooffset_t start, vm_ooffset_t end,
      int flags)
  {
  	vm_page_t np, p;
  	vm_pindex_t pi, tend, tstart;
 -	int clearobjflags, curgeneration, n, pagerflags;
 +	int curgeneration, n, pagerflags;
 +	boolean_t clearobjflags, eio, res;
  
  	mtx_assert(&vm_page_queue_mtx, MA_NOTOWNED);
  	VM_OBJECT_LOCK_ASSERT(object, MA_OWNED);
  	KASSERT(object->type == OBJT_VNODE, ("Not a vnode object"));
  	if ((object->flags & OBJ_MIGHTBEDIRTY) == 0 ||
  	    object->resident_page_count == 0)
 -		return;
 +		return (TRUE);
  
  	pagerflags = (flags & (OBJPC_SYNC | OBJPC_INVAL)) != 0 ?
  	    VM_PAGER_PUT_SYNC : VM_PAGER_CLUSTER_OK;
 @@ -828,6 +833,7 @@ vm_object_page_clean(vm_object_t object,
  	tstart = OFF_TO_IDX(start);
  	tend = (end == 0) ? object->size : OFF_TO_IDX(end + PAGE_MASK);
  	clearobjflags = tstart == 0 && tend >= object->size;
 +	res = TRUE;
  
  rescan:
  	curgeneration = object->generation;
 @@ -844,7 +850,7 @@ rescan:
  				if ((flags & OBJPC_SYNC) != 0)
  					goto rescan;
  				else
 -					clearobjflags = 0;
 +					clearobjflags = FALSE;
  			}
  			np = vm_page_find_least(object, pi);
  			continue;
 @@ -853,12 +859,16 @@ rescan:
  			continue;
  
  		n = vm_object_page_collect_flush(object, p, pagerflags,
 -		    flags, &clearobjflags);
 +		    flags, &clearobjflags, &eio);
 +		if (eio) {
 +			res = FALSE;
 +			clearobjflags = FALSE;
 +		}
  		if (object->generation != curgeneration) {
  			if ((flags & OBJPC_SYNC) != 0)
  				goto rescan;
  			else
 -				clearobjflags = 0;
 +				clearobjflags = FALSE;
  		}
  
  		/*
 @@ -873,8 +883,10 @@ rescan:
  		 * behind, but there is not much we can do there if
  		 * filesystem refuses to write it.
  		 */
 -		if (n == 0)
 +		if (n == 0) {
  			n = 1;
 +			clearobjflags = FALSE;
 +		}
  		np = vm_page_find_least(object, pi + n);
  	}
  #if 0
 @@ -883,11 +895,12 @@ rescan:
  
  	if (clearobjflags)
  		vm_object_clear_flag(object, OBJ_MIGHTBEDIRTY);
 +	return (res);
  }
  
  static int
  vm_object_page_collect_flush(vm_object_t object, vm_page_t p, int pagerflags,
 -    int flags, int *clearobjflags)
 +    int flags, boolean_t *clearobjflags, boolean_t *eio)
  {
  	vm_page_t ma[vm_pageout_page_count], p_first, tp;
  	int count, i, mreq, runlen;
 @@ -920,7 +933,7 @@ vm_object_page_collect_flush(vm_object_t
  	for (tp = p_first, i = 0; i < count; tp = TAILQ_NEXT(tp, listq), i++)
  		ma[i] = tp;
  
 -	vm_pageout_flush(ma, count, pagerflags, mreq, &runlen);
 +	vm_pageout_flush(ma, count, pagerflags, mreq, &runlen, eio);
  	return (runlen);
  }
  
 @@ -938,17 +951,20 @@ vm_object_page_collect_flush(vm_object_t
   * Note: certain anonymous maps, such as MAP_NOSYNC maps,
   * may start out with a NULL object.
   */
 -void
 +boolean_t
  vm_object_sync(vm_object_t object, vm_ooffset_t offset, vm_size_t size,
      boolean_t syncio, boolean_t invalidate)
  {
  	vm_object_t backing_object;
  	struct vnode *vp;
  	struct mount *mp;
 -	int flags, fsync_after;
 +	int error, flags, fsync_after;
 +	boolean_t res;
  
  	if (object == NULL)
 -		return;
 +		return (TRUE);
 +	res = TRUE;
 +	error = 0;
  	VM_OBJECT_LOCK(object);
  	while ((backing_object = object->backing_object) != NULL) {
  		VM_OBJECT_LOCK(backing_object);
 @@ -994,13 +1010,16 @@ vm_object_sync(vm_object_t object, vm_oo
  			fsync_after = FALSE;
  		}
  		VM_OBJECT_LOCK(object);
 -		vm_object_page_clean(object, offset, offset + size, flags);
 +		res = vm_object_page_clean(object, offset, offset + size,
 +		    flags);
  		VM_OBJECT_UNLOCK(object);
  		if (fsync_after)
 -			(void) VOP_FSYNC(vp, MNT_WAIT, curthread);
 +			error = VOP_FSYNC(vp, MNT_WAIT, curthread);
  		VOP_UNLOCK(vp, 0);
  		VFS_UNLOCK_GIANT(vfslocked);
  		vn_finished_write(mp);
 +		if (error != 0)
 +			res = FALSE;
  		VM_OBJECT_LOCK(object);
  	}
  	if ((object->type == OBJT_VNODE ||
 @@ -1020,6 +1039,7 @@ vm_object_sync(vm_object_t object, vm_oo
  		    OFF_TO_IDX(offset + size + PAGE_MASK), flags);
  	}
  	VM_OBJECT_UNLOCK(object);
 +	return (res);
  }
  
  /*
 
 Modified: stable/9/sys/vm/vm_object.h
 ==============================================================================
 --- stable/9/sys/vm/vm_object.h	Sat Mar 31 01:21:54 2012	(r233727)
 +++ stable/9/sys/vm/vm_object.h	Sat Mar 31 06:44:48 2012	(r233728)
 @@ -228,7 +228,7 @@ void vm_object_set_writeable_dirty (vm_o
  void vm_object_init (void);
  void vm_object_page_cache(vm_object_t object, vm_pindex_t start,
      vm_pindex_t end);
 -void vm_object_page_clean(vm_object_t object, vm_ooffset_t start,
 +boolean_t vm_object_page_clean(vm_object_t object, vm_ooffset_t start,
      vm_ooffset_t end, int flags);
  void vm_object_page_remove(vm_object_t object, vm_pindex_t start,
      vm_pindex_t end, int options);
 @@ -239,7 +239,7 @@ void vm_object_reference_locked(vm_objec
  int  vm_object_set_memattr(vm_object_t object, vm_memattr_t memattr);
  void vm_object_shadow (vm_object_t *, vm_ooffset_t *, vm_size_t);
  void vm_object_split(vm_map_entry_t);
 -void vm_object_sync(vm_object_t, vm_ooffset_t, vm_size_t, boolean_t,
 +boolean_t vm_object_sync(vm_object_t, vm_ooffset_t, vm_size_t, boolean_t,
      boolean_t);
  void vm_object_madvise (vm_object_t, vm_pindex_t, int, int);
  #endif				/* _KERNEL */
 
 Modified: stable/9/sys/vm/vm_pageout.c
 ==============================================================================
 --- stable/9/sys/vm/vm_pageout.c	Sat Mar 31 01:21:54 2012	(r233727)
 +++ stable/9/sys/vm/vm_pageout.c	Sat Mar 31 06:44:48 2012	(r233728)
 @@ -445,7 +445,8 @@ more:
  	/*
  	 * we allow reads during pageouts...
  	 */
 -	return (vm_pageout_flush(&mc[page_base], pageout_count, 0, 0, NULL));
 +	return (vm_pageout_flush(&mc[page_base], pageout_count, 0, 0, NULL,
 +	    NULL));
  }
  
  /*
 @@ -459,9 +460,12 @@ more:
   *
   *	Returned runlen is the count of pages between mreq and first
   *	page after mreq with status VM_PAGER_AGAIN.
 + *	*eio is set to TRUE if pager returned VM_PAGER_ERROR or VM_PAGER_FAIL
 + *	for any page in runlen set.
   */
  int
 -vm_pageout_flush(vm_page_t *mc, int count, int flags, int mreq, int *prunlen)
 +vm_pageout_flush(vm_page_t *mc, int count, int flags, int mreq, int *prunlen,
 +    boolean_t *eio)
  {
  	vm_object_t object = mc[0]->object;
  	int pageout_status[count];
 @@ -493,6 +497,8 @@ vm_pageout_flush(vm_page_t *mc, int coun
  	vm_pager_put_pages(object, mc, count, flags, pageout_status);
  
  	runlen = count - mreq;
 +	if (eio != NULL)
 +		*eio = FALSE;
  	for (i = 0; i < count; i++) {
  		vm_page_t mt = mc[i];
  
 @@ -522,6 +528,8 @@ vm_pageout_flush(vm_page_t *mc, int coun
  			vm_page_lock(mt);
  			vm_page_activate(mt);
  			vm_page_unlock(mt);
 +			if (eio != NULL && i >= mreq && i - mreq < runlen)
 +				*eio = TRUE;
  			break;
  		case VM_PAGER_AGAIN:
  			if (i >= mreq && i - mreq < runlen)
 
 Modified: stable/9/sys/vm/vm_pageout.h
 ==============================================================================
 --- stable/9/sys/vm/vm_pageout.h	Sat Mar 31 01:21:54 2012	(r233727)
 +++ stable/9/sys/vm/vm_pageout.h	Sat Mar 31 06:44:48 2012	(r233728)
 @@ -102,7 +102,7 @@ extern void vm_waitpfault(void);
  
  #ifdef _KERNEL
  boolean_t vm_pageout_fallback_object_lock(vm_page_t, vm_page_t *);
 -int vm_pageout_flush(vm_page_t *, int, int, int, int *);
 +int vm_pageout_flush(vm_page_t *, int, int, int, int *, boolean_t *);
  void vm_pageout_oom(int shortage);
  boolean_t vm_pageout_page_lock(vm_page_t, vm_page_t *);
  void vm_contig_grow_cache(int, vm_paddr_t, vm_paddr_t);
 _______________________________________________
 svn-src-all@freebsd.org mailing list
 http://lists.freebsd.org/mailman/listinfo/svn-src-all
 To unsubscribe, send any mail to "svn-src-all-unsubscribe@freebsd.org"
 



Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?201203310650.q2V6oFDn034907>