From owner-svn-src-head@freebsd.org Tue Sep 3 15:42:14 2019 Return-Path: Delivered-To: svn-src-head@mailman.nyi.freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2610:1c1:1:606c::19:1]) by mailman.nyi.freebsd.org (Postfix) with ESMTP id 33505DF374; Tue, 3 Sep 2019 15:42:14 +0000 (UTC) (envelope-from mjg@FreeBSD.org) Received: from mxrelay.nyi.freebsd.org (mxrelay.nyi.freebsd.org [IPv6:2610:1c1:1:606c::19:3]) (using TLSv1.3 with cipher TLS_AES_256_GCM_SHA384 (256/256 bits) server-signature RSA-PSS (4096 bits) client-signature RSA-PSS (4096 bits) client-digest SHA256) (Client CN "mxrelay.nyi.freebsd.org", Issuer "Let's Encrypt Authority X3" (verified OK)) by mx1.freebsd.org (Postfix) with ESMTPS id 46NB5y0J0Pz3F8r; Tue, 3 Sep 2019 15:42:14 +0000 (UTC) (envelope-from mjg@FreeBSD.org) Received: from repo.freebsd.org (repo.freebsd.org [IPv6:2610:1c1:1:6068::e6a:0]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (Client did not present a certificate) by mxrelay.nyi.freebsd.org (Postfix) with ESMTPS id E1E8F51AE; Tue, 3 Sep 2019 15:42:13 +0000 (UTC) (envelope-from mjg@FreeBSD.org) Received: from repo.freebsd.org ([127.0.1.37]) by repo.freebsd.org (8.15.2/8.15.2) with ESMTP id x83FgDWj011812; Tue, 3 Sep 2019 15:42:13 GMT (envelope-from mjg@FreeBSD.org) Received: (from mjg@localhost) by repo.freebsd.org (8.15.2/8.15.2/Submit) id x83FgCNu011401; Tue, 3 Sep 2019 15:42:12 GMT (envelope-from mjg@FreeBSD.org) Message-Id: <201909031542.x83FgCNu011401@repo.freebsd.org> X-Authentication-Warning: repo.freebsd.org: mjg set sender to mjg@FreeBSD.org using -f From: Mateusz Guzik Date: Tue, 3 Sep 2019 15:42:12 +0000 (UTC) To: src-committers@freebsd.org, svn-src-all@freebsd.org, svn-src-head@freebsd.org Subject: svn commit: r351748 - in head/sys: cddl/contrib/opensolaris/uts/common/fs cddl/contrib/opensolaris/uts/common/fs/zfs kern sys X-SVN-Group: head X-SVN-Commit-Author: mjg X-SVN-Commit-Paths: in head/sys: cddl/contrib/opensolaris/uts/common/fs cddl/contrib/opensolaris/uts/common/fs/zfs kern sys X-SVN-Commit-Revision: 351748 X-SVN-Commit-Repository: base MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit X-BeenThere: svn-src-head@freebsd.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: SVN commit messages for the src tree for head/-current List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Tue, 03 Sep 2019 15:42:14 -0000 Author: mjg Date: Tue Sep 3 15:42:11 2019 New Revision: 351748 URL: https://svnweb.freebsd.org/changeset/base/351748 Log: vfs: implement usecount implying holdcnt vnodes have 2 reference counts - holdcnt to keep the vnode itself from getting freed and usecount to denote it is actively used. Previously all operations bumping usecount would also bump holdcnt, which is not necessary. We can detect if usecount is already > 1 (in which case holdcnt is also > 1) and utilize it to avoid bumping holdcnt on our own. This saves on atomic ops. Reviewed by: kib Tested by: pho (previous version) Sponsored by: The FreeBSD Foundation Differential Revision: https://reviews.freebsd.org/D21471 Modified: head/sys/cddl/contrib/opensolaris/uts/common/fs/vnode.c head/sys/cddl/contrib/opensolaris/uts/common/fs/zfs/zfs_ctldir.c head/sys/cddl/contrib/opensolaris/uts/common/fs/zfs/zfs_vnops.c head/sys/kern/vfs_cache.c head/sys/kern/vfs_hash.c head/sys/kern/vfs_subr.c head/sys/sys/lockmgr.h head/sys/sys/vnode.h Modified: head/sys/cddl/contrib/opensolaris/uts/common/fs/vnode.c ============================================================================== --- head/sys/cddl/contrib/opensolaris/uts/common/fs/vnode.c Tue Sep 3 15:26:11 2019 (r351747) +++ head/sys/cddl/contrib/opensolaris/uts/common/fs/vnode.c Tue Sep 3 15:42:11 2019 (r351748) @@ -87,7 +87,6 @@ vn_rele_async(vnode_t *vp, taskq_t *taskq) { VERIFY(vp->v_count > 0); if (refcount_release_if_not_last(&vp->v_usecount)) { - vdrop(vp); return; } VERIFY(taskq_dispatch((taskq_t *)taskq, Modified: head/sys/cddl/contrib/opensolaris/uts/common/fs/zfs/zfs_ctldir.c ============================================================================== --- head/sys/cddl/contrib/opensolaris/uts/common/fs/zfs/zfs_ctldir.c Tue Sep 3 15:26:11 2019 (r351747) +++ head/sys/cddl/contrib/opensolaris/uts/common/fs/zfs/zfs_ctldir.c Tue Sep 3 15:42:11 2019 (r351748) @@ -1196,6 +1196,7 @@ zfsctl_snapshot_vptocnp(struct vop_vptocnp_args *ap) vnode_t *vp; sfs_node_t *node; size_t len; + enum vgetstate vs; int locked; int error; @@ -1224,7 +1225,7 @@ zfsctl_snapshot_vptocnp(struct vop_vptocnp_args *ap) * before we can lock the vnode again. */ locked = VOP_ISLOCKED(vp); - vhold(vp); + vs = vget_prep(vp); vput(vp); /* Look up .zfs/snapshot, our parent. */ @@ -1236,7 +1237,7 @@ zfsctl_snapshot_vptocnp(struct vop_vptocnp_args *ap) bcopy(node->sn_name, ap->a_buf + *ap->a_buflen, len); } vfs_unbusy(mp); - vget(vp, locked | LK_VNHELD | LK_RETRY, curthread); + vget_finish(vp, locked | LK_RETRY, vs); return (error); } Modified: head/sys/cddl/contrib/opensolaris/uts/common/fs/zfs/zfs_vnops.c ============================================================================== --- head/sys/cddl/contrib/opensolaris/uts/common/fs/zfs/zfs_vnops.c Tue Sep 3 15:26:11 2019 (r351747) +++ head/sys/cddl/contrib/opensolaris/uts/common/fs/zfs/zfs_vnops.c Tue Sep 3 15:42:11 2019 (r351748) @@ -5916,6 +5916,7 @@ zfs_vptocnp(struct vop_vptocnp_args *ap) vnode_t *vp = ap->a_vp;; zfsvfs_t *zfsvfs = vp->v_vfsp->vfs_data; znode_t *zp = VTOZ(vp); + enum vgetstate vs; int ltype; int error; @@ -5948,10 +5949,10 @@ zfs_vptocnp(struct vop_vptocnp_args *ap) ZFS_EXIT(zfsvfs); covered_vp = vp->v_mount->mnt_vnodecovered; - vhold(covered_vp); + vs = vget_prep(covered_vp); ltype = VOP_ISLOCKED(vp); VOP_UNLOCK(vp, 0); - error = vget(covered_vp, LK_SHARED | LK_VNHELD, curthread); + error = vget_finish(covered_vp, LK_SHARED, vs); if (error == 0) { error = VOP_VPTOCNP(covered_vp, ap->a_vpp, ap->a_cred, ap->a_buf, ap->a_buflen); Modified: head/sys/kern/vfs_cache.c ============================================================================== --- head/sys/kern/vfs_cache.c Tue Sep 3 15:26:11 2019 (r351747) +++ head/sys/kern/vfs_cache.c Tue Sep 3 15:42:11 2019 (r351748) @@ -1255,6 +1255,7 @@ cache_lookup(struct vnode *dvp, struct vnode **vpp, st struct rwlock *blp; struct mtx *dvlp; uint32_t hash; + enum vgetstate vs; int error, ltype; if (__predict_false(!doingcache)) { @@ -1369,9 +1370,9 @@ success: ltype = VOP_ISLOCKED(dvp); VOP_UNLOCK(dvp, 0); } - vhold(*vpp); + vs = vget_prep(*vpp); cache_lookup_unlock(blp, dvlp); - error = vget(*vpp, cnp->cn_lkflags | LK_VNHELD, cnp->cn_thread); + error = vget_finish(*vpp, cnp->cn_lkflags, vs); if (cnp->cn_flags & ISDOTDOT) { vn_lock(dvp, ltype | LK_RETRY); if (dvp->v_iflag & VI_DOOMED) { @@ -2430,6 +2431,7 @@ vn_dir_dd_ino(struct vnode *vp) struct namecache *ncp; struct vnode *ddvp; struct mtx *vlp; + enum vgetstate vs; ASSERT_VOP_LOCKED(vp, "vn_dir_dd_ino"); vlp = VP2VNODELOCK(vp); @@ -2438,9 +2440,9 @@ vn_dir_dd_ino(struct vnode *vp) if ((ncp->nc_flag & NCF_ISDOTDOT) != 0) continue; ddvp = ncp->nc_dvp; - vhold(ddvp); + vs = vget_prep(ddvp); mtx_unlock(vlp); - if (vget(ddvp, LK_SHARED | LK_NOWAIT | LK_VNHELD, curthread)) + if (vget_finish(ddvp, LK_SHARED | LK_NOWAIT, vs)) return (NULL); return (ddvp); } Modified: head/sys/kern/vfs_hash.c ============================================================================== --- head/sys/kern/vfs_hash.c Tue Sep 3 15:26:11 2019 (r351747) +++ head/sys/kern/vfs_hash.c Tue Sep 3 15:42:11 2019 (r351748) @@ -76,6 +76,7 @@ vfs_hash_get(const struct mount *mp, u_int hash, int f struct vnode **vpp, vfs_hash_cmp_t *fn, void *arg) { struct vnode *vp; + enum vgetstate vs; int error; while (1) { @@ -87,9 +88,9 @@ vfs_hash_get(const struct mount *mp, u_int hash, int f continue; if (fn != NULL && fn(vp, arg)) continue; - vhold(vp); + vs = vget_prep(vp); rw_runlock(&vfs_hash_lock); - error = vget(vp, flags | LK_VNHELD, td); + error = vget_finish(vp, flags, vs); if (error == ENOENT && (flags & LK_NOWAIT) == 0) break; if (error) @@ -149,6 +150,7 @@ vfs_hash_insert(struct vnode *vp, u_int hash, int flag struct vnode **vpp, vfs_hash_cmp_t *fn, void *arg) { struct vnode *vp2; + enum vgetstate vs; int error; *vpp = NULL; @@ -162,9 +164,9 @@ vfs_hash_insert(struct vnode *vp, u_int hash, int flag continue; if (fn != NULL && fn(vp2, arg)) continue; - vhold(vp2); + vs = vget_prep(vp2); rw_wunlock(&vfs_hash_lock); - error = vget(vp2, flags | LK_VNHELD, td); + error = vget_finish(vp2, flags, vs); if (error == ENOENT && (flags & LK_NOWAIT) == 0) break; rw_wlock(&vfs_hash_lock); Modified: head/sys/kern/vfs_subr.c ============================================================================== --- head/sys/kern/vfs_subr.c Tue Sep 3 15:26:11 2019 (r351747) +++ head/sys/kern/vfs_subr.c Tue Sep 3 15:42:11 2019 (r351748) @@ -107,8 +107,6 @@ static int flushbuflist(struct bufv *bufv, int flags, static void syncer_shutdown(void *arg, int howto); static int vtryrecycle(struct vnode *vp); static void v_init_counters(struct vnode *); -static void v_incr_usecount(struct vnode *); -static void v_incr_usecount_locked(struct vnode *); static void v_incr_devcount(struct vnode *); static void v_decr_devcount(struct vnode *); static void vgonel(struct vnode *); @@ -2632,45 +2630,7 @@ v_init_counters(struct vnode *vp) refcount_init(&vp->v_usecount, 1); } -static void -v_incr_usecount_locked(struct vnode *vp) -{ - - ASSERT_VI_LOCKED(vp, __func__); - if ((vp->v_iflag & VI_OWEINACT) != 0) { - VNASSERT(vp->v_usecount == 0, vp, - ("vnode with usecount and VI_OWEINACT set")); - vp->v_iflag &= ~VI_OWEINACT; - VNODE_REFCOUNT_FENCE_REL(); - } - refcount_acquire(&vp->v_usecount); - v_incr_devcount(vp); -} - /* - * Increment the use count on the vnode, taking care to reference - * the driver's usecount if this is a chardev. - */ -static void -v_incr_usecount(struct vnode *vp) -{ - - ASSERT_VI_UNLOCKED(vp, __func__); - CTR2(KTR_VFS, "%s: vp %p", __func__, vp); - - if (vp->v_type != VCHR && - refcount_acquire_if_not_zero(&vp->v_usecount)) { - VNODE_REFCOUNT_FENCE_ACQ(); - VNASSERT((vp->v_iflag & VI_OWEINACT) == 0, vp, - ("vnode with usecount and VI_OWEINACT set")); - } else { - VI_LOCK(vp); - v_incr_usecount_locked(vp); - VI_UNLOCK(vp); - } -} - -/* * Increment si_usecount of the associated device, if any. */ static void @@ -2714,36 +2674,108 @@ v_decr_devcount(struct vnode *vp) * with atomic operations. Otherwise the interlock is taken covering * both the atomic and additional actions. */ +static enum vgetstate +_vget_prep(struct vnode *vp, bool interlock) +{ + enum vgetstate vs; + + if (__predict_true(vp->v_type != VCHR)) { + if (refcount_acquire_if_not_zero(&vp->v_usecount)) { + vs = VGET_USECOUNT; + } else { + _vhold(vp, interlock); + vs = VGET_HOLDCNT; + } + } else { + if (!interlock) + VI_LOCK(vp); + if (vp->v_usecount == 0) { + vholdl(vp); + vs = VGET_HOLDCNT; + } else { + v_incr_devcount(vp); + refcount_acquire(&vp->v_usecount); + vs = VGET_USECOUNT; + } + if (!interlock) + VI_UNLOCK(vp); + } + return (vs); +} + +enum vgetstate +vget_prep(struct vnode *vp) +{ + + return (_vget_prep(vp, false)); +} + int vget(struct vnode *vp, int flags, struct thread *td) { + enum vgetstate vs; + + MPASS(td == curthread); + + vs = _vget_prep(vp, (flags & LK_INTERLOCK) != 0); + return (vget_finish(vp, flags, vs)); +} + +int +vget_finish(struct vnode *vp, int flags, enum vgetstate vs) +{ int error, oweinact; VNASSERT((flags & LK_TYPE_MASK) != 0, vp, - ("vget: invalid lock operation")); + ("%s: invalid lock operation", __func__)); if ((flags & LK_INTERLOCK) != 0) ASSERT_VI_LOCKED(vp, __func__); else ASSERT_VI_UNLOCKED(vp, __func__); - if ((flags & LK_VNHELD) != 0) - VNASSERT((vp->v_holdcnt > 0), vp, - ("vget: LK_VNHELD passed but vnode not held")); + VNASSERT(vp->v_holdcnt > 0, vp, ("%s: vnode not held", __func__)); + if (vs == VGET_USECOUNT) { + VNASSERT(vp->v_usecount > 0, vp, + ("%s: vnode without usecount when VGET_USECOUNT was passed", + __func__)); + } - CTR3(KTR_VFS, "%s: vp %p with flags %d", __func__, vp, flags); - - if ((flags & LK_VNHELD) == 0) - _vhold(vp, (flags & LK_INTERLOCK) != 0); - if ((error = vn_lock(vp, flags)) != 0) { - vdrop(vp); + if (vs == VGET_USECOUNT) + vrele(vp); + else + vdrop(vp); CTR2(KTR_VFS, "%s: impossible to lock vnode %p", __func__, vp); return (error); } - if (vp->v_iflag & VI_DOOMED && (flags & LK_RETRY) == 0) - panic("vget: vn_lock failed to return ENOENT\n"); + + if (vs == VGET_USECOUNT) { + VNASSERT((vp->v_iflag & VI_OWEINACT) == 0, vp, + ("%s: vnode with usecount and VI_OWEINACT set", __func__)); + return (0); + } + /* + * We hold the vnode. If the usecount is 0 it will be utilized to keep + * the vnode around. Otherwise someone else lended their hold count and + * we have to drop ours. + */ + if (vp->v_type != VCHR && + refcount_acquire_if_not_zero(&vp->v_usecount)) { +#ifdef INVARIANTS + int old = atomic_fetchadd_int(&vp->v_holdcnt, -1) - 1; + VNASSERT(old > 0, vp, ("%s: wrong hold count", __func__)); +#else + refcount_release(&vp->v_holdcnt); +#endif + VNODE_REFCOUNT_FENCE_ACQ(); + VNASSERT((vp->v_iflag & VI_OWEINACT) == 0, vp, + ("%s: vnode with usecount and VI_OWEINACT set", __func__)); + return (0); + } + + /* * We don't guarantee that any particular close will * trigger inactive processing so just make a best effort * here at preventing a reference to a removed file. If @@ -2751,23 +2783,22 @@ vget(struct vnode *vp, int flags, struct thread *td) * * Upgrade our holdcnt to a usecount. */ - if (vp->v_type == VCHR || - !refcount_acquire_if_not_zero(&vp->v_usecount)) { - VI_LOCK(vp); - if ((vp->v_iflag & VI_OWEINACT) == 0) { - oweinact = 0; - } else { - oweinact = 1; - vp->v_iflag &= ~VI_OWEINACT; - VNODE_REFCOUNT_FENCE_REL(); - } - refcount_acquire(&vp->v_usecount); - v_incr_devcount(vp); - if (oweinact && VOP_ISLOCKED(vp) == LK_EXCLUSIVE && - (flags & LK_NOWAIT) == 0) - vinactive(vp, td); - VI_UNLOCK(vp); + VI_LOCK(vp); + if ((vp->v_iflag & VI_OWEINACT) == 0) { + oweinact = 0; + } else { + oweinact = 1; + vp->v_iflag &= ~VI_OWEINACT; + VNODE_REFCOUNT_FENCE_REL(); } + if (vp->v_usecount > 0) + refcount_release(&vp->v_holdcnt); + v_incr_devcount(vp); + refcount_acquire(&vp->v_usecount); + if (oweinact && VOP_ISLOCKED(vp) == LK_EXCLUSIVE && + (flags & LK_NOWAIT) == 0) + vinactive(vp, curthread); + VI_UNLOCK(vp); return (0); } @@ -2781,8 +2812,18 @@ vref(struct vnode *vp) ASSERT_VI_UNLOCKED(vp, __func__); CTR2(KTR_VFS, "%s: vp %p", __func__, vp); - _vhold(vp, false); - v_incr_usecount(vp); + if (vp->v_type != VCHR && + refcount_acquire_if_not_zero(&vp->v_usecount)) { + VNODE_REFCOUNT_FENCE_ACQ(); + VNASSERT(vp->v_holdcnt > 0, vp, + ("%s: active vnode not held", __func__)); + VNASSERT((vp->v_iflag & VI_OWEINACT) == 0, vp, + ("%s: vnode with usecount and VI_OWEINACT set", __func__)); + return; + } + VI_LOCK(vp); + vrefl(vp); + VI_UNLOCK(vp); } void @@ -2791,8 +2832,14 @@ vrefl(struct vnode *vp) ASSERT_VI_LOCKED(vp, __func__); CTR2(KTR_VFS, "%s: vp %p", __func__, vp); - _vhold(vp, true); - v_incr_usecount_locked(vp); + if (vp->v_usecount == 0) + vholdl(vp); + if ((vp->v_iflag & VI_OWEINACT) != 0) { + vp->v_iflag &= ~VI_OWEINACT; + VNODE_REFCOUNT_FENCE_REL(); + } + v_incr_devcount(vp); + refcount_acquire(&vp->v_usecount); } void @@ -2807,12 +2854,9 @@ vrefact(struct vnode *vp) return; } #ifdef INVARIANTS - int old = atomic_fetchadd_int(&vp->v_holdcnt, 1); - VNASSERT(old > 0, vp, ("%s: wrong hold count", __func__)); - old = atomic_fetchadd_int(&vp->v_usecount, 1); + int old = atomic_fetchadd_int(&vp->v_usecount, 1); VNASSERT(old > 0, vp, ("%s: wrong use count", __func__)); #else - refcount_acquire(&vp->v_holdcnt); refcount_acquire(&vp->v_usecount); #endif } @@ -2860,13 +2904,22 @@ vputx(struct vnode *vp, int func) CTR2(KTR_VFS, "%s: vp %p", __func__, vp); + /* + * It is an invariant that all VOP_* calls operate on a held vnode. + * We may be only having an implicit hold stemming from our usecount, + * which we are about to release. If we unlock the vnode afterwards we + * open a time window where someone else dropped the last usecount and + * proceeded to free the vnode before our unlock finished. For this + * reason we unlock the vnode early. This is a little bit wasteful as + * it may be the vnode is exclusively locked and inactive processing is + * needed, in which case we are adding work. + */ + if (func == VPUTX_VPUT) + VOP_UNLOCK(vp, 0); + if (vp->v_type != VCHR && - refcount_release_if_not_last(&vp->v_usecount)) { - if (func == VPUTX_VPUT) - VOP_UNLOCK(vp, 0); - vdrop(vp); + refcount_release_if_not_last(&vp->v_usecount)) return; - } VI_LOCK(vp); @@ -2875,17 +2928,16 @@ vputx(struct vnode *vp, int func) * prevent vgone() races. We drop the use count here and the * hold count below when we're done. */ - if (!refcount_release(&vp->v_usecount) || - (vp->v_iflag & VI_DOINGINACT)) { - if (func == VPUTX_VPUT) - VOP_UNLOCK(vp, 0); - v_decr_devcount(vp); + v_decr_devcount(vp); + if (!refcount_release(&vp->v_usecount)) { + VI_UNLOCK(vp); + return; + } + if (vp->v_iflag & VI_DOINGINACT) { vdropl(vp); return; } - v_decr_devcount(vp); - error = 0; if (vp->v_usecount != 0) { @@ -2904,8 +2956,6 @@ vputx(struct vnode *vp, int func) */ if (__predict_false(vp->v_iflag & VI_DOOMED) || VOP_NEED_INACTIVE(vp) == 0) { - if (func == VPUTX_VPUT) - VOP_UNLOCK(vp, 0); vdropl(vp); return; } @@ -2921,11 +2971,8 @@ vputx(struct vnode *vp, int func) VI_LOCK(vp); break; case VPUTX_VPUT: - if (VOP_ISLOCKED(vp) != LK_EXCLUSIVE) { - error = VOP_LOCK(vp, LK_UPGRADE | LK_INTERLOCK | - LK_NOWAIT); - VI_LOCK(vp); - } + error = VOP_LOCK(vp, LK_EXCLUSIVE | LK_INTERLOCK | LK_NOWAIT); + VI_LOCK(vp); break; case VPUTX_VUNREF: if (VOP_ISLOCKED(vp) != LK_EXCLUSIVE) { Modified: head/sys/sys/lockmgr.h ============================================================================== --- head/sys/sys/lockmgr.h Tue Sep 3 15:26:11 2019 (r351747) +++ head/sys/sys/lockmgr.h Tue Sep 3 15:42:11 2019 (r351748) @@ -164,7 +164,6 @@ _lockmgr_args_rw(struct lock *lk, u_int flags, struct #define LK_SLEEPFAIL 0x000800 #define LK_TIMELOCK 0x001000 #define LK_NODDLKTREAT 0x002000 -#define LK_VNHELD 0x004000 /* * Operations for lockmgr(). Modified: head/sys/sys/vnode.h ============================================================================== --- head/sys/sys/vnode.h Tue Sep 3 15:26:11 2019 (r351747) +++ head/sys/sys/vnode.h Tue Sep 3 15:42:11 2019 (r351748) @@ -58,6 +58,7 @@ enum vtype { VNON, VREG, VDIR, VBLK, VCHR, VLNK, VSOCK, VFIFO, VBAD, VMARKER }; +enum vgetstate { VGET_HOLDCNT, VGET_USECOUNT }; /* * Each underlying filesystem allocates its own private area and hangs * it from v_data. If non-null, this area is freed in getnewvnode(). @@ -652,7 +653,9 @@ int vcount(struct vnode *vp); #define vdropl(vp) _vdrop((vp), 1) void _vdrop(struct vnode *, bool); int vflush(struct mount *mp, int rootrefs, int flags, struct thread *td); -int vget(struct vnode *vp, int lockflag, struct thread *td); +int vget(struct vnode *vp, int flags, struct thread *td); +enum vgetstate vget_prep(struct vnode *vp); +int vget_finish(struct vnode *vp, int flags, enum vgetstate vs); void vgone(struct vnode *vp); #define vhold(vp) _vhold((vp), 0) #define vholdl(vp) _vhold((vp), 1)