|  | /* | 
|  | * Copyright (c) 2000-2003,2005 Silicon Graphics, Inc. | 
|  | * All Rights Reserved. | 
|  | * | 
|  | * This program is free software; you can redistribute it and/or | 
|  | * modify it under the terms of the GNU General Public License as | 
|  | * published by the Free Software Foundation. | 
|  | * | 
|  | * This program is distributed in the hope that it would be useful, | 
|  | * but WITHOUT ANY WARRANTY; without even the implied warranty of | 
|  | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the | 
|  | * GNU General Public License for more details. | 
|  | * | 
|  | * You should have received a copy of the GNU General Public License | 
|  | * along with this program; if not, write the Free Software Foundation, | 
|  | * Inc.,  51 Franklin St, Fifth Floor, Boston, MA  02110-1301  USA | 
|  | */ | 
|  | #include "xfs.h" | 
|  |  | 
|  | uint64_t vn_generation;		/* vnode generation number */ | 
|  | DEFINE_SPINLOCK(vnumber_lock); | 
|  |  | 
|  | /* | 
|  | * Dedicated vnode inactive/reclaim sync semaphores. | 
|  | * Prime number of hash buckets since address is used as the key. | 
|  | */ | 
|  | #define NVSYNC                  37 | 
|  | #define vptosync(v)             (&vsync[((unsigned long)v) % NVSYNC]) | 
|  | static wait_queue_head_t vsync[NVSYNC]; | 
|  |  | 
|  | void | 
|  | vn_init(void) | 
|  | { | 
|  | int i; | 
|  |  | 
|  | for (i = 0; i < NVSYNC; i++) | 
|  | init_waitqueue_head(&vsync[i]); | 
|  | } | 
|  |  | 
|  | void | 
|  | vn_iowait( | 
|  | bhv_vnode_t	*vp) | 
|  | { | 
|  | wait_queue_head_t *wq = vptosync(vp); | 
|  |  | 
|  | wait_event(*wq, (atomic_read(&vp->v_iocount) == 0)); | 
|  | } | 
|  |  | 
|  | void | 
|  | vn_iowake( | 
|  | bhv_vnode_t	*vp) | 
|  | { | 
|  | if (atomic_dec_and_test(&vp->v_iocount)) | 
|  | wake_up(vptosync(vp)); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Volume managers supporting multiple paths can send back ENODEV when the | 
|  | * final path disappears.  In this case continuing to fill the page cache | 
|  | * with dirty data which cannot be written out is evil, so prevent that. | 
|  | */ | 
|  | void | 
|  | vn_ioerror( | 
|  | bhv_vnode_t	*vp, | 
|  | int		error, | 
|  | char		*f, | 
|  | int		l) | 
|  | { | 
|  | if (unlikely(error == -ENODEV)) | 
|  | bhv_vfs_force_shutdown(vp->v_vfsp, SHUTDOWN_DEVICE_REQ, f, l); | 
|  | } | 
|  |  | 
|  | bhv_vnode_t * | 
|  | vn_initialize( | 
|  | struct inode	*inode) | 
|  | { | 
|  | bhv_vnode_t	*vp = vn_from_inode(inode); | 
|  |  | 
|  | XFS_STATS_INC(vn_active); | 
|  | XFS_STATS_INC(vn_alloc); | 
|  |  | 
|  | vp->v_flag = VMODIFIED; | 
|  | spinlock_init(&vp->v_lock, "v_lock"); | 
|  |  | 
|  | spin_lock(&vnumber_lock); | 
|  | if (!++vn_generation)	/* v_number shouldn't be zero */ | 
|  | vn_generation++; | 
|  | vp->v_number = vn_generation; | 
|  | spin_unlock(&vnumber_lock); | 
|  |  | 
|  | ASSERT(VN_CACHED(vp) == 0); | 
|  |  | 
|  | /* Initialize the first behavior and the behavior chain head. */ | 
|  | vn_bhv_head_init(VN_BHV_HEAD(vp), "vnode"); | 
|  |  | 
|  | atomic_set(&vp->v_iocount, 0); | 
|  |  | 
|  | #ifdef	XFS_VNODE_TRACE | 
|  | vp->v_trace = ktrace_alloc(VNODE_TRACE_SIZE, KM_SLEEP); | 
|  | #endif	/* XFS_VNODE_TRACE */ | 
|  |  | 
|  | vn_trace_exit(vp, __FUNCTION__, (inst_t *)__return_address); | 
|  | return vp; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Revalidate the Linux inode from the vattr. | 
|  | * Note: i_size _not_ updated; we must hold the inode | 
|  | * semaphore when doing that - callers responsibility. | 
|  | */ | 
|  | void | 
|  | vn_revalidate_core( | 
|  | bhv_vnode_t	*vp, | 
|  | bhv_vattr_t	*vap) | 
|  | { | 
|  | struct inode	*inode = vn_to_inode(vp); | 
|  |  | 
|  | inode->i_mode	    = vap->va_mode; | 
|  | inode->i_nlink	    = vap->va_nlink; | 
|  | inode->i_uid	    = vap->va_uid; | 
|  | inode->i_gid	    = vap->va_gid; | 
|  | inode->i_blocks	    = vap->va_nblocks; | 
|  | inode->i_mtime	    = vap->va_mtime; | 
|  | inode->i_ctime	    = vap->va_ctime; | 
|  | if (vap->va_xflags & XFS_XFLAG_IMMUTABLE) | 
|  | inode->i_flags |= S_IMMUTABLE; | 
|  | else | 
|  | inode->i_flags &= ~S_IMMUTABLE; | 
|  | if (vap->va_xflags & XFS_XFLAG_APPEND) | 
|  | inode->i_flags |= S_APPEND; | 
|  | else | 
|  | inode->i_flags &= ~S_APPEND; | 
|  | if (vap->va_xflags & XFS_XFLAG_SYNC) | 
|  | inode->i_flags |= S_SYNC; | 
|  | else | 
|  | inode->i_flags &= ~S_SYNC; | 
|  | if (vap->va_xflags & XFS_XFLAG_NOATIME) | 
|  | inode->i_flags |= S_NOATIME; | 
|  | else | 
|  | inode->i_flags &= ~S_NOATIME; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Revalidate the Linux inode from the vnode. | 
|  | */ | 
|  | int | 
|  | __vn_revalidate( | 
|  | bhv_vnode_t	*vp, | 
|  | bhv_vattr_t	*vattr) | 
|  | { | 
|  | int		error; | 
|  |  | 
|  | vn_trace_entry(vp, __FUNCTION__, (inst_t *)__return_address); | 
|  | vattr->va_mask = XFS_AT_STAT | XFS_AT_XFLAGS; | 
|  | error = bhv_vop_getattr(vp, vattr, 0, NULL); | 
|  | if (likely(!error)) { | 
|  | vn_revalidate_core(vp, vattr); | 
|  | VUNMODIFY(vp); | 
|  | } | 
|  | return -error; | 
|  | } | 
|  |  | 
|  | int | 
|  | vn_revalidate( | 
|  | bhv_vnode_t	*vp) | 
|  | { | 
|  | bhv_vattr_t	vattr; | 
|  |  | 
|  | return __vn_revalidate(vp, &vattr); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Add a reference to a referenced vnode. | 
|  | */ | 
|  | bhv_vnode_t * | 
|  | vn_hold( | 
|  | bhv_vnode_t	*vp) | 
|  | { | 
|  | struct inode	*inode; | 
|  |  | 
|  | XFS_STATS_INC(vn_hold); | 
|  |  | 
|  | VN_LOCK(vp); | 
|  | inode = igrab(vn_to_inode(vp)); | 
|  | ASSERT(inode); | 
|  | VN_UNLOCK(vp, 0); | 
|  |  | 
|  | return vp; | 
|  | } | 
|  |  | 
|  | #ifdef	XFS_VNODE_TRACE | 
|  |  | 
|  | #define KTRACE_ENTER(vp, vk, s, line, ra)			\ | 
|  | ktrace_enter(	(vp)->v_trace,				\ | 
|  | /*  0 */		(void *)(__psint_t)(vk),		\ | 
|  | /*  1 */		(void *)(s),				\ | 
|  | /*  2 */		(void *)(__psint_t) line,		\ | 
|  | /*  3 */		(void *)(__psint_t)(vn_count(vp)),	\ | 
|  | /*  4 */		(void *)(ra),				\ | 
|  | /*  5 */		(void *)(__psunsigned_t)(vp)->v_flag,	\ | 
|  | /*  6 */		(void *)(__psint_t)current_cpu(),	\ | 
|  | /*  7 */		(void *)(__psint_t)current_pid(),	\ | 
|  | /*  8 */		(void *)__return_address,		\ | 
|  | /*  9 */		NULL, NULL, NULL, NULL, NULL, NULL, NULL) | 
|  |  | 
|  | /* | 
|  | * Vnode tracing code. | 
|  | */ | 
|  | void | 
|  | vn_trace_entry(bhv_vnode_t *vp, const char *func, inst_t *ra) | 
|  | { | 
|  | KTRACE_ENTER(vp, VNODE_KTRACE_ENTRY, func, 0, ra); | 
|  | } | 
|  |  | 
|  | void | 
|  | vn_trace_exit(bhv_vnode_t *vp, const char *func, inst_t *ra) | 
|  | { | 
|  | KTRACE_ENTER(vp, VNODE_KTRACE_EXIT, func, 0, ra); | 
|  | } | 
|  |  | 
|  | void | 
|  | vn_trace_hold(bhv_vnode_t *vp, char *file, int line, inst_t *ra) | 
|  | { | 
|  | KTRACE_ENTER(vp, VNODE_KTRACE_HOLD, file, line, ra); | 
|  | } | 
|  |  | 
|  | void | 
|  | vn_trace_ref(bhv_vnode_t *vp, char *file, int line, inst_t *ra) | 
|  | { | 
|  | KTRACE_ENTER(vp, VNODE_KTRACE_REF, file, line, ra); | 
|  | } | 
|  |  | 
|  | void | 
|  | vn_trace_rele(bhv_vnode_t *vp, char *file, int line, inst_t *ra) | 
|  | { | 
|  | KTRACE_ENTER(vp, VNODE_KTRACE_RELE, file, line, ra); | 
|  | } | 
|  | #endif	/* XFS_VNODE_TRACE */ |