| /* | 
 |  * Copyright (c) 2000-2003,2005 Silicon Graphics, Inc. | 
 |  * All Rights Reserved. | 
 |  * | 
 |  * This program is free software; you can redistribute it and/or | 
 |  * modify it under the terms of the GNU General Public License as | 
 |  * published by the Free Software Foundation. | 
 |  * | 
 |  * This program is distributed in the hope that it would be useful, | 
 |  * but WITHOUT ANY WARRANTY; without even the implied warranty of | 
 |  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the | 
 |  * GNU General Public License for more details. | 
 |  * | 
 |  * You should have received a copy of the GNU General Public License | 
 |  * along with this program; if not, write the Free Software Foundation, | 
 |  * Inc.,  51 Franklin St, Fifth Floor, Boston, MA  02110-1301  USA | 
 |  */ | 
 | #include "xfs.h" | 
 |  | 
 | uint64_t vn_generation;		/* vnode generation number */ | 
 | DEFINE_SPINLOCK(vnumber_lock); | 
 |  | 
 | /* | 
 |  * Dedicated vnode inactive/reclaim sync semaphores. | 
 |  * Prime number of hash buckets since address is used as the key. | 
 |  */ | 
 | #define NVSYNC                  37 | 
 | #define vptosync(v)             (&vsync[((unsigned long)v) % NVSYNC]) | 
 | STATIC wait_queue_head_t vsync[NVSYNC]; | 
 |  | 
 | void | 
 | vn_init(void) | 
 | { | 
 | 	int i; | 
 |  | 
 | 	for (i = 0; i < NVSYNC; i++) | 
 | 		init_waitqueue_head(&vsync[i]); | 
 | } | 
 |  | 
 | void | 
 | vn_iowait( | 
 | 	bhv_vnode_t	*vp) | 
 | { | 
 | 	wait_queue_head_t *wq = vptosync(vp); | 
 |  | 
 | 	wait_event(*wq, (atomic_read(&vp->v_iocount) == 0)); | 
 | } | 
 |  | 
 | void | 
 | vn_iowake( | 
 | 	bhv_vnode_t	*vp) | 
 | { | 
 | 	if (atomic_dec_and_test(&vp->v_iocount)) | 
 | 		wake_up(vptosync(vp)); | 
 | } | 
 |  | 
 | /* | 
 |  * Volume managers supporting multiple paths can send back ENODEV when the | 
 |  * final path disappears.  In this case continuing to fill the page cache | 
 |  * with dirty data which cannot be written out is evil, so prevent that. | 
 |  */ | 
 | void | 
 | vn_ioerror( | 
 | 	bhv_vnode_t	*vp, | 
 | 	int		error, | 
 | 	char		*f, | 
 | 	int		l) | 
 | { | 
 | 	if (unlikely(error == -ENODEV)) | 
 | 		bhv_vfs_force_shutdown(vp->v_vfsp, SHUTDOWN_DEVICE_REQ, f, l); | 
 | } | 
 |  | 
 | bhv_vnode_t * | 
 | vn_initialize( | 
 | 	struct inode	*inode) | 
 | { | 
 | 	bhv_vnode_t	*vp = vn_from_inode(inode); | 
 |  | 
 | 	XFS_STATS_INC(vn_active); | 
 | 	XFS_STATS_INC(vn_alloc); | 
 |  | 
 | 	vp->v_flag = VMODIFIED; | 
 | 	spinlock_init(&vp->v_lock, "v_lock"); | 
 |  | 
 | 	spin_lock(&vnumber_lock); | 
 | 	if (!++vn_generation)	/* v_number shouldn't be zero */ | 
 | 		vn_generation++; | 
 | 	vp->v_number = vn_generation; | 
 | 	spin_unlock(&vnumber_lock); | 
 |  | 
 | 	ASSERT(VN_CACHED(vp) == 0); | 
 |  | 
 | 	/* Initialize the first behavior and the behavior chain head. */ | 
 | 	vn_bhv_head_init(VN_BHV_HEAD(vp), "vnode"); | 
 |  | 
 | 	atomic_set(&vp->v_iocount, 0); | 
 |  | 
 | #ifdef	XFS_VNODE_TRACE | 
 | 	vp->v_trace = ktrace_alloc(VNODE_TRACE_SIZE, KM_SLEEP); | 
 | #endif	/* XFS_VNODE_TRACE */ | 
 |  | 
 | 	vn_trace_exit(vp, __FUNCTION__, (inst_t *)__return_address); | 
 | 	return vp; | 
 | } | 
 |  | 
 | /* | 
 |  * Revalidate the Linux inode from the vattr. | 
 |  * Note: i_size _not_ updated; we must hold the inode | 
 |  * semaphore when doing that - callers responsibility. | 
 |  */ | 
 | void | 
 | vn_revalidate_core( | 
 | 	bhv_vnode_t	*vp, | 
 | 	bhv_vattr_t	*vap) | 
 | { | 
 | 	struct inode	*inode = vn_to_inode(vp); | 
 |  | 
 | 	inode->i_mode	    = vap->va_mode; | 
 | 	inode->i_nlink	    = vap->va_nlink; | 
 | 	inode->i_uid	    = vap->va_uid; | 
 | 	inode->i_gid	    = vap->va_gid; | 
 | 	inode->i_blocks	    = vap->va_nblocks; | 
 | 	inode->i_mtime	    = vap->va_mtime; | 
 | 	inode->i_ctime	    = vap->va_ctime; | 
 | 	if (vap->va_xflags & XFS_XFLAG_IMMUTABLE) | 
 | 		inode->i_flags |= S_IMMUTABLE; | 
 | 	else | 
 | 		inode->i_flags &= ~S_IMMUTABLE; | 
 | 	if (vap->va_xflags & XFS_XFLAG_APPEND) | 
 | 		inode->i_flags |= S_APPEND; | 
 | 	else | 
 | 		inode->i_flags &= ~S_APPEND; | 
 | 	if (vap->va_xflags & XFS_XFLAG_SYNC) | 
 | 		inode->i_flags |= S_SYNC; | 
 | 	else | 
 | 		inode->i_flags &= ~S_SYNC; | 
 | 	if (vap->va_xflags & XFS_XFLAG_NOATIME) | 
 | 		inode->i_flags |= S_NOATIME; | 
 | 	else | 
 | 		inode->i_flags &= ~S_NOATIME; | 
 | } | 
 |  | 
 | /* | 
 |  * Revalidate the Linux inode from the vnode. | 
 |  */ | 
 | int | 
 | __vn_revalidate( | 
 | 	bhv_vnode_t	*vp, | 
 | 	bhv_vattr_t	*vattr) | 
 | { | 
 | 	int		error; | 
 |  | 
 | 	vn_trace_entry(vp, __FUNCTION__, (inst_t *)__return_address); | 
 | 	vattr->va_mask = XFS_AT_STAT | XFS_AT_XFLAGS; | 
 | 	error = bhv_vop_getattr(vp, vattr, 0, NULL); | 
 | 	if (likely(!error)) { | 
 | 		vn_revalidate_core(vp, vattr); | 
 | 		VUNMODIFY(vp); | 
 | 	} | 
 | 	return -error; | 
 | } | 
 |  | 
 | int | 
 | vn_revalidate( | 
 | 	bhv_vnode_t	*vp) | 
 | { | 
 | 	bhv_vattr_t	vattr; | 
 |  | 
 | 	return __vn_revalidate(vp, &vattr); | 
 | } | 
 |  | 
 | /* | 
 |  * Add a reference to a referenced vnode. | 
 |  */ | 
 | bhv_vnode_t * | 
 | vn_hold( | 
 | 	bhv_vnode_t	*vp) | 
 | { | 
 | 	struct inode	*inode; | 
 |  | 
 | 	XFS_STATS_INC(vn_hold); | 
 |  | 
 | 	VN_LOCK(vp); | 
 | 	inode = igrab(vn_to_inode(vp)); | 
 | 	ASSERT(inode); | 
 | 	VN_UNLOCK(vp, 0); | 
 |  | 
 | 	return vp; | 
 | } | 
 |  | 
 | #ifdef	XFS_VNODE_TRACE | 
 |  | 
 | #define KTRACE_ENTER(vp, vk, s, line, ra)			\ | 
 | 	ktrace_enter(	(vp)->v_trace,				\ | 
 | /*  0 */		(void *)(__psint_t)(vk),		\ | 
 | /*  1 */		(void *)(s),				\ | 
 | /*  2 */		(void *)(__psint_t) line,		\ | 
 | /*  3 */		(void *)(__psint_t)(vn_count(vp)),	\ | 
 | /*  4 */		(void *)(ra),				\ | 
 | /*  5 */		(void *)(__psunsigned_t)(vp)->v_flag,	\ | 
 | /*  6 */		(void *)(__psint_t)current_cpu(),	\ | 
 | /*  7 */		(void *)(__psint_t)current_pid(),	\ | 
 | /*  8 */		(void *)__return_address,		\ | 
 | /*  9 */		NULL, NULL, NULL, NULL, NULL, NULL, NULL) | 
 |  | 
 | /* | 
 |  * Vnode tracing code. | 
 |  */ | 
 | void | 
 | vn_trace_entry(bhv_vnode_t *vp, const char *func, inst_t *ra) | 
 | { | 
 | 	KTRACE_ENTER(vp, VNODE_KTRACE_ENTRY, func, 0, ra); | 
 | } | 
 |  | 
 | void | 
 | vn_trace_exit(bhv_vnode_t *vp, const char *func, inst_t *ra) | 
 | { | 
 | 	KTRACE_ENTER(vp, VNODE_KTRACE_EXIT, func, 0, ra); | 
 | } | 
 |  | 
 | void | 
 | vn_trace_hold(bhv_vnode_t *vp, char *file, int line, inst_t *ra) | 
 | { | 
 | 	KTRACE_ENTER(vp, VNODE_KTRACE_HOLD, file, line, ra); | 
 | } | 
 |  | 
 | void | 
 | vn_trace_ref(bhv_vnode_t *vp, char *file, int line, inst_t *ra) | 
 | { | 
 | 	KTRACE_ENTER(vp, VNODE_KTRACE_REF, file, line, ra); | 
 | } | 
 |  | 
 | void | 
 | vn_trace_rele(bhv_vnode_t *vp, char *file, int line, inst_t *ra) | 
 | { | 
 | 	KTRACE_ENTER(vp, VNODE_KTRACE_RELE, file, line, ra); | 
 | } | 
 | #endif	/* XFS_VNODE_TRACE */ |