|  | /* | 
|  | *  net/dccp/ccid.c | 
|  | * | 
|  | *  An implementation of the DCCP protocol | 
|  | *  Arnaldo Carvalho de Melo <acme@conectiva.com.br> | 
|  | * | 
|  | *  CCID infrastructure | 
|  | * | 
|  | *	This program is free software; you can redistribute it and/or modify it | 
|  | *	under the terms of the GNU General Public License version 2 as | 
|  | *	published by the Free Software Foundation. | 
|  | */ | 
|  |  | 
|  | #include "ccid.h" | 
|  |  | 
|  | static struct ccid_operations *ccids[CCID_MAX]; | 
|  | #if defined(CONFIG_SMP) || defined(CONFIG_PREEMPT) | 
|  | static atomic_t ccids_lockct = ATOMIC_INIT(0); | 
|  | static DEFINE_SPINLOCK(ccids_lock); | 
|  |  | 
|  | /* | 
|  | * The strategy is: modifications ccids vector are short, do not sleep and | 
|  | * veeery rare, but read access should be free of any exclusive locks. | 
|  | */ | 
|  | static void ccids_write_lock(void) | 
|  | { | 
|  | spin_lock(&ccids_lock); | 
|  | while (atomic_read(&ccids_lockct) != 0) { | 
|  | spin_unlock(&ccids_lock); | 
|  | yield(); | 
|  | spin_lock(&ccids_lock); | 
|  | } | 
|  | } | 
|  |  | 
|  | static inline void ccids_write_unlock(void) | 
|  | { | 
|  | spin_unlock(&ccids_lock); | 
|  | } | 
|  |  | 
|  | static inline void ccids_read_lock(void) | 
|  | { | 
|  | atomic_inc(&ccids_lockct); | 
|  | smp_mb__after_atomic_inc(); | 
|  | spin_unlock_wait(&ccids_lock); | 
|  | } | 
|  |  | 
|  | static inline void ccids_read_unlock(void) | 
|  | { | 
|  | atomic_dec(&ccids_lockct); | 
|  | } | 
|  |  | 
|  | #else | 
|  | #define ccids_write_lock() do { } while(0) | 
|  | #define ccids_write_unlock() do { } while(0) | 
|  | #define ccids_read_lock() do { } while(0) | 
|  | #define ccids_read_unlock() do { } while(0) | 
|  | #endif | 
|  |  | 
|  | static struct kmem_cache *ccid_kmem_cache_create(int obj_size, const char *fmt,...) | 
|  | { | 
|  | struct kmem_cache *slab; | 
|  | char slab_name_fmt[32], *slab_name; | 
|  | va_list args; | 
|  |  | 
|  | va_start(args, fmt); | 
|  | vsnprintf(slab_name_fmt, sizeof(slab_name_fmt), fmt, args); | 
|  | va_end(args); | 
|  |  | 
|  | slab_name = kstrdup(slab_name_fmt, GFP_KERNEL); | 
|  | if (slab_name == NULL) | 
|  | return NULL; | 
|  | slab = kmem_cache_create(slab_name, sizeof(struct ccid) + obj_size, 0, | 
|  | SLAB_HWCACHE_ALIGN, NULL); | 
|  | if (slab == NULL) | 
|  | kfree(slab_name); | 
|  | return slab; | 
|  | } | 
|  |  | 
|  | static void ccid_kmem_cache_destroy(struct kmem_cache *slab) | 
|  | { | 
|  | if (slab != NULL) { | 
|  | const char *name = kmem_cache_name(slab); | 
|  |  | 
|  | kmem_cache_destroy(slab); | 
|  | kfree(name); | 
|  | } | 
|  | } | 
|  |  | 
|  | int ccid_register(struct ccid_operations *ccid_ops) | 
|  | { | 
|  | int err = -ENOBUFS; | 
|  |  | 
|  | ccid_ops->ccid_hc_rx_slab = | 
|  | ccid_kmem_cache_create(ccid_ops->ccid_hc_rx_obj_size, | 
|  | "%s_hc_rx_sock", | 
|  | ccid_ops->ccid_name); | 
|  | if (ccid_ops->ccid_hc_rx_slab == NULL) | 
|  | goto out; | 
|  |  | 
|  | ccid_ops->ccid_hc_tx_slab = | 
|  | ccid_kmem_cache_create(ccid_ops->ccid_hc_tx_obj_size, | 
|  | "%s_hc_tx_sock", | 
|  | ccid_ops->ccid_name); | 
|  | if (ccid_ops->ccid_hc_tx_slab == NULL) | 
|  | goto out_free_rx_slab; | 
|  |  | 
|  | ccids_write_lock(); | 
|  | err = -EEXIST; | 
|  | if (ccids[ccid_ops->ccid_id] == NULL) { | 
|  | ccids[ccid_ops->ccid_id] = ccid_ops; | 
|  | err = 0; | 
|  | } | 
|  | ccids_write_unlock(); | 
|  | if (err != 0) | 
|  | goto out_free_tx_slab; | 
|  |  | 
|  | pr_info("CCID: Registered CCID %d (%s)\n", | 
|  | ccid_ops->ccid_id, ccid_ops->ccid_name); | 
|  | out: | 
|  | return err; | 
|  | out_free_tx_slab: | 
|  | ccid_kmem_cache_destroy(ccid_ops->ccid_hc_tx_slab); | 
|  | ccid_ops->ccid_hc_tx_slab = NULL; | 
|  | goto out; | 
|  | out_free_rx_slab: | 
|  | ccid_kmem_cache_destroy(ccid_ops->ccid_hc_rx_slab); | 
|  | ccid_ops->ccid_hc_rx_slab = NULL; | 
|  | goto out; | 
|  | } | 
|  |  | 
|  | EXPORT_SYMBOL_GPL(ccid_register); | 
|  |  | 
|  | int ccid_unregister(struct ccid_operations *ccid_ops) | 
|  | { | 
|  | ccids_write_lock(); | 
|  | ccids[ccid_ops->ccid_id] = NULL; | 
|  | ccids_write_unlock(); | 
|  |  | 
|  | ccid_kmem_cache_destroy(ccid_ops->ccid_hc_tx_slab); | 
|  | ccid_ops->ccid_hc_tx_slab = NULL; | 
|  | ccid_kmem_cache_destroy(ccid_ops->ccid_hc_rx_slab); | 
|  | ccid_ops->ccid_hc_rx_slab = NULL; | 
|  |  | 
|  | pr_info("CCID: Unregistered CCID %d (%s)\n", | 
|  | ccid_ops->ccid_id, ccid_ops->ccid_name); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | EXPORT_SYMBOL_GPL(ccid_unregister); | 
|  |  | 
|  | struct ccid *ccid_new(unsigned char id, struct sock *sk, int rx, gfp_t gfp) | 
|  | { | 
|  | struct ccid_operations *ccid_ops; | 
|  | struct ccid *ccid = NULL; | 
|  |  | 
|  | ccids_read_lock(); | 
|  | #ifdef CONFIG_KMOD | 
|  | if (ccids[id] == NULL) { | 
|  | /* We only try to load if in process context */ | 
|  | ccids_read_unlock(); | 
|  | if (gfp & GFP_ATOMIC) | 
|  | goto out; | 
|  | request_module("net-dccp-ccid-%d", id); | 
|  | ccids_read_lock(); | 
|  | } | 
|  | #endif | 
|  | ccid_ops = ccids[id]; | 
|  | if (ccid_ops == NULL) | 
|  | goto out_unlock; | 
|  |  | 
|  | if (!try_module_get(ccid_ops->ccid_owner)) | 
|  | goto out_unlock; | 
|  |  | 
|  | ccids_read_unlock(); | 
|  |  | 
|  | ccid = kmem_cache_alloc(rx ? ccid_ops->ccid_hc_rx_slab : | 
|  | ccid_ops->ccid_hc_tx_slab, gfp); | 
|  | if (ccid == NULL) | 
|  | goto out_module_put; | 
|  | ccid->ccid_ops = ccid_ops; | 
|  | if (rx) { | 
|  | memset(ccid + 1, 0, ccid_ops->ccid_hc_rx_obj_size); | 
|  | if (ccid->ccid_ops->ccid_hc_rx_init != NULL && | 
|  | ccid->ccid_ops->ccid_hc_rx_init(ccid, sk) != 0) | 
|  | goto out_free_ccid; | 
|  | } else { | 
|  | memset(ccid + 1, 0, ccid_ops->ccid_hc_tx_obj_size); | 
|  | if (ccid->ccid_ops->ccid_hc_tx_init != NULL && | 
|  | ccid->ccid_ops->ccid_hc_tx_init(ccid, sk) != 0) | 
|  | goto out_free_ccid; | 
|  | } | 
|  | out: | 
|  | return ccid; | 
|  | out_unlock: | 
|  | ccids_read_unlock(); | 
|  | goto out; | 
|  | out_free_ccid: | 
|  | kmem_cache_free(rx ? ccid_ops->ccid_hc_rx_slab : | 
|  | ccid_ops->ccid_hc_tx_slab, ccid); | 
|  | ccid = NULL; | 
|  | out_module_put: | 
|  | module_put(ccid_ops->ccid_owner); | 
|  | goto out; | 
|  | } | 
|  |  | 
|  | EXPORT_SYMBOL_GPL(ccid_new); | 
|  |  | 
|  | struct ccid *ccid_hc_rx_new(unsigned char id, struct sock *sk, gfp_t gfp) | 
|  | { | 
|  | return ccid_new(id, sk, 1, gfp); | 
|  | } | 
|  |  | 
|  | EXPORT_SYMBOL_GPL(ccid_hc_rx_new); | 
|  |  | 
|  | struct ccid *ccid_hc_tx_new(unsigned char id,struct sock *sk,  gfp_t gfp) | 
|  | { | 
|  | return ccid_new(id, sk, 0, gfp); | 
|  | } | 
|  |  | 
|  | EXPORT_SYMBOL_GPL(ccid_hc_tx_new); | 
|  |  | 
|  | static void ccid_delete(struct ccid *ccid, struct sock *sk, int rx) | 
|  | { | 
|  | struct ccid_operations *ccid_ops; | 
|  |  | 
|  | if (ccid == NULL) | 
|  | return; | 
|  |  | 
|  | ccid_ops = ccid->ccid_ops; | 
|  | if (rx) { | 
|  | if (ccid_ops->ccid_hc_rx_exit != NULL) | 
|  | ccid_ops->ccid_hc_rx_exit(sk); | 
|  | kmem_cache_free(ccid_ops->ccid_hc_rx_slab,  ccid); | 
|  | } else { | 
|  | if (ccid_ops->ccid_hc_tx_exit != NULL) | 
|  | ccid_ops->ccid_hc_tx_exit(sk); | 
|  | kmem_cache_free(ccid_ops->ccid_hc_tx_slab,  ccid); | 
|  | } | 
|  | ccids_read_lock(); | 
|  | if (ccids[ccid_ops->ccid_id] != NULL) | 
|  | module_put(ccid_ops->ccid_owner); | 
|  | ccids_read_unlock(); | 
|  | } | 
|  |  | 
|  | void ccid_hc_rx_delete(struct ccid *ccid, struct sock *sk) | 
|  | { | 
|  | ccid_delete(ccid, sk, 1); | 
|  | } | 
|  |  | 
|  | EXPORT_SYMBOL_GPL(ccid_hc_rx_delete); | 
|  |  | 
|  | void ccid_hc_tx_delete(struct ccid *ccid, struct sock *sk) | 
|  | { | 
|  | ccid_delete(ccid, sk, 0); | 
|  | } | 
|  |  | 
|  | EXPORT_SYMBOL_GPL(ccid_hc_tx_delete); |