/* * net/dccp/ccid.c * * An implementation of the DCCP protocol * Arnaldo Carvalho de Melo * * CCID infrastructure * * This program is free software; you can redistribute it and/or modify it * under the terms of the GNU General Public License version 2 as * published by the Free Software Foundation. */ #include "ccid.h" static struct ccid_operations *ccids[CCID_MAX]; #if defined(CONFIG_SMP) || defined(CONFIG_PREEMPT) static atomic_t ccids_lockct = ATOMIC_INIT(0); static DEFINE_SPINLOCK(ccids_lock); /* * The strategy is: modifications ccids vector are short, do not sleep and * veeery rare, but read access should be free of any exclusive locks. */ static void ccids_write_lock(void) { spin_lock(&ccids_lock); while (atomic_read(&ccids_lockct) != 0) { spin_unlock(&ccids_lock); yield(); spin_lock(&ccids_lock); } } static inline void ccids_write_unlock(void) { spin_unlock(&ccids_lock); } static inline void ccids_read_lock(void) { atomic_inc(&ccids_lockct); smp_mb__after_atomic_inc(); spin_unlock_wait(&ccids_lock); } static inline void ccids_read_unlock(void) { atomic_dec(&ccids_lockct); } #else #define ccids_write_lock() do { } while(0) #define ccids_write_unlock() do { } while(0) #define ccids_read_lock() do { } while(0) #define ccids_read_unlock() do { } while(0) #endif static struct kmem_cache *ccid_kmem_cache_create(int obj_size, const char *fmt,...) { struct kmem_cache *slab; char slab_name_fmt[32], *slab_name; va_list args; va_start(args, fmt); vsnprintf(slab_name_fmt, sizeof(slab_name_fmt), fmt, args); va_end(args); slab_name = kstrdup(slab_name_fmt, GFP_KERNEL); if (slab_name == NULL) return NULL; slab = kmem_cache_create(slab_name, sizeof(struct ccid) + obj_size, 0, SLAB_HWCACHE_ALIGN, NULL); if (slab == NULL) kfree(slab_name); return slab; } static void ccid_kmem_cache_destroy(struct kmem_cache *slab) { if (slab != NULL) { const char *name = kmem_cache_name(slab); kmem_cache_destroy(slab); kfree(name); } } int ccid_register(struct ccid_operations *ccid_ops) { int err = -ENOBUFS; ccid_ops->ccid_hc_rx_slab = ccid_kmem_cache_create(ccid_ops->ccid_hc_rx_obj_size, "ccid%u_hc_rx_sock", ccid_ops->ccid_id); if (ccid_ops->ccid_hc_rx_slab == NULL) goto out; ccid_ops->ccid_hc_tx_slab = ccid_kmem_cache_create(ccid_ops->ccid_hc_tx_obj_size, "ccid%u_hc_tx_sock", ccid_ops->ccid_id); if (ccid_ops->ccid_hc_tx_slab == NULL) goto out_free_rx_slab; ccids_write_lock(); err = -EEXIST; if (ccids[ccid_ops->ccid_id] == NULL) { ccids[ccid_ops->ccid_id] = ccid_ops; err = 0; } ccids_write_unlock(); if (err != 0) goto out_free_tx_slab; pr_info("CCID: Registered CCID %d (%s)\n", ccid_ops->ccid_id, ccid_ops->ccid_name); out: return err; out_free_tx_slab: ccid_kmem_cache_destroy(ccid_ops->ccid_hc_tx_slab); ccid_ops->ccid_hc_tx_slab = NULL; goto out; out_free_rx_slab: ccid_kmem_cache_destroy(ccid_ops->ccid_hc_rx_slab); ccid_ops->ccid_hc_rx_slab = NULL; goto out; } EXPORT_SYMBOL_GPL(ccid_register); int ccid_unregister(struct ccid_operations *ccid_ops) { ccids_write_lock(); ccids[ccid_ops->ccid_id] = NULL; ccids_write_unlock(); ccid_kmem_cache_destroy(ccid_ops->ccid_hc_tx_slab); ccid_ops->ccid_hc_tx_slab = NULL; ccid_kmem_cache_destroy(ccid_ops->ccid_hc_rx_slab); ccid_ops->ccid_hc_rx_slab = NULL; pr_info("CCID: Unregistered CCID %d (%s)\n", ccid_ops->ccid_id, ccid_ops->ccid_name); return 0; } EXPORT_SYMBOL_GPL(ccid_unregister); struct ccid *ccid_new(unsigned char id, struct sock *sk, int rx, gfp_t gfp) { struct ccid_operations *ccid_ops; struct ccid *ccid = NULL; ccids_read_lock(); #ifdef CONFIG_KMOD if (ccids[id] == NULL) { /* We only try to load if in process context */ ccids_read_unlock(); if (gfp & GFP_ATOMIC) goto out; request_module("net-dccp-ccid-%d", id); ccids_read_lock(); } #endif ccid_ops = ccids[id]; if (ccid_ops == NULL) goto out_unlock; if (!try_module_get(ccid_ops->ccid_owner)) goto out_unlock; ccids_read_unlock(); ccid = kmem_cache_alloc(rx ? ccid_ops->ccid_hc_rx_slab : ccid_ops->ccid_hc_tx_slab, gfp); if (ccid == NULL) goto out_module_put; ccid->ccid_ops = ccid_ops; if (rx) { memset(ccid + 1, 0, ccid_ops->ccid_hc_rx_obj_size); if (ccid->ccid_ops->ccid_hc_rx_init != NULL && ccid->ccid_ops->ccid_hc_rx_init(ccid, sk) != 0) goto out_free_ccid; } else { memset(ccid + 1, 0, ccid_ops->ccid_hc_tx_obj_size); if (ccid->ccid_ops->ccid_hc_tx_init != NULL && ccid->ccid_ops->ccid_hc_tx_init(ccid, sk) != 0) goto out_free_ccid; } out: return ccid; out_unlock: ccids_read_unlock(); goto out; out_free_ccid: kmem_cache_free(rx ? ccid_ops->ccid_hc_rx_slab : ccid_ops->ccid_hc_tx_slab, ccid); ccid = NULL; out_module_put: module_put(ccid_ops->ccid_owner); goto out; } EXPORT_SYMBOL_GPL(ccid_new); struct ccid *ccid_hc_rx_new(unsigned char id, struct sock *sk, gfp_t gfp) { return ccid_new(id, sk, 1, gfp); } EXPORT_SYMBOL_GPL(ccid_hc_rx_new); struct ccid *ccid_hc_tx_new(unsigned char id,struct sock *sk, gfp_t gfp) { return ccid_new(id, sk, 0, gfp); } EXPORT_SYMBOL_GPL(ccid_hc_tx_new); static void ccid_delete(struct ccid *ccid, struct sock *sk, int rx) { struct ccid_operations *ccid_ops; if (ccid == NULL) return; ccid_ops = ccid->ccid_ops; if (rx) { if (ccid_ops->ccid_hc_rx_exit != NULL) ccid_ops->ccid_hc_rx_exit(sk); kmem_cache_free(ccid_ops->ccid_hc_rx_slab, ccid); } else { if (ccid_ops->ccid_hc_tx_exit != NULL) ccid_ops->ccid_hc_tx_exit(sk); kmem_cache_free(ccid_ops->ccid_hc_tx_slab, ccid); } ccids_read_lock(); if (ccids[ccid_ops->ccid_id] != NULL) module_put(ccid_ops->ccid_owner); ccids_read_unlock(); } void ccid_hc_rx_delete(struct ccid *ccid, struct sock *sk) { ccid_delete(ccid, sk, 1); } EXPORT_SYMBOL_GPL(ccid_hc_rx_delete); void ccid_hc_tx_delete(struct ccid *ccid, struct sock *sk) { ccid_delete(ccid, sk, 0); } EXPORT_SYMBOL_GPL(ccid_hc_tx_delete);