summaryrefslogtreecommitdiffstats
path: root/net/tipc/ref.c
diff options
context:
space:
mode:
authorJon Paul Maloy <jon.maloy@ericsson.com>2014-08-22 18:09:16 -0400
committerDavid S. Miller <davem@davemloft.net>2014-08-23 11:18:34 -0700
commit6c9808ce09f778a1de7b207b82cfc36a59cda2d3 (patch)
treedce467a4bd4be90896e7243f87335e1b41cded75 /net/tipc/ref.c
parent9b50fd087a9f1454d6a8b613fff376dfb6d6ea93 (diff)
downloadlinux-6c9808ce09f778a1de7b207b82cfc36a59cda2d3.tar.gz
linux-6c9808ce09f778a1de7b207b82cfc36a59cda2d3.tar.bz2
linux-6c9808ce09f778a1de7b207b82cfc36a59cda2d3.zip
tipc: remove port_lock
In previous commits we have reduced usage of port_lock to a minimum, and complemented it with usage of bh_lock_sock() at the remaining locations. The purpose has been to remove this lock altogether, since it largely duplicates the role of bh_lock_sock. We are now ready to do this. However, we still need to protect the BH callers from inadvertent release of the socket while they hold a reference to it. We do this by replacing port_lock by a combination of a rw-lock protecting the reference table as such, and updating the socket reference counter while the socket is referenced from BH. This technique is more standard and comprehensible than the previous approach, and turns out to have a positive effect on overall performance. Signed-off-by: Jon Maloy <jon.maloy@ericsson.com> Reviewed-by: Erik Hugne <erik.hugne@ericsson.com> Reviewed-by: Ying Xue <ying.xue@windriver.com> Signed-off-by: David S. Miller <davem@davemloft.net>
Diffstat (limited to 'net/tipc/ref.c')
-rw-r--r--net/tipc/ref.c158
1 files changed, 71 insertions, 87 deletions
diff --git a/net/tipc/ref.c b/net/tipc/ref.c
index 7fc2740846e3..ea981bed967b 100644
--- a/net/tipc/ref.c
+++ b/net/tipc/ref.c
@@ -1,7 +1,7 @@
/*
- * net/tipc/ref.c: TIPC object registry code
+ * net/tipc/ref.c: TIPC socket registry code
*
- * Copyright (c) 1991-2006, Ericsson AB
+ * Copyright (c) 1991-2006, 2014, Ericsson AB
* Copyright (c) 2004-2007, Wind River Systems
* All rights reserved.
*
@@ -38,24 +38,22 @@
#include "ref.h"
/**
- * struct reference - TIPC object reference entry
- * @object: pointer to object associated with reference entry
- * @lock: spinlock controlling access to object
- * @ref: reference value for object (combines instance & array index info)
+ * struct reference - TIPC socket reference entry
+ * @tsk: pointer to socket associated with reference entry
+ * @ref: reference value for socket (combines instance & array index info)
*/
struct reference {
- void *object;
- spinlock_t lock;
+ struct tipc_sock *tsk;
u32 ref;
};
/**
- * struct tipc_ref_table - table of TIPC object reference entries
+ * struct tipc_ref_table - table of TIPC socket reference entries
* @entries: pointer to array of reference entries
* @capacity: array index of first unusable entry
* @init_point: array index of first uninitialized entry
- * @first_free: array index of first unused object reference entry
- * @last_free: array index of last unused object reference entry
+ * @first_free: array index of first unused socket reference entry
+ * @last_free: array index of last unused socket reference entry
* @index_mask: bitmask for array index portion of reference values
* @start_mask: initial value for instance value portion of reference values
*/
@@ -70,9 +68,9 @@ struct ref_table {
};
/*
- * Object reference table consists of 2**N entries.
+ * Socket reference table consists of 2**N entries.
*
- * State Object ptr Reference
+ * State Socket ptr Reference
* ----- ---------- ---------
* In use non-NULL XXXX|own index
* (XXXX changes each time entry is acquired)
@@ -89,10 +87,10 @@ struct ref_table {
static struct ref_table tipc_ref_table;
-static DEFINE_SPINLOCK(ref_table_lock);
+static DEFINE_RWLOCK(ref_table_lock);
/**
- * tipc_ref_table_init - create reference table for objects
+ * tipc_ref_table_init - create reference table for sockets
*/
int tipc_ref_table_init(u32 requested_size, u32 start)
{
@@ -122,84 +120,69 @@ int tipc_ref_table_init(u32 requested_size, u32 start)
}
/**
- * tipc_ref_table_stop - destroy reference table for objects
+ * tipc_ref_table_stop - destroy reference table for sockets
*/
void tipc_ref_table_stop(void)
{
+ if (!tipc_ref_table.entries)
+ return;
vfree(tipc_ref_table.entries);
tipc_ref_table.entries = NULL;
}
-/**
- * tipc_ref_acquire - create reference to an object
+/* tipc_ref_acquire - create reference to a socket
*
- * Register an object pointer in reference table and lock the object.
+ * Register an socket pointer in the reference table.
* Returns a unique reference value that is used from then on to retrieve the
- * object pointer, or to determine that the object has been deregistered.
- *
- * Note: The object is returned in the locked state so that the caller can
- * register a partially initialized object, without running the risk that
- * the object will be accessed before initialization is complete.
+ * socket pointer, or to determine if the socket has been deregistered.
*/
-u32 tipc_ref_acquire(void *object, spinlock_t **lock)
+u32 tipc_ref_acquire(struct tipc_sock *tsk)
{
u32 index;
u32 index_mask;
u32 next_plus_upper;
- u32 ref;
- struct reference *entry = NULL;
+ u32 ref = 0;
+ struct reference *entry;
- if (!object) {
+ if (unlikely(!tsk)) {
pr_err("Attempt to acquire ref. to non-existent obj\n");
return 0;
}
- if (!tipc_ref_table.entries) {
+ if (unlikely(!tipc_ref_table.entries)) {
pr_err("Ref. table not found in acquisition attempt\n");
return 0;
}
- /* take a free entry, if available; otherwise initialize a new entry */
- spin_lock_bh(&ref_table_lock);
- if (tipc_ref_table.first_free) {
+ /* Take a free entry, if available; otherwise initialize a new one */
+ write_lock_bh(&ref_table_lock);
+ index = tipc_ref_table.first_free;
+ entry = &tipc_ref_table.entries[index];
+
+ if (likely(index)) {
index = tipc_ref_table.first_free;
entry = &(tipc_ref_table.entries[index]);
index_mask = tipc_ref_table.index_mask;
next_plus_upper = entry->ref;
tipc_ref_table.first_free = next_plus_upper & index_mask;
ref = (next_plus_upper & ~index_mask) + index;
+ entry->tsk = tsk;
} else if (tipc_ref_table.init_point < tipc_ref_table.capacity) {
index = tipc_ref_table.init_point++;
entry = &(tipc_ref_table.entries[index]);
- spin_lock_init(&entry->lock);
ref = tipc_ref_table.start_mask + index;
- } else {
- ref = 0;
}
- spin_unlock_bh(&ref_table_lock);
- /*
- * Grab the lock so no one else can modify this entry
- * While we assign its ref value & object pointer
- */
- if (entry) {
- spin_lock_bh(&entry->lock);
+ if (ref) {
entry->ref = ref;
- entry->object = object;
- *lock = &entry->lock;
- /*
- * keep it locked, the caller is responsible
- * for unlocking this when they're done with it
- */
+ entry->tsk = tsk;
}
-
+ write_unlock_bh(&ref_table_lock);
return ref;
}
-/**
- * tipc_ref_discard - invalidate references to an object
+/* tipc_ref_discard - invalidate reference to an socket
*
- * Disallow future references to an object and free up the entry for re-use.
- * Note: The entry's spin_lock may still be busy after discard
+ * Disallow future references to an socket and free up the entry for re-use.
*/
void tipc_ref_discard(u32 ref)
{
@@ -207,7 +190,7 @@ void tipc_ref_discard(u32 ref)
u32 index;
u32 index_mask;
- if (!tipc_ref_table.entries) {
+ if (unlikely(!tipc_ref_table.entries)) {
pr_err("Ref. table not found during discard attempt\n");
return;
}
@@ -216,71 +199,72 @@ void tipc_ref_discard(u32 ref)
index = ref & index_mask;
entry = &(tipc_ref_table.entries[index]);
- spin_lock_bh(&ref_table_lock);
+ write_lock_bh(&ref_table_lock);
- if (!entry->object) {
- pr_err("Attempt to discard ref. to non-existent obj\n");
+ if (unlikely(!entry->tsk)) {
+ pr_err("Attempt to discard ref. to non-existent socket\n");
goto exit;
}
- if (entry->ref != ref) {
+ if (unlikely(entry->ref != ref)) {
pr_err("Attempt to discard non-existent reference\n");
goto exit;
}
/*
- * mark entry as unused; increment instance part of entry's reference
+ * Mark entry as unused; increment instance part of entry's reference
* to invalidate any subsequent references
*/
- entry->object = NULL;
+ entry->tsk = NULL;
entry->ref = (ref & ~index_mask) + (index_mask + 1);
- /* append entry to free entry list */
- if (tipc_ref_table.first_free == 0)
+ /* Append entry to free entry list */
+ if (unlikely(tipc_ref_table.first_free == 0))
tipc_ref_table.first_free = index;
else
tipc_ref_table.entries[tipc_ref_table.last_free].ref |= index;
tipc_ref_table.last_free = index;
-
exit:
- spin_unlock_bh(&ref_table_lock);
+ write_unlock_bh(&ref_table_lock);
}
-/**
- * tipc_ref_lock - lock referenced object and return pointer to it
+/* tipc_sk_get - find referenced socket and return pointer to it
*/
-void *tipc_ref_lock(u32 ref)
+struct tipc_sock *tipc_sk_get(u32 ref)
{
- if (likely(tipc_ref_table.entries)) {
- struct reference *entry;
+ struct reference *entry;
+ struct tipc_sock *tsk;
- entry = &tipc_ref_table.entries[ref &
- tipc_ref_table.index_mask];
- if (likely(entry->ref != 0)) {
- spin_lock_bh(&entry->lock);
- if (likely((entry->ref == ref) && (entry->object)))
- return entry->object;
- spin_unlock_bh(&entry->lock);
- }
- }
- return NULL;
+ if (unlikely(!tipc_ref_table.entries))
+ return NULL;
+ read_lock_bh(&ref_table_lock);
+ entry = &tipc_ref_table.entries[ref & tipc_ref_table.index_mask];
+ tsk = entry->tsk;
+ if (likely(tsk && (entry->ref == ref)))
+ sock_hold(&tsk->sk);
+ else
+ tsk = NULL;
+ read_unlock_bh(&ref_table_lock);
+ return tsk;
}
-/* tipc_ref_lock_next - lock & return next object after referenced one
+/* tipc_sk_get_next - lock & return next socket after referenced one
*/
-void *tipc_ref_lock_next(u32 *ref)
+struct tipc_sock *tipc_sk_get_next(u32 *ref)
{
struct reference *entry;
+ struct tipc_sock *tsk = NULL;
uint index = *ref & tipc_ref_table.index_mask;
+ read_lock_bh(&ref_table_lock);
while (++index < tipc_ref_table.capacity) {
entry = &tipc_ref_table.entries[index];
- if (!entry->object)
+ if (!entry->tsk)
continue;
- spin_lock_bh(&entry->lock);
+ tsk = entry->tsk;
+ sock_hold(&tsk->sk);
*ref = entry->ref;
- if (entry->object)
- return entry->object;
- spin_unlock_bh(&entry->lock);
+ break;
}
- return NULL;
+ read_unlock_bh(&ref_table_lock);
+ return tsk;
}