summaryrefslogtreecommitdiffstats
path: root/fs/nfs
diff options
context:
space:
mode:
authorTrond Myklebust <Trond.Myklebust@netapp.com>2009-12-05 19:32:19 -0500
committerTrond Myklebust <Trond.Myklebust@netapp.com>2009-12-05 19:32:19 -0500
commit35dc1d74a8d97a302a202ccb6751bf2bdbf5173e (patch)
tree0791d69bc97fac30b9f52e7f96479e31386bb6f8 /fs/nfs
parentd61e612a728fb9bf848c4383f8f6645e822d5b57 (diff)
downloadlinux-stable-35dc1d74a8d97a302a202ccb6751bf2bdbf5173e.tar.gz
linux-stable-35dc1d74a8d97a302a202ccb6751bf2bdbf5173e.tar.bz2
linux-stable-35dc1d74a8d97a302a202ccb6751bf2bdbf5173e.zip
NFSv41: Fix up some bugs in the NFS4CLNT_SESSION_DRAINING code
Signed-off-by: Trond Myklebust <Trond.Myklebust@netapp.com>
Diffstat (limited to 'fs/nfs')
-rw-r--r--fs/nfs/nfs4proc.c17
-rw-r--r--fs/nfs/nfs4state.c4
2 files changed, 10 insertions, 11 deletions
diff --git a/fs/nfs/nfs4proc.c b/fs/nfs/nfs4proc.c
index c06a2bade59e..9da7a872ee0e 100644
--- a/fs/nfs/nfs4proc.c
+++ b/fs/nfs/nfs4proc.c
@@ -318,13 +318,14 @@ static void renew_lease(const struct nfs_server *server, unsigned long timestamp
* so we need to scan down from highest_used_slotid to 0 looking for the now
* highest slotid in use.
* If none found, highest_used_slotid is set to -1.
+ *
+ * Must be called while holding tbl->slot_tbl_lock
*/
static void
nfs4_free_slot(struct nfs4_slot_table *tbl, u8 free_slotid)
{
int slotid = free_slotid;
- spin_lock(&tbl->slot_tbl_lock);
/* clear used bit in bitmap */
__clear_bit(slotid, tbl->used_slots);
@@ -336,7 +337,6 @@ nfs4_free_slot(struct nfs4_slot_table *tbl, u8 free_slotid)
else
tbl->highest_used_slotid = -1;
}
- spin_unlock(&tbl->slot_tbl_lock);
dprintk("%s: free_slotid %u highest_used_slotid %d\n", __func__,
free_slotid, tbl->highest_used_slotid);
}
@@ -351,22 +351,23 @@ static void nfs41_sequence_free_slot(const struct nfs_client *clp,
/* just wake up the next guy waiting since
* we may have not consumed a slot after all */
dprintk("%s: No slot\n", __func__);
- } else {
- nfs4_free_slot(tbl, res->sr_slotid);
- res->sr_slotid = NFS4_MAX_SLOT_TABLE;
+ return;
}
+ spin_lock(&tbl->slot_tbl_lock);
+ nfs4_free_slot(tbl, res->sr_slotid);
+
/* Signal state manager thread if session is drained */
if (test_bit(NFS4CLNT_SESSION_DRAINING, &clp->cl_state)) {
- spin_lock(&tbl->slot_tbl_lock);
if (tbl->highest_used_slotid == -1) {
dprintk("%s COMPLETE: Session Drained\n", __func__);
complete(&clp->cl_session->complete);
}
- spin_unlock(&tbl->slot_tbl_lock);
} else {
rpc_wake_up_next(&tbl->slot_tbl_waitq);
}
+ spin_unlock(&tbl->slot_tbl_lock);
+ res->sr_slotid = NFS4_MAX_SLOT_TABLE;
}
static void nfs41_sequence_done(struct nfs_client *clp,
@@ -470,7 +471,6 @@ static int nfs41_setup_sequence(struct nfs4_session *session,
*/
dprintk("%s Schedule Session Reset\n", __func__);
rpc_sleep_on(&tbl->slot_tbl_waitq, task, NULL);
- nfs4_schedule_state_manager(session->clp);
spin_unlock(&tbl->slot_tbl_lock);
return -EAGAIN;
}
@@ -4489,7 +4489,6 @@ static int nfs4_reset_slot_tables(struct nfs4_session *session)
1);
if (status)
return status;
- init_completion(&session->complete);
status = nfs4_reset_slot_table(&session->bc_slot_table,
session->bc_attrs.max_reqs,
diff --git a/fs/nfs/nfs4state.c b/fs/nfs/nfs4state.c
index 37f020eb92f2..ef9622e500e4 100644
--- a/fs/nfs/nfs4state.c
+++ b/fs/nfs/nfs4state.c
@@ -1220,10 +1220,10 @@ static int nfs4_reset_session(struct nfs_client *clp)
struct nfs4_slot_table *tbl = &ses->fc_slot_table;
int status;
- INIT_COMPLETION(ses->complete);
spin_lock(&tbl->slot_tbl_lock);
set_bit(NFS4CLNT_SESSION_DRAINING, &clp->cl_state);
if (tbl->highest_used_slotid != -1) {
+ INIT_COMPLETION(ses->complete);
spin_unlock(&tbl->slot_tbl_lock);
status = wait_for_completion_interruptible(&ses->complete);
if (status) /* -ERESTARTSYS */
@@ -1247,7 +1247,7 @@ static int nfs4_reset_session(struct nfs_client *clp)
out:
/* Wake up the next rpc task even on error */
clear_bit(NFS4CLNT_SESSION_DRAINING, &clp->cl_state);
- rpc_wake_up_next(&clp->cl_session->fc_slot_table.slot_tbl_waitq);
+ rpc_wake_up(&clp->cl_session->fc_slot_table.slot_tbl_waitq);
return status;
}