From: Al Viro Date: Sat, 13 Feb 2016 15:15:22 +0000 (-0500) Subject: orangefs: get rid of loop in wait_for_matching_downcall() X-Git-Url: https://git.karo-electronics.de/?a=commitdiff_plain;h=d2d87a3b6df3088a991e277d42cd6a549ff2bc66;p=linux-beck.git orangefs: get rid of loop in wait_for_matching_downcall() turn op->waitq into struct completion... Signed-off-by: Al Viro Signed-off-by: Mike Marshall --- diff --git a/fs/orangefs/orangefs-cache.c b/fs/orangefs/orangefs-cache.c index 59ab0c207e90..09194e69875f 100644 --- a/fs/orangefs/orangefs-cache.c +++ b/fs/orangefs/orangefs-cache.c @@ -118,7 +118,7 @@ struct orangefs_kernel_op_s *op_alloc(__s32 type) if (new_op) { INIT_LIST_HEAD(&new_op->list); spin_lock_init(&new_op->lock); - init_waitqueue_head(&new_op->waitq); + init_completion(&new_op->waitq); atomic_set(&new_op->ref_count, 1); diff --git a/fs/orangefs/orangefs-kernel.h b/fs/orangefs/orangefs-kernel.h index 3ceeeaed4143..de898bda7859 100644 --- a/fs/orangefs/orangefs-kernel.h +++ b/fs/orangefs/orangefs-kernel.h @@ -202,7 +202,7 @@ struct orangefs_kernel_op_s { struct orangefs_upcall_s upcall; struct orangefs_downcall_s downcall; - wait_queue_head_t waitq; + struct completion waitq; spinlock_t lock; struct completion done; @@ -222,7 +222,7 @@ struct orangefs_kernel_op_s { static inline void set_op_state_serviced(struct orangefs_kernel_op_s *op) { op->op_state = OP_VFS_STATE_SERVICED; - wake_up_interruptible(&op->waitq); + complete(&op->waitq); } #define op_state_waiting(op) ((op)->op_state & OP_VFS_STATE_WAITING) @@ -266,7 +266,7 @@ static inline void set_op_state_purged(struct orangefs_kernel_op_s *op) put_cancel(op); } else { op->op_state |= OP_VFS_STATE_PURGED; - wake_up_interruptible(&op->waitq); + complete(&op->waitq); spin_unlock(&op->lock); } } diff --git a/fs/orangefs/waitqueue.c b/fs/orangefs/waitqueue.c index 3ea1665efdf0..89622717a06d 100644 --- a/fs/orangefs/waitqueue.c +++ b/fs/orangefs/waitqueue.c @@ -17,6 +17,7 @@ #include "orangefs-bufmap.h" static int wait_for_matching_downcall(struct orangefs_kernel_op_s *); +static void orangefs_clean_up_interrupted_operation(struct orangefs_kernel_op_s *); /* * What we do in this function is to walk the list of operations that are @@ -170,8 +171,10 @@ retry_servicing: gossip_err("orangefs: %s -- wait timed out; aborting attempt.\n", op_name); } + orangefs_clean_up_interrupted_operation(op); op->downcall.status = ret; } else { + spin_unlock(&op->lock); /* got matching downcall; make sure status is in errno format */ op->downcall.status = orangefs_normalize_to_errno(op->downcall.status); @@ -343,6 +346,7 @@ static void orangefs_clean_up_interrupted_operation(struct orangefs_kernel_op_s gossip_err("%s: can't get here.\n", __func__); spin_unlock(&op->lock); } + reinit_completion(&op->waitq); } /* @@ -359,95 +363,52 @@ static void orangefs_clean_up_interrupted_operation(struct orangefs_kernel_op_s * EINTR/EIO/ETIMEDOUT indicating we are done trying to service this * operation since client-core seems to be exiting too often * or if we were interrupted. + * + * Returns with op->lock taken. */ static int wait_for_matching_downcall(struct orangefs_kernel_op_s *op) { - int ret = -EINVAL; - DEFINE_WAIT(wait_entry); + long timeout, n; - while (1) { - spin_lock(&op->lock); - prepare_to_wait(&op->waitq, &wait_entry, TASK_INTERRUPTIBLE); - if (op_state_serviced(op)) { - spin_unlock(&op->lock); - ret = 0; - break; - } + timeout = op->attempts ? op_timeout_secs * HZ : MAX_SCHEDULE_TIMEOUT; + n = wait_for_completion_interruptible_timeout(&op->waitq, timeout); + spin_lock(&op->lock); - if (unlikely(signal_pending(current))) { - gossip_debug(GOSSIP_WAIT_DEBUG, - "*** %s:" - " operation interrupted by a signal (tag " - "%llu, op %p)\n", - __func__, - llu(op->tag), - op); - orangefs_clean_up_interrupted_operation(op); - ret = -EINTR; - break; - } + if (op_state_serviced(op)) + return 0; - /* - * if this was our first attempt and client-core - * has not purged our operation, we are happy to - * simply wait - */ - if (op->attempts == 0 && !op_state_purged(op)) { - spin_unlock(&op->lock); - schedule(); - } else { - spin_unlock(&op->lock); - /* - * subsequent attempts, we retry exactly once - * with timeouts - */ - if (!schedule_timeout(op_timeout_secs * HZ)) { - gossip_debug(GOSSIP_WAIT_DEBUG, - "*** %s:" - " operation timed out (tag" - " %llu, %p, att %d)\n", - __func__, - llu(op->tag), - op, - op->attempts); - ret = -ETIMEDOUT; - spin_lock(&op->lock); - orangefs_clean_up_interrupted_operation(op); - break; - } - } - spin_lock(&op->lock); - op->attempts++; - /* - * if the operation was purged in the meantime, it - * is better to requeue it afresh but ensure that - * we have not been purged repeatedly. This could - * happen if client-core crashes when an op - * is being serviced, so we requeue the op, client - * core crashes again so we requeue the op, client - * core starts, and so on... - */ - if (op_state_purged(op)) { - ret = (op->attempts < ORANGEFS_PURGE_RETRY_COUNT) ? - -EAGAIN : - -EIO; - gossip_debug(GOSSIP_WAIT_DEBUG, - "*** %s:" - " operation purged (tag " - "%llu, %p, att %d)\n", - __func__, - llu(op->tag), - op, - op->attempts); - orangefs_clean_up_interrupted_operation(op); - break; - } - spin_unlock(&op->lock); + if (unlikely(n < 0)) { + gossip_debug(GOSSIP_WAIT_DEBUG, + "*** %s:" + " operation interrupted by a signal (tag " + "%llu, op %p)\n", + __func__, + llu(op->tag), + op); + return -EINTR; } - - spin_lock(&op->lock); - finish_wait(&op->waitq, &wait_entry); - spin_unlock(&op->lock); - - return ret; + op->attempts++; + if (op_state_purged(op)) { + gossip_debug(GOSSIP_WAIT_DEBUG, + "*** %s:" + " operation purged (tag " + "%llu, %p, att %d)\n", + __func__, + llu(op->tag), + op, + op->attempts); + return (op->attempts < ORANGEFS_PURGE_RETRY_COUNT) ? + -EAGAIN : + -EIO; + } + /* must have timed out, then... */ + gossip_debug(GOSSIP_WAIT_DEBUG, + "*** %s:" + " operation timed out (tag" + " %llu, %p, att %d)\n", + __func__, + llu(op->tag), + op, + op->attempts); + return -ETIMEDOUT; }