[PATCH] nvme: don't retry request marked as NVME_REQ_CANCELLED
jianchao.wang
jianchao.w.wang at oracle.com
Thu Jan 25 00:52:31 PST 2018
Hi Ming
Thanks for your patch and detailed comment.
On 01/25/2018 04:10 PM, Ming Lei wrote:
> If request is marked as NVME_REQ_CANCELLED, we don't need to retry for
> requeuing it, and it should be completed immediately. Even simply from
> the flag name, it needn't to be requeued.
>
> Otherwise, it is easy to cause IO hang when IO is timed out in case of
> PCI NVMe:
>
> 1) IO timeout is triggered, and nvme_timeout() tries to disable
> device(nvme_dev_disable) and reset controller(nvme_reset_ctrl)
>
> 2) inside nvme_dev_disable(), queue is frozen and quiesced, and
> try to cancel every request, but the timeout request can't be canceled
> since it is completed by __blk_mq_complete_request() in blk_mq_rq_timed_out().
>
> 3) this timeout req is requeued via nvme_complete_rq(), but can't be
> dispatched at all because queue is quiesced and hardware isn't ready,
> finally nvme_wait_freeze() waits for ever in nvme_reset_work().
The queue will be unquiesced by nvme_start_queues() before start to wait freeze.
nvme_reset_work
....
nvme_start_queues(&dev->ctrl);
nvme_wait_freeze(&dev->ctrl);
/* hit this only when allocate tagset fails */
if (nvme_dev_add(dev))
new_state = NVME_CTRL_ADMIN_ONLY;
nvme_unfreeze(&dev->ctrl);
....
And the relationship between nvme_timeout and nvme_dev_disable is really complicated.
The nvme_timeout may run with nvme_dev_disable in parallel and also could invoke it.
nvme_dev_disbale may sleep holding shutdown_lock, nvme_timeout will try to get it when invokes nvme_dev_disable....
On the other hand, can we provide a helper interface to quiesce the request_work timeout_work ?
Such as following:
diff --git a/block/blk-core.c b/block/blk-core.c
index b888175..60469cd 100644
--- a/block/blk-core.c
+++ b/block/blk-core.c
@@ -616,6 +616,44 @@ void blk_queue_bypass_end(struct request_queue *q)
}
EXPORT_SYMBOL_GPL(blk_queue_bypass_end);
+/*
+ * When returns, the timeout_work will be invoked any more.
+ * The caller must take the charge of the outstanding IOs.
+ */
+void blk_quiesce_queue_timeout(struct request_queue *q)
+{
+ unsigned long flags;
+
+ spin_lock_irqsave(q->queue_lock, flags);
+ queue_flag_set(QUEUE_FLAG_TIMEOUT_QUIESCED, q);
+ spin_unlock_irqrestore(q->queue_lock, flags);
+
+ synchronize_rcu_bh();
+ cancel_work_sync(&q->timeout_work);
+ /*
+ * The timeout work will not be invoked anymore.
+ */
+ del_timer_sync(&q->timeout);
+ /*
+ * If the caller doesn't want the timer to be re-armed,
+ * it has to stop/quiesce the queue.
+ */
+}
+EXPORT_SYMBOL_GPL(blk_quiesce_queue_timeout);
+
+/*
+ * Caller must ensure all the outstanding IOs has been handled.
+ */
+void blk_unquiesce_queue_timeout(struct request_queue *q)
+{
+ unsigned long flags;
+
+ spin_lock_irqsave(q->queue_lock, flags);
+ queue_flag_clear(QUEUE_FLAG_TIMEOUT_QUIESCED, q);
+ spin_unlock_irqrestore(q->queue_lock, flags);
+}
+
+EXPORT_SYMBOL_GPL(blk_unquiesce_queue_timeout);
void blk_set_queue_dying(struct request_queue *q)
{
spin_lock_irq(q->queue_lock);
@@ -867,6 +905,9 @@ static void blk_rq_timed_out_timer(struct timer_list *t)
{
struct request_queue *q = from_timer(q, t, timeout);
+ if (blk_queue_timeout_quiesced(q))
+ return;
+
kblockd_schedule_work(&q->timeout_work);
}
More information about the Linux-nvme
mailing list