[PATCH V5 0/9] nvme: pci: fix & improve timeout handling
Ming Lei
ming.lei at redhat.com
Tue May 15 05:56:41 PDT 2018
On Tue, May 15, 2018 at 05:56:14PM +0800, jianchao.wang wrote:
> Hi ming
>
> On 05/15/2018 08:33 AM, Ming Lei wrote:
> > We still have to quiesce admin queue before canceling request, so looks
> > the following patch is better, so please ignore the above patch and try
> > the following one and see if your hang can be addressed:
> >
> > diff --git a/drivers/nvme/host/pci.c b/drivers/nvme/host/pci.c
> > index f509d37b2fb8..c2adc76472a8 100644
> > --- a/drivers/nvme/host/pci.c
> > +++ b/drivers/nvme/host/pci.c
> > @@ -1741,8 +1741,7 @@ static int nvme_alloc_admin_tags(struct nvme_dev *dev)
> > dev->ctrl.admin_q = NULL;
> > return -ENODEV;
> > }
> > - } else
> > - blk_mq_unquiesce_queue(dev->ctrl.admin_q);
> > + }
> >
> > return 0;
> > }
> > @@ -2520,6 +2519,12 @@ static void nvme_dev_disable(struct nvme_dev *dev, bool shutdown, bool
> > */
> > if (shutdown)
> > nvme_start_queues(&dev->ctrl);
> > +
> > + /*
> > + * Avoid to suck reset because timeout may happen during reset and
> > + * reset may hang forever if admin queue is kept as quiesced
> > + */
> > + blk_mq_unquiesce_queue(dev->ctrl.admin_q);
> > mutex_unlock(&dev->shutdown_lock);
> > }
>
> w/ patch above and patch below, both the warning and io hung issue didn't reproduce till now.
>
That is great to see another issue which can be covered now, :-)
>
> @@ -1450,6 +1648,7 @@ static int nvme_create_queue(struct nvme_queue *nvmeq, int qid)
> {
> struct nvme_dev *dev = nvmeq->dev;
> int result;
> + int cq_vector;
>
> if (dev->cmb && use_cmb_sqes && (dev->cmbsz & NVME_CMBSZ_SQS)) {
> unsigned offset = (qid - 1) * roundup(SQ_SIZE(nvmeq->q_depth),
> @@ -1462,15 +1661,16 @@ static int nvme_create_queue(struct nvme_queue *nvmeq, int qid)
> * A queue's vector matches the queue identifier unless the controller
> * has only one vector available.
> */
> - nvmeq->cq_vector = dev->num_vecs == 1 ? 0 : qid;
> - result = adapter_alloc_cq(dev, qid, nvmeq);
> + cq_vector = dev->num_vecs == 1 ? 0 : qid;
> + result = adapter_alloc_cq(dev, qid, nvmeq, cq_vector);
> if (result < 0)
> - goto release_vector;
> + goto out;
>
> result = adapter_alloc_sq(dev, qid, nvmeq);
> if (result < 0)
> goto release_cq;
> -
> +
> + nvmeq->cq_vector = cq_vector;
> nvme_init_queue(nvmeq, qid);
> result = queue_request_irq(nvmeq);
> if (result < 0)
> @@ -1479,12 +1679,12 @@ static int nvme_create_queue(struct nvme_queue *nvmeq, int qid)
> return result;
>
> release_sq:
> + nvmeq->cq_vector = -1;
> dev->online_queues--;
> adapter_delete_sq(dev, qid);
> release_cq:
> adapter_delete_cq(dev, qid);
> - release_vector:
> - nvmeq->cq_vector = -1;
> + out:
> return result;
> }
>
Looks a nice fix on nvme_create_queue(), but seems the change on
adapter_alloc_cq() is missed in above patch.
Could you prepare a formal one so that I may integrate it to V6?
Thanks,
Ming
More information about the Linux-nvme
mailing list