[PATCH 2/3] nvme: Sync queues on controller resets

jianchao.wang jianchao.w.wang at oracle.com
Sat Feb 10 17:53:03 PST 2018


Hi Keith

On 02/10/2018 09:55 AM, jianchao.wang wrote:
> Hi Keith
> 
> Thanks for your time and patch for this.
> 
> However, as I shared last time, there is defect here. Please refer to.
> 
> There could be a circular pattern here. Please consider the following scenario:
> 
> timeout_work context                    reset_work context
> nvme_timeout                            nvme_reset_work
>   -> nvme_dev_disable                     -> nvme_sync_queues // hold namespace_mutex
>     -> nvme_stop_queues                     -> blk_sync_queue
>       -> require namespaces_mutex               -> cancel_work_sync(&q->timeout_work)
> 

Looks like we could use rwsem to replace namespaces_mutex.

Thanks
Jianchao


> On the other hand, the blk_mq_kick_requeue_list() should be also added in nvme_kill_queues
> for the case of queue_count < 2
> 
> Thanks
> Jianchao
> 
> On 02/10/2018 01:41 AM, Keith Busch wrote:
>> This patch has the nvme pci driver synchronize request queues to ensure
>> starting the controller is not racing with a previously running timeout
>> handler.
>>
>> Reported-by: Jianchao Wang <jianchao.w.wang at oracle.com>
>> Signed-off-by: Keith Busch <keith.busch at intel.com>
>> ---
>>  drivers/nvme/host/core.c | 15 ++++++++++++++-
>>  drivers/nvme/host/nvme.h |  1 +
>>  drivers/nvme/host/pci.c  |  1 +
>>  3 files changed, 16 insertions(+), 1 deletion(-)
>>
>> diff --git a/drivers/nvme/host/core.c b/drivers/nvme/host/core.c
>> index 2fd8688cfa47..a9bce23a991f 100644
>> --- a/drivers/nvme/host/core.c
>> +++ b/drivers/nvme/host/core.c
>> @@ -3542,12 +3542,25 @@ void nvme_start_queues(struct nvme_ctrl *ctrl)
>>  	struct nvme_ns *ns;
>>  
>>  	mutex_lock(&ctrl->namespaces_mutex);
>> -	list_for_each_entry(ns, &ctrl->namespaces, list)
>> +	list_for_each_entry(ns, &ctrl->namespaces, list) {
>>  		blk_mq_unquiesce_queue(ns->queue);
>> +		blk_mq_kick_requeue_list(ns->queue);
>> +	}
>>  	mutex_unlock(&ctrl->namespaces_mutex);
>>  }
>>  EXPORT_SYMBOL_GPL(nvme_start_queues);
>>  
>> +void nvme_sync_queues(struct nvme_ctrl *ctrl)
>> +{
>> +	struct nvme_ns *ns;
>> +
>> +	mutex_lock(&ctrl->namespaces_mutex);
>> +	list_for_each_entry(ns, &ctrl->namespaces, list)
>> +		blk_sync_queue(ns->queue);
>> +	mutex_unlock(&ctrl->namespaces_mutex);
>> +}
>> +EXPORT_SYMBOL_GPL(nvme_sync_queues);
>> +
>>  int nvme_reinit_tagset(struct nvme_ctrl *ctrl, struct blk_mq_tag_set *set)
>>  {
>>  	if (!ctrl->ops->reinit_request)
>> diff --git a/drivers/nvme/host/nvme.h b/drivers/nvme/host/nvme.h
>> index 27e31c00b306..466081e5f680 100644
>> --- a/drivers/nvme/host/nvme.h
>> +++ b/drivers/nvme/host/nvme.h
>> @@ -374,6 +374,7 @@ void nvme_complete_async_event(struct nvme_ctrl *ctrl, __le16 status,
>>  
>>  void nvme_stop_queues(struct nvme_ctrl *ctrl);
>>  void nvme_start_queues(struct nvme_ctrl *ctrl);
>> +void nvme_sync_queues(struct nvme_ctrl *ctrl);
>>  void nvme_kill_queues(struct nvme_ctrl *ctrl);
>>  void nvme_unfreeze(struct nvme_ctrl *ctrl);
>>  void nvme_wait_freeze(struct nvme_ctrl *ctrl);
>> diff --git a/drivers/nvme/host/pci.c b/drivers/nvme/host/pci.c
>> index 90e276c05f79..7a2e4383c468 100644
>> --- a/drivers/nvme/host/pci.c
>> +++ b/drivers/nvme/host/pci.c
>> @@ -2290,6 +2290,7 @@ static void nvme_reset_work(struct work_struct *work)
>>  	 */
>>  	if (dev->ctrl.ctrl_config & NVME_CC_ENABLE)
>>  		nvme_dev_disable(dev, false);
>> +	nvme_sync_queues(&dev->ctrl);
>>  
>>  	/*
>>  	 * Introduce CONNECTING state from nvme-fc/rdma transports to mark the
>>
> 
> _______________________________________________
> Linux-nvme mailing list
> Linux-nvme at lists.infradead.org
> https://urldefense.proofpoint.com/v2/url?u=http-3A__lists.infradead.org_mailman_listinfo_linux-2Dnvme&d=DwICAg&c=RoP1YumCXCgaWHvlZYR8PZh8Bv7qIrMUB65eapI_JnE&r=7WdAxUBeiTUTCy8v-7zXyr4qk7sx26ATvfo6QSTvZyQ&m=-ffcRNVB1ERVD0CZrnJUyZOFBETbuy0WHX7ksyM8Dcs&s=KWBywqLDbcong1rA6oK0dD_hff4FSr7Y86aUkY-gskE&e=
> 



More information about the Linux-nvme mailing list