NVMe over Fabrics RDMA transport drivers V2
Jens Axboe
axboe at kernel.dk
Fri Jul 8 07:40:20 PDT 2016
On 07/06/2016 06:55 AM, Christoph Hellwig wrote:
> This patch set implements the NVMe over Fabrics RDMA host and the target
> drivers.
>
> The host driver is tied into the NVMe host stack and implements the RDMA
> transport under the NVMe core and Fabrics modules. The NVMe over Fabrics
> RDMA host module is responsible for establishing a connection against a
> given target/controller, RDMA event handling and data-plane command
> processing.
>
> The target driver hooks into the NVMe target core stack and implements
> the RDMA transport. The module is responsible for RDMA connection
> establishment, RDMA event handling and data-plane RDMA commands
> processing.
>
> RDMA connection establishment is done using RDMA/CM and IP resolution.
> The data-plane command sequence follows the classic storage model where
> the target pushes/pulls the data.
>
> Changes since V1:
> - updates for req_op changes in for-next (me)
> - validate adrfam in nvmet-rdma (Ming)
> - don't leak rsp structures on connect failure in nvmet-rdma (Steve)
> - don't use RDMA/CM errors codes in reject path in nvmet-rdma (Steve)
> - fix nvmet_rdma_delete_ctrl (me)
> - invoke fatal error on error completion in nvmet-rdma (Sagi)
> - don't leak rsp structure on disconnected queue in nvmet-rdma (Ming)
> - properly set the SGL flag on AERs in nvme-rdma (me)
> - correctly stop the keep alive timer on reconnect in nvme-rdma (Ming)
> - stop and drain queues before freeing the tagset in nvet-rdma (Steve)
Added for 4.8, thanks.
--
Jens Axboe
More information about the Linux-nvme
mailing list