[RFC 0/8] Copy Offload with Peer-to-Peer PCI Memory
Jason Gunthorpe
jgunthorpe at obsidianresearch.com
Tue Apr 18 12:48:45 PDT 2017
On Tue, Apr 18, 2017 at 01:35:32PM -0600, Logan Gunthorpe wrote:
> > Ultimately every dma_ops will need special code to support P2P with
> > the special hardware that ops is controlling, so it makes some sense
> > to start by pushing the check down there in the first place. This
> > advice is partially motivated by how dma_map_sg is just a small
> > wrapper around the function pointer call...
>
> Yes, I noticed this problem too and that makes sense. It just means
> every dma_ops will probably need to be modified to either support p2p
> pages or fail on them. Though, the only real difficulty there is that it
> will be a lot of work.
I think this is why progress on this keeps getting stuck - every
solution is a lot of work.
> > Where p2p_same_segment_map_page checks if the two devices are on the
> > 'same switch' and if so returns the address translated to match the
> > bus address programmed into the BAR or fails. We knows this case is
> > required to work by the PCI spec, so it makes sense to use it as the
> > first canned helper.
>
> I've also suggested that this check should probably be done (or perhaps
> duplicated) before we even get to the map stage.
Since the mechanics of the check is essentially unique to every
dma-ops I would not hoist it out of the map function without a really
good reason.
> In the case of nvme-fabrics we'd probably want to let the user know
> when they try to configure it or at least fall back to allocating
> regular memory instead.
You could try to do a dummy mapping / create a MR early on to detect
this.
FWIW, I wonder if from a RDMA perspective we have another
problem.. Should we allow P2P memory to be used with the local DMA
lkey? There are potential designs around virtualization that would not
allow that. Should we mandate that P2P memory be in its own MR?
Jason
More information about the Linux-nvme
mailing list