[PATCH 02/13] irq: Introduce IRQD_AFFINITY_MANAGED flag
Christoph Hellwig
hch at lst.de
Mon Jun 20 05:22:19 PDT 2016
On Thu, Jun 16, 2016 at 05:39:07PM +0200, Bart Van Assche wrote:
> On 06/16/2016 05:20 PM, Christoph Hellwig wrote:
>> On Wed, Jun 15, 2016 at 09:36:54PM +0200, Bart Van Assche wrote:
>>> Do you agree that - ignoring other interrupt assignments - that the latter
>>> interrupt assignment scheme would result in higher throughput and lower
>>> interrupt processing latency?
>>
>> Probably. Once we've got it in the core IRQ code we can tweak the
>> algorithm to be optimal.
>
> Sorry but I'm afraid that we are embedding policy in the kernel, something
> we should not do. I know that there are workloads for which dedicating some
> CPU cores to interrupt processing and other CPU cores to running kernel
> threads improves throughput, probably because this results in less cache
> eviction on the CPU cores that run kernel threads and some degree of
> interrupt coalescing on the CPU cores that process interrupts.
And you can still easily set this use case up by chosing less queues
(aka interrupts) than CPUs and assining your workload to the other
cores.
> My concern
> is that I doubt that there is an interrupt assignment scheme that works
> optimally for all workloads. Hence my request to preserve the ability to
> modify interrupt affinity from user space.
I'd say let's do such an interface incrementall based on the use
case - especially after we get networking over to use common code
to distribute the interrupts. If you were doing something like this
with the current blk-mq code it wouldn't work very well due to the
fact that you'd have a mismatch between the assigned interrupt and
the blk-mq queue mapping anyway.
It might be a good idea to start brainstorming how we'd want to handle
this change - we'd basically need a per-device notification that the
interrupt mapping changes so that we can rebuild the queue mapping,
which is somewhat similar to the lib/cpu_rmap.c code used by a few
networking drivers. This would also help with dealing with cpu
hotplug events that change the cpu mapping.
More information about the Linux-nvme
mailing list