Higher block layer latency in kernel v4.8-r6 vs. v4.4.16 for NVMe
Alana Alexander-Rutledge
Alana.Alexander-Rutledge at microsemi.com
Thu Nov 10 14:14:26 PST 2016
Thanks for the info. Yes, maybe that change in ordering could explain the increased Q2D latency then.
I am actually using blk-mq for the 12G SAS. The latencies appear to have been pretty similar in v4.4.16 but not in v4.8-rc6. It does seem odd that they are different in v4.8-rc6. I've noticed that the average submission latencies between the SAS and NVMe are also around 3 or so us higher for NVMe at queue depth=1. I thought those paths would also be the same so that doesn't really make sense to me either.
The overall latency does look like it increased for v4.8-rc6 as well, mainly for queue depth <= 4. In the fio reports, I can see that both the submission and completion latencies for these cases are higher for v4.8-rc6. Below are just the fio reported average latencies (us).
Queue Depth v4.4.16 v4.8-rc6
1 91.64 119.65
2 91.38 112.42
4 91.56 112.39
8 94.57 95.29
16 106.25 107.90
32 181.36 173.40
64 263.58 265.89
128 512.82 519.96
Thanks,
Alana
-----Original Message-----
From: Keith Busch [mailto:keith.busch at intel.com]
Sent: Thursday, November 10, 2016 11:05 AM
To: Alana Alexander-Rutledge <Alana.Alexander-Rutledge at microsemi.com>
Cc: linux-block at vger.kernel.org; linux-nvme at lists.infradead.org; Stephen Bates <stephen.bates at microsemi.com>
Subject: Re: Higher block layer latency in kernel v4.8-r6 vs. v4.4.16 for NVMe
EXTERNAL EMAIL
On Wed, Nov 09, 2016 at 01:43:55AM +0000, Alana Alexander-Rutledge wrote:
> Hi,
>
> I have been profiling the performance of the NVMe and SAS IO stacks on Linux. I used blktrace and blkparse to collect block layer trace points and a custom analysis script on the trace points to average out the latencies of each trace point interval of each IO.
>
> I started with Linux kernel v4.4.16 but then switched to v4.8-r6. One thing that stood out is that for measurements at queue depth = 1, the average Q2D latency was quite a bit higher in the NVMe path with the newer version of the kernel.
>
> The Q, G, I, and D below refer to blktrace/blkparse trace points (queued, get request, inserted, and issued).
>
> Queue Depth = 1
> Interval Average - v4.4.16 (us) Average - v4.8-rc6 (us)
> Q2G 0.212 0.573
> G2I 0.944 1.507
> I2D 0.435 0.837
> Q2D 1.592 2.917
>
> For other queue depths, Q2D was similar for both versions of the kernel.
>
> Queue Depth Average Q2D - v4.4.16 (us) Average Q2D - v4.8-rc6 (us)
> 2 1.893 1.736
> 4 1.289 1.38
> 8 1.223 1.162
> 16 1.14 1.178
> 32 1.007 1.425
> 64 0.964 0.978
> 128 0.915 0.941
>
> I did not see this as a problem with the 12G SAS SSD that I measured.
>
> Queue Depth = 1
> Interval Average - v4.4.16 (us) Average - v4.8-rc6 (us)
> Q2G 0.264 0.301
> G2I 0.917 0.864
> I2D 0.432 0.397
> Q2D 1.613 1.561
>
> Is this a known change or do you know what the reason for this is?
Are you using blk-mq for the 12G SAS? I assume not since most of these intervals would have executed through the same code path and shouldn't show a difference from to the underlying driver.
My guess for at least part of the additional latency to D/issued, the nvme driver in 4.1 used to call blk_mq_start_request (marks the "issued" trace
point) before it constructed the nvme command. 4.8 calls it after.
Have you noticed a difference in over-all latency?
More information about the Linux-nvme
mailing list