Sequential read from NVMe/XFS twice slower on Fedora 42 than on Rocky 9.5

Laurence Oberman loberman at redhat.com
Mon May 5 05:29:21 PDT 2025


On Mon, 2025-05-05 at 07:50 +1000, Dave Chinner wrote:
> [cc linux-block]
> 
> [original bug report:
> https://lore.kernel.org/linux-xfs/CAAiJnjoo0--yp47UKZhbu8sNSZN6DZ-QzmZBMmtr1oC=fOOgAQ@mail.gmail.com/
>  ]
> 
> On Sun, May 04, 2025 at 10:22:58AM +0300, Anton Gavriliuk wrote:
> > > What's the comparitive performance of an identical read profile
> > > directly on the raw MD raid0 device?
> > 
> > Rocky 9.5 (5.14.0-503.40.1.el9_5.x86_64)
> > 
> > [root at localhost ~]# df -mh /mnt
> > Filesystem      Size  Used Avail Use% Mounted on
> > /dev/md127       35T  1.3T   34T   4% /mnt
> > 
> > [root at localhost ~]# fio --name=test --rw=read --bs=256k
> > --filename=/dev/md127 --direct=1 --numjobs=1 --iodepth=64 --exitall
> > --group_reporting --ioengine=libaio --runtime=30 --time_based
> > test: (g=0): rw=read, bs=(R) 256KiB-256KiB, (W) 256KiB-256KiB, (T)
> > 256KiB-256KiB, ioengine=libaio, iodepth=64
> > fio-3.39-44-g19d9
> > Starting 1 process
> > Jobs: 1 (f=1): [R(1)][100.0%][r=81.4GiB/s][r=334k IOPS][eta
> > 00m:00s]
> > test: (groupid=0, jobs=1): err= 0: pid=43189: Sun May  4 08:22:12
> > 2025
> >   read: IOPS=363k, BW=88.5GiB/s (95.1GB/s)(2656GiB/30001msec)
> >     slat (nsec): min=971, max=312380, avg=1817.92, stdev=1367.75
> >     clat (usec): min=78, max=1351, avg=174.46, stdev=28.86
> >      lat (usec): min=80, max=1352, avg=176.27, stdev=28.81
> > 
> > Fedora 42 (6.14.5-300.fc42.x86_64)
> > 
> > [root at localhost anton]# df -mh /mnt
> > Filesystem      Size  Used Avail Use% Mounted on
> > /dev/md127       35T  1.3T   34T   4% /mnt
> > 
> > [root at localhost ~]# fio --name=test --rw=read --bs=256k
> > --filename=/dev/md127 --direct=1 --numjobs=1 --iodepth=64 --exitall
> > --group_reporting --ioengine=libaio --runtime=30 --time_based
> > test: (g=0): rw=read, bs=(R) 256KiB-256KiB, (W) 256KiB-256KiB, (T)
> > 256KiB-256KiB, ioengine=libaio, iodepth=64
> > fio-3.39-44-g19d9
> > Starting 1 process
> > Jobs: 1 (f=1): [R(1)][100.0%][r=41.0GiB/s][r=168k IOPS][eta
> > 00m:00s]
> > test: (groupid=0, jobs=1): err= 0: pid=5685: Sun May  4 10:14:00
> > 2025
> >   read: IOPS=168k, BW=41.0GiB/s (44.1GB/s)(1231GiB/30001msec)
> >     slat (usec): min=3, max=273, avg= 5.63, stdev= 1.48
> >     clat (usec): min=67, max=2800, avg=374.99, stdev=29.90
> >      lat (usec): min=72, max=2914, avg=380.62, stdev=30.22
> 
> So the MD block device shows the same read performance as the
> filesystem on top of it. That means this is a regression at the MD
> device layer or in the block/driver layers below it. i.e. it is not
> an XFS of filesystem issue at all.
> 
> -Dave.

I have a lab setup, let me see if I can also reproduce and then trace
this to see where it is spending the time




More information about the Linux-nvme mailing list