Sequential read from NVMe/XFS twice slower on Fedora 42 than on Rocky 9.5

Anton Gavriliuk antosha20xx at gmail.com
Sun May 4 00:22:58 PDT 2025


> What's the comparitive performance of an identical read profile
> directly on the raw MD raid0 device?

Rocky 9.5 (5.14.0-503.40.1.el9_5.x86_64)

[root at localhost ~]# df -mh /mnt
Filesystem      Size  Used Avail Use% Mounted on
/dev/md127       35T  1.3T   34T   4% /mnt

[root at localhost ~]# fio --name=test --rw=read --bs=256k
--filename=/dev/md127 --direct=1 --numjobs=1 --iodepth=64 --exitall
--group_reporting --ioengine=libaio --runtime=30 --time_based
test: (g=0): rw=read, bs=(R) 256KiB-256KiB, (W) 256KiB-256KiB, (T)
256KiB-256KiB, ioengine=libaio, iodepth=64
fio-3.39-44-g19d9
Starting 1 process
Jobs: 1 (f=1): [R(1)][100.0%][r=81.4GiB/s][r=334k IOPS][eta 00m:00s]
test: (groupid=0, jobs=1): err= 0: pid=43189: Sun May  4 08:22:12 2025
  read: IOPS=363k, BW=88.5GiB/s (95.1GB/s)(2656GiB/30001msec)
    slat (nsec): min=971, max=312380, avg=1817.92, stdev=1367.75
    clat (usec): min=78, max=1351, avg=174.46, stdev=28.86
     lat (usec): min=80, max=1352, avg=176.27, stdev=28.81

Fedora 42 (6.14.5-300.fc42.x86_64)

[root at localhost anton]# df -mh /mnt
Filesystem      Size  Used Avail Use% Mounted on
/dev/md127       35T  1.3T   34T   4% /mnt

[root at localhost ~]# fio --name=test --rw=read --bs=256k
--filename=/dev/md127 --direct=1 --numjobs=1 --iodepth=64 --exitall
--group_reporting --ioengine=libaio --runtime=30 --time_based
test: (g=0): rw=read, bs=(R) 256KiB-256KiB, (W) 256KiB-256KiB, (T)
256KiB-256KiB, ioengine=libaio, iodepth=64
fio-3.39-44-g19d9
Starting 1 process
Jobs: 1 (f=1): [R(1)][100.0%][r=41.0GiB/s][r=168k IOPS][eta 00m:00s]
test: (groupid=0, jobs=1): err= 0: pid=5685: Sun May  4 10:14:00 2025
  read: IOPS=168k, BW=41.0GiB/s (44.1GB/s)(1231GiB/30001msec)
    slat (usec): min=3, max=273, avg= 5.63, stdev= 1.48
    clat (usec): min=67, max=2800, avg=374.99, stdev=29.90
     lat (usec): min=72, max=2914, avg=380.62, stdev=30.22


Anton

вс, 4 мая 2025 г. в 01:16, Dave Chinner <david at fromorbit.com>:
>
> On Sun, May 04, 2025 at 12:04:16AM +0300, Anton Gavriliuk wrote:
> > There are 12 Kioxia CM-7 NVMe SSDs configured in mdadm/raid0 and
> > mounted to /mnt.
> >
> > Exactly the same fio command running under Fedora 42
> > (6.14.5-300.fc42.x86_64) and then under Rocky 9.5
> > (5.14.0-503.40.1.el9_5.x86_64) shows twice the performance difference.
> >
> > /mnt/testfile size 1TB
> > server's total dram 192GB
> >
> > Fedora 42
> >
> > [root at localhost ~]# fio --name=test --rw=read --bs=256k
> > --filename=/mnt/testfile --direct=1 --numjobs=1 --iodepth=64 --exitall
> > --group_reporting --ioengine=libaio --runtime=30 --time_based
> > test: (g=0): rw=read, bs=(R) 256KiB-256KiB, (W) 256KiB-256KiB, (T)
> > 256KiB-256KiB, ioengine=libaio, iodepth=64
> > fio-3.39-44-g19d9
> > Starting 1 process
> > Jobs: 1 (f=1): [R(1)][100.0%][r=49.6GiB/s][r=203k IOPS][eta 00m:00s]
> > test: (groupid=0, jobs=1): err= 0: pid=2465: Sat May  3 17:51:24 2025
> >   read: IOPS=203k, BW=49.6GiB/s (53.2GB/s)(1487GiB/30001msec)
> >     slat (usec): min=3, max=1053, avg= 4.60, stdev= 1.76
> >     clat (usec): min=104, max=4776, avg=310.53, stdev=29.49
> >      lat (usec): min=110, max=4850, avg=315.13, stdev=29.82
>
> > Rocky 9.5
> >
> > [root at localhost ~]# fio --name=test --rw=read --bs=256k
> > --filename=/mnt/testfile --direct=1 --numjobs=1 --iodepth=64 --exitall
> > --group_reporting --ioengine=libaio --runtime=30 --time_based
> > test: (g=0): rw=read, bs=(R) 256KiB-256KiB, (W) 256KiB-256KiB, (T)
> > 256KiB-256KiB, ioengine=libaio, iodepth=64
> > fio-3.39-44-g19d9
> > Starting 1 process
> > Jobs: 1 (f=1): [R(1)][100.0%][r=96.0GiB/s][r=393k IOPS][eta 00m:00s]
> > test: (groupid=0, jobs=1): err= 0: pid=15467: Sun May  4 00:00:39 2025
> >   read: IOPS=390k, BW=95.3GiB/s (102GB/s)(2860GiB/30001msec)
> >     slat (nsec): min=1111, max=183816, avg=2117.94, stdev=1412.34
> >     clat (usec): min=81, max=1086, avg=161.60, stdev=19.67
> >      lat (usec): min=82, max=1240, avg=163.72, stdev=19.73
> >
>
> Completely latency has doubled on the fc42 kernel. For a read, there
> isn't much in terms of filesystem work to be done on direct IO
> completion, so I'm not sure this is a filesystem issue...
>
> What's the comparitive performance of an identical read profile
> directly on the raw MD raid0 device?
>
> -Dave.
> --
> Dave Chinner
> david at fromorbit.com



More information about the Linux-nvme mailing list