NVMe scalability issue

Jens Axboe axboe at fb.com
Tue Jun 2 11:22:23 PDT 2015


On 06/02/2015 11:24 AM, Ming Lin wrote:
> On Mon, Jun 1, 2015 at 8:30 PM, Keith Busch <keith.busch at intel.com> wrote:
>> On Mon, 1 Jun 2015, Ming Lin wrote:
>>>
>>> On Mon, Jun 1, 2015 at 4:02 PM, Keith Busch <keith.busch at intel.com> wrote:
>>>>
>>>> There was a demo at SC'14 with a heck of a lot more NVMe drives than
>>>> that,
>>>> and performance scaled quite linearly. Are your devices sharing PCI-e
>>>> lanes?
>>>
>>>
>>> Is there a way to check it via, for example, /sys?
>>
>>
>>    # lspci -tv
>
> Each 4 drives share x16 lane.
>
>>
>>>> You could try setting "cpus_allowed" on each job to the CPU's on the
>>>> socket local to the nvme device. That should get a measurable
>>>> improvement,
>>>> and if your irq's are appropriately affinitized.
>>>
>>>
>>> How to know which socket is local to which nvme device?
>>
>>
>>    # cat /sys/class/nvme/nvme<#>/device/numa_node
>
> # grep . /sys/class/nvme/nvme*/device/numa_node
> /sys/class/nvme/nvme0/device/numa_node:1
> /sys/class/nvme/nvme1/device/numa_node:1
> /sys/class/nvme/nvme2/device/numa_node:1
> /sys/class/nvme/nvme3/device/numa_node:1
> /sys/class/nvme/nvme4/device/numa_node:2
> /sys/class/nvme/nvme5/device/numa_node:2
> /sys/class/nvme/nvme6/device/numa_node:2
> /sys/class/nvme/nvme7/device/numa_node:2
>
> With correct numa_node binding, now I can get 5010K IOPS with 8 drives.
> It's better now, but still not linear scaled to 5864K
>
> I'll check if irq's are appropriately affinitized.

Just a thought, but one thing that fio is pretty intensive on is time 
keeping. Depending on the platform, there's some shared state between 
the fio IO threads. Does the picture change if you add gtod_reduce=0?
In general, I'd also turn off strict random tracking. Either add 
'norandommap' as an option, or use random_generator=lfsr instead.

-- 
Jens Axboe




More information about the Linux-nvme mailing list