slab-use-after-free in __ib_process_cq

Sagi Grimberg sagi at grimberg.me
Thu May 4 01:32:36 PDT 2023


> Hi,
> 
> While testing Jens' for-next branch I encountered a use-after-free
> issue, triggered by test nvmeof-mp/002. This is not the first time I see
> this issue - I had already observed this several weeks ago but I had not
> yet had the time to report this.

That is surprising because this area did not change for quite a while now.

CCing linux-rdma as well, I'm assuming that this is with rxe?
Does this happen with siw as well?
> 
> Thanks,
> 
> Bart.
> 
> [ 1377.113215] 
> ==================================================================
> [ 1377.119731] BUG: KASAN: slab-use-after-free in 
> __ib_process_cq+0x11c/0x300 [ib_core]
> [ 1377.120242] nvmet_rdma:nvmet_rdma_cm_handler: nvmet_rdma: 
> disconnected (10): status 0 id 00000000bcabdd1c
> [ 1377.123791] Read of size 8 at addr ffff88813e60e000 by task 
> ksoftirqd/24/140
> [ 1377.123799]
> [ 1377.123803] CPU: 24 PID: 140 Comm: ksoftirqd/24 Tainted: G        W   
> E      6.3.0-dbg #36
> [ 1377.123808] Hardware name: QEMU Standard PC (Q35 + ICH9, 2009), BIOS 
> 1.16.0-debian-1.16.0-5 04/01/2014
> [ 1377.123812] Call Trace:
> [ 1377.123815]  <TASK>
> [ 1377.123819]  dump_stack_lvl+0x4a/0x80
> [ 1377.123832]  print_address_description.constprop.0+0x33/0x3c0
> [ 1377.123841]  ? preempt_count_sub+0x18/0xc0
> [ 1377.129660] nvmet_rdma:nvmet_rdma_free_queue: nvmet_rdma: freeing 
> queue 64
> [ 1377.132413]  print_report+0xb6/0x260
> [ 1377.133455] nvmet_rdma:__nvmet_rdma_queue_disconnect: nvmet_rdma: 
> cm_id= 00000000bcabdd1c queue->state= 1
> [ 1377.135496]  ? kasan_complete_mode_report_info+0x5c/0x190
> [ 1377.138029] nvmet_rdma:nvmet_rdma_free_queue: nvmet_rdma: freeing 
> queue 65
> [ 1377.138700]  kasan_report+0xc2/0xf0
> [ 1377.139328] nvmet_rdma:nvmet_rdma_cm_handler: nvmet_rdma: 
> disconnected (10): status 0 id 000000005baa5f8f
> [ 1377.140304]  ? __ib_process_cq+0x11c/0x300 [ib_core]
> [ 1377.141857] nvmet_rdma:__nvmet_rdma_queue_disconnect: nvmet_rdma: 
> cm_id= 000000005baa5f8f queue->state= 1
> [ 1377.142982]  ? __ib_process_cq+0x11c/0x300 [ib_core]
> [ 1377.144873] nvmet_rdma:nvmet_rdma_free_queue: nvmet_rdma: freeing 
> queue 66
> [ 1377.145854]  __asan_load8+0x69/0x90
> [ 1377.150901] nvmet_rdma:nvmet_rdma_cm_handler: nvmet_rdma: 
> disconnected (10): status 0 id 000000007181ad76
> [ 1377.152350]  __ib_process_cq+0x11c/0x300 [ib_core]
> [ 1377.154247] nvmet_rdma:__nvmet_rdma_queue_disconnect: nvmet_rdma: 
> cm_id= 000000007181ad76 queue->state= 1
> [ 1377.155204]  ib_poll_handler+0x46/0x1b0 [ib_core]
> [ 1377.158327] nvmet_rdma:nvmet_rdma_cm_handler: nvmet_rdma: 
> disconnected (10): status 0 id 000000005378c1fe
> [ 1377.159140]  irq_poll_softirq+0x1cb/0x2c0
> [ 1377.161765] nvmet_rdma:nvmet_rdma_free_queue: nvmet_rdma: freeing 
> queue 67
> [ 1377.163108]  __do_softirq+0x115/0x645
> [ 1377.163119]  run_ksoftirqd+0x37/0x60
> [ 1377.163305] nvmet_rdma:__nvmet_rdma_queue_disconnect: nvmet_rdma: 
> cm_id= 000000005378c1fe queue->state= 1
> [ 1377.163361] nvmet_rdma:nvmet_rdma_free_queue: nvmet_rdma: freeing 
> queue 68
> [ 1377.163476] nvmet_rdma:nvmet_rdma_cm_handler: nvmet_rdma: 
> disconnected (10): status 0 id 00000000a8044607
> [ 1377.163490] nvmet_rdma:__nvmet_rdma_queue_disconnect: nvmet_rdma: 
> cm_id= 00000000a8044607 queue->state= 1
> [ 1377.163527] nvmet_rdma:nvmet_rdma_free_queue: nvmet_rdma: freeing 
> queue 69
> [ 1377.163602] nvmet_rdma:nvmet_rdma_cm_handler: nvmet_rdma: 
> disconnected (10): status 0 id 00000000cb366712
> [ 1377.163615] nvmet_rdma:__nvmet_rdma_queue_disconnect: nvmet_rdma: 
> cm_id= 00000000cb366712 queue->state= 1
> [ 1377.163739] nvmet_rdma:nvmet_rdma_cm_handler: nvmet_rdma: 
> disconnected (10): status 0 id 00000000f3c4860c
> [ 1377.163752] nvmet_rdma:__nvmet_rdma_queue_disconnect: nvmet_rdma: 
> cm_id= 00000000f3c4860c queue->state= 1
> [ 1377.163818] nvmet_rdma:nvmet_rdma_cm_handler: nvmet_rdma: 
> disconnected (10): status 0 id 00000000e4bc69e3
> [ 1377.163829] nvmet_rdma:__nvmet_rdma_queue_disconnect: nvmet_rdma: 
> cm_id= 00000000e4bc69e3 queue->state= 1
> [ 1377.164168] nvmet_rdma:nvmet_rdma_free_queue: nvmet_rdma: freeing 
> queue 70
> [ 1377.164644] nvmet_rdma:nvmet_rdma_free_queue: nvmet_rdma: freeing 
> queue 71
> [ 1377.165068] nvmet_rdma:nvmet_rdma_free_queue: nvmet_rdma: freeing 
> queue 72
> [ 1377.165961]  smpboot_thread_fn+0x2ef/0x400
> [ 1377.165967]  ? _local_bh_enable+0x80/0x80
> [ 1377.339612]  ? sort_range+0x30/0x30
> [ 1377.343292]  kthread+0x185/0x1c0
> [ 1377.347334]  ? kthread_complete_and_exit+0x30/0x30
> [ 1377.351696]  ret_from_fork+0x1f/0x30
> [ 1377.355575]  </TASK>
> [ 1377.359060]
> [ 1377.391814] Allocated by task 15493:
> [ 1377.401149]  kasan_save_stack+0x26/0x50
> [ 1377.408029]  kasan_set_track+0x25/0x30
> [ 1377.412759]  kasan_save_alloc_info+0x1e/0x30
> [ 1377.417555]  __kasan_kmalloc+0x90/0xa0
> [ 1377.421267]  __kmalloc+0x62/0x120
> [ 1377.424637]  nvme_rdma_create_queue_ib+0x1a7/0x710 [nvme_rdma]
> [ 1377.428692]  nvme_rdma_cm_handler+0x4fa/0x6b0 [nvme_rdma]
> [ 1377.432600]  cma_cm_event_handler+0x92/0x290 [rdma_cm]
> [ 1377.436409]  addr_handler+0x16f/0x280 [rdma_cm]
> [ 1377.440030]  process_one_req+0x91/0x290 [ib_core]
> [ 1377.443707]  process_one_work+0x552/0xa30
> [ 1377.447081]  worker_thread+0x90/0x650
> [ 1377.450319]  kthread+0x185/0x1c0
> [ 1377.453412]  ret_from_fork+0x1f/0x30
> [ 1377.456543]
> [ 1377.459034] Freed by task 16086:
> [ 1377.461988]  kasan_save_stack+0x26/0x50
> [ 1377.465075]  kasan_set_track+0x25/0x30
> [ 1377.468085]  kasan_save_free_info+0x2e/0x50
> [ 1377.471143]  ____kasan_slab_free+0x14f/0x1b0
> [ 1377.474226]  __kasan_slab_free+0x12/0x20
> [ 1377.477137]  __kmem_cache_free+0x1b6/0x2d0
> [ 1377.480045]  kfree+0x99/0x160
> [ 1377.482601]  nvme_rdma_destroy_queue_ib+0x1b8/0x250 [nvme_rdma]
> [ 1377.485962]  nvme_rdma_free_queue+0x3e/0x50 [nvme_rdma]
> [ 1377.489117]  nvme_rdma_configure_io_queues+0x138/0x390 [nvme_rdma]
> [ 1377.492549]  nvme_rdma_setup_ctrl+0x137/0x4a0 [nvme_rdma]
> [ 1377.495733]  nvme_rdma_create_ctrl+0x4bd/0x6c0 [nvme_rdma]
> [ 1377.498900]  nvmf_create_ctrl+0x1c8/0x4c0 [nvme_fabrics]
> [ 1377.502024]  nvmf_dev_write+0xb6/0x120 [nvme_fabrics]
> [ 1377.505228]  vfs_write+0x1b7/0x7b0
> [ 1377.507779]  ksys_write+0xbb/0x140
> [ 1377.510274]  __x64_sys_write+0x42/0x50
> [ 1377.512833]  do_syscall_64+0x34/0x80
> [ 1377.515301]  entry_SYSCALL_64_after_hwframe+0x46/0xb0
> [ 1377.518157]
> [ 1377.520010] The buggy address belongs to the object at ffff88813e60e000
> [ 1377.520010]  which belongs to the cache kmalloc-4k of size 4096
> [ 1377.526186] The buggy address is located 0 bytes inside of
> [ 1377.526186]  freed 4096-byte region [ffff88813e60e000, ffff88813e60f000)
> [ 1377.532364]
> [ 1377.534252] The buggy address belongs to the physical page:
> [ 1377.537272] page:0000000098da500d refcount:1 mapcount:0 
> mapping:0000000000000000 index:0x0 pfn:0x13e608
> [ 1377.541367] head:0000000098da500d order:3 entire_mapcount:0 
> nr_pages_mapped:0 pincount:0
> [ 1377.545159] flags: 0x4000000000010200(slab|head|zone=2)
> [ 1377.548197] raw: 4000000000010200 ffff88810004d040 dead000000000122 
> 0000000000000000
> [ 1377.551942] raw: 0000000000000000 0000000000040004 00000001ffffffff 
> 0000000000000000
> [ 1377.555683] page dumped because: kasan: bad access detected
> [ 1377.558864]
> [ 1377.560967] Memory state around the buggy address:
> [ 1377.564075]  ffff88813e60df00: fc fc fc fc fc fc fc fc fc fc fc fc fc 
> fc fc fc
> [ 1377.567783]  ffff88813e60df80: fc fc fc fc fc fc fc fc fc fc fc fc fc 
> fc fc fc
> [ 1377.571461] >ffff88813e60e000: fa fb fb fb fb fb fb fb fb fb fb fb fb 
> fb fb fb
> [ 1377.575138]                    ^
> [ 1377.577761]  ffff88813e60e080: fb fb fb fb fb fb fb fb fb fb fb fb fb 
> fb fb fb
> [ 1377.581487]  ffff88813e60e100: fb fb fb fb fb fb fb fb fb fb fb fb fb 
> fb fb fb
> [ 1377.585194] 
> ==================================================================
> 



More information about the Linux-nvme mailing list