WARN_ON() when booting with nvme loopback over zram

Johannes Thumshirn jthumshirn at suse.de
Fri Mar 3 08:18:27 PST 2017


Hi,

I get the following WARN_ON when trying to establish a nvmf loopback device
backed by zram.

My topmost commit is c82be9d2244aacea9851c86f4fb74694c99cd874


+ nvmet_cfs=/sys/kernel/config/nvmet/
+ nvmet_subsystem=nvmf-test
+ mkdir -p /sys/kernel/config/nvmet//subsystems/nvmf-test
+ echo 1
+ mkdir /sys/kernel/config/nvmet//subsystems/nvmf-test/namespaces/1
[    6.163905] nvmet: adding nsid 1 to subsystem nvmf-test
+ echo -n /dev/zram1
+ echo -n 1
+ mkdir /sys/kernel/config/nvmet//ports/1
+ echo loop
+ ln -s /sys/kernel/config/nvmet//subsystems/nvmf-test /sys/kernel/config/nvmet//ports/1/subsystems/nvmf-test
+ echo transport=loop,nqn=nvmf-test
[    6.175710] ------------[ cut here ]------------
[    6.176181] WARNING: CPU: 1 PID: 207 at block/blk-mq-tag.c:114 blk_mq_get_tag+0x4ec/0x580
[    6.176922] Modules linked in: nvme_loop nvmet nvme_fabrics nvme_core
[    6.177523] CPU: 1 PID: 207 Comm: sh Not tainted 4.10.0+ #413
[    6.178048] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS rel-1.9.1-0-gb3ef39f-prebuilt.qemu-project.org 04/01/2014
[    6.179089] Call Trace:
[    6.179331]  dump_stack+0x4d/0x68
[    6.179656]  __warn+0x107/0x130
[    6.179950]  warn_slowpath_null+0x18/0x20
[    6.180326]  blk_mq_get_tag+0x4ec/0x580
[    6.180679]  ? __blk_mq_tag_idle+0x40/0x40
[    6.181069]  ? save_stack_trace+0x16/0x20
[    6.181442]  ? wake_atomic_t_function+0x90/0x90
[    6.181512]  ? __vfs_write+0xd1/0x330
[    6.181512]  ? vfs_write+0x10e/0x270
[    6.181512]  ? SyS_write+0xa5/0x130
[    6.181512]  ? entry_SYSCALL_64_fastpath+0x13/0x94
[    6.181512]  ? blk_queue_enter+0x1ac/0x280
[    6.181512]  __blk_mq_alloc_request+0x1c/0x1b0
[    6.181512]  blk_mq_sched_get_request+0x333/0x420
[    6.181512]  ? __list_add_valid+0x2e/0xd0
[    6.181512]  blk_mq_alloc_request+0xb9/0x120
[    6.181512]  ? __blk_mq_alloc_request+0x1b0/0x1b0
[    6.181512]  ? kmemleak_disable+0x70/0x70
[    6.181512]  nvme_alloc_request+0x98/0xb0 [nvme_core]
[    6.181512]  __nvme_submit_sync_cmd+0x2c/0x110 [nvme_core]
[    6.181512]  nvmf_connect_admin_queue+0x216/0x2b0 [nvme_fabrics]
[    6.181512]  ? nvmf_log_connect_error.isra.0+0x130/0x130 [nvme_fabrics]
[    6.181512]  ? blk_mq_sched_init+0x2e/0x40
[    6.181512]  ? blk_mq_init_allocated_queue+0x791/0x7c0
[    6.181512]  nvme_loop_configure_admin_queue+0x168/0x270 [nvme_loop]
[    6.181512]  nvme_loop_create_ctrl+0x23e/0x8f8 [nvme_loop]
[    6.181512]  ? __delete_object+0x59/0xa0
[    6.181512]  ? delete_object_full+0x18/0x20
[    6.181512]  nvmf_dev_write+0xbb1/0xd77 [nvme_fabrics]
[    6.181512]  ? nvmf_check_required_opts.isra.2+0xa0/0xa0 [nvme_fabrics]
[    6.181512]  ? kasan_slab_free+0x12f/0x180
[    6.181512]  ? save_stack_trace+0x16/0x20
[    6.181512]  ? kasan_slab_free+0xae/0x180
[    6.181512]  ? kmem_cache_free+0x84/0x150
[    6.181512]  ? putname+0x7b/0x80
[    6.181512]  ? do_sys_open+0x23f/0x290
[    6.181512]  ? SyS_open+0x19/0x20
[    6.181512]  ? entry_SYSCALL_64_fastpath+0x13/0x94
[    6.181512]  ? __save_stack_trace+0x7e/0xd0
[    6.181512]  __vfs_write+0xd1/0x330
[    6.181512]  ? restore_nameidata+0x7a/0xa0
[    6.181512]  ? __vfs_read+0x320/0x320
[    6.181512]  ? ptep_set_access_flags+0x2b/0x50
[    6.181512]  ? __handle_mm_fault+0xc9d/0x14e0
[    6.181512]  ? vm_insert_page+0x320/0x320
[    6.181512]  ? locks_remove_posix+0x38/0x70
[    6.181512]  vfs_write+0x10e/0x270
[    6.181512]  SyS_write+0xa5/0x130
[    6.181512]  ? SyS_read+0x130/0x130
[    6.181512]  entry_SYSCALL_64_fastpath+0x13/0x94
[    6.181512] RIP: 0033:0x7fd6d0de8560
[    6.181512] RSP: 002b:00007ffd284d2278 EFLAGS: 00000246 ORIG_RAX: 0000000000000001
[    6.181512] RAX: ffffffffffffffda RBX: 000000000000001c RCX: 00007fd6d0de8560
[    6.181512] RDX: 000000000000001d RSI: 00007fd6d194a000 RDI: 0000000000000001
[    6.181512] RBP: 00007fd6d10a9280 R08: 000000000000000a R09: 00007fd6d194c700
[    6.181512] R10: 0000000001094b50 R11: 0000000000000246 R12: 0000000001094b50
[    6.181512] R13: 000000000000001c R14: 0000000000000000 R15: 00007ffd284d2228
[    6.203886] ---[ end trace 96b98033c328af9c ]---
[    6.204314] nvme nvme0: Connect command failed, error wo/DNR bit: -16395
sh: echo: write error: Resource temporarily unavailable
+ _fatal
+ echo 1
+ e[    6.226208] sysrq: SysRq : Power Off
cho o
+ sleep 2
[    7.880759] ACPI: Preparing to enter system sleep state S5
[    7.881327] reboot: Power down

Is this known or shall I re-test with Jens' latest tree?

Nice weekend,
	Joahnnes

-- 
Johannes Thumshirn                                          Storage
jthumshirn at suse.de                                +49 911 74053 689
SUSE LINUX GmbH, Maxfeldstr. 5, 90409 Nürnberg
GF: Felix Imendörffer, Jane Smithard, Graham Norton
HRB 21284 (AG Nürnberg)
Key fingerprint = EC38 9CAB C2C4 F25D 8600 D0D0 0393 969D 2D76 0850



More information about the Linux-nvme mailing list