======================================================
WARNING: possible circular locking dependency detected
syzkaller #0 Not tainted
------------------------------------------------------
syz.2.10971/12057 is trying to acquire lock:
ffff888161ea0418 (&hugetlb_fault_mutex_table[i]){+.+.}-{4:4}, at: hugetlb_fault+0x3d8/0x1510

but task is already holding lock:
ffff888023a4c078 (&mm->mmap_lock){++++}-{4:4}, at: lock_mm_and_find_vma+0x36/0x340

which lock already depends on the new lock.


the existing dependency chain (in reverse order) is:

-> #8 (&mm->mmap_lock){++++}-{4:4}:
       __might_fault+0xcb/0x130
       _copy_from_user+0x28/0xb0
       csum_and_copy_from_iter_full+0x1e7/0x1f00
       ip_generic_getfrag+0x149/0x2d0
       __ip6_append_data+0x39cd/0x3f60
       ip6_append_data+0x10f/0x280
       rawv6_sendmsg+0x12d3/0x18e0
       ____sys_sendmsg+0x80a/0x9f0
       ___sys_sendmsg+0x2a5/0x360
       __x64_sys_sendmsg+0x1bd/0x2a0
       do_syscall_64+0x15f/0xf80
       entry_SYSCALL_64_after_hwframe+0x77/0x7f

-> #7 (sk_lock-AF_INET6){+.+.}-{0:0}:
       lock_sock_nested+0x41/0x100
       inet_shutdown+0x6a/0x390
       nbd_mark_nsock_dead+0x2e9/0x560
       recv_work+0x1c2e/0x1d40
       process_scheduled_works+0xb5d/0x1860
       worker_thread+0xa53/0xfc0
       kthread+0x388/0x470
       ret_from_fork+0x514/0xb70
       ret_from_fork_asm+0x1a/0x30

-> #6 (&nsock->tx_lock){+.+.}-{4:4}:
       __mutex_lock+0x1a3/0x1550
       nbd_queue_rq+0x37b/0x1100
       blk_mq_dispatch_rq_list+0xa70/0x1910
       __blk_mq_sched_dispatch_requests+0xdcc/0x1600
       blk_mq_sched_dispatch_requests+0xd7/0x190
       blk_mq_run_work_fn+0x16c/0x300
       process_scheduled_works+0xb5d/0x1860
       worker_thread+0xa53/0xfc0
       kthread+0x388/0x470
       ret_from_fork+0x514/0xb70
       ret_from_fork_asm+0x1a/0x30

-> #5 (&cmd->lock){+.+.}-{4:4}:
       __mutex_lock+0x1a3/0x1550
       nbd_queue_rq+0xc6/0x1100
       blk_mq_dispatch_rq_list+0xa70/0x1910
       __blk_mq_sched_dispatch_requests+0xdcc/0x1600
       blk_mq_sched_dispatch_requests+0xd7/0x190
       blk_mq_run_work_fn+0x16c/0x300
       process_scheduled_works+0xb5d/0x1860
       worker_thread+0xa53/0xfc0
       kthread+0x388/0x470
       ret_from_fork+0x514/0xb70
       ret_from_fork_asm+0x1a/0x30

-> #4 (set->srcu){.+.+}-{0:0}:
       __synchronize_srcu+0xca/0x300
       elevator_switch+0x1e8/0x7a0
       elevator_change+0x2cc/0x450
       elevator_set_default+0x36c/0x430
       blk_register_queue+0x3e9/0x4e0
       __add_disk+0x677/0xd50
       add_disk_fwnode+0xfb/0x480
       nbd_dev_add+0x72c/0xb50
       nbd_init+0x168/0x1f0
       do_one_initcall+0x250/0x870
       do_initcall_level+0x104/0x190
       do_initcalls+0x59/0xa0
       kernel_init_freeable+0x2a6/0x3e0
       kernel_init+0x1d/0x1d0
       ret_from_fork+0x514/0xb70
       ret_from_fork_asm+0x1a/0x30

-> #3 (&q->elevator_lock){+.+.}-{4:4}:
       __mutex_lock+0x1a3/0x1550
       elevator_change+0x1b3/0x450
       elevator_set_none+0xb5/0x140
       blk_mq_update_nr_hw_queues+0x5e7/0x1a60
       nbd_start_device+0x17f/0xb10
       nbd_genl_connect+0x165b/0x1cf0
       genl_family_rcv_msg_doit+0x22a/0x330
       genl_rcv_msg+0x61c/0x7a0
       netlink_rcv_skb+0x232/0x4b0
       genl_rcv+0x28/0x40
       netlink_unicast+0x75c/0x8e0
       netlink_sendmsg+0x813/0xb40
       ____sys_sendmsg+0x972/0x9f0
       ___sys_sendmsg+0x2a5/0x360
       __x64_sys_sendmsg+0x1bd/0x2a0
       do_syscall_64+0x15f/0xf80
       entry_SYSCALL_64_after_hwframe+0x77/0x7f

-> #2 (&q->q_usage_counter(io)#49){++++}-{0:0}:
       blk_alloc_queue+0x546/0x680
       __blk_mq_alloc_disk+0x197/0x390
       nbd_dev_add+0x499/0xb50
       nbd_init+0x168/0x1f0
       do_one_initcall+0x250/0x870
       do_initcall_level+0x104/0x190
       do_initcalls+0x59/0xa0
       kernel_init_freeable+0x2a6/0x3e0
       kernel_init+0x1d/0x1d0
       ret_from_fork+0x514/0xb70
       ret_from_fork_asm+0x1a/0x30

-> #1 (fs_reclaim){+.+.}-{0:0}:
       fs_reclaim_acquire+0x71/0x100
       kmem_cache_alloc_noprof+0x40/0x650
       __anon_vma_prepare+0xcb/0x4a0
       __vmf_anon_prepare+0xe1/0x1b0
       hugetlb_no_page+0x399/0x2100
       hugetlb_fault+0x747/0x1510
       handle_mm_fault+0x2007/0x3170
       do_user_addr_fault+0xa73/0x1340
       exc_page_fault+0x6a/0xc0
       asm_exc_page_fault+0x26/0x30

-> #0 (&hugetlb_fault_mutex_table[i]){+.+.}-{4:4}:
       __lock_acquire+0x15a5/0x2cf0
       lock_acquire+0x106/0x350
       __mutex_lock+0x1a3/0x1550
       hugetlb_fault+0x3d8/0x1510
       handle_mm_fault+0x2007/0x3170
       do_user_addr_fault+0x75b/0x1340
       exc_page_fault+0x6a/0xc0
       asm_exc_page_fault+0x26/0x30
       rep_movs_alternative+0x4a/0x90
       _copy_from_user+0x7a/0xb0
       do_mcast_group_source+0x261/0x550
       do_ip_setsockopt+0x1af6/0x2ea0
       ip_setsockopt+0x66/0x110
       do_sock_setsockopt+0x17c/0x1b0
       __x64_sys_setsockopt+0x13d/0x1b0
       do_syscall_64+0x15f/0xf80
       entry_SYSCALL_64_after_hwframe+0x77/0x7f

other info that might help us debug this:

Chain exists of:
  &hugetlb_fault_mutex_table[i] --> sk_lock-AF_INET6 --> &mm->mmap_lock

 Possible unsafe locking scenario:

       CPU0                    CPU1
       ----                    ----
  rlock(&mm->mmap_lock);
                               lock(sk_lock-AF_INET6);
                               lock(&mm->mmap_lock);
  lock(&hugetlb_fault_mutex_table[i]);

 *** DEADLOCK ***

3 locks held by syz.2.10971/12057:
 #0: ffffffff8fdcf900 (rtnl_mutex){+.+.}-{4:4}, at: do_ip_setsockopt+0xefa/0x2ea0
 #1: ffff8881a2775060 (sk_lock-AF_INET){+.+.}-{0:0}, at: do_ip_setsockopt+0x1623/0x2ea0
 #2: ffff888023a4c078 (&mm->mmap_lock){++++}-{4:4}, at: lock_mm_and_find_vma+0x36/0x340

stack backtrace:
CPU: 0 UID: 0 PID: 12057 Comm: syz.2.10971 Not tainted syzkaller #0 PREEMPT(full) 
Hardware name: QEMU Standard PC (Q35 + ICH9, 2009), BIOS 1.16.2-debian-1.16.2-1 04/01/2014
Call Trace:
 <TASK>
 dump_stack_lvl+0xe8/0x150
 print_circular_bug+0x2e1/0x300
 check_noncircular+0x12e/0x150
 __lock_acquire+0x15a5/0x2cf0
 lock_acquire+0x106/0x350
 __mutex_lock+0x1a3/0x1550
 hugetlb_fault+0x3d8/0x1510
 handle_mm_fault+0x2007/0x3170
 do_user_addr_fault+0x75b/0x1340
 exc_page_fault+0x6a/0xc0
 asm_exc_page_fault+0x26/0x30
RIP: 0010:rep_movs_alternative+0x4a/0x90
Code: cc cc cc 66 2e 0f 1f 84 00 00 00 00 00 0f 1f 00 48 8b 06 48 89 07 48 83 c6 08 48 83 c7 08 83 e9 08 74 db 83 f9 08 73 e8 eb c5 <f3> a4 c3 cc cc cc cc 48 8b 06 48 89 07 48 8d 47 08 48 83 e0 f8 48
RSP: 0018:ffffc9000acd78d8 EFLAGS: 00050202
RAX: 00007ffffffff001 RBX: 0000000000000108 RCX: 0000000000000108
RDX: 0000000000000001 RSI: 0000200000000240 RDI: ffffc9000acd7a90
RBP: ffffc9000acd7c70 R08: ffffc9000acd7b97 R09: 1ffff9200159af72
R10: dffffc0000000000 R11: fffff5200159af73 R12: f8f8f8f8f8f8f8f8
R13: ffff8881a3184a10 R14: ffffc9000acd7a90 R15: 0000200000000240
 _copy_from_user+0x7a/0xb0
 do_mcast_group_source+0x261/0x550
 do_ip_setsockopt+0x1af6/0x2ea0
 ip_setsockopt+0x66/0x110
 do_sock_setsockopt+0x17c/0x1b0
 __x64_sys_setsockopt+0x13d/0x1b0
 do_syscall_64+0x15f/0xf80
 entry_SYSCALL_64_after_hwframe+0x77/0x7f
RIP: 0033:0x7fa030d9cdd9
Code: ff c3 66 2e 0f 1f 84 00 00 00 00 00 0f 1f 44 00 00 48 89 f8 48 89 f7 48 89 d6 48 89 ca 4d 89 c2 4d 89 c8 4c 8b 4c 24 08 0f 05 <48> 3d 01 f0 ff ff 73 01 c3 48 c7 c1 e8 ff ff ff f7 d8 64 89 01 48
RSP: 002b:00007fa031c06028 EFLAGS: 00000246 ORIG_RAX: 0000000000000036
RAX: ffffffffffffffda RBX: 00007fa031015fa0 RCX: 00007fa030d9cdd9
RDX: 000000000000002c RSI: 0000000000000000 RDI: 0000000000000008
RBP: 00007fa030e32d69 R08: 0000000000000108 R09: 0000000000000000
R10: 0000200000000240 R11: 0000000000000246 R12: 0000000000000000
R13: 00007fa031016038 R14: 00007fa031015fa0 R15: 00007ffd97889bc8
 </TASK>
