==================================================================
BUG: KASAN: slab-use-after-free in txEnd+0x32d/0x530
Write of size 8 at addr ffff88816d158040 by task jfsCommit/115

CPU: 1 UID: 0 PID: 115 Comm: jfsCommit Not tainted syzkaller #0 PREEMPT(full) 
Hardware name: QEMU Standard PC (Q35 + ICH9, 2009), BIOS 1.16.2-debian-1.16.2-1 04/01/2014
Call Trace:
 <TASK>
 dump_stack_lvl+0x189/0x250
 print_report+0xca/0x240
 kasan_report+0x118/0x150
 kasan_check_range+0x2b0/0x2c0
 txEnd+0x32d/0x530
 jfs_lazycommit+0x5ba/0xa90
 kthread+0x711/0x8a0
 ret_from_fork+0x4bc/0x870
 ret_from_fork_asm+0x1a/0x30
 </TASK>

Allocated by task 8071:
 kasan_save_track+0x3e/0x80
 __kasan_kmalloc+0x93/0xb0
 __kmalloc_cache_noprof+0x3e2/0x700
 lmLogOpen+0x2d1/0xfa0
 jfs_mount_rw+0xe9/0x670
 jfs_fill_super+0x754/0xd80
 get_tree_bdev_flags+0x40e/0x4d0
 vfs_get_tree+0x92/0x2b0
 do_new_mount+0x302/0xa10
 __se_sys_mount+0x313/0x410
 do_syscall_64+0xfa/0xfa0
 entry_SYSCALL_64_after_hwframe+0x77/0x7f

Freed by task 6530:
 kasan_save_track+0x3e/0x80
 kasan_save_free_info+0x46/0x50
 __kasan_slab_free+0x5c/0x80
 kfree+0x19a/0x6d0
 lmLogClose+0x297/0x520
 jfs_umount+0x2ef/0x3c0
 jfs_put_super+0x8c/0x190
 generic_shutdown_super+0x135/0x2c0
 kill_block_super+0x44/0x90
 deactivate_locked_super+0xbc/0x130
 cleanup_mnt+0x425/0x4c0
 task_work_run+0x1d4/0x260
 exit_to_user_mode_loop+0xe9/0x130
 do_syscall_64+0x2bd/0xfa0
 entry_SYSCALL_64_after_hwframe+0x77/0x7f

The buggy address belongs to the object at ffff88816d158000
 which belongs to the cache kmalloc-1k of size 1024
The buggy address is located 64 bytes inside of
 freed 1024-byte region [ffff88816d158000, ffff88816d158400)

The buggy address belongs to the physical page:
page: refcount:0 mapcount:0 mapping:0000000000000000 index:0x0 pfn:0x16d158
head: order:3 mapcount:0 entire_mapcount:0 nr_pages_mapped:0 pincount:0
flags: 0x57ff00000000040(head|node=1|zone=2|lastcpupid=0x7ff)
page_type: f5(slab)
raw: 057ff00000000040 ffff888100041dc0 ffffea0005a74e00 dead000000000002
raw: 0000000000000000 0000000000100010 00000000f5000000 0000000000000000
head: 057ff00000000040 ffff888100041dc0 ffffea0005a74e00 dead000000000002
head: 0000000000000000 0000000000100010 00000000f5000000 0000000000000000
head: 057ff00000000003 ffffea0005b45601 00000000ffffffff 00000000ffffffff
head: ffffffffffffffff 0000000000000000 00000000ffffffff 0000000000000008
page dumped because: kasan: bad access detected
page_owner tracks the page as allocated
page last allocated via order 3, migratetype Unmovable, gfp_mask 0x52820(GFP_ATOMIC|__GFP_NOWARN|__GFP_NORETRY|__GFP_COMP), pid 27, tgid 27 (kworker/u10:0), ts 120068204222, free_ts 119986330028
 post_alloc_hook+0x240/0x2a0
 get_page_from_freelist+0x2365/0x2440
 __alloc_frozen_pages_noprof+0x181/0x370
 alloc_pages_mpol+0x232/0x4a0
 allocate_slab+0x96/0x350
 ___slab_alloc+0xe94/0x18a0
 __slab_alloc+0x65/0x100
 __kmalloc_noprof+0x47d/0x800
 ieee802_11_parse_elems_full+0x152/0x2ab0
 ieee80211_ibss_rx_queued_mgmt+0x47a/0x2af0
 ieee80211_iface_work+0x85f/0x12d0
 cfg80211_wiphy_work+0x2bb/0x470
 process_scheduled_works+0xae1/0x17b0
 worker_thread+0x8a0/0xda0
 kthread+0x711/0x8a0
 ret_from_fork+0x4bc/0x870
page last free pid 5872 tgid 5872 stack trace:
 __free_frozen_pages+0xbc8/0xd30
 __put_partials+0x146/0x170
 put_cpu_partial+0x1f2/0x2e0
 __slab_free+0x2b9/0x390
 qlist_free_all+0x97/0x140
 kasan_quarantine_reduce+0x148/0x160
 __kasan_kmalloc+0x22/0xb0
 __kmalloc_noprof+0x41d/0x800
 blk_rq_map_kern+0x2a7/0x650
 scsi_execute_cmd+0x2fb/0x1130
 sr_check_events+0x1b1/0xb30
 cdrom_check_events+0x64/0xe0
 disk_check_events+0x117/0x5b0
 process_scheduled_works+0xae1/0x17b0
 worker_thread+0x8a0/0xda0
 kthread+0x711/0x8a0

Memory state around the buggy address:
 ffff88816d157f00: fc fc fc fc fc fc fc fc fc fc fc fc fc fc fc fc
 ffff88816d157f80: fc fc fc fc fc fc fc fc fc fc fc fc fc fc fc fc
>ffff88816d158000: fa fb fb fb fb fb fb fb fb fb fb fb fb fb fb fb
                                           ^
 ffff88816d158080: fb fb fb fb fb fb fb fb fb fb fb fb fb fb fb fb
 ffff88816d158100: fb fb fb fb fb fb fb fb fb fb fb fb fb fb fb fb
==================================================================
