2026/01/30 05:06:47 extracted 325008 text symbol hashes for base and 325010 for patched 2026/01/30 05:06:47 symbol "__do_trace_mm_compaction_begin.__UNIQUE_ID_addressable___SCK__tp_func_mm_compaction_begin_1434" has different values in base vs patch 2026/01/30 05:06:47 binaries are different, continuing fuzzing 2026/01/30 05:06:47 adding modified_functions to focus areas: ["___free_pages" "__access_remote_vm" "__add_to_free_list" "__alloc_frozen_pages_noprof" "__alloc_pages_cpuset_fallback" "__alloc_pages_direct_compact" "__alloc_pages_direct_reclaim" "__alloc_pages_may_oom" "__alloc_pages_slowpath" "__btrfs_prealloc_file_range" "__build_all_zonelists" "__cow_file_range_inline" "__del_page_from_free_list" "__exchange_data_block" "__ext4_get_inode_loc" "__f2fs_cluster_blocks" "__f2fs_ioctl" "__folio_alloc_noprof" "__folio_freeze_and_split_unmapped" "__folio_put" "__folio_split" "__free_contig_frozen_range" "__free_frozen_pages" "__free_one_page" "__free_pages_ok" "__get_obj_cgroup_from_memcg" "__gfs2_jdata_write_folio" "__has_merged_page" "__ia32_sys_set_mempolicy" "__iomap_write_begin" "__isolate_free_page" "__kasan_poison_slab" "__linger_cancel" "__mem_cgroup_charge" "__mem_cgroup_handle_over_high" "__mem_cgroup_try_charge_swap" "__memcg_kmem_charge_page" "__memcg_kmem_uncharge_page" "__memcg_slab_free_hook" "__memcg_slab_post_alloc_hook" "__move_freepages_block_isolate" "__node_reclaim" "__nr_hugepages_store_common" "__ocfs2_rotate_tree_left" "__offline_isolated_pages" "__p4d_alloc" "__pagetable_ctor" "__pfx___pagetable_ctor" "__pmd_alloc" "__pte_alloc" "__pte_alloc_kernel" "__pud_alloc" "__rmqueue_pcplist" "__se_sys_mbind" "__se_sys_migrate_pages" "__se_sys_move_pages" "__se_sys_set_mempolicy_home_node" "__shmem_file_setup" "__submit_request" "__swap_writepage" "__tlb_remove_table" "__unmap_hugepage_range" "__vm_insert_mixed" "__vma_reservation_common" "__write_extent_buffer" "__x64_sys_set_mempolicy" "abort_on_full_fn" "adjust_managed_page_count" "adjust_pool_surplus" "affs_read_folio_ofs" "affs_write_begin" "alloc_contig_frozen_range_noprof" "alloc_contig_pages_noprof" "alloc_contig_range_noprof" "alloc_dummy_extent_buffer" "alloc_extent_buffer" "alloc_frozen_pages_nolock_noprof" "alloc_frozen_pages_noprof" "alloc_hugetlb_folio" "alloc_migration_target_by_mpol" "alloc_pages_bulk_mempolicy_noprof" "alloc_pages_exact_nid_noprof" "alloc_pages_nolock_noprof" "alloc_pages_noprof" "alloc_surplus_hugetlb_folio" "bio_associate_blkg_from_page" "bl_read_pagelist" "btrfs_add_link" "btrfs_begin_enable_verity" "btrfs_calc_input_length" "btrfs_clone_extent_buffer" "btrfs_compr_pool_scan" "btrfs_compress_heuristic" "btrfs_cont_expand" "btrfs_create_new_inode" "btrfs_do_readpage" "btrfs_end_enable_verity" "btrfs_finish_one_ordered" "btrfs_get_extent" "btrfs_iget" "btrfs_link" "btrfs_lookup_dentry" "btrfs_page_mkwrite" "btrfs_read_locked_inode" "btrfs_read_merkle_tree_page" "btrfs_release_extent_buffer_folios" "btrfs_rename2" "btrfs_rmdir" "btrfs_swap_activate" "btrfs_symlink" "btrfs_unlink" "btrfs_update_inode_item" "btrfs_writepages" "calc_target" "can_change_pmd_writable" "ceph_copy_file_range" "ceph_fallocate" "ceph_filemap_fault" "ceph_invalidate_folio" "ceph_netfs_check_write_begin" "ceph_netfs_issue_read" "ceph_osdc_call" "ceph_osdc_handle_map" "ceph_osdc_list_watchers" "ceph_osdc_notify" "ceph_osdc_notify_ack" "ceph_osdc_start_request" "ceph_osdc_unwatch" "ceph_osdc_watch" "ceph_write_iter" "ceph_writepages_start" "ceph_zero_partial_page" "charge_memcg" "check_linger_pool_dne" "check_move_unevictable_folios" "check_new_pages" "check_pool_dne" "clear_refs_write" "clear_shadow_entries" "close_osd" "compact_node" "compaction_alloc" "compaction_proactiveness_sysctl_handler" "compress_file_range" "copy_extent_buffer" "copy_huge_pmd" "copy_page_range" "copy_pmd_range" "copy_remote_vm_str" "count_memcg_folio_events" "count_swpout_vm_event" "cow_file_range" "current_objcg_update" "default_hugepagesz_setupargs" "deferred_split_folio" "deferred_split_scan" "do_huge_pmd_anonymous_page" "do_huge_pmd_wp_page" "do_migrate_pages" "do_pte_missing" "do_swap_page" "do_try_to_free_pages" "do_watch_error" "do_watch_notify" "do_wp_page" "drain_local_memcg_stock" "drain_local_obj_stock" "drain_obj_stock" "end_bbio_data_read" "end_bbio_meta_write" "end_folio_read" "evict_folios" "ext4_bio_write_folio" "ext4_block_zero_page_range" "ext4_da_get_block_prep" "ext4_da_write_end" "ext4_dax_writepages" "ext4_do_writepages" "ext4_fill_raw_inode" "ext4_iomap_begin" "ext4_iomap_begin_report" "ext4_journalled_write_end" "ext4_journalled_zero_new_buffers" "ext4_mark_iloc_dirty" "ext4_move_extents" "ext4_page_mkwrite" "ext4_readpage_inline" "ext4_release_io_end" "ext4_try_add_inline_entry" "ext4_wait_for_tail_page_commit" "ext4_writepages" "extent_buffer_get_byte" "extent_writepage_io" "f2fs_alloc_dic" "f2fs_compat_ioc_gc_range" "f2fs_compat_ioc_move_range" "f2fs_compat_ioctl" "f2fs_compress_write_end" "f2fs_decompress_end_io" "f2fs_defragment_range" "f2fs_encrypt_one_page" "f2fs_end_read_compressed_page" "f2fs_fallocate" "f2fs_fiemap" "f2fs_file_fadvise" "f2fs_file_read_iter" "f2fs_file_splice_read" "f2fs_file_write_iter" "f2fs_fileattr_set" "f2fs_get_new_data_folio" "f2fs_get_read_data_folio" "f2fs_ioc_start_atomic_write" "f2fs_ioctl" "f2fs_map_blocks" "f2fs_mpage_readpages" "f2fs_prepare_compress_overwrite" "f2fs_prepare_decomp_mem" "f2fs_put_rpages_wbc" "f2fs_read_inline_data" "f2fs_read_multi_pages" "f2fs_reserve_block" "f2fs_submit_page_write" "f2fs_truncate_partial_cluster" "f2fs_vm_page_mkwrite" "f2fs_write_begin" "f2fs_write_data_pages" "f2fs_write_multi_pages" "f2fs_write_single_data_page" "f2fs_zero_range" "f2fs_zone_write_end_io" "fallback_to_cow" "fill_zero" "find_lock_delalloc_range" "flush_memcg_stats_dwork" "folio_alloc_noprof" "folio_isolate_lru" "folio_split_memcg_refs" "folio_split_unmapped" "folio_try_dup_anon_rmap_pmd" "folio_unmap_invalidate" "folio_unqueue_deferred_split" "folio_zero_range" "folio_zero_segment" "folios_put_refs" "follow_pfnmap_start" "free_contig_frozen_range" "free_frozen_page_commit" "free_hpage_workfn" "free_one_page" "free_page_is_bad" "free_pages_prepare" "free_pcppages_bulk" "free_tail_page_prepare" "free_unref_folios" "fuse_copy_folio" "fuse_dev_do_read" "fuse_dev_do_write" "fuse_dev_read" "fuse_dev_splice_read" "fuse_dev_splice_write" "fuse_dev_write" "fuse_read_forget" "fuse_read_interrupt" "get_mem_cgroup_from_current" "get_mem_cgroup_from_folio" "get_mem_cgroup_from_mm" "get_mem_cgroup_from_objcg" "get_obj_cgroup_from_folio" "get_page_from_freelist" "get_pte_pfn" "gfs2_qd_dispose" "gfs2_qd_shrink_scan" "gfs2_quota_cleanup" "gfs2_write_buf_to_page" "gfs2_write_cache_jdata" "handle_mm_fault" "handle_pte_marker" "handle_timeout" "hfs_btree_open" "high_work_func" "hugepages_setupargs" "hugepagesz_setupargs" "hugetlb_acct_memory" "hugetlb_add_to_page_cache" "hugetlb_cma_alloc_frozen_folio" "hugetlb_fault" "hugetlb_folio_init_vmemmap" "hugetlb_hstate_alloc_pages" "hugetlb_no_page" "hugetlb_resv_map_add" "hugetlbfs_evict_inode" "hugetlbfs_zero_partial_page" "inc_valid_block_count" "insert_page" "iomap_file_buffered_write" "iomap_file_unshare" "iomap_write_begin" "iomap_write_delalloc_release" "iomap_write_end" "iomap_writeback_folio" "iomap_zero_range" "isolate_single_pageblock" "kcompactd" "kick_osd_requests" "kswapd" "linger_commit_cb" "linger_map_check_cb" "linger_ping_cb" "linger_reconnect_cb" "linger_submit" "link_linger" "lowmem_reserve_ratio_sysctl_handler" "lru_gen_seq_write" "madvise_free_huge_pmd" "make_alloc_exact" "map_check_cb" "mem_cgroup_charge_hugetlb" "mem_cgroup_css_free" "mem_cgroup_css_offline" "mem_cgroup_css_online" "mem_cgroup_exit" "mem_cgroup_get_from_id" "mem_cgroup_get_oom_group" "mem_cgroup_iter" "mem_cgroup_iter_break" "mem_cgroup_oom_synchronize" "mem_cgroup_replace_folio" "mem_cgroup_scan_tasks" "mem_cgroup_swap_full" "mem_cgroup_swapin_charge_folio" "memcg1_check_events" "memcg1_swapout" "memcg1_uncharge_batch" "memcg_event_remove" "memcg_write_event_control" "memcmp_extent_buffer" "memcpy_extent_buffer" "memmap_init_compound" "memmove_extent_buffer" "memory_numa_stat_show" "mempolicy_slab_node" "memzero_extent_buffer" "migrate_device_unmap" "migrate_folio_add" "migrate_folio_done" "migrate_pages" "migrate_pages_batch" "mm_get_huge_zero_folio" "mod_objcg_mlstate" "move_folios_to_lru" "move_to_free_list" "netfs_page_mkwrite" "netfs_perform_write" "netfs_prefetch_for_write" "netfs_read_to_pagecache" "netfs_write_begin" "nfs_read_completion" "nfs_write_begin" "nfs_write_end" "ni_readpage_cmpr" "ntfs_compress_write" "ntfs_fallocate" "ntfs_file_write_iter" "ntfs_get_block_bmap" "ntfs_get_block_direct_IO_R" "ntfs_get_block_direct_IO_W" "ntfs_get_link" "ntfs_zero_range" "numa_migrate_check" "obj_cgroup_charge_account" "obj_cgroup_charge_pages" "obj_cgroup_charge_zswap" "obj_cgroup_may_zswap" "obj_cgroup_release" "obj_cgroup_uncharge_zswap" "ocfs2_append_rec_to_path" "ocfs2_commit_truncate" "ocfs2_convert_inline_data_to_extents" "ocfs2_dio_wr_get_block" "ocfs2_do_insert_extent" "ocfs2_grow_tree" "ocfs2_insert_extent" "ocfs2_map_and_dirty_folio" "ocfs2_merge_rec_left" "ocfs2_merge_rec_right" "ocfs2_remove_rightmost_path" "ocfs2_rotate_tree_right" "ocfs2_split_extent" "ocfs2_truncate_inline" "ocfs2_truncate_rec" "ocfs2_try_to_merge_extent" "ocfs2_write_begin_nolock" "only_alloc_fresh_hugetlb_folio" "osd_dispatch" "osd_get_con" "osd_sparse_read" "page_alloc_cpu_dead" "page_array_alloc" "page_rmappable_folio" "pagemap_pmd_range" "pagemap_read" "pagetable_dtor_free" "percpu_pagelist_high_fraction_sysctl_handler" "perf_trace_mm_lru_insertion" "pgd_alloc" "pmd_free_pte_page" "policy_nodemask" "prep_and_add_bootmem_folios" "prep_compound_page" "prepare_kswapd_sleep" "pte_alloc_one" "pte_free" "pte_free_kernel" "pte_free_now" "qd_put" "queue_folios_hugetlb" "read_extent_buffer" "read_inline_extent" "refill_obj_stock" "region_add" "region_chg" "release_compress_blocks" "release_extent_buffer" "remove_device_exclusive_entry" "remove_pool_hugetlb_folio" "reserve_highatomic_pageblock" "restore_reserve_on_error" "run_delalloc_nocow" "secretmem_free_folio" "send_linger" "send_request" "set_pageblock_migratetype" "set_pmd_migration_entry" "setup_per_zone_lowmem_reserve" "setup_per_zone_wmarks" "shmem_alloc_and_add_folio" "shmem_enabled_store" "shmem_evict_inode" "shmem_fallocate" "shmem_fault" "shmem_file_read_iter" "shmem_file_splice_read" "shmem_get_folio_gfp" "shmem_get_link" "shmem_getattr" "shmem_symlink" "shmem_write_begin" "shmem_write_end" "shmem_writeout" "should_reclaim_retry" "show_smaps_rollup" "shrink_active_list" "shrink_folio_list" "shrink_lruvec" "shrink_node" "sio_read_complete" "smaps_account" "smaps_hugetlb_range" "split_folio_to_list" "split_free_frozen_pages" "split_huge_pages_all" "split_huge_pages_in_file" "split_huge_pages_write" "split_huge_pmd_locked" "split_large_buddy" "split_page_memcg" "squashfs_read_folio" "start_isolate_page_range" "submit_extent_folio" "submit_uncompressed_range" "swap_read_folio" "swap_writeout" "sysctl_compaction_handler" "sysctl_min_slab_ratio_sysctl_handler" "sysctl_min_unmapped_ratio_sysctl_handler" "tlb_remove_table_rcu" "trace_event_raw_event_mm_lru_insertion" "truncate_cleanup_folio" "truncate_folio_batch_exceptionals" "try_folio_split_or_unmap" "try_restore_exclusive_pte" "try_to_free_pages" "ubifs_write_begin" "ubifs_write_end" "uncharge_batch" "uncharge_folio" "uncompress_inline" "unlock_delalloc_folio" "unmap_huge_pmd_locked" "unmap_page_range" "update_and_free_pages_bulk" "virtio_fs_probe" "virtio_fs_request_complete" "vm_insert_pages" "vma_alloc_folio_noprof" "vmemmap_pmd_entry" "walk_pmd_range_locked" "walk_pud_range" "walk_update_folio" "weighted_interleave_auto_store" "weighted_interleave_nodes" "write_begin_slow" "write_one_eb" "writepage_delalloc" "z_erofs_decompress_queue" "z_erofs_read_folio" "z_erofs_readahead" "z_erofs_scan_folio" "zap_huge_pmd" "zero_pmd_populate" "zero_user_segment" "zero_user_segments" "zlib_compress_folios" "zone_pcp_disable" "zswap_current_read"] 2026/01/30 05:06:47 adding directly modified files to focus areas: ["include/linux/mm.h" "io_uring/memmap.c" "mm/huge_memory.c" "mm/hugetlb.c" "mm/hugetlb_cma.c" "mm/internal.h" "mm/mm_init.c" "mm/page_alloc.c"] 2026/01/30 05:06:47 downloading corpus #1: "https://storage.googleapis.com/syzkaller/corpus/ci-upstream-kasan-gce-root-corpus.db" 2026/01/30 05:07:47 runner 6 connected 2026/01/30 05:07:47 runner 1 connected 2026/01/30 05:07:47 runner 8 connected 2026/01/30 05:07:48 runner 5 connected 2026/01/30 05:07:48 runner 3 connected 2026/01/30 05:07:48 runner 2 connected 2026/01/30 05:07:48 runner 1 connected 2026/01/30 05:07:48 runner 2 connected 2026/01/30 05:07:48 runner 0 connected 2026/01/30 05:07:48 runner 0 connected 2026/01/30 05:07:48 runner 4 connected 2026/01/30 05:07:53 initializing coverage information... 2026/01/30 05:07:54 executor cover filter: 0 PCs 2026/01/30 05:07:56 runner 7 connected 2026/01/30 05:07:57 machine check: disabled the following syscalls: openat$acpi_thermal_rel : failed to open /dev/acpi_thermal_rel: no such file or directory openat$ashmem : failed to open /dev/ashmem: no such file or directory openat$bifrost : failed to open /dev/bifrost: no such file or directory openat$binder : failed to open /dev/binder: no such file or directory openat$camx : failed to open /dev/v4l/by-path/platform-soc@0:qcom_cam-req-mgr-video-index0: no such file or directory openat$capi20 : failed to open /dev/capi20: no such file or directory openat$cdrom1 : failed to open /dev/cdrom1: no such file or directory openat$damon_attrs : failed to open /sys/kernel/debug/damon/attrs: no such file or directory openat$damon_init_regions : failed to open /sys/kernel/debug/damon/init_regions: no such file or directory openat$damon_kdamond_pid : failed to open /sys/kernel/debug/damon/kdamond_pid: no such file or directory openat$damon_mk_contexts : failed to open /sys/kernel/debug/damon/mk_contexts: no such file or directory openat$damon_monitor_on : failed to open /sys/kernel/debug/damon/monitor_on: no such file or directory openat$damon_rm_contexts : failed to open /sys/kernel/debug/damon/rm_contexts: no such file or directory openat$damon_schemes : failed to open /sys/kernel/debug/damon/schemes: no such file or directory openat$damon_target_ids : failed to open /sys/kernel/debug/damon/target_ids: no such file or directory openat$hwbinder : failed to open /dev/hwbinder: no such file or directory openat$i915 : failed to open /dev/i915: no such file or directory openat$img_rogue : failed to open /dev/img-rogue: no such file or directory openat$irnet : failed to open /dev/irnet: no such file or directory openat$keychord : failed to open /dev/keychord: no such file or directory openat$kvm : failed to open /dev/kvm: no such file or directory openat$lightnvm : failed to open /dev/lightnvm/control: no such file or directory openat$mali : failed to open /dev/mali0: no such file or directory openat$md : failed to open /dev/md0: no such file or directory openat$msm : failed to open /dev/msm: no such file or directory openat$ndctl0 : failed to open /dev/ndctl0: no such file or directory openat$nmem0 : failed to open /dev/nmem0: no such file or directory openat$pktcdvd : failed to open /dev/pktcdvd/control: no such file or directory openat$pmem0 : failed to open /dev/pmem0: no such file or directory openat$proc_capi20 : failed to open /proc/capi/capi20: no such file or directory openat$proc_capi20ncci : failed to open /proc/capi/capi20ncci: no such file or directory openat$proc_reclaim : failed to open /proc/self/reclaim: no such file or directory openat$ptp1 : failed to open /dev/ptp1: no such file or directory openat$rnullb : failed to open /dev/rnullb0: no such file or directory openat$selinux_access : failed to open /selinux/access: no such file or directory openat$selinux_attr : selinux is not enabled openat$selinux_avc_cache_stats : failed to open /selinux/avc/cache_stats: no such file or directory openat$selinux_avc_cache_threshold : failed to open /selinux/avc/cache_threshold: no such file or directory openat$selinux_avc_hash_stats : failed to open /selinux/avc/hash_stats: no such file or directory openat$selinux_checkreqprot : failed to open /selinux/checkreqprot: no such file or directory openat$selinux_commit_pending_bools : failed to open /selinux/commit_pending_bools: no such file or directory openat$selinux_context : failed to open /selinux/context: no such file or directory openat$selinux_create : failed to open /selinux/create: no such file or directory openat$selinux_enforce : failed to open /selinux/enforce: no such file or directory openat$selinux_load : failed to open /selinux/load: no such file or directory openat$selinux_member : failed to open /selinux/member: no such file or directory openat$selinux_mls : failed to open /selinux/mls: no such file or directory openat$selinux_policy : failed to open /selinux/policy: no such file or directory openat$selinux_relabel : failed to open /selinux/relabel: no such file or directory openat$selinux_status : failed to open /selinux/status: no such file or directory openat$selinux_user : failed to open /selinux/user: no such file or directory openat$selinux_validatetrans : failed to open /selinux/validatetrans: no such file or directory openat$sev : failed to open /dev/sev: no such file or directory openat$sgx_provision : failed to open /dev/sgx_provision: no such file or directory openat$smack_task_current : smack is not enabled openat$smack_thread_current : smack is not enabled openat$smackfs_access : failed to open /sys/fs/smackfs/access: no such file or directory openat$smackfs_ambient : failed to open /sys/fs/smackfs/ambient: no such file or directory openat$smackfs_change_rule : failed to open /sys/fs/smackfs/change-rule: no such file or directory openat$smackfs_cipso : failed to open /sys/fs/smackfs/cipso: no such file or directory openat$smackfs_cipsonum : failed to open /sys/fs/smackfs/direct: no such file or directory openat$smackfs_ipv6host : failed to open /sys/fs/smackfs/ipv6host: no such file or directory openat$smackfs_load : failed to open /sys/fs/smackfs/load: no such file or directory openat$smackfs_logging : failed to open /sys/fs/smackfs/logging: no such file or directory openat$smackfs_netlabel : failed to open /sys/fs/smackfs/netlabel: no such file or directory openat$smackfs_onlycap : failed to open /sys/fs/smackfs/onlycap: no such file or directory openat$smackfs_ptrace : failed to open /sys/fs/smackfs/ptrace: no such file or directory openat$smackfs_relabel_self : failed to open /sys/fs/smackfs/relabel-self: no such file or directory openat$smackfs_revoke_subject : failed to open /sys/fs/smackfs/revoke-subject: no such file or directory openat$smackfs_syslog : failed to open /sys/fs/smackfs/syslog: no such file or directory openat$smackfs_unconfined : failed to open /sys/fs/smackfs/unconfined: no such file or directory openat$tlk_device : failed to open /dev/tlk_device: no such file or directory openat$trusty : failed to open /dev/trusty-ipc-dev0: no such file or directory openat$trusty_avb : failed to open /dev/trusty-ipc-dev0: no such file or directory openat$trusty_gatekeeper : failed to open /dev/trusty-ipc-dev0: no such file or directory openat$trusty_hwkey : failed to open /dev/trusty-ipc-dev0: no such file or directory openat$trusty_hwrng : failed to open /dev/trusty-ipc-dev0: no such file or directory openat$trusty_km : failed to open /dev/trusty-ipc-dev0: no such file or directory openat$trusty_km_secure : failed to open /dev/trusty-ipc-dev0: no such file or directory openat$trusty_storage : failed to open /dev/trusty-ipc-dev0: no such file or directory openat$tty : failed to open /dev/tty: no such device or address openat$uverbs0 : failed to open /dev/infiniband/uverbs0: no such file or directory openat$vfio : failed to open /dev/vfio/vfio: no such file or directory openat$vndbinder : failed to open /dev/vndbinder: no such file or directory openat$vtpm : failed to open /dev/vtpmx: no such file or directory openat$xenevtchn : failed to open /dev/xen/evtchn: no such file or directory openat$zygote : failed to open /dev/socket/zygote: no such file or directory socket$hf : socket$hf(0x13, 0x2, 0x0) failed: address family not supported by protocol socket$inet6_dccp : socket$inet6_dccp(0xa, 0x6, 0x0) failed: socket type not supported socket$inet_dccp : socket$inet_dccp(0x2, 0x6, 0x0) failed: socket type not supported socket$vsock_dgram : socket$vsock_dgram(0x28, 0x2, 0x0) failed: no such device transitively disabled the following syscalls (missing resource [creating syscalls]): accept$ax25 : sock_ax25 [accept$ax25 accept4$ax25 syz_init_net_socket$ax25] accept$netrom : sock_netrom [accept$netrom accept4$netrom syz_init_net_socket$netrom] accept$nfc_llcp : sock_nfc_llcp [accept$nfc_llcp accept4$nfc_llcp syz_init_net_socket$nfc_llcp] close$binfmt : fd_binfmt [openat$binfmt] close$fd_v4l2_buffer : fd_v4l2_buffer [ioctl$VIDIOC_QUERYBUF_DMABUF] close$ibv_device : fd_rdma [openat$uverbs0] mmap$DRM_I915 : fd_i915 [openat$i915] mmap$DRM_MSM : fd_msm [openat$msm] mmap$KVM_VCPU : vcpu_mmap_size [ioctl$KVM_GET_VCPU_MMAP_SIZE] mmap$bifrost : fd_bifrost [openat$bifrost openat$mali] mmap$perf : fd_perf [perf_event_open perf_event_open$cgroup] mmap$snddsp : fd_snd_dsp [syz_open_dev$sndpcmc syz_open_dev$sndpcmp] mmap$snddsp_control : fd_snd_dsp [syz_open_dev$sndpcmc syz_open_dev$sndpcmp] mmap$snddsp_status : fd_snd_dsp [syz_open_dev$sndpcmc syz_open_dev$sndpcmp] mmap$usbfs : fd_usbfs [syz_open_dev$usbfs] mmap$usbmon : fd_usbmon [syz_open_dev$usbmon] openat$binfmt : ptr_binfmt_file [syz_create_resource$binfmt] setsockopt$EBT_SO_SET_ENTRIES : uid [fstat fstat$auto geteuid ...] setsockopt$IP6T_SO_SET_REPLACE : fd_bpf_prog [bpf$BPF_PROG_GET_FD_BY_ID bpf$BPF_PROG_RAW_TRACEPOINT_LOAD bpf$BPF_PROG_WITH_BTFID_LOAD ...] setsockopt$IPT_SO_SET_REPLACE : fd_bpf_prog [bpf$BPF_PROG_GET_FD_BY_ID bpf$BPF_PROG_RAW_TRACEPOINT_LOAD bpf$BPF_PROG_WITH_BTFID_LOAD ...] setsockopt$SO_VM_SOCKETS_CONNECT_TIMEOUT_OLD: time_usec [getitimer getrusage getsockopt$sock_timeval ...] setsockopt$WPAN_SECURITY : sock_802154_dgram [syz_init_net_socket$802154_dgram] setsockopt$WPAN_SECURITY_LEVEL : sock_802154_dgram [syz_init_net_socket$802154_dgram] setsockopt$WPAN_WANTACK : sock_802154_dgram [syz_init_net_socket$802154_dgram] setsockopt$WPAN_WANTLQI : sock_802154_dgram [syz_init_net_socket$802154_dgram] setsockopt$X25_QBITINCL : sock_x25 [accept4$x25 syz_init_net_socket$x25] setsockopt$ax25_SO_BINDTODEVICE : sock_ax25 [accept$ax25 accept4$ax25 syz_init_net_socket$ax25] setsockopt$ax25_int : sock_ax25 [accept$ax25 accept4$ax25 syz_init_net_socket$ax25] setsockopt$bt_BT_CHANNEL_POLICY : sock_bt [accept4$bt_l2cap syz_init_net_socket$bt_bnep syz_init_net_socket$bt_cmtp ...] setsockopt$bt_BT_DEFER_SETUP : sock_bt [accept4$bt_l2cap syz_init_net_socket$bt_bnep syz_init_net_socket$bt_cmtp ...] setsockopt$bt_BT_FLUSHABLE : sock_bt [accept4$bt_l2cap syz_init_net_socket$bt_bnep syz_init_net_socket$bt_cmtp ...] setsockopt$bt_BT_POWER : sock_bt [accept4$bt_l2cap syz_init_net_socket$bt_bnep syz_init_net_socket$bt_cmtp ...] setsockopt$bt_BT_RCVMTU : sock_bt [accept4$bt_l2cap syz_init_net_socket$bt_bnep syz_init_net_socket$bt_cmtp ...] setsockopt$bt_BT_SECURITY : sock_bt [accept4$bt_l2cap syz_init_net_socket$bt_bnep syz_init_net_socket$bt_cmtp ...] setsockopt$bt_BT_SNDMTU : sock_bt [accept4$bt_l2cap syz_init_net_socket$bt_bnep syz_init_net_socket$bt_cmtp ...] setsockopt$bt_BT_VOICE : sock_bt [accept4$bt_l2cap syz_init_net_socket$bt_bnep syz_init_net_socket$bt_cmtp ...] setsockopt$bt_hci_HCI_DATA_DIR : sock_bt_hci [syz_init_net_socket$bt_hci] setsockopt$bt_hci_HCI_FILTER : sock_bt_hci [syz_init_net_socket$bt_hci] setsockopt$bt_hci_HCI_TIME_STAMP : sock_bt_hci [syz_init_net_socket$bt_hci] setsockopt$bt_l2cap_L2CAP_CONNINFO : sock_bt_l2cap [accept4$bt_l2cap syz_init_net_socket$bt_l2cap] setsockopt$bt_l2cap_L2CAP_LM : sock_bt_l2cap [accept4$bt_l2cap syz_init_net_socket$bt_l2cap] setsockopt$bt_l2cap_L2CAP_OPTIONS : sock_bt_l2cap [accept4$bt_l2cap syz_init_net_socket$bt_l2cap] setsockopt$bt_rfcomm_RFCOMM_LM : sock_bt_rfcomm [syz_init_net_socket$bt_rfcomm] setsockopt$inet6_IPV6_IPSEC_POLICY : uid [fstat fstat$auto geteuid ...] setsockopt$inet6_IPV6_XFRM_POLICY : uid [fstat fstat$auto geteuid ...] setsockopt$inet6_dccp_buf : sock_dccp6 [socket$inet6_dccp] setsockopt$inet6_dccp_int : sock_dccp6 [socket$inet6_dccp] setsockopt$inet_IP_IPSEC_POLICY : uid [fstat fstat$auto geteuid ...] setsockopt$inet_IP_XFRM_POLICY : uid [fstat fstat$auto geteuid ...] setsockopt$inet_dccp_buf : sock_dccp [socket$inet_dccp] setsockopt$inet_dccp_int : sock_dccp [socket$inet_dccp] setsockopt$inet_sctp6_SCTP_ADD_STREAMS : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_ASSOCINFO : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_AUTH_DEACTIVATE_KEY: assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_AUTH_DELETE_KEY : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_AUTH_KEY : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_CONTEXT : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_DEFAULT_PRINFO : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_DEFAULT_SEND_PARAM: assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_DEFAULT_SNDINFO : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_DELAYED_SACK : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_ENABLE_STREAM_RESET: assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_MAXSEG : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_MAX_BURST : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_PEER_ADDR_PARAMS : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_PEER_ADDR_THLDS : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_PRIMARY_ADDR : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_PR_SUPPORTED : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_RECONFIG_SUPPORTED: assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_RESET_ASSOC : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_RESET_STREAMS : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_RTOINFO : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_SET_PEER_PRIMARY_ADDR: assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_STREAM_SCHEDULER : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_STREAM_SCHEDULER_VALUE: assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_ADD_STREAMS : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_ASSOCINFO : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_AUTH_ACTIVE_KEY : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_AUTH_DEACTIVATE_KEY: assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_AUTH_DELETE_KEY : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_AUTH_KEY : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_CONTEXT : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_DEFAULT_PRINFO : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_DEFAULT_SEND_PARAM: assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_DEFAULT_SNDINFO : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_DELAYED_SACK : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_ENABLE_STREAM_RESET: assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_MAXSEG : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_MAX_BURST : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_PEER_ADDR_PARAMS : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_PEER_ADDR_THLDS : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_PRIMARY_ADDR : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_PR_SUPPORTED : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_RECONFIG_SUPPORTED: assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_RESET_ASSOC : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_RESET_STREAMS : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_RTOINFO : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_SET_PEER_PRIMARY_ADDR: assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_STREAM_SCHEDULER : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_STREAM_SCHEDULER_VALUE: assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$llc_int : sock_llc [accept4$llc syz_init_net_socket$llc] setsockopt$netrom_NETROM_IDLE : sock_netrom [accept$netrom accept4$netrom syz_init_net_socket$netrom] setsockopt$netrom_NETROM_N2 : sock_netrom [accept$netrom accept4$netrom syz_init_net_socket$netrom] setsockopt$netrom_NETROM_T1 : sock_netrom [accept$netrom accept4$netrom syz_init_net_socket$netrom] setsockopt$netrom_NETROM_T2 : sock_netrom [accept$netrom accept4$netrom syz_init_net_socket$netrom] setsockopt$netrom_NETROM_T4 : sock_netrom [accept$netrom accept4$netrom syz_init_net_socket$netrom] setsockopt$nfc_llcp_NFC_LLCP_MIUX : sock_nfc_llcp [accept$nfc_llcp accept4$nfc_llcp syz_init_net_socket$nfc_llcp] setsockopt$nfc_llcp_NFC_LLCP_RW : sock_nfc_llcp [accept$nfc_llcp accept4$nfc_llcp syz_init_net_socket$nfc_llcp] setsockopt$rose : sock_rose [accept4$rose syz_init_net_socket$rose] setsockopt$sock_attach_bpf : fd_bpf_prog [bpf$BPF_PROG_GET_FD_BY_ID bpf$BPF_PROG_RAW_TRACEPOINT_LOAD bpf$BPF_PROG_WITH_BTFID_LOAD ...] setsockopt$sock_cred : pid [capget$auto capset$auto clone$auto ...] setsockopt$sock_timeval : time_usec [getitimer getrusage getsockopt$sock_timeval ...] syz_memcpy_off$KVM_EXIT_HYPERCALL : kvm_run_ptr [mmap$KVM_VCPU] syz_memcpy_off$KVM_EXIT_MMIO : kvm_run_ptr [mmap$KVM_VCPU] BinFmtMisc : enabled Comparisons : enabled Coverage : enabled DelayKcovMmap : enabled DevlinkPCI : PCI device 0000:00:10.0 is not available ExtraCoverage : enabled Fault : enabled KCSAN : write(/sys/kernel/debug/kcsan, on) failed KcovResetIoctl : kernel does not support ioctl(KCOV_RESET_TRACE) LRWPANEmulation : enabled Leak : failed to write(kmemleak, "scan=off") NetDevices : enabled NetInjection : enabled NicVF : PCI device 0000:00:11.0 is not available SandboxAndroid : setfilecon: setxattr failed. (errno 1: Operation not permitted). . process exited with status 67. SandboxNamespace : enabled SandboxNone : enabled SandboxSetuid : enabled Swap : enabled USBEmulation : enabled VhciInjection : enabled WifiEmulation : enabled syscalls : 450/8071 2026/01/30 05:07:57 base: machine check complete 2026/01/30 05:07:57 discovered 7637 source files, 336366 symbols 2026/01/30 05:07:59 coverage filter: ^___free_pages$: [] 2026/01/30 05:07:59 coverage filter: ^__access_remote_vm$: [__access_remote_vm] 2026/01/30 05:07:59 coverage filter: ^__add_to_free_list$: [] 2026/01/30 05:07:59 coverage filter: ^__alloc_frozen_pages_noprof$: [] 2026/01/30 05:07:59 coverage filter: ^__alloc_pages_cpuset_fallback$: [] 2026/01/30 05:07:59 coverage filter: ^__alloc_pages_direct_compact$: [] 2026/01/30 05:07:59 coverage filter: ^__alloc_pages_direct_reclaim$: [] 2026/01/30 05:07:59 coverage filter: ^__alloc_pages_may_oom$: [] 2026/01/30 05:07:59 coverage filter: ^__alloc_pages_slowpath$: [] 2026/01/30 05:07:59 coverage filter: ^__btrfs_prealloc_file_range$: [__btrfs_prealloc_file_range] 2026/01/30 05:07:59 coverage filter: ^__build_all_zonelists$: [] 2026/01/30 05:07:59 coverage filter: ^__cow_file_range_inline$: [__cow_file_range_inline] 2026/01/30 05:07:59 coverage filter: ^__del_page_from_free_list$: [] 2026/01/30 05:07:59 coverage filter: ^__exchange_data_block$: [__exchange_data_block] 2026/01/30 05:07:59 coverage filter: ^__ext4_get_inode_loc$: [__ext4_get_inode_loc] 2026/01/30 05:07:59 coverage filter: ^__f2fs_cluster_blocks$: [__f2fs_cluster_blocks] 2026/01/30 05:07:59 coverage filter: ^__f2fs_ioctl$: [__f2fs_ioctl] 2026/01/30 05:07:59 coverage filter: ^__folio_alloc_noprof$: [] 2026/01/30 05:07:59 coverage filter: ^__folio_freeze_and_split_unmapped$: [__folio_freeze_and_split_unmapped] 2026/01/30 05:07:59 coverage filter: ^__folio_put$: [__folio_put] 2026/01/30 05:07:59 coverage filter: ^__folio_split$: [__folio_split] 2026/01/30 05:07:59 coverage filter: ^__free_contig_frozen_range$: [] 2026/01/30 05:07:59 coverage filter: ^__free_frozen_pages$: [] 2026/01/30 05:07:59 coverage filter: ^__free_one_page$: [] 2026/01/30 05:07:59 coverage filter: ^__free_pages_ok$: [] 2026/01/30 05:07:59 coverage filter: ^__get_obj_cgroup_from_memcg$: [] 2026/01/30 05:07:59 coverage filter: ^__gfs2_jdata_write_folio$: [__gfs2_jdata_write_folio] 2026/01/30 05:07:59 coverage filter: ^__has_merged_page$: [__has_merged_page] 2026/01/30 05:07:59 coverage filter: ^__ia32_sys_set_mempolicy$: [__ia32_sys_set_mempolicy] 2026/01/30 05:07:59 coverage filter: ^__iomap_write_begin$: [__iomap_write_begin] 2026/01/30 05:07:59 coverage filter: ^__isolate_free_page$: [] 2026/01/30 05:07:59 coverage filter: ^__kasan_poison_slab$: [] 2026/01/30 05:07:59 coverage filter: ^__linger_cancel$: [__linger_cancel] 2026/01/30 05:07:59 coverage filter: ^__mem_cgroup_charge$: [] 2026/01/30 05:07:59 coverage filter: ^__mem_cgroup_handle_over_high$: [] 2026/01/30 05:07:59 coverage filter: ^__mem_cgroup_try_charge_swap$: [] 2026/01/30 05:07:59 coverage filter: ^__memcg_kmem_charge_page$: [] 2026/01/30 05:07:59 coverage filter: ^__memcg_kmem_uncharge_page$: [] 2026/01/30 05:07:59 coverage filter: ^__memcg_slab_free_hook$: [] 2026/01/30 05:07:59 coverage filter: ^__memcg_slab_post_alloc_hook$: [] 2026/01/30 05:07:59 coverage filter: ^__move_freepages_block_isolate$: [] 2026/01/30 05:07:59 coverage filter: ^__node_reclaim$: [__node_reclaim] 2026/01/30 05:07:59 coverage filter: ^__nr_hugepages_store_common$: [__nr_hugepages_store_common] 2026/01/30 05:07:59 coverage filter: ^__ocfs2_rotate_tree_left$: [__ocfs2_rotate_tree_left] 2026/01/30 05:07:59 coverage filter: ^__offline_isolated_pages$: [] 2026/01/30 05:07:59 coverage filter: ^__p4d_alloc$: [__p4d_alloc] 2026/01/30 05:07:59 coverage filter: ^__pagetable_ctor$: [__pagetable_ctor __pagetable_ctor] 2026/01/30 05:07:59 coverage filter: ^__pfx___pagetable_ctor$: [] 2026/01/30 05:07:59 coverage filter: ^__pmd_alloc$: [__pmd_alloc] 2026/01/30 05:07:59 coverage filter: ^__pte_alloc$: [__pte_alloc] 2026/01/30 05:07:59 coverage filter: ^__pte_alloc_kernel$: [__pte_alloc_kernel] 2026/01/30 05:07:59 coverage filter: ^__pud_alloc$: [__pud_alloc] 2026/01/30 05:07:59 coverage filter: ^__rmqueue_pcplist$: [] 2026/01/30 05:07:59 coverage filter: ^__se_sys_mbind$: [__se_sys_mbind] 2026/01/30 05:07:59 coverage filter: ^__se_sys_migrate_pages$: [__se_sys_migrate_pages] 2026/01/30 05:07:59 coverage filter: ^__se_sys_move_pages$: [__se_sys_move_pages] 2026/01/30 05:07:59 coverage filter: ^__se_sys_set_mempolicy_home_node$: [__se_sys_set_mempolicy_home_node] 2026/01/30 05:07:59 coverage filter: ^__shmem_file_setup$: [__shmem_file_setup] 2026/01/30 05:07:59 coverage filter: ^__submit_request$: [__submit_request] 2026/01/30 05:07:59 coverage filter: ^__swap_writepage$: [__swap_writepage] 2026/01/30 05:07:59 coverage filter: ^__tlb_remove_table$: [__tlb_remove_table] 2026/01/30 05:07:59 coverage filter: ^__unmap_hugepage_range$: [__unmap_hugepage_range] 2026/01/30 05:07:59 coverage filter: ^__vm_insert_mixed$: [__vm_insert_mixed] 2026/01/30 05:07:59 coverage filter: ^__vma_reservation_common$: [__vma_reservation_common] 2026/01/30 05:07:59 coverage filter: ^__write_extent_buffer$: [__write_extent_buffer] 2026/01/30 05:07:59 coverage filter: ^__x64_sys_set_mempolicy$: [__x64_sys_set_mempolicy] 2026/01/30 05:07:59 coverage filter: ^abort_on_full_fn$: [abort_on_full_fn] 2026/01/30 05:07:59 coverage filter: ^adjust_managed_page_count$: [] 2026/01/30 05:07:59 coverage filter: ^adjust_pool_surplus$: [adjust_pool_surplus] 2026/01/30 05:07:59 coverage filter: ^affs_read_folio_ofs$: [affs_read_folio_ofs] 2026/01/30 05:07:59 coverage filter: ^affs_write_begin$: [affs_write_begin] 2026/01/30 05:07:59 coverage filter: ^alloc_contig_frozen_range_noprof$: [] 2026/01/30 05:07:59 coverage filter: ^alloc_contig_pages_noprof$: [] 2026/01/30 05:07:59 coverage filter: ^alloc_contig_range_noprof$: [] 2026/01/30 05:07:59 coverage filter: ^alloc_dummy_extent_buffer$: [alloc_dummy_extent_buffer] 2026/01/30 05:07:59 coverage filter: ^alloc_extent_buffer$: [alloc_extent_buffer] 2026/01/30 05:07:59 coverage filter: ^alloc_frozen_pages_nolock_noprof$: [] 2026/01/30 05:07:59 coverage filter: ^alloc_frozen_pages_noprof$: [alloc_frozen_pages_noprof] 2026/01/30 05:07:59 coverage filter: ^alloc_hugetlb_folio$: [alloc_hugetlb_folio] 2026/01/30 05:07:59 coverage filter: ^alloc_migration_target_by_mpol$: [alloc_migration_target_by_mpol] 2026/01/30 05:07:59 coverage filter: ^alloc_pages_bulk_mempolicy_noprof$: [alloc_pages_bulk_mempolicy_noprof] 2026/01/30 05:07:59 coverage filter: ^alloc_pages_exact_nid_noprof$: [] 2026/01/30 05:07:59 coverage filter: ^alloc_pages_nolock_noprof$: [] 2026/01/30 05:07:59 coverage filter: ^alloc_pages_noprof$: [alloc_pages_noprof] 2026/01/30 05:07:59 coverage filter: ^alloc_surplus_hugetlb_folio$: [alloc_surplus_hugetlb_folio] 2026/01/30 05:07:59 coverage filter: ^bio_associate_blkg_from_page$: [bio_associate_blkg_from_page] 2026/01/30 05:07:59 coverage filter: ^bl_read_pagelist$: [bl_read_pagelist] 2026/01/30 05:07:59 coverage filter: ^btrfs_add_link$: [btrfs_add_link] 2026/01/30 05:07:59 coverage filter: ^btrfs_begin_enable_verity$: [btrfs_begin_enable_verity] 2026/01/30 05:07:59 coverage filter: ^btrfs_calc_input_length$: [btrfs_calc_input_length] 2026/01/30 05:07:59 coverage filter: ^btrfs_clone_extent_buffer$: [btrfs_clone_extent_buffer] 2026/01/30 05:07:59 coverage filter: ^btrfs_compr_pool_scan$: [btrfs_compr_pool_scan] 2026/01/30 05:07:59 coverage filter: ^btrfs_compress_heuristic$: [btrfs_compress_heuristic] 2026/01/30 05:07:59 coverage filter: ^btrfs_cont_expand$: [btrfs_cont_expand] 2026/01/30 05:07:59 coverage filter: ^btrfs_create_new_inode$: [btrfs_create_new_inode] 2026/01/30 05:07:59 coverage filter: ^btrfs_do_readpage$: [btrfs_do_readpage] 2026/01/30 05:07:59 coverage filter: ^btrfs_end_enable_verity$: [btrfs_end_enable_verity] 2026/01/30 05:07:59 coverage filter: ^btrfs_finish_one_ordered$: [btrfs_finish_one_ordered] 2026/01/30 05:07:59 coverage filter: ^btrfs_get_extent$: [btrfs_get_extent] 2026/01/30 05:07:59 coverage filter: ^btrfs_iget$: [btrfs_iget] 2026/01/30 05:07:59 coverage filter: ^btrfs_link$: [btrfs_link] 2026/01/30 05:07:59 coverage filter: ^btrfs_lookup_dentry$: [btrfs_lookup_dentry] 2026/01/30 05:07:59 coverage filter: ^btrfs_page_mkwrite$: [btrfs_page_mkwrite] 2026/01/30 05:07:59 coverage filter: ^btrfs_read_locked_inode$: [btrfs_read_locked_inode] 2026/01/30 05:07:59 coverage filter: ^btrfs_read_merkle_tree_page$: [btrfs_read_merkle_tree_page] 2026/01/30 05:07:59 coverage filter: ^btrfs_release_extent_buffer_folios$: [btrfs_release_extent_buffer_folios] 2026/01/30 05:07:59 coverage filter: ^btrfs_rename2$: [btrfs_rename2] 2026/01/30 05:07:59 coverage filter: ^btrfs_rmdir$: [btrfs_rmdir] 2026/01/30 05:07:59 coverage filter: ^btrfs_swap_activate$: [btrfs_swap_activate] 2026/01/30 05:07:59 coverage filter: ^btrfs_symlink$: [btrfs_symlink] 2026/01/30 05:07:59 coverage filter: ^btrfs_unlink$: [btrfs_unlink] 2026/01/30 05:07:59 coverage filter: ^btrfs_update_inode_item$: [btrfs_update_inode_item] 2026/01/30 05:07:59 coverage filter: ^btrfs_writepages$: [btrfs_writepages] 2026/01/30 05:07:59 coverage filter: ^calc_target$: [calc_target] 2026/01/30 05:07:59 coverage filter: ^can_change_pmd_writable$: [can_change_pmd_writable] 2026/01/30 05:07:59 coverage filter: ^ceph_copy_file_range$: [ceph_copy_file_range] 2026/01/30 05:07:59 coverage filter: ^ceph_fallocate$: [ceph_fallocate] 2026/01/30 05:07:59 coverage filter: ^ceph_filemap_fault$: [ceph_filemap_fault] 2026/01/30 05:07:59 coverage filter: ^ceph_invalidate_folio$: [ceph_invalidate_folio] 2026/01/30 05:07:59 coverage filter: ^ceph_netfs_check_write_begin$: [ceph_netfs_check_write_begin] 2026/01/30 05:07:59 coverage filter: ^ceph_netfs_issue_read$: [ceph_netfs_issue_read] 2026/01/30 05:07:59 coverage filter: ^ceph_osdc_call$: [ceph_osdc_call] 2026/01/30 05:07:59 coverage filter: ^ceph_osdc_handle_map$: [ceph_osdc_handle_map] 2026/01/30 05:07:59 coverage filter: ^ceph_osdc_list_watchers$: [ceph_osdc_list_watchers] 2026/01/30 05:07:59 coverage filter: ^ceph_osdc_notify$: [ceph_osdc_notify] 2026/01/30 05:07:59 coverage filter: ^ceph_osdc_notify_ack$: [ceph_osdc_notify_ack] 2026/01/30 05:07:59 coverage filter: ^ceph_osdc_start_request$: [ceph_osdc_start_request] 2026/01/30 05:07:59 coverage filter: ^ceph_osdc_unwatch$: [ceph_osdc_unwatch] 2026/01/30 05:07:59 coverage filter: ^ceph_osdc_watch$: [ceph_osdc_watch] 2026/01/30 05:07:59 coverage filter: ^ceph_write_iter$: [ceph_write_iter] 2026/01/30 05:07:59 coverage filter: ^ceph_writepages_start$: [ceph_writepages_start] 2026/01/30 05:07:59 coverage filter: ^ceph_zero_partial_page$: [ceph_zero_partial_page] 2026/01/30 05:07:59 coverage filter: ^charge_memcg$: [] 2026/01/30 05:07:59 coverage filter: ^check_linger_pool_dne$: [check_linger_pool_dne] 2026/01/30 05:07:59 coverage filter: ^check_move_unevictable_folios$: [check_move_unevictable_folios] 2026/01/30 05:07:59 coverage filter: ^check_new_pages$: [] 2026/01/30 05:07:59 coverage filter: ^check_pool_dne$: [check_pool_dne] 2026/01/30 05:07:59 coverage filter: ^clear_refs_write$: [clear_refs_write] 2026/01/30 05:07:59 coverage filter: ^clear_shadow_entries$: [clear_shadow_entries] 2026/01/30 05:07:59 coverage filter: ^close_osd$: [close_osd] 2026/01/30 05:07:59 coverage filter: ^compact_node$: [compact_node] 2026/01/30 05:07:59 coverage filter: ^compaction_alloc$: [compaction_alloc] 2026/01/30 05:07:59 coverage filter: ^compaction_proactiveness_sysctl_handler$: [compaction_proactiveness_sysctl_handler] 2026/01/30 05:07:59 coverage filter: ^compress_file_range$: [compress_file_range] 2026/01/30 05:07:59 coverage filter: ^copy_extent_buffer$: [copy_extent_buffer] 2026/01/30 05:07:59 coverage filter: ^copy_huge_pmd$: [copy_huge_pmd] 2026/01/30 05:07:59 coverage filter: ^copy_page_range$: [copy_page_range] 2026/01/30 05:07:59 coverage filter: ^copy_pmd_range$: [copy_pmd_range] 2026/01/30 05:07:59 coverage filter: ^copy_remote_vm_str$: [copy_remote_vm_str] 2026/01/30 05:07:59 coverage filter: ^count_memcg_folio_events$: [count_memcg_folio_events count_memcg_folio_events] 2026/01/30 05:07:59 coverage filter: ^count_swpout_vm_event$: [count_swpout_vm_event] 2026/01/30 05:07:59 coverage filter: ^cow_file_range$: [cow_file_range] 2026/01/30 05:07:59 coverage filter: ^current_objcg_update$: [] 2026/01/30 05:07:59 coverage filter: ^default_hugepagesz_setupargs$: [] 2026/01/30 05:07:59 coverage filter: ^deferred_split_folio$: [deferred_split_folio] 2026/01/30 05:07:59 coverage filter: ^deferred_split_scan$: [deferred_split_scan] 2026/01/30 05:07:59 coverage filter: ^do_huge_pmd_anonymous_page$: [do_huge_pmd_anonymous_page] 2026/01/30 05:07:59 coverage filter: ^do_huge_pmd_wp_page$: [do_huge_pmd_wp_page] 2026/01/30 05:07:59 coverage filter: ^do_migrate_pages$: [do_migrate_pages] 2026/01/30 05:07:59 coverage filter: ^do_pte_missing$: [do_pte_missing] 2026/01/30 05:07:59 coverage filter: ^do_swap_page$: [do_swap_page] 2026/01/30 05:07:59 coverage filter: ^do_try_to_free_pages$: [do_try_to_free_pages] 2026/01/30 05:07:59 coverage filter: ^do_watch_error$: [do_watch_error] 2026/01/30 05:07:59 coverage filter: ^do_watch_notify$: [do_watch_notify] 2026/01/30 05:07:59 coverage filter: ^do_wp_page$: [do_wp_page] 2026/01/30 05:07:59 coverage filter: ^drain_local_memcg_stock$: [] 2026/01/30 05:07:59 coverage filter: ^drain_local_obj_stock$: [] 2026/01/30 05:07:59 coverage filter: ^drain_obj_stock$: [] 2026/01/30 05:07:59 coverage filter: ^end_bbio_data_read$: [end_bbio_data_read] 2026/01/30 05:07:59 coverage filter: ^end_bbio_meta_write$: [end_bbio_meta_write] 2026/01/30 05:07:59 coverage filter: ^end_folio_read$: [end_folio_read] 2026/01/30 05:07:59 coverage filter: ^evict_folios$: [evict_folios] 2026/01/30 05:07:59 coverage filter: ^ext4_bio_write_folio$: [ext4_bio_write_folio] 2026/01/30 05:07:59 coverage filter: ^ext4_block_zero_page_range$: [ext4_block_zero_page_range] 2026/01/30 05:07:59 coverage filter: ^ext4_da_get_block_prep$: [ext4_da_get_block_prep] 2026/01/30 05:07:59 coverage filter: ^ext4_da_write_end$: [ext4_da_write_end] 2026/01/30 05:07:59 coverage filter: ^ext4_dax_writepages$: [ext4_dax_writepages] 2026/01/30 05:07:59 coverage filter: ^ext4_do_writepages$: [ext4_do_writepages] 2026/01/30 05:07:59 coverage filter: ^ext4_fill_raw_inode$: [ext4_fill_raw_inode] 2026/01/30 05:07:59 coverage filter: ^ext4_iomap_begin$: [ext4_iomap_begin] 2026/01/30 05:07:59 coverage filter: ^ext4_iomap_begin_report$: [ext4_iomap_begin_report] 2026/01/30 05:07:59 coverage filter: ^ext4_journalled_write_end$: [ext4_journalled_write_end] 2026/01/30 05:07:59 coverage filter: ^ext4_journalled_zero_new_buffers$: [ext4_journalled_zero_new_buffers] 2026/01/30 05:07:59 coverage filter: ^ext4_mark_iloc_dirty$: [ext4_mark_iloc_dirty] 2026/01/30 05:07:59 coverage filter: ^ext4_move_extents$: [ext4_move_extents] 2026/01/30 05:07:59 coverage filter: ^ext4_page_mkwrite$: [ext4_page_mkwrite] 2026/01/30 05:07:59 coverage filter: ^ext4_readpage_inline$: [ext4_readpage_inline] 2026/01/30 05:07:59 coverage filter: ^ext4_release_io_end$: [ext4_release_io_end] 2026/01/30 05:07:59 coverage filter: ^ext4_try_add_inline_entry$: [ext4_try_add_inline_entry] 2026/01/30 05:07:59 coverage filter: ^ext4_wait_for_tail_page_commit$: [ext4_wait_for_tail_page_commit] 2026/01/30 05:07:59 coverage filter: ^ext4_writepages$: [ext4_writepages] 2026/01/30 05:07:59 coverage filter: ^extent_buffer_get_byte$: [extent_buffer_get_byte] 2026/01/30 05:07:59 coverage filter: ^extent_writepage_io$: [extent_writepage_io] 2026/01/30 05:07:59 coverage filter: ^f2fs_alloc_dic$: [f2fs_alloc_dic] 2026/01/30 05:07:59 coverage filter: ^f2fs_compat_ioc_gc_range$: [f2fs_compat_ioc_gc_range] 2026/01/30 05:07:59 coverage filter: ^f2fs_compat_ioc_move_range$: [f2fs_compat_ioc_move_range] 2026/01/30 05:07:59 coverage filter: ^f2fs_compat_ioctl$: [f2fs_compat_ioctl] 2026/01/30 05:07:59 coverage filter: ^f2fs_compress_write_end$: [f2fs_compress_write_end] 2026/01/30 05:07:59 coverage filter: ^f2fs_decompress_end_io$: [f2fs_decompress_end_io] 2026/01/30 05:07:59 coverage filter: ^f2fs_defragment_range$: [f2fs_defragment_range] 2026/01/30 05:07:59 coverage filter: ^f2fs_encrypt_one_page$: [f2fs_encrypt_one_page] 2026/01/30 05:07:59 coverage filter: ^f2fs_end_read_compressed_page$: [f2fs_end_read_compressed_page] 2026/01/30 05:07:59 coverage filter: ^f2fs_fallocate$: [f2fs_fallocate] 2026/01/30 05:07:59 coverage filter: ^f2fs_fiemap$: [f2fs_fiemap] 2026/01/30 05:07:59 coverage filter: ^f2fs_file_fadvise$: [f2fs_file_fadvise] 2026/01/30 05:07:59 coverage filter: ^f2fs_file_read_iter$: [f2fs_file_read_iter] 2026/01/30 05:07:59 coverage filter: ^f2fs_file_splice_read$: [f2fs_file_splice_read] 2026/01/30 05:07:59 coverage filter: ^f2fs_file_write_iter$: [f2fs_file_write_iter] 2026/01/30 05:07:59 coverage filter: ^f2fs_fileattr_set$: [f2fs_fileattr_set] 2026/01/30 05:07:59 coverage filter: ^f2fs_get_new_data_folio$: [f2fs_get_new_data_folio] 2026/01/30 05:07:59 coverage filter: ^f2fs_get_read_data_folio$: [f2fs_get_read_data_folio] 2026/01/30 05:07:59 coverage filter: ^f2fs_ioc_start_atomic_write$: [f2fs_ioc_start_atomic_write] 2026/01/30 05:07:59 coverage filter: ^f2fs_ioctl$: [f2fs_ioctl] 2026/01/30 05:07:59 coverage filter: ^f2fs_map_blocks$: [f2fs_map_blocks] 2026/01/30 05:07:59 coverage filter: ^f2fs_mpage_readpages$: [f2fs_mpage_readpages] 2026/01/30 05:07:59 coverage filter: ^f2fs_prepare_compress_overwrite$: [f2fs_prepare_compress_overwrite] 2026/01/30 05:07:59 coverage filter: ^f2fs_prepare_decomp_mem$: [f2fs_prepare_decomp_mem] 2026/01/30 05:07:59 coverage filter: ^f2fs_put_rpages_wbc$: [f2fs_put_rpages_wbc] 2026/01/30 05:07:59 coverage filter: ^f2fs_read_inline_data$: [f2fs_read_inline_data] 2026/01/30 05:07:59 coverage filter: ^f2fs_read_multi_pages$: [f2fs_read_multi_pages] 2026/01/30 05:07:59 coverage filter: ^f2fs_reserve_block$: [f2fs_reserve_block] 2026/01/30 05:07:59 coverage filter: ^f2fs_submit_page_write$: [f2fs_submit_page_write] 2026/01/30 05:07:59 coverage filter: ^f2fs_truncate_partial_cluster$: [f2fs_truncate_partial_cluster] 2026/01/30 05:07:59 coverage filter: ^f2fs_vm_page_mkwrite$: [f2fs_vm_page_mkwrite] 2026/01/30 05:07:59 coverage filter: ^f2fs_write_begin$: [f2fs_write_begin] 2026/01/30 05:07:59 coverage filter: ^f2fs_write_data_pages$: [f2fs_write_data_pages] 2026/01/30 05:07:59 coverage filter: ^f2fs_write_multi_pages$: [f2fs_write_multi_pages] 2026/01/30 05:07:59 coverage filter: ^f2fs_write_single_data_page$: [f2fs_write_single_data_page] 2026/01/30 05:07:59 coverage filter: ^f2fs_zero_range$: [f2fs_zero_range] 2026/01/30 05:07:59 coverage filter: ^f2fs_zone_write_end_io$: [f2fs_zone_write_end_io] 2026/01/30 05:07:59 coverage filter: ^fallback_to_cow$: [fallback_to_cow] 2026/01/30 05:07:59 coverage filter: ^fill_zero$: [fill_zero] 2026/01/30 05:07:59 coverage filter: ^find_lock_delalloc_range$: [find_lock_delalloc_range] 2026/01/30 05:07:59 coverage filter: ^flush_memcg_stats_dwork$: [] 2026/01/30 05:07:59 coverage filter: ^folio_alloc_noprof$: [folio_alloc_noprof] 2026/01/30 05:07:59 coverage filter: ^folio_isolate_lru$: [folio_isolate_lru] 2026/01/30 05:07:59 coverage filter: ^folio_split_memcg_refs$: [] 2026/01/30 05:07:59 coverage filter: ^folio_split_unmapped$: [folio_split_unmapped] 2026/01/30 05:07:59 coverage filter: ^folio_try_dup_anon_rmap_pmd$: [folio_try_dup_anon_rmap_pmd] 2026/01/30 05:07:59 coverage filter: ^folio_unmap_invalidate$: [folio_unmap_invalidate] 2026/01/30 05:07:59 coverage filter: ^folio_unqueue_deferred_split$: [folio_unqueue_deferred_split folio_unqueue_deferred_split folio_unqueue_deferred_split] 2026/01/30 05:07:59 coverage filter: ^folio_zero_range$: [folio_zero_range folio_zero_range folio_zero_range folio_zero_range folio_zero_range folio_zero_range folio_zero_range] 2026/01/30 05:07:59 coverage filter: ^folio_zero_segment$: [folio_zero_segment folio_zero_segment folio_zero_segment folio_zero_segment folio_zero_segment folio_zero_segment folio_zero_segment folio_zero_segment] 2026/01/30 05:07:59 coverage filter: ^folios_put_refs$: [folios_put_refs] 2026/01/30 05:07:59 coverage filter: ^follow_pfnmap_start$: [follow_pfnmap_start] 2026/01/30 05:07:59 coverage filter: ^free_contig_frozen_range$: [] 2026/01/30 05:07:59 coverage filter: ^free_frozen_page_commit$: [] 2026/01/30 05:07:59 coverage filter: ^free_hpage_workfn$: [free_hpage_workfn] 2026/01/30 05:07:59 coverage filter: ^free_one_page$: [] 2026/01/30 05:07:59 coverage filter: ^free_page_is_bad$: [] 2026/01/30 05:07:59 coverage filter: ^free_pages_prepare$: [] 2026/01/30 05:07:59 coverage filter: ^free_pcppages_bulk$: [] 2026/01/30 05:07:59 coverage filter: ^free_tail_page_prepare$: [] 2026/01/30 05:07:59 coverage filter: ^free_unref_folios$: [] 2026/01/30 05:07:59 coverage filter: ^fuse_copy_folio$: [fuse_copy_folio] 2026/01/30 05:07:59 coverage filter: ^fuse_dev_do_read$: [fuse_dev_do_read] 2026/01/30 05:07:59 coverage filter: ^fuse_dev_do_write$: [fuse_dev_do_write] 2026/01/30 05:07:59 coverage filter: ^fuse_dev_read$: [fuse_dev_read] 2026/01/30 05:07:59 coverage filter: ^fuse_dev_splice_read$: [fuse_dev_splice_read] 2026/01/30 05:07:59 coverage filter: ^fuse_dev_splice_write$: [fuse_dev_splice_write] 2026/01/30 05:07:59 coverage filter: ^fuse_dev_write$: [fuse_dev_write] 2026/01/30 05:07:59 coverage filter: ^fuse_read_forget$: [fuse_read_forget] 2026/01/30 05:07:59 coverage filter: ^fuse_read_interrupt$: [fuse_read_interrupt] 2026/01/30 05:07:59 coverage filter: ^get_mem_cgroup_from_current$: [] 2026/01/30 05:07:59 coverage filter: ^get_mem_cgroup_from_folio$: [] 2026/01/30 05:07:59 coverage filter: ^get_mem_cgroup_from_mm$: [] 2026/01/30 05:07:59 coverage filter: ^get_mem_cgroup_from_objcg$: [get_mem_cgroup_from_objcg] 2026/01/30 05:07:59 coverage filter: ^get_obj_cgroup_from_folio$: [] 2026/01/30 05:07:59 coverage filter: ^get_page_from_freelist$: [] 2026/01/30 05:07:59 coverage filter: ^get_pte_pfn$: [get_pte_pfn] 2026/01/30 05:07:59 coverage filter: ^gfs2_qd_dispose$: [gfs2_qd_dispose] 2026/01/30 05:07:59 coverage filter: ^gfs2_qd_shrink_scan$: [gfs2_qd_shrink_scan] 2026/01/30 05:07:59 coverage filter: ^gfs2_quota_cleanup$: [gfs2_quota_cleanup] 2026/01/30 05:07:59 coverage filter: ^gfs2_write_buf_to_page$: [gfs2_write_buf_to_page] 2026/01/30 05:07:59 coverage filter: ^gfs2_write_cache_jdata$: [gfs2_write_cache_jdata] 2026/01/30 05:07:59 coverage filter: ^handle_mm_fault$: [handle_mm_fault] 2026/01/30 05:07:59 coverage filter: ^handle_pte_marker$: [handle_pte_marker] 2026/01/30 05:07:59 coverage filter: ^handle_timeout$: [handle_timeout] 2026/01/30 05:07:59 coverage filter: ^hfs_btree_open$: [hfs_btree_open] 2026/01/30 05:07:59 coverage filter: ^high_work_func$: [] 2026/01/30 05:07:59 coverage filter: ^hugepages_setupargs$: [] 2026/01/30 05:07:59 coverage filter: ^hugepagesz_setupargs$: [] 2026/01/30 05:07:59 coverage filter: ^hugetlb_acct_memory$: [hugetlb_acct_memory] 2026/01/30 05:07:59 coverage filter: ^hugetlb_add_to_page_cache$: [hugetlb_add_to_page_cache] 2026/01/30 05:07:59 coverage filter: ^hugetlb_cma_alloc_frozen_folio$: [hugetlb_cma_alloc_frozen_folio] 2026/01/30 05:07:59 coverage filter: ^hugetlb_fault$: [hugetlb_fault] 2026/01/30 05:07:59 coverage filter: ^hugetlb_folio_init_vmemmap$: [] 2026/01/30 05:07:59 coverage filter: ^hugetlb_hstate_alloc_pages$: [] 2026/01/30 05:07:59 coverage filter: ^hugetlb_no_page$: [hugetlb_no_page] 2026/01/30 05:07:59 coverage filter: ^hugetlb_resv_map_add$: [hugetlb_resv_map_add] 2026/01/30 05:07:59 coverage filter: ^hugetlbfs_evict_inode$: [hugetlbfs_evict_inode] 2026/01/30 05:07:59 coverage filter: ^hugetlbfs_zero_partial_page$: [hugetlbfs_zero_partial_page] 2026/01/30 05:07:59 coverage filter: ^inc_valid_block_count$: [inc_valid_block_count] 2026/01/30 05:07:59 coverage filter: ^insert_page$: [insert_page] 2026/01/30 05:07:59 coverage filter: ^iomap_file_buffered_write$: [iomap_file_buffered_write] 2026/01/30 05:07:59 coverage filter: ^iomap_file_unshare$: [iomap_file_unshare] 2026/01/30 05:07:59 coverage filter: ^iomap_write_begin$: [iomap_write_begin] 2026/01/30 05:07:59 coverage filter: ^iomap_write_delalloc_release$: [iomap_write_delalloc_release] 2026/01/30 05:07:59 coverage filter: ^iomap_write_end$: [iomap_write_end] 2026/01/30 05:07:59 coverage filter: ^iomap_writeback_folio$: [iomap_writeback_folio] 2026/01/30 05:07:59 coverage filter: ^iomap_zero_range$: [iomap_zero_range] 2026/01/30 05:07:59 coverage filter: ^isolate_single_pageblock$: [isolate_single_pageblock] 2026/01/30 05:07:59 coverage filter: ^kcompactd$: [kcompactd] 2026/01/30 05:07:59 coverage filter: ^kick_osd_requests$: [kick_osd_requests] 2026/01/30 05:07:59 coverage filter: ^kswapd$: [kswapd] 2026/01/30 05:07:59 coverage filter: ^linger_commit_cb$: [linger_commit_cb] 2026/01/30 05:07:59 coverage filter: ^linger_map_check_cb$: [linger_map_check_cb] 2026/01/30 05:07:59 coverage filter: ^linger_ping_cb$: [linger_ping_cb] 2026/01/30 05:07:59 coverage filter: ^linger_reconnect_cb$: [linger_reconnect_cb] 2026/01/30 05:07:59 coverage filter: ^linger_submit$: [linger_submit] 2026/01/30 05:07:59 coverage filter: ^link_linger$: [link_linger] 2026/01/30 05:07:59 coverage filter: ^lowmem_reserve_ratio_sysctl_handler$: [] 2026/01/30 05:07:59 coverage filter: ^lru_gen_seq_write$: [lru_gen_seq_write] 2026/01/30 05:07:59 coverage filter: ^madvise_free_huge_pmd$: [madvise_free_huge_pmd] 2026/01/30 05:07:59 coverage filter: ^make_alloc_exact$: [] 2026/01/30 05:07:59 coverage filter: ^map_check_cb$: [map_check_cb] 2026/01/30 05:07:59 coverage filter: ^mem_cgroup_charge_hugetlb$: [] 2026/01/30 05:07:59 coverage filter: ^mem_cgroup_css_free$: [] 2026/01/30 05:07:59 coverage filter: ^mem_cgroup_css_offline$: [] 2026/01/30 05:07:59 coverage filter: ^mem_cgroup_css_online$: [] 2026/01/30 05:07:59 coverage filter: ^mem_cgroup_exit$: [] 2026/01/30 05:07:59 coverage filter: ^mem_cgroup_get_from_id$: [] 2026/01/30 05:07:59 coverage filter: ^mem_cgroup_get_oom_group$: [] 2026/01/30 05:07:59 coverage filter: ^mem_cgroup_iter$: [] 2026/01/30 05:07:59 coverage filter: ^mem_cgroup_iter_break$: [] 2026/01/30 05:07:59 coverage filter: ^mem_cgroup_oom_synchronize$: [] 2026/01/30 05:07:59 coverage filter: ^mem_cgroup_replace_folio$: [] 2026/01/30 05:07:59 coverage filter: ^mem_cgroup_scan_tasks$: [] 2026/01/30 05:07:59 coverage filter: ^mem_cgroup_swap_full$: [] 2026/01/30 05:07:59 coverage filter: ^mem_cgroup_swapin_charge_folio$: [] 2026/01/30 05:07:59 coverage filter: ^memcg1_check_events$: [] 2026/01/30 05:07:59 coverage filter: ^memcg1_swapout$: [] 2026/01/30 05:07:59 coverage filter: ^memcg1_uncharge_batch$: [] 2026/01/30 05:07:59 coverage filter: ^memcg_event_remove$: [] 2026/01/30 05:07:59 coverage filter: ^memcg_write_event_control$: [] 2026/01/30 05:07:59 coverage filter: ^memcmp_extent_buffer$: [memcmp_extent_buffer] 2026/01/30 05:07:59 coverage filter: ^memcpy_extent_buffer$: [memcpy_extent_buffer] 2026/01/30 05:07:59 coverage filter: ^memmap_init_compound$: [memmap_init_compound] 2026/01/30 05:07:59 coverage filter: ^memmove_extent_buffer$: [memmove_extent_buffer] 2026/01/30 05:07:59 coverage filter: ^memory_numa_stat_show$: [] 2026/01/30 05:07:59 coverage filter: ^mempolicy_slab_node$: [mempolicy_slab_node] 2026/01/30 05:07:59 coverage filter: ^memzero_extent_buffer$: [memzero_extent_buffer] 2026/01/30 05:07:59 coverage filter: ^migrate_device_unmap$: [migrate_device_unmap] 2026/01/30 05:07:59 coverage filter: ^migrate_folio_add$: [migrate_folio_add] 2026/01/30 05:07:59 coverage filter: ^migrate_folio_done$: [migrate_folio_done] 2026/01/30 05:07:59 coverage filter: ^migrate_pages$: [migrate_pages] 2026/01/30 05:07:59 coverage filter: ^migrate_pages_batch$: [migrate_pages_batch] 2026/01/30 05:07:59 coverage filter: ^mm_get_huge_zero_folio$: [mm_get_huge_zero_folio] 2026/01/30 05:07:59 coverage filter: ^mod_objcg_mlstate$: [] 2026/01/30 05:07:59 coverage filter: ^move_folios_to_lru$: [move_folios_to_lru] 2026/01/30 05:07:59 coverage filter: ^move_to_free_list$: [] 2026/01/30 05:07:59 coverage filter: ^netfs_page_mkwrite$: [netfs_page_mkwrite] 2026/01/30 05:07:59 coverage filter: ^netfs_perform_write$: [netfs_perform_write] 2026/01/30 05:07:59 coverage filter: ^netfs_prefetch_for_write$: [netfs_prefetch_for_write] 2026/01/30 05:07:59 coverage filter: ^netfs_read_to_pagecache$: [netfs_read_to_pagecache] 2026/01/30 05:07:59 coverage filter: ^netfs_write_begin$: [netfs_write_begin] 2026/01/30 05:07:59 coverage filter: ^nfs_read_completion$: [nfs_read_completion] 2026/01/30 05:07:59 coverage filter: ^nfs_write_begin$: [nfs_write_begin] 2026/01/30 05:07:59 coverage filter: ^nfs_write_end$: [nfs_write_end] 2026/01/30 05:07:59 coverage filter: ^ni_readpage_cmpr$: [ni_readpage_cmpr] 2026/01/30 05:07:59 coverage filter: ^ntfs_compress_write$: [ntfs_compress_write] 2026/01/30 05:07:59 coverage filter: ^ntfs_fallocate$: [ntfs_fallocate] 2026/01/30 05:07:59 coverage filter: ^ntfs_file_write_iter$: [ntfs_file_write_iter] 2026/01/30 05:07:59 coverage filter: ^ntfs_get_block_bmap$: [ntfs_get_block_bmap] 2026/01/30 05:07:59 coverage filter: ^ntfs_get_block_direct_IO_R$: [ntfs_get_block_direct_IO_R] 2026/01/30 05:07:59 coverage filter: ^ntfs_get_block_direct_IO_W$: [ntfs_get_block_direct_IO_W] 2026/01/30 05:07:59 coverage filter: ^ntfs_get_link$: [ntfs_get_link] 2026/01/30 05:07:59 coverage filter: ^ntfs_zero_range$: [ntfs_zero_range] 2026/01/30 05:07:59 coverage filter: ^numa_migrate_check$: [numa_migrate_check] 2026/01/30 05:07:59 coverage filter: ^obj_cgroup_charge_account$: [] 2026/01/30 05:07:59 coverage filter: ^obj_cgroup_charge_pages$: [] 2026/01/30 05:07:59 coverage filter: ^obj_cgroup_charge_zswap$: [] 2026/01/30 05:07:59 coverage filter: ^obj_cgroup_may_zswap$: [] 2026/01/30 05:07:59 coverage filter: ^obj_cgroup_release$: [] 2026/01/30 05:07:59 coverage filter: ^obj_cgroup_uncharge_zswap$: [] 2026/01/30 05:07:59 coverage filter: ^ocfs2_append_rec_to_path$: [ocfs2_append_rec_to_path] 2026/01/30 05:07:59 coverage filter: ^ocfs2_commit_truncate$: [ocfs2_commit_truncate] 2026/01/30 05:07:59 coverage filter: ^ocfs2_convert_inline_data_to_extents$: [ocfs2_convert_inline_data_to_extents] 2026/01/30 05:07:59 coverage filter: ^ocfs2_dio_wr_get_block$: [ocfs2_dio_wr_get_block] 2026/01/30 05:07:59 coverage filter: ^ocfs2_do_insert_extent$: [ocfs2_do_insert_extent] 2026/01/30 05:07:59 coverage filter: ^ocfs2_grow_tree$: [ocfs2_grow_tree] 2026/01/30 05:07:59 coverage filter: ^ocfs2_insert_extent$: [ocfs2_insert_extent] 2026/01/30 05:07:59 coverage filter: ^ocfs2_map_and_dirty_folio$: [ocfs2_map_and_dirty_folio] 2026/01/30 05:07:59 coverage filter: ^ocfs2_merge_rec_left$: [ocfs2_merge_rec_left] 2026/01/30 05:07:59 coverage filter: ^ocfs2_merge_rec_right$: [ocfs2_merge_rec_right] 2026/01/30 05:07:59 coverage filter: ^ocfs2_remove_rightmost_path$: [ocfs2_remove_rightmost_path] 2026/01/30 05:07:59 coverage filter: ^ocfs2_rotate_tree_right$: [ocfs2_rotate_tree_right] 2026/01/30 05:07:59 coverage filter: ^ocfs2_split_extent$: [ocfs2_split_extent] 2026/01/30 05:07:59 coverage filter: ^ocfs2_truncate_inline$: [ocfs2_truncate_inline] 2026/01/30 05:07:59 coverage filter: ^ocfs2_truncate_rec$: [ocfs2_truncate_rec] 2026/01/30 05:07:59 coverage filter: ^ocfs2_try_to_merge_extent$: [ocfs2_try_to_merge_extent] 2026/01/30 05:07:59 coverage filter: ^ocfs2_write_begin_nolock$: [ocfs2_write_begin_nolock] 2026/01/30 05:07:59 coverage filter: ^only_alloc_fresh_hugetlb_folio$: [only_alloc_fresh_hugetlb_folio] 2026/01/30 05:07:59 coverage filter: ^osd_dispatch$: [osd_dispatch] 2026/01/30 05:07:59 coverage filter: ^osd_get_con$: [osd_get_con] 2026/01/30 05:07:59 coverage filter: ^osd_sparse_read$: [osd_sparse_read] 2026/01/30 05:07:59 coverage filter: ^page_alloc_cpu_dead$: [] 2026/01/30 05:07:59 coverage filter: ^page_array_alloc$: [page_array_alloc] 2026/01/30 05:07:59 coverage filter: ^page_rmappable_folio$: [page_rmappable_folio page_rmappable_folio] 2026/01/30 05:07:59 coverage filter: ^pagemap_pmd_range$: [pagemap_pmd_range] 2026/01/30 05:07:59 coverage filter: ^pagemap_read$: [pagemap_read] 2026/01/30 05:07:59 coverage filter: ^pagetable_dtor_free$: [pagetable_dtor_free pagetable_dtor_free pagetable_dtor_free] 2026/01/30 05:07:59 coverage filter: ^percpu_pagelist_high_fraction_sysctl_handler$: [] 2026/01/30 05:07:59 coverage filter: ^perf_trace_mm_lru_insertion$: [perf_trace_mm_lru_insertion] 2026/01/30 05:07:59 coverage filter: ^pgd_alloc$: [pgd_alloc] 2026/01/30 05:07:59 coverage filter: ^pmd_free_pte_page$: [pmd_free_pte_page] 2026/01/30 05:07:59 coverage filter: ^policy_nodemask$: [policy_nodemask] 2026/01/30 05:07:59 coverage filter: ^prep_and_add_bootmem_folios$: [] 2026/01/30 05:07:59 coverage filter: ^prep_compound_page$: [] 2026/01/30 05:07:59 coverage filter: ^prepare_kswapd_sleep$: [prepare_kswapd_sleep] 2026/01/30 05:07:59 coverage filter: ^pte_alloc_one$: [pte_alloc_one] 2026/01/30 05:07:59 coverage filter: ^pte_free$: [pte_free pte_free] 2026/01/30 05:07:59 coverage filter: ^pte_free_kernel$: [pte_free_kernel] 2026/01/30 05:07:59 coverage filter: ^pte_free_now$: [pte_free_now] 2026/01/30 05:07:59 coverage filter: ^qd_put$: [qd_put] 2026/01/30 05:07:59 coverage filter: ^queue_folios_hugetlb$: [queue_folios_hugetlb] 2026/01/30 05:07:59 coverage filter: ^read_extent_buffer$: [read_extent_buffer] 2026/01/30 05:07:59 coverage filter: ^read_inline_extent$: [read_inline_extent] 2026/01/30 05:07:59 coverage filter: ^refill_obj_stock$: [] 2026/01/30 05:07:59 coverage filter: ^region_add$: [region_add] 2026/01/30 05:07:59 coverage filter: ^region_chg$: [region_chg] 2026/01/30 05:07:59 coverage filter: ^release_compress_blocks$: [release_compress_blocks] 2026/01/30 05:07:59 coverage filter: ^release_extent_buffer$: [release_extent_buffer] 2026/01/30 05:07:59 coverage filter: ^remove_device_exclusive_entry$: [remove_device_exclusive_entry] 2026/01/30 05:07:59 coverage filter: ^remove_pool_hugetlb_folio$: [remove_pool_hugetlb_folio] 2026/01/30 05:07:59 coverage filter: ^reserve_highatomic_pageblock$: [] 2026/01/30 05:07:59 coverage filter: ^restore_reserve_on_error$: [restore_reserve_on_error] 2026/01/30 05:07:59 coverage filter: ^run_delalloc_nocow$: [run_delalloc_nocow] 2026/01/30 05:07:59 coverage filter: ^secretmem_free_folio$: [secretmem_free_folio] 2026/01/30 05:07:59 coverage filter: ^send_linger$: [send_linger] 2026/01/30 05:07:59 coverage filter: ^send_request$: [send_request] 2026/01/30 05:07:59 coverage filter: ^set_pageblock_migratetype$: [] 2026/01/30 05:07:59 coverage filter: ^set_pmd_migration_entry$: [set_pmd_migration_entry] 2026/01/30 05:07:59 coverage filter: ^setup_per_zone_lowmem_reserve$: [] 2026/01/30 05:07:59 coverage filter: ^setup_per_zone_wmarks$: [] 2026/01/30 05:07:59 coverage filter: ^shmem_alloc_and_add_folio$: [shmem_alloc_and_add_folio] 2026/01/30 05:07:59 coverage filter: ^shmem_enabled_store$: [shmem_enabled_store] 2026/01/30 05:07:59 coverage filter: ^shmem_evict_inode$: [shmem_evict_inode] 2026/01/30 05:07:59 coverage filter: ^shmem_fallocate$: [shmem_fallocate] 2026/01/30 05:07:59 coverage filter: ^shmem_fault$: [shmem_fault] 2026/01/30 05:07:59 coverage filter: ^shmem_file_read_iter$: [shmem_file_read_iter] 2026/01/30 05:07:59 coverage filter: ^shmem_file_splice_read$: [shmem_file_splice_read] 2026/01/30 05:07:59 coverage filter: ^shmem_get_folio_gfp$: [shmem_get_folio_gfp] 2026/01/30 05:07:59 coverage filter: ^shmem_get_link$: [shmem_get_link] 2026/01/30 05:07:59 coverage filter: ^shmem_getattr$: [shmem_getattr] 2026/01/30 05:07:59 coverage filter: ^shmem_symlink$: [shmem_symlink] 2026/01/30 05:07:59 coverage filter: ^shmem_write_begin$: [shmem_write_begin] 2026/01/30 05:07:59 coverage filter: ^shmem_write_end$: [shmem_write_end] 2026/01/30 05:07:59 coverage filter: ^shmem_writeout$: [shmem_writeout] 2026/01/30 05:07:59 coverage filter: ^should_reclaim_retry$: [] 2026/01/30 05:07:59 coverage filter: ^show_smaps_rollup$: [show_smaps_rollup] 2026/01/30 05:07:59 coverage filter: ^shrink_active_list$: [shrink_active_list] 2026/01/30 05:07:59 coverage filter: ^shrink_folio_list$: [shrink_folio_list] 2026/01/30 05:07:59 coverage filter: ^shrink_lruvec$: [shrink_lruvec] 2026/01/30 05:07:59 coverage filter: ^shrink_node$: [shrink_node] 2026/01/30 05:07:59 coverage filter: ^sio_read_complete$: [sio_read_complete] 2026/01/30 05:07:59 coverage filter: ^smaps_account$: [smaps_account] 2026/01/30 05:07:59 coverage filter: ^smaps_hugetlb_range$: [smaps_hugetlb_range] 2026/01/30 05:07:59 coverage filter: ^split_folio_to_list$: [split_folio_to_list] 2026/01/30 05:07:59 coverage filter: ^split_free_frozen_pages$: [] 2026/01/30 05:07:59 coverage filter: ^split_huge_pages_all$: [split_huge_pages_all] 2026/01/30 05:07:59 coverage filter: ^split_huge_pages_in_file$: [split_huge_pages_in_file] 2026/01/30 05:07:59 coverage filter: ^split_huge_pages_write$: [split_huge_pages_write] 2026/01/30 05:07:59 coverage filter: ^split_huge_pmd_locked$: [split_huge_pmd_locked] 2026/01/30 05:07:59 coverage filter: ^split_large_buddy$: [] 2026/01/30 05:07:59 coverage filter: ^split_page_memcg$: [] 2026/01/30 05:07:59 coverage filter: ^squashfs_read_folio$: [squashfs_read_folio] 2026/01/30 05:07:59 coverage filter: ^start_isolate_page_range$: [start_isolate_page_range] 2026/01/30 05:07:59 coverage filter: ^submit_extent_folio$: [submit_extent_folio] 2026/01/30 05:07:59 coverage filter: ^submit_uncompressed_range$: [submit_uncompressed_range] 2026/01/30 05:07:59 coverage filter: ^swap_read_folio$: [swap_read_folio] 2026/01/30 05:07:59 coverage filter: ^swap_writeout$: [swap_writeout] 2026/01/30 05:07:59 coverage filter: ^sysctl_compaction_handler$: [sysctl_compaction_handler] 2026/01/30 05:07:59 coverage filter: ^sysctl_min_slab_ratio_sysctl_handler$: [] 2026/01/30 05:07:59 coverage filter: ^sysctl_min_unmapped_ratio_sysctl_handler$: [] 2026/01/30 05:07:59 coverage filter: ^tlb_remove_table_rcu$: [tlb_remove_table_rcu] 2026/01/30 05:07:59 coverage filter: ^trace_event_raw_event_mm_lru_insertion$: [trace_event_raw_event_mm_lru_insertion] 2026/01/30 05:07:59 coverage filter: ^truncate_cleanup_folio$: [truncate_cleanup_folio] 2026/01/30 05:07:59 coverage filter: ^truncate_folio_batch_exceptionals$: [truncate_folio_batch_exceptionals] 2026/01/30 05:07:59 coverage filter: ^try_folio_split_or_unmap$: [try_folio_split_or_unmap] 2026/01/30 05:07:59 coverage filter: ^try_restore_exclusive_pte$: [try_restore_exclusive_pte] 2026/01/30 05:07:59 coverage filter: ^try_to_free_pages$: [try_to_free_pages] 2026/01/30 05:07:59 coverage filter: ^ubifs_write_begin$: [ubifs_write_begin] 2026/01/30 05:07:59 coverage filter: ^ubifs_write_end$: [ubifs_write_end] 2026/01/30 05:07:59 coverage filter: ^uncharge_batch$: [] 2026/01/30 05:07:59 coverage filter: ^uncharge_folio$: [] 2026/01/30 05:07:59 coverage filter: ^uncompress_inline$: [uncompress_inline] 2026/01/30 05:07:59 coverage filter: ^unlock_delalloc_folio$: [unlock_delalloc_folio] 2026/01/30 05:07:59 coverage filter: ^unmap_huge_pmd_locked$: [unmap_huge_pmd_locked] 2026/01/30 05:07:59 coverage filter: ^unmap_page_range$: [unmap_page_range] 2026/01/30 05:07:59 coverage filter: ^update_and_free_pages_bulk$: [update_and_free_pages_bulk] 2026/01/30 05:07:59 coverage filter: ^virtio_fs_probe$: [virtio_fs_probe] 2026/01/30 05:07:59 coverage filter: ^virtio_fs_request_complete$: [virtio_fs_request_complete] 2026/01/30 05:07:59 coverage filter: ^vm_insert_pages$: [vm_insert_pages] 2026/01/30 05:07:59 coverage filter: ^vma_alloc_folio_noprof$: [vma_alloc_folio_noprof] 2026/01/30 05:07:59 coverage filter: ^vmemmap_pmd_entry$: [vmemmap_pmd_entry] 2026/01/30 05:07:59 coverage filter: ^walk_pmd_range_locked$: [walk_pmd_range_locked] 2026/01/30 05:07:59 coverage filter: ^walk_pud_range$: [walk_pud_range] 2026/01/30 05:07:59 coverage filter: ^walk_update_folio$: [walk_update_folio] 2026/01/30 05:07:59 coverage filter: ^weighted_interleave_auto_store$: [weighted_interleave_auto_store] 2026/01/30 05:07:59 coverage filter: ^weighted_interleave_nodes$: [weighted_interleave_nodes] 2026/01/30 05:07:59 coverage filter: ^write_begin_slow$: [write_begin_slow] 2026/01/30 05:07:59 coverage filter: ^write_one_eb$: [write_one_eb] 2026/01/30 05:07:59 coverage filter: ^writepage_delalloc$: [writepage_delalloc] 2026/01/30 05:07:59 coverage filter: ^z_erofs_decompress_queue$: [z_erofs_decompress_queue] 2026/01/30 05:07:59 coverage filter: ^z_erofs_read_folio$: [z_erofs_read_folio] 2026/01/30 05:07:59 coverage filter: ^z_erofs_readahead$: [z_erofs_readahead] 2026/01/30 05:07:59 coverage filter: ^z_erofs_scan_folio$: [z_erofs_scan_folio] 2026/01/30 05:07:59 coverage filter: ^zap_huge_pmd$: [zap_huge_pmd] 2026/01/30 05:07:59 coverage filter: ^zero_pmd_populate$: [] 2026/01/30 05:07:59 coverage filter: ^zero_user_segment$: [zero_user_segment zero_user_segment] 2026/01/30 05:07:59 coverage filter: ^zero_user_segments$: [zero_user_segments zero_user_segments zero_user_segments zero_user_segments zero_user_segments zero_user_segments zero_user_segments zero_user_segments zero_user_segments zero_user_segments zero_user_segments zero_user_segments zero_user_segments zero_user_segments zero_user_segments] 2026/01/30 05:07:59 coverage filter: ^zlib_compress_folios$: [zlib_compress_folios] 2026/01/30 05:07:59 coverage filter: ^zone_pcp_disable$: [] 2026/01/30 05:07:59 coverage filter: ^zswap_current_read$: [] 2026/01/30 05:07:59 coverage filter: include/linux/mm.h: [] 2026/01/30 05:07:59 coverage filter: io_uring/memmap.c: [io_uring/memmap.c] 2026/01/30 05:07:59 coverage filter: mm/huge_memory.c: [mm/huge_memory.c] 2026/01/30 05:07:59 coverage filter: mm/hugetlb.c: [mm/hugetlb.c mm/hugetlb_cgroup.c mm/hugetlb_cma.c] 2026/01/30 05:07:59 coverage filter: mm/hugetlb_cma.c: [] 2026/01/30 05:07:59 coverage filter: mm/internal.h: [] 2026/01/30 05:07:59 coverage filter: mm/mm_init.c: [mm/mm_init.c] 2026/01/30 05:07:59 coverage filter: mm/page_alloc.c: [] 2026/01/30 05:07:59 area "symbols": 33667 PCs in the cover filter 2026/01/30 05:07:59 area "files": 9153 PCs in the cover filter 2026/01/30 05:07:59 area "": 0 PCs in the cover filter 2026/01/30 05:07:59 executor cover filter: 0 PCs 2026/01/30 05:08:01 machine check: disabled the following syscalls: openat$acpi_thermal_rel : failed to open /dev/acpi_thermal_rel: no such file or directory openat$ashmem : failed to open /dev/ashmem: no such file or directory openat$bifrost : failed to open /dev/bifrost: no such file or directory openat$binder : failed to open /dev/binder: no such file or directory openat$camx : failed to open /dev/v4l/by-path/platform-soc@0:qcom_cam-req-mgr-video-index0: no such file or directory openat$capi20 : failed to open /dev/capi20: no such file or directory openat$cdrom1 : failed to open /dev/cdrom1: no such file or directory openat$damon_attrs : failed to open /sys/kernel/debug/damon/attrs: no such file or directory openat$damon_init_regions : failed to open /sys/kernel/debug/damon/init_regions: no such file or directory openat$damon_kdamond_pid : failed to open /sys/kernel/debug/damon/kdamond_pid: no such file or directory openat$damon_mk_contexts : failed to open /sys/kernel/debug/damon/mk_contexts: no such file or directory openat$damon_monitor_on : failed to open /sys/kernel/debug/damon/monitor_on: no such file or directory openat$damon_rm_contexts : failed to open /sys/kernel/debug/damon/rm_contexts: no such file or directory openat$damon_schemes : failed to open /sys/kernel/debug/damon/schemes: no such file or directory openat$damon_target_ids : failed to open /sys/kernel/debug/damon/target_ids: no such file or directory openat$hwbinder : failed to open /dev/hwbinder: no such file or directory openat$i915 : failed to open /dev/i915: no such file or directory openat$img_rogue : failed to open /dev/img-rogue: no such file or directory openat$irnet : failed to open /dev/irnet: no such file or directory openat$keychord : failed to open /dev/keychord: no such file or directory openat$kvm : failed to open /dev/kvm: no such file or directory openat$lightnvm : failed to open /dev/lightnvm/control: no such file or directory openat$mali : failed to open /dev/mali0: no such file or directory openat$md : failed to open /dev/md0: no such file or directory openat$msm : failed to open /dev/msm: no such file or directory openat$ndctl0 : failed to open /dev/ndctl0: no such file or directory openat$nmem0 : failed to open /dev/nmem0: no such file or directory openat$pktcdvd : failed to open /dev/pktcdvd/control: no such file or directory openat$pmem0 : failed to open /dev/pmem0: no such file or directory openat$proc_capi20 : failed to open /proc/capi/capi20: no such file or directory openat$proc_capi20ncci : failed to open /proc/capi/capi20ncci: no such file or directory openat$proc_reclaim : failed to open /proc/self/reclaim: no such file or directory openat$ptp1 : failed to open /dev/ptp1: no such file or directory openat$rnullb : failed to open /dev/rnullb0: no such file or directory openat$selinux_access : failed to open /selinux/access: no such file or directory openat$selinux_attr : selinux is not enabled openat$selinux_avc_cache_stats : failed to open /selinux/avc/cache_stats: no such file or directory openat$selinux_avc_cache_threshold : failed to open /selinux/avc/cache_threshold: no such file or directory openat$selinux_avc_hash_stats : failed to open /selinux/avc/hash_stats: no such file or directory openat$selinux_checkreqprot : failed to open /selinux/checkreqprot: no such file or directory openat$selinux_commit_pending_bools : failed to open /selinux/commit_pending_bools: no such file or directory openat$selinux_context : failed to open /selinux/context: no such file or directory openat$selinux_create : failed to open /selinux/create: no such file or directory openat$selinux_enforce : failed to open /selinux/enforce: no such file or directory openat$selinux_load : failed to open /selinux/load: no such file or directory openat$selinux_member : failed to open /selinux/member: no such file or directory openat$selinux_mls : failed to open /selinux/mls: no such file or directory openat$selinux_policy : failed to open /selinux/policy: no such file or directory openat$selinux_relabel : failed to open /selinux/relabel: no such file or directory openat$selinux_status : failed to open /selinux/status: no such file or directory openat$selinux_user : failed to open /selinux/user: no such file or directory openat$selinux_validatetrans : failed to open /selinux/validatetrans: no such file or directory openat$sev : failed to open /dev/sev: no such file or directory openat$sgx_provision : failed to open /dev/sgx_provision: no such file or directory openat$smack_task_current : smack is not enabled openat$smack_thread_current : smack is not enabled openat$smackfs_access : failed to open /sys/fs/smackfs/access: no such file or directory openat$smackfs_ambient : failed to open /sys/fs/smackfs/ambient: no such file or directory openat$smackfs_change_rule : failed to open /sys/fs/smackfs/change-rule: no such file or directory openat$smackfs_cipso : failed to open /sys/fs/smackfs/cipso: no such file or directory openat$smackfs_cipsonum : failed to open /sys/fs/smackfs/direct: no such file or directory openat$smackfs_ipv6host : failed to open /sys/fs/smackfs/ipv6host: no such file or directory openat$smackfs_load : failed to open /sys/fs/smackfs/load: no such file or directory openat$smackfs_logging : failed to open /sys/fs/smackfs/logging: no such file or directory openat$smackfs_netlabel : failed to open /sys/fs/smackfs/netlabel: no such file or directory openat$smackfs_onlycap : failed to open /sys/fs/smackfs/onlycap: no such file or directory openat$smackfs_ptrace : failed to open /sys/fs/smackfs/ptrace: no such file or directory openat$smackfs_relabel_self : failed to open /sys/fs/smackfs/relabel-self: no such file or directory openat$smackfs_revoke_subject : failed to open /sys/fs/smackfs/revoke-subject: no such file or directory openat$smackfs_syslog : failed to open /sys/fs/smackfs/syslog: no such file or directory openat$smackfs_unconfined : failed to open /sys/fs/smackfs/unconfined: no such file or directory openat$tlk_device : failed to open /dev/tlk_device: no such file or directory openat$trusty : failed to open /dev/trusty-ipc-dev0: no such file or directory openat$trusty_avb : failed to open /dev/trusty-ipc-dev0: no such file or directory openat$trusty_gatekeeper : failed to open /dev/trusty-ipc-dev0: no such file or directory openat$trusty_hwkey : failed to open /dev/trusty-ipc-dev0: no such file or directory openat$trusty_hwrng : failed to open /dev/trusty-ipc-dev0: no such file or directory openat$trusty_km : failed to open /dev/trusty-ipc-dev0: no such file or directory openat$trusty_km_secure : failed to open /dev/trusty-ipc-dev0: no such file or directory openat$trusty_storage : failed to open /dev/trusty-ipc-dev0: no such file or directory openat$tty : failed to open /dev/tty: no such device or address openat$uverbs0 : failed to open /dev/infiniband/uverbs0: no such file or directory openat$vfio : failed to open /dev/vfio/vfio: no such file or directory openat$vndbinder : failed to open /dev/vndbinder: no such file or directory openat$vtpm : failed to open /dev/vtpmx: no such file or directory openat$xenevtchn : failed to open /dev/xen/evtchn: no such file or directory openat$zygote : failed to open /dev/socket/zygote: no such file or directory socket$hf : socket$hf(0x13, 0x2, 0x0) failed: address family not supported by protocol socket$inet6_dccp : socket$inet6_dccp(0xa, 0x6, 0x0) failed: socket type not supported socket$inet_dccp : socket$inet_dccp(0x2, 0x6, 0x0) failed: socket type not supported socket$vsock_dgram : socket$vsock_dgram(0x28, 0x2, 0x0) failed: no such device transitively disabled the following syscalls (missing resource [creating syscalls]): accept$ax25 : sock_ax25 [accept$ax25 accept4$ax25 syz_init_net_socket$ax25] accept$netrom : sock_netrom [accept$netrom accept4$netrom syz_init_net_socket$netrom] accept$nfc_llcp : sock_nfc_llcp [accept$nfc_llcp accept4$nfc_llcp syz_init_net_socket$nfc_llcp] close$binfmt : fd_binfmt [openat$binfmt] close$fd_v4l2_buffer : fd_v4l2_buffer [ioctl$VIDIOC_QUERYBUF_DMABUF] close$ibv_device : fd_rdma [openat$uverbs0] mmap$DRM_I915 : fd_i915 [openat$i915] mmap$DRM_MSM : fd_msm [openat$msm] mmap$KVM_VCPU : vcpu_mmap_size [ioctl$KVM_GET_VCPU_MMAP_SIZE] mmap$bifrost : fd_bifrost [openat$bifrost openat$mali] mmap$perf : fd_perf [perf_event_open perf_event_open$cgroup] mmap$snddsp : fd_snd_dsp [syz_open_dev$sndpcmc syz_open_dev$sndpcmp] mmap$snddsp_control : fd_snd_dsp [syz_open_dev$sndpcmc syz_open_dev$sndpcmp] mmap$snddsp_status : fd_snd_dsp [syz_open_dev$sndpcmc syz_open_dev$sndpcmp] mmap$usbfs : fd_usbfs [syz_open_dev$usbfs] mmap$usbmon : fd_usbmon [syz_open_dev$usbmon] openat$binfmt : ptr_binfmt_file [syz_create_resource$binfmt] setsockopt$EBT_SO_SET_ENTRIES : uid [fstat fstat$auto geteuid ...] setsockopt$IP6T_SO_SET_REPLACE : fd_bpf_prog [bpf$BPF_PROG_GET_FD_BY_ID bpf$BPF_PROG_RAW_TRACEPOINT_LOAD bpf$BPF_PROG_WITH_BTFID_LOAD ...] setsockopt$IPT_SO_SET_REPLACE : fd_bpf_prog [bpf$BPF_PROG_GET_FD_BY_ID bpf$BPF_PROG_RAW_TRACEPOINT_LOAD bpf$BPF_PROG_WITH_BTFID_LOAD ...] setsockopt$SO_VM_SOCKETS_CONNECT_TIMEOUT_OLD: time_usec [getitimer getrusage getsockopt$sock_timeval ...] setsockopt$WPAN_SECURITY : sock_802154_dgram [syz_init_net_socket$802154_dgram] setsockopt$WPAN_SECURITY_LEVEL : sock_802154_dgram [syz_init_net_socket$802154_dgram] setsockopt$WPAN_WANTACK : sock_802154_dgram [syz_init_net_socket$802154_dgram] setsockopt$WPAN_WANTLQI : sock_802154_dgram [syz_init_net_socket$802154_dgram] setsockopt$X25_QBITINCL : sock_x25 [accept4$x25 syz_init_net_socket$x25] setsockopt$ax25_SO_BINDTODEVICE : sock_ax25 [accept$ax25 accept4$ax25 syz_init_net_socket$ax25] setsockopt$ax25_int : sock_ax25 [accept$ax25 accept4$ax25 syz_init_net_socket$ax25] setsockopt$bt_BT_CHANNEL_POLICY : sock_bt [accept4$bt_l2cap syz_init_net_socket$bt_bnep syz_init_net_socket$bt_cmtp ...] setsockopt$bt_BT_DEFER_SETUP : sock_bt [accept4$bt_l2cap syz_init_net_socket$bt_bnep syz_init_net_socket$bt_cmtp ...] setsockopt$bt_BT_FLUSHABLE : sock_bt [accept4$bt_l2cap syz_init_net_socket$bt_bnep syz_init_net_socket$bt_cmtp ...] setsockopt$bt_BT_POWER : sock_bt [accept4$bt_l2cap syz_init_net_socket$bt_bnep syz_init_net_socket$bt_cmtp ...] setsockopt$bt_BT_RCVMTU : sock_bt [accept4$bt_l2cap syz_init_net_socket$bt_bnep syz_init_net_socket$bt_cmtp ...] setsockopt$bt_BT_SECURITY : sock_bt [accept4$bt_l2cap syz_init_net_socket$bt_bnep syz_init_net_socket$bt_cmtp ...] setsockopt$bt_BT_SNDMTU : sock_bt [accept4$bt_l2cap syz_init_net_socket$bt_bnep syz_init_net_socket$bt_cmtp ...] setsockopt$bt_BT_VOICE : sock_bt [accept4$bt_l2cap syz_init_net_socket$bt_bnep syz_init_net_socket$bt_cmtp ...] setsockopt$bt_hci_HCI_DATA_DIR : sock_bt_hci [syz_init_net_socket$bt_hci] setsockopt$bt_hci_HCI_FILTER : sock_bt_hci [syz_init_net_socket$bt_hci] setsockopt$bt_hci_HCI_TIME_STAMP : sock_bt_hci [syz_init_net_socket$bt_hci] setsockopt$bt_l2cap_L2CAP_CONNINFO : sock_bt_l2cap [accept4$bt_l2cap syz_init_net_socket$bt_l2cap] setsockopt$bt_l2cap_L2CAP_LM : sock_bt_l2cap [accept4$bt_l2cap syz_init_net_socket$bt_l2cap] setsockopt$bt_l2cap_L2CAP_OPTIONS : sock_bt_l2cap [accept4$bt_l2cap syz_init_net_socket$bt_l2cap] setsockopt$bt_rfcomm_RFCOMM_LM : sock_bt_rfcomm [syz_init_net_socket$bt_rfcomm] setsockopt$inet6_IPV6_IPSEC_POLICY : uid [fstat fstat$auto geteuid ...] setsockopt$inet6_IPV6_XFRM_POLICY : uid [fstat fstat$auto geteuid ...] setsockopt$inet6_dccp_buf : sock_dccp6 [socket$inet6_dccp] setsockopt$inet6_dccp_int : sock_dccp6 [socket$inet6_dccp] setsockopt$inet_IP_IPSEC_POLICY : uid [fstat fstat$auto geteuid ...] setsockopt$inet_IP_XFRM_POLICY : uid [fstat fstat$auto geteuid ...] setsockopt$inet_dccp_buf : sock_dccp [socket$inet_dccp] setsockopt$inet_dccp_int : sock_dccp [socket$inet_dccp] setsockopt$inet_sctp6_SCTP_ADD_STREAMS : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_ASSOCINFO : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_AUTH_DEACTIVATE_KEY: assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_AUTH_DELETE_KEY : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_AUTH_KEY : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_CONTEXT : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_DEFAULT_PRINFO : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_DEFAULT_SEND_PARAM: assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_DEFAULT_SNDINFO : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_DELAYED_SACK : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_ENABLE_STREAM_RESET: assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_MAXSEG : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_MAX_BURST : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_PEER_ADDR_PARAMS : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_PEER_ADDR_THLDS : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_PRIMARY_ADDR : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_PR_SUPPORTED : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_RECONFIG_SUPPORTED: assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_RESET_ASSOC : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_RESET_STREAMS : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_RTOINFO : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_SET_PEER_PRIMARY_ADDR: assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_STREAM_SCHEDULER : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp6_SCTP_STREAM_SCHEDULER_VALUE: assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_ADD_STREAMS : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_ASSOCINFO : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_AUTH_ACTIVE_KEY : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_AUTH_DEACTIVATE_KEY: assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_AUTH_DELETE_KEY : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_AUTH_KEY : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_CONTEXT : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_DEFAULT_PRINFO : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_DEFAULT_SEND_PARAM: assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_DEFAULT_SNDINFO : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_DELAYED_SACK : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_ENABLE_STREAM_RESET: assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_MAXSEG : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_MAX_BURST : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_PEER_ADDR_PARAMS : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_PEER_ADDR_THLDS : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_PRIMARY_ADDR : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_PR_SUPPORTED : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_RECONFIG_SUPPORTED: assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_RESET_ASSOC : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_RESET_STREAMS : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_RTOINFO : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_SET_PEER_PRIMARY_ADDR: assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_STREAM_SCHEDULER : assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$inet_sctp_SCTP_STREAM_SCHEDULER_VALUE: assoc_id [getsockopt$inet_sctp6_SCTP_ASSOCINFO getsockopt$inet_sctp6_SCTP_AUTH_ACTIVE_KEY getsockopt$inet_sctp6_SCTP_CONTEXT ...] setsockopt$llc_int : sock_llc [accept4$llc syz_init_net_socket$llc] setsockopt$netrom_NETROM_IDLE : sock_netrom [accept$netrom accept4$netrom syz_init_net_socket$netrom] setsockopt$netrom_NETROM_N2 : sock_netrom [accept$netrom accept4$netrom syz_init_net_socket$netrom] setsockopt$netrom_NETROM_T1 : sock_netrom [accept$netrom accept4$netrom syz_init_net_socket$netrom] setsockopt$netrom_NETROM_T2 : sock_netrom [accept$netrom accept4$netrom syz_init_net_socket$netrom] setsockopt$netrom_NETROM_T4 : sock_netrom [accept$netrom accept4$netrom syz_init_net_socket$netrom] setsockopt$nfc_llcp_NFC_LLCP_MIUX : sock_nfc_llcp [accept$nfc_llcp accept4$nfc_llcp syz_init_net_socket$nfc_llcp] setsockopt$nfc_llcp_NFC_LLCP_RW : sock_nfc_llcp [accept$nfc_llcp accept4$nfc_llcp syz_init_net_socket$nfc_llcp] setsockopt$rose : sock_rose [accept4$rose syz_init_net_socket$rose] setsockopt$sock_attach_bpf : fd_bpf_prog [bpf$BPF_PROG_GET_FD_BY_ID bpf$BPF_PROG_RAW_TRACEPOINT_LOAD bpf$BPF_PROG_WITH_BTFID_LOAD ...] setsockopt$sock_cred : pid [capget$auto capset$auto clone$auto ...] setsockopt$sock_timeval : time_usec [getitimer getrusage getsockopt$sock_timeval ...] syz_memcpy_off$KVM_EXIT_HYPERCALL : kvm_run_ptr [mmap$KVM_VCPU] syz_memcpy_off$KVM_EXIT_MMIO : kvm_run_ptr [mmap$KVM_VCPU] BinFmtMisc : enabled Comparisons : enabled Coverage : enabled DelayKcovMmap : enabled DevlinkPCI : PCI device 0000:00:10.0 is not available ExtraCoverage : enabled Fault : enabled KCSAN : write(/sys/kernel/debug/kcsan, on) failed KcovResetIoctl : kernel does not support ioctl(KCOV_RESET_TRACE) LRWPANEmulation : enabled Leak : failed to write(kmemleak, "scan=off") NetDevices : enabled NetInjection : enabled NicVF : PCI device 0000:00:11.0 is not available SandboxAndroid : setfilecon: setxattr failed. (errno 1: Operation not permitted). . process exited with status 67. SandboxNamespace : enabled SandboxNone : enabled SandboxSetuid : enabled Swap : enabled USBEmulation : enabled VhciInjection : enabled WifiEmulation : enabled syscalls : 450/8071 2026/01/30 05:08:01 new: machine check complete 2026/01/30 05:08:04 new: adding 54308 seeds 2026/01/30 05:08:26 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:08:26 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:08:36 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:08:36 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:08:38 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:08:38 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:08:39 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:08:39 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:08:41 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:08:41 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:08:42 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:08:42 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:08:48 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:08:48 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:08:49 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:08:49 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:08:52 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:08:52 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:09:23 runner 4 connected 2026/01/30 05:09:26 runner 6 connected 2026/01/30 05:09:30 runner 5 connected 2026/01/30 05:09:31 runner 1 connected 2026/01/30 05:09:32 runner 8 connected 2026/01/30 05:09:33 runner 3 connected 2026/01/30 05:09:38 runner 2 connected 2026/01/30 05:09:39 runner 7 connected 2026/01/30 05:09:40 runner 0 connected 2026/01/30 05:09:44 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:09:44 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:09:45 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:09:45 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:10:01 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:10:01 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:10:09 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:10:09 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:10:11 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:10:11 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:10:14 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:10:14 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:10:15 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:10:15 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:10:19 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:10:19 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:10:20 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:10:20 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:10:42 runner 4 connected 2026/01/30 05:10:42 runner 6 connected 2026/01/30 05:10:50 runner 7 connected 2026/01/30 05:10:58 runner 3 connected 2026/01/30 05:11:01 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:11:01 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:11:01 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:11:01 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:11:02 runner 5 connected 2026/01/30 05:11:03 runner 2 connected 2026/01/30 05:11:04 runner 8 connected 2026/01/30 05:11:07 runner 1 connected 2026/01/30 05:11:09 runner 0 connected 2026/01/30 05:11:11 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:11:11 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:11:17 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:11:17 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:11:22 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:11:22 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:11:40 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:11:40 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:11:50 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:11:50 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:11:51 STAT { "buffer too small": 0, "candidate triage jobs": 53, "candidates": 52736, "comps overflows": 0, "corpus": 1489, "corpus [files]": 200, "corpus [symbols]": 729, "cover overflows": 1156, "coverage": 29086, "distributor delayed": 1819, "distributor undelayed": 1772, "distributor violated": 65, "exec candidate": 1572, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 0, "exec seeds": 0, "exec smash": 0, "exec total [base]": 18260, "exec total [new]": 18375, "exec triage": 4773, "executor restarts [base]": 41, "executor restarts [new]": 121, "fault jobs": 0, "fuzzer jobs": 53, "fuzzing VMs [base]": 3, "fuzzing VMs [new]": 2, "hints jobs": 0, "max signal": 29453, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 0, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 1572, "no exec duration": 174506000000, "no exec requests": 1231, "pending": 25, "prog exec time": 109, "reproducing": 0, "rpc recv": 1220351664, "rpc sent": 13009104, "signal": 28861, "smash jobs": 0, "triage jobs": 0, "vm output": 456679, "vm restarts [base]": 3, "vm restarts [new]": 27 } 2026/01/30 05:11:59 runner 6 connected 2026/01/30 05:11:59 runner 4 connected 2026/01/30 05:12:00 runner 7 connected 2026/01/30 05:12:07 runner 3 connected 2026/01/30 05:12:11 runner 5 connected 2026/01/30 05:12:26 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:12:26 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:12:28 runner 0 connected 2026/01/30 05:12:36 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:12:36 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:12:36 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:12:36 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:12:46 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:12:46 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:12:46 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:12:46 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:12:48 runner 1 connected 2026/01/30 05:12:56 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:12:56 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:12:56 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:12:56 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:13:06 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:13:06 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:13:15 runner 2 connected 2026/01/30 05:13:17 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:13:17 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:13:26 runner 3 connected 2026/01/30 05:13:33 runner 4 connected 2026/01/30 05:13:36 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:13:36 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:13:36 runner 6 connected 2026/01/30 05:13:36 runner 5 connected 2026/01/30 05:13:45 runner 7 connected 2026/01/30 05:13:46 runner 0 connected 2026/01/30 05:13:46 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:13:46 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:13:54 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:13:54 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:13:54 runner 8 connected 2026/01/30 05:14:04 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:14:04 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:14:05 runner 1 connected 2026/01/30 05:14:14 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:14:14 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:14:25 runner 2 connected 2026/01/30 05:14:35 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:14:35 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:14:37 runner 3 connected 2026/01/30 05:14:45 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:14:45 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:14:50 runner 4 connected 2026/01/30 05:15:02 runner 5 connected 2026/01/30 05:15:12 runner 6 connected 2026/01/30 05:15:19 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:15:19 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:15:29 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:15:29 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:15:32 runner 7 connected 2026/01/30 05:15:37 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:15:37 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:15:42 runner 0 connected 2026/01/30 05:15:43 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:15:43 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:15:47 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:15:47 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:15:53 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:15:53 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:15:57 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:15:57 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:16:03 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:16:03 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:16:10 runner 2 connected 2026/01/30 05:16:27 runner 4 connected 2026/01/30 05:16:33 runner 6 connected 2026/01/30 05:16:36 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:16:36 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:16:39 runner 8 connected 2026/01/30 05:16:42 runner 5 connected 2026/01/30 05:16:43 runner 1 connected 2026/01/30 05:16:46 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:16:46 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:16:47 runner 3 connected 2026/01/30 05:16:51 STAT { "buffer too small": 0, "candidate triage jobs": 46, "candidates": 51392, "comps overflows": 0, "corpus": 2821, "corpus [files]": 325, "corpus [symbols]": 1203, "cover overflows": 3214, "coverage": 35386, "distributor delayed": 3945, "distributor undelayed": 3912, "distributor violated": 288, "exec candidate": 2916, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 0, "exec seeds": 0, "exec smash": 0, "exec total [base]": 51775, "exec total [new]": 51951, "exec triage": 8932, "executor restarts [base]": 41, "executor restarts [new]": 202, "fault jobs": 0, "fuzzer jobs": 46, "fuzzing VMs [base]": 3, "fuzzing VMs [new]": 5, "hints jobs": 0, "max signal": 35742, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 0, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 2916, "no exec duration": 646165000000, "no exec requests": 5711, "pending": 51, "prog exec time": 80, "reproducing": 0, "rpc recv": 2496864104, "rpc sent": 33721240, "signal": 34923, "smash jobs": 0, "triage jobs": 0, "vm output": 940960, "vm restarts [base]": 3, "vm restarts [new]": 57 } 2026/01/30 05:16:52 runner 7 connected 2026/01/30 05:17:05 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:17:05 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:17:14 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:17:14 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:17:15 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:17:15 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:17:24 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:17:24 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:17:26 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:17:26 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:17:32 runner 0 connected 2026/01/30 05:17:34 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:17:34 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:17:36 runner 2 connected 2026/01/30 05:17:36 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:17:36 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:17:52 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:17:52 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:18:02 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:18:02 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:18:02 runner 4 connected 2026/01/30 05:18:04 runner 5 connected 2026/01/30 05:18:05 runner 1 connected 2026/01/30 05:18:13 runner 3 connected 2026/01/30 05:18:15 runner 6 connected 2026/01/30 05:18:23 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:18:23 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:18:24 runner 7 connected 2026/01/30 05:18:24 runner 8 connected 2026/01/30 05:18:41 runner 0 connected 2026/01/30 05:18:52 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:18:52 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:18:58 runner 2 connected 2026/01/30 05:19:01 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:19:01 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:19:02 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:19:02 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:19:11 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:19:11 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:19:12 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:19:12 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:19:20 runner 5 connected 2026/01/30 05:19:22 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:19:22 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:19:22 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:19:22 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:19:28 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:19:28 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:19:39 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:19:39 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:19:42 runner 8 connected 2026/01/30 05:19:51 runner 1 connected 2026/01/30 05:19:51 runner 4 connected 2026/01/30 05:20:00 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:20:00 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:20:00 runner 6 connected 2026/01/30 05:20:01 runner 3 connected 2026/01/30 05:20:10 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:20:10 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:20:11 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:20:11 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:20:12 runner 0 connected 2026/01/30 05:20:13 runner 2 connected 2026/01/30 05:20:16 runner 7 connected 2026/01/30 05:20:21 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:20:21 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:20:21 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:20:21 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:20:28 runner 5 connected 2026/01/30 05:20:33 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:20:33 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:20:35 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:20:35 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:20:43 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:20:43 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:20:48 runner 8 connected 2026/01/30 05:20:59 runner 4 connected 2026/01/30 05:21:01 runner 1 connected 2026/01/30 05:21:10 runner 3 connected 2026/01/30 05:21:11 runner 6 connected 2026/01/30 05:21:19 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:21:19 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:21:20 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:21:20 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:21:21 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:21:21 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:21:23 runner 7 connected 2026/01/30 05:21:28 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:21:28 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:21:30 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:21:30 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:21:30 runner 0 connected 2026/01/30 05:21:31 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:21:31 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:21:33 runner 2 connected 2026/01/30 05:21:42 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:21:42 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:21:51 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 51047, "comps overflows": 0, "corpus": 3201, "corpus [files]": 353, "corpus [symbols]": 1342, "cover overflows": 4426, "coverage": 36376, "distributor delayed": 4446, "distributor undelayed": 4446, "distributor violated": 291, "exec candidate": 3261, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 0, "exec seeds": 0, "exec smash": 0, "exec total [base]": 73644, "exec total [new]": 73906, "exec triage": 10050, "executor restarts [base]": 41, "executor restarts [new]": 293, "fault jobs": 0, "fuzzer jobs": 0, "fuzzing VMs [base]": 3, "fuzzing VMs [new]": 1, "hints jobs": 0, "max signal": 36485, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 0, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 3261, "no exec duration": 1277522000000, "no exec requests": 10303, "pending": 85, "prog exec time": 62, "reproducing": 0, "rpc recv": 3661601100, "rpc sent": 50718080, "signal": 35898, "smash jobs": 0, "triage jobs": 0, "vm output": 1457419, "vm restarts [base]": 3, "vm restarts [new]": 87 } 2026/01/30 05:21:54 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:21:54 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:22:08 runner 5 connected 2026/01/30 05:22:08 runner 8 connected 2026/01/30 05:22:09 runner 4 connected 2026/01/30 05:22:17 runner 1 connected 2026/01/30 05:22:19 runner 3 connected 2026/01/30 05:22:22 runner 6 connected 2026/01/30 05:22:32 runner 7 connected 2026/01/30 05:22:32 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:22:32 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:22:42 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:22:42 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:22:45 runner 2 connected 2026/01/30 05:22:53 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:22:53 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:23:03 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:23:03 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:23:07 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:23:07 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:23:08 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:23:08 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:23:18 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:23:18 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:23:29 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:23:29 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:23:30 runner 0 connected 2026/01/30 05:23:39 runner 4 connected 2026/01/30 05:23:50 runner 5 connected 2026/01/30 05:23:59 runner 3 connected 2026/01/30 05:24:03 runner 2 connected 2026/01/30 05:24:04 runner 8 connected 2026/01/30 05:24:07 runner 1 connected 2026/01/30 05:24:08 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:24:08 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:24:18 runner 7 connected 2026/01/30 05:24:41 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:24:41 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:24:49 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:24:49 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:24:50 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:24:50 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:24:51 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:24:51 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:25:00 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:25:00 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:25:02 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:25:02 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:25:05 runner 4 connected 2026/01/30 05:25:10 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:25:10 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:25:39 runner 6 connected 2026/01/30 05:25:46 runner 0 connected 2026/01/30 05:25:47 runner 1 connected 2026/01/30 05:25:48 runner 2 connected 2026/01/30 05:25:56 runner 3 connected 2026/01/30 05:25:57 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:25:57 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:25:58 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:25:58 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:25:58 runner 8 connected 2026/01/30 05:25:59 runner 5 connected 2026/01/30 05:26:07 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:26:07 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:26:08 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:26:08 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:26:13 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:26:13 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:26:18 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:26:18 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:26:23 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:26:23 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:26:33 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:26:33 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:26:44 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:26:44 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:26:47 runner 4 connected 2026/01/30 05:26:47 runner 7 connected 2026/01/30 05:26:51 STAT { "buffer too small": 0, "candidate triage jobs": 2, "candidates": 51026, "comps overflows": 0, "corpus": 3212, "corpus [files]": 355, "corpus [symbols]": 1351, "cover overflows": 6690, "coverage": 36426, "distributor delayed": 4484, "distributor undelayed": 4482, "distributor violated": 292, "exec candidate": 3282, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 0, "exec seeds": 0, "exec smash": 0, "exec total [base]": 111902, "exec total [new]": 113250, "exec triage": 10130, "executor restarts [base]": 41, "executor restarts [new]": 365, "fault jobs": 0, "fuzzer jobs": 2, "fuzzing VMs [base]": 3, "fuzzing VMs [new]": 0, "hints jobs": 0, "max signal": 36554, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 0, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 3282, "no exec duration": 1766862000000, "no exec requests": 15389, "pending": 111, "prog exec time": 0, "reproducing": 0, "rpc recv": 4685872764, "rpc sent": 70935848, "signal": 35944, "smash jobs": 0, "triage jobs": 0, "vm output": 1853139, "vm restarts [base]": 3, "vm restarts [new]": 113 } 2026/01/30 05:26:57 runner 2 connected 2026/01/30 05:26:57 runner 6 connected 2026/01/30 05:27:02 runner 0 connected 2026/01/30 05:27:08 runner 3 connected 2026/01/30 05:27:12 runner 1 connected 2026/01/30 05:27:31 runner 5 connected 2026/01/30 05:27:34 runner 8 connected 2026/01/30 05:27:37 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:27:37 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:27:39 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:27:39 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:27:40 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:27:40 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:27:43 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:27:43 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:27:44 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:27:44 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:27:50 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:27:50 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:28:05 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:28:05 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:28:06 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:28:06 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:28:06 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:28:06 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:28:26 runner 2 connected 2026/01/30 05:28:29 runner 0 connected 2026/01/30 05:28:30 runner 1 connected 2026/01/30 05:28:32 runner 6 connected 2026/01/30 05:28:33 runner 4 connected 2026/01/30 05:28:40 runner 7 connected 2026/01/30 05:28:47 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:28:47 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:28:54 runner 3 connected 2026/01/30 05:28:54 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:28:54 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:28:55 runner 8 connected 2026/01/30 05:28:57 runner 5 connected 2026/01/30 05:28:58 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:28:58 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:29:44 runner 2 connected 2026/01/30 05:29:44 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:29:44 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:29:46 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:29:46 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:29:52 runner 6 connected 2026/01/30 05:29:52 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:29:52 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:29:55 runner 4 connected 2026/01/30 05:29:56 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:29:56 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:29:57 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:29:57 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:30:02 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:30:02 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:30:07 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:30:07 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:30:17 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:30:17 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:30:37 runner 3 connected 2026/01/30 05:30:41 runner 8 connected 2026/01/30 05:30:46 runner 1 connected 2026/01/30 05:30:49 runner 0 connected 2026/01/30 05:30:51 runner 7 connected 2026/01/30 05:30:53 runner 5 connected 2026/01/30 05:30:58 runner 2 connected 2026/01/30 05:31:07 runner 6 connected 2026/01/30 05:31:07 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:31:07 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:31:18 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:31:18 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:31:47 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:31:47 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:31:49 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:31:49 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:31:51 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 50861, "comps overflows": 0, "corpus": 3214, "corpus [files]": 355, "corpus [symbols]": 1351, "cover overflows": 9085, "coverage": 36429, "distributor delayed": 4513, "distributor undelayed": 4513, "distributor violated": 293, "exec candidate": 3447, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 0, "exec seeds": 0, "exec smash": 0, "exec total [base]": 150123, "exec total [new]": 157595, "exec triage": 10202, "executor restarts [base]": 41, "executor restarts [new]": 452, "fault jobs": 0, "fuzzer jobs": 0, "fuzzing VMs [base]": 3, "fuzzing VMs [new]": 3, "hints jobs": 0, "max signal": 36579, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 0, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 3295, "no exec duration": 2114265000000, "no exec requests": 19006, "pending": 135, "prog exec time": 53, "reproducing": 0, "rpc recv": 5862841084, "rpc sent": 93327496, "signal": 35947, "smash jobs": 0, "triage jobs": 0, "vm output": 2324803, "vm restarts [base]": 3, "vm restarts [new]": 140 } 2026/01/30 05:31:54 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:31:54 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:31:56 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:31:56 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:32:04 runner 4 connected 2026/01/30 05:32:15 runner 7 connected 2026/01/30 05:32:27 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:32:27 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:32:35 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:32:35 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:32:37 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:32:37 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:32:45 runner 0 connected 2026/01/30 05:32:45 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:32:45 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:32:46 runner 2 connected 2026/01/30 05:32:48 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:32:48 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:32:50 runner 8 connected 2026/01/30 05:32:53 runner 3 connected 2026/01/30 05:33:06 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:33:06 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:33:16 runner 4 connected 2026/01/30 05:33:24 runner 1 connected 2026/01/30 05:33:27 runner 6 connected 2026/01/30 05:33:38 runner 7 connected 2026/01/30 05:33:42 runner 5 connected 2026/01/30 05:33:53 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:33:53 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:33:55 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:33:55 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:34:02 runner 2 connected 2026/01/30 05:34:03 patched crashed: WARNING in __folio_large_mapcount_sanity_checks [need repro = true] 2026/01/30 05:34:03 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:34:53 runner 0 connected 2026/01/30 05:34:58 runner 4 connected 2026/01/30 05:35:00 runner 5 connected 2026/01/30 05:35:51 triaged 94.9% of the corpus 2026/01/30 05:35:51 starting bug reproductions 2026/01/30 05:35:51 starting bug reproductions (max 6 VMs, 4 repros) 2026/01/30 05:35:51 start reproducing 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:36:51 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 0, "comps overflows": 83, "corpus": 3369, "corpus [files]": 396, "corpus [symbols]": 1439, "cover overflows": 13784, "coverage": 37429, "distributor delayed": 4672, "distributor undelayed": 4672, "distributor violated": 295, "exec candidate": 54308, "exec collide": 459, "exec fuzz": 773, "exec gen": 40, "exec hints": 120, "exec inject": 0, "exec minimize": 2932, "exec retries": 0, "exec seeds": 314, "exec smash": 526, "exec total [base]": 193201, "exec total [new]": 223691, "exec triage": 10812, "executor restarts [base]": 41, "executor restarts [new]": 494, "fault jobs": 0, "fuzzer jobs": 312, "fuzzing VMs [base]": 3, "fuzzing VMs [new]": 7, "hints jobs": 123, "max signal": 38271, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 1705, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 3536, "no exec duration": 2182466000000, "no exec requests": 20105, "pending": 145, "prog exec time": 149, "reproducing": 1, "rpc recv": 6784373308, "rpc sent": 120288368, "signal": 36842, "smash jobs": 159, "triage jobs": 30, "vm output": 2630814, "vm restarts [base]": 3, "vm restarts [new]": 155 } 2026/01/30 05:37:19 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 05:40:42 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 05:41:27 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 05:41:51 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 0, "comps overflows": 671, "corpus": 3856, "corpus [files]": 483, "corpus [symbols]": 1696, "cover overflows": 22653, "coverage": 39612, "distributor delayed": 5106, "distributor undelayed": 5106, "distributor violated": 295, "exec candidate": 54308, "exec collide": 3396, "exec fuzz": 6353, "exec gen": 334, "exec hints": 1737, "exec inject": 0, "exec minimize": 12191, "exec retries": 0, "exec seeds": 1806, "exec smash": 6228, "exec total [base]": 205931, "exec total [new]": 252132, "exec triage": 12372, "executor restarts [base]": 41, "executor restarts [new]": 495, "fault jobs": 0, "fuzzer jobs": 907, "fuzzing VMs [base]": 3, "fuzzing VMs [new]": 7, "hints jobs": 244, "max signal": 40769, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 7200, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 4179, "no exec duration": 2182471000000, "no exec requests": 20106, "pending": 145, "prog exec time": 241, "reproducing": 1, "rpc recv": 7796910040, "rpc sent": 155245768, "signal": 38858, "smash jobs": 646, "triage jobs": 17, "vm output": 3200448, "vm restarts [base]": 3, "vm restarts [new]": 155 } 2026/01/30 05:42:09 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 05:42:09 repro finished 'WARNING in __folio_large_mapcount_sanity_checks', repro=true crepro=false desc='WARNING in __folio_large_mapcount_sanity_checks' hub=false from_dashboard=false 2026/01/30 05:42:09 found repro for "WARNING in __folio_large_mapcount_sanity_checks" (orig title: "-SAME-", reliability: 1), took 6.31 minutes 2026/01/30 05:42:09 "WARNING in __folio_large_mapcount_sanity_checks": saved crash log into 1769751729.crash.log 2026/01/30 05:42:09 "WARNING in __folio_large_mapcount_sanity_checks": saved repro log into 1769751729.repro.log 2026/01/30 05:42:10 start reproducing 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:42:59 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 05:44:19 attempt #0 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 05:46:01 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 05:46:26 attempt #1 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 05:46:51 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 0, "comps overflows": 1006, "corpus": 4153, "corpus [files]": 529, "corpus [symbols]": 1843, "cover overflows": 31420, "coverage": 41186, "distributor delayed": 5319, "distributor undelayed": 5319, "distributor violated": 295, "exec candidate": 54308, "exec collide": 7331, "exec fuzz": 13819, "exec gen": 719, "exec hints": 3802, "exec inject": 0, "exec minimize": 17391, "exec retries": 0, "exec seeds": 2722, "exec smash": 15028, "exec total [base]": 214444, "exec total [new]": 281780, "exec triage": 13254, "executor restarts [base]": 41, "executor restarts [new]": 495, "fault jobs": 0, "fuzzer jobs": 1007, "fuzzing VMs [base]": 2, "fuzzing VMs [new]": 7, "hints jobs": 238, "max signal": 42385, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 10244, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 4555, "no exec duration": 2182476000000, "no exec requests": 20107, "pending": 144, "prog exec time": 216, "reproducing": 1, "rpc recv": 8397971012, "rpc sent": 190762152, "signal": 40353, "smash jobs": 759, "triage jobs": 10, "vm output": 3820304, "vm restarts [base]": 3, "vm restarts [new]": 155 } 2026/01/30 05:46:56 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 05:47:26 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 05:47:26 repro finished 'WARNING in __folio_large_mapcount_sanity_checks', repro=true crepro=false desc='WARNING in __folio_large_mapcount_sanity_checks' hub=false from_dashboard=false 2026/01/30 05:47:26 found repro for "WARNING in __folio_large_mapcount_sanity_checks" (orig title: "-SAME-", reliability: 1), took 5.28 minutes 2026/01/30 05:47:26 "WARNING in __folio_large_mapcount_sanity_checks": saved crash log into 1769752046.crash.log 2026/01/30 05:47:26 "WARNING in __folio_large_mapcount_sanity_checks": saved repro log into 1769752046.repro.log 2026/01/30 05:47:26 start reproducing 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:48:21 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 05:48:27 attempt #2 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 05:48:27 patched-only: WARNING in __folio_large_mapcount_sanity_checks 2026/01/30 05:48:27 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks (full)' 2026/01/30 05:48:27 start reproducing 'WARNING in __folio_large_mapcount_sanity_checks (full)' 2026/01/30 05:48:51 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 05:49:24 runner 0 connected 2026/01/30 05:49:29 attempt #0 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 05:49:52 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 05:51:30 attempt #1 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 05:51:51 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 0, "comps overflows": 1222, "corpus": 4394, "corpus [files]": 547, "corpus [symbols]": 1946, "cover overflows": 38696, "coverage": 42432, "distributor delayed": 5501, "distributor undelayed": 5501, "distributor violated": 295, "exec candidate": 54308, "exec collide": 10852, "exec fuzz": 20717, "exec gen": 1077, "exec hints": 5744, "exec inject": 0, "exec minimize": 21952, "exec retries": 0, "exec seeds": 3475, "exec smash": 23118, "exec total [base]": 221003, "exec total [new]": 308576, "exec triage": 13929, "executor restarts [base]": 44, "executor restarts [new]": 496, "fault jobs": 0, "fuzzer jobs": 785, "fuzzing VMs [base]": 2, "fuzzing VMs [new]": 6, "hints jobs": 206, "max signal": 43750, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 12769, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 4850, "no exec duration": 2182575000000, "no exec requests": 20108, "pending": 143, "prog exec time": 211, "reproducing": 2, "rpc recv": 8885796296, "rpc sent": 223411160, "signal": 41512, "smash jobs": 570, "triage jobs": 9, "vm output": 4422621, "vm restarts [base]": 4, "vm restarts [new]": 155 } 2026/01/30 05:53:11 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 05:53:32 attempt #2 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 05:53:32 patched-only: WARNING in __folio_large_mapcount_sanity_checks 2026/01/30 05:53:32 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks (full)' 2026/01/30 05:54:29 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 05:54:30 runner 1 connected 2026/01/30 05:54:37 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 05:55:07 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 05:55:52 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 05:55:59 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 05:55:59 repro finished 'WARNING in __folio_large_mapcount_sanity_checks', repro=true crepro=false desc='WARNING in __folio_large_mapcount_sanity_checks' hub=false from_dashboard=false 2026/01/30 05:55:59 found repro for "WARNING in __folio_large_mapcount_sanity_checks" (orig title: "-SAME-", reliability: 1), took 8.54 minutes 2026/01/30 05:55:59 "WARNING in __folio_large_mapcount_sanity_checks": saved crash log into 1769752559.crash.log 2026/01/30 05:55:59 "WARNING in __folio_large_mapcount_sanity_checks": saved repro log into 1769752559.repro.log 2026/01/30 05:55:59 start reproducing 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 05:56:18 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 05:56:51 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 0, "comps overflows": 1441, "corpus": 4632, "corpus [files]": 559, "corpus [symbols]": 2034, "cover overflows": 45185, "coverage": 43992, "distributor delayed": 5682, "distributor undelayed": 5682, "distributor violated": 295, "exec candidate": 54308, "exec collide": 14428, "exec fuzz": 27485, "exec gen": 1454, "exec hints": 7735, "exec inject": 0, "exec minimize": 26216, "exec retries": 0, "exec seeds": 4260, "exec smash": 31056, "exec total [base]": 230747, "exec total [new]": 334895, "exec triage": 14545, "executor restarts [base]": 47, "executor restarts [new]": 498, "fault jobs": 0, "fuzzer jobs": 568, "fuzzing VMs [base]": 2, "fuzzing VMs [new]": 6, "hints jobs": 138, "max signal": 45375, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 15137, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 1, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 5130, "no exec duration": 2182650000000, "no exec requests": 20111, "pending": 143, "prog exec time": 209, "reproducing": 2, "rpc recv": 9452493996, "rpc sent": 258089120, "signal": 42970, "smash jobs": 421, "triage jobs": 9, "vm output": 5224073, "vm restarts [base]": 5, "vm restarts [new]": 155 } 2026/01/30 05:57:10 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 05:57:16 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 05:57:30 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 05:58:00 attempt #0 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 05:58:29 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 05:58:44 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 05:59:55 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:00:00 attempt #1 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 06:00:28 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:01:51 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 0, "comps overflows": 1590, "corpus": 4834, "corpus [files]": 563, "corpus [symbols]": 2100, "cover overflows": 51102, "coverage": 45531, "distributor delayed": 5827, "distributor undelayed": 5827, "distributor violated": 295, "exec candidate": 54308, "exec collide": 17907, "exec fuzz": 34149, "exec gen": 1821, "exec hints": 9793, "exec inject": 0, "exec minimize": 29862, "exec retries": 0, "exec seeds": 4894, "exec smash": 38876, "exec total [base]": 239071, "exec total [new]": 360102, "exec triage": 15085, "executor restarts [base]": 48, "executor restarts [new]": 501, "fault jobs": 0, "fuzzer jobs": 280, "fuzzing VMs [base]": 2, "fuzzing VMs [new]": 6, "hints jobs": 81, "max signal": 46928, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 17122, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 1, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 5367, "no exec duration": 2182650000000, "no exec requests": 20111, "pending": 143, "prog exec time": 203, "reproducing": 2, "rpc recv": 9895316976, "rpc sent": 290419872, "signal": 44184, "smash jobs": 195, "triage jobs": 4, "vm output": 6170150, "vm restarts [base]": 5, "vm restarts [new]": 155 } 2026/01/30 06:01:54 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:01:54 repro finished 'WARNING in __folio_large_mapcount_sanity_checks (full)', repro=true crepro=true desc='WARNING in __folio_large_mapcount_sanity_checks' hub=false from_dashboard=false 2026/01/30 06:01:54 found repro for "WARNING in __folio_large_mapcount_sanity_checks" (orig title: "-SAME-", reliability: 1), took 13.45 minutes 2026/01/30 06:01:54 "WARNING in __folio_large_mapcount_sanity_checks": saved crash log into 1769752914.crash.log 2026/01/30 06:01:54 start reproducing 'WARNING in __folio_large_mapcount_sanity_checks (full)' 2026/01/30 06:01:54 "WARNING in __folio_large_mapcount_sanity_checks": saved repro log into 1769752914.repro.log 2026/01/30 06:02:03 attempt #2 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 06:02:03 patched-only: WARNING in __folio_large_mapcount_sanity_checks 2026/01/30 06:02:03 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks (full)' 2026/01/30 06:02:04 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:02:59 runner 0 connected 2026/01/30 06:03:00 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:03:20 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:03:54 attempt #0 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 06:04:24 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:04:24 repro finished 'WARNING in __folio_large_mapcount_sanity_checks', repro=true crepro=false desc='WARNING in __folio_large_mapcount_sanity_checks' hub=false from_dashboard=false 2026/01/30 06:04:24 found repro for "WARNING in __folio_large_mapcount_sanity_checks" (orig title: "-SAME-", reliability: 1), took 8.43 minutes 2026/01/30 06:04:24 "WARNING in __folio_large_mapcount_sanity_checks": saved crash log into 1769753064.crash.log 2026/01/30 06:04:24 "WARNING in __folio_large_mapcount_sanity_checks": saved repro log into 1769753064.repro.log 2026/01/30 06:04:24 start reproducing 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 06:05:13 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:05:54 attempt #1 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 06:06:25 attempt #0 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 06:06:30 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:06:51 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 0, "comps overflows": 1653, "corpus": 4974, "corpus [files]": 570, "corpus [symbols]": 2152, "cover overflows": 57275, "coverage": 46184, "distributor delayed": 5927, "distributor undelayed": 5927, "distributor violated": 295, "exec candidate": 54308, "exec collide": 22701, "exec fuzz": 43236, "exec gen": 2287, "exec hints": 14615, "exec inject": 0, "exec minimize": 32593, "exec retries": 0, "exec seeds": 5345, "exec smash": 44487, "exec total [base]": 244896, "exec total [new]": 388443, "exec triage": 15465, "executor restarts [base]": 51, "executor restarts [new]": 501, "fault jobs": 0, "fuzzer jobs": 16, "fuzzing VMs [base]": 1, "fuzzing VMs [new]": 6, "hints jobs": 6, "max signal": 47807, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 18590, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 2, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 5537, "no exec duration": 2182666000000, "no exec requests": 20112, "pending": 142, "prog exec time": 200, "reproducing": 2, "rpc recv": 10212587408, "rpc sent": 324145336, "signal": 44812, "smash jobs": 5, "triage jobs": 5, "vm output": 6919936, "vm restarts [base]": 6, "vm restarts [new]": 155 } 2026/01/30 06:07:32 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:07:46 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:07:57 attempt #2 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 06:07:57 patched-only: WARNING in __folio_large_mapcount_sanity_checks 2026/01/30 06:08:27 attempt #1 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 06:08:44 runner 2 connected 2026/01/30 06:08:49 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:09:23 runner 0 connected 2026/01/30 06:09:48 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:09:58 attempt #2 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 06:09:58 patched-only: WARNING in __folio_large_mapcount_sanity_checks 2026/01/30 06:09:58 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks (full)' 2026/01/30 06:10:39 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:10:55 runner 1 connected 2026/01/30 06:11:13 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:11:38 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:11:51 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 0, "comps overflows": 1775, "corpus": 5042, "corpus [files]": 579, "corpus [symbols]": 2190, "cover overflows": 64233, "coverage": 46469, "distributor delayed": 5990, "distributor undelayed": 5990, "distributor violated": 295, "exec candidate": 54308, "exec collide": 29318, "exec fuzz": 56048, "exec gen": 2973, "exec hints": 15207, "exec inject": 0, "exec minimize": 34201, "exec retries": 0, "exec seeds": 5557, "exec smash": 46205, "exec total [base]": 250241, "exec total [new]": 412918, "exec triage": 15697, "executor restarts [base]": 60, "executor restarts [new]": 501, "fault jobs": 0, "fuzzer jobs": 18, "fuzzing VMs [base]": 3, "fuzzing VMs [new]": 6, "hints jobs": 3, "max signal": 48191, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 19386, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 2, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 5632, "no exec duration": 2182666000000, "no exec requests": 20112, "pending": 143, "prog exec time": 223, "reproducing": 2, "rpc recv": 10584523384, "rpc sent": 357637344, "signal": 45037, "smash jobs": 8, "triage jobs": 7, "vm output": 7737733, "vm restarts [base]": 9, "vm restarts [new]": 155 } 2026/01/30 06:12:09 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:12:33 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:13:03 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:13:03 repro finished 'WARNING in __folio_large_mapcount_sanity_checks', repro=true crepro=false desc='WARNING in __folio_large_mapcount_sanity_checks' hub=false from_dashboard=false 2026/01/30 06:13:03 found repro for "WARNING in __folio_large_mapcount_sanity_checks" (orig title: "-SAME-", reliability: 1), took 8.65 minutes 2026/01/30 06:13:03 "WARNING in __folio_large_mapcount_sanity_checks": saved crash log into 1769753583.crash.log 2026/01/30 06:13:03 "WARNING in __folio_large_mapcount_sanity_checks": saved repro log into 1769753583.repro.log 2026/01/30 06:13:03 start reproducing 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 06:13:20 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:14:02 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:14:16 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:14:45 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:15:05 attempt #0 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 06:15:30 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:15:42 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:16:12 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:16:51 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 0, "comps overflows": 1871, "corpus": 5134, "corpus [files]": 581, "corpus [symbols]": 2233, "cover overflows": 70865, "coverage": 46902, "distributor delayed": 6069, "distributor undelayed": 6069, "distributor violated": 295, "exec candidate": 54308, "exec collide": 35377, "exec fuzz": 67482, "exec gen": 3604, "exec hints": 15543, "exec inject": 0, "exec minimize": 36264, "exec retries": 0, "exec seeds": 5841, "exec smash": 48646, "exec total [base]": 259027, "exec total [new]": 436439, "exec triage": 15968, "executor restarts [base]": 60, "executor restarts [new]": 502, "fault jobs": 0, "fuzzer jobs": 9, "fuzzing VMs [base]": 2, "fuzzing VMs [new]": 6, "hints jobs": 0, "max signal": 48733, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 20468, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 2, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 5751, "no exec duration": 2182666000000, "no exec requests": 20112, "pending": 142, "prog exec time": 209, "reproducing": 2, "rpc recv": 10986117652, "rpc sent": 397241096, "signal": 45454, "smash jobs": 2, "triage jobs": 7, "vm output": 8448712, "vm restarts [base]": 9, "vm restarts [new]": 155 } 2026/01/30 06:17:06 attempt #1 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 06:17:10 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:17:10 repro finished 'WARNING in __folio_large_mapcount_sanity_checks (full)', repro=true crepro=true desc='WARNING in __folio_large_mapcount_sanity_checks' hub=false from_dashboard=false 2026/01/30 06:17:10 found repro for "WARNING in __folio_large_mapcount_sanity_checks" (orig title: "-SAME-", reliability: 1), took 15.26 minutes 2026/01/30 06:17:10 start reproducing 'WARNING in __folio_large_mapcount_sanity_checks (full)' 2026/01/30 06:17:10 "WARNING in __folio_large_mapcount_sanity_checks": saved crash log into 1769753830.crash.log 2026/01/30 06:17:10 "WARNING in __folio_large_mapcount_sanity_checks": saved repro log into 1769753830.repro.log 2026/01/30 06:17:39 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:19:06 attempt #2 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 06:19:06 patched-only: WARNING in __folio_large_mapcount_sanity_checks 2026/01/30 06:19:06 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks (full)' 2026/01/30 06:19:11 attempt #0 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 06:20:11 runner 0 connected 2026/01/30 06:21:12 attempt #1 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 06:21:50 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:21:51 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 0, "comps overflows": 1958, "corpus": 5190, "corpus [files]": 586, "corpus [symbols]": 2259, "cover overflows": 77584, "coverage": 47131, "distributor delayed": 6102, "distributor undelayed": 6102, "distributor violated": 295, "exec candidate": 54308, "exec collide": 42243, "exec fuzz": 80408, "exec gen": 4288, "exec hints": 15628, "exec inject": 0, "exec minimize": 37976, "exec retries": 0, "exec seeds": 6015, "exec smash": 50053, "exec total [base]": 264379, "exec total [new]": 460438, "exec triage": 16113, "executor restarts [base]": 63, "executor restarts [new]": 503, "fault jobs": 0, "fuzzer jobs": 12, "fuzzing VMs [base]": 2, "fuzzing VMs [new]": 6, "hints jobs": 2, "max signal": 48980, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 21305, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 3, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 5816, "no exec duration": 2182666000000, "no exec requests": 20112, "pending": 142, "prog exec time": 239, "reproducing": 2, "rpc recv": 11289983932, "rpc sent": 430509256, "signal": 45653, "smash jobs": 5, "triage jobs": 5, "vm output": 9314874, "vm restarts [base]": 10, "vm restarts [new]": 155 } 2026/01/30 06:21:55 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:22:24 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:23:10 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:23:13 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:23:14 attempt #2 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 06:23:14 patched-only: WARNING in __folio_large_mapcount_sanity_checks 2026/01/30 06:23:36 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:24:11 runner 1 connected 2026/01/30 06:24:25 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:24:37 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:24:37 repro finished 'WARNING in __folio_large_mapcount_sanity_checks', repro=true crepro=false desc='WARNING in __folio_large_mapcount_sanity_checks' hub=false from_dashboard=false 2026/01/30 06:24:37 found repro for "WARNING in __folio_large_mapcount_sanity_checks" (orig title: "-SAME-", reliability: 1), took 11.56 minutes 2026/01/30 06:24:37 "WARNING in __folio_large_mapcount_sanity_checks": saved crash log into 1769754277.crash.log 2026/01/30 06:24:37 "WARNING in __folio_large_mapcount_sanity_checks": saved repro log into 1769754277.repro.log 2026/01/30 06:24:37 start reproducing 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 06:24:48 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:25:47 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:25:51 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:26:03 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:26:37 attempt #0 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 06:26:51 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 0, "comps overflows": 2012, "corpus": 5222, "corpus [files]": 588, "corpus [symbols]": 2277, "cover overflows": 83416, "coverage": 47228, "distributor delayed": 6137, "distributor undelayed": 6137, "distributor violated": 295, "exec candidate": 54308, "exec collide": 48549, "exec fuzz": 92651, "exec gen": 4919, "exec hints": 15808, "exec inject": 0, "exec minimize": 38888, "exec retries": 0, "exec seeds": 6111, "exec smash": 50924, "exec total [base]": 272513, "exec total [new]": 481786, "exec triage": 16225, "executor restarts [base]": 66, "executor restarts [new]": 504, "fault jobs": 0, "fuzzer jobs": 5, "fuzzing VMs [base]": 2, "fuzzing VMs [new]": 6, "hints jobs": 0, "max signal": 49185, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 21769, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 3, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 5862, "no exec duration": 2182666000000, "no exec requests": 20112, "pending": 141, "prog exec time": 239, "reproducing": 2, "rpc recv": 11661185652, "rpc sent": 464042608, "signal": 45748, "smash jobs": 1, "triage jobs": 4, "vm output": 9984494, "vm restarts [base]": 11, "vm restarts [new]": 155 } 2026/01/30 06:27:15 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:27:21 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:28:38 attempt #1 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 06:28:43 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:28:47 patched crashed: no output from test machine [need repro = false] 2026/01/30 06:29:27 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:29:27 repro finished 'WARNING in __folio_large_mapcount_sanity_checks (full)', repro=true crepro=true desc='WARNING in __folio_large_mapcount_sanity_checks' hub=false from_dashboard=false 2026/01/30 06:29:27 found repro for "WARNING in __folio_large_mapcount_sanity_checks" (orig title: "-SAME-", reliability: 1), took 12.29 minutes 2026/01/30 06:29:27 start reproducing 'WARNING in __folio_large_mapcount_sanity_checks (full)' 2026/01/30 06:29:27 "WARNING in __folio_large_mapcount_sanity_checks": saved crash log into 1769754567.crash.log 2026/01/30 06:29:27 "WARNING in __folio_large_mapcount_sanity_checks": saved repro log into 1769754567.repro.log 2026/01/30 06:29:45 runner 4 connected 2026/01/30 06:30:27 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:30:40 attempt #2 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 06:30:40 patched-only: WARNING in __folio_large_mapcount_sanity_checks 2026/01/30 06:30:40 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks (full)' 2026/01/30 06:30:51 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:31:29 attempt #0 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 06:31:38 runner 0 connected 2026/01/30 06:31:51 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 0, "comps overflows": 2063, "corpus": 5259, "corpus [files]": 589, "corpus [symbols]": 2303, "cover overflows": 89091, "coverage": 47403, "distributor delayed": 6191, "distributor undelayed": 6191, "distributor violated": 295, "exec candidate": 54308, "exec collide": 54947, "exec fuzz": 104783, "exec gen": 5538, "exec hints": 15885, "exec inject": 0, "exec minimize": 39720, "exec retries": 0, "exec seeds": 6225, "exec smash": 51839, "exec total [base]": 278532, "exec total [new]": 503033, "exec triage": 16382, "executor restarts [base]": 69, "executor restarts [new]": 509, "fault jobs": 0, "fuzzer jobs": 9, "fuzzing VMs [base]": 2, "fuzzing VMs [new]": 6, "hints jobs": 3, "max signal": 49322, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 22225, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 3, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 5923, "no exec duration": 2182666000000, "no exec requests": 20112, "pending": 141, "prog exec time": 209, "reproducing": 2, "rpc recv": 11984777292, "rpc sent": 495472176, "signal": 45921, "smash jobs": 3, "triage jobs": 3, "vm output": 10714200, "vm restarts [base]": 12, "vm restarts [new]": 156 } 2026/01/30 06:31:55 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:32:29 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:32:29 repro finished 'WARNING in __folio_large_mapcount_sanity_checks', repro=true crepro=false desc='WARNING in __folio_large_mapcount_sanity_checks' hub=false from_dashboard=false 2026/01/30 06:32:29 found repro for "WARNING in __folio_large_mapcount_sanity_checks" (orig title: "-SAME-", reliability: 1), took 7.87 minutes 2026/01/30 06:32:29 "WARNING in __folio_large_mapcount_sanity_checks": saved crash log into 1769754749.crash.log 2026/01/30 06:32:29 "WARNING in __folio_large_mapcount_sanity_checks": saved repro log into 1769754749.repro.log 2026/01/30 06:32:29 start reproducing 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 06:33:26 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:33:32 attempt #1 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 06:33:47 runner 8 connected 2026/01/30 06:33:49 runner 3 connected 2026/01/30 06:34:24 runner 6 connected 2026/01/30 06:34:29 attempt #0 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 06:34:34 runner 7 connected 2026/01/30 06:35:32 attempt #2 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 06:35:32 patched-only: WARNING in __folio_large_mapcount_sanity_checks 2026/01/30 06:35:56 runner 5 connected 2026/01/30 06:36:29 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:36:30 attempt #1 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 06:36:51 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 0, "comps overflows": 2098, "corpus": 5296, "corpus [files]": 591, "corpus [symbols]": 2320, "cover overflows": 94697, "coverage": 47516, "distributor delayed": 6244, "distributor undelayed": 6244, "distributor violated": 295, "exec candidate": 54308, "exec collide": 60933, "exec fuzz": 116220, "exec gen": 6133, "exec hints": 16060, "exec inject": 0, "exec minimize": 40845, "exec retries": 0, "exec seeds": 6338, "exec smash": 52776, "exec total [base]": 283368, "exec total [new]": 523552, "exec triage": 16522, "executor restarts [base]": 69, "executor restarts [new]": 524, "fault jobs": 0, "fuzzer jobs": 7, "fuzzing VMs [base]": 1, "fuzzing VMs [new]": 6, "hints jobs": 1, "max signal": 49565, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 22810, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 4, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 5981, "no exec duration": 2182666000000, "no exec requests": 20112, "pending": 140, "prog exec time": 206, "reproducing": 2, "rpc recv": 12361514556, "rpc sent": 526912888, "signal": 46036, "smash jobs": 4, "triage jobs": 2, "vm output": 11550686, "vm restarts [base]": 12, "vm restarts [new]": 161 } 2026/01/30 06:37:06 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:37:27 runner 0 connected 2026/01/30 06:37:33 attempt #2 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 06:37:34 patched-only: WARNING in __folio_large_mapcount_sanity_checks 2026/01/30 06:37:34 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks (full)' 2026/01/30 06:37:54 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:37:55 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:38:31 runner 1 connected 2026/01/30 06:38:34 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:39:05 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:39:23 reproducing crash 'WARNING in __folio_large_mapcount_sanity_checks': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:39:23 repro finished 'WARNING in __folio_large_mapcount_sanity_checks', repro=true crepro=false desc='WARNING in __folio_large_mapcount_sanity_checks' hub=false from_dashboard=false 2026/01/30 06:39:23 found repro for "WARNING in __folio_large_mapcount_sanity_checks" (orig title: "-SAME-", reliability: 1), took 6.90 minutes 2026/01/30 06:39:23 "WARNING in __folio_large_mapcount_sanity_checks": saved crash log into 1769755163.crash.log 2026/01/30 06:39:23 "WARNING in __folio_large_mapcount_sanity_checks": saved repro log into 1769755163.repro.log 2026/01/30 06:39:23 start reproducing 'WARNING in __folio_large_mapcount_sanity_checks' 2026/01/30 06:39:47 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:40:34 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:40:59 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:41:25 attempt #0 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 06:41:29 patched crashed: BUG: Bad page state in mm_get_huge_zero_folio [need repro = true] 2026/01/30 06:41:29 scheduled a reproduction of 'BUG: Bad page state in mm_get_huge_zero_folio' 2026/01/30 06:41:29 start reproducing 'BUG: Bad page state in mm_get_huge_zero_folio' 2026/01/30 06:41:49 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:41:51 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 0, "comps overflows": 2150, "corpus": 5334, "corpus [files]": 593, "corpus [symbols]": 2342, "cover overflows": 101001, "coverage": 47751, "distributor delayed": 6289, "distributor undelayed": 6289, "distributor violated": 295, "exec candidate": 54308, "exec collide": 67817, "exec fuzz": 129332, "exec gen": 6862, "exec hints": 16231, "exec inject": 0, "exec minimize": 41805, "exec retries": 0, "exec seeds": 6462, "exec smash": 53790, "exec total [base]": 291202, "exec total [new]": 546680, "exec triage": 16658, "executor restarts [base]": 76, "executor restarts [new]": 526, "fault jobs": 0, "fuzzer jobs": 8, "fuzzing VMs [base]": 2, "fuzzing VMs [new]": 4, "hints jobs": 3, "max signal": 49778, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 23410, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 4, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 6038, "no exec duration": 2182666000000, "no exec requests": 20112, "pending": 140, "prog exec time": 188, "reproducing": 3, "rpc recv": 12750184788, "rpc sent": 563320000, "signal": 46241, "smash jobs": 4, "triage jobs": 1, "vm output": 12319516, "vm restarts [base]": 14, "vm restarts [new]": 161 } 2026/01/30 06:42:27 runner 5 connected 2026/01/30 06:42:53 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:43:25 attempt #1 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 06:43:57 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:45:01 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:45:01 repro finished 'WARNING in __folio_large_mapcount_sanity_checks (full)', repro=true crepro=true desc='WARNING in __folio_large_mapcount_sanity_checks' hub=false from_dashboard=false 2026/01/30 06:45:01 start reproducing 'WARNING in __folio_large_mapcount_sanity_checks (full)' 2026/01/30 06:45:01 found repro for "WARNING in __folio_large_mapcount_sanity_checks" (orig title: "-SAME-", reliability: 1), took 15.56 minutes 2026/01/30 06:45:01 "WARNING in __folio_large_mapcount_sanity_checks": saved crash log into 1769755501.crash.log 2026/01/30 06:45:01 "WARNING in __folio_large_mapcount_sanity_checks": saved repro log into 1769755501.repro.log 2026/01/30 06:45:25 attempt #2 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 06:45:25 patched-only: WARNING in __folio_large_mapcount_sanity_checks 2026/01/30 06:45:25 scheduled a reproduction of 'WARNING in __folio_large_mapcount_sanity_checks (full)' 2026/01/30 06:46:16 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:46:22 runner 0 connected 2026/01/30 06:46:51 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 0, "comps overflows": 2182, "corpus": 5363, "corpus [files]": 594, "corpus [symbols]": 2355, "cover overflows": 106652, "coverage": 47819, "distributor delayed": 6320, "distributor undelayed": 6320, "distributor violated": 295, "exec candidate": 54308, "exec collide": 74349, "exec fuzz": 141368, "exec gen": 7542, "exec hints": 16327, "exec inject": 0, "exec minimize": 42514, "exec retries": 0, "exec seeds": 6551, "exec smash": 54578, "exec total [base]": 298113, "exec total [new]": 567712, "exec triage": 16757, "executor restarts [base]": 79, "executor restarts [new]": 529, "fault jobs": 0, "fuzzer jobs": 4, "fuzzing VMs [base]": 2, "fuzzing VMs [new]": 5, "hints jobs": 1, "max signal": 49868, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 23780, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 4, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 6080, "no exec duration": 2182670000000, "no exec requests": 20113, "pending": 140, "prog exec time": 232, "reproducing": 3, "rpc recv": 13085328148, "rpc sent": 596436912, "signal": 46309, "smash jobs": 1, "triage jobs": 2, "vm output": 13238771, "vm restarts [base]": 15, "vm restarts [new]": 162 } 2026/01/30 06:47:02 attempt #0 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 06:49:08 attempt #1 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 06:50:13 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:51:09 attempt #2 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 06:51:09 patched-only: WARNING in __folio_large_mapcount_sanity_checks 2026/01/30 06:51:14 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:51:23 patched crashed: BUG: Bad page state in mm_get_huge_zero_folio [need repro = true] 2026/01/30 06:51:23 scheduled a reproduction of 'BUG: Bad page state in mm_get_huge_zero_folio' 2026/01/30 06:51:32 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:51:51 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 0, "comps overflows": 2221, "corpus": 5386, "corpus [files]": 596, "corpus [symbols]": 2370, "cover overflows": 112233, "coverage": 47899, "distributor delayed": 6357, "distributor undelayed": 6357, "distributor violated": 295, "exec candidate": 54308, "exec collide": 80488, "exec fuzz": 152828, "exec gen": 8163, "exec hints": 16417, "exec inject": 0, "exec minimize": 43074, "exec retries": 0, "exec seeds": 6621, "exec smash": 55156, "exec total [base]": 305851, "exec total [new]": 587320, "exec triage": 16850, "executor restarts [base]": 82, "executor restarts [new]": 531, "fault jobs": 0, "fuzzer jobs": 3, "fuzzing VMs [base]": 2, "fuzzing VMs [new]": 4, "hints jobs": 0, "max signal": 49986, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 24113, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 5, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 6116, "no exec duration": 2182670000000, "no exec requests": 20113, "pending": 141, "prog exec time": 229, "reproducing": 3, "rpc recv": 13390044796, "rpc sent": 629264336, "signal": 46378, "smash jobs": 2, "triage jobs": 1, "vm output": 14175137, "vm restarts [base]": 15, "vm restarts [new]": 162 } 2026/01/30 06:52:06 runner 1 connected 2026/01/30 06:52:20 runner 5 connected 2026/01/30 06:52:26 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:52:43 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:53:36 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:53:55 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:54:46 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:56:03 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:56:51 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 0, "comps overflows": 2281, "corpus": 5409, "corpus [files]": 597, "corpus [symbols]": 2385, "cover overflows": 117699, "coverage": 48212, "distributor delayed": 6394, "distributor undelayed": 6394, "distributor violated": 295, "exec candidate": 54308, "exec collide": 86323, "exec fuzz": 163841, "exec gen": 8729, "exec hints": 16540, "exec inject": 0, "exec minimize": 43713, "exec retries": 0, "exec seeds": 6696, "exec smash": 55749, "exec total [base]": 316902, "exec total [new]": 606260, "exec triage": 16941, "executor restarts [base]": 85, "executor restarts [new]": 536, "fault jobs": 0, "fuzzer jobs": 7, "fuzzing VMs [base]": 3, "fuzzing VMs [new]": 5, "hints jobs": 3, "max signal": 50327, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 24438, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 5, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 6153, "no exec duration": 2182670000000, "no exec requests": 20113, "pending": 141, "prog exec time": 231, "reproducing": 3, "rpc recv": 13898922128, "rpc sent": 664942056, "signal": 46688, "smash jobs": 4, "triage jobs": 0, "vm output": 15070217, "vm restarts [base]": 16, "vm restarts [new]": 163 } 2026/01/30 06:57:01 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:57:33 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:57:33 repro finished 'WARNING in __folio_large_mapcount_sanity_checks (full)', repro=true crepro=true desc='WARNING in __folio_large_mapcount_sanity_checks' hub=false from_dashboard=false 2026/01/30 06:57:33 found repro for "WARNING in __folio_large_mapcount_sanity_checks" (orig title: "-SAME-", reliability: 1), took 12.54 minutes 2026/01/30 06:57:33 start reproducing 'WARNING in __folio_large_mapcount_sanity_checks (full)' 2026/01/30 06:57:33 "WARNING in __folio_large_mapcount_sanity_checks": saved crash log into 1769756253.crash.log 2026/01/30 06:57:33 "WARNING in __folio_large_mapcount_sanity_checks": saved repro log into 1769756253.repro.log 2026/01/30 06:58:51 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 06:59:36 attempt #0 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 07:01:23 patched crashed: kernel BUG in __filemap_add_folio [need repro = true] 2026/01/30 07:01:23 scheduled a reproduction of 'kernel BUG in __filemap_add_folio' 2026/01/30 07:01:23 start reproducing 'kernel BUG in __filemap_add_folio' 2026/01/30 07:01:38 attempt #1 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 07:01:51 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 0, "comps overflows": 2316, "corpus": 5436, "corpus [files]": 600, "corpus [symbols]": 2401, "cover overflows": 122972, "coverage": 48289, "distributor delayed": 6441, "distributor undelayed": 6441, "distributor violated": 295, "exec candidate": 54308, "exec collide": 92263, "exec fuzz": 174806, "exec gen": 9306, "exec hints": 16654, "exec inject": 0, "exec minimize": 44360, "exec retries": 0, "exec seeds": 6772, "exec smash": 56403, "exec total [base]": 325629, "exec total [new]": 625347, "exec triage": 17057, "executor restarts [base]": 85, "executor restarts [new]": 537, "fault jobs": 0, "fuzzer jobs": 11, "fuzzing VMs [base]": 2, "fuzzing VMs [new]": 3, "hints jobs": 3, "max signal": 50511, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 24770, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 6, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 6199, "no exec duration": 2182670000000, "no exec requests": 20113, "pending": 140, "prog exec time": 223, "reproducing": 4, "rpc recv": 14242286696, "rpc sent": 697676720, "signal": 46760, "smash jobs": 4, "triage jobs": 4, "vm output": 16173878, "vm restarts [base]": 16, "vm restarts [new]": 163 } 2026/01/30 07:03:21 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:03:40 attempt #2 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 07:03:40 patched-only: WARNING in __folio_large_mapcount_sanity_checks 2026/01/30 07:04:15 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:04:34 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:04:38 runner 0 connected 2026/01/30 07:04:49 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:05:26 patched crashed: INFO: rcu detected stall in worker_thread [need repro = false] 2026/01/30 07:05:36 repro finished 'WARNING in __folio_large_mapcount_sanity_checks', repro=false crepro=false desc='' hub=false from_dashboard=false 2026/01/30 07:05:36 failed repro for "WARNING in __folio_large_mapcount_sanity_checks", err=%!s() 2026/01/30 07:05:36 "WARNING in __folio_large_mapcount_sanity_checks": saved crash log into 1769756736.crash.log 2026/01/30 07:05:36 "WARNING in __folio_large_mapcount_sanity_checks": saved repro log into 1769756736.repro.log 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:36 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:37 reproduction of "WARNING in __folio_large_mapcount_sanity_checks" aborted: it's no longer needed 2026/01/30 07:05:41 runner 1 connected 2026/01/30 07:05:45 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:06:00 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:06:24 runner 8 connected 2026/01/30 07:06:33 runner 0 connected 2026/01/30 07:06:51 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 0, "comps overflows": 2380, "corpus": 5471, "corpus [files]": 600, "corpus [symbols]": 2417, "cover overflows": 125907, "coverage": 48392, "distributor delayed": 6483, "distributor undelayed": 6483, "distributor violated": 299, "exec candidate": 54308, "exec collide": 94700, "exec fuzz": 179595, "exec gen": 9556, "exec hints": 16847, "exec inject": 0, "exec minimize": 45486, "exec retries": 0, "exec seeds": 6885, "exec smash": 57337, "exec total [base]": 334359, "exec total [new]": 635279, "exec triage": 17141, "executor restarts [base]": 92, "executor restarts [new]": 547, "fault jobs": 0, "fuzzer jobs": 4, "fuzzing VMs [base]": 3, "fuzzing VMs [new]": 5, "hints jobs": 0, "max signal": 50611, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 25234, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 6, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 6238, "no exec duration": 2182670000000, "no exec requests": 20113, "pending": 3, "prog exec time": 218, "reproducing": 3, "rpc recv": 14748795412, "rpc sent": 725234464, "signal": 46858, "smash jobs": 1, "triage jobs": 3, "vm output": 16995792, "vm restarts [base]": 17, "vm restarts [new]": 166 } 2026/01/30 07:06:56 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:07:14 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:08:25 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:08:58 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:09:40 runner 2 connected 2026/01/30 07:09:53 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:09:53 repro finished 'WARNING in __folio_large_mapcount_sanity_checks (full)', repro=true crepro=true desc='WARNING in __folio_large_mapcount_sanity_checks' hub=false from_dashboard=false 2026/01/30 07:09:53 start reproducing 'WARNING in __folio_large_mapcount_sanity_checks (full)' 2026/01/30 07:09:53 found repro for "WARNING in __folio_large_mapcount_sanity_checks" (orig title: "-SAME-", reliability: 1), took 12.32 minutes 2026/01/30 07:09:53 "WARNING in __folio_large_mapcount_sanity_checks": saved crash log into 1769756993.crash.log 2026/01/30 07:09:53 "WARNING in __folio_large_mapcount_sanity_checks": saved repro log into 1769756993.repro.log 2026/01/30 07:10:24 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:11:22 patched crashed: INFO: rcu detected stall in corrupted [need repro = false] 2026/01/30 07:11:51 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 0, "comps overflows": 2434, "corpus": 5498, "corpus [files]": 600, "corpus [symbols]": 2436, "cover overflows": 131246, "coverage": 48473, "distributor delayed": 6510, "distributor undelayed": 6510, "distributor violated": 299, "exec candidate": 54308, "exec collide": 100501, "exec fuzz": 190499, "exec gen": 10164, "exec hints": 16937, "exec inject": 0, "exec minimize": 46040, "exec retries": 0, "exec seeds": 6969, "exec smash": 58033, "exec total [base]": 343124, "exec total [new]": 654092, "exec triage": 17219, "executor restarts [base]": 95, "executor restarts [new]": 547, "fault jobs": 0, "fuzzer jobs": 5, "fuzzing VMs [base]": 2, "fuzzing VMs [new]": 4, "hints jobs": 1, "max signal": 50713, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 25526, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 6, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 6272, "no exec duration": 2182706000000, "no exec requests": 20115, "pending": 2, "prog exec time": 216, "reproducing": 3, "rpc recv": 15135637700, "rpc sent": 759985744, "signal": 46937, "smash jobs": 2, "triage jobs": 2, "vm output": 18191855, "vm restarts [base]": 18, "vm restarts [new]": 166 } 2026/01/30 07:11:55 attempt #0 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 07:12:19 runner 7 connected 2026/01/30 07:13:21 repro finished 'kernel BUG in __filemap_add_folio', repro=false crepro=false desc='' hub=false from_dashboard=false 2026/01/30 07:13:21 failed repro for "kernel BUG in __filemap_add_folio", err=%!s() 2026/01/30 07:13:21 "kernel BUG in __filemap_add_folio": saved crash log into 1769757201.crash.log 2026/01/30 07:13:21 "kernel BUG in __filemap_add_folio": saved repro log into 1769757201.repro.log 2026/01/30 07:13:31 runner 2 connected 2026/01/30 07:13:55 attempt #1 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 07:15:40 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:15:58 attempt #2 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 07:15:58 patched-only: WARNING in __folio_large_mapcount_sanity_checks 2026/01/30 07:16:23 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:16:51 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 0, "comps overflows": 2467, "corpus": 5523, "corpus [files]": 602, "corpus [symbols]": 2448, "cover overflows": 137720, "coverage": 48546, "distributor delayed": 6552, "distributor undelayed": 6552, "distributor violated": 299, "exec candidate": 54308, "exec collide": 107336, "exec fuzz": 203420, "exec gen": 10821, "exec hints": 16972, "exec inject": 0, "exec minimize": 46742, "exec retries": 0, "exec seeds": 7044, "exec smash": 58652, "exec total [base]": 350175, "exec total [new]": 676055, "exec triage": 17334, "executor restarts [base]": 97, "executor restarts [new]": 553, "fault jobs": 0, "fuzzer jobs": 5, "fuzzing VMs [base]": 2, "fuzzing VMs [new]": 6, "hints jobs": 0, "max signal": 50788, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 25890, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 7, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 6317, "no exec duration": 2182706000000, "no exec requests": 20115, "pending": 2, "prog exec time": 228, "reproducing": 2, "rpc recv": 15488793488, "rpc sent": 796688920, "signal": 46993, "smash jobs": 1, "triage jobs": 4, "vm output": 19249437, "vm restarts [base]": 18, "vm restarts [new]": 168 } 2026/01/30 07:16:55 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:16:56 runner 0 connected 2026/01/30 07:17:36 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:18:08 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:18:33 base crash: kernel BUG in mpage_readahead 2026/01/30 07:19:15 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:19:31 runner 0 connected 2026/01/30 07:19:32 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:20:26 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:21:02 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:21:51 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 0, "comps overflows": 2533, "corpus": 5546, "corpus [files]": 607, "corpus [symbols]": 2464, "cover overflows": 144211, "coverage": 48683, "distributor delayed": 6582, "distributor undelayed": 6582, "distributor violated": 299, "exec candidate": 54308, "exec collide": 114335, "exec fuzz": 216913, "exec gen": 11533, "exec hints": 17199, "exec inject": 0, "exec minimize": 47564, "exec retries": 0, "exec seeds": 7116, "exec smash": 59258, "exec total [base]": 359216, "exec total [new]": 699092, "exec triage": 17438, "executor restarts [base]": 104, "executor restarts [new]": 554, "fault jobs": 0, "fuzzer jobs": 5, "fuzzing VMs [base]": 3, "fuzzing VMs [new]": 6, "hints jobs": 1, "max signal": 50953, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 26272, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 7, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 6355, "no exec duration": 2182706000000, "no exec requests": 20115, "pending": 2, "prog exec time": 250, "reproducing": 2, "rpc recv": 15936076900, "rpc sent": 838010456, "signal": 47170, "smash jobs": 2, "triage jobs": 2, "vm output": 20517155, "vm restarts [base]": 20, "vm restarts [new]": 168 } 2026/01/30 07:22:01 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:22:35 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:22:35 repro finished 'WARNING in __folio_large_mapcount_sanity_checks (full)', repro=true crepro=true desc='WARNING in __folio_large_mapcount_sanity_checks' hub=false from_dashboard=false 2026/01/30 07:22:35 start reproducing 'WARNING in __folio_large_mapcount_sanity_checks (full)' 2026/01/30 07:22:35 found repro for "WARNING in __folio_large_mapcount_sanity_checks" (orig title: "-SAME-", reliability: 1), took 12.71 minutes 2026/01/30 07:22:35 "WARNING in __folio_large_mapcount_sanity_checks": saved crash log into 1769757755.crash.log 2026/01/30 07:22:35 "WARNING in __folio_large_mapcount_sanity_checks": saved repro log into 1769757755.repro.log 2026/01/30 07:23:29 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:24:41 attempt #0 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 07:26:44 attempt #1 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 07:26:51 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 0, "comps overflows": 2584, "corpus": 5566, "corpus [files]": 609, "corpus [symbols]": 2476, "cover overflows": 151010, "coverage": 48722, "distributor delayed": 6601, "distributor undelayed": 6601, "distributor violated": 299, "exec candidate": 54308, "exec collide": 121658, "exec fuzz": 230707, "exec gen": 12246, "exec hints": 17345, "exec inject": 0, "exec minimize": 48076, "exec retries": 0, "exec seeds": 7176, "exec smash": 59774, "exec total [base]": 366525, "exec total [new]": 722231, "exec triage": 17514, "executor restarts [base]": 104, "executor restarts [new]": 556, "fault jobs": 0, "fuzzer jobs": 2, "fuzzing VMs [base]": 2, "fuzzing VMs [new]": 6, "hints jobs": 0, "max signal": 51052, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 26559, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 8, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 6386, "no exec duration": 2182706000000, "no exec requests": 20115, "pending": 1, "prog exec time": 232, "reproducing": 2, "rpc recv": 16233237412, "rpc sent": 876936136, "signal": 47204, "smash jobs": 0, "triage jobs": 2, "vm output": 21760913, "vm restarts [base]": 20, "vm restarts [new]": 168 } 2026/01/30 07:27:25 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:28:17 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:28:41 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:28:46 attempt #2 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 07:28:46 patched-only: WARNING in __folio_large_mapcount_sanity_checks 2026/01/30 07:29:29 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:29:52 runner 0 connected 2026/01/30 07:30:41 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:30:57 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:31:51 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 0, "comps overflows": 2657, "corpus": 5586, "corpus [files]": 611, "corpus [symbols]": 2489, "cover overflows": 158028, "coverage": 48782, "distributor delayed": 6624, "distributor undelayed": 6624, "distributor violated": 299, "exec candidate": 54308, "exec collide": 128687, "exec fuzz": 244306, "exec gen": 12952, "exec hints": 17379, "exec inject": 0, "exec minimize": 48521, "exec retries": 0, "exec seeds": 7236, "exec smash": 60269, "exec total [base]": 374370, "exec total [new]": 744682, "exec triage": 17595, "executor restarts [base]": 107, "executor restarts [new]": 556, "fault jobs": 0, "fuzzer jobs": 2, "fuzzing VMs [base]": 3, "fuzzing VMs [new]": 6, "hints jobs": 1, "max signal": 51194, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 26785, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 8, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 6417, "no exec duration": 2182706000000, "no exec requests": 20115, "pending": 1, "prog exec time": 240, "reproducing": 2, "rpc recv": 16585127480, "rpc sent": 917373112, "signal": 47241, "smash jobs": 1, "triage jobs": 0, "vm output": 22786219, "vm restarts [base]": 21, "vm restarts [new]": 168 } 2026/01/30 07:31:52 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:32:09 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:33:20 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:33:52 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:34:49 reproducing crash 'no output/lost connection': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/rmap.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2026/01/30 07:34:49 repro finished 'WARNING in __folio_large_mapcount_sanity_checks (full)', repro=true crepro=true desc='WARNING in __folio_large_mapcount_sanity_checks' hub=false from_dashboard=false 2026/01/30 07:34:49 found repro for "WARNING in __folio_large_mapcount_sanity_checks" (orig title: "-SAME-", reliability: 1), took 12.23 minutes 2026/01/30 07:34:49 "WARNING in __folio_large_mapcount_sanity_checks": saved crash log into 1769758489.crash.log 2026/01/30 07:34:49 "WARNING in __folio_large_mapcount_sanity_checks": saved repro log into 1769758489.repro.log 2026/01/30 07:34:57 runner 3 connected 2026/01/30 07:35:23 runner 6 connected 2026/01/30 07:36:51 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 0, "comps overflows": 2681, "corpus": 5600, "corpus [files]": 611, "corpus [symbols]": 2501, "cover overflows": 164396, "coverage": 48939, "distributor delayed": 6640, "distributor undelayed": 6640, "distributor violated": 299, "exec candidate": 54308, "exec collide": 135994, "exec fuzz": 257955, "exec gen": 13645, "exec hints": 17531, "exec inject": 0, "exec minimize": 48892, "exec retries": 0, "exec seeds": 7281, "exec smash": 60649, "exec total [base]": 382908, "exec total [new]": 767345, "exec triage": 17658, "executor restarts [base]": 108, "executor restarts [new]": 565, "fault jobs": 0, "fuzzer jobs": 1, "fuzzing VMs [base]": 2, "fuzzing VMs [new]": 7, "hints jobs": 0, "max signal": 51367, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 26985, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 8, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 6443, "no exec duration": 2182706000000, "no exec requests": 20115, "pending": 1, "prog exec time": 229, "reproducing": 1, "rpc recv": 16978943088, "rpc sent": 958462048, "signal": 47386, "smash jobs": 0, "triage jobs": 1, "vm output": 23892901, "vm restarts [base]": 21, "vm restarts [new]": 170 } 2026/01/30 07:36:51 attempt #0 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 07:38:52 attempt #1 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 07:40:53 attempt #2 to run "WARNING in __folio_large_mapcount_sanity_checks" on base: did not crash 2026/01/30 07:40:53 patched-only: WARNING in __folio_large_mapcount_sanity_checks 2026/01/30 07:41:51 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 0, "comps overflows": 2716, "corpus": 5657, "corpus [files]": 614, "corpus [symbols]": 2519, "cover overflows": 172019, "coverage": 49080, "distributor delayed": 6693, "distributor undelayed": 6693, "distributor violated": 299, "exec candidate": 54308, "exec collide": 144208, "exec fuzz": 273573, "exec gen": 14405, "exec hints": 17610, "exec inject": 0, "exec minimize": 50210, "exec retries": 0, "exec seeds": 7455, "exec smash": 62099, "exec total [base]": 389497, "exec total [new]": 795139, "exec triage": 17839, "executor restarts [base]": 108, "executor restarts [new]": 565, "fault jobs": 0, "fuzzer jobs": 0, "fuzzing VMs [base]": 2, "fuzzing VMs [new]": 7, "hints jobs": 0, "max signal": 51599, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 27600, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 8, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 6519, "no exec duration": 2182706000000, "no exec requests": 20115, "pending": 1, "prog exec time": 226, "reproducing": 1, "rpc recv": 17289883116, "rpc sent": 1002573848, "signal": 47516, "smash jobs": 0, "triage jobs": 0, "vm output": 25102953, "vm restarts [base]": 21, "vm restarts [new]": 170 } 2026/01/30 07:41:51 runner 0 connected 2026/01/30 07:46:51 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 0, "comps overflows": 2757, "corpus": 5677, "corpus [files]": 614, "corpus [symbols]": 2530, "cover overflows": 179562, "coverage": 49112, "distributor delayed": 6708, "distributor undelayed": 6708, "distributor violated": 299, "exec candidate": 54308, "exec collide": 152366, "exec fuzz": 289057, "exec gen": 15219, "exec hints": 17650, "exec inject": 0, "exec minimize": 50716, "exec retries": 0, "exec seeds": 7515, "exec smash": 62591, "exec total [base]": 399193, "exec total [new]": 820772, "exec triage": 17919, "executor restarts [base]": 112, "executor restarts [new]": 567, "fault jobs": 0, "fuzzer jobs": 3, "fuzzing VMs [base]": 3, "fuzzing VMs [new]": 7, "hints jobs": 0, "max signal": 51638, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 27875, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 8, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 6551, "no exec duration": 2182767000000, "no exec requests": 20118, "pending": 1, "prog exec time": 239, "reproducing": 1, "rpc recv": 17698692904, "rpc sent": 1049001880, "signal": 47714, "smash jobs": 1, "triage jobs": 2, "vm output": 26338919, "vm restarts [base]": 22, "vm restarts [new]": 170 } 2026/01/30 07:47:29 repro finished 'BUG: Bad page state in mm_get_huge_zero_folio', repro=true crepro=false desc='SYZFAIL: failed to recv rpc' hub=false from_dashboard=false 2026/01/30 07:47:29 found repro for "SYZFAIL: failed to recv rpc" (orig title: "BUG: Bad page state in mm_get_huge_zero_folio", reliability: 0), took 66.00 minutes 2026/01/30 07:47:29 "SYZFAIL: failed to recv rpc": saved crash log into 1769759249.crash.log 2026/01/30 07:47:29 "SYZFAIL: failed to recv rpc": saved repro log into 1769759249.repro.log 2026/01/30 07:47:29 start reproducing 'BUG: Bad page state in mm_get_huge_zero_folio' 2026/01/30 07:50:10 attempt #0 to run "SYZFAIL: failed to recv rpc" on base: crashed with SYZFAIL: failed to recv rpc 2026/01/30 07:50:10 crashes both: SYZFAIL: failed to recv rpc / SYZFAIL: failed to recv rpc 2026/01/30 07:51:06 runner 0 connected 2026/01/30 07:51:46 patched crashed: INFO: rcu detected stall in corrupted [need repro = false] 2026/01/30 07:51:51 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 0, "comps overflows": 2796, "corpus": 5692, "corpus [files]": 614, "corpus [symbols]": 2534, "cover overflows": 186721, "coverage": 49172, "distributor delayed": 6721, "distributor undelayed": 6721, "distributor violated": 299, "exec candidate": 54308, "exec collide": 160529, "exec fuzz": 304323, "exec gen": 16040, "exec hints": 17690, "exec inject": 0, "exec minimize": 51080, "exec retries": 0, "exec seeds": 7560, "exec smash": 62960, "exec total [base]": 406730, "exec total [new]": 845895, "exec triage": 17975, "executor restarts [base]": 116, "executor restarts [new]": 568, "fault jobs": 0, "fuzzer jobs": 3, "fuzzing VMs [base]": 3, "fuzzing VMs [new]": 6, "hints jobs": 0, "max signal": 51715, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 28067, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 8, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 6574, "no exec duration": 2182767000000, "no exec requests": 20118, "pending": 0, "prog exec time": 237, "reproducing": 1, "rpc recv": 18020536200, "rpc sent": 1091976400, "signal": 47767, "smash jobs": 1, "triage jobs": 2, "vm output": 27234405, "vm restarts [base]": 23, "vm restarts [new]": 170 } 2026/01/30 07:52:44 runner 3 connected 2026/01/30 07:53:04 runner 1 connected 2026/01/30 07:56:51 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 0, "comps overflows": 2823, "corpus": 5707, "corpus [files]": 616, "corpus [symbols]": 2543, "cover overflows": 193865, "coverage": 49225, "distributor delayed": 6733, "distributor undelayed": 6733, "distributor violated": 299, "exec candidate": 54308, "exec collide": 168607, "exec fuzz": 319536, "exec gen": 16812, "exec hints": 17722, "exec inject": 0, "exec minimize": 51474, "exec retries": 0, "exec seeds": 7605, "exec smash": 63349, "exec total [base]": 416777, "exec total [new]": 870879, "exec triage": 18034, "executor restarts [base]": 119, "executor restarts [new]": 571, "fault jobs": 0, "fuzzer jobs": 0, "fuzzing VMs [base]": 3, "fuzzing VMs [new]": 7, "hints jobs": 0, "max signal": 51826, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 28298, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 8, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 6597, "no exec duration": 2182767000000, "no exec requests": 20118, "pending": 0, "prog exec time": 238, "reproducing": 1, "rpc recv": 18461144208, "rpc sent": 1138236000, "signal": 47818, "smash jobs": 0, "triage jobs": 0, "vm output": 28104830, "vm restarts [base]": 24, "vm restarts [new]": 171 } 2026/01/30 08:01:51 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 0, "comps overflows": 2841, "corpus": 5734, "corpus [files]": 618, "corpus [symbols]": 2555, "cover overflows": 200920, "coverage": 49292, "distributor delayed": 6754, "distributor undelayed": 6754, "distributor violated": 299, "exec candidate": 54308, "exec collide": 176205, "exec fuzz": 334021, "exec gen": 17577, "exec hints": 17801, "exec inject": 0, "exec minimize": 52056, "exec retries": 0, "exec seeds": 7684, "exec smash": 64024, "exec total [base]": 427380, "exec total [new]": 895240, "exec triage": 18131, "executor restarts [base]": 120, "executor restarts [new]": 573, "fault jobs": 0, "fuzzer jobs": 2, "fuzzing VMs [base]": 3, "fuzzing VMs [new]": 7, "hints jobs": 0, "max signal": 51941, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 28592, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 8, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 6637, "no exec duration": 2182767000000, "no exec requests": 20118, "pending": 0, "prog exec time": 273, "reproducing": 1, "rpc recv": 18860634248, "rpc sent": 1185760336, "signal": 47885, "smash jobs": 0, "triage jobs": 2, "vm output": 29045152, "vm restarts [base]": 24, "vm restarts [new]": 171 } 2026/01/30 08:06:45 bug reporting terminated 2026/01/30 08:06:45 status reporting terminated 2026/01/30 08:06:45 base: rpc server terminaled 2026/01/30 08:06:45 new: rpc server terminaled 2026/01/30 08:06:45 base: pool terminated 2026/01/30 08:06:45 base: kernel context loop terminated 2026/01/30 08:08:20 repro finished 'BUG: Bad page state in mm_get_huge_zero_folio', repro=false crepro=false desc='' hub=false from_dashboard=false 2026/01/30 08:08:20 repro loop terminated 2026/01/30 08:08:20 new: pool terminated 2026/01/30 08:08:20 new: kernel context loop terminated 2026/01/30 08:08:20 diff fuzzing terminated 2026/01/30 08:08:20 fuzzing is finished 2026/01/30 08:08:20 status at the end: Title On-Base On-Patched WARNING in __folio_large_mapcount_sanity_checks 146 crashes[reproduced] BUG: Bad page state in mm_get_huge_zero_folio 2 crashes INFO: rcu detected stall in corrupted 2 crashes INFO: rcu detected stall in worker_thread 1 crashes SYZFAIL: failed to recv rpc 1 crashes [reproduced] kernel BUG in __filemap_add_folio 1 crashes kernel BUG in mpage_readahead 1 crashes no output from test machine 1 crashes 2026/01/30 08:08:20 possibly patched-only: WARNING in __folio_large_mapcount_sanity_checks