2025/08/19 02:29:32 extracted 303749 symbol hashes for base and 303751 for patched 2025/08/19 02:29:32 adding modified_functions to focus areas: ["__UNIQUE_ID_nop1452" "__UNIQUE_ID_nop1453" "__bio_release_pages" "__dec_node_state" "__f2fs_commit_super" "__filemap_fdatawait_range" "__filemap_get_folio" "__folio_put" "__get_meta_folio" "__get_metapage" "__get_node_folio" "__inc_node_state" "__mem_cgroup_try_charge_swap" "__mem_cgroup_uncharge_swap" "__memcg_kmem_charge_page" "__memcg_kmem_uncharge_page" "__mod_node_page_state" "__node_get_memory_tier" "__oom_kill_process" "__pfx_filemap_mod_uncharged_vmstat" "__se_sys_cachestat" "attr_wof_frame_info" "bch2_mark_pagecache_reserved" "bch2_mark_pagecache_unallocated" "bch2_page_mkwrite" "bch2_seek_pagecache_data" "bio_set_pages_dirty" "block_page_mkwrite" "btrfs_cleanup_bg_io" "btrfs_cleanup_one_transaction" "btrfs_cleanup_transaction" "btrfs_clear_buffer_dirty" "btrfs_defrag_file" "btrfs_drop_and_free_fs_root" "btrfs_mark_buffer_dirty" "btrfs_page_mkwrite" "btrfs_read_merkle_tree_page" "btrfs_replay_log" "btrfs_truncate_block" "btrfs_writepage_fixup_worker" "btrfs_writepages" "ceph_page_mkwrite" "ceph_uninline_data" "ceph_writepages_start" "cgroup_id_from_mm" "change_prot_numa" "clean_bdev_aliases" "compact_zone" "compaction_alloc" "compaction_proactiveness_sysctl_handler" "count_memcg_event_mm" "cpu_vm_stats_fold" "dax_iomap_fault" "delete_from_page_cache_batch" "do_convert_inline_dir" "do_huge_pmd_wp_page" "do_read_cache_folio" "do_swap_page" "do_wp_page" "drain_obj_stock" "ecryptfs_write" "ext4_page_mkwrite" "ext4_write_begin" "f2fs_delete_entry" "f2fs_delete_inline_entry" "f2fs_flush_inline_data" "f2fs_fsync_node_pages" "f2fs_get_lock_data_folio" "f2fs_init_inode_metadata" "f2fs_invalidate_compress_pages" "f2fs_quota_read" "f2fs_recover_fsync_data" "f2fs_set_link" "f2fs_sync_meta_pages" "f2fs_sync_node_pages" "f2fs_truncate_inode_blocks" "f2fs_vm_page_mkwrite" "f2fs_write_begin" "f2fs_write_data_pages" "f2fs_write_multi_pages" "fb_deferred_io_mkwrite" "filemap_add_folio" "filemap_fault" "filemap_get_entry" "filemap_get_folios_contig" "filemap_get_folios_tag" "filemap_get_read_batch" "filemap_mod_uncharged_vmstat" "filemap_page_mkwrite" "filemap_read" "filemap_splice_read" "filemap_unaccount_folio" "find_get_block_common" "find_get_entries" "find_lock_entries" "flush_memcg_stats_dwork" "folio_batch_move_lru" "folio_deactivate" "folio_lock" "folio_lruvec" "folio_lruvec_lock" "folio_lruvec_lock_irq" "folio_lruvec_lock_irqsave" "folio_mark_accessed" "folio_mark_dirty_lock" "folio_mark_lazyfree" "folios_put_refs" "force_metapage" "free_percpu" "fuse_page_mkwrite" "generic_perform_write" "generic_pipe_buf_try_steal" "get_mem_cgroup_from_current" "get_mem_cgroup_from_folio" "get_mem_cgroup_from_mm" "gfs2_page_mkwrite" "gfs2_trans_add_meta" "gfs2_write_cache_jdata" "grab_metapage" "hold_metapage" "invalid_folio_referenced_vma" "invalidate_inode_pages2_range" "io_ctl_prepare_pages" "iomap_page_mkwrite" "isolate_migratepages_block" "kcompactd" "ksm_do_scan" "ksm_get_folio" "lock_delalloc_folios" "lock_metapage" "lock_page" "lru_activate" "lru_add" "lru_deactivate" "lru_deactivate_file" "lru_gen_clear_refs" "lru_lazyfree" "lru_move_tail" "lruvec_page_state" "lruvec_page_state_local" "make_device_exclusive" "map_anon_folio_pmd" "mapping_seek_hole_data" "mem_cgroup_charge_skmem" "mem_cgroup_iter" "mem_cgroup_iter_break" "mem_cgroup_migrate" "mem_cgroup_replace_folio" "mem_cgroup_scan_tasks" "mem_cgroup_swap_full" "mem_cgroup_track_foreign_dirty_slowpath" "mem_cgroup_uncharge_skmem" "mem_cgroup_wb_stats" "memcg1_stat_format" "memcg1_swapout" "memcg_page_state" "memcg_page_state_local" "memcg_page_state_local_output" "memcg_page_state_output" "memory_numa_stat_show" "memory_stat_format" "memtier_hotplug_callback" "migrate_device_coherent_folio" "migrate_pages" "migrate_pages_batch" "minix_add_link" "minix_delete_entry" "minix_set_link" "mod_memcg_lruvec_state" "mod_memcg_page_state" "mod_memcg_state" "mod_node_state" "mod_objcg_mlstate" "move_pages_huge_pmd" "move_pages_pte" "mpage_prepare_extent_to_map" "need_update" "netfs_page_mkwrite" "next_uptodate_folio" "nfs_vm_page_mkwrite" "nilfs_add_link" "nilfs_begin_folio_io" "nilfs_btnode_delete" "nilfs_btnode_prepare_change_key" "nilfs_clear_dirty_pages" "nilfs_copy_back_pages" "nilfs_copy_dirty_pages" "nilfs_delete_entry" "nilfs_end_folio_io" "nilfs_find_uncommitted_extent" "nilfs_lookup_dirty_data_buffers" "nilfs_page_mkwrite" "nilfs_segctor_do_construct" "nilfs_set_link" "node_get_allowed_targets" "node_is_toptier" "node_page_state" "node_page_state_add" "node_page_state_pages" "node_read_vmstat" "ntfs_bio_fill_1" "obj_cgroup_charge_pages" "obj_cgroup_charge_zswap" "obj_cgroup_may_zswap" "obj_cgroup_release" "obj_cgroup_uncharge_zswap" "ocfs2_duplicate_clusters_by_page" "ocfs2_grab_folios_for_write" "orangefs_page_mkwrite" "page_cache_pipe_buf_confirm" "page_cache_pipe_buf_try_steal" "pagetypeinfo_showfree_print" "pcpu_memcg_post_alloc_hook" "perf_trace_mm_lru_insertion" "prepare_one_folio" "putback_movable_ops_page" "refill_obj_stock" "refresh_cpu_vm_stats" "release_metapage" "relocate_file_extent_cluster" "remove_device_exclusive_entry" "remove_inode_hugepages" "send_extent_data" "set_extent_buffer_dirty" "set_node_memory_tier" "shmem_get_folio_gfp" "shmem_get_partial_folio" "shmem_swapin_folio" "shmem_undo_range" "shrink_folio_list" "split_huge_pages_all" "trace_event_raw_event_mm_lru_insertion" "truncate_inode_pages_range" "try_split_folio" "try_to_compact_pages" "try_to_unuse" "ttm_backup_backup_page" "txLock" "ubifs_vm_page_mkwrite" "udf_expand_file_adinicb" "udf_page_mkwrite" "ufs_add_link" "ufs_delete_entry" "ufs_get_locked_folio" "ufs_set_link" "uncharge_batch" "unpin_user_page_range_dirty_lock" "unpin_user_pages_dirty_lock" "vfs_dedupe_file_range_compare" "vmstat_next" "vmstat_refresh" "vmstat_shepherd" "vmstat_show" "vmstat_start" "workingset_activation" "workingset_eviction" "workingset_refault" "write_all_supers" "write_one_eb" "writeback_iter" "z_erofs_runqueue" "zone_device_page_init" "zoneinfo_show_print" "zswap_current_read" "zswap_folio_swapin" "zswap_shrinker_count"] 2025/08/19 02:29:32 adding directly modified files to focus areas: ["fs/btrfs/disk-io.c" "include/linux/memcontrol.h" "include/linux/mmzone.h" "include/linux/pagemap.h" "mm/filemap.c" "mm/vmstat.c"] 2025/08/19 02:29:33 downloaded the corpus from https://storage.googleapis.com/syzkaller/corpus/ci-upstream-kasan-gce-root-corpus.db 2025/08/19 02:30:30 runner 6 connected 2025/08/19 02:30:31 runner 2 connected 2025/08/19 02:30:31 runner 5 connected 2025/08/19 02:30:31 runner 3 connected 2025/08/19 02:30:31 runner 0 connected 2025/08/19 02:30:31 runner 3 connected 2025/08/19 02:30:31 runner 1 connected 2025/08/19 02:30:31 runner 0 connected 2025/08/19 02:30:31 runner 2 connected 2025/08/19 02:30:31 runner 7 connected 2025/08/19 02:30:31 runner 9 connected 2025/08/19 02:30:31 runner 8 connected 2025/08/19 02:30:37 initializing coverage information... 2025/08/19 02:30:37 executor cover filter: 0 PCs 2025/08/19 02:30:41 discovered 7699 source files, 338620 symbols 2025/08/19 02:30:41 machine check: disabled the following syscalls: fsetxattr$security_selinux : selinux is not enabled fsetxattr$security_smack_transmute : smack is not enabled fsetxattr$smack_xattr_label : smack is not enabled get_thread_area : syscall get_thread_area is not present lookup_dcookie : syscall lookup_dcookie is not present lsetxattr$security_selinux : selinux is not enabled lsetxattr$security_smack_transmute : smack is not enabled lsetxattr$smack_xattr_label : smack is not enabled mount$esdfs : /proc/filesystems does not contain esdfs mount$incfs : /proc/filesystems does not contain incremental-fs openat$acpi_thermal_rel : failed to open /dev/acpi_thermal_rel: no such file or directory openat$ashmem : failed to open /dev/ashmem: no such file or directory openat$bifrost : failed to open /dev/bifrost: no such file or directory openat$binder : failed to open /dev/binder: no such file or directory openat$camx : failed to open /dev/v4l/by-path/platform-soc@0:qcom_cam-req-mgr-video-index0: no such file or directory openat$capi20 : failed to open /dev/capi20: no such file or directory openat$cdrom1 : failed to open /dev/cdrom1: no such file or directory openat$damon_attrs : failed to open /sys/kernel/debug/damon/attrs: no such file or directory openat$damon_init_regions : failed to open /sys/kernel/debug/damon/init_regions: no such file or directory openat$damon_kdamond_pid : failed to open /sys/kernel/debug/damon/kdamond_pid: no such file or directory openat$damon_mk_contexts : failed to open /sys/kernel/debug/damon/mk_contexts: no such file or directory openat$damon_monitor_on : failed to open /sys/kernel/debug/damon/monitor_on: no such file or directory openat$damon_rm_contexts : failed to open /sys/kernel/debug/damon/rm_contexts: no such file or directory openat$damon_schemes : failed to open /sys/kernel/debug/damon/schemes: no such file or directory openat$damon_target_ids : failed to open /sys/kernel/debug/damon/target_ids: no such file or directory openat$hwbinder : failed to open /dev/hwbinder: no such file or directory openat$i915 : failed to open /dev/i915: no such file or directory openat$img_rogue : failed to open /dev/img-rogue: no such file or directory openat$irnet : failed to open /dev/irnet: no such file or directory openat$keychord : failed to open /dev/keychord: no such file or directory openat$kvm : failed to open /dev/kvm: no such file or directory openat$lightnvm : failed to open /dev/lightnvm/control: no such file or directory openat$mali : failed to open /dev/mali0: no such file or directory openat$md : failed to open /dev/md0: no such file or directory openat$msm : failed to open /dev/msm: no such file or directory openat$ndctl0 : failed to open /dev/ndctl0: no such file or directory openat$nmem0 : failed to open /dev/nmem0: no such file or directory openat$pktcdvd : failed to open /dev/pktcdvd/control: no such file or directory openat$pmem0 : failed to open /dev/pmem0: no such file or directory openat$proc_capi20 : failed to open /proc/capi/capi20: no such file or directory openat$proc_capi20ncci : failed to open /proc/capi/capi20ncci: no such file or directory openat$proc_reclaim : failed to open /proc/self/reclaim: no such file or directory openat$ptp1 : failed to open /dev/ptp1: no such file or directory openat$rnullb : failed to open /dev/rnullb0: no such file or directory openat$selinux_access : failed to open /selinux/access: no such file or directory openat$selinux_attr : selinux is not enabled openat$selinux_avc_cache_stats : failed to open /selinux/avc/cache_stats: no such file or directory openat$selinux_avc_cache_threshold : failed to open /selinux/avc/cache_threshold: no such file or directory openat$selinux_avc_hash_stats : failed to open /selinux/avc/hash_stats: no such file or directory openat$selinux_checkreqprot : failed to open /selinux/checkreqprot: no such file or directory openat$selinux_commit_pending_bools : failed to open /selinux/commit_pending_bools: no such file or directory openat$selinux_context : failed to open /selinux/context: no such file or directory openat$selinux_create : failed to open /selinux/create: no such file or directory openat$selinux_enforce : failed to open /selinux/enforce: no such file or directory openat$selinux_load : failed to open /selinux/load: no such file or directory openat$selinux_member : failed to open /selinux/member: no such file or directory openat$selinux_mls : failed to open /selinux/mls: no such file or directory openat$selinux_policy : failed to open /selinux/policy: no such file or directory openat$selinux_relabel : failed to open /selinux/relabel: no such file or directory openat$selinux_status : failed to open /selinux/status: no such file or directory openat$selinux_user : failed to open /selinux/user: no such file or directory openat$selinux_validatetrans : failed to open /selinux/validatetrans: no such file or directory openat$sev : failed to open /dev/sev: no such file or directory openat$sgx_provision : failed to open /dev/sgx_provision: no such file or directory openat$smack_task_current : smack is not enabled openat$smack_thread_current : smack is not enabled openat$smackfs_access : failed to open /sys/fs/smackfs/access: no such file or directory openat$smackfs_ambient : failed to open /sys/fs/smackfs/ambient: no such file or directory openat$smackfs_change_rule : failed to open /sys/fs/smackfs/change-rule: no such file or directory openat$smackfs_cipso : failed to open /sys/fs/smackfs/cipso: no such file or directory openat$smackfs_cipsonum : failed to open /sys/fs/smackfs/direct: no such file or directory openat$smackfs_ipv6host : failed to open /sys/fs/smackfs/ipv6host: no such file or directory openat$smackfs_load : failed to open /sys/fs/smackfs/load: no such file or directory openat$smackfs_logging : failed to open /sys/fs/smackfs/logging: no such file or directory openat$smackfs_netlabel : failed to open /sys/fs/smackfs/netlabel: no such file or directory openat$smackfs_onlycap : failed to open /sys/fs/smackfs/onlycap: no such file or directory openat$smackfs_ptrace : failed to open /sys/fs/smackfs/ptrace: no such file or directory openat$smackfs_relabel_self : failed to open /sys/fs/smackfs/relabel-self: no such file or directory openat$smackfs_revoke_subject : failed to open /sys/fs/smackfs/revoke-subject: no such file or directory openat$smackfs_syslog : failed to open /sys/fs/smackfs/syslog: no such file or directory openat$smackfs_unconfined : failed to open /sys/fs/smackfs/unconfined: no such file or directory openat$tlk_device : failed to open /dev/tlk_device: no such file or directory openat$trusty : failed to open /dev/trusty-ipc-dev0: no such file or directory openat$trusty_avb : failed to open /dev/trusty-ipc-dev0: no such file or directory openat$trusty_gatekeeper : failed to open /dev/trusty-ipc-dev0: no such file or directory openat$trusty_hwkey : failed to open /dev/trusty-ipc-dev0: no such file or directory openat$trusty_hwrng : failed to open /dev/trusty-ipc-dev0: no such file or directory openat$trusty_km : failed to open /dev/trusty-ipc-dev0: no such file or directory openat$trusty_km_secure : failed to open /dev/trusty-ipc-dev0: no such file or directory openat$trusty_storage : failed to open /dev/trusty-ipc-dev0: no such file or directory openat$tty : failed to open /dev/tty: no such device or address openat$uverbs0 : failed to open /dev/infiniband/uverbs0: no such file or directory openat$vfio : failed to open /dev/vfio/vfio: no such file or directory openat$vndbinder : failed to open /dev/vndbinder: no such file or directory openat$vtpm : failed to open /dev/vtpmx: no such file or directory openat$xenevtchn : failed to open /dev/xen/evtchn: no such file or directory openat$zygote : failed to open /dev/socket/zygote: no such file or directory pkey_alloc : pkey_alloc(0x0, 0x0) failed: no space left on device read$smackfs_access : smack is not enabled read$smackfs_cipsonum : smack is not enabled read$smackfs_logging : smack is not enabled read$smackfs_ptrace : smack is not enabled set_thread_area : syscall set_thread_area is not present setxattr$security_selinux : selinux is not enabled setxattr$security_smack_transmute : smack is not enabled setxattr$smack_xattr_label : smack is not enabled socket$hf : socket$hf(0x13, 0x2, 0x0) failed: address family not supported by protocol socket$inet6_dccp : socket$inet6_dccp(0xa, 0x6, 0x0) failed: socket type not supported socket$inet_dccp : socket$inet_dccp(0x2, 0x6, 0x0) failed: socket type not supported socket$vsock_dgram : socket$vsock_dgram(0x28, 0x2, 0x0) failed: no such device syz_btf_id_by_name$bpf_lsm : failed to open /sys/kernel/btf/vmlinux: no such file or directory syz_init_net_socket$bt_cmtp : syz_init_net_socket$bt_cmtp(0x1f, 0x3, 0x5) failed: protocol not supported syz_kvm_setup_cpu$ppc64 : unsupported arch syz_mount_image$ntfs : /proc/filesystems does not contain ntfs syz_mount_image$reiserfs : /proc/filesystems does not contain reiserfs syz_mount_image$sysv : /proc/filesystems does not contain sysv syz_mount_image$v7 : /proc/filesystems does not contain v7 syz_open_dev$dricontrol : failed to open /dev/dri/controlD#: no such file or directory syz_open_dev$drirender : failed to open /dev/dri/renderD#: no such file or directory syz_open_dev$floppy : failed to open /dev/fd#: no such file or directory syz_open_dev$ircomm : failed to open /dev/ircomm#: no such file or directory syz_open_dev$sndhw : failed to open /dev/snd/hwC#D#: no such file or directory syz_pkey_set : pkey_alloc(0x0, 0x0) failed: no space left on device uselib : syscall uselib is not present write$selinux_access : selinux is not enabled write$selinux_attr : selinux is not enabled write$selinux_context : selinux is not enabled write$selinux_create : selinux is not enabled write$selinux_load : selinux is not enabled write$selinux_user : selinux is not enabled write$selinux_validatetrans : selinux is not enabled write$smack_current : smack is not enabled write$smackfs_access : smack is not enabled write$smackfs_change_rule : smack is not enabled write$smackfs_cipso : smack is not enabled write$smackfs_cipsonum : smack is not enabled write$smackfs_ipv6host : smack is not enabled write$smackfs_label : smack is not enabled write$smackfs_labels_list : smack is not enabled write$smackfs_load : smack is not enabled write$smackfs_logging : smack is not enabled write$smackfs_netlabel : smack is not enabled write$smackfs_ptrace : smack is not enabled transitively disabled the following syscalls (missing resource [creating syscalls]): bind$vsock_dgram : sock_vsock_dgram [socket$vsock_dgram] close$ibv_device : fd_rdma [openat$uverbs0] connect$hf : sock_hf [socket$hf] connect$vsock_dgram : sock_vsock_dgram [socket$vsock_dgram] getsockopt$inet6_dccp_buf : sock_dccp6 [socket$inet6_dccp] getsockopt$inet6_dccp_int : sock_dccp6 [socket$inet6_dccp] getsockopt$inet_dccp_buf : sock_dccp [socket$inet_dccp] getsockopt$inet_dccp_int : sock_dccp [socket$inet_dccp] ioctl$ACPI_THERMAL_GET_ART : fd_acpi_thermal_rel [openat$acpi_thermal_rel] ioctl$ACPI_THERMAL_GET_ART_COUNT : fd_acpi_thermal_rel [openat$acpi_thermal_rel] ioctl$ACPI_THERMAL_GET_ART_LEN : fd_acpi_thermal_rel [openat$acpi_thermal_rel] ioctl$ACPI_THERMAL_GET_TRT : fd_acpi_thermal_rel [openat$acpi_thermal_rel] ioctl$ACPI_THERMAL_GET_TRT_COUNT : fd_acpi_thermal_rel [openat$acpi_thermal_rel] ioctl$ACPI_THERMAL_GET_TRT_LEN : fd_acpi_thermal_rel [openat$acpi_thermal_rel] ioctl$ASHMEM_GET_NAME : fd_ashmem [openat$ashmem] ioctl$ASHMEM_GET_PIN_STATUS : fd_ashmem [openat$ashmem] ioctl$ASHMEM_GET_PROT_MASK : fd_ashmem [openat$ashmem] ioctl$ASHMEM_GET_SIZE : fd_ashmem [openat$ashmem] ioctl$ASHMEM_PURGE_ALL_CACHES : fd_ashmem [openat$ashmem] ioctl$ASHMEM_SET_NAME : fd_ashmem [openat$ashmem] ioctl$ASHMEM_SET_PROT_MASK : fd_ashmem [openat$ashmem] ioctl$ASHMEM_SET_SIZE : fd_ashmem [openat$ashmem] ioctl$CAPI_CLR_FLAGS : fd_capi20 [openat$capi20] ioctl$CAPI_GET_ERRCODE : fd_capi20 [openat$capi20] ioctl$CAPI_GET_FLAGS : fd_capi20 [openat$capi20] ioctl$CAPI_GET_MANUFACTURER : fd_capi20 [openat$capi20] ioctl$CAPI_GET_PROFILE : fd_capi20 [openat$capi20] ioctl$CAPI_GET_SERIAL : fd_capi20 [openat$capi20] ioctl$CAPI_INSTALLED : fd_capi20 [openat$capi20] ioctl$CAPI_MANUFACTURER_CMD : fd_capi20 [openat$capi20] ioctl$CAPI_NCCI_GETUNIT : fd_capi20 [openat$capi20] ioctl$CAPI_NCCI_OPENCOUNT : fd_capi20 [openat$capi20] ioctl$CAPI_REGISTER : fd_capi20 [openat$capi20] ioctl$CAPI_SET_FLAGS : fd_capi20 [openat$capi20] ioctl$CREATE_COUNTERS : fd_rdma [openat$uverbs0] ioctl$DESTROY_COUNTERS : fd_rdma [openat$uverbs0] ioctl$DRM_IOCTL_I915_GEM_BUSY : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_CONTEXT_CREATE : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_CONTEXT_DESTROY : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_CONTEXT_GETPARAM : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_CONTEXT_SETPARAM : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_CREATE : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_EXECBUFFER : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_EXECBUFFER2 : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_EXECBUFFER2_WR : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_GET_APERTURE : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_GET_CACHING : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_GET_TILING : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_MADVISE : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_MMAP : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_MMAP_GTT : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_MMAP_OFFSET : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_PIN : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_PREAD : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_PWRITE : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_SET_CACHING : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_SET_DOMAIN : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_SET_TILING : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_SW_FINISH : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_THROTTLE : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_UNPIN : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_USERPTR : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_VM_CREATE : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_VM_DESTROY : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_WAIT : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GETPARAM : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GET_PIPE_FROM_CRTC_ID : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GET_RESET_STATS : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_OVERLAY_ATTRS : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_OVERLAY_PUT_IMAGE : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_PERF_ADD_CONFIG : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_PERF_OPEN : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_PERF_REMOVE_CONFIG : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_QUERY : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_REG_READ : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_SET_SPRITE_COLORKEY : fd_i915 [openat$i915] ioctl$DRM_IOCTL_MSM_GEM_CPU_FINI : fd_msm [openat$msm] ioctl$DRM_IOCTL_MSM_GEM_CPU_PREP : fd_msm [openat$msm] ioctl$DRM_IOCTL_MSM_GEM_INFO : fd_msm [openat$msm] ioctl$DRM_IOCTL_MSM_GEM_MADVISE : fd_msm [openat$msm] ioctl$DRM_IOCTL_MSM_GEM_NEW : fd_msm [openat$msm] ioctl$DRM_IOCTL_MSM_GEM_SUBMIT : fd_msm [openat$msm] ioctl$DRM_IOCTL_MSM_GET_PARAM : fd_msm [openat$msm] ioctl$DRM_IOCTL_MSM_SET_PARAM : fd_msm [openat$msm] ioctl$DRM_IOCTL_MSM_SUBMITQUEUE_CLOSE : fd_msm [openat$msm] ioctl$DRM_IOCTL_MSM_SUBMITQUEUE_NEW : fd_msm [openat$msm] ioctl$DRM_IOCTL_MSM_SUBMITQUEUE_QUERY : fd_msm [openat$msm] ioctl$DRM_IOCTL_MSM_WAIT_FENCE : fd_msm [openat$msm] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_CACHE_CACHEOPEXEC: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_CACHE_CACHEOPLOG: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_CACHE_CACHEOPQUEUE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_CMM_DEVMEMINTACQUIREREMOTECTX: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_CMM_DEVMEMINTEXPORTCTX: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_CMM_DEVMEMINTUNEXPORTCTX: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_DEVICEMEMHISTORY_DEVICEMEMHISTORYMAP: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_DEVICEMEMHISTORY_DEVICEMEMHISTORYMAPVRANGE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_DEVICEMEMHISTORY_DEVICEMEMHISTORYSPARSECHANGE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_DEVICEMEMHISTORY_DEVICEMEMHISTORYUNMAP: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_DEVICEMEMHISTORY_DEVICEMEMHISTORYUNMAPVRANGE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_DMABUF_PHYSMEMEXPORTDMABUF: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_DMABUF_PHYSMEMIMPORTDMABUF: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_DMABUF_PHYSMEMIMPORTSPARSEDMABUF: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_HTBUFFER_HTBCONTROL: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_HTBUFFER_HTBLOG: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_CHANGESPARSEMEM: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMFLUSHDEVSLCRANGE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMGETFAULTADDRESS: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMINTCTXCREATE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMINTCTXDESTROY: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMINTHEAPCREATE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMINTHEAPDESTROY: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMINTMAPPAGES: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMINTMAPPMR: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMINTPIN: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMINTPINVALIDATE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMINTREGISTERPFNOTIFYKM: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMINTRESERVERANGE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMINTUNMAPPAGES: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMINTUNMAPPMR: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMINTUNPIN: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMINTUNPININVALIDATE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMINTUNRESERVERANGE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMINVALIDATEFBSCTABLE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMISVDEVADDRVALID: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_GETMAXDEVMEMSIZE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_HEAPCFGHEAPCONFIGCOUNT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_HEAPCFGHEAPCONFIGNAME: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_HEAPCFGHEAPCOUNT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_HEAPCFGHEAPDETAILS: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_PHYSMEMNEWRAMBACKEDLOCKEDPMR: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_PHYSMEMNEWRAMBACKEDPMR: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_PMREXPORTPMR: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_PMRGETUID: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_PMRIMPORTPMR: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_PMRLOCALIMPORTPMR: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_PMRMAKELOCALIMPORTHANDLE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_PMRUNEXPORTPMR: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_PMRUNMAKELOCALIMPORTHANDLE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_PMRUNREFPMR: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_PMRUNREFUNLOCKPMR: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_PVRSRVUPDATEOOMSTATS: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_PVRTL_TLACQUIREDATA: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_PVRTL_TLCLOSESTREAM: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_PVRTL_TLCOMMITSTREAM: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_PVRTL_TLDISCOVERSTREAMS: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_PVRTL_TLOPENSTREAM: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_PVRTL_TLRELEASEDATA: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_PVRTL_TLRESERVESTREAM: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_PVRTL_TLWRITEDATA: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXBREAKPOINT_RGXCLEARBREAKPOINT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXBREAKPOINT_RGXDISABLEBREAKPOINT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXBREAKPOINT_RGXENABLEBREAKPOINT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXBREAKPOINT_RGXOVERALLOCATEBPREGISTERS: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXBREAKPOINT_RGXSETBREAKPOINT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXCMP_RGXCREATECOMPUTECONTEXT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXCMP_RGXDESTROYCOMPUTECONTEXT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXCMP_RGXFLUSHCOMPUTEDATA: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXCMP_RGXGETLASTCOMPUTECONTEXTRESETREASON: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXCMP_RGXKICKCDM2: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXCMP_RGXNOTIFYCOMPUTEWRITEOFFSETUPDATE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXCMP_RGXSETCOMPUTECONTEXTPRIORITY: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXCMP_RGXSETCOMPUTECONTEXTPROPERTY: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXFWDBG_RGXCURRENTTIME: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXFWDBG_RGXFWDEBUGDUMPFREELISTPAGELIST: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXFWDBG_RGXFWDEBUGPHRCONFIGURE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXFWDBG_RGXFWDEBUGSETFWLOG: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXFWDBG_RGXFWDEBUGSETHCSDEADLINE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXFWDBG_RGXFWDEBUGSETOSIDPRIORITY: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXFWDBG_RGXFWDEBUGSETOSNEWONLINESTATE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXHWPERF_RGXCONFIGCUSTOMCOUNTERS: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXHWPERF_RGXCONFIGENABLEHWPERFCOUNTERS: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXHWPERF_RGXCTRLHWPERF: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXHWPERF_RGXCTRLHWPERFCOUNTERS: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXHWPERF_RGXGETHWPERFBVNCFEATUREFLAGS: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXKICKSYNC_RGXCREATEKICKSYNCCONTEXT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXKICKSYNC_RGXDESTROYKICKSYNCCONTEXT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXKICKSYNC_RGXKICKSYNC2: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXKICKSYNC_RGXSETKICKSYNCCONTEXTPROPERTY: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXREGCONFIG_RGXADDREGCONFIG: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXREGCONFIG_RGXCLEARREGCONFIG: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXREGCONFIG_RGXDISABLEREGCONFIG: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXREGCONFIG_RGXENABLEREGCONFIG: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXREGCONFIG_RGXSETREGCONFIGTYPE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXSIGNALS_RGXNOTIFYSIGNALUPDATE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTA3D_RGXCREATEFREELIST: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTA3D_RGXCREATEHWRTDATASET: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTA3D_RGXCREATERENDERCONTEXT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTA3D_RGXCREATEZSBUFFER: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTA3D_RGXDESTROYFREELIST: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTA3D_RGXDESTROYHWRTDATASET: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTA3D_RGXDESTROYRENDERCONTEXT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTA3D_RGXDESTROYZSBUFFER: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTA3D_RGXGETLASTRENDERCONTEXTRESETREASON: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTA3D_RGXKICKTA3D2: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTA3D_RGXPOPULATEZSBUFFER: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTA3D_RGXRENDERCONTEXTSTALLED: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTA3D_RGXSETRENDERCONTEXTPRIORITY: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTA3D_RGXSETRENDERCONTEXTPROPERTY: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTA3D_RGXUNPOPULATEZSBUFFER: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTQ2_RGXTDMCREATETRANSFERCONTEXT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTQ2_RGXTDMDESTROYTRANSFERCONTEXT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTQ2_RGXTDMGETSHAREDMEMORY: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTQ2_RGXTDMNOTIFYWRITEOFFSETUPDATE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTQ2_RGXTDMRELEASESHAREDMEMORY: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTQ2_RGXTDMSETTRANSFERCONTEXTPRIORITY: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTQ2_RGXTDMSETTRANSFERCONTEXTPROPERTY: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTQ2_RGXTDMSUBMITTRANSFER2: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTQ_RGXCREATETRANSFERCONTEXT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTQ_RGXDESTROYTRANSFERCONTEXT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTQ_RGXSETTRANSFERCONTEXTPRIORITY: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTQ_RGXSETTRANSFERCONTEXTPROPERTY: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTQ_RGXSUBMITTRANSFER2: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_ACQUIREGLOBALEVENTOBJECT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_ACQUIREINFOPAGE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_ALIGNMENTCHECK: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_CONNECT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_DISCONNECT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_DUMPDEBUGINFO: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_EVENTOBJECTCLOSE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_EVENTOBJECTOPEN: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_EVENTOBJECTWAIT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_EVENTOBJECTWAITTIMEOUT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_FINDPROCESSMEMSTATS: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_GETDEVCLOCKSPEED: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_GETDEVICESTATUS: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_GETMULTICOREINFO: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_HWOPTIMEOUT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_RELEASEGLOBALEVENTOBJECT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_RELEASEINFOPAGE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SYNCTRACKING_SYNCRECORDADD: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SYNCTRACKING_SYNCRECORDREMOVEBYHANDLE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SYNC_ALLOCSYNCPRIMITIVEBLOCK: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SYNC_FREESYNCPRIMITIVEBLOCK: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SYNC_SYNCALLOCEVENT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SYNC_SYNCCHECKPOINTSIGNALLEDPDUMPPOL: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SYNC_SYNCFREEEVENT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SYNC_SYNCPRIMPDUMP: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SYNC_SYNCPRIMPDUMPCBP: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SYNC_SYNCPRIMPDUMPPOL: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SYNC_SYNCPRIMPDUMPVALUE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SYNC_SYNCPRIMSET: fd_rogue [openat$img_rogue] ioctl$FLOPPY_FDCLRPRM : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDDEFPRM : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDEJECT : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDFLUSH : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDFMTBEG : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDFMTEND : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDFMTTRK : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDGETDRVPRM : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDGETDRVSTAT : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDGETDRVTYP : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDGETFDCSTAT : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDGETMAXERRS : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDGETPRM : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDMSGOFF : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDMSGON : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDPOLLDRVSTAT : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDRAWCMD : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDRESET : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDSETDRVPRM : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDSETEMSGTRESH : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDSETMAXERRS : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDSETPRM : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDTWADDLE : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDWERRORCLR : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDWERRORGET : fd_floppy [syz_open_dev$floppy] ioctl$KBASE_HWCNT_READER_CLEAR : fd_hwcnt [ioctl$KBASE_IOCTL_HWCNT_READER_SETUP] ioctl$KBASE_HWCNT_READER_DISABLE_EVENT : fd_hwcnt [ioctl$KBASE_IOCTL_HWCNT_READER_SETUP] ioctl$KBASE_HWCNT_READER_DUMP : fd_hwcnt [ioctl$KBASE_IOCTL_HWCNT_READER_SETUP] ioctl$KBASE_HWCNT_READER_ENABLE_EVENT : fd_hwcnt [ioctl$KBASE_IOCTL_HWCNT_READER_SETUP] ioctl$KBASE_HWCNT_READER_GET_API_VERSION : fd_hwcnt [ioctl$KBASE_IOCTL_HWCNT_READER_SETUP] ioctl$KBASE_HWCNT_READER_GET_API_VERSION_WITH_FEATURES: fd_hwcnt [ioctl$KBASE_IOCTL_HWCNT_READER_SETUP] ioctl$KBASE_HWCNT_READER_GET_BUFFER : fd_hwcnt [ioctl$KBASE_IOCTL_HWCNT_READER_SETUP] ioctl$KBASE_HWCNT_READER_GET_BUFFER_SIZE : fd_hwcnt [ioctl$KBASE_IOCTL_HWCNT_READER_SETUP] ioctl$KBASE_HWCNT_READER_GET_BUFFER_WITH_CYCLES: fd_hwcnt [ioctl$KBASE_IOCTL_HWCNT_READER_SETUP] ioctl$KBASE_HWCNT_READER_GET_HWVER : fd_hwcnt [ioctl$KBASE_IOCTL_HWCNT_READER_SETUP] ioctl$KBASE_HWCNT_READER_PUT_BUFFER : fd_hwcnt [ioctl$KBASE_IOCTL_HWCNT_READER_SETUP] ioctl$KBASE_HWCNT_READER_PUT_BUFFER_WITH_CYCLES: fd_hwcnt [ioctl$KBASE_IOCTL_HWCNT_READER_SETUP] ioctl$KBASE_HWCNT_READER_SET_INTERVAL : fd_hwcnt [ioctl$KBASE_IOCTL_HWCNT_READER_SETUP] ioctl$KBASE_IOCTL_BUFFER_LIVENESS_UPDATE : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_CONTEXT_PRIORITY_CHECK : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_CS_CPU_QUEUE_DUMP : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_CS_EVENT_SIGNAL : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_CS_GET_GLB_IFACE : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_CS_QUEUE_BIND : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_CS_QUEUE_GROUP_CREATE : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_CS_QUEUE_GROUP_CREATE_1_6 : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_CS_QUEUE_GROUP_TERMINATE : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_CS_QUEUE_KICK : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_CS_QUEUE_REGISTER : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_CS_QUEUE_REGISTER_EX : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_CS_QUEUE_TERMINATE : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_CS_TILER_HEAP_INIT : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_CS_TILER_HEAP_INIT_1_13 : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_CS_TILER_HEAP_TERM : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_DISJOINT_QUERY : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_FENCE_VALIDATE : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_GET_CONTEXT_ID : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_GET_CPU_GPU_TIMEINFO : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_GET_DDK_VERSION : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_GET_GPUPROPS : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_HWCNT_CLEAR : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_HWCNT_DUMP : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_HWCNT_ENABLE : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_HWCNT_READER_SETUP : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_HWCNT_SET : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_JOB_SUBMIT : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_KCPU_QUEUE_CREATE : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_KCPU_QUEUE_DELETE : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_KCPU_QUEUE_ENQUEUE : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_KINSTR_PRFCNT_CMD : fd_kinstr [ioctl$KBASE_IOCTL_KINSTR_PRFCNT_SETUP] ioctl$KBASE_IOCTL_KINSTR_PRFCNT_ENUM_INFO : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_KINSTR_PRFCNT_GET_SAMPLE : fd_kinstr [ioctl$KBASE_IOCTL_KINSTR_PRFCNT_SETUP] ioctl$KBASE_IOCTL_KINSTR_PRFCNT_PUT_SAMPLE : fd_kinstr [ioctl$KBASE_IOCTL_KINSTR_PRFCNT_SETUP] ioctl$KBASE_IOCTL_KINSTR_PRFCNT_SETUP : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_MEM_ALIAS : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_MEM_ALLOC : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_MEM_ALLOC_EX : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_MEM_COMMIT : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_MEM_EXEC_INIT : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_MEM_FIND_CPU_OFFSET : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_MEM_FIND_GPU_START_AND_OFFSET: fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_MEM_FLAGS_CHANGE : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_MEM_FREE : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_MEM_IMPORT : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_MEM_JIT_INIT : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_MEM_JIT_INIT_10_2 : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_MEM_JIT_INIT_11_5 : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_MEM_PROFILE_ADD : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_MEM_QUERY : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_MEM_SYNC : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_POST_TERM : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_READ_USER_PAGE : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_SET_FLAGS : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_SET_LIMITED_CORE_COUNT : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_SOFT_EVENT_UPDATE : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_STICKY_RESOURCE_MAP : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_STICKY_RESOURCE_UNMAP : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_STREAM_CREATE : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_TLSTREAM_ACQUIRE : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_TLSTREAM_FLUSH : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_VERSION_CHECK : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_VERSION_CHECK_RESERVED : fd_bifrost [openat$bifrost openat$mali] ioctl$KVM_ASSIGN_SET_MSIX_ENTRY : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_ASSIGN_SET_MSIX_NR : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_DIRTY_LOG_RING : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_DIRTY_LOG_RING_ACQ_REL : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_DISABLE_QUIRKS : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_DISABLE_QUIRKS2 : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_ENFORCE_PV_FEATURE_CPUID : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_CAP_EXCEPTION_PAYLOAD : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_EXIT_HYPERCALL : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_EXIT_ON_EMULATION_FAILURE : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_HALT_POLL : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_HYPERV_DIRECT_TLBFLUSH : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_CAP_HYPERV_ENFORCE_CPUID : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_CAP_HYPERV_ENLIGHTENED_VMCS : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_CAP_HYPERV_SEND_IPI : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_HYPERV_SYNIC : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_CAP_HYPERV_SYNIC2 : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_CAP_HYPERV_TLBFLUSH : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_HYPERV_VP_INDEX : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_MANUAL_DIRTY_LOG_PROTECT2 : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_MAX_VCPU_ID : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_MEMORY_FAULT_INFO : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_MSR_PLATFORM_INFO : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_PMU_CAPABILITY : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_PTP_KVM : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_SGX_ATTRIBUTE : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_SPLIT_IRQCHIP : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_STEAL_TIME : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_SYNC_REGS : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_CAP_VM_COPY_ENC_CONTEXT_FROM : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_VM_DISABLE_NX_HUGE_PAGES : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_VM_MOVE_ENC_CONTEXT_FROM : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_VM_TYPES : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_X2APIC_API : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_X86_APIC_BUS_CYCLES_NS : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_X86_BUS_LOCK_EXIT : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_X86_DISABLE_EXITS : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_X86_GUEST_MODE : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_X86_NOTIFY_VMEXIT : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_X86_USER_SPACE_MSR : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_XEN_HVM : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CHECK_EXTENSION : fd_kvm [openat$kvm] ioctl$KVM_CHECK_EXTENSION_VM : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CLEAR_DIRTY_LOG : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CREATE_DEVICE : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CREATE_GUEST_MEMFD : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CREATE_IRQCHIP : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CREATE_PIT2 : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CREATE_VCPU : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CREATE_VM : fd_kvm [openat$kvm] ioctl$KVM_DIRTY_TLB : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_API_VERSION : fd_kvm [openat$kvm] ioctl$KVM_GET_CLOCK : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_GET_CPUID2 : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_DEBUGREGS : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_DEVICE_ATTR : fd_kvmdev [ioctl$KVM_CREATE_DEVICE] ioctl$KVM_GET_DEVICE_ATTR_vcpu : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_DEVICE_ATTR_vm : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_GET_DIRTY_LOG : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_GET_EMULATED_CPUID : fd_kvm [openat$kvm] ioctl$KVM_GET_FPU : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_IRQCHIP : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_GET_LAPIC : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_MP_STATE : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_MSRS_cpu : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_MSRS_sys : fd_kvm [openat$kvm] ioctl$KVM_GET_MSR_FEATURE_INDEX_LIST : fd_kvm [openat$kvm] ioctl$KVM_GET_MSR_INDEX_LIST : fd_kvm [openat$kvm] ioctl$KVM_GET_NESTED_STATE : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_NR_MMU_PAGES : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_GET_ONE_REG : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_PIT : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_GET_PIT2 : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_GET_REGS : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_REG_LIST : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_SREGS : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_SREGS2 : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_STATS_FD_cpu : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_STATS_FD_vm : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_GET_SUPPORTED_CPUID : fd_kvm [openat$kvm] ioctl$KVM_GET_SUPPORTED_HV_CPUID_cpu : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_SUPPORTED_HV_CPUID_sys : fd_kvm [openat$kvm] ioctl$KVM_GET_TSC_KHZ_cpu : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_TSC_KHZ_vm : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_GET_VCPU_EVENTS : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_VCPU_MMAP_SIZE : fd_kvm [openat$kvm] ioctl$KVM_GET_XCRS : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_XSAVE : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_XSAVE2 : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_HAS_DEVICE_ATTR : fd_kvmdev [ioctl$KVM_CREATE_DEVICE] ioctl$KVM_HAS_DEVICE_ATTR_vcpu : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_HAS_DEVICE_ATTR_vm : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_HYPERV_EVENTFD : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_INTERRUPT : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_IOEVENTFD : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_IRQFD : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_IRQ_LINE : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_IRQ_LINE_STATUS : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_KVMCLOCK_CTRL : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_MEMORY_ENCRYPT_REG_REGION : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_MEMORY_ENCRYPT_UNREG_REGION : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_NMI : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_PPC_ALLOCATE_HTAB : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_PRE_FAULT_MEMORY : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_REGISTER_COALESCED_MMIO : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_REINJECT_CONTROL : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_RESET_DIRTY_RINGS : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_RUN : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_S390_VCPU_FAULT : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_BOOT_CPU_ID : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SET_CLOCK : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SET_CPUID : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_CPUID2 : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_DEBUGREGS : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_DEVICE_ATTR : fd_kvmdev [ioctl$KVM_CREATE_DEVICE] ioctl$KVM_SET_DEVICE_ATTR_vcpu : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_DEVICE_ATTR_vm : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SET_FPU : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_GSI_ROUTING : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SET_GUEST_DEBUG : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_IDENTITY_MAP_ADDR : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SET_IRQCHIP : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SET_LAPIC : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_MEMORY_ATTRIBUTES : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SET_MP_STATE : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_MSRS : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_NESTED_STATE : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_NR_MMU_PAGES : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SET_ONE_REG : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_PIT : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SET_PIT2 : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SET_REGS : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_SIGNAL_MASK : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_SREGS : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_SREGS2 : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_TSC_KHZ_cpu : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_TSC_KHZ_vm : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SET_TSS_ADDR : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SET_USER_MEMORY_REGION : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SET_USER_MEMORY_REGION2 : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SET_VAPIC_ADDR : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_VCPU_EVENTS : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_XCRS : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_XSAVE : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SEV_CERT_EXPORT : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_DBG_DECRYPT : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_DBG_ENCRYPT : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_ES_INIT : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_GET_ATTESTATION_REPORT : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_GUEST_STATUS : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_INIT : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_INIT2 : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_LAUNCH_FINISH : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_LAUNCH_MEASURE : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_LAUNCH_SECRET : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_LAUNCH_START : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_LAUNCH_UPDATE_DATA : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_LAUNCH_UPDATE_VMSA : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_RECEIVE_FINISH : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_RECEIVE_START : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_RECEIVE_UPDATE_DATA : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_RECEIVE_UPDATE_VMSA : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_SEND_CANCEL : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_SEND_FINISH : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_SEND_START : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_SEND_UPDATE_DATA : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_SEND_UPDATE_VMSA : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_SNP_LAUNCH_FINISH : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_SNP_LAUNCH_START : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_SNP_LAUNCH_UPDATE : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SIGNAL_MSI : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SMI : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_TPR_ACCESS_REPORTING : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_TRANSLATE : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_UNREGISTER_COALESCED_MMIO : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_X86_GET_MCE_CAP_SUPPORTED : fd_kvm [openat$kvm] ioctl$KVM_X86_SETUP_MCE : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_X86_SET_MCE : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_X86_SET_MSR_FILTER : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_XEN_HVM_CONFIG : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$PERF_EVENT_IOC_DISABLE : fd_perf [perf_event_open perf_event_open$cgroup] ioctl$PERF_EVENT_IOC_ENABLE : fd_perf [perf_event_open perf_event_open$cgroup] ioctl$PERF_EVENT_IOC_ID : fd_perf [perf_event_open perf_event_open$cgroup] ioctl$PERF_EVENT_IOC_MODIFY_ATTRIBUTES : fd_perf [perf_event_open perf_event_open$cgroup] ioctl$PERF_EVENT_IOC_PAUSE_OUTPUT : fd_perf [perf_event_open perf_event_open$cgroup] ioctl$PERF_EVENT_IOC_PERIOD : fd_perf [perf_event_open perf_event_open$cgroup] ioctl$PERF_EVENT_IOC_QUERY_BPF : fd_perf [perf_event_open perf_event_open$cgroup] ioctl$PERF_EVENT_IOC_REFRESH : fd_perf [perf_event_open perf_event_open$cgroup] ioctl$PERF_EVENT_IOC_RESET : fd_perf [perf_event_open perf_event_open$cgroup] ioctl$PERF_EVENT_IOC_SET_BPF : fd_perf [perf_event_open perf_event_open$cgroup] ioctl$PERF_EVENT_IOC_SET_FILTER : fd_perf [perf_event_open perf_event_open$cgroup] ioctl$PERF_EVENT_IOC_SET_OUTPUT : fd_perf [perf_event_open perf_event_open$cgroup] ioctl$READ_COUNTERS : fd_rdma [openat$uverbs0] ioctl$SNDRV_FIREWIRE_IOCTL_GET_INFO : fd_snd_hw [syz_open_dev$sndhw] ioctl$SNDRV_FIREWIRE_IOCTL_LOCK : fd_snd_hw [syz_open_dev$sndhw] ioctl$SNDRV_FIREWIRE_IOCTL_TASCAM_STATE : fd_snd_hw [syz_open_dev$sndhw] ioctl$SNDRV_FIREWIRE_IOCTL_UNLOCK : fd_snd_hw [syz_open_dev$sndhw] ioctl$SNDRV_HWDEP_IOCTL_DSP_LOAD : fd_snd_hw [syz_open_dev$sndhw] ioctl$SNDRV_HWDEP_IOCTL_DSP_STATUS : fd_snd_hw [syz_open_dev$sndhw] ioctl$SNDRV_HWDEP_IOCTL_INFO : fd_snd_hw [syz_open_dev$sndhw] ioctl$SNDRV_HWDEP_IOCTL_PVERSION : fd_snd_hw [syz_open_dev$sndhw] ioctl$TE_IOCTL_CLOSE_CLIENT_SESSION : fd_tlk [openat$tlk_device] ioctl$TE_IOCTL_LAUNCH_OPERATION : fd_tlk [openat$tlk_device] ioctl$TE_IOCTL_OPEN_CLIENT_SESSION : fd_tlk [openat$tlk_device] ioctl$TE_IOCTL_SS_CMD : fd_tlk [openat$tlk_device] ioctl$TIPC_IOC_CONNECT : fd_trusty [openat$trusty openat$trusty_avb openat$trusty_gatekeeper ...] ioctl$TIPC_IOC_CONNECT_avb : fd_trusty_avb [openat$trusty_avb] ioctl$TIPC_IOC_CONNECT_gatekeeper : fd_trusty_gatekeeper [openat$trusty_gatekeeper] ioctl$TIPC_IOC_CONNECT_hwkey : fd_trusty_hwkey [openat$trusty_hwkey] ioctl$TIPC_IOC_CONNECT_hwrng : fd_trusty_hwrng [openat$trusty_hwrng] ioctl$TIPC_IOC_CONNECT_keymaster_secure : fd_trusty_km_secure [openat$trusty_km_secure] ioctl$TIPC_IOC_CONNECT_km : fd_trusty_km [openat$trusty_km] ioctl$TIPC_IOC_CONNECT_storage : fd_trusty_storage [openat$trusty_storage] ioctl$VFIO_CHECK_EXTENSION : fd_vfio [openat$vfio] ioctl$VFIO_GET_API_VERSION : fd_vfio [openat$vfio] ioctl$VFIO_IOMMU_GET_INFO : fd_vfio [openat$vfio] ioctl$VFIO_IOMMU_MAP_DMA : fd_vfio [openat$vfio] ioctl$VFIO_IOMMU_UNMAP_DMA : fd_vfio [openat$vfio] ioctl$VFIO_SET_IOMMU : fd_vfio [openat$vfio] ioctl$VTPM_PROXY_IOC_NEW_DEV : fd_vtpm [openat$vtpm] ioctl$sock_bt_cmtp_CMTPCONNADD : sock_bt_cmtp [syz_init_net_socket$bt_cmtp] ioctl$sock_bt_cmtp_CMTPCONNDEL : sock_bt_cmtp [syz_init_net_socket$bt_cmtp] ioctl$sock_bt_cmtp_CMTPGETCONNINFO : sock_bt_cmtp [syz_init_net_socket$bt_cmtp] ioctl$sock_bt_cmtp_CMTPGETCONNLIST : sock_bt_cmtp [syz_init_net_socket$bt_cmtp] mmap$DRM_I915 : fd_i915 [openat$i915] mmap$DRM_MSM : fd_msm [openat$msm] mmap$KVM_VCPU : vcpu_mmap_size [ioctl$KVM_GET_VCPU_MMAP_SIZE] mmap$bifrost : fd_bifrost [openat$bifrost openat$mali] mmap$perf : fd_perf [perf_event_open perf_event_open$cgroup] pkey_free : pkey [pkey_alloc] pkey_mprotect : pkey [pkey_alloc] read$sndhw : fd_snd_hw [syz_open_dev$sndhw] read$trusty : fd_trusty [openat$trusty openat$trusty_avb openat$trusty_gatekeeper ...] recvmsg$hf : sock_hf [socket$hf] sendmsg$hf : sock_hf [socket$hf] setsockopt$inet6_dccp_buf : sock_dccp6 [socket$inet6_dccp] setsockopt$inet6_dccp_int : sock_dccp6 [socket$inet6_dccp] setsockopt$inet_dccp_buf : sock_dccp [socket$inet_dccp] setsockopt$inet_dccp_int : sock_dccp [socket$inet_dccp] syz_kvm_add_vcpu$x86 : kvm_syz_vm$x86 [syz_kvm_setup_syzos_vm$x86] syz_kvm_assert_syzos_uexit$x86 : kvm_run_ptr [mmap$KVM_VCPU] syz_kvm_setup_cpu$x86 : fd_kvmvm [ioctl$KVM_CREATE_VM] syz_kvm_setup_syzos_vm$x86 : fd_kvmvm [ioctl$KVM_CREATE_VM] syz_memcpy_off$KVM_EXIT_HYPERCALL : kvm_run_ptr [mmap$KVM_VCPU] syz_memcpy_off$KVM_EXIT_MMIO : kvm_run_ptr [mmap$KVM_VCPU] write$ALLOC_MW : fd_rdma [openat$uverbs0] write$ALLOC_PD : fd_rdma [openat$uverbs0] write$ATTACH_MCAST : fd_rdma [openat$uverbs0] write$CLOSE_XRCD : fd_rdma [openat$uverbs0] write$CREATE_AH : fd_rdma [openat$uverbs0] write$CREATE_COMP_CHANNEL : fd_rdma [openat$uverbs0] write$CREATE_CQ : fd_rdma [openat$uverbs0] write$CREATE_CQ_EX : fd_rdma [openat$uverbs0] write$CREATE_FLOW : fd_rdma [openat$uverbs0] write$CREATE_QP : fd_rdma [openat$uverbs0] write$CREATE_RWQ_IND_TBL : fd_rdma [openat$uverbs0] write$CREATE_SRQ : fd_rdma [openat$uverbs0] write$CREATE_WQ : fd_rdma [openat$uverbs0] write$DEALLOC_MW : fd_rdma [openat$uverbs0] write$DEALLOC_PD : fd_rdma [openat$uverbs0] write$DEREG_MR : fd_rdma [openat$uverbs0] write$DESTROY_AH : fd_rdma [openat$uverbs0] write$DESTROY_CQ : fd_rdma [openat$uverbs0] write$DESTROY_FLOW : fd_rdma [openat$uverbs0] write$DESTROY_QP : fd_rdma [openat$uverbs0] write$DESTROY_RWQ_IND_TBL : fd_rdma [openat$uverbs0] write$DESTROY_SRQ : fd_rdma [openat$uverbs0] write$DESTROY_WQ : fd_rdma [openat$uverbs0] write$DETACH_MCAST : fd_rdma [openat$uverbs0] write$MLX5_ALLOC_PD : fd_rdma [openat$uverbs0] write$MLX5_CREATE_CQ : fd_rdma [openat$uverbs0] write$MLX5_CREATE_DV_QP : fd_rdma [openat$uverbs0] write$MLX5_CREATE_QP : fd_rdma [openat$uverbs0] write$MLX5_CREATE_SRQ : fd_rdma [openat$uverbs0] write$MLX5_CREATE_WQ : fd_rdma [openat$uverbs0] write$MLX5_GET_CONTEXT : fd_rdma [openat$uverbs0] write$MLX5_MODIFY_WQ : fd_rdma [openat$uverbs0] write$MODIFY_QP : fd_rdma [openat$uverbs0] write$MODIFY_SRQ : fd_rdma [openat$uverbs0] write$OPEN_XRCD : fd_rdma [openat$uverbs0] write$POLL_CQ : fd_rdma [openat$uverbs0] write$POST_RECV : fd_rdma [openat$uverbs0] write$POST_SEND : fd_rdma [openat$uverbs0] write$POST_SRQ_RECV : fd_rdma [openat$uverbs0] write$QUERY_DEVICE_EX : fd_rdma [openat$uverbs0] write$QUERY_PORT : fd_rdma [openat$uverbs0] write$QUERY_QP : fd_rdma [openat$uverbs0] write$QUERY_SRQ : fd_rdma [openat$uverbs0] write$REG_MR : fd_rdma [openat$uverbs0] write$REQ_NOTIFY_CQ : fd_rdma [openat$uverbs0] write$REREG_MR : fd_rdma [openat$uverbs0] write$RESIZE_CQ : fd_rdma [openat$uverbs0] write$capi20 : fd_capi20 [openat$capi20] write$capi20_data : fd_capi20 [openat$capi20] write$damon_attrs : fd_damon_attrs [openat$damon_attrs] write$damon_contexts : fd_damon_contexts [openat$damon_mk_contexts openat$damon_rm_contexts] write$damon_init_regions : fd_damon_init_regions [openat$damon_init_regions] write$damon_monitor_on : fd_damon_monitor_on [openat$damon_monitor_on] write$damon_schemes : fd_damon_schemes [openat$damon_schemes] write$damon_target_ids : fd_damon_target_ids [openat$damon_target_ids] write$proc_reclaim : fd_proc_reclaim [openat$proc_reclaim] write$sndhw : fd_snd_hw [syz_open_dev$sndhw] write$sndhw_fireworks : fd_snd_hw [syz_open_dev$sndhw] write$trusty : fd_trusty [openat$trusty openat$trusty_avb openat$trusty_gatekeeper ...] write$trusty_avb : fd_trusty_avb [openat$trusty_avb] write$trusty_gatekeeper : fd_trusty_gatekeeper [openat$trusty_gatekeeper] write$trusty_hwkey : fd_trusty_hwkey [openat$trusty_hwkey] write$trusty_hwrng : fd_trusty_hwrng [openat$trusty_hwrng] write$trusty_km : fd_trusty_km [openat$trusty_km] write$trusty_km_secure : fd_trusty_km_secure [openat$trusty_km_secure] write$trusty_storage : fd_trusty_storage [openat$trusty_storage] BinFmtMisc : enabled Comparisons : enabled Coverage : enabled DelayKcovMmap : enabled DevlinkPCI : PCI device 0000:00:10.0 is not available ExtraCoverage : enabled Fault : enabled KCSAN : write(/sys/kernel/debug/kcsan, on) failed KcovResetIoctl : kernel does not support ioctl(KCOV_RESET_TRACE) LRWPANEmulation : enabled Leak : failed to write(kmemleak, "scan=off") NetDevices : enabled NetInjection : enabled NicVF : PCI device 0000:00:11.0 is not available SandboxAndroid : setfilecon: setxattr failed. (errno 1: Operation not permitted). . process exited with status 67. SandboxNamespace : enabled SandboxNone : enabled SandboxSetuid : enabled Swap : enabled USBEmulation : enabled VhciInjection : enabled WifiEmulation : enabled syscalls : 3832/8048 2025/08/19 02:30:41 base: machine check complete 2025/08/19 02:30:42 coverage filter: __UNIQUE_ID_nop1452: [] 2025/08/19 02:30:42 coverage filter: __UNIQUE_ID_nop1453: [] 2025/08/19 02:30:42 coverage filter: __bio_release_pages: [__bio_release_pages] 2025/08/19 02:30:42 coverage filter: __dec_node_state: [] 2025/08/19 02:30:42 coverage filter: __f2fs_commit_super: [__f2fs_commit_super] 2025/08/19 02:30:42 coverage filter: __filemap_fdatawait_range: [__filemap_fdatawait_range] 2025/08/19 02:30:42 coverage filter: __filemap_get_folio: [__filemap_get_folio] 2025/08/19 02:30:42 coverage filter: __folio_put: [__folio_put] 2025/08/19 02:30:42 coverage filter: __get_meta_folio: [__get_meta_folio] 2025/08/19 02:30:42 coverage filter: __get_metapage: [__get_metapage] 2025/08/19 02:30:42 coverage filter: __get_node_folio: [__get_node_folio] 2025/08/19 02:30:42 coverage filter: __inc_node_state: [] 2025/08/19 02:30:42 coverage filter: __mem_cgroup_try_charge_swap: [] 2025/08/19 02:30:42 coverage filter: __mem_cgroup_uncharge_swap: [] 2025/08/19 02:30:42 coverage filter: __memcg_kmem_charge_page: [] 2025/08/19 02:30:42 coverage filter: __memcg_kmem_uncharge_page: [] 2025/08/19 02:30:42 coverage filter: __mod_node_page_state: [] 2025/08/19 02:30:42 coverage filter: __node_get_memory_tier: [__node_get_memory_tier] 2025/08/19 02:30:42 coverage filter: __oom_kill_process: [__oom_kill_process] 2025/08/19 02:30:42 coverage filter: __pfx_filemap_mod_uncharged_vmstat: [] 2025/08/19 02:30:42 coverage filter: __se_sys_cachestat: [__se_sys_cachestat] 2025/08/19 02:30:42 coverage filter: attr_wof_frame_info: [attr_wof_frame_info] 2025/08/19 02:30:42 coverage filter: bch2_mark_pagecache_reserved: [bch2_mark_pagecache_reserved] 2025/08/19 02:30:42 coverage filter: bch2_mark_pagecache_unallocated: [bch2_mark_pagecache_unallocated] 2025/08/19 02:30:42 coverage filter: bch2_page_mkwrite: [bch2_page_mkwrite] 2025/08/19 02:30:42 coverage filter: bch2_seek_pagecache_data: [bch2_seek_pagecache_data] 2025/08/19 02:30:42 coverage filter: bio_set_pages_dirty: [bio_set_pages_dirty] 2025/08/19 02:30:42 coverage filter: block_page_mkwrite: [block_page_mkwrite] 2025/08/19 02:30:42 coverage filter: btrfs_cleanup_bg_io: [btrfs_cleanup_bg_io] 2025/08/19 02:30:42 coverage filter: btrfs_cleanup_one_transaction: [btrfs_cleanup_one_transaction] 2025/08/19 02:30:42 coverage filter: btrfs_cleanup_transaction: [btrfs_cleanup_transaction] 2025/08/19 02:30:42 coverage filter: btrfs_clear_buffer_dirty: [btrfs_clear_buffer_dirty] 2025/08/19 02:30:42 coverage filter: btrfs_defrag_file: [btrfs_defrag_file] 2025/08/19 02:30:42 coverage filter: btrfs_drop_and_free_fs_root: [btrfs_drop_and_free_fs_root] 2025/08/19 02:30:42 coverage filter: btrfs_mark_buffer_dirty: [btrfs_mark_buffer_dirty] 2025/08/19 02:30:42 coverage filter: btrfs_page_mkwrite: [btrfs_page_mkwrite] 2025/08/19 02:30:42 coverage filter: btrfs_read_merkle_tree_page: [btrfs_read_merkle_tree_page] 2025/08/19 02:30:42 coverage filter: btrfs_replay_log: [btrfs_replay_log] 2025/08/19 02:30:42 coverage filter: btrfs_truncate_block: [btrfs_truncate_block] 2025/08/19 02:30:42 coverage filter: btrfs_writepage_fixup_worker: [btrfs_writepage_fixup_worker] 2025/08/19 02:30:42 coverage filter: btrfs_writepages: [btrfs_writepages] 2025/08/19 02:30:42 coverage filter: ceph_page_mkwrite: [ceph_page_mkwrite] 2025/08/19 02:30:42 coverage filter: ceph_uninline_data: [ceph_uninline_data] 2025/08/19 02:30:42 coverage filter: ceph_writepages_start: [ceph_writepages_start] 2025/08/19 02:30:42 coverage filter: cgroup_id_from_mm: [cgroup_id_from_mm] 2025/08/19 02:30:42 coverage filter: change_prot_numa: [change_prot_numa] 2025/08/19 02:30:42 coverage filter: clean_bdev_aliases: [clean_bdev_aliases] 2025/08/19 02:30:42 coverage filter: compact_zone: [compact_zone] 2025/08/19 02:30:42 coverage filter: compaction_alloc: [compaction_alloc] 2025/08/19 02:30:42 coverage filter: compaction_proactiveness_sysctl_handler: [compaction_proactiveness_sysctl_handler] 2025/08/19 02:30:42 coverage filter: count_memcg_event_mm: [count_memcg_event_mm count_memcg_event_mm] 2025/08/19 02:30:42 coverage filter: cpu_vm_stats_fold: [] 2025/08/19 02:30:42 coverage filter: dax_iomap_fault: [dax_iomap_fault] 2025/08/19 02:30:42 coverage filter: delete_from_page_cache_batch: [delete_from_page_cache_batch] 2025/08/19 02:30:42 coverage filter: do_convert_inline_dir: [do_convert_inline_dir] 2025/08/19 02:30:42 coverage filter: do_huge_pmd_wp_page: [do_huge_pmd_wp_page] 2025/08/19 02:30:42 coverage filter: do_read_cache_folio: [do_read_cache_folio] 2025/08/19 02:30:42 coverage filter: do_swap_page: [do_swap_page] 2025/08/19 02:30:42 coverage filter: do_wp_page: [do_wp_page] 2025/08/19 02:30:42 coverage filter: drain_obj_stock: [] 2025/08/19 02:30:42 coverage filter: ecryptfs_write: [ecryptfs_write ecryptfs_write_begin ecryptfs_write_crypt_stat_flags ecryptfs_write_end ecryptfs_write_header_metadata ecryptfs_write_inode_size_to_metadata ecryptfs_write_lower ecryptfs_write_lower_page_segment ecryptfs_write_metadata ecryptfs_write_packet_length ecryptfs_write_tag_70_packet ecryptfs_writepages] 2025/08/19 02:30:42 coverage filter: ext4_page_mkwrite: [ext4_page_mkwrite] 2025/08/19 02:30:42 coverage filter: ext4_write_begin: [__probestub_ext4_write_begin __traceiter_ext4_write_begin ext4_write_begin] 2025/08/19 02:30:42 coverage filter: f2fs_delete_entry: [f2fs_delete_entry] 2025/08/19 02:30:42 coverage filter: f2fs_delete_inline_entry: [f2fs_delete_inline_entry] 2025/08/19 02:30:42 coverage filter: f2fs_flush_inline_data: [f2fs_flush_inline_data] 2025/08/19 02:30:42 coverage filter: f2fs_fsync_node_pages: [f2fs_fsync_node_pages] 2025/08/19 02:30:42 coverage filter: f2fs_get_lock_data_folio: [f2fs_get_lock_data_folio] 2025/08/19 02:30:42 coverage filter: f2fs_init_inode_metadata: [f2fs_init_inode_metadata] 2025/08/19 02:30:42 coverage filter: f2fs_invalidate_compress_pages: [f2fs_invalidate_compress_pages f2fs_invalidate_compress_pages_range] 2025/08/19 02:30:42 coverage filter: f2fs_quota_read: [f2fs_quota_read] 2025/08/19 02:30:42 coverage filter: f2fs_recover_fsync_data: [f2fs_recover_fsync_data] 2025/08/19 02:30:42 coverage filter: f2fs_set_link: [f2fs_set_link] 2025/08/19 02:30:42 coverage filter: f2fs_sync_meta_pages: [f2fs_sync_meta_pages] 2025/08/19 02:30:42 coverage filter: f2fs_sync_node_pages: [f2fs_sync_node_pages] 2025/08/19 02:30:42 coverage filter: f2fs_truncate_inode_blocks: [__probestub_f2fs_truncate_inode_blocks_enter __probestub_f2fs_truncate_inode_blocks_exit __traceiter_f2fs_truncate_inode_blocks_enter __traceiter_f2fs_truncate_inode_blocks_exit f2fs_truncate_inode_blocks trace_f2fs_truncate_inode_blocks_exit] 2025/08/19 02:30:42 coverage filter: f2fs_vm_page_mkwrite: [__probestub_f2fs_vm_page_mkwrite __traceiter_f2fs_vm_page_mkwrite f2fs_vm_page_mkwrite] 2025/08/19 02:30:42 coverage filter: f2fs_write_begin: [__bpf_trace_f2fs_write_begin __probestub_f2fs_write_begin __traceiter_f2fs_write_begin f2fs_write_begin perf_trace_f2fs_write_begin trace_event_raw_event_f2fs_write_begin trace_raw_output_f2fs_write_begin] 2025/08/19 02:30:42 coverage filter: f2fs_write_data_pages: [f2fs_write_data_pages] 2025/08/19 02:30:42 coverage filter: f2fs_write_multi_pages: [f2fs_write_multi_pages] 2025/08/19 02:30:42 coverage filter: fb_deferred_io_mkwrite: [fb_deferred_io_mkwrite] 2025/08/19 02:30:42 coverage filter: filemap_add_folio: [__filemap_add_folio filemap_add_folio] 2025/08/19 02:30:42 coverage filter: filemap_fault: [__bpf_trace_mm_filemap_fault __probestub_f2fs_filemap_fault __probestub_mm_filemap_fault __traceiter_f2fs_filemap_fault __traceiter_mm_filemap_fault ceph_filemap_fault f2fs_filemap_fault filemap_fault filemap_fault_recheck_pte_none perf_trace_mm_filemap_fault trace_event_raw_event_mm_filemap_fault trace_raw_output_mm_filemap_fault xfs_filemap_fault] 2025/08/19 02:30:42 coverage filter: filemap_get_entry: [filemap_get_entry] 2025/08/19 02:30:42 coverage filter: filemap_get_folios_contig: [filemap_get_folios_contig] 2025/08/19 02:30:42 coverage filter: filemap_get_folios_tag: [filemap_get_folios_tag] 2025/08/19 02:30:42 coverage filter: filemap_get_read_batch: [filemap_get_read_batch] 2025/08/19 02:30:42 coverage filter: filemap_mod_uncharged_vmstat: [filemap_mod_uncharged_vmstat] 2025/08/19 02:30:42 coverage filter: filemap_page_mkwrite: [filemap_page_mkwrite xfs_filemap_page_mkwrite zonefs_filemap_page_mkwrite] 2025/08/19 02:30:42 coverage filter: filemap_read: [filemap_read filemap_read_folio] 2025/08/19 02:30:42 coverage filter: filemap_splice_read: [__probestub_filemap_splice_read_ret __traceiter_filemap_splice_read_ret filemap_splice_read] 2025/08/19 02:30:42 coverage filter: filemap_unaccount_folio: [filemap_unaccount_folio] 2025/08/19 02:30:42 coverage filter: find_get_block_common: [find_get_block_common] 2025/08/19 02:30:42 coverage filter: find_get_entries: [find_get_entries] 2025/08/19 02:30:42 coverage filter: find_lock_entries: [find_lock_entries] 2025/08/19 02:30:42 coverage filter: flush_memcg_stats_dwork: [] 2025/08/19 02:30:42 coverage filter: folio_batch_move_lru: [folio_batch_move_lru] 2025/08/19 02:30:42 coverage filter: folio_deactivate: [folio_deactivate] 2025/08/19 02:30:42 coverage filter: folio_lock: [__folio_lock __folio_lock_killable __folio_lock_or_retry folio_lock folio_lock folio_lock folio_lock folio_lock folio_lock folio_lock folio_lock_anon_vma_read] 2025/08/19 02:30:42 coverage filter: folio_lruvec: [folio_lruvec folio_lruvec_relock_irq] 2025/08/19 02:30:42 coverage filter: folio_lruvec_lock: [] 2025/08/19 02:30:42 coverage filter: folio_lruvec_lock_irq: [] 2025/08/19 02:30:42 coverage filter: folio_lruvec_lock_irqsave: [] 2025/08/19 02:30:42 coverage filter: folio_mark_accessed: [folio_mark_accessed] 2025/08/19 02:30:42 coverage filter: folio_mark_dirty_lock: [folio_mark_dirty_lock] 2025/08/19 02:30:42 coverage filter: folio_mark_lazyfree: [folio_mark_lazyfree] 2025/08/19 02:30:42 coverage filter: folios_put_refs: [folios_put_refs] 2025/08/19 02:30:42 coverage filter: force_metapage: [force_metapage] 2025/08/19 02:30:42 coverage filter: free_percpu: [__bpf_trace_percpu_free_percpu __free_percpu_irq __probestub_percpu_free_percpu __traceiter_percpu_free_percpu devm_free_percpu free_percpu free_percpu_irq free_percpu_nmi metadata_dst_free_percpu perf_trace_percpu_free_percpu trace_event_raw_event_percpu_free_percpu trace_percpu_free_percpu trace_raw_output_percpu_free_percpu] 2025/08/19 02:30:42 coverage filter: fuse_page_mkwrite: [fuse_page_mkwrite] 2025/08/19 02:30:42 coverage filter: generic_perform_write: [generic_perform_write] 2025/08/19 02:30:42 coverage filter: generic_pipe_buf_try_steal: [generic_pipe_buf_try_steal] 2025/08/19 02:30:42 coverage filter: get_mem_cgroup_from_current: [] 2025/08/19 02:30:42 coverage filter: get_mem_cgroup_from_folio: [] 2025/08/19 02:30:42 coverage filter: get_mem_cgroup_from_mm: [] 2025/08/19 02:30:42 coverage filter: gfs2_page_mkwrite: [gfs2_page_mkwrite] 2025/08/19 02:30:42 coverage filter: gfs2_trans_add_meta: [gfs2_trans_add_meta] 2025/08/19 02:30:42 coverage filter: gfs2_write_cache_jdata: [gfs2_write_cache_jdata] 2025/08/19 02:30:42 coverage filter: grab_metapage: [grab_metapage] 2025/08/19 02:30:42 coverage filter: hold_metapage: [hold_metapage] 2025/08/19 02:30:42 coverage filter: invalid_folio_referenced_vma: [invalid_folio_referenced_vma] 2025/08/19 02:30:42 coverage filter: invalidate_inode_pages2_range: [invalidate_inode_pages2_range] 2025/08/19 02:30:42 coverage filter: io_ctl_prepare_pages: [io_ctl_prepare_pages] 2025/08/19 02:30:42 coverage filter: iomap_page_mkwrite: [iomap_page_mkwrite] 2025/08/19 02:30:42 coverage filter: isolate_migratepages_block: [isolate_migratepages_block] 2025/08/19 02:30:42 coverage filter: kcompactd: [__bpf_trace_kcompactd_wake_template __bpf_trace_mm_compaction_kcompactd_sleep __probestub_mm_compaction_kcompactd_sleep __probestub_mm_compaction_kcompactd_wake __probestub_mm_compaction_wakeup_kcompactd __traceiter_mm_compaction_kcompactd_sleep __traceiter_mm_compaction_kcompactd_wake __traceiter_mm_compaction_wakeup_kcompactd kcompactd kcompactd_run kcompactd_stop perf_trace_kcompactd_wake_template perf_trace_mm_compaction_kcompactd_sleep trace_event_raw_event_kcompactd_wake_template trace_event_raw_event_mm_compaction_kcompactd_sleep trace_mm_compaction_wakeup_kcompactd trace_raw_output_kcompactd_wake_template trace_raw_output_mm_compaction_kcompactd_sleep wakeup_kcompactd] 2025/08/19 02:30:42 coverage filter: ksm_do_scan: [ksm_do_scan] 2025/08/19 02:30:42 coverage filter: ksm_get_folio: [ksm_get_folio] 2025/08/19 02:30:42 coverage filter: lock_delalloc_folios: [lock_delalloc_folios] 2025/08/19 02:30:42 coverage filter: lock_metapage: [lock_metapage] 2025/08/19 02:30:42 coverage filter: lock_page: [lock_page lock_page lock_page trylock_page unlock_page] 2025/08/19 02:30:42 coverage filter: lru_activate: [__bpf_trace_mm_lru_activate __probestub_mm_lru_activate __traceiter_mm_lru_activate lru_activate perf_trace_mm_lru_activate trace_event_raw_event_mm_lru_activate trace_raw_output_mm_lru_activate] 2025/08/19 02:30:42 coverage filter: lru_add: [__lru_add_drain_all __probestub_nfsd_file_lru_add __traceiter_nfsd_file_lru_add d_lru_add list_lru_add list_lru_add_obj lru_add lru_add_drain lru_add_drain_all lru_add_drain_cpu lru_add_drain_cpu_zone lru_add_drain_per_cpu lru_add_split_folio] 2025/08/19 02:30:42 coverage filter: lru_deactivate: [lru_deactivate lru_deactivate_file] 2025/08/19 02:30:42 coverage filter: lru_deactivate_file: [] 2025/08/19 02:30:42 coverage filter: lru_gen_clear_refs: [lru_gen_clear_refs] 2025/08/19 02:30:42 coverage filter: lru_lazyfree: [lru_lazyfree] 2025/08/19 02:30:42 coverage filter: lru_move_tail: [drm_gem_lru_move_tail drm_gem_lru_move_tail_locked lru_move_tail] 2025/08/19 02:30:42 coverage filter: lruvec_page_state: [] 2025/08/19 02:30:42 coverage filter: lruvec_page_state_local: [] 2025/08/19 02:30:42 coverage filter: make_device_exclusive: [make_device_exclusive] 2025/08/19 02:30:42 coverage filter: map_anon_folio_pmd: [map_anon_folio_pmd] 2025/08/19 02:30:42 coverage filter: mapping_seek_hole_data: [mapping_seek_hole_data] 2025/08/19 02:30:42 coverage filter: mem_cgroup_charge_skmem: [] 2025/08/19 02:30:42 coverage filter: mem_cgroup_iter: [] 2025/08/19 02:30:42 coverage filter: mem_cgroup_iter_break: [] 2025/08/19 02:30:42 coverage filter: mem_cgroup_migrate: [] 2025/08/19 02:30:42 coverage filter: mem_cgroup_replace_folio: [] 2025/08/19 02:30:42 coverage filter: mem_cgroup_scan_tasks: [] 2025/08/19 02:30:42 coverage filter: mem_cgroup_swap_full: [] 2025/08/19 02:30:42 coverage filter: mem_cgroup_track_foreign_dirty_slowpath: [] 2025/08/19 02:30:42 coverage filter: mem_cgroup_uncharge_skmem: [] 2025/08/19 02:30:42 coverage filter: mem_cgroup_wb_stats: [] 2025/08/19 02:30:42 coverage filter: memcg1_stat_format: [] 2025/08/19 02:30:42 coverage filter: memcg1_swapout: [] 2025/08/19 02:30:42 coverage filter: memcg_page_state: [mod_memcg_page_state] 2025/08/19 02:30:42 coverage filter: memcg_page_state_local: [] 2025/08/19 02:30:42 coverage filter: memcg_page_state_local_output: [] 2025/08/19 02:30:42 coverage filter: memcg_page_state_output: [] 2025/08/19 02:30:42 coverage filter: memory_numa_stat_show: [] 2025/08/19 02:30:42 coverage filter: memory_stat_format: [] 2025/08/19 02:30:42 coverage filter: memtier_hotplug_callback: [memtier_hotplug_callback] 2025/08/19 02:30:42 coverage filter: migrate_device_coherent_folio: [migrate_device_coherent_folio] 2025/08/19 02:30:42 coverage filter: migrate_pages: [__bpf_trace_mm_migrate_pages __bpf_trace_mm_migrate_pages_start __ia32_sys_migrate_pages __probestub_mm_migrate_pages __probestub_mm_migrate_pages_start __se_sys_migrate_pages __traceiter_mm_migrate_pages __traceiter_mm_migrate_pages_start __x64_sys_migrate_pages damon_migrate_pages do_migrate_pages migrate_pages migrate_pages_batch perf_trace_mm_migrate_pages perf_trace_mm_migrate_pages_start trace_event_raw_event_mm_migrate_pages trace_event_raw_event_mm_migrate_pages_start trace_raw_output_mm_migrate_pages trace_raw_output_mm_migrate_pages_start] 2025/08/19 02:30:42 coverage filter: migrate_pages_batch: [] 2025/08/19 02:30:42 coverage filter: minix_add_link: [minix_add_link] 2025/08/19 02:30:42 coverage filter: minix_delete_entry: [minix_delete_entry] 2025/08/19 02:30:42 coverage filter: minix_set_link: [minix_set_link] 2025/08/19 02:30:42 coverage filter: mod_memcg_lruvec_state: [] 2025/08/19 02:30:42 coverage filter: mod_memcg_page_state: [] 2025/08/19 02:30:42 coverage filter: mod_memcg_state: [] 2025/08/19 02:30:42 coverage filter: mod_node_state: [] 2025/08/19 02:30:42 coverage filter: mod_objcg_mlstate: [] 2025/08/19 02:30:42 coverage filter: move_pages_huge_pmd: [move_pages_huge_pmd] 2025/08/19 02:30:42 coverage filter: move_pages_pte: [move_pages_pte] 2025/08/19 02:30:42 coverage filter: mpage_prepare_extent_to_map: [mpage_prepare_extent_to_map] 2025/08/19 02:30:42 coverage filter: need_update: [] 2025/08/19 02:30:42 coverage filter: netfs_page_mkwrite: [netfs_page_mkwrite] 2025/08/19 02:30:42 coverage filter: next_uptodate_folio: [next_uptodate_folio] 2025/08/19 02:30:42 coverage filter: nfs_vm_page_mkwrite: [nfs_vm_page_mkwrite] 2025/08/19 02:30:42 coverage filter: nilfs_add_link: [nilfs_add_link] 2025/08/19 02:30:42 coverage filter: nilfs_begin_folio_io: [nilfs_begin_folio_io] 2025/08/19 02:30:42 coverage filter: nilfs_btnode_delete: [nilfs_btnode_delete] 2025/08/19 02:30:42 coverage filter: nilfs_btnode_prepare_change_key: [nilfs_btnode_prepare_change_key] 2025/08/19 02:30:42 coverage filter: nilfs_clear_dirty_pages: [nilfs_clear_dirty_pages] 2025/08/19 02:30:42 coverage filter: nilfs_copy_back_pages: [nilfs_copy_back_pages] 2025/08/19 02:30:42 coverage filter: nilfs_copy_dirty_pages: [nilfs_copy_dirty_pages] 2025/08/19 02:30:42 coverage filter: nilfs_delete_entry: [nilfs_delete_entry] 2025/08/19 02:30:42 coverage filter: nilfs_end_folio_io: [nilfs_end_folio_io] 2025/08/19 02:30:42 coverage filter: nilfs_find_uncommitted_extent: [nilfs_find_uncommitted_extent] 2025/08/19 02:30:42 coverage filter: nilfs_lookup_dirty_data_buffers: [nilfs_lookup_dirty_data_buffers] 2025/08/19 02:30:42 coverage filter: nilfs_page_mkwrite: [nilfs_page_mkwrite] 2025/08/19 02:30:42 coverage filter: nilfs_segctor_do_construct: [nilfs_segctor_do_construct] 2025/08/19 02:30:42 coverage filter: nilfs_set_link: [nilfs_set_link] 2025/08/19 02:30:42 coverage filter: node_get_allowed_targets: [node_get_allowed_targets] 2025/08/19 02:30:42 coverage filter: node_is_toptier: [node_is_toptier] 2025/08/19 02:30:42 coverage filter: node_page_state: [] 2025/08/19 02:30:42 coverage filter: node_page_state_add: [] 2025/08/19 02:30:42 coverage filter: node_page_state_pages: [] 2025/08/19 02:30:42 coverage filter: node_read_vmstat: [node_read_vmstat] 2025/08/19 02:30:42 coverage filter: ntfs_bio_fill_1: [ntfs_bio_fill_1] 2025/08/19 02:30:42 coverage filter: obj_cgroup_charge_pages: [] 2025/08/19 02:30:42 coverage filter: obj_cgroup_charge_zswap: [] 2025/08/19 02:30:42 coverage filter: obj_cgroup_may_zswap: [] 2025/08/19 02:30:42 coverage filter: obj_cgroup_release: [] 2025/08/19 02:30:42 coverage filter: obj_cgroup_uncharge_zswap: [] 2025/08/19 02:30:42 coverage filter: ocfs2_duplicate_clusters_by_page: [__probestub_ocfs2_duplicate_clusters_by_page __traceiter_ocfs2_duplicate_clusters_by_page ocfs2_duplicate_clusters_by_page trace_ocfs2_duplicate_clusters_by_page] 2025/08/19 02:30:42 coverage filter: ocfs2_grab_folios_for_write: [ocfs2_grab_folios_for_write] 2025/08/19 02:30:42 coverage filter: orangefs_page_mkwrite: [orangefs_page_mkwrite] 2025/08/19 02:30:42 coverage filter: page_cache_pipe_buf_confirm: [page_cache_pipe_buf_confirm] 2025/08/19 02:30:42 coverage filter: page_cache_pipe_buf_try_steal: [page_cache_pipe_buf_try_steal] 2025/08/19 02:30:42 coverage filter: pagetypeinfo_showfree_print: [] 2025/08/19 02:30:42 coverage filter: pcpu_memcg_post_alloc_hook: [pcpu_memcg_post_alloc_hook] 2025/08/19 02:30:42 coverage filter: perf_trace_mm_lru_insertion: [perf_trace_mm_lru_insertion] 2025/08/19 02:30:42 coverage filter: prepare_one_folio: [prepare_one_folio] 2025/08/19 02:30:42 coverage filter: putback_movable_ops_page: [putback_movable_ops_page] 2025/08/19 02:30:42 coverage filter: refill_obj_stock: [] 2025/08/19 02:30:42 coverage filter: refresh_cpu_vm_stats: [] 2025/08/19 02:30:42 coverage filter: release_metapage: [release_metapage] 2025/08/19 02:30:42 coverage filter: relocate_file_extent_cluster: [relocate_file_extent_cluster] 2025/08/19 02:30:42 coverage filter: remove_device_exclusive_entry: [remove_device_exclusive_entry] 2025/08/19 02:30:42 coverage filter: remove_inode_hugepages: [remove_inode_hugepages] 2025/08/19 02:30:42 coverage filter: send_extent_data: [send_extent_data] 2025/08/19 02:30:42 coverage filter: set_extent_buffer_dirty: [set_extent_buffer_dirty] 2025/08/19 02:30:42 coverage filter: set_node_memory_tier: [set_node_memory_tier] 2025/08/19 02:30:42 coverage filter: shmem_get_folio_gfp: [shmem_get_folio_gfp] 2025/08/19 02:30:42 coverage filter: shmem_get_partial_folio: [shmem_get_partial_folio] 2025/08/19 02:30:42 coverage filter: shmem_swapin_folio: [shmem_swapin_folio] 2025/08/19 02:30:42 coverage filter: shmem_undo_range: [shmem_undo_range] 2025/08/19 02:30:42 coverage filter: shrink_folio_list: [shrink_folio_list] 2025/08/19 02:30:42 coverage filter: split_huge_pages_all: [split_huge_pages_all] 2025/08/19 02:30:42 coverage filter: trace_event_raw_event_mm_lru_insertion: [trace_event_raw_event_mm_lru_insertion] 2025/08/19 02:30:42 coverage filter: truncate_inode_pages_range: [truncate_inode_pages_range] 2025/08/19 02:30:42 coverage filter: try_split_folio: [try_split_folio] 2025/08/19 02:30:42 coverage filter: try_to_compact_pages: [__bpf_trace_mm_compaction_try_to_compact_pages __probestub_mm_compaction_try_to_compact_pages __traceiter_mm_compaction_try_to_compact_pages perf_trace_mm_compaction_try_to_compact_pages trace_event_raw_event_mm_compaction_try_to_compact_pages trace_raw_output_mm_compaction_try_to_compact_pages try_to_compact_pages] 2025/08/19 02:30:42 coverage filter: try_to_unuse: [try_to_unuse] 2025/08/19 02:30:42 coverage filter: ttm_backup_backup_page: [ttm_backup_backup_page] 2025/08/19 02:30:42 coverage filter: txLock: [txLock txLockAlloc] 2025/08/19 02:30:42 coverage filter: ubifs_vm_page_mkwrite: [ubifs_vm_page_mkwrite] 2025/08/19 02:30:42 coverage filter: udf_expand_file_adinicb: [udf_expand_file_adinicb] 2025/08/19 02:30:42 coverage filter: udf_page_mkwrite: [udf_page_mkwrite] 2025/08/19 02:30:42 coverage filter: ufs_add_link: [ufs_add_link] 2025/08/19 02:30:42 coverage filter: ufs_delete_entry: [ufs_delete_entry] 2025/08/19 02:30:42 coverage filter: ufs_get_locked_folio: [ufs_get_locked_folio] 2025/08/19 02:30:42 coverage filter: ufs_set_link: [ufs_set_link] 2025/08/19 02:30:42 coverage filter: uncharge_batch: [] 2025/08/19 02:30:42 coverage filter: unpin_user_page_range_dirty_lock: [unpin_user_page_range_dirty_lock] 2025/08/19 02:30:42 coverage filter: unpin_user_pages_dirty_lock: [unpin_user_pages_dirty_lock] 2025/08/19 02:30:42 coverage filter: vfs_dedupe_file_range_compare: [vfs_dedupe_file_range_compare] 2025/08/19 02:30:42 coverage filter: vmstat_next: [] 2025/08/19 02:30:42 coverage filter: vmstat_refresh: [] 2025/08/19 02:30:42 coverage filter: vmstat_shepherd: [] 2025/08/19 02:30:42 coverage filter: vmstat_show: [] 2025/08/19 02:30:42 coverage filter: vmstat_start: [] 2025/08/19 02:30:42 coverage filter: workingset_activation: [workingset_activation] 2025/08/19 02:30:42 coverage filter: workingset_eviction: [workingset_eviction] 2025/08/19 02:30:42 coverage filter: workingset_refault: [workingset_refault] 2025/08/19 02:30:42 coverage filter: write_all_supers: [write_all_supers] 2025/08/19 02:30:42 coverage filter: write_one_eb: [write_one_eb] 2025/08/19 02:30:42 coverage filter: writeback_iter: [writeback_iter] 2025/08/19 02:30:42 coverage filter: z_erofs_runqueue: [z_erofs_runqueue] 2025/08/19 02:30:42 coverage filter: zone_device_page_init: [zone_device_page_init] 2025/08/19 02:30:42 coverage filter: zoneinfo_show_print: [] 2025/08/19 02:30:42 coverage filter: zswap_current_read: [] 2025/08/19 02:30:42 coverage filter: zswap_folio_swapin: [zswap_folio_swapin] 2025/08/19 02:30:42 coverage filter: zswap_shrinker_count: [zswap_shrinker_count] 2025/08/19 02:30:42 coverage filter: fs/btrfs/disk-io.c: [fs/btrfs/disk-io.c] 2025/08/19 02:30:42 coverage filter: include/linux/memcontrol.h: [] 2025/08/19 02:30:42 coverage filter: include/linux/mmzone.h: [] 2025/08/19 02:30:42 coverage filter: include/linux/pagemap.h: [] 2025/08/19 02:30:42 coverage filter: mm/filemap.c: [mm/filemap.c] 2025/08/19 02:30:42 coverage filter: mm/vmstat.c: [] 2025/08/19 02:30:42 area "symbols": 19253 PCs in the cover filter 2025/08/19 02:30:42 area "files": 5769 PCs in the cover filter 2025/08/19 02:30:42 area "": 0 PCs in the cover filter 2025/08/19 02:30:42 executor cover filter: 0 PCs 2025/08/19 02:30:46 machine check: disabled the following syscalls: fsetxattr$security_selinux : selinux is not enabled fsetxattr$security_smack_transmute : smack is not enabled fsetxattr$smack_xattr_label : smack is not enabled get_thread_area : syscall get_thread_area is not present lookup_dcookie : syscall lookup_dcookie is not present lsetxattr$security_selinux : selinux is not enabled lsetxattr$security_smack_transmute : smack is not enabled lsetxattr$smack_xattr_label : smack is not enabled mount$esdfs : /proc/filesystems does not contain esdfs mount$incfs : /proc/filesystems does not contain incremental-fs openat$acpi_thermal_rel : failed to open /dev/acpi_thermal_rel: no such file or directory openat$ashmem : failed to open /dev/ashmem: no such file or directory openat$bifrost : failed to open /dev/bifrost: no such file or directory openat$binder : failed to open /dev/binder: no such file or directory openat$camx : failed to open /dev/v4l/by-path/platform-soc@0:qcom_cam-req-mgr-video-index0: no such file or directory openat$capi20 : failed to open /dev/capi20: no such file or directory openat$cdrom1 : failed to open /dev/cdrom1: no such file or directory openat$damon_attrs : failed to open /sys/kernel/debug/damon/attrs: no such file or directory openat$damon_init_regions : failed to open /sys/kernel/debug/damon/init_regions: no such file or directory openat$damon_kdamond_pid : failed to open /sys/kernel/debug/damon/kdamond_pid: no such file or directory openat$damon_mk_contexts : failed to open /sys/kernel/debug/damon/mk_contexts: no such file or directory openat$damon_monitor_on : failed to open /sys/kernel/debug/damon/monitor_on: no such file or directory openat$damon_rm_contexts : failed to open /sys/kernel/debug/damon/rm_contexts: no such file or directory openat$damon_schemes : failed to open /sys/kernel/debug/damon/schemes: no such file or directory openat$damon_target_ids : failed to open /sys/kernel/debug/damon/target_ids: no such file or directory openat$hwbinder : failed to open /dev/hwbinder: no such file or directory openat$i915 : failed to open /dev/i915: no such file or directory openat$img_rogue : failed to open /dev/img-rogue: no such file or directory openat$irnet : failed to open /dev/irnet: no such file or directory openat$keychord : failed to open /dev/keychord: no such file or directory openat$kvm : failed to open /dev/kvm: no such file or directory openat$lightnvm : failed to open /dev/lightnvm/control: no such file or directory openat$mali : failed to open /dev/mali0: no such file or directory openat$md : failed to open /dev/md0: no such file or directory openat$msm : failed to open /dev/msm: no such file or directory openat$ndctl0 : failed to open /dev/ndctl0: no such file or directory openat$nmem0 : failed to open /dev/nmem0: no such file or directory openat$pktcdvd : failed to open /dev/pktcdvd/control: no such file or directory openat$pmem0 : failed to open /dev/pmem0: no such file or directory openat$proc_capi20 : failed to open /proc/capi/capi20: no such file or directory openat$proc_capi20ncci : failed to open /proc/capi/capi20ncci: no such file or directory openat$proc_reclaim : failed to open /proc/self/reclaim: no such file or directory openat$ptp1 : failed to open /dev/ptp1: no such file or directory openat$rnullb : failed to open /dev/rnullb0: no such file or directory openat$selinux_access : failed to open /selinux/access: no such file or directory openat$selinux_attr : selinux is not enabled openat$selinux_avc_cache_stats : failed to open /selinux/avc/cache_stats: no such file or directory openat$selinux_avc_cache_threshold : failed to open /selinux/avc/cache_threshold: no such file or directory openat$selinux_avc_hash_stats : failed to open /selinux/avc/hash_stats: no such file or directory openat$selinux_checkreqprot : failed to open /selinux/checkreqprot: no such file or directory openat$selinux_commit_pending_bools : failed to open /selinux/commit_pending_bools: no such file or directory openat$selinux_context : failed to open /selinux/context: no such file or directory openat$selinux_create : failed to open /selinux/create: no such file or directory openat$selinux_enforce : failed to open /selinux/enforce: no such file or directory openat$selinux_load : failed to open /selinux/load: no such file or directory openat$selinux_member : failed to open /selinux/member: no such file or directory openat$selinux_mls : failed to open /selinux/mls: no such file or directory openat$selinux_policy : failed to open /selinux/policy: no such file or directory openat$selinux_relabel : failed to open /selinux/relabel: no such file or directory openat$selinux_status : failed to open /selinux/status: no such file or directory openat$selinux_user : failed to open /selinux/user: no such file or directory openat$selinux_validatetrans : failed to open /selinux/validatetrans: no such file or directory openat$sev : failed to open /dev/sev: no such file or directory openat$sgx_provision : failed to open /dev/sgx_provision: no such file or directory openat$smack_task_current : smack is not enabled openat$smack_thread_current : smack is not enabled openat$smackfs_access : failed to open /sys/fs/smackfs/access: no such file or directory openat$smackfs_ambient : failed to open /sys/fs/smackfs/ambient: no such file or directory openat$smackfs_change_rule : failed to open /sys/fs/smackfs/change-rule: no such file or directory openat$smackfs_cipso : failed to open /sys/fs/smackfs/cipso: no such file or directory openat$smackfs_cipsonum : failed to open /sys/fs/smackfs/direct: no such file or directory openat$smackfs_ipv6host : failed to open /sys/fs/smackfs/ipv6host: no such file or directory openat$smackfs_load : failed to open /sys/fs/smackfs/load: no such file or directory openat$smackfs_logging : failed to open /sys/fs/smackfs/logging: no such file or directory openat$smackfs_netlabel : failed to open /sys/fs/smackfs/netlabel: no such file or directory openat$smackfs_onlycap : failed to open /sys/fs/smackfs/onlycap: no such file or directory openat$smackfs_ptrace : failed to open /sys/fs/smackfs/ptrace: no such file or directory openat$smackfs_relabel_self : failed to open /sys/fs/smackfs/relabel-self: no such file or directory openat$smackfs_revoke_subject : failed to open /sys/fs/smackfs/revoke-subject: no such file or directory openat$smackfs_syslog : failed to open /sys/fs/smackfs/syslog: no such file or directory openat$smackfs_unconfined : failed to open /sys/fs/smackfs/unconfined: no such file or directory openat$tlk_device : failed to open /dev/tlk_device: no such file or directory openat$trusty : failed to open /dev/trusty-ipc-dev0: no such file or directory openat$trusty_avb : failed to open /dev/trusty-ipc-dev0: no such file or directory openat$trusty_gatekeeper : failed to open /dev/trusty-ipc-dev0: no such file or directory openat$trusty_hwkey : failed to open /dev/trusty-ipc-dev0: no such file or directory openat$trusty_hwrng : failed to open /dev/trusty-ipc-dev0: no such file or directory openat$trusty_km : failed to open /dev/trusty-ipc-dev0: no such file or directory openat$trusty_km_secure : failed to open /dev/trusty-ipc-dev0: no such file or directory openat$trusty_storage : failed to open /dev/trusty-ipc-dev0: no such file or directory openat$tty : failed to open /dev/tty: no such device or address openat$uverbs0 : failed to open /dev/infiniband/uverbs0: no such file or directory openat$vfio : failed to open /dev/vfio/vfio: no such file or directory openat$vndbinder : failed to open /dev/vndbinder: no such file or directory openat$vtpm : failed to open /dev/vtpmx: no such file or directory openat$xenevtchn : failed to open /dev/xen/evtchn: no such file or directory openat$zygote : failed to open /dev/socket/zygote: no such file or directory pkey_alloc : pkey_alloc(0x0, 0x0) failed: no space left on device read$smackfs_access : smack is not enabled read$smackfs_cipsonum : smack is not enabled read$smackfs_logging : smack is not enabled read$smackfs_ptrace : smack is not enabled set_thread_area : syscall set_thread_area is not present setxattr$security_selinux : selinux is not enabled setxattr$security_smack_transmute : smack is not enabled setxattr$smack_xattr_label : smack is not enabled socket$hf : socket$hf(0x13, 0x2, 0x0) failed: address family not supported by protocol socket$inet6_dccp : socket$inet6_dccp(0xa, 0x6, 0x0) failed: socket type not supported socket$inet_dccp : socket$inet_dccp(0x2, 0x6, 0x0) failed: socket type not supported socket$vsock_dgram : socket$vsock_dgram(0x28, 0x2, 0x0) failed: no such device syz_btf_id_by_name$bpf_lsm : failed to open /sys/kernel/btf/vmlinux: no such file or directory syz_init_net_socket$bt_cmtp : syz_init_net_socket$bt_cmtp(0x1f, 0x3, 0x5) failed: protocol not supported syz_kvm_setup_cpu$ppc64 : unsupported arch syz_mount_image$ntfs : /proc/filesystems does not contain ntfs syz_mount_image$reiserfs : /proc/filesystems does not contain reiserfs syz_mount_image$sysv : /proc/filesystems does not contain sysv syz_mount_image$v7 : /proc/filesystems does not contain v7 syz_open_dev$dricontrol : failed to open /dev/dri/controlD#: no such file or directory syz_open_dev$drirender : failed to open /dev/dri/renderD#: no such file or directory syz_open_dev$floppy : failed to open /dev/fd#: no such file or directory syz_open_dev$ircomm : failed to open /dev/ircomm#: no such file or directory syz_open_dev$sndhw : failed to open /dev/snd/hwC#D#: no such file or directory syz_pkey_set : pkey_alloc(0x0, 0x0) failed: no space left on device uselib : syscall uselib is not present write$selinux_access : selinux is not enabled write$selinux_attr : selinux is not enabled write$selinux_context : selinux is not enabled write$selinux_create : selinux is not enabled write$selinux_load : selinux is not enabled write$selinux_user : selinux is not enabled write$selinux_validatetrans : selinux is not enabled write$smack_current : smack is not enabled write$smackfs_access : smack is not enabled write$smackfs_change_rule : smack is not enabled write$smackfs_cipso : smack is not enabled write$smackfs_cipsonum : smack is not enabled write$smackfs_ipv6host : smack is not enabled write$smackfs_label : smack is not enabled write$smackfs_labels_list : smack is not enabled write$smackfs_load : smack is not enabled write$smackfs_logging : smack is not enabled write$smackfs_netlabel : smack is not enabled write$smackfs_ptrace : smack is not enabled transitively disabled the following syscalls (missing resource [creating syscalls]): bind$vsock_dgram : sock_vsock_dgram [socket$vsock_dgram] close$ibv_device : fd_rdma [openat$uverbs0] connect$hf : sock_hf [socket$hf] connect$vsock_dgram : sock_vsock_dgram [socket$vsock_dgram] getsockopt$inet6_dccp_buf : sock_dccp6 [socket$inet6_dccp] getsockopt$inet6_dccp_int : sock_dccp6 [socket$inet6_dccp] getsockopt$inet_dccp_buf : sock_dccp [socket$inet_dccp] getsockopt$inet_dccp_int : sock_dccp [socket$inet_dccp] ioctl$ACPI_THERMAL_GET_ART : fd_acpi_thermal_rel [openat$acpi_thermal_rel] ioctl$ACPI_THERMAL_GET_ART_COUNT : fd_acpi_thermal_rel [openat$acpi_thermal_rel] ioctl$ACPI_THERMAL_GET_ART_LEN : fd_acpi_thermal_rel [openat$acpi_thermal_rel] ioctl$ACPI_THERMAL_GET_TRT : fd_acpi_thermal_rel [openat$acpi_thermal_rel] ioctl$ACPI_THERMAL_GET_TRT_COUNT : fd_acpi_thermal_rel [openat$acpi_thermal_rel] ioctl$ACPI_THERMAL_GET_TRT_LEN : fd_acpi_thermal_rel [openat$acpi_thermal_rel] ioctl$ASHMEM_GET_NAME : fd_ashmem [openat$ashmem] ioctl$ASHMEM_GET_PIN_STATUS : fd_ashmem [openat$ashmem] ioctl$ASHMEM_GET_PROT_MASK : fd_ashmem [openat$ashmem] ioctl$ASHMEM_GET_SIZE : fd_ashmem [openat$ashmem] ioctl$ASHMEM_PURGE_ALL_CACHES : fd_ashmem [openat$ashmem] ioctl$ASHMEM_SET_NAME : fd_ashmem [openat$ashmem] ioctl$ASHMEM_SET_PROT_MASK : fd_ashmem [openat$ashmem] ioctl$ASHMEM_SET_SIZE : fd_ashmem [openat$ashmem] ioctl$CAPI_CLR_FLAGS : fd_capi20 [openat$capi20] ioctl$CAPI_GET_ERRCODE : fd_capi20 [openat$capi20] ioctl$CAPI_GET_FLAGS : fd_capi20 [openat$capi20] ioctl$CAPI_GET_MANUFACTURER : fd_capi20 [openat$capi20] ioctl$CAPI_GET_PROFILE : fd_capi20 [openat$capi20] ioctl$CAPI_GET_SERIAL : fd_capi20 [openat$capi20] ioctl$CAPI_INSTALLED : fd_capi20 [openat$capi20] ioctl$CAPI_MANUFACTURER_CMD : fd_capi20 [openat$capi20] ioctl$CAPI_NCCI_GETUNIT : fd_capi20 [openat$capi20] ioctl$CAPI_NCCI_OPENCOUNT : fd_capi20 [openat$capi20] ioctl$CAPI_REGISTER : fd_capi20 [openat$capi20] ioctl$CAPI_SET_FLAGS : fd_capi20 [openat$capi20] ioctl$CREATE_COUNTERS : fd_rdma [openat$uverbs0] ioctl$DESTROY_COUNTERS : fd_rdma [openat$uverbs0] ioctl$DRM_IOCTL_I915_GEM_BUSY : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_CONTEXT_CREATE : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_CONTEXT_DESTROY : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_CONTEXT_GETPARAM : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_CONTEXT_SETPARAM : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_CREATE : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_EXECBUFFER : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_EXECBUFFER2 : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_EXECBUFFER2_WR : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_GET_APERTURE : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_GET_CACHING : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_GET_TILING : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_MADVISE : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_MMAP : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_MMAP_GTT : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_MMAP_OFFSET : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_PIN : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_PREAD : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_PWRITE : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_SET_CACHING : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_SET_DOMAIN : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_SET_TILING : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_SW_FINISH : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_THROTTLE : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_UNPIN : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_USERPTR : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_VM_CREATE : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_VM_DESTROY : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GEM_WAIT : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GETPARAM : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GET_PIPE_FROM_CRTC_ID : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_GET_RESET_STATS : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_OVERLAY_ATTRS : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_OVERLAY_PUT_IMAGE : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_PERF_ADD_CONFIG : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_PERF_OPEN : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_PERF_REMOVE_CONFIG : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_QUERY : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_REG_READ : fd_i915 [openat$i915] ioctl$DRM_IOCTL_I915_SET_SPRITE_COLORKEY : fd_i915 [openat$i915] ioctl$DRM_IOCTL_MSM_GEM_CPU_FINI : fd_msm [openat$msm] ioctl$DRM_IOCTL_MSM_GEM_CPU_PREP : fd_msm [openat$msm] ioctl$DRM_IOCTL_MSM_GEM_INFO : fd_msm [openat$msm] ioctl$DRM_IOCTL_MSM_GEM_MADVISE : fd_msm [openat$msm] ioctl$DRM_IOCTL_MSM_GEM_NEW : fd_msm [openat$msm] ioctl$DRM_IOCTL_MSM_GEM_SUBMIT : fd_msm [openat$msm] ioctl$DRM_IOCTL_MSM_GET_PARAM : fd_msm [openat$msm] ioctl$DRM_IOCTL_MSM_SET_PARAM : fd_msm [openat$msm] ioctl$DRM_IOCTL_MSM_SUBMITQUEUE_CLOSE : fd_msm [openat$msm] ioctl$DRM_IOCTL_MSM_SUBMITQUEUE_NEW : fd_msm [openat$msm] ioctl$DRM_IOCTL_MSM_SUBMITQUEUE_QUERY : fd_msm [openat$msm] ioctl$DRM_IOCTL_MSM_WAIT_FENCE : fd_msm [openat$msm] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_CACHE_CACHEOPEXEC: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_CACHE_CACHEOPLOG: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_CACHE_CACHEOPQUEUE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_CMM_DEVMEMINTACQUIREREMOTECTX: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_CMM_DEVMEMINTEXPORTCTX: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_CMM_DEVMEMINTUNEXPORTCTX: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_DEVICEMEMHISTORY_DEVICEMEMHISTORYMAP: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_DEVICEMEMHISTORY_DEVICEMEMHISTORYMAPVRANGE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_DEVICEMEMHISTORY_DEVICEMEMHISTORYSPARSECHANGE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_DEVICEMEMHISTORY_DEVICEMEMHISTORYUNMAP: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_DEVICEMEMHISTORY_DEVICEMEMHISTORYUNMAPVRANGE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_DMABUF_PHYSMEMEXPORTDMABUF: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_DMABUF_PHYSMEMIMPORTDMABUF: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_DMABUF_PHYSMEMIMPORTSPARSEDMABUF: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_HTBUFFER_HTBCONTROL: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_HTBUFFER_HTBLOG: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_CHANGESPARSEMEM: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMFLUSHDEVSLCRANGE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMGETFAULTADDRESS: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMINTCTXCREATE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMINTCTXDESTROY: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMINTHEAPCREATE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMINTHEAPDESTROY: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMINTMAPPAGES: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMINTMAPPMR: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMINTPIN: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMINTPINVALIDATE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMINTREGISTERPFNOTIFYKM: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMINTRESERVERANGE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMINTUNMAPPAGES: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMINTUNMAPPMR: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMINTUNPIN: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMINTUNPININVALIDATE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMINTUNRESERVERANGE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMINVALIDATEFBSCTABLE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_DEVMEMISVDEVADDRVALID: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_GETMAXDEVMEMSIZE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_HEAPCFGHEAPCONFIGCOUNT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_HEAPCFGHEAPCONFIGNAME: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_HEAPCFGHEAPCOUNT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_HEAPCFGHEAPDETAILS: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_PHYSMEMNEWRAMBACKEDLOCKEDPMR: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_PHYSMEMNEWRAMBACKEDPMR: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_PMREXPORTPMR: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_PMRGETUID: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_PMRIMPORTPMR: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_PMRLOCALIMPORTPMR: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_PMRMAKELOCALIMPORTHANDLE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_PMRUNEXPORTPMR: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_PMRUNMAKELOCALIMPORTHANDLE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_PMRUNREFPMR: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_PMRUNREFUNLOCKPMR: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_MM_PVRSRVUPDATEOOMSTATS: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_PVRTL_TLACQUIREDATA: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_PVRTL_TLCLOSESTREAM: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_PVRTL_TLCOMMITSTREAM: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_PVRTL_TLDISCOVERSTREAMS: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_PVRTL_TLOPENSTREAM: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_PVRTL_TLRELEASEDATA: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_PVRTL_TLRESERVESTREAM: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_PVRTL_TLWRITEDATA: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXBREAKPOINT_RGXCLEARBREAKPOINT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXBREAKPOINT_RGXDISABLEBREAKPOINT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXBREAKPOINT_RGXENABLEBREAKPOINT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXBREAKPOINT_RGXOVERALLOCATEBPREGISTERS: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXBREAKPOINT_RGXSETBREAKPOINT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXCMP_RGXCREATECOMPUTECONTEXT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXCMP_RGXDESTROYCOMPUTECONTEXT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXCMP_RGXFLUSHCOMPUTEDATA: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXCMP_RGXGETLASTCOMPUTECONTEXTRESETREASON: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXCMP_RGXKICKCDM2: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXCMP_RGXNOTIFYCOMPUTEWRITEOFFSETUPDATE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXCMP_RGXSETCOMPUTECONTEXTPRIORITY: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXCMP_RGXSETCOMPUTECONTEXTPROPERTY: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXFWDBG_RGXCURRENTTIME: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXFWDBG_RGXFWDEBUGDUMPFREELISTPAGELIST: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXFWDBG_RGXFWDEBUGPHRCONFIGURE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXFWDBG_RGXFWDEBUGSETFWLOG: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXFWDBG_RGXFWDEBUGSETHCSDEADLINE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXFWDBG_RGXFWDEBUGSETOSIDPRIORITY: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXFWDBG_RGXFWDEBUGSETOSNEWONLINESTATE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXHWPERF_RGXCONFIGCUSTOMCOUNTERS: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXHWPERF_RGXCONFIGENABLEHWPERFCOUNTERS: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXHWPERF_RGXCTRLHWPERF: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXHWPERF_RGXCTRLHWPERFCOUNTERS: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXHWPERF_RGXGETHWPERFBVNCFEATUREFLAGS: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXKICKSYNC_RGXCREATEKICKSYNCCONTEXT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXKICKSYNC_RGXDESTROYKICKSYNCCONTEXT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXKICKSYNC_RGXKICKSYNC2: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXKICKSYNC_RGXSETKICKSYNCCONTEXTPROPERTY: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXREGCONFIG_RGXADDREGCONFIG: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXREGCONFIG_RGXCLEARREGCONFIG: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXREGCONFIG_RGXDISABLEREGCONFIG: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXREGCONFIG_RGXENABLEREGCONFIG: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXREGCONFIG_RGXSETREGCONFIGTYPE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXSIGNALS_RGXNOTIFYSIGNALUPDATE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTA3D_RGXCREATEFREELIST: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTA3D_RGXCREATEHWRTDATASET: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTA3D_RGXCREATERENDERCONTEXT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTA3D_RGXCREATEZSBUFFER: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTA3D_RGXDESTROYFREELIST: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTA3D_RGXDESTROYHWRTDATASET: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTA3D_RGXDESTROYRENDERCONTEXT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTA3D_RGXDESTROYZSBUFFER: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTA3D_RGXGETLASTRENDERCONTEXTRESETREASON: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTA3D_RGXKICKTA3D2: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTA3D_RGXPOPULATEZSBUFFER: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTA3D_RGXRENDERCONTEXTSTALLED: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTA3D_RGXSETRENDERCONTEXTPRIORITY: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTA3D_RGXSETRENDERCONTEXTPROPERTY: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTA3D_RGXUNPOPULATEZSBUFFER: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTQ2_RGXTDMCREATETRANSFERCONTEXT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTQ2_RGXTDMDESTROYTRANSFERCONTEXT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTQ2_RGXTDMGETSHAREDMEMORY: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTQ2_RGXTDMNOTIFYWRITEOFFSETUPDATE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTQ2_RGXTDMRELEASESHAREDMEMORY: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTQ2_RGXTDMSETTRANSFERCONTEXTPRIORITY: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTQ2_RGXTDMSETTRANSFERCONTEXTPROPERTY: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTQ2_RGXTDMSUBMITTRANSFER2: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTQ_RGXCREATETRANSFERCONTEXT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTQ_RGXDESTROYTRANSFERCONTEXT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTQ_RGXSETTRANSFERCONTEXTPRIORITY: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTQ_RGXSETTRANSFERCONTEXTPROPERTY: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_RGXTQ_RGXSUBMITTRANSFER2: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_ACQUIREGLOBALEVENTOBJECT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_ACQUIREINFOPAGE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_ALIGNMENTCHECK: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_CONNECT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_DISCONNECT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_DUMPDEBUGINFO: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_EVENTOBJECTCLOSE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_EVENTOBJECTOPEN: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_EVENTOBJECTWAIT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_EVENTOBJECTWAITTIMEOUT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_FINDPROCESSMEMSTATS: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_GETDEVCLOCKSPEED: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_GETDEVICESTATUS: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_GETMULTICOREINFO: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_HWOPTIMEOUT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_RELEASEGLOBALEVENTOBJECT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SRVCORE_RELEASEINFOPAGE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SYNCTRACKING_SYNCRECORDADD: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SYNCTRACKING_SYNCRECORDREMOVEBYHANDLE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SYNC_ALLOCSYNCPRIMITIVEBLOCK: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SYNC_FREESYNCPRIMITIVEBLOCK: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SYNC_SYNCALLOCEVENT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SYNC_SYNCCHECKPOINTSIGNALLEDPDUMPPOL: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SYNC_SYNCFREEEVENT: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SYNC_SYNCPRIMPDUMP: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SYNC_SYNCPRIMPDUMPCBP: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SYNC_SYNCPRIMPDUMPPOL: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SYNC_SYNCPRIMPDUMPVALUE: fd_rogue [openat$img_rogue] ioctl$DRM_IOCTL_PVR_SRVKM_CMD_PVRSRV_BRIDGE_SYNC_SYNCPRIMSET: fd_rogue [openat$img_rogue] ioctl$FLOPPY_FDCLRPRM : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDDEFPRM : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDEJECT : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDFLUSH : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDFMTBEG : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDFMTEND : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDFMTTRK : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDGETDRVPRM : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDGETDRVSTAT : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDGETDRVTYP : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDGETFDCSTAT : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDGETMAXERRS : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDGETPRM : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDMSGOFF : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDMSGON : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDPOLLDRVSTAT : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDRAWCMD : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDRESET : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDSETDRVPRM : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDSETEMSGTRESH : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDSETMAXERRS : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDSETPRM : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDTWADDLE : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDWERRORCLR : fd_floppy [syz_open_dev$floppy] ioctl$FLOPPY_FDWERRORGET : fd_floppy [syz_open_dev$floppy] ioctl$KBASE_HWCNT_READER_CLEAR : fd_hwcnt [ioctl$KBASE_IOCTL_HWCNT_READER_SETUP] ioctl$KBASE_HWCNT_READER_DISABLE_EVENT : fd_hwcnt [ioctl$KBASE_IOCTL_HWCNT_READER_SETUP] ioctl$KBASE_HWCNT_READER_DUMP : fd_hwcnt [ioctl$KBASE_IOCTL_HWCNT_READER_SETUP] ioctl$KBASE_HWCNT_READER_ENABLE_EVENT : fd_hwcnt [ioctl$KBASE_IOCTL_HWCNT_READER_SETUP] ioctl$KBASE_HWCNT_READER_GET_API_VERSION : fd_hwcnt [ioctl$KBASE_IOCTL_HWCNT_READER_SETUP] ioctl$KBASE_HWCNT_READER_GET_API_VERSION_WITH_FEATURES: fd_hwcnt [ioctl$KBASE_IOCTL_HWCNT_READER_SETUP] ioctl$KBASE_HWCNT_READER_GET_BUFFER : fd_hwcnt [ioctl$KBASE_IOCTL_HWCNT_READER_SETUP] ioctl$KBASE_HWCNT_READER_GET_BUFFER_SIZE : fd_hwcnt [ioctl$KBASE_IOCTL_HWCNT_READER_SETUP] ioctl$KBASE_HWCNT_READER_GET_BUFFER_WITH_CYCLES: fd_hwcnt [ioctl$KBASE_IOCTL_HWCNT_READER_SETUP] ioctl$KBASE_HWCNT_READER_GET_HWVER : fd_hwcnt [ioctl$KBASE_IOCTL_HWCNT_READER_SETUP] ioctl$KBASE_HWCNT_READER_PUT_BUFFER : fd_hwcnt [ioctl$KBASE_IOCTL_HWCNT_READER_SETUP] ioctl$KBASE_HWCNT_READER_PUT_BUFFER_WITH_CYCLES: fd_hwcnt [ioctl$KBASE_IOCTL_HWCNT_READER_SETUP] ioctl$KBASE_HWCNT_READER_SET_INTERVAL : fd_hwcnt [ioctl$KBASE_IOCTL_HWCNT_READER_SETUP] ioctl$KBASE_IOCTL_BUFFER_LIVENESS_UPDATE : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_CONTEXT_PRIORITY_CHECK : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_CS_CPU_QUEUE_DUMP : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_CS_EVENT_SIGNAL : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_CS_GET_GLB_IFACE : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_CS_QUEUE_BIND : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_CS_QUEUE_GROUP_CREATE : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_CS_QUEUE_GROUP_CREATE_1_6 : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_CS_QUEUE_GROUP_TERMINATE : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_CS_QUEUE_KICK : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_CS_QUEUE_REGISTER : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_CS_QUEUE_REGISTER_EX : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_CS_QUEUE_TERMINATE : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_CS_TILER_HEAP_INIT : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_CS_TILER_HEAP_INIT_1_13 : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_CS_TILER_HEAP_TERM : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_DISJOINT_QUERY : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_FENCE_VALIDATE : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_GET_CONTEXT_ID : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_GET_CPU_GPU_TIMEINFO : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_GET_DDK_VERSION : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_GET_GPUPROPS : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_HWCNT_CLEAR : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_HWCNT_DUMP : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_HWCNT_ENABLE : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_HWCNT_READER_SETUP : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_HWCNT_SET : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_JOB_SUBMIT : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_KCPU_QUEUE_CREATE : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_KCPU_QUEUE_DELETE : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_KCPU_QUEUE_ENQUEUE : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_KINSTR_PRFCNT_CMD : fd_kinstr [ioctl$KBASE_IOCTL_KINSTR_PRFCNT_SETUP] ioctl$KBASE_IOCTL_KINSTR_PRFCNT_ENUM_INFO : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_KINSTR_PRFCNT_GET_SAMPLE : fd_kinstr [ioctl$KBASE_IOCTL_KINSTR_PRFCNT_SETUP] ioctl$KBASE_IOCTL_KINSTR_PRFCNT_PUT_SAMPLE : fd_kinstr [ioctl$KBASE_IOCTL_KINSTR_PRFCNT_SETUP] ioctl$KBASE_IOCTL_KINSTR_PRFCNT_SETUP : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_MEM_ALIAS : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_MEM_ALLOC : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_MEM_ALLOC_EX : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_MEM_COMMIT : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_MEM_EXEC_INIT : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_MEM_FIND_CPU_OFFSET : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_MEM_FIND_GPU_START_AND_OFFSET: fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_MEM_FLAGS_CHANGE : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_MEM_FREE : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_MEM_IMPORT : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_MEM_JIT_INIT : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_MEM_JIT_INIT_10_2 : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_MEM_JIT_INIT_11_5 : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_MEM_PROFILE_ADD : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_MEM_QUERY : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_MEM_SYNC : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_POST_TERM : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_READ_USER_PAGE : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_SET_FLAGS : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_SET_LIMITED_CORE_COUNT : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_SOFT_EVENT_UPDATE : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_STICKY_RESOURCE_MAP : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_STICKY_RESOURCE_UNMAP : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_STREAM_CREATE : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_TLSTREAM_ACQUIRE : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_TLSTREAM_FLUSH : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_VERSION_CHECK : fd_bifrost [openat$bifrost openat$mali] ioctl$KBASE_IOCTL_VERSION_CHECK_RESERVED : fd_bifrost [openat$bifrost openat$mali] ioctl$KVM_ASSIGN_SET_MSIX_ENTRY : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_ASSIGN_SET_MSIX_NR : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_DIRTY_LOG_RING : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_DIRTY_LOG_RING_ACQ_REL : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_DISABLE_QUIRKS : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_DISABLE_QUIRKS2 : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_ENFORCE_PV_FEATURE_CPUID : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_CAP_EXCEPTION_PAYLOAD : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_EXIT_HYPERCALL : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_EXIT_ON_EMULATION_FAILURE : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_HALT_POLL : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_HYPERV_DIRECT_TLBFLUSH : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_CAP_HYPERV_ENFORCE_CPUID : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_CAP_HYPERV_ENLIGHTENED_VMCS : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_CAP_HYPERV_SEND_IPI : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_HYPERV_SYNIC : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_CAP_HYPERV_SYNIC2 : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_CAP_HYPERV_TLBFLUSH : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_HYPERV_VP_INDEX : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_MANUAL_DIRTY_LOG_PROTECT2 : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_MAX_VCPU_ID : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_MEMORY_FAULT_INFO : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_MSR_PLATFORM_INFO : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_PMU_CAPABILITY : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_PTP_KVM : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_SGX_ATTRIBUTE : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_SPLIT_IRQCHIP : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_STEAL_TIME : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_SYNC_REGS : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_CAP_VM_COPY_ENC_CONTEXT_FROM : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_VM_DISABLE_NX_HUGE_PAGES : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_VM_MOVE_ENC_CONTEXT_FROM : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_VM_TYPES : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_X2APIC_API : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_X86_APIC_BUS_CYCLES_NS : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_X86_BUS_LOCK_EXIT : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_X86_DISABLE_EXITS : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_X86_GUEST_MODE : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_X86_NOTIFY_VMEXIT : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_X86_USER_SPACE_MSR : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CAP_XEN_HVM : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CHECK_EXTENSION : fd_kvm [openat$kvm] ioctl$KVM_CHECK_EXTENSION_VM : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CLEAR_DIRTY_LOG : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CREATE_DEVICE : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CREATE_GUEST_MEMFD : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CREATE_IRQCHIP : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CREATE_PIT2 : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CREATE_VCPU : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_CREATE_VM : fd_kvm [openat$kvm] ioctl$KVM_DIRTY_TLB : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_API_VERSION : fd_kvm [openat$kvm] ioctl$KVM_GET_CLOCK : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_GET_CPUID2 : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_DEBUGREGS : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_DEVICE_ATTR : fd_kvmdev [ioctl$KVM_CREATE_DEVICE] ioctl$KVM_GET_DEVICE_ATTR_vcpu : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_DEVICE_ATTR_vm : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_GET_DIRTY_LOG : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_GET_EMULATED_CPUID : fd_kvm [openat$kvm] ioctl$KVM_GET_FPU : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_IRQCHIP : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_GET_LAPIC : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_MP_STATE : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_MSRS_cpu : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_MSRS_sys : fd_kvm [openat$kvm] ioctl$KVM_GET_MSR_FEATURE_INDEX_LIST : fd_kvm [openat$kvm] ioctl$KVM_GET_MSR_INDEX_LIST : fd_kvm [openat$kvm] ioctl$KVM_GET_NESTED_STATE : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_NR_MMU_PAGES : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_GET_ONE_REG : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_PIT : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_GET_PIT2 : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_GET_REGS : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_REG_LIST : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_SREGS : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_SREGS2 : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_STATS_FD_cpu : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_STATS_FD_vm : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_GET_SUPPORTED_CPUID : fd_kvm [openat$kvm] ioctl$KVM_GET_SUPPORTED_HV_CPUID_cpu : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_SUPPORTED_HV_CPUID_sys : fd_kvm [openat$kvm] ioctl$KVM_GET_TSC_KHZ_cpu : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_TSC_KHZ_vm : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_GET_VCPU_EVENTS : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_VCPU_MMAP_SIZE : fd_kvm [openat$kvm] ioctl$KVM_GET_XCRS : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_XSAVE : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_GET_XSAVE2 : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_HAS_DEVICE_ATTR : fd_kvmdev [ioctl$KVM_CREATE_DEVICE] ioctl$KVM_HAS_DEVICE_ATTR_vcpu : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_HAS_DEVICE_ATTR_vm : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_HYPERV_EVENTFD : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_INTERRUPT : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_IOEVENTFD : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_IRQFD : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_IRQ_LINE : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_IRQ_LINE_STATUS : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_KVMCLOCK_CTRL : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_MEMORY_ENCRYPT_REG_REGION : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_MEMORY_ENCRYPT_UNREG_REGION : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_NMI : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_PPC_ALLOCATE_HTAB : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_PRE_FAULT_MEMORY : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_REGISTER_COALESCED_MMIO : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_REINJECT_CONTROL : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_RESET_DIRTY_RINGS : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_RUN : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_S390_VCPU_FAULT : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_BOOT_CPU_ID : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SET_CLOCK : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SET_CPUID : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_CPUID2 : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_DEBUGREGS : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_DEVICE_ATTR : fd_kvmdev [ioctl$KVM_CREATE_DEVICE] ioctl$KVM_SET_DEVICE_ATTR_vcpu : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_DEVICE_ATTR_vm : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SET_FPU : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_GSI_ROUTING : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SET_GUEST_DEBUG : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_IDENTITY_MAP_ADDR : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SET_IRQCHIP : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SET_LAPIC : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_MEMORY_ATTRIBUTES : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SET_MP_STATE : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_MSRS : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_NESTED_STATE : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_NR_MMU_PAGES : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SET_ONE_REG : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_PIT : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SET_PIT2 : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SET_REGS : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_SIGNAL_MASK : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_SREGS : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_SREGS2 : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_TSC_KHZ_cpu : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_TSC_KHZ_vm : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SET_TSS_ADDR : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SET_USER_MEMORY_REGION : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SET_USER_MEMORY_REGION2 : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SET_VAPIC_ADDR : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_VCPU_EVENTS : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_XCRS : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SET_XSAVE : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_SEV_CERT_EXPORT : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_DBG_DECRYPT : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_DBG_ENCRYPT : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_ES_INIT : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_GET_ATTESTATION_REPORT : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_GUEST_STATUS : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_INIT : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_INIT2 : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_LAUNCH_FINISH : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_LAUNCH_MEASURE : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_LAUNCH_SECRET : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_LAUNCH_START : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_LAUNCH_UPDATE_DATA : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_LAUNCH_UPDATE_VMSA : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_RECEIVE_FINISH : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_RECEIVE_START : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_RECEIVE_UPDATE_DATA : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_RECEIVE_UPDATE_VMSA : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_SEND_CANCEL : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_SEND_FINISH : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_SEND_START : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_SEND_UPDATE_DATA : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_SEND_UPDATE_VMSA : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_SNP_LAUNCH_FINISH : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_SNP_LAUNCH_START : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SEV_SNP_LAUNCH_UPDATE : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SIGNAL_MSI : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_SMI : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_TPR_ACCESS_REPORTING : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_TRANSLATE : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_UNREGISTER_COALESCED_MMIO : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_X86_GET_MCE_CAP_SUPPORTED : fd_kvm [openat$kvm] ioctl$KVM_X86_SETUP_MCE : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_X86_SET_MCE : fd_kvmcpu [ioctl$KVM_CREATE_VCPU syz_kvm_add_vcpu$x86] ioctl$KVM_X86_SET_MSR_FILTER : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$KVM_XEN_HVM_CONFIG : fd_kvmvm [ioctl$KVM_CREATE_VM] ioctl$PERF_EVENT_IOC_DISABLE : fd_perf [perf_event_open perf_event_open$cgroup] ioctl$PERF_EVENT_IOC_ENABLE : fd_perf [perf_event_open perf_event_open$cgroup] ioctl$PERF_EVENT_IOC_ID : fd_perf [perf_event_open perf_event_open$cgroup] ioctl$PERF_EVENT_IOC_MODIFY_ATTRIBUTES : fd_perf [perf_event_open perf_event_open$cgroup] ioctl$PERF_EVENT_IOC_PAUSE_OUTPUT : fd_perf [perf_event_open perf_event_open$cgroup] ioctl$PERF_EVENT_IOC_PERIOD : fd_perf [perf_event_open perf_event_open$cgroup] ioctl$PERF_EVENT_IOC_QUERY_BPF : fd_perf [perf_event_open perf_event_open$cgroup] ioctl$PERF_EVENT_IOC_REFRESH : fd_perf [perf_event_open perf_event_open$cgroup] ioctl$PERF_EVENT_IOC_RESET : fd_perf [perf_event_open perf_event_open$cgroup] ioctl$PERF_EVENT_IOC_SET_BPF : fd_perf [perf_event_open perf_event_open$cgroup] ioctl$PERF_EVENT_IOC_SET_FILTER : fd_perf [perf_event_open perf_event_open$cgroup] ioctl$PERF_EVENT_IOC_SET_OUTPUT : fd_perf [perf_event_open perf_event_open$cgroup] ioctl$READ_COUNTERS : fd_rdma [openat$uverbs0] ioctl$SNDRV_FIREWIRE_IOCTL_GET_INFO : fd_snd_hw [syz_open_dev$sndhw] ioctl$SNDRV_FIREWIRE_IOCTL_LOCK : fd_snd_hw [syz_open_dev$sndhw] ioctl$SNDRV_FIREWIRE_IOCTL_TASCAM_STATE : fd_snd_hw [syz_open_dev$sndhw] ioctl$SNDRV_FIREWIRE_IOCTL_UNLOCK : fd_snd_hw [syz_open_dev$sndhw] ioctl$SNDRV_HWDEP_IOCTL_DSP_LOAD : fd_snd_hw [syz_open_dev$sndhw] ioctl$SNDRV_HWDEP_IOCTL_DSP_STATUS : fd_snd_hw [syz_open_dev$sndhw] ioctl$SNDRV_HWDEP_IOCTL_INFO : fd_snd_hw [syz_open_dev$sndhw] ioctl$SNDRV_HWDEP_IOCTL_PVERSION : fd_snd_hw [syz_open_dev$sndhw] ioctl$TE_IOCTL_CLOSE_CLIENT_SESSION : fd_tlk [openat$tlk_device] ioctl$TE_IOCTL_LAUNCH_OPERATION : fd_tlk [openat$tlk_device] ioctl$TE_IOCTL_OPEN_CLIENT_SESSION : fd_tlk [openat$tlk_device] ioctl$TE_IOCTL_SS_CMD : fd_tlk [openat$tlk_device] ioctl$TIPC_IOC_CONNECT : fd_trusty [openat$trusty openat$trusty_avb openat$trusty_gatekeeper ...] ioctl$TIPC_IOC_CONNECT_avb : fd_trusty_avb [openat$trusty_avb] ioctl$TIPC_IOC_CONNECT_gatekeeper : fd_trusty_gatekeeper [openat$trusty_gatekeeper] ioctl$TIPC_IOC_CONNECT_hwkey : fd_trusty_hwkey [openat$trusty_hwkey] ioctl$TIPC_IOC_CONNECT_hwrng : fd_trusty_hwrng [openat$trusty_hwrng] ioctl$TIPC_IOC_CONNECT_keymaster_secure : fd_trusty_km_secure [openat$trusty_km_secure] ioctl$TIPC_IOC_CONNECT_km : fd_trusty_km [openat$trusty_km] ioctl$TIPC_IOC_CONNECT_storage : fd_trusty_storage [openat$trusty_storage] ioctl$VFIO_CHECK_EXTENSION : fd_vfio [openat$vfio] ioctl$VFIO_GET_API_VERSION : fd_vfio [openat$vfio] ioctl$VFIO_IOMMU_GET_INFO : fd_vfio [openat$vfio] ioctl$VFIO_IOMMU_MAP_DMA : fd_vfio [openat$vfio] ioctl$VFIO_IOMMU_UNMAP_DMA : fd_vfio [openat$vfio] ioctl$VFIO_SET_IOMMU : fd_vfio [openat$vfio] ioctl$VTPM_PROXY_IOC_NEW_DEV : fd_vtpm [openat$vtpm] ioctl$sock_bt_cmtp_CMTPCONNADD : sock_bt_cmtp [syz_init_net_socket$bt_cmtp] ioctl$sock_bt_cmtp_CMTPCONNDEL : sock_bt_cmtp [syz_init_net_socket$bt_cmtp] ioctl$sock_bt_cmtp_CMTPGETCONNINFO : sock_bt_cmtp [syz_init_net_socket$bt_cmtp] ioctl$sock_bt_cmtp_CMTPGETCONNLIST : sock_bt_cmtp [syz_init_net_socket$bt_cmtp] mmap$DRM_I915 : fd_i915 [openat$i915] mmap$DRM_MSM : fd_msm [openat$msm] mmap$KVM_VCPU : vcpu_mmap_size [ioctl$KVM_GET_VCPU_MMAP_SIZE] mmap$bifrost : fd_bifrost [openat$bifrost openat$mali] mmap$perf : fd_perf [perf_event_open perf_event_open$cgroup] pkey_free : pkey [pkey_alloc] pkey_mprotect : pkey [pkey_alloc] read$sndhw : fd_snd_hw [syz_open_dev$sndhw] read$trusty : fd_trusty [openat$trusty openat$trusty_avb openat$trusty_gatekeeper ...] recvmsg$hf : sock_hf [socket$hf] sendmsg$hf : sock_hf [socket$hf] setsockopt$inet6_dccp_buf : sock_dccp6 [socket$inet6_dccp] setsockopt$inet6_dccp_int : sock_dccp6 [socket$inet6_dccp] setsockopt$inet_dccp_buf : sock_dccp [socket$inet_dccp] setsockopt$inet_dccp_int : sock_dccp [socket$inet_dccp] syz_kvm_add_vcpu$x86 : kvm_syz_vm$x86 [syz_kvm_setup_syzos_vm$x86] syz_kvm_assert_syzos_uexit$x86 : kvm_run_ptr [mmap$KVM_VCPU] syz_kvm_setup_cpu$x86 : fd_kvmvm [ioctl$KVM_CREATE_VM] syz_kvm_setup_syzos_vm$x86 : fd_kvmvm [ioctl$KVM_CREATE_VM] syz_memcpy_off$KVM_EXIT_HYPERCALL : kvm_run_ptr [mmap$KVM_VCPU] syz_memcpy_off$KVM_EXIT_MMIO : kvm_run_ptr [mmap$KVM_VCPU] write$ALLOC_MW : fd_rdma [openat$uverbs0] write$ALLOC_PD : fd_rdma [openat$uverbs0] write$ATTACH_MCAST : fd_rdma [openat$uverbs0] write$CLOSE_XRCD : fd_rdma [openat$uverbs0] write$CREATE_AH : fd_rdma [openat$uverbs0] write$CREATE_COMP_CHANNEL : fd_rdma [openat$uverbs0] write$CREATE_CQ : fd_rdma [openat$uverbs0] write$CREATE_CQ_EX : fd_rdma [openat$uverbs0] write$CREATE_FLOW : fd_rdma [openat$uverbs0] write$CREATE_QP : fd_rdma [openat$uverbs0] write$CREATE_RWQ_IND_TBL : fd_rdma [openat$uverbs0] write$CREATE_SRQ : fd_rdma [openat$uverbs0] write$CREATE_WQ : fd_rdma [openat$uverbs0] write$DEALLOC_MW : fd_rdma [openat$uverbs0] write$DEALLOC_PD : fd_rdma [openat$uverbs0] write$DEREG_MR : fd_rdma [openat$uverbs0] write$DESTROY_AH : fd_rdma [openat$uverbs0] write$DESTROY_CQ : fd_rdma [openat$uverbs0] write$DESTROY_FLOW : fd_rdma [openat$uverbs0] write$DESTROY_QP : fd_rdma [openat$uverbs0] write$DESTROY_RWQ_IND_TBL : fd_rdma [openat$uverbs0] write$DESTROY_SRQ : fd_rdma [openat$uverbs0] write$DESTROY_WQ : fd_rdma [openat$uverbs0] write$DETACH_MCAST : fd_rdma [openat$uverbs0] write$MLX5_ALLOC_PD : fd_rdma [openat$uverbs0] write$MLX5_CREATE_CQ : fd_rdma [openat$uverbs0] write$MLX5_CREATE_DV_QP : fd_rdma [openat$uverbs0] write$MLX5_CREATE_QP : fd_rdma [openat$uverbs0] write$MLX5_CREATE_SRQ : fd_rdma [openat$uverbs0] write$MLX5_CREATE_WQ : fd_rdma [openat$uverbs0] write$MLX5_GET_CONTEXT : fd_rdma [openat$uverbs0] write$MLX5_MODIFY_WQ : fd_rdma [openat$uverbs0] write$MODIFY_QP : fd_rdma [openat$uverbs0] write$MODIFY_SRQ : fd_rdma [openat$uverbs0] write$OPEN_XRCD : fd_rdma [openat$uverbs0] write$POLL_CQ : fd_rdma [openat$uverbs0] write$POST_RECV : fd_rdma [openat$uverbs0] write$POST_SEND : fd_rdma [openat$uverbs0] write$POST_SRQ_RECV : fd_rdma [openat$uverbs0] write$QUERY_DEVICE_EX : fd_rdma [openat$uverbs0] write$QUERY_PORT : fd_rdma [openat$uverbs0] write$QUERY_QP : fd_rdma [openat$uverbs0] write$QUERY_SRQ : fd_rdma [openat$uverbs0] write$REG_MR : fd_rdma [openat$uverbs0] write$REQ_NOTIFY_CQ : fd_rdma [openat$uverbs0] write$REREG_MR : fd_rdma [openat$uverbs0] write$RESIZE_CQ : fd_rdma [openat$uverbs0] write$capi20 : fd_capi20 [openat$capi20] write$capi20_data : fd_capi20 [openat$capi20] write$damon_attrs : fd_damon_attrs [openat$damon_attrs] write$damon_contexts : fd_damon_contexts [openat$damon_mk_contexts openat$damon_rm_contexts] write$damon_init_regions : fd_damon_init_regions [openat$damon_init_regions] write$damon_monitor_on : fd_damon_monitor_on [openat$damon_monitor_on] write$damon_schemes : fd_damon_schemes [openat$damon_schemes] write$damon_target_ids : fd_damon_target_ids [openat$damon_target_ids] write$proc_reclaim : fd_proc_reclaim [openat$proc_reclaim] write$sndhw : fd_snd_hw [syz_open_dev$sndhw] write$sndhw_fireworks : fd_snd_hw [syz_open_dev$sndhw] write$trusty : fd_trusty [openat$trusty openat$trusty_avb openat$trusty_gatekeeper ...] write$trusty_avb : fd_trusty_avb [openat$trusty_avb] write$trusty_gatekeeper : fd_trusty_gatekeeper [openat$trusty_gatekeeper] write$trusty_hwkey : fd_trusty_hwkey [openat$trusty_hwkey] write$trusty_hwrng : fd_trusty_hwrng [openat$trusty_hwrng] write$trusty_km : fd_trusty_km [openat$trusty_km] write$trusty_km_secure : fd_trusty_km_secure [openat$trusty_km_secure] write$trusty_storage : fd_trusty_storage [openat$trusty_storage] BinFmtMisc : enabled Comparisons : enabled Coverage : enabled DelayKcovMmap : enabled DevlinkPCI : PCI device 0000:00:10.0 is not available ExtraCoverage : enabled Fault : enabled KCSAN : write(/sys/kernel/debug/kcsan, on) failed KcovResetIoctl : kernel does not support ioctl(KCOV_RESET_TRACE) LRWPANEmulation : enabled Leak : failed to write(kmemleak, "scan=off") NetDevices : enabled NetInjection : enabled NicVF : PCI device 0000:00:11.0 is not available SandboxAndroid : setfilecon: setxattr failed. (errno 1: Operation not permitted). . process exited with status 67. SandboxNamespace : enabled SandboxNone : enabled SandboxSetuid : enabled Swap : enabled USBEmulation : enabled VhciInjection : enabled WifiEmulation : enabled syscalls : 3832/8048 2025/08/19 02:30:46 new: machine check complete 2025/08/19 02:30:47 new: adding 81150 seeds 2025/08/19 02:31:29 patched crashed: WARNING in mem_cgroup_migrate [need repro = true] 2025/08/19 02:31:29 scheduled a reproduction of 'WARNING in mem_cgroup_migrate' 2025/08/19 02:31:54 patched crashed: possible deadlock in dqget [need repro = true] 2025/08/19 02:31:54 scheduled a reproduction of 'possible deadlock in dqget' 2025/08/19 02:32:20 runner 8 connected 2025/08/19 02:32:30 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = true] 2025/08/19 02:32:30 scheduled a reproduction of 'possible deadlock in ocfs2_reserve_suballoc_bits' 2025/08/19 02:32:39 patched crashed: possible deadlock in ocfs2_try_remove_refcount_tree [need repro = true] 2025/08/19 02:32:39 scheduled a reproduction of 'possible deadlock in ocfs2_try_remove_refcount_tree' 2025/08/19 02:32:41 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = true] 2025/08/19 02:32:41 scheduled a reproduction of 'possible deadlock in ocfs2_reserve_suballoc_bits' 2025/08/19 02:32:47 base crash: possible deadlock in ocfs2_try_remove_refcount_tree 2025/08/19 02:32:54 patched crashed: possible deadlock in ocfs2_try_remove_refcount_tree [need repro = false] 2025/08/19 02:33:05 patched crashed: possible deadlock in ocfs2_try_remove_refcount_tree [need repro = false] 2025/08/19 02:33:18 runner 2 connected 2025/08/19 02:33:27 runner 3 connected 2025/08/19 02:33:31 runner 7 connected 2025/08/19 02:33:36 runner 2 connected 2025/08/19 02:33:40 patched crashed: possible deadlock in ocfs2_init_acl [need repro = true] 2025/08/19 02:33:40 scheduled a reproduction of 'possible deadlock in ocfs2_init_acl' 2025/08/19 02:33:44 runner 0 connected 2025/08/19 02:33:55 runner 5 connected 2025/08/19 02:34:29 runner 8 connected 2025/08/19 02:34:34 STAT { "buffer too small": 0, "candidate triage jobs": 47, "candidates": 78510, "comps overflows": 0, "corpus": 2567, "corpus [files]": 825, "corpus [symbols]": 1793, "cover overflows": 1548, "coverage": 143279, "distributor delayed": 4539, "distributor undelayed": 4538, "distributor violated": 122, "exec candidate": 2640, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 0, "exec seeds": 0, "exec smash": 0, "exec total [base]": 9607, "exec total [new]": 11654, "exec triage": 8208, "executor restarts": 115, "fault jobs": 0, "fuzzer jobs": 47, "fuzzing VMs [base]": 4, "fuzzing VMs [new]": 6, "hints jobs": 0, "max signal": 146111, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 0, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 2640, "no exec duration": 43642000000, "no exec requests": 373, "pending": 6, "prog exec time": 183, "reproducing": 0, "rpc recv": 881761676, "rpc sent": 88495704, "signal": 141280, "smash jobs": 0, "triage jobs": 0, "vm output": 1881893, "vm restarts [base]": 5, "vm restarts [new]": 15 } 2025/08/19 02:35:31 base crash: possible deadlock in ocfs2_init_acl 2025/08/19 02:35:59 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = true] 2025/08/19 02:35:59 scheduled a reproduction of 'possible deadlock in ocfs2_reserve_suballoc_bits' 2025/08/19 02:36:12 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = true] 2025/08/19 02:36:12 scheduled a reproduction of 'possible deadlock in ocfs2_reserve_suballoc_bits' 2025/08/19 02:36:15 base crash: lost connection to test machine 2025/08/19 02:36:20 runner 2 connected 2025/08/19 02:36:24 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = true] 2025/08/19 02:36:24 scheduled a reproduction of 'possible deadlock in ocfs2_reserve_suballoc_bits' 2025/08/19 02:36:43 patched crashed: WARNING in __linkwatch_sync_dev [need repro = true] 2025/08/19 02:36:43 scheduled a reproduction of 'WARNING in __linkwatch_sync_dev' 2025/08/19 02:36:46 base crash: possible deadlock in ocfs2_reserve_suballoc_bits 2025/08/19 02:36:48 runner 8 connected 2025/08/19 02:36:54 patched crashed: WARNING in __linkwatch_sync_dev [need repro = true] 2025/08/19 02:36:54 scheduled a reproduction of 'WARNING in __linkwatch_sync_dev' 2025/08/19 02:37:00 runner 7 connected 2025/08/19 02:37:03 runner 3 connected 2025/08/19 02:37:09 base crash: lost connection to test machine 2025/08/19 02:37:14 runner 5 connected 2025/08/19 02:37:35 runner 1 connected 2025/08/19 02:37:50 patched crashed: possible deadlock in ocfs2_try_remove_refcount_tree [need repro = false] 2025/08/19 02:37:58 runner 0 connected 2025/08/19 02:38:23 base crash: possible deadlock in ocfs2_reserve_suballoc_bits 2025/08/19 02:38:27 base crash: possible deadlock in ocfs2_try_remove_refcount_tree 2025/08/19 02:38:28 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 02:38:39 runner 8 connected 2025/08/19 02:38:40 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 02:39:12 runner 0 connected 2025/08/19 02:39:15 runner 1 connected 2025/08/19 02:39:16 runner 5 connected 2025/08/19 02:39:29 runner 7 connected 2025/08/19 02:39:30 base crash: KASAN: slab-use-after-free Read in xfrm_alloc_spi 2025/08/19 02:39:34 STAT { "buffer too small": 0, "candidate triage jobs": 100, "candidates": 74870, "comps overflows": 0, "corpus": 6133, "corpus [files]": 1542, "corpus [symbols]": 3585, "cover overflows": 4129, "coverage": 187731, "distributor delayed": 11484, "distributor undelayed": 11411, "distributor violated": 894, "exec candidate": 6280, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 0, "exec seeds": 0, "exec smash": 0, "exec total [base]": 19536, "exec total [new]": 28115, "exec triage": 19542, "executor restarts": 146, "fault jobs": 0, "fuzzer jobs": 100, "fuzzing VMs [base]": 3, "fuzzing VMs [new]": 4, "hints jobs": 0, "max signal": 190090, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 0, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 6280, "no exec duration": 43821000000, "no exec requests": 379, "pending": 11, "prog exec time": 193, "reproducing": 0, "rpc recv": 1578171312, "rpc sent": 187712848, "signal": 184774, "smash jobs": 0, "triage jobs": 0, "vm output": 3704519, "vm restarts [base]": 11, "vm restarts [new]": 21 } 2025/08/19 02:39:40 new: boot error: can't ssh into the instance 2025/08/19 02:39:40 new: boot error: can't ssh into the instance 2025/08/19 02:39:45 base crash: possible deadlock in ocfs2_init_acl 2025/08/19 02:40:19 runner 2 connected 2025/08/19 02:40:28 runner 4 connected 2025/08/19 02:40:30 runner 1 connected 2025/08/19 02:40:34 runner 3 connected 2025/08/19 02:40:34 patched crashed: KASAN: slab-use-after-free Read in xfrm_alloc_spi [need repro = false] 2025/08/19 02:40:44 patched crashed: KASAN: slab-use-after-free Read in xfrm_alloc_spi [need repro = false] 2025/08/19 02:41:30 runner 2 connected 2025/08/19 02:41:31 base crash: KASAN: slab-use-after-free Read in __xfrm_state_lookup 2025/08/19 02:41:32 runner 0 connected 2025/08/19 02:41:33 patched crashed: possible deadlock in ocfs2_try_remove_refcount_tree [need repro = false] 2025/08/19 02:41:53 patched crashed: possible deadlock in ocfs2_xattr_set [need repro = true] 2025/08/19 02:41:53 scheduled a reproduction of 'possible deadlock in ocfs2_xattr_set' 2025/08/19 02:42:00 new: boot error: can't ssh into the instance 2025/08/19 02:42:04 patched crashed: possible deadlock in ocfs2_xattr_set [need repro = true] 2025/08/19 02:42:04 scheduled a reproduction of 'possible deadlock in ocfs2_xattr_set' 2025/08/19 02:42:21 runner 1 connected 2025/08/19 02:42:21 patched crashed: possible deadlock in ocfs2_try_remove_refcount_tree [need repro = false] 2025/08/19 02:42:22 runner 1 connected 2025/08/19 02:42:34 patched crashed: possible deadlock in ocfs2_try_remove_refcount_tree [need repro = false] 2025/08/19 02:42:49 runner 6 connected 2025/08/19 02:42:50 base crash: possible deadlock in ocfs2_try_remove_refcount_tree 2025/08/19 02:42:52 runner 2 connected 2025/08/19 02:43:01 patched crashed: possible deadlock in ocfs2_try_remove_refcount_tree [need repro = false] 2025/08/19 02:43:23 runner 0 connected 2025/08/19 02:43:39 runner 1 connected 2025/08/19 02:43:50 runner 1 connected 2025/08/19 02:43:56 patched crashed: WARNING in xfrm_state_fini [need repro = true] 2025/08/19 02:43:56 scheduled a reproduction of 'WARNING in xfrm_state_fini' 2025/08/19 02:44:34 STAT { "buffer too small": 0, "candidate triage jobs": 23, "candidates": 71046, "comps overflows": 0, "corpus": 9991, "corpus [files]": 2024, "corpus [symbols]": 5131, "cover overflows": 7211, "coverage": 211064, "distributor delayed": 19111, "distributor undelayed": 19110, "distributor violated": 1245, "exec candidate": 10104, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 0, "exec seeds": 0, "exec smash": 0, "exec total [base]": 34586, "exec total [new]": 47043, "exec triage": 31575, "executor restarts": 190, "fault jobs": 0, "fuzzer jobs": 23, "fuzzing VMs [base]": 4, "fuzzing VMs [new]": 5, "hints jobs": 0, "max signal": 213103, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 0, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 10104, "no exec duration": 44075000000, "no exec requests": 397, "pending": 14, "prog exec time": 160, "reproducing": 0, "rpc recv": 2311981416, "rpc sent": 310301840, "signal": 207851, "smash jobs": 0, "triage jobs": 0, "vm output": 5058437, "vm restarts [base]": 15, "vm restarts [new]": 30 } 2025/08/19 02:44:45 runner 2 connected 2025/08/19 02:45:54 patched crashed: KASAN: slab-use-after-free Read in __xfrm_state_lookup [need repro = false] 2025/08/19 02:46:03 patched crashed: KASAN: slab-use-after-free Read in xfrm_state_find [need repro = true] 2025/08/19 02:46:03 scheduled a reproduction of 'KASAN: slab-use-after-free Read in xfrm_state_find' 2025/08/19 02:46:29 base crash: lost connection to test machine 2025/08/19 02:46:43 runner 7 connected 2025/08/19 02:46:45 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 02:46:49 new: boot error: can't ssh into the instance 2025/08/19 02:46:52 runner 0 connected 2025/08/19 02:47:00 new: boot error: can't ssh into the instance 2025/08/19 02:47:18 runner 0 connected 2025/08/19 02:47:28 patched crashed: lost connection to test machine [need repro = false] 2025/08/19 02:47:34 runner 2 connected 2025/08/19 02:47:38 runner 9 connected 2025/08/19 02:47:49 runner 3 connected 2025/08/19 02:47:52 base crash: lost connection to test machine 2025/08/19 02:48:17 runner 0 connected 2025/08/19 02:48:43 runner 1 connected 2025/08/19 02:49:01 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 02:49:34 STAT { "buffer too small": 0, "candidate triage jobs": 40, "candidates": 66686, "comps overflows": 0, "corpus": 14273, "corpus [files]": 2562, "corpus [symbols]": 6729, "cover overflows": 10610, "coverage": 230759, "distributor delayed": 25827, "distributor undelayed": 25827, "distributor violated": 1405, "exec candidate": 14464, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 0, "exec seeds": 0, "exec smash": 0, "exec total [base]": 51805, "exec total [new]": 69154, "exec triage": 45294, "executor restarts": 242, "fault jobs": 0, "fuzzer jobs": 40, "fuzzing VMs [base]": 4, "fuzzing VMs [new]": 7, "hints jobs": 0, "max signal": 233412, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 0, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 14464, "no exec duration": 46736000000, "no exec requests": 404, "pending": 15, "prog exec time": 208, "reproducing": 0, "rpc recv": 2931628564, "rpc sent": 447610216, "signal": 227116, "smash jobs": 0, "triage jobs": 0, "vm output": 6956911, "vm restarts [base]": 17, "vm restarts [new]": 37 } 2025/08/19 02:49:49 runner 9 connected 2025/08/19 02:50:04 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 02:50:52 runner 3 connected 2025/08/19 02:51:03 patched crashed: KASAN: slab-use-after-free Read in xfrm_alloc_spi [need repro = false] 2025/08/19 02:51:13 patched crashed: KASAN: slab-use-after-free Read in xfrm_alloc_spi [need repro = false] 2025/08/19 02:51:24 patched crashed: KASAN: slab-use-after-free Read in xfrm_alloc_spi [need repro = false] 2025/08/19 02:51:38 patched crashed: KASAN: slab-use-after-free Read in __xfrm_state_lookup [need repro = false] 2025/08/19 02:51:46 base crash: INFO: task hung in disable_device 2025/08/19 02:51:52 runner 7 connected 2025/08/19 02:51:58 new: boot error: can't ssh into the instance 2025/08/19 02:52:03 runner 2 connected 2025/08/19 02:52:13 runner 1 connected 2025/08/19 02:52:27 new: boot error: can't ssh into the instance 2025/08/19 02:52:28 runner 8 connected 2025/08/19 02:52:37 runner 2 connected 2025/08/19 02:52:48 runner 4 connected 2025/08/19 02:53:01 patched crashed: WARNING in drv_unassign_vif_chanctx [need repro = true] 2025/08/19 02:53:01 scheduled a reproduction of 'WARNING in drv_unassign_vif_chanctx' 2025/08/19 02:53:09 patched crashed: WARNING in xfrm_state_fini [need repro = true] 2025/08/19 02:53:09 scheduled a reproduction of 'WARNING in xfrm_state_fini' 2025/08/19 02:53:16 runner 5 connected 2025/08/19 02:53:28 base crash: WARNING in xfrm_state_fini 2025/08/19 02:53:58 runner 9 connected 2025/08/19 02:54:06 runner 8 connected 2025/08/19 02:54:15 runner 1 connected 2025/08/19 02:54:34 STAT { "buffer too small": 0, "candidate triage jobs": 51, "candidates": 61823, "comps overflows": 0, "corpus": 19069, "corpus [files]": 3210, "corpus [symbols]": 8425, "cover overflows": 14388, "coverage": 248401, "distributor delayed": 31750, "distributor undelayed": 31750, "distributor violated": 1416, "exec candidate": 19327, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 0, "exec seeds": 0, "exec smash": 0, "exec total [base]": 66931, "exec total [new]": 94115, "exec triage": 60433, "executor restarts": 293, "fault jobs": 0, "fuzzer jobs": 51, "fuzzing VMs [base]": 4, "fuzzing VMs [new]": 10, "hints jobs": 0, "max signal": 251336, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 0, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 19327, "no exec duration": 46856000000, "no exec requests": 409, "pending": 17, "prog exec time": 301, "reproducing": 0, "rpc recv": 3680317812, "rpc sent": 600182136, "signal": 244190, "smash jobs": 0, "triage jobs": 0, "vm output": 9315898, "vm restarts [base]": 19, "vm restarts [new]": 47 } 2025/08/19 02:55:22 patched crashed: WARNING in xfrm_state_fini [need repro = false] 2025/08/19 02:55:26 base crash: INFO: task hung in disable_device 2025/08/19 02:55:47 patched crashed: KASAN: slab-use-after-free Read in xfrm_alloc_spi [need repro = false] 2025/08/19 02:55:52 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 02:56:18 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 02:56:18 patched crashed: KASAN: slab-use-after-free Read in xfrm_alloc_spi [need repro = false] 2025/08/19 02:56:19 runner 7 connected 2025/08/19 02:56:24 runner 3 connected 2025/08/19 02:56:28 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 02:56:36 runner 6 connected 2025/08/19 02:56:41 patched crashed: KASAN: slab-use-after-free Read in xfrm_alloc_spi [need repro = false] 2025/08/19 02:56:43 runner 9 connected 2025/08/19 02:57:02 patched crashed: WARNING: suspicious RCU usage in get_callchain_entry [need repro = true] 2025/08/19 02:57:02 scheduled a reproduction of 'WARNING: suspicious RCU usage in get_callchain_entry' 2025/08/19 02:57:04 base crash: possible deadlock in ocfs2_reserve_suballoc_bits 2025/08/19 02:57:07 runner 1 connected 2025/08/19 02:57:09 runner 8 connected 2025/08/19 02:57:12 patched crashed: WARNING: suspicious RCU usage in get_callchain_entry [need repro = true] 2025/08/19 02:57:12 scheduled a reproduction of 'WARNING: suspicious RCU usage in get_callchain_entry' 2025/08/19 02:57:18 runner 5 connected 2025/08/19 02:57:31 runner 2 connected 2025/08/19 02:58:00 runner 2 connected 2025/08/19 02:58:01 runner 7 connected 2025/08/19 02:58:09 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 02:58:50 patched crashed: WARNING in drv_unassign_vif_chanctx [need repro = true] 2025/08/19 02:58:50 scheduled a reproduction of 'WARNING in drv_unassign_vif_chanctx' 2025/08/19 02:59:06 runner 2 connected 2025/08/19 02:59:17 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 02:59:28 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 02:59:34 STAT { "buffer too small": 0, "candidate triage jobs": 32, "candidates": 56928, "comps overflows": 0, "corpus": 23899, "corpus [files]": 3813, "corpus [symbols]": 10081, "cover overflows": 18252, "coverage": 263019, "distributor delayed": 37504, "distributor undelayed": 37502, "distributor violated": 1432, "exec candidate": 24222, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 0, "exec seeds": 0, "exec smash": 0, "exec total [base]": 81402, "exec total [new]": 121058, "exec triage": 75766, "executor restarts": 347, "fault jobs": 0, "fuzzer jobs": 32, "fuzzing VMs [base]": 4, "fuzzing VMs [new]": 6, "hints jobs": 0, "max signal": 266083, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 0, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 24222, "no exec duration": 46909000000, "no exec requests": 412, "pending": 20, "prog exec time": 193, "reproducing": 0, "rpc recv": 4397982412, "rpc sent": 767131928, "signal": 258510, "smash jobs": 0, "triage jobs": 0, "vm output": 12048599, "vm restarts [base]": 21, "vm restarts [new]": 56 } 2025/08/19 02:59:39 runner 4 connected 2025/08/19 03:00:17 runner 9 connected 2025/08/19 03:00:29 base crash: KASAN: slab-use-after-free Write in __xfrm_state_delete 2025/08/19 03:00:33 base crash: possible deadlock in ocfs2_try_remove_refcount_tree 2025/08/19 03:00:40 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:00:52 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:01:09 base crash: lost connection to test machine 2025/08/19 03:01:18 runner 2 connected 2025/08/19 03:01:22 runner 3 connected 2025/08/19 03:01:24 patched crashed: possible deadlock in ocfs2_setattr [need repro = true] 2025/08/19 03:01:24 scheduled a reproduction of 'possible deadlock in ocfs2_setattr' 2025/08/19 03:01:31 runner 8 connected 2025/08/19 03:01:34 patched crashed: possible deadlock in ocfs2_try_remove_refcount_tree [need repro = false] 2025/08/19 03:01:36 patched crashed: lost connection to test machine [need repro = false] 2025/08/19 03:01:49 base crash: KASAN: slab-use-after-free Read in jfs_lazycommit 2025/08/19 03:01:51 base crash: possible deadlock in ocfs2_setattr 2025/08/19 03:01:58 patched crashed: lost connection to test machine [need repro = false] 2025/08/19 03:02:05 runner 5 connected 2025/08/19 03:02:22 runner 1 connected 2025/08/19 03:02:25 runner 4 connected 2025/08/19 03:02:38 runner 2 connected 2025/08/19 03:02:41 runner 3 connected 2025/08/19 03:02:48 runner 9 connected 2025/08/19 03:03:36 base crash: KASAN: slab-use-after-free Read in xfrm_alloc_spi 2025/08/19 03:03:58 patched crashed: possible deadlock in ocfs2_try_remove_refcount_tree [need repro = false] 2025/08/19 03:04:09 patched crashed: possible deadlock in ocfs2_try_remove_refcount_tree [need repro = false] 2025/08/19 03:04:25 runner 0 connected 2025/08/19 03:04:34 STAT { "buffer too small": 0, "candidate triage jobs": 23, "candidates": 52993, "comps overflows": 0, "corpus": 27768, "corpus [files]": 4324, "corpus [symbols]": 11455, "cover overflows": 21382, "coverage": 273452, "distributor delayed": 42994, "distributor undelayed": 42993, "distributor violated": 1608, "exec candidate": 28157, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 1, "exec seeds": 0, "exec smash": 0, "exec total [base]": 92131, "exec total [new]": 143866, "exec triage": 88038, "executor restarts": 408, "fault jobs": 0, "fuzzer jobs": 23, "fuzzing VMs [base]": 3, "fuzzing VMs [new]": 5, "hints jobs": 0, "max signal": 276825, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 0, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 28157, "no exec duration": 46956000000, "no exec requests": 417, "pending": 21, "prog exec time": 184, "reproducing": 0, "rpc recv": 5071491132, "rpc sent": 908147120, "signal": 268836, "smash jobs": 0, "triage jobs": 0, "vm output": 14421387, "vm restarts [base]": 26, "vm restarts [new]": 63 } 2025/08/19 03:04:47 runner 8 connected 2025/08/19 03:04:58 runner 9 connected 2025/08/19 03:04:59 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 03:05:48 runner 5 connected 2025/08/19 03:06:15 patched crashed: general protection fault in pcl818_ai_cancel [need repro = true] 2025/08/19 03:06:15 scheduled a reproduction of 'general protection fault in pcl818_ai_cancel' 2025/08/19 03:06:40 patched crashed: WARNING in xfrm6_tunnel_net_exit [need repro = true] 2025/08/19 03:06:40 scheduled a reproduction of 'WARNING in xfrm6_tunnel_net_exit' 2025/08/19 03:07:05 runner 8 connected 2025/08/19 03:07:07 new: boot error: can't ssh into the instance 2025/08/19 03:07:30 runner 3 connected 2025/08/19 03:07:41 base crash: general protection fault in pcl818_ai_cancel 2025/08/19 03:07:58 runner 0 connected 2025/08/19 03:08:25 patched crashed: possible deadlock in ocfs2_setattr [need repro = false] 2025/08/19 03:09:22 new: boot error: can't ssh into the instance 2025/08/19 03:09:34 STAT { "buffer too small": 0, "candidate triage jobs": 37, "candidates": 49787, "comps overflows": 0, "corpus": 30897, "corpus [files]": 4727, "corpus [symbols]": 12558, "cover overflows": 24053, "coverage": 280710, "distributor delayed": 47895, "distributor undelayed": 47894, "distributor violated": 1784, "exec candidate": 31363, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 3, "exec seeds": 0, "exec smash": 0, "exec total [base]": 101317, "exec total [new]": 163699, "exec triage": 98030, "executor restarts": 448, "fault jobs": 0, "fuzzer jobs": 37, "fuzzing VMs [base]": 2, "fuzzing VMs [new]": 7, "hints jobs": 0, "max signal": 284180, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 0, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 31363, "no exec duration": 47517000000, "no exec requests": 423, "pending": 23, "prog exec time": 248, "reproducing": 0, "rpc recv": 5495216252, "rpc sent": 1033587192, "signal": 275987, "smash jobs": 0, "triage jobs": 0, "vm output": 16589258, "vm restarts [base]": 26, "vm restarts [new]": 69 } 2025/08/19 03:10:02 patched crashed: INFO: task hung in corrupted [need repro = true] 2025/08/19 03:10:02 scheduled a reproduction of 'INFO: task hung in corrupted' 2025/08/19 03:10:11 runner 6 connected 2025/08/19 03:10:39 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 03:10:50 runner 5 connected 2025/08/19 03:10:58 new: boot error: can't ssh into the instance 2025/08/19 03:11:15 base: boot error: can't ssh into the instance 2025/08/19 03:11:29 runner 8 connected 2025/08/19 03:11:33 patched crashed: KASAN: slab-use-after-free Read in xfrm_alloc_spi [need repro = false] 2025/08/19 03:11:43 patched crashed: KASAN: slab-use-after-free Read in xfrm_alloc_spi [need repro = false] 2025/08/19 03:11:48 runner 2 connected 2025/08/19 03:12:04 runner 1 connected 2025/08/19 03:12:11 patched crashed: possible deadlock in ocfs2_xattr_set [need repro = true] 2025/08/19 03:12:11 scheduled a reproduction of 'possible deadlock in ocfs2_xattr_set' 2025/08/19 03:12:23 runner 5 connected 2025/08/19 03:12:41 patched crashed: WARNING in xfrm_state_fini [need repro = false] 2025/08/19 03:12:57 base crash: possible deadlock in ocfs2_reserve_suballoc_bits 2025/08/19 03:12:59 base crash: possible deadlock in attr_data_get_block 2025/08/19 03:13:00 runner 8 connected 2025/08/19 03:13:02 patched crashed: KASAN: slab-use-after-free Read in __xfrm_state_lookup [need repro = false] 2025/08/19 03:13:19 patched crashed: possible deadlock in ocfs2_xattr_set [need repro = true] 2025/08/19 03:13:19 scheduled a reproduction of 'possible deadlock in ocfs2_xattr_set' 2025/08/19 03:13:20 patched crashed: WARNING in xfrm_state_fini [need repro = false] 2025/08/19 03:13:31 runner 6 connected 2025/08/19 03:13:40 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:13:49 runner 2 connected 2025/08/19 03:13:52 runner 7 connected 2025/08/19 03:14:04 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:14:08 runner 1 connected 2025/08/19 03:14:09 runner 3 connected 2025/08/19 03:14:28 runner 8 connected 2025/08/19 03:14:34 STAT { "buffer too small": 0, "candidate triage jobs": 30, "candidates": 46229, "comps overflows": 0, "corpus": 34390, "corpus [files]": 5132, "corpus [symbols]": 13694, "cover overflows": 26958, "coverage": 288597, "distributor delayed": 53030, "distributor undelayed": 53030, "distributor violated": 1836, "exec candidate": 34921, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 3, "exec seeds": 0, "exec smash": 0, "exec total [base]": 107782, "exec total [new]": 185851, "exec triage": 109116, "executor restarts": 501, "fault jobs": 0, "fuzzer jobs": 30, "fuzzing VMs [base]": 2, "fuzzing VMs [new]": 6, "hints jobs": 0, "max signal": 291996, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 0, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 34921, "no exec duration": 47801000000, "no exec requests": 426, "pending": 26, "prog exec time": 255, "reproducing": 0, "rpc recv": 6161037872, "rpc sent": 1164862976, "signal": 283706, "smash jobs": 0, "triage jobs": 0, "vm output": 18911527, "vm restarts [base]": 28, "vm restarts [new]": 80 } 2025/08/19 03:14:47 patched crashed: possible deadlock in ocfs2_try_remove_refcount_tree [need repro = false] 2025/08/19 03:14:48 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:14:53 runner 2 connected 2025/08/19 03:14:54 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:14:58 patched crashed: possible deadlock in ocfs2_try_remove_refcount_tree [need repro = false] 2025/08/19 03:15:01 base crash: possible deadlock in ocfs2_init_acl 2025/08/19 03:15:37 runner 6 connected 2025/08/19 03:15:38 runner 7 connected 2025/08/19 03:15:42 runner 3 connected 2025/08/19 03:15:47 runner 8 connected 2025/08/19 03:15:50 runner 2 connected 2025/08/19 03:16:09 base crash: unregister_netdevice: waiting for DEV to become free 2025/08/19 03:16:33 patched crashed: lost connection to test machine [need repro = false] 2025/08/19 03:16:57 runner 3 connected 2025/08/19 03:16:59 patched crashed: KASAN: slab-use-after-free Read in xfrm_alloc_spi [need repro = false] 2025/08/19 03:17:23 runner 3 connected 2025/08/19 03:17:45 patched crashed: possible deadlock in ocfs2_xattr_set [need repro = true] 2025/08/19 03:17:45 scheduled a reproduction of 'possible deadlock in ocfs2_xattr_set' 2025/08/19 03:17:46 base: boot error: can't ssh into the instance 2025/08/19 03:17:47 runner 9 connected 2025/08/19 03:17:57 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 03:18:11 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 03:18:20 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 03:18:22 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 03:18:30 patched crashed: possible deadlock in ocfs2_xattr_set [need repro = true] 2025/08/19 03:18:30 scheduled a reproduction of 'possible deadlock in ocfs2_xattr_set' 2025/08/19 03:18:31 new: boot error: can't ssh into the instance 2025/08/19 03:18:35 runner 0 connected 2025/08/19 03:18:36 runner 2 connected 2025/08/19 03:18:45 runner 3 connected 2025/08/19 03:18:46 patched crashed: unregister_netdevice: waiting for DEV to become free [need repro = false] 2025/08/19 03:18:59 runner 8 connected 2025/08/19 03:19:08 runner 6 connected 2025/08/19 03:19:11 runner 7 connected 2025/08/19 03:19:19 runner 1 connected 2025/08/19 03:19:34 STAT { "buffer too small": 0, "candidate triage jobs": 45, "candidates": 43813, "comps overflows": 0, "corpus": 36761, "corpus [files]": 5489, "corpus [symbols]": 14528, "cover overflows": 28658, "coverage": 294441, "distributor delayed": 56687, "distributor undelayed": 56687, "distributor violated": 1857, "exec candidate": 37337, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 7, "exec seeds": 0, "exec smash": 0, "exec total [base]": 113254, "exec total [new]": 199721, "exec triage": 116412, "executor restarts": 570, "fault jobs": 0, "fuzzer jobs": 45, "fuzzing VMs [base]": 3, "fuzzing VMs [new]": 7, "hints jobs": 0, "max signal": 297883, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 0, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 37337, "no exec duration": 47801000000, "no exec requests": 426, "pending": 28, "prog exec time": 443, "reproducing": 0, "rpc recv": 6884225276, "rpc sent": 1293598496, "signal": 289577, "smash jobs": 0, "triage jobs": 0, "vm output": 21899038, "vm restarts [base]": 31, "vm restarts [new]": 93 } 2025/08/19 03:19:35 runner 5 connected 2025/08/19 03:19:50 base crash: possible deadlock in ocfs2_xattr_set 2025/08/19 03:20:00 base crash: possible deadlock in ocfs2_try_remove_refcount_tree 2025/08/19 03:20:17 base crash: WARNING in dbAdjTree 2025/08/19 03:20:34 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:20:39 runner 3 connected 2025/08/19 03:20:45 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:20:49 runner 0 connected 2025/08/19 03:21:05 runner 2 connected 2025/08/19 03:21:06 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:21:19 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:21:23 runner 8 connected 2025/08/19 03:21:28 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 03:21:31 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:21:35 base crash: possible deadlock in ocfs2_try_remove_refcount_tree 2025/08/19 03:21:35 runner 9 connected 2025/08/19 03:21:40 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 03:21:49 new: boot error: can't ssh into the instance 2025/08/19 03:21:54 runner 5 connected 2025/08/19 03:22:09 runner 1 connected 2025/08/19 03:22:17 runner 6 connected 2025/08/19 03:22:19 runner 3 connected 2025/08/19 03:22:24 runner 0 connected 2025/08/19 03:22:24 base crash: possible deadlock in ocfs2_reserve_suballoc_bits 2025/08/19 03:22:28 runner 7 connected 2025/08/19 03:22:30 patched crashed: possible deadlock in ntfs_fiemap [need repro = true] 2025/08/19 03:22:30 scheduled a reproduction of 'possible deadlock in ntfs_fiemap' 2025/08/19 03:22:38 runner 4 connected 2025/08/19 03:23:03 base: boot error: can't ssh into the instance 2025/08/19 03:23:19 runner 2 connected 2025/08/19 03:23:20 runner 3 connected 2025/08/19 03:23:51 runner 1 connected 2025/08/19 03:24:06 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 03:24:27 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 03:24:34 STAT { "buffer too small": 0, "candidate triage jobs": 22, "candidates": 41384, "comps overflows": 0, "corpus": 39163, "corpus [files]": 6001, "corpus [symbols]": 15442, "cover overflows": 30476, "coverage": 300339, "distributor delayed": 59562, "distributor undelayed": 59561, "distributor violated": 1857, "exec candidate": 39766, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 8, "exec seeds": 0, "exec smash": 0, "exec total [base]": 117983, "exec total [new]": 215700, "exec triage": 123837, "executor restarts": 656, "fault jobs": 0, "fuzzer jobs": 22, "fuzzing VMs [base]": 4, "fuzzing VMs [new]": 7, "hints jobs": 0, "max signal": 303645, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 0, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 39766, "no exec duration": 47801000000, "no exec requests": 426, "pending": 29, "prog exec time": 338, "reproducing": 0, "rpc recv": 7648127260, "rpc sent": 1441848864, "signal": 295490, "smash jobs": 0, "triage jobs": 0, "vm output": 26162632, "vm restarts [base]": 37, "vm restarts [new]": 103 } 2025/08/19 03:24:56 runner 9 connected 2025/08/19 03:25:00 patched crashed: INFO: task hung in sync_bdevs [need repro = true] 2025/08/19 03:25:00 scheduled a reproduction of 'INFO: task hung in sync_bdevs' 2025/08/19 03:25:23 base crash: INFO: task hung in __iterate_supers 2025/08/19 03:25:23 runner 4 connected 2025/08/19 03:25:57 patched crashed: possible deadlock in ocfs2_xattr_set [need repro = false] 2025/08/19 03:25:58 runner 8 connected 2025/08/19 03:26:07 patched crashed: possible deadlock in ocfs2_xattr_set [need repro = false] 2025/08/19 03:26:11 runner 2 connected 2025/08/19 03:26:53 runner 7 connected 2025/08/19 03:26:56 runner 2 connected 2025/08/19 03:27:04 patched crashed: possible deadlock in ocfs2_xattr_set [need repro = false] 2025/08/19 03:27:42 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 03:27:45 patched crashed: kernel BUG in txUnlock [need repro = true] 2025/08/19 03:27:45 scheduled a reproduction of 'kernel BUG in txUnlock' 2025/08/19 03:27:46 patched crashed: kernel BUG in txUnlock [need repro = true] 2025/08/19 03:27:46 scheduled a reproduction of 'kernel BUG in txUnlock' 2025/08/19 03:27:47 patched crashed: kernel BUG in txUnlock [need repro = true] 2025/08/19 03:27:47 scheduled a reproduction of 'kernel BUG in txUnlock' 2025/08/19 03:27:48 patched crashed: kernel BUG in txUnlock [need repro = true] 2025/08/19 03:27:48 scheduled a reproduction of 'kernel BUG in txUnlock' 2025/08/19 03:28:00 runner 1 connected 2025/08/19 03:28:13 base crash: kernel BUG in txUnlock 2025/08/19 03:28:25 base crash: lost connection to test machine 2025/08/19 03:28:27 base crash: kernel BUG in txUnlock 2025/08/19 03:28:30 runner 4 connected 2025/08/19 03:28:33 runner 7 connected 2025/08/19 03:28:35 runner 8 connected 2025/08/19 03:28:35 patched crashed: possible deadlock in ocfs2_xattr_set [need repro = false] 2025/08/19 03:28:36 runner 3 connected 2025/08/19 03:28:36 new: boot error: can't ssh into the instance 2025/08/19 03:28:37 runner 6 connected 2025/08/19 03:29:02 runner 2 connected 2025/08/19 03:29:09 patched crashed: WARNING: suspicious RCU usage in get_callchain_entry [need repro = true] 2025/08/19 03:29:09 scheduled a reproduction of 'WARNING: suspicious RCU usage in get_callchain_entry' 2025/08/19 03:29:13 runner 0 connected 2025/08/19 03:29:15 runner 1 connected 2025/08/19 03:29:23 patched crashed: possible deadlock in ocfs2_xattr_set [need repro = false] 2025/08/19 03:29:25 runner 0 connected 2025/08/19 03:29:26 runner 2 connected 2025/08/19 03:29:34 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 03:29:34 STAT { "buffer too small": 0, "candidate triage jobs": 26, "candidates": 39620, "comps overflows": 0, "corpus": 40882, "corpus [files]": 6426, "corpus [symbols]": 16146, "cover overflows": 32452, "coverage": 304331, "distributor delayed": 61786, "distributor undelayed": 61786, "distributor violated": 1914, "exec candidate": 41530, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 10, "exec seeds": 0, "exec smash": 0, "exec total [base]": 127962, "exec total [new]": 229856, "exec triage": 129191, "executor restarts": 741, "fault jobs": 0, "fuzzer jobs": 26, "fuzzing VMs [base]": 4, "fuzzing VMs [new]": 7, "hints jobs": 0, "max signal": 307634, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 0, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 41530, "no exec duration": 48038000000, "no exec requests": 428, "pending": 35, "prog exec time": 332, "reproducing": 0, "rpc recv": 8380889444, "rpc sent": 1608692816, "signal": 299562, "smash jobs": 0, "triage jobs": 0, "vm output": 29621190, "vm restarts [base]": 41, "vm restarts [new]": 116 } 2025/08/19 03:29:47 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 03:29:58 runner 9 connected 2025/08/19 03:30:08 patched crashed: possible deadlock in ocfs2_write_begin_nolock [need repro = true] 2025/08/19 03:30:08 scheduled a reproduction of 'possible deadlock in ocfs2_write_begin_nolock' 2025/08/19 03:30:12 runner 5 connected 2025/08/19 03:30:13 patched crashed: possible deadlock in ocfs2_try_remove_refcount_tree [need repro = false] 2025/08/19 03:30:14 base crash: kernel BUG in txUnlock 2025/08/19 03:30:21 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:30:24 runner 7 connected 2025/08/19 03:30:33 patched crashed: possible deadlock in ocfs2_try_remove_refcount_tree [need repro = false] 2025/08/19 03:30:37 runner 6 connected 2025/08/19 03:30:43 base crash: possible deadlock in ocfs2_try_remove_refcount_tree 2025/08/19 03:30:52 base crash: possible deadlock in ocfs2_try_remove_refcount_tree 2025/08/19 03:30:52 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:30:59 runner 0 connected 2025/08/19 03:31:02 runner 1 connected 2025/08/19 03:31:04 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:31:04 runner 2 connected 2025/08/19 03:31:10 runner 8 connected 2025/08/19 03:31:21 runner 3 connected 2025/08/19 03:31:33 runner 1 connected 2025/08/19 03:31:40 runner 3 connected 2025/08/19 03:31:42 runner 5 connected 2025/08/19 03:31:47 base crash: possible deadlock in ocfs2_init_acl 2025/08/19 03:31:54 runner 2 connected 2025/08/19 03:31:59 patched crashed: lost connection to test machine [need repro = false] 2025/08/19 03:32:07 patched crashed: possible deadlock in ocfs2_setattr [need repro = false] 2025/08/19 03:32:19 patched crashed: possible deadlock in ocfs2_xattr_set [need repro = false] 2025/08/19 03:32:44 runner 2 connected 2025/08/19 03:32:48 runner 4 connected 2025/08/19 03:32:57 runner 8 connected 2025/08/19 03:33:01 base crash: possible deadlock in ocfs2_reserve_suballoc_bits 2025/08/19 03:33:01 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:33:08 patched crashed: possible deadlock in ocfs2_xattr_set [need repro = false] 2025/08/19 03:33:12 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:33:12 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:33:23 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:33:27 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:33:38 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:33:39 base crash: possible deadlock in ocfs2_xattr_set 2025/08/19 03:33:49 runner 3 connected 2025/08/19 03:33:49 runner 5 connected 2025/08/19 03:33:57 runner 3 connected 2025/08/19 03:34:00 runner 7 connected 2025/08/19 03:34:00 runner 6 connected 2025/08/19 03:34:07 patched crashed: possible deadlock in ocfs2_try_remove_refcount_tree [need repro = false] 2025/08/19 03:34:08 base crash: possible deadlock in ocfs2_xattr_set 2025/08/19 03:34:13 runner 4 connected 2025/08/19 03:34:15 runner 8 connected 2025/08/19 03:34:27 runner 9 connected 2025/08/19 03:34:29 runner 0 connected 2025/08/19 03:34:33 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:34:34 STAT { "buffer too small": 0, "candidate triage jobs": 35, "candidates": 38205, "comps overflows": 0, "corpus": 42249, "corpus [files]": 6714, "corpus [symbols]": 16647, "cover overflows": 34147, "coverage": 307194, "distributor delayed": 63694, "distributor undelayed": 63685, "distributor violated": 1915, "exec candidate": 42945, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 10, "exec seeds": 0, "exec smash": 0, "exec total [base]": 134296, "exec total [new]": 242520, "exec triage": 133455, "executor restarts": 849, "fault jobs": 0, "fuzzer jobs": 35, "fuzzing VMs [base]": 2, "fuzzing VMs [new]": 2, "hints jobs": 0, "max signal": 310629, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 0, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 42945, "no exec duration": 48259000000, "no exec requests": 429, "pending": 36, "prog exec time": 282, "reproducing": 0, "rpc recv": 9287923532, "rpc sent": 1760966408, "signal": 302479, "smash jobs": 0, "triage jobs": 0, "vm output": 32937864, "vm restarts [base]": 47, "vm restarts [new]": 135 } 2025/08/19 03:34:36 patched crashed: possible deadlock in ocfs2_try_remove_refcount_tree [need repro = false] 2025/08/19 03:34:38 patched crashed: WARNING in ext4_xattr_inode_lookup_create [need repro = true] 2025/08/19 03:34:38 scheduled a reproduction of 'WARNING in ext4_xattr_inode_lookup_create' 2025/08/19 03:34:38 patched crashed: WARNING in ext4_xattr_inode_lookup_create [need repro = true] 2025/08/19 03:34:38 scheduled a reproduction of 'WARNING in ext4_xattr_inode_lookup_create' 2025/08/19 03:34:40 patched crashed: WARNING in ext4_xattr_inode_lookup_create [need repro = true] 2025/08/19 03:34:40 scheduled a reproduction of 'WARNING in ext4_xattr_inode_lookup_create' 2025/08/19 03:34:40 patched crashed: WARNING in ext4_xattr_inode_lookup_create [need repro = true] 2025/08/19 03:34:40 scheduled a reproduction of 'WARNING in ext4_xattr_inode_lookup_create' 2025/08/19 03:34:47 patched crashed: possible deadlock in ocfs2_try_remove_refcount_tree [need repro = false] 2025/08/19 03:34:56 runner 2 connected 2025/08/19 03:34:57 base crash: possible deadlock in ocfs2_try_remove_refcount_tree 2025/08/19 03:35:22 runner 1 connected 2025/08/19 03:35:25 runner 3 connected 2025/08/19 03:35:25 base crash: possible deadlock in ocfs2_init_acl 2025/08/19 03:35:27 runner 4 connected 2025/08/19 03:35:27 runner 6 connected 2025/08/19 03:35:28 runner 7 connected 2025/08/19 03:35:36 runner 8 connected 2025/08/19 03:35:45 runner 3 connected 2025/08/19 03:35:47 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 03:35:59 base crash: WARNING in xfrm6_tunnel_net_exit 2025/08/19 03:36:15 runner 0 connected 2025/08/19 03:36:36 runner 9 connected 2025/08/19 03:36:48 runner 2 connected 2025/08/19 03:37:46 base crash: possible deadlock in ocfs2_reserve_suballoc_bits 2025/08/19 03:37:47 patched crashed: kernel BUG in txUnlock [need repro = false] 2025/08/19 03:37:48 patched crashed: kernel BUG in txUnlock [need repro = false] 2025/08/19 03:37:49 patched crashed: kernel BUG in txUnlock [need repro = false] 2025/08/19 03:37:49 patched crashed: kernel BUG in txUnlock [need repro = false] 2025/08/19 03:37:53 patched crashed: possible deadlock in ocfs2_try_remove_refcount_tree [need repro = false] 2025/08/19 03:38:34 runner 0 connected 2025/08/19 03:38:37 runner 6 connected 2025/08/19 03:38:38 runner 7 connected 2025/08/19 03:38:39 base crash: kernel BUG in txUnlock 2025/08/19 03:38:42 runner 2 connected 2025/08/19 03:38:50 patched crashed: WARNING in xfrm_state_fini [need repro = false] 2025/08/19 03:39:08 base crash: possible deadlock in ocfs2_reserve_suballoc_bits 2025/08/19 03:39:16 patched crashed: lost connection to test machine [need repro = false] 2025/08/19 03:39:20 runner 2 connected 2025/08/19 03:39:34 STAT { "buffer too small": 0, "candidate triage jobs": 14, "candidates": 36894, "comps overflows": 0, "corpus": 43546, "corpus [files]": 6980, "corpus [symbols]": 17121, "cover overflows": 35633, "coverage": 310014, "distributor delayed": 65583, "distributor undelayed": 65581, "distributor violated": 1987, "exec candidate": 44256, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 14, "exec seeds": 0, "exec smash": 0, "exec total [base]": 138981, "exec total [new]": 254440, "exec triage": 137493, "executor restarts": 923, "fault jobs": 0, "fuzzer jobs": 14, "fuzzing VMs [base]": 2, "fuzzing VMs [new]": 4, "hints jobs": 0, "max signal": 313439, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 0, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 44256, "no exec duration": 48364000000, "no exec requests": 431, "pending": 40, "prog exec time": 245, "reproducing": 0, "rpc recv": 9972409564, "rpc sent": 1868247136, "signal": 305326, "smash jobs": 0, "triage jobs": 0, "vm output": 36151805, "vm restarts [base]": 52, "vm restarts [new]": 146 } 2025/08/19 03:39:39 runner 9 connected 2025/08/19 03:39:44 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 03:39:58 runner 3 connected 2025/08/19 03:40:04 runner 6 connected 2025/08/19 03:40:10 base crash: kernel BUG in txUnlock 2025/08/19 03:40:21 patched crashed: possible deadlock in ocfs2_try_remove_refcount_tree [need repro = false] 2025/08/19 03:40:23 base crash: kernel BUG in txUnlock 2025/08/19 03:40:33 runner 7 connected 2025/08/19 03:40:40 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 03:40:58 base crash: possible deadlock in ocfs2_reserve_suballoc_bits 2025/08/19 03:41:00 runner 0 connected 2025/08/19 03:41:12 runner 3 connected 2025/08/19 03:41:30 runner 9 connected 2025/08/19 03:41:43 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 03:41:47 runner 2 connected 2025/08/19 03:41:54 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 03:42:25 new: boot error: can't ssh into the instance 2025/08/19 03:42:32 runner 4 connected 2025/08/19 03:42:34 base crash: possible deadlock in ocfs2_reserve_suballoc_bits 2025/08/19 03:42:44 runner 8 connected 2025/08/19 03:42:52 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 03:43:06 runner 0 connected 2025/08/19 03:43:08 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:43:11 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:43:17 base crash: possible deadlock in ocfs2_reserve_suballoc_bits 2025/08/19 03:43:22 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:43:23 runner 3 connected 2025/08/19 03:43:31 patched crashed: WARNING in xfrm_state_fini [need repro = false] 2025/08/19 03:43:41 runner 9 connected 2025/08/19 03:43:49 runner 7 connected 2025/08/19 03:44:00 runner 6 connected 2025/08/19 03:44:00 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:44:04 runner 4 connected 2025/08/19 03:44:14 base: boot error: can't ssh into the instance 2025/08/19 03:44:19 runner 8 connected 2025/08/19 03:44:20 base crash: possible deadlock in ocfs2_init_acl 2025/08/19 03:44:34 STAT { "buffer too small": 0, "candidate triage jobs": 10, "candidates": 36077, "comps overflows": 0, "corpus": 44336, "corpus [files]": 7161, "corpus [symbols]": 17419, "cover overflows": 37037, "coverage": 312113, "distributor delayed": 67274, "distributor undelayed": 67274, "distributor violated": 2071, "exec candidate": 45073, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 14, "exec seeds": 0, "exec smash": 0, "exec total [base]": 143838, "exec total [new]": 263378, "exec triage": 140022, "executor restarts": 989, "fault jobs": 0, "fuzzer jobs": 10, "fuzzing VMs [base]": 1, "fuzzing VMs [new]": 5, "hints jobs": 0, "max signal": 315563, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 0, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 45073, "no exec duration": 48364000000, "no exec requests": 431, "pending": 40, "prog exec time": 321, "reproducing": 0, "rpc recv": 10589589636, "rpc sent": 1970067272, "signal": 307480, "smash jobs": 0, "triage jobs": 0, "vm output": 37985007, "vm restarts [base]": 57, "vm restarts [new]": 158 } 2025/08/19 03:44:46 new: boot error: can't ssh into the instance 2025/08/19 03:44:49 runner 0 connected 2025/08/19 03:45:02 runner 1 connected 2025/08/19 03:45:34 runner 5 connected 2025/08/19 03:45:34 base crash: INFO: task hung in v9fs_evict_inode 2025/08/19 03:45:40 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:45:50 base crash: WARNING in xfrm_state_fini 2025/08/19 03:45:54 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 03:45:59 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:46:06 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 03:46:15 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:46:17 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 03:46:22 runner 0 connected 2025/08/19 03:46:29 runner 9 connected 2025/08/19 03:46:38 runner 1 connected 2025/08/19 03:46:40 runner 6 connected 2025/08/19 03:46:44 runner 8 connected 2025/08/19 03:46:47 runner 7 connected 2025/08/19 03:46:56 runner 5 connected 2025/08/19 03:47:05 runner 0 connected 2025/08/19 03:47:35 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 03:47:41 patched crashed: possible deadlock in ocfs2_try_remove_refcount_tree [need repro = false] 2025/08/19 03:47:52 patched crashed: possible deadlock in ocfs2_try_remove_refcount_tree [need repro = false] 2025/08/19 03:47:52 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:47:53 new: boot error: can't ssh into the instance 2025/08/19 03:47:54 new: boot error: can't ssh into the instance 2025/08/19 03:48:08 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 03:48:24 runner 8 connected 2025/08/19 03:48:29 runner 4 connected 2025/08/19 03:48:40 runner 7 connected 2025/08/19 03:48:41 runner 6 connected 2025/08/19 03:48:42 runner 3 connected 2025/08/19 03:48:43 runner 1 connected 2025/08/19 03:49:18 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 03:49:28 patched crashed: possible deadlock in ocfs2_xattr_set [need repro = false] 2025/08/19 03:49:29 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 03:49:34 STAT { "buffer too small": 0, "candidate triage jobs": 11, "candidates": 35427, "comps overflows": 0, "corpus": 44926, "corpus [files]": 7316, "corpus [symbols]": 17671, "cover overflows": 39105, "coverage": 313332, "distributor delayed": 68245, "distributor undelayed": 68245, "distributor violated": 2078, "exec candidate": 45723, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 14, "exec seeds": 0, "exec smash": 0, "exec total [base]": 147962, "exec total [new]": 275771, "exec triage": 141935, "executor restarts": 1065, "fault jobs": 0, "fuzzer jobs": 11, "fuzzing VMs [base]": 1, "fuzzing VMs [new]": 5, "hints jobs": 0, "max signal": 316882, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 0, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 45693, "no exec duration": 48364000000, "no exec requests": 431, "pending": 40, "prog exec time": 332, "reproducing": 0, "rpc recv": 11194463448, "rpc sent": 2085620392, "signal": 308692, "smash jobs": 0, "triage jobs": 0, "vm output": 40512114, "vm restarts [base]": 60, "vm restarts [new]": 172 } 2025/08/19 03:49:39 base crash: possible deadlock in ocfs2_reserve_suballoc_bits 2025/08/19 03:49:51 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:50:02 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:50:06 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:50:06 runner 4 connected 2025/08/19 03:50:18 runner 7 connected 2025/08/19 03:50:18 runner 9 connected 2025/08/19 03:50:27 new: boot error: can't ssh into the instance 2025/08/19 03:50:29 runner 1 connected 2025/08/19 03:50:38 base crash: WARNING in xfrm_state_fini 2025/08/19 03:50:40 runner 3 connected 2025/08/19 03:50:51 runner 8 connected 2025/08/19 03:50:54 runner 1 connected 2025/08/19 03:50:54 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 03:51:15 runner 2 connected 2025/08/19 03:51:16 base crash: possible deadlock in ocfs2_reserve_suballoc_bits 2025/08/19 03:51:28 runner 0 connected 2025/08/19 03:51:38 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 03:51:39 patched crashed: WARNING in xfrm_state_fini [need repro = false] 2025/08/19 03:51:44 runner 6 connected 2025/08/19 03:51:46 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:51:53 patched crashed: possible deadlock in ocfs2_try_remove_refcount_tree [need repro = false] 2025/08/19 03:52:04 patched crashed: possible deadlock in ocfs2_try_remove_refcount_tree [need repro = false] 2025/08/19 03:52:05 runner 1 connected 2025/08/19 03:52:10 patched crashed: WARNING: suspicious RCU usage in get_callchain_entry [need repro = true] 2025/08/19 03:52:10 scheduled a reproduction of 'WARNING: suspicious RCU usage in get_callchain_entry' 2025/08/19 03:52:21 patched crashed: WARNING: suspicious RCU usage in get_callchain_entry [need repro = true] 2025/08/19 03:52:21 scheduled a reproduction of 'WARNING: suspicious RCU usage in get_callchain_entry' 2025/08/19 03:52:28 runner 8 connected 2025/08/19 03:52:29 base crash: possible deadlock in ocfs2_reserve_suballoc_bits 2025/08/19 03:52:36 runner 9 connected 2025/08/19 03:52:40 runner 1 connected 2025/08/19 03:52:49 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 03:52:53 runner 7 connected 2025/08/19 03:52:59 runner 2 connected 2025/08/19 03:53:04 runner 6 connected 2025/08/19 03:53:17 runner 0 connected 2025/08/19 03:53:22 base: boot error: can't ssh into the instance 2025/08/19 03:53:38 runner 4 connected 2025/08/19 03:53:58 patched crashed: possible deadlock in ocfs2_setattr [need repro = false] 2025/08/19 03:54:03 patched crashed: WARNING in xfrm_state_fini [need repro = false] 2025/08/19 03:54:12 runner 2 connected 2025/08/19 03:54:26 base: boot error: can't ssh into the instance 2025/08/19 03:54:34 STAT { "buffer too small": 0, "candidate triage jobs": 12, "candidates": 34966, "comps overflows": 0, "corpus": 45299, "corpus [files]": 7379, "corpus [symbols]": 17798, "cover overflows": 41737, "coverage": 314100, "distributor delayed": 68833, "distributor undelayed": 68833, "distributor violated": 2078, "exec candidate": 46184, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 15, "exec seeds": 0, "exec smash": 0, "exec total [base]": 151294, "exec total [new]": 290102, "exec triage": 143210, "executor restarts": 1132, "fault jobs": 0, "fuzzer jobs": 12, "fuzzing VMs [base]": 3, "fuzzing VMs [new]": 5, "hints jobs": 0, "max signal": 317707, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 0, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 46097, "no exec duration": 48364000000, "no exec requests": 431, "pending": 42, "prog exec time": 298, "reproducing": 0, "rpc recv": 11867435640, "rpc sent": 2191525840, "signal": 309428, "smash jobs": 0, "triage jobs": 0, "vm output": 42842597, "vm restarts [base]": 65, "vm restarts [new]": 187 } 2025/08/19 03:54:43 patched crashed: possible deadlock in ocfs2_try_remove_refcount_tree [need repro = false] 2025/08/19 03:54:44 base crash: possible deadlock in ocfs2_reserve_suballoc_bits 2025/08/19 03:54:47 runner 2 connected 2025/08/19 03:54:52 runner 9 connected 2025/08/19 03:54:56 patched crashed: possible deadlock in ocfs2_try_remove_refcount_tree [need repro = false] 2025/08/19 03:55:04 base crash: possible deadlock in ocfs2_try_remove_refcount_tree 2025/08/19 03:55:15 runner 3 connected 2025/08/19 03:55:32 runner 6 connected 2025/08/19 03:55:34 runner 1 connected 2025/08/19 03:55:45 runner 1 connected 2025/08/19 03:55:53 runner 0 connected 2025/08/19 03:56:05 patched crashed: lost connection to test machine [need repro = false] 2025/08/19 03:56:11 patched crashed: possible deadlock in ocfs2_try_remove_refcount_tree [need repro = false] 2025/08/19 03:56:54 runner 3 connected 2025/08/19 03:57:08 runner 9 connected 2025/08/19 03:57:11 patched crashed: possible deadlock in ocfs2_try_remove_refcount_tree [need repro = false] 2025/08/19 03:57:17 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 03:57:28 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:57:32 base crash: possible deadlock in ocfs2_reserve_suballoc_bits 2025/08/19 03:57:39 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:57:43 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:57:48 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:57:50 patched crashed: possible deadlock in ocfs2_init_acl [need repro = false] 2025/08/19 03:58:01 runner 7 connected 2025/08/19 03:58:07 runner 1 connected 2025/08/19 03:58:14 new: boot error: can't ssh into the instance 2025/08/19 03:58:18 runner 6 connected 2025/08/19 03:58:20 runner 2 connected 2025/08/19 03:58:27 runner 4 connected 2025/08/19 03:58:30 base crash: possible deadlock in ocfs2_reserve_suballoc_bits 2025/08/19 03:58:37 runner 8 connected 2025/08/19 03:58:39 runner 2 connected 2025/08/19 03:58:40 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 03:58:47 base crash: possible deadlock in ocfs2_init_acl 2025/08/19 03:58:52 base crash: possible deadlock in ocfs2_try_remove_refcount_tree 2025/08/19 03:58:59 base crash: possible deadlock in ocfs2_init_acl 2025/08/19 03:59:03 runner 0 connected 2025/08/19 03:59:30 runner 1 connected 2025/08/19 03:59:34 timed out waiting for coprus triage 2025/08/19 03:59:34 starting bug reproductions 2025/08/19 03:59:34 starting bug reproductions (max 10 VMs, 7 repros) 2025/08/19 03:59:34 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 34560, "comps overflows": 0, "corpus": 45634, "corpus [files]": 7444, "corpus [symbols]": 17921, "cover overflows": 44367, "coverage": 314738, "distributor delayed": 69360, "distributor undelayed": 69360, "distributor violated": 2078, "exec candidate": 46590, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 17, "exec seeds": 0, "exec smash": 0, "exec total [base]": 158678, "exec total [new]": 304843, "exec triage": 144366, "executor restarts": 1198, "fault jobs": 0, "fuzzer jobs": 0, "fuzzing VMs [base]": 0, "fuzzing VMs [new]": 7, "hints jobs": 0, "max signal": 318386, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 0, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 46448, "no exec duration": 48364000000, "no exec requests": 431, "pending": 42, "prog exec time": 275, "reproducing": 0, "rpc recv": 12442020732, "rpc sent": 2305581152, "signal": 310083, "smash jobs": 0, "triage jobs": 0, "vm output": 45122579, "vm restarts [base]": 69, "vm restarts [new]": 201 } 2025/08/19 03:59:34 reproduction of "possible deadlock in ocfs2_reserve_suballoc_bits" aborted: it's no longer needed 2025/08/19 03:59:34 reproduction of "possible deadlock in ocfs2_try_remove_refcount_tree" aborted: it's no longer needed 2025/08/19 03:59:34 reproduction of "possible deadlock in ocfs2_reserve_suballoc_bits" aborted: it's no longer needed 2025/08/19 03:59:34 reproduction of "possible deadlock in ocfs2_init_acl" aborted: it's no longer needed 2025/08/19 03:59:34 reproduction of "possible deadlock in ocfs2_reserve_suballoc_bits" aborted: it's no longer needed 2025/08/19 03:59:34 reproduction of "possible deadlock in ocfs2_reserve_suballoc_bits" aborted: it's no longer needed 2025/08/19 03:59:34 reproduction of "possible deadlock in ocfs2_reserve_suballoc_bits" aborted: it's no longer needed 2025/08/19 03:59:34 reproduction of "possible deadlock in ocfs2_xattr_set" aborted: it's no longer needed 2025/08/19 03:59:34 reproduction of "possible deadlock in ocfs2_xattr_set" aborted: it's no longer needed 2025/08/19 03:59:34 reproduction of "WARNING in xfrm_state_fini" aborted: it's no longer needed 2025/08/19 03:59:34 start reproducing 'WARNING in mem_cgroup_migrate' 2025/08/19 03:59:34 start reproducing 'WARNING in drv_unassign_vif_chanctx' 2025/08/19 03:59:34 failed to recv *flatrpc.InfoRequestRawT: unexpected EOF 2025/08/19 03:59:34 start reproducing 'WARNING in __linkwatch_sync_dev' 2025/08/19 03:59:34 start reproducing 'possible deadlock in dqget' 2025/08/19 03:59:34 start reproducing 'KASAN: slab-use-after-free Read in xfrm_state_find' 2025/08/19 03:59:34 reproduction of "WARNING in xfrm_state_fini" aborted: it's no longer needed 2025/08/19 03:59:34 reproduction of "possible deadlock in ocfs2_setattr" aborted: it's no longer needed 2025/08/19 03:59:34 reproduction of "general protection fault in pcl818_ai_cancel" aborted: it's no longer needed 2025/08/19 03:59:34 start reproducing 'WARNING: suspicious RCU usage in get_callchain_entry' 2025/08/19 03:59:34 reproduction of "WARNING in xfrm6_tunnel_net_exit" aborted: it's no longer needed 2025/08/19 03:59:34 reproduction of "possible deadlock in ocfs2_xattr_set" aborted: it's no longer needed 2025/08/19 03:59:34 reproduction of "possible deadlock in ocfs2_xattr_set" aborted: it's no longer needed 2025/08/19 03:59:34 reproduction of "possible deadlock in ocfs2_xattr_set" aborted: it's no longer needed 2025/08/19 03:59:34 reproduction of "possible deadlock in ocfs2_xattr_set" aborted: it's no longer needed 2025/08/19 03:59:34 start reproducing 'INFO: task hung in corrupted' 2025/08/19 03:59:35 runner 2 connected 2025/08/19 03:59:40 runner 3 connected 2025/08/19 03:59:48 runner 0 connected 2025/08/19 04:00:08 base crash: possible deadlock in ocfs2_init_acl 2025/08/19 04:00:52 reproducing crash 'WARNING: suspicious RCU usage in get_callchain_entry': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f kernel/events/callchain.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 04:01:43 new: boot error: can't ssh into the instance 2025/08/19 04:02:03 base crash: possible deadlock in ocfs2_reserve_suballoc_bits 2025/08/19 04:02:52 runner 2 connected 2025/08/19 04:04:34 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 34560, "comps overflows": 0, "corpus": 45634, "corpus [files]": 7444, "corpus [symbols]": 17921, "cover overflows": 44367, "coverage": 314738, "distributor delayed": 69360, "distributor undelayed": 69360, "distributor violated": 2078, "exec candidate": 46590, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 17, "exec seeds": 0, "exec smash": 0, "exec total [base]": 162786, "exec total [new]": 304843, "exec triage": 144366, "executor restarts": 1198, "fault jobs": 0, "fuzzer jobs": 0, "fuzzing VMs [base]": 2, "fuzzing VMs [new]": 0, "hints jobs": 0, "max signal": 318386, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 3, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 46448, "no exec duration": 48369000000, "no exec requests": 432, "pending": 16, "prog exec time": 0, "reproducing": 7, "rpc recv": 12566665676, "rpc sent": 2320520984, "signal": 310083, "smash jobs": 0, "triage jobs": 0, "vm output": 46767755, "vm restarts [base]": 73, "vm restarts [new]": 201 } 2025/08/19 04:07:48 base crash: no output from test machine 2025/08/19 04:07:49 new: boot error: can't ssh into the instance 2025/08/19 04:07:52 base crash: no output from test machine 2025/08/19 04:08:29 runner 0 connected 2025/08/19 04:08:36 base: boot error: can't ssh into the instance 2025/08/19 04:08:41 runner 2 connected 2025/08/19 04:08:53 reproducing crash 'WARNING: suspicious RCU usage in get_callchain_entry': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f kernel/events/callchain.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 04:09:17 runner 1 connected 2025/08/19 04:09:27 reproducing crash 'WARNING in mem_cgroup_migrate': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/memcontrol.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 04:09:34 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 34560, "comps overflows": 0, "corpus": 45634, "corpus [files]": 7444, "corpus [symbols]": 17921, "cover overflows": 44367, "coverage": 314738, "distributor delayed": 69360, "distributor undelayed": 69360, "distributor violated": 2078, "exec candidate": 46590, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 17, "exec seeds": 0, "exec smash": 0, "exec total [base]": 162786, "exec total [new]": 304843, "exec triage": 144366, "executor restarts": 1198, "fault jobs": 0, "fuzzer jobs": 0, "fuzzing VMs [base]": 3, "fuzzing VMs [new]": 0, "hints jobs": 0, "max signal": 318386, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 5, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 46448, "no exec duration": 48369000000, "no exec requests": 432, "pending": 16, "prog exec time": 0, "reproducing": 7, "rpc recv": 12659353860, "rpc sent": 2320521824, "signal": 310083, "smash jobs": 0, "triage jobs": 0, "vm output": 48986915, "vm restarts [base]": 76, "vm restarts [new]": 201 } 2025/08/19 04:09:40 new: boot error: can't ssh into the instance 2025/08/19 04:09:40 new: boot error: can't ssh into the instance 2025/08/19 04:10:14 base: boot error: can't ssh into the instance 2025/08/19 04:10:15 reproducing crash 'WARNING: suspicious RCU usage in get_callchain_entry': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f kernel/events/callchain.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 04:10:55 runner 3 connected 2025/08/19 04:11:18 reproducing crash 'WARNING: suspicious RCU usage in get_callchain_entry': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f kernel/events/callchain.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 04:11:38 reproducing crash 'WARNING in mem_cgroup_migrate': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/memcontrol.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 04:12:02 reproducing crash 'WARNING: suspicious RCU usage in get_callchain_entry': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f kernel/events/callchain.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 04:12:02 repro finished 'WARNING: suspicious RCU usage in get_callchain_entry', repro=true crepro=false desc='WARNING: suspicious RCU usage in get_callchain_entry' hub=false from_dashboard=false 2025/08/19 04:12:02 found repro for "WARNING: suspicious RCU usage in get_callchain_entry" (orig title: "-SAME-", reliability: 1), took 12.45 minutes 2025/08/19 04:12:02 start reproducing 'possible deadlock in ntfs_fiemap' 2025/08/19 04:12:02 "WARNING: suspicious RCU usage in get_callchain_entry": saved crash log into 1755576722.crash.log 2025/08/19 04:12:02 "WARNING: suspicious RCU usage in get_callchain_entry": saved repro log into 1755576722.repro.log 2025/08/19 04:12:55 reproducing crash 'WARNING in mem_cgroup_migrate': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/memcontrol.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 04:13:14 attempt #0 to run "WARNING: suspicious RCU usage in get_callchain_entry" on base: crashed with WARNING: suspicious RCU usage in get_callchain_entry 2025/08/19 04:13:14 crashes both: WARNING: suspicious RCU usage in get_callchain_entry / WARNING: suspicious RCU usage in get_callchain_entry 2025/08/19 04:13:40 base crash: no output from test machine 2025/08/19 04:13:47 repro finished 'possible deadlock in dqget', repro=false crepro=false desc='' hub=false from_dashboard=false 2025/08/19 04:13:47 reproduction of "kernel BUG in txUnlock" aborted: it's no longer needed 2025/08/19 04:13:47 reproduction of "kernel BUG in txUnlock" aborted: it's no longer needed 2025/08/19 04:13:47 failed repro for "possible deadlock in dqget", err=%!s() 2025/08/19 04:13:47 reproduction of "kernel BUG in txUnlock" aborted: it's no longer needed 2025/08/19 04:13:47 reproduction of "kernel BUG in txUnlock" aborted: it's no longer needed 2025/08/19 04:13:47 start reproducing 'INFO: task hung in sync_bdevs' 2025/08/19 04:13:47 "possible deadlock in dqget": saved crash log into 1755576827.crash.log 2025/08/19 04:13:47 "possible deadlock in dqget": saved repro log into 1755576827.repro.log 2025/08/19 04:13:55 runner 0 connected 2025/08/19 04:14:17 base crash: no output from test machine 2025/08/19 04:14:28 runner 2 connected 2025/08/19 04:14:34 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 34560, "comps overflows": 0, "corpus": 45634, "corpus [files]": 7444, "corpus [symbols]": 17921, "cover overflows": 44367, "coverage": 314738, "distributor delayed": 69360, "distributor undelayed": 69360, "distributor violated": 2078, "exec candidate": 46590, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 17, "exec seeds": 0, "exec smash": 0, "exec total [base]": 162786, "exec total [new]": 304843, "exec triage": 144366, "executor restarts": 1198, "fault jobs": 0, "fuzzer jobs": 0, "fuzzing VMs [base]": 2, "fuzzing VMs [new]": 0, "hints jobs": 0, "max signal": 318386, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 5, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 46448, "no exec duration": 48369000000, "no exec requests": 432, "pending": 10, "prog exec time": 0, "reproducing": 7, "rpc recv": 12721146144, "rpc sent": 2320522648, "signal": 310083, "smash jobs": 0, "triage jobs": 0, "vm output": 51712350, "vm restarts [base]": 79, "vm restarts [new]": 201 } 2025/08/19 04:14:59 runner 1 connected 2025/08/19 04:15:04 new: boot error: can't ssh into the instance 2025/08/19 04:15:55 base crash: no output from test machine 2025/08/19 04:16:36 runner 3 connected 2025/08/19 04:17:24 new: boot error: can't ssh into the instance 2025/08/19 04:18:55 base crash: no output from test machine 2025/08/19 04:19:28 base crash: no output from test machine 2025/08/19 04:19:34 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 34560, "comps overflows": 0, "corpus": 45634, "corpus [files]": 7444, "corpus [symbols]": 17921, "cover overflows": 44367, "coverage": 314738, "distributor delayed": 69360, "distributor undelayed": 69360, "distributor violated": 2078, "exec candidate": 46590, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 17, "exec seeds": 0, "exec smash": 0, "exec total [base]": 162786, "exec total [new]": 304843, "exec triage": 144366, "executor restarts": 1198, "fault jobs": 0, "fuzzer jobs": 0, "fuzzing VMs [base]": 2, "fuzzing VMs [new]": 0, "hints jobs": 0, "max signal": 318386, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 5, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 46448, "no exec duration": 48369000000, "no exec requests": 432, "pending": 10, "prog exec time": 0, "reproducing": 7, "rpc recv": 12813834172, "rpc sent": 2320523224, "signal": 310083, "smash jobs": 0, "triage jobs": 0, "vm output": 56237637, "vm restarts [base]": 81, "vm restarts [new]": 201 } 2025/08/19 04:19:43 runner 0 connected 2025/08/19 04:19:58 base crash: no output from test machine 2025/08/19 04:20:48 runner 1 connected 2025/08/19 04:21:36 base crash: no output from test machine 2025/08/19 04:22:25 runner 3 connected 2025/08/19 04:23:48 reproducing crash 'WARNING in mem_cgroup_migrate': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/memcontrol.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 04:24:34 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 34560, "comps overflows": 0, "corpus": 45634, "corpus [files]": 7444, "corpus [symbols]": 17921, "cover overflows": 44367, "coverage": 314738, "distributor delayed": 69360, "distributor undelayed": 69360, "distributor violated": 2078, "exec candidate": 46590, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 17, "exec seeds": 0, "exec smash": 0, "exec total [base]": 162786, "exec total [new]": 304843, "exec triage": 144366, "executor restarts": 1198, "fault jobs": 0, "fuzzer jobs": 0, "fuzzing VMs [base]": 3, "fuzzing VMs [new]": 0, "hints jobs": 0, "max signal": 318386, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 5, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 46448, "no exec duration": 48369000000, "no exec requests": 432, "pending": 10, "prog exec time": 0, "reproducing": 7, "rpc recv": 12906522356, "rpc sent": 2320524064, "signal": 310083, "smash jobs": 0, "triage jobs": 0, "vm output": 60357983, "vm restarts [base]": 84, "vm restarts [new]": 201 } 2025/08/19 04:24:35 repro finished 'possible deadlock in ntfs_fiemap', repro=false crepro=false desc='' hub=false from_dashboard=false 2025/08/19 04:24:35 failed repro for "possible deadlock in ntfs_fiemap", err=%!s() 2025/08/19 04:24:35 start reproducing 'possible deadlock in ocfs2_write_begin_nolock' 2025/08/19 04:24:35 "possible deadlock in ntfs_fiemap": saved crash log into 1755577475.crash.log 2025/08/19 04:24:35 "possible deadlock in ntfs_fiemap": saved repro log into 1755577475.repro.log 2025/08/19 04:24:43 base crash: no output from test machine 2025/08/19 04:25:31 runner 0 connected 2025/08/19 04:25:47 base crash: no output from test machine 2025/08/19 04:25:50 new: boot error: can't ssh into the instance 2025/08/19 04:26:29 runner 1 connected 2025/08/19 04:27:24 base crash: no output from test machine 2025/08/19 04:28:07 runner 3 connected 2025/08/19 04:29:34 base: boot error: can't ssh into the instance 2025/08/19 04:29:34 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 34560, "comps overflows": 0, "corpus": 45634, "corpus [files]": 7444, "corpus [symbols]": 17921, "cover overflows": 44367, "coverage": 314738, "distributor delayed": 69360, "distributor undelayed": 69360, "distributor violated": 2078, "exec candidate": 46590, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 17, "exec seeds": 0, "exec smash": 0, "exec total [base]": 162786, "exec total [new]": 304843, "exec triage": 144366, "executor restarts": 1198, "fault jobs": 0, "fuzzer jobs": 0, "fuzzing VMs [base]": 3, "fuzzing VMs [new]": 0, "hints jobs": 0, "max signal": 318386, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 5, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 46448, "no exec duration": 48369000000, "no exec requests": 432, "pending": 9, "prog exec time": 0, "reproducing": 7, "rpc recv": 12999210540, "rpc sent": 2320524904, "signal": 310083, "smash jobs": 0, "triage jobs": 0, "vm output": 63451370, "vm restarts [base]": 87, "vm restarts [new]": 201 } 2025/08/19 04:30:16 runner 2 connected 2025/08/19 04:30:25 reproducing crash 'WARNING in mem_cgroup_migrate': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/memcontrol.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 04:30:31 base crash: no output from test machine 2025/08/19 04:31:13 runner 0 connected 2025/08/19 04:31:28 base crash: no output from test machine 2025/08/19 04:31:38 reproducing crash 'WARNING in mem_cgroup_migrate': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/memcontrol.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 04:32:10 runner 1 connected 2025/08/19 04:32:31 reproducing crash 'WARNING in mem_cgroup_migrate': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/memcontrol.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 04:33:06 base crash: no output from test machine 2025/08/19 04:33:47 runner 3 connected 2025/08/19 04:34:24 new: boot error: can't ssh into the instance 2025/08/19 04:34:27 reproducing crash 'WARNING in mem_cgroup_migrate': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/memcontrol.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 04:34:34 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 34560, "comps overflows": 0, "corpus": 45634, "corpus [files]": 7444, "corpus [symbols]": 17921, "cover overflows": 44367, "coverage": 314738, "distributor delayed": 69360, "distributor undelayed": 69360, "distributor violated": 2078, "exec candidate": 46590, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 17, "exec seeds": 0, "exec smash": 0, "exec total [base]": 162786, "exec total [new]": 304843, "exec triage": 144366, "executor restarts": 1198, "fault jobs": 0, "fuzzer jobs": 0, "fuzzing VMs [base]": 4, "fuzzing VMs [new]": 0, "hints jobs": 0, "max signal": 318386, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 8, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 46448, "no exec duration": 48369000000, "no exec requests": 432, "pending": 9, "prog exec time": 0, "reproducing": 7, "rpc recv": 13122794788, "rpc sent": 2320526024, "signal": 310083, "smash jobs": 0, "triage jobs": 0, "vm output": 65652549, "vm restarts [base]": 91, "vm restarts [new]": 201 } 2025/08/19 04:35:16 base crash: no output from test machine 2025/08/19 04:36:04 runner 2 connected 2025/08/19 04:36:13 base crash: no output from test machine 2025/08/19 04:36:34 repro finished 'possible deadlock in ocfs2_write_begin_nolock', repro=false crepro=false desc='' hub=false from_dashboard=false 2025/08/19 04:36:34 failed repro for "possible deadlock in ocfs2_write_begin_nolock", err=%!s() 2025/08/19 04:36:34 start reproducing 'WARNING in ext4_xattr_inode_lookup_create' 2025/08/19 04:36:34 reproduction of "WARNING: suspicious RCU usage in get_callchain_entry" aborted: it's no longer needed 2025/08/19 04:36:34 reproduction of "WARNING: suspicious RCU usage in get_callchain_entry" aborted: it's no longer needed 2025/08/19 04:36:34 reproduction of "WARNING: suspicious RCU usage in get_callchain_entry" aborted: it's no longer needed 2025/08/19 04:36:34 reproduction of "WARNING: suspicious RCU usage in get_callchain_entry" aborted: it's no longer needed 2025/08/19 04:36:34 "possible deadlock in ocfs2_write_begin_nolock": saved crash log into 1755578194.crash.log 2025/08/19 04:36:34 "possible deadlock in ocfs2_write_begin_nolock": saved repro log into 1755578194.repro.log 2025/08/19 04:36:54 runner 0 connected 2025/08/19 04:37:09 base crash: no output from test machine 2025/08/19 04:37:35 reproducing crash 'WARNING in ext4_xattr_inode_lookup_create': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f fs/ext4/xattr.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 04:37:52 runner 1 connected 2025/08/19 04:38:47 base crash: no output from test machine 2025/08/19 04:39:29 runner 3 connected 2025/08/19 04:39:34 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 34560, "comps overflows": 0, "corpus": 45634, "corpus [files]": 7444, "corpus [symbols]": 17921, "cover overflows": 44367, "coverage": 314738, "distributor delayed": 69360, "distributor undelayed": 69360, "distributor violated": 2078, "exec candidate": 46590, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 17, "exec seeds": 0, "exec smash": 0, "exec total [base]": 162786, "exec total [new]": 304843, "exec triage": 144366, "executor restarts": 1198, "fault jobs": 0, "fuzzer jobs": 0, "fuzzing VMs [base]": 3, "fuzzing VMs [new]": 0, "hints jobs": 0, "max signal": 318386, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 11, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 46448, "no exec duration": 48369000000, "no exec requests": 432, "pending": 5, "prog exec time": 0, "reproducing": 7, "rpc recv": 13215483136, "rpc sent": 2320527128, "signal": 310083, "smash jobs": 0, "triage jobs": 0, "vm output": 67756567, "vm restarts [base]": 95, "vm restarts [new]": 201 } 2025/08/19 04:39:49 new: boot error: can't ssh into the instance 2025/08/19 04:40:03 reproducing crash 'WARNING in ext4_xattr_inode_lookup_create': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f fs/ext4/xattr.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 04:40:27 reproducing crash 'WARNING in ext4_xattr_inode_lookup_create': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f fs/ext4/xattr.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 04:41:04 base crash: no output from test machine 2025/08/19 04:41:13 reproducing crash 'WARNING in ext4_xattr_inode_lookup_create': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f fs/ext4/xattr.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 04:41:13 repro finished 'WARNING in ext4_xattr_inode_lookup_create', repro=true crepro=false desc='WARNING in ext4_xattr_inode_lookup_create' hub=false from_dashboard=false 2025/08/19 04:41:13 found repro for "WARNING in ext4_xattr_inode_lookup_create" (orig title: "-SAME-", reliability: 1), took 4.66 minutes 2025/08/19 04:41:13 start reproducing 'WARNING in ext4_xattr_inode_lookup_create' 2025/08/19 04:41:13 "WARNING in ext4_xattr_inode_lookup_create": saved crash log into 1755578473.crash.log 2025/08/19 04:41:13 "WARNING in ext4_xattr_inode_lookup_create": saved repro log into 1755578473.repro.log 2025/08/19 04:41:52 runner 2 connected 2025/08/19 04:42:33 attempt #0 to run "WARNING in ext4_xattr_inode_lookup_create" on base: crashed with WARNING in ext4_xattr_inode_lookup_create 2025/08/19 04:42:33 crashes both: WARNING in ext4_xattr_inode_lookup_create / WARNING in ext4_xattr_inode_lookup_create 2025/08/19 04:42:43 reproducing crash 'WARNING in mem_cgroup_migrate': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/memcontrol.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 04:42:51 base crash: no output from test machine 2025/08/19 04:43:21 reproducing crash 'INFO: task hung in corrupted': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f drivers/net/netdevsim/bus.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 04:43:22 runner 0 connected 2025/08/19 04:43:39 runner 1 connected 2025/08/19 04:43:56 reproducing crash 'WARNING in mem_cgroup_migrate': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/memcontrol.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 04:44:28 base crash: no output from test machine 2025/08/19 04:44:34 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 34560, "comps overflows": 0, "corpus": 45634, "corpus [files]": 7444, "corpus [symbols]": 17921, "cover overflows": 44367, "coverage": 314738, "distributor delayed": 69360, "distributor undelayed": 69360, "distributor violated": 2078, "exec candidate": 46590, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 17, "exec seeds": 0, "exec smash": 0, "exec total [base]": 162786, "exec total [new]": 304843, "exec triage": 144366, "executor restarts": 1198, "fault jobs": 0, "fuzzer jobs": 0, "fuzzing VMs [base]": 3, "fuzzing VMs [new]": 0, "hints jobs": 0, "max signal": 318386, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 14, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 46448, "no exec duration": 48369000000, "no exec requests": 432, "pending": 4, "prog exec time": 0, "reproducing": 7, "rpc recv": 13339067220, "rpc sent": 2320527984, "signal": 310083, "smash jobs": 0, "triage jobs": 0, "vm output": 69513165, "vm restarts [base]": 98, "vm restarts [new]": 201 } 2025/08/19 04:44:55 reproducing crash 'WARNING in mem_cgroup_migrate': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/memcontrol.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 04:45:10 runner 3 connected 2025/08/19 04:45:52 repro finished 'WARNING in drv_unassign_vif_chanctx', repro=false crepro=false desc='' hub=false from_dashboard=false 2025/08/19 04:45:52 failed repro for "WARNING in drv_unassign_vif_chanctx", err=%!s() 2025/08/19 04:45:52 start reproducing 'WARNING in drv_unassign_vif_chanctx' 2025/08/19 04:45:52 "WARNING in drv_unassign_vif_chanctx": saved crash log into 1755578752.crash.log 2025/08/19 04:45:52 "WARNING in drv_unassign_vif_chanctx": saved repro log into 1755578752.repro.log 2025/08/19 04:46:52 base crash: no output from test machine 2025/08/19 04:47:40 runner 2 connected 2025/08/19 04:48:22 base crash: no output from test machine 2025/08/19 04:48:39 base crash: no output from test machine 2025/08/19 04:49:10 runner 0 connected 2025/08/19 04:49:27 runner 1 connected 2025/08/19 04:49:34 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 34560, "comps overflows": 0, "corpus": 45634, "corpus [files]": 7444, "corpus [symbols]": 17921, "cover overflows": 44367, "coverage": 314738, "distributor delayed": 69360, "distributor undelayed": 69360, "distributor violated": 2078, "exec candidate": 46590, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 17, "exec seeds": 0, "exec smash": 0, "exec total [base]": 162786, "exec total [new]": 304843, "exec triage": 144366, "executor restarts": 1198, "fault jobs": 0, "fuzzer jobs": 0, "fuzzing VMs [base]": 4, "fuzzing VMs [new]": 0, "hints jobs": 0, "max signal": 318386, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 15, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 46448, "no exec duration": 48369000000, "no exec requests": 432, "pending": 3, "prog exec time": 0, "reproducing": 7, "rpc recv": 13462651468, "rpc sent": 2320529104, "signal": 310083, "smash jobs": 0, "triage jobs": 0, "vm output": 72348742, "vm restarts [base]": 102, "vm restarts [new]": 201 } 2025/08/19 04:50:09 base crash: no output from test machine 2025/08/19 04:50:15 new: boot error: can't ssh into the instance 2025/08/19 04:50:50 runner 3 connected 2025/08/19 04:52:28 reproducing crash 'INFO: task hung in corrupted': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f drivers/base/core.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 04:52:39 base crash: no output from test machine 2025/08/19 04:53:21 runner 2 connected 2025/08/19 04:54:10 base crash: no output from test machine 2025/08/19 04:54:26 base crash: no output from test machine 2025/08/19 04:54:34 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 34560, "comps overflows": 0, "corpus": 45634, "corpus [files]": 7444, "corpus [symbols]": 17921, "cover overflows": 44367, "coverage": 314738, "distributor delayed": 69360, "distributor undelayed": 69360, "distributor violated": 2078, "exec candidate": 46590, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 17, "exec seeds": 0, "exec smash": 0, "exec total [base]": 162786, "exec total [new]": 304843, "exec triage": 144366, "executor restarts": 1198, "fault jobs": 0, "fuzzer jobs": 0, "fuzzing VMs [base]": 2, "fuzzing VMs [new]": 0, "hints jobs": 0, "max signal": 318386, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 16, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 46448, "no exec duration": 48369000000, "no exec requests": 432, "pending": 3, "prog exec time": 0, "reproducing": 7, "rpc recv": 13524443596, "rpc sent": 2320529664, "signal": 310083, "smash jobs": 0, "triage jobs": 0, "vm output": 75592477, "vm restarts [base]": 104, "vm restarts [new]": 201 } 2025/08/19 04:54:59 runner 0 connected 2025/08/19 04:55:00 new: boot error: can't ssh into the instance 2025/08/19 04:55:15 runner 1 connected 2025/08/19 04:55:50 base crash: no output from test machine 2025/08/19 04:56:54 new: boot error: can't ssh into the instance 2025/08/19 04:58:20 base crash: no output from test machine 2025/08/19 04:59:09 runner 2 connected 2025/08/19 04:59:34 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 34560, "comps overflows": 0, "corpus": 45634, "corpus [files]": 7444, "corpus [symbols]": 17921, "cover overflows": 44367, "coverage": 314738, "distributor delayed": 69360, "distributor undelayed": 69360, "distributor violated": 2078, "exec candidate": 46590, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 17, "exec seeds": 0, "exec smash": 0, "exec total [base]": 162786, "exec total [new]": 304843, "exec triage": 144366, "executor restarts": 1198, "fault jobs": 0, "fuzzer jobs": 0, "fuzzing VMs [base]": 3, "fuzzing VMs [new]": 0, "hints jobs": 0, "max signal": 318386, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 16, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 46448, "no exec duration": 48369000000, "no exec requests": 432, "pending": 3, "prog exec time": 0, "reproducing": 7, "rpc recv": 13617131780, "rpc sent": 2320530504, "signal": 310083, "smash jobs": 0, "triage jobs": 0, "vm output": 78167235, "vm restarts [base]": 107, "vm restarts [new]": 201 } 2025/08/19 04:59:58 base crash: no output from test machine 2025/08/19 05:00:15 base crash: no output from test machine 2025/08/19 05:00:40 reproducing crash 'WARNING in ext4_xattr_inode_lookup_create': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f fs/ext4/xattr.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 05:00:47 runner 0 connected 2025/08/19 05:02:07 reproducing crash 'INFO: task hung in corrupted': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f drivers/base/core.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 05:02:55 reproducing crash 'WARNING in ext4_xattr_inode_lookup_create': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f fs/ext4/xattr.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 05:04:09 base crash: no output from test machine 2025/08/19 05:04:09 reproducing crash 'WARNING in ext4_xattr_inode_lookup_create': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f fs/ext4/xattr.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 05:04:28 repro finished 'KASAN: slab-use-after-free Read in xfrm_state_find', repro=false crepro=false desc='' hub=false from_dashboard=false 2025/08/19 05:04:28 failed repro for "KASAN: slab-use-after-free Read in xfrm_state_find", err=%!s() 2025/08/19 05:04:28 "KASAN: slab-use-after-free Read in xfrm_state_find": saved crash log into 1755579868.crash.log 2025/08/19 05:04:28 "KASAN: slab-use-after-free Read in xfrm_state_find": saved repro log into 1755579868.repro.log 2025/08/19 05:04:34 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 34560, "comps overflows": 0, "corpus": 45634, "corpus [files]": 7444, "corpus [symbols]": 17921, "cover overflows": 44367, "coverage": 314738, "distributor delayed": 69360, "distributor undelayed": 69360, "distributor violated": 2078, "exec candidate": 46590, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 17, "exec seeds": 0, "exec smash": 0, "exec total [base]": 162786, "exec total [new]": 304843, "exec triage": 144366, "executor restarts": 1198, "fault jobs": 0, "fuzzer jobs": 0, "fuzzing VMs [base]": 1, "fuzzing VMs [new]": 0, "hints jobs": 0, "max signal": 318386, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 17, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 46448, "no exec duration": 48369000000, "no exec requests": 432, "pending": 3, "prog exec time": 0, "reproducing": 6, "rpc recv": 13648027836, "rpc sent": 2320530784, "signal": 310083, "smash jobs": 0, "triage jobs": 0, "vm output": 81147033, "vm restarts [base]": 108, "vm restarts [new]": 201 } 2025/08/19 05:04:49 new: boot error: can't ssh into the instance 2025/08/19 05:04:57 runner 2 connected 2025/08/19 05:05:06 new: boot error: can't ssh into the instance 2025/08/19 05:05:46 base crash: no output from test machine 2025/08/19 05:05:56 base: boot error: can't ssh into the instance 2025/08/19 05:06:35 runner 0 connected 2025/08/19 05:06:46 runner 3 connected 2025/08/19 05:07:00 new: boot error: can't ssh into the instance 2025/08/19 05:09:34 reproducing crash 'WARNING in mem_cgroup_migrate': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/memcontrol.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 05:09:34 STAT { "buffer too small": 0, "candidate triage jobs": 0, "candidates": 34560, "comps overflows": 0, "corpus": 45634, "corpus [files]": 7444, "corpus [symbols]": 17921, "cover overflows": 44367, "coverage": 314738, "distributor delayed": 69360, "distributor undelayed": 69360, "distributor violated": 2078, "exec candidate": 46590, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 17, "exec seeds": 0, "exec smash": 0, "exec total [base]": 162786, "exec total [new]": 304843, "exec triage": 144366, "executor restarts": 1198, "fault jobs": 0, "fuzzer jobs": 0, "fuzzing VMs [base]": 3, "fuzzing VMs [new]": 0, "hints jobs": 0, "max signal": 318386, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 19, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 46448, "no exec duration": 48369000000, "no exec requests": 432, "pending": 3, "prog exec time": 0, "reproducing": 6, "rpc recv": 13740716020, "rpc sent": 2320531624, "signal": 310083, "smash jobs": 0, "triage jobs": 0, "vm output": 83542136, "vm restarts [base]": 111, "vm restarts [new]": 201 } 2025/08/19 05:09:41 runner 0 connected 2025/08/19 05:09:57 base crash: no output from test machine 2025/08/19 05:10:11 reproducing crash 'WARNING in mem_cgroup_migrate': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/memcontrol.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 05:10:18 new: boot error: can't ssh into the instance 2025/08/19 05:10:20 base: boot error: can't ssh into the instance 2025/08/19 05:10:45 runner 2 connected 2025/08/19 05:11:07 runner 1 connected 2025/08/19 05:11:14 reproducing crash 'WARNING in ext4_xattr_inode_lookup_create': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f fs/ext4/xattr.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 05:11:30 reproducing crash 'WARNING in mem_cgroup_migrate': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/memcontrol.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 05:11:38 reproducing crash 'INFO: task hung in corrupted': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f drivers/net/netdevsim/bus.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 05:12:13 reproducing crash 'WARNING in ext4_xattr_inode_lookup_create': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f fs/ext4/xattr.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 05:12:40 reproducing crash 'WARNING in mem_cgroup_migrate': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/memcontrol.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 05:12:44 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 05:12:44 base crash: possible deadlock in ocfs2_reserve_suballoc_bits 2025/08/19 05:13:14 patched crashed: possible deadlock in ocfs2_xattr_set [need repro = false] 2025/08/19 05:13:27 reproducing crash 'WARNING in ext4_xattr_inode_lookup_create': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f fs/ext4/xattr.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 05:13:32 runner 1 connected 2025/08/19 05:14:01 base crash: WARNING in dbAdjTree 2025/08/19 05:14:02 patched crashed: WARNING in dbAdjTree [need repro = false] 2025/08/19 05:14:03 runner 0 connected 2025/08/19 05:14:09 reproducing crash 'WARNING in mem_cgroup_migrate': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/memcontrol.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 05:14:23 patched crashed: WARNING in dbAdjTree [need repro = false] 2025/08/19 05:14:34 STAT { "buffer too small": 0, "candidate triage jobs": 2, "candidates": 33816, "comps overflows": 0, "corpus": 45650, "corpus [files]": 7448, "corpus [symbols]": 17928, "cover overflows": 44834, "coverage": 314759, "distributor delayed": 69385, "distributor undelayed": 69385, "distributor violated": 2084, "exec candidate": 47334, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 17, "exec seeds": 0, "exec smash": 0, "exec total [base]": 165411, "exec total [new]": 307531, "exec triage": 144457, "executor restarts": 1212, "fault jobs": 0, "fuzzer jobs": 2, "fuzzing VMs [base]": 1, "fuzzing VMs [new]": 0, "hints jobs": 0, "max signal": 318426, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 24, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 46473, "no exec duration": 568773000000, "no exec requests": 2137, "pending": 3, "prog exec time": 363, "reproducing": 6, "rpc recv": 13900550352, "rpc sent": 2349178648, "signal": 310104, "smash jobs": 0, "triage jobs": 0, "vm output": 85669710, "vm restarts [base]": 112, "vm restarts [new]": 205 } 2025/08/19 05:14:50 runner 2 connected 2025/08/19 05:14:52 runner 1 connected 2025/08/19 05:14:54 new: boot error: can't ssh into the instance 2025/08/19 05:15:13 runner 0 connected 2025/08/19 05:15:30 reproducing crash 'WARNING in mem_cgroup_migrate': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/memcontrol.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 05:16:49 base crash: possible deadlock in ocfs2_try_remove_refcount_tree 2025/08/19 05:16:50 patched crashed: possible deadlock in ocfs2_try_remove_refcount_tree [need repro = false] 2025/08/19 05:16:59 reproducing crash 'WARNING in mem_cgroup_migrate': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/memcontrol.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 05:17:38 runner 0 connected 2025/08/19 05:17:39 runner 3 connected 2025/08/19 05:17:47 reproducing crash 'WARNING in mem_cgroup_migrate': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/memcontrol.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 05:18:48 reproducing crash 'WARNING in ext4_xattr_inode_lookup_create': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f fs/ext4/xattr.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 05:19:00 patched crashed: possible deadlock in ocfs2_reserve_suballoc_bits [need repro = false] 2025/08/19 05:19:01 base crash: possible deadlock in ocfs2_xattr_set 2025/08/19 05:19:11 patched crashed: possible deadlock in ocfs2_xattr_set [need repro = false] 2025/08/19 05:19:34 STAT { "buffer too small": 0, "candidate triage jobs": 2, "candidates": 30262, "comps overflows": 0, "corpus": 45684, "corpus [files]": 7455, "corpus [symbols]": 17940, "cover overflows": 45486, "coverage": 314839, "distributor delayed": 69418, "distributor undelayed": 69416, "distributor violated": 2084, "exec candidate": 50888, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 18, "exec seeds": 0, "exec smash": 0, "exec total [base]": 169105, "exec total [new]": 311229, "exec triage": 144594, "executor restarts": 1231, "fault jobs": 0, "fuzzer jobs": 2, "fuzzing VMs [base]": 1, "fuzzing VMs [new]": 0, "hints jobs": 0, "max signal": 318510, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 29, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 46513, "no exec duration": 725419000000, "no exec requests": 2749, "pending": 3, "prog exec time": 0, "reproducing": 6, "rpc recv": 14062476756, "rpc sent": 2391086040, "signal": 310168, "smash jobs": 0, "triage jobs": 0, "vm output": 88730904, "vm restarts [base]": 114, "vm restarts [new]": 208 } 2025/08/19 05:19:49 runner 2 connected 2025/08/19 05:19:50 runner 0 connected 2025/08/19 05:20:00 runner 1 connected 2025/08/19 05:20:03 reproducing crash 'WARNING in ext4_xattr_inode_lookup_create': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f fs/ext4/xattr.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 05:20:26 base: boot error: can't ssh into the instance 2025/08/19 05:20:51 reproducing crash 'INFO: task hung in corrupted': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f drivers/net/netdevsim/bus.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 05:21:20 reproducing crash 'WARNING in ext4_xattr_inode_lookup_create': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f fs/ext4/xattr.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 05:21:20 repro finished 'WARNING in ext4_xattr_inode_lookup_create', repro=true crepro=false desc='WARNING in ext4_xattr_inode_lookup_create' hub=false from_dashboard=false 2025/08/19 05:21:20 reproduction of "WARNING in ext4_xattr_inode_lookup_create" aborted: it's no longer needed 2025/08/19 05:21:20 reproduction of "WARNING in ext4_xattr_inode_lookup_create" aborted: it's no longer needed 2025/08/19 05:21:20 found repro for "WARNING in ext4_xattr_inode_lookup_create" (orig title: "-SAME-", reliability: 1), took 40.10 minutes 2025/08/19 05:21:20 "WARNING in ext4_xattr_inode_lookup_create": saved crash log into 1755580880.crash.log 2025/08/19 05:21:20 "WARNING in ext4_xattr_inode_lookup_create": saved repro log into 1755580880.repro.log 2025/08/19 05:21:35 new: boot error: can't ssh into the instance 2025/08/19 05:22:12 runner 2 connected 2025/08/19 05:22:50 base: boot error: can't ssh into the instance 2025/08/19 05:24:25 reproducing crash 'WARNING in mem_cgroup_migrate': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/memcontrol.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 05:24:34 STAT { "buffer too small": 0, "candidate triage jobs": 1, "candidates": 22819, "comps overflows": 0, "corpus": 45711, "corpus [files]": 7463, "corpus [symbols]": 17954, "cover overflows": 46857, "coverage": 314883, "distributor delayed": 69461, "distributor undelayed": 69461, "distributor violated": 2084, "exec candidate": 58331, "exec collide": 0, "exec fuzz": 0, "exec gen": 0, "exec hints": 0, "exec inject": 0, "exec minimize": 0, "exec retries": 19, "exec seeds": 0, "exec smash": 0, "exec total [base]": 175356, "exec total [new]": 318853, "exec triage": 144767, "executor restarts": 1252, "fault jobs": 0, "fuzzer jobs": 1, "fuzzing VMs [base]": 2, "fuzzing VMs [new]": 2, "hints jobs": 0, "max signal": 318620, "minimize: array": 0, "minimize: buffer": 0, "minimize: call": 31, "minimize: filename": 0, "minimize: integer": 0, "minimize: pointer": 0, "minimize: props": 0, "minimize: resource": 0, "modules [base]": 1, "modules [new]": 1, "new inputs": 46555, "no exec duration": 784714000000, "no exec requests": 2999, "pending": 1, "prog exec time": 219, "reproducing": 5, "rpc recv": 14197289628, "rpc sent": 2447191808, "signal": 310243, "smash jobs": 0, "triage jobs": 0, "vm output": 91778176, "vm restarts [base]": 115, "vm restarts [new]": 211 } 2025/08/19 05:24:53 patched crashed: lost connection to test machine [need repro = false] 2025/08/19 05:25:10 reproducing crash 'WARNING in mem_cgroup_migrate': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/memcontrol.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 05:25:27 patched crashed: possible deadlock in attr_data_get_block [need repro = false] 2025/08/19 05:25:42 runner 1 connected 2025/08/19 05:26:14 reproducing crash 'INFO: task hung in corrupted': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f net/ipv6/addrconf.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 05:26:15 runner 2 connected 2025/08/19 05:27:35 reproducing crash 'WARNING in mem_cgroup_migrate': failed to symbolize report: failed to start scripts/get_maintainer.pl [scripts/get_maintainer.pl --git-min-percent=15 -f mm/memcontrol.c]: fork/exec scripts/get_maintainer.pl: no such file or directory 2025/08/19 05:27:35 repro finished 'WARNING in mem_cgroup_migrate', repro=true crepro=false desc='WARNING in mem_cgroup_migrate' hub=false from_dashboard=false 2025/08/19 05:27:35 found repro for "WARNING in mem_cgroup_migrate" (orig title: "-SAME-", reliability: 1), took 88.01 minutes 2025/08/19 05:27:35 "WARNING in mem_cgroup_migrate": saved crash log into 1755581255.crash.log 2025/08/19 05:27:35 "WARNING in mem_cgroup_migrate": saved repro log into 1755581255.repro.log 2025/08/19 05:27:53 new: boot error: can't ssh into the instance 2025/08/19 05:28:23 runner 3 connected 2025/08/19 05:28:45 base crash: INFO: rcu detected stall in corrupted 2025/08/19 05:29:31 status reporting terminated 2025/08/19 05:29:31 bug reporting terminated 2025/08/19 05:29:31 attempt #0 to run "WARNING in mem_cgroup_migrate" on base: skipping due to errors: context deadline exceeded / 2025/08/19 05:29:31 attempt #0 to run "WARNING in ext4_xattr_inode_lookup_create" on base: skipping due to errors: context deadline exceeded / 2025/08/19 05:29:31 repro finished 'WARNING in drv_unassign_vif_chanctx', repro=false crepro=false desc='' hub=false from_dashboard=false 2025/08/19 05:31:03 repro finished 'WARNING in __linkwatch_sync_dev', repro=false crepro=false desc='' hub=false from_dashboard=false 2025/08/19 05:32:56 syz-diff (base): kernel context loop terminated 2025/08/19 05:33:45 repro finished 'INFO: task hung in sync_bdevs', repro=false crepro=false desc='' hub=false from_dashboard=false 2025/08/19 05:35:09 reproducing crash 'INFO: task hung in corrupted': concatenation step failed with context deadline exceeded 2025/08/19 05:35:09 repro finished 'INFO: task hung in corrupted', repro=false crepro=false desc='' hub=false from_dashboard=false 2025/08/19 05:37:19 syz-diff (new): kernel context loop terminated 2025/08/19 05:37:19 diff fuzzing terminated 2025/08/19 05:37:19 fuzzing is finished 2025/08/19 05:37:19 status at the end: Title On-Base On-Patched INFO: rcu detected stall in corrupted 1 crashes INFO: task hung in __iterate_supers 1 crashes INFO: task hung in corrupted 1 crashes INFO: task hung in disable_device 2 crashes INFO: task hung in sync_bdevs 1 crashes INFO: task hung in v9fs_evict_inode 1 crashes KASAN: slab-use-after-free Read in __xfrm_state_lookup 1 crashes 3 crashes KASAN: slab-use-after-free Read in jfs_lazycommit 1 crashes KASAN: slab-use-after-free Read in xfrm_alloc_spi 2 crashes 11 crashes KASAN: slab-use-after-free Read in xfrm_state_find 1 crashes KASAN: slab-use-after-free Write in __xfrm_state_delete 1 crashes WARNING in __linkwatch_sync_dev 2 crashes WARNING in dbAdjTree 2 crashes 2 crashes WARNING in drv_unassign_vif_chanctx 2 crashes WARNING in ext4_xattr_inode_lookup_create 1 crashes 4 crashes[reproduced] WARNING in mem_cgroup_migrate 1 crashes[reproduced] WARNING in xfrm6_tunnel_net_exit 1 crashes 1 crashes WARNING in xfrm_state_fini 3 crashes 9 crashes WARNING: suspicious RCU usage in get_callchain_entry 1 crashes 5 crashes[reproduced] general protection fault in pcl818_ai_cancel 1 crashes 1 crashes kernel BUG in txUnlock 6 crashes 8 crashes lost connection to test machine 6 crashes 8 crashes no output from test machine 36 crashes possible deadlock in attr_data_get_block 1 crashes 1 crashes possible deadlock in dqget 1 crashes possible deadlock in ntfs_fiemap 1 crashes possible deadlock in ocfs2_init_acl 9 crashes 42 crashes possible deadlock in ocfs2_reserve_suballoc_bits 19 crashes 46 crashes possible deadlock in ocfs2_setattr 1 crashes 4 crashes possible deadlock in ocfs2_try_remove_refcount_tree 12 crashes 29 crashes possible deadlock in ocfs2_write_begin_nolock 1 crashes possible deadlock in ocfs2_xattr_set 4 crashes 16 crashes unregister_netdevice: waiting for DEV to become free 1 crashes 1 crashes