BUG: workqueue lockup - pool cpus=0 node=0 flags=0x0 nice=0 stuck for 147s! Showing busy workqueues and worker pools: workqueue events: flags=0x0 pwq 2: cpus=1 node=0 flags=0x0 nice=0 active=15/256 refcnt=16 pending: 7*nsim_dev_hwstats_traffic_work, 3*ovs_dp_masks_rebalance, 3*psi_avgs_work, switchdev_deferred_process_work, rht_deferred_worker pwq 0: cpus=0 node=0 flags=0x0 nice=0 active=8/256 refcnt=9 pending: nsim_dev_hwstats_traffic_work, free_obj_work, psi_avgs_work, vmstat_shepherd, 2*ovs_dp_masks_rebalance, psi_avgs_work, kfree_rcu_monitor workqueue events_long: flags=0x0 pwq 2: cpus=1 node=0 flags=0x0 nice=0 active=5/256 refcnt=6 pending: 3*defense_work_handler, 2*br_multicast_gc_work pwq 0: cpus=0 node=0 flags=0x0 nice=0 active=2/256 refcnt=3 pending: 2*defense_work_handler workqueue events_unbound: flags=0x2 pwq 4: cpus=0-1 flags=0x4 nice=0 active=7/512 refcnt=8 pending: 2*nsim_dev_trap_report_work, cfg80211_wiphy_work, toggle_allocation_gate, 2*nsim_dev_trap_report_work, macvlan_process_broadcast pwq 4: cpus=0-1 flags=0x4 nice=0 active=13/512 refcnt=14 pending: nsim_dev_trap_report_work, 3*cfg80211_wiphy_work, 3*nsim_dev_trap_report_work, flush_memcg_stats_dwork, macvlan_process_broadcast, crng_reseed, macvlan_process_broadcast, cfg80211_wiphy_work, macvlan_process_broadcast workqueue events_freezable: flags=0x4 pwq 0: cpus=0 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: update_balloon_stats_func workqueue events_power_efficient: flags=0x80 pwq 2: cpus=1 node=0 flags=0x0 nice=0 active=3/256 refcnt=4 pending: wg_ratelimiter_gc_entries, gc_worker, check_lifetime pwq 0: cpus=0 node=0 flags=0x0 nice=0 active=6/256 refcnt=7 pending: nf_flow_offload_work_gc, 2*neigh_managed_work, do_cache_clean, 2*neigh_periodic_work workqueue mm_percpu_wq: flags=0x8 pwq 2: cpus=1 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: vmstat_update pwq 0: cpus=0 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: vmstat_update workqueue writeback: flags=0x4a pwq 4: cpus=0-1 flags=0x4 nice=0 active=1/256 refcnt=2 pending: wb_workfn workqueue kblockd: flags=0x18 pwq 3: cpus=1 node=0 flags=0x0 nice=-20 active=1/256 refcnt=2 pending: blk_mq_timeout_work pwq 1: cpus=0 node=0 flags=0x0 nice=-20 active=1/256 refcnt=2 pending: blk_mq_timeout_work workqueue dm_bufio_cache: flags=0x8 pwq 2: cpus=1 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: work_fn workqueue ipv6_addrconf: flags=0xe000a pwq 4: cpus=0-1 flags=0x4 nice=0 active=1/1 refcnt=8 pending: addrconf_verify_work inactive: 4*addrconf_verify_work workqueue krxrpcd: flags=0xa001a pwq 5: cpus=0-1 node=0 flags=0x5 nice=-20 active=1/1 refcnt=9 MAYDAY pending: rxrpc_peer_keepalive_worker inactive: 4*rxrpc_peer_keepalive_worker workqueue bat_events: flags=0xe000a pwq 4: cpus=0-1 flags=0x4 nice=0 active=1/1 refcnt=39 pending: batadv_nc_worker inactive: 3*batadv_nc_worker, 4*batadv_mcast_mla_update, 4*batadv_tt_purge, batadv_purge_orig, 11*batadv_iv_send_outstanding_bat_ogm_packet, 3*batadv_purge_orig, batadv_iv_send_outstanding_bat_ogm_packet, 2*batadv_dat_purge, 3*batadv_bla_periodic_work, 2*batadv_dat_purge, batadv_bla_periodic_work workqueue wg-crypt-wg0: flags=0x28 pwq 2: cpus=1 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: wg_packet_encrypt_worker pwq 0: cpus=0 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: wg_packet_encrypt_worker workqueue wg-kex-wg1: flags=0x6 pwq 4: cpus=0-1 flags=0x4 nice=0 active=1/256 refcnt=2 pending: wg_packet_handshake_send_worker workqueue wg-crypt-wg1: flags=0x28 pwq 2: cpus=1 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: wg_packet_encrypt_worker workqueue wg-kex-wg2: flags=0x6 pwq 4: cpus=0-1 flags=0x4 nice=0 active=1/256 refcnt=2 pending: wg_packet_handshake_send_worker workqueue wg-crypt-wg2: flags=0x28 pwq 2: cpus=1 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: wg_packet_encrypt_worker workqueue wg-kex-wg0: flags=0x6 pwq 4: cpus=0-1 flags=0x4 nice=0 active=1/256 refcnt=2 pending: wg_packet_handshake_send_worker pwq 4: cpus=0-1 flags=0x4 nice=0 active=1/256 refcnt=2 pending: wg_packet_handshake_send_worker workqueue wg-kex-wg1: flags=0x6 pwq 4: cpus=0-1 flags=0x4 nice=0 active=2/256 refcnt=3 pending: 2*wg_packet_handshake_send_worker workqueue wg-kex-wg2: flags=0x6 pwq 4: cpus=0-1 flags=0x4 nice=0 active=2/256 refcnt=3 pending: 2*wg_packet_handshake_send_worker workqueue hci0: flags=0xa0012 pwq 5: cpus=0-1 node=0 flags=0x5 nice=-20 active=1/1 refcnt=4 in-flight: 12715:hci_conn_timeout workqueue hci0: flags=0xa0012 pwq 5: cpus=0-1 node=0 flags=0x5 nice=-20 active=1/1 refcnt=4 pending: hci_cmd_sync_work workqueue wg-crypt-wg0: flags=0x28 pwq 2: cpus=1 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: wg_packet_encrypt_worker pwq 0: cpus=0 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: wg_packet_encrypt_worker workqueue wg-crypt-wg1: flags=0x28 pwq 2: cpus=1 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: wg_packet_encrypt_worker pwq 0: cpus=0 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: wg_packet_encrypt_worker workqueue wg-crypt-wg2: flags=0x28 pwq 2: cpus=1 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: wg_packet_encrypt_worker pwq 0: cpus=0 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 pending: wg_packet_encrypt_worker workqueue bond1: flags=0xa000a pwq 4: cpus=0-1 flags=0x4 nice=0 active=1/1 refcnt=4 pending: bond_mii_monitor workqueue bond1: flags=0xa000a pwq 4: cpus=0-1 flags=0x4 nice=0 active=1/1 refcnt=5 pending: bond_alb_monitor inactive: bond_mii_monitor pool 5: cpus=0-1 node=0 flags=0x5 nice=-20 hung=98s workers=2 manager: 5773 Showing backtraces of running workers in stalled CPU-bound worker pools: