BUG: workqueue lockup - pool cpus=0 node=0 flags=0x0 nice=0 stuck for 148s! Showing busy workqueues and worker pools: workqueue events: flags=0x100 pwq 2: cpus=0 node=0 flags=0x0 nice=0 active=13 refcnt=14 pending: 3*nsim_dev_hwstats_traffic_work, ovs_dp_masks_rebalance, perf_sched_delayed, psi_avgs_work, 3*ovs_dp_masks_rebalance, debugfs_reap_work, vmstat_shepherd, switchdev_deferred_process_work, rht_deferred_worker pwq 6: cpus=1 node=0 flags=0x0 nice=0 active=11 refcnt=12 in-flight: 7121:xfrm_state_gc_task pending: 7*nsim_dev_hwstats_traffic_work, 3*ovs_dp_masks_rebalance workqueue events_long: flags=0x100 pwq 2: cpus=0 node=0 flags=0x0 nice=0 active=5 refcnt=6 pending: 4*defense_work_handler, br_multicast_gc_work workqueue events_unbound: flags=0x2 pwq 8: cpus=0-1 flags=0x4 nice=0 active=2 refcnt=3 in-flight: 6656:idle_cull_fn ,6624:idle_cull_fn pwq 8: cpus=0-1 flags=0x4 nice=0 active=3 refcnt=4 in-flight: 6610:toggle_allocation_gate ,6630:linkwatch_event ,6636:fsnotify_mark_destroy_workfn workqueue events_freezable: flags=0x104 pwq 2: cpus=0 node=0 flags=0x0 nice=0 active=1 refcnt=2 pending: update_balloon_stats_func workqueue events_power_efficient: flags=0x180 pwq 2: cpus=0 node=0 flags=0x0 nice=0 active=11 refcnt=12 pending: crda_timeout_work, gc_worker, wg_ratelimiter_gc_entries, do_cache_clean, 2*neigh_managed_work, 2*neigh_periodic_work, reg_check_chans_work, 2*check_lifetime workqueue rcu_gp: flags=0x108 pwq 2: cpus=0 node=0 flags=0x0 nice=0 active=1 refcnt=2 pending: srcu_invoke_callbacks workqueue netns: flags=0x6000a pwq 8: cpus=0-1 flags=0x4 nice=0 active=1 refcnt=5 in-flight: 6623:cleanup_net inactive: cleanup_net workqueue mm_percpu_wq: flags=0x108 pwq 2: cpus=0 node=0 flags=0x0 nice=0 active=1 refcnt=2 pending: vmstat_update workqueue writeback: flags=0x4a pwq 8: cpus=0-1 flags=0x4 nice=0 active=1 refcnt=2 in-flight: 6631:wb_workfn workqueue mld: flags=0x40108 pwq 2: cpus=0 node=0 flags=0x0 nice=0 active=1 refcnt=3 pending: mld_ifc_work inactive: mld_ifc_work workqueue ipv6_addrconf: flags=0x6000a pwq 8: cpus=0-1 flags=0x4 nice=0 active=1 refcnt=54 in-flight: 6611:addrconf_dad_work inactive: 42*addrconf_dad_work, 8*addrconf_verify_work workqueue bat_events: flags=0x6000a pwq 8: cpus=0-1 flags=0x4 nice=0 active=1 refcnt=37 in-flight: 6620:batadv_tt_purge inactive: batadv_tt_purge, batadv_dat_purge, batadv_bla_periodic_work, 2*batadv_dat_purge, 2*batadv_bla_periodic_work, batadv_dat_purge, batadv_tt_purge, batadv_bla_periodic_work, batadv_dat_purge, batadv_bla_periodic_work, batadv_dat_purge, batadv_bla_periodic_work, batadv_dat_purge, 7*batadv_mcast_mla_update, batadv_iv_send_outstanding_bat_ogm_packet, 3*batadv_purge_orig, 2*batadv_iv_send_outstanding_bat_ogm_packet, batadv_purge_orig, 3*batadv_iv_send_outstanding_bat_ogm_packet, batadv_tt_purge workqueue wg-crypt-wg0: flags=0x128 pwq 2: cpus=0 node=0 flags=0x0 nice=0 active=3 refcnt=4 pending: wg_packet_encrypt_worker, wg_packet_tx_worker, wg_packet_decrypt_worker workqueue wg-crypt-wg1: flags=0x128 pwq 2: cpus=0 node=0 flags=0x0 nice=0 active=3 refcnt=4 pending: wg_packet_encrypt_worker, wg_packet_tx_worker, wg_packet_decrypt_worker workqueue wg-crypt-wg2: flags=0x128 pwq 2: cpus=0 node=0 flags=0x0 nice=0 active=2 refcnt=3 pending: wg_packet_encrypt_worker, wg_packet_tx_worker workqueue hci0: flags=0x20012 pwq 9: cpus=0-1 node=0 flags=0x4 nice=-20 active=1 refcnt=4 in-flight: 21401:hci_cmd_sync_work workqueue wg-crypt-wg0: flags=0x128 pwq 2: cpus=0 node=0 flags=0x0 nice=0 active=2 refcnt=3 pending: wg_packet_encrypt_worker, wg_packet_tx_worker workqueue wg-crypt-wg1: flags=0x128 pwq 2: cpus=0 node=0 flags=0x0 nice=0 active=2 refcnt=3 pending: wg_packet_encrypt_worker, wg_packet_tx_worker workqueue wg-crypt-wg2: flags=0x128 pwq 2: cpus=0 node=0 flags=0x0 nice=0 active=2 refcnt=3 pending: wg_packet_encrypt_worker, wg_packet_tx_worker workqueue hci3: flags=0x20012 pwq 9: cpus=0-1 node=0 flags=0x4 nice=-20 active=1 refcnt=4 in-flight: 5142:hci_cmd_sync_work workqueue wg-crypt-wg0: flags=0x128 pwq 2: cpus=0 node=0 flags=0x0 nice=0 active=3 refcnt=4 pending: wg_packet_encrypt_worker, wg_packet_tx_worker, wg_packet_decrypt_worker workqueue wg-crypt-wg1: flags=0x128 pwq 2: cpus=0 node=0 flags=0x0 nice=0 active=3 refcnt=4 pending: wg_packet_encrypt_worker, wg_packet_tx_worker, wg_packet_decrypt_worker workqueue wg-crypt-wg2: flags=0x128 pwq 2: cpus=0 node=0 flags=0x0 nice=0 active=2 refcnt=3 pending: wg_packet_encrypt_worker, wg_packet_tx_worker workqueue hci2: flags=0x20012 pwq 9: cpus=0-1 node=0 flags=0x4 nice=-20 active=1 refcnt=4 in-flight: 28421:hci_cmd_sync_work workqueue hci1: flags=0x20012 pwq 9: cpus=0-1 node=0 flags=0x4 nice=-20 active=1 refcnt=4 in-flight: 5835:hci_cmd_sync_work pool 6: cpus=1 node=0 flags=0x0 nice=0 hung=0s workers=7 idle: 24 5904 7124 43 7125 7122 pool 8: cpus=0-1 flags=0x4 nice=0 hung=0s workers=12 idle: 6656 6631 6607 6661 28425 pool 9: cpus=0-1 node=0 flags=0x4 nice=-20 hung=0s workers=6 idle: 28424 28423 Showing backtraces of running workers in stalled CPU-bound worker pools: