Home
last modified time | relevance | path

Searched refs:sync_work (Results 1 – 8 of 8) sorted by relevance

/linux/fs/hfsplus/
H A Dsuper.c292 sbi = container_of(work, struct hfsplus_sb_info, sync_work.work); in delayed_sync_fs()
314 queue_delayed_work(system_long_wq, &sbi->sync_work, delay); in hfsplus_mark_mdb_dirty()
334 cancel_delayed_work_sync(&sbi->sync_work); in hfsplus_put_super()
443 INIT_DELAYED_WORK(&sbi->sync_work, delayed_sync_fs); in hfsplus_fill_super()
638 cancel_delayed_work_sync(&sbi->sync_work); in hfsplus_fill_super()
/linux/drivers/net/ethernet/mellanox/mlx5/core/lib/
H A Dcrypto.c48 struct work_struct sync_work; member
446 if (mlx5_crypto_dek_need_sync(pool) && schedule_work(&pool->sync_work)) in mlx5_crypto_dek_free_locked()
592 container_of(work, struct mlx5_crypto_dek_pool, sync_work); in mlx5_crypto_dek_sync_work_fn()
699 INIT_WORK(&pool->sync_work, mlx5_crypto_dek_sync_work_fn); in mlx5_crypto_dek_pool_create()
711 cancel_work_sync(&pool->sync_work); in mlx5_crypto_dek_pool_destroy()
/linux/net/rfkill/
H A Dcore.c66 struct work_struct sync_work; member
1053 struct rfkill *rfkill = container_of(work, struct rfkill, sync_work); in rfkill_sync_work()
1096 INIT_WORK(&rfkill->sync_work, rfkill_sync_work); in rfkill_register()
1105 schedule_work(&rfkill->sync_work); in rfkill_register()
1139 cancel_work_sync(&rfkill->sync_work); in rfkill_unregister()
/linux/fs/ufs/
H A Dufs.h31 struct delayed_work sync_work; /* FS sync delayed work */ member
H A Dsuper.c658 sbi = container_of(work, struct ufs_sb_info, sync_work.work); in delayed_sync_fs()
675 queue_delayed_work(system_long_wq, &sbi->sync_work, delay); in ufs_mark_sb_dirty()
689 cancel_delayed_work_sync(&sbi->sync_work); in ufs_put_super()
756 INIT_DELAYED_WORK(&sbi->sync_work, delayed_sync_fs); in ufs_fill_super()
/linux/kernel/time/
H A Dntp.c495 static DECLARE_WORK(sync_work, sync_hw_clock);
501 queue_work(system_freezable_power_efficient_wq, &sync_work); in sync_timer_callback()
684 queue_work(system_freezable_power_efficient_wq, &sync_work); in ntp_notify_cmos_timer()
/linux/drivers/ptp/
H A Dptp_ocp.c368 struct delayed_work sync_work; member
4721 bp = container_of(work, struct ptp_ocp, sync_work.work); in ptp_ocp_sync_work()
4729 queue_delayed_work(system_power_efficient_wq, &bp->sync_work, HZ); in ptp_ocp_sync_work()
4757 INIT_DELAYED_WORK(&bp->sync_work, ptp_ocp_sync_work); in ptp_ocp_probe()
4819 queue_delayed_work(system_power_efficient_wq, &bp->sync_work, HZ); in ptp_ocp_probe()
4844 cancel_delayed_work_sync(&bp->sync_work); in ptp_ocp_remove()
/linux/drivers/md/
H A Dmd.c789 INIT_WORK(&mddev->sync_work, md_start_sync); in mddev_init()
5174 if (work_pending(&mddev->sync_work)) in stop_sync_thread()
5175 flush_work(&mddev->sync_work); in stop_sync_thread()
5252 if (work_busy(&mddev->sync_work)) in action_store()
5253 flush_work(&mddev->sync_work); in action_store()
5259 if (work_busy(&mddev->sync_work)) { in action_store()
8205 flush_work(&mddev->sync_work); in md_ioctl()
10019 struct mddev *mddev = container_of(ws, struct mddev, sync_work); in md_start_sync()
10243 queue_work(md_misc_wq, &mddev->sync_work); in md_check_recovery()
10299 queue_work(md_misc_wq, &mddev->sync_work); in md_check_recovery()