Searched refs:work_queued (Results 1 – 7 of 7) sorted by relevance
66 sbi->work_queued = 0; in flush_mdb()83 if (!sbi->work_queued) { in hfs_mark_mdb_dirty()86 sbi->work_queued = 1; in hfs_mark_mdb_dirty()
579 bool work_queued = false; in amdgpu_dm_irq_schedule_work() local586 work_queued = true; in amdgpu_dm_irq_schedule_work()591 if (!work_queued) { in amdgpu_dm_irq_schedule_work()
74 sbi->work_queued = 0; in flush_superblock()89 if (!sbi->work_queued) { in affs_mark_sb_dirty()92 sbi->work_queued = 1; in affs_mark_sb_dirty()
30 int work_queued; /* non-zero if the delayed work is queued */ member
661 sbi->work_queued = 0; in delayed_sync_fs()673 if (!sbi->work_queued) { in ufs_mark_sb_dirty()676 sbi->work_queued = 1; in ufs_mark_sb_dirty()
295 sbi->work_queued = 0; in delayed_sync_fs()312 if (!sbi->work_queued) { in hfsplus_mark_mdb_dirty()315 sbi->work_queued = 1; in hfsplus_mark_mdb_dirty()
1541 unsigned int work_queued; in xennet_handle_rx() local1548 work_queued = XEN_RING_NR_UNCONSUMED_RESPONSES(&queue->rx); in xennet_handle_rx()1549 if (work_queued > queue->rx_rsp_unconsumed) { in xennet_handle_rx()1550 queue->rx_rsp_unconsumed = work_queued; in xennet_handle_rx()1552 } else if (unlikely(work_queued < queue->rx_rsp_unconsumed)) { in xennet_handle_rx()1563 if (likely(netif_carrier_ok(queue->info->netdev) && work_queued)) in xennet_handle_rx()