| /linux/mm/ |
| H A D | backing-dev.c | 63 struct bdi_writeback *wb) in collect_wb_stats() argument 67 spin_lock(&wb->list_lock); in collect_wb_stats() 68 list_for_each_entry(inode, &wb->b_dirty, i_io_list) in collect_wb_stats() 70 list_for_each_entry(inode, &wb->b_io, i_io_list) in collect_wb_stats() 72 list_for_each_entry(inode, &wb->b_more_io, i_io_list) in collect_wb_stats() 74 list_for_each_entry(inode, &wb->b_dirty_time, i_io_list) in collect_wb_stats() 77 spin_unlock(&wb->list_lock); in collect_wb_stats() 79 stats->nr_writeback += wb_stat(wb, WB_WRITEBACK); in collect_wb_stats() 80 stats->nr_reclaimable += wb_stat(wb, WB_RECLAIMABLE); in collect_wb_stats() 81 stats->nr_dirtied += wb_stat(wb, WB_DIRTIED); in collect_wb_stats() [all …]
|
| H A D | page-writeback.c | 125 #define GDTC_INIT(__wb) .wb = (__wb), \ 131 #define MDTC_INIT(__wb, __gdtc) .wb = (__wb), \ 151 static struct fprop_local_percpu *wb_memcg_completions(struct bdi_writeback *wb) in dtc_dom() 153 return &wb->memcg_completions; 156 static void wb_min_max_ratio(struct bdi_writeback *wb, in mdtc_gdtc() 159 unsigned long this_bw = READ_ONCE(wb->avg_write_bandwidth); in wb_memcg_completions() argument 160 unsigned long tot_bw = atomic_long_read(&wb->bdi->tot_write_bandwidth); in wb_memcg_completions() 161 unsigned long long min = wb->bdi->min_ratio; in wb_memcg_completions() 162 unsigned long long max = wb->bdi->max_ratio; in wb_memcg_completions() 165 * @wb ma in wb_min_max_ratio() 164 wb_min_max_ratio(struct bdi_writeback * wb,unsigned long * minp,unsigned long * maxp) wb_min_max_ratio() argument 213 wb_memcg_completions(struct bdi_writeback * wb) wb_memcg_completions() argument 218 wb_min_max_ratio(struct bdi_writeback * wb,unsigned long * minp,unsigned long * maxp) wb_min_max_ratio() argument 581 __wb_writeout_add(struct bdi_writeback * wb,long nr) __wb_writeout_add() argument 595 wb_writeout_inc(struct bdi_writeback * wb) wb_writeout_inc() argument 904 struct bdi_writeback *wb = dtc->wb; __wb_calc_thresh() local 945 wb_calc_thresh(struct bdi_writeback * wb,unsigned long thresh) wb_calc_thresh() argument 953 cgwb_calc_thresh(struct bdi_writeback * wb) cgwb_calc_thresh() argument 1073 struct bdi_writeback *wb = dtc->wb; wb_position_ratio() local 1234 wb_update_write_bandwidth(struct bdi_writeback * wb,unsigned long elapsed,unsigned long written) wb_update_write_bandwidth() argument 1342 struct bdi_writeback *wb = dtc->wb; wb_update_dirty_ratelimit() local 1493 struct bdi_writeback *wb = gdtc->wb; __wb_update_bandwidth() local 1532 wb_update_bandwidth(struct bdi_writeback * wb) wb_update_bandwidth() argument 1542 wb_bandwidth_estimate_start(struct bdi_writeback * wb) wb_bandwidth_estimate_start() argument 1574 wb_max_pause(struct bdi_writeback * wb,unsigned long wb_dirty) wb_max_pause() argument 1593 wb_min_pause(struct bdi_writeback * wb,long max_pause,unsigned long task_ratelimit,unsigned long dirty_ratelimit,int * nr_dirtied_pause) wb_min_pause() argument 1670 struct bdi_writeback *wb = dtc->wb; wb_dirty_limits() local 1810 balance_dirty_pages(struct bdi_writeback * wb,unsigned long pages_dirtied,unsigned int flags) balance_dirty_pages() argument 2051 struct bdi_writeback *wb = NULL; balance_dirty_pages_ratelimited_flags() local 2128 struct bdi_writeback *wb = dtc->wb; wb_bg_dirty_limits() local 2160 wb_over_bg_thresh(struct bdi_writeback * wb) wb_over_bg_thresh() argument 2590 struct bdi_writeback *wb; do_writepages() local 2649 struct bdi_writeback *wb; folio_account_dirtied() local 2672 folio_account_cleaned(struct folio * folio,struct bdi_writeback * wb) folio_account_cleaned() argument 2775 struct bdi_writeback *wb; folio_redirty_for_writepage() local 2866 struct bdi_writeback *wb; __folio_cancel_dirty() local 2904 struct bdi_writeback *wb; folio_clear_dirty_for_io() local 2957 wb_inode_writeback_start(struct bdi_writeback * wb) wb_inode_writeback_start() argument 2962 wb_inode_writeback_end(struct bdi_writeback * wb) wb_inode_writeback_end() argument 2987 struct bdi_writeback *wb; __folio_end_writeback() local 3028 struct bdi_writeback *wb; __folio_start_writeback() local [all...] |
| /linux/fs/ |
| H A D | fs-writeback.c | 81 static bool wb_io_lists_populated(struct bdi_writeback *wb) in wb_io_lists_populated() argument 83 if (wb_has_dirty_io(wb)) { in wb_io_lists_populated() 86 set_bit(WB_has_dirty_io, &wb->state); in wb_io_lists_populated() 87 WARN_ON_ONCE(!wb->avg_write_bandwidth); in wb_io_lists_populated() 88 atomic_long_add(wb->avg_write_bandwidth, in wb_io_lists_populated() 89 &wb->bdi->tot_write_bandwidth); in wb_io_lists_populated() 94 static void wb_io_lists_depopulated(struct bdi_writeback *wb) in wb_io_lists_depopulated() argument 96 if (wb_has_dirty_io(wb) && list_empty(&wb->b_dirty) && in wb_io_lists_depopulated() 97 list_empty(&wb->b_io) && list_empty(&wb->b_more_io)) { in wb_io_lists_depopulated() 98 clear_bit(WB_has_dirty_io, &wb->state); in wb_io_lists_depopulated() [all …]
|
| /linux/include/linux/ |
| H A D | backing-dev.h | 39 void wb_start_background_writeback(struct bdi_writeback *wb); 49 static inline bool wb_has_dirty_io(struct bdi_writeback *wb) in wb_has_dirty_io() argument 51 return test_bit(WB_has_dirty_io, &wb->state); in wb_has_dirty_io() 63 static inline void wb_stat_mod(struct bdi_writeback *wb, in wb_stat_mod() argument 66 percpu_counter_add_batch(&wb->stat[item], amount, WB_STAT_BATCH); in wb_stat_mod() 69 static inline s64 wb_stat(struct bdi_writeback *wb, enum wb_stat_item item) in wb_stat() argument 71 return percpu_counter_read_positive(&wb->stat[item]); in wb_stat() 74 static inline s64 wb_stat_sum(struct bdi_writeback *wb, enum wb_stat_item item) in wb_stat_sum() argument 76 return percpu_counter_sum_positive(&wb->stat[item]); in wb_stat_sum() 79 extern void wb_writeout_inc(struct bdi_writeback *wb); [all …]
|
| H A D | writeback.h | 80 struct bdi_writeback *wb; /* wb this writeback is issued under */ member 84 int wb_id; /* current wb id */ 85 int wb_lcand_id; /* last foreign candidate wb id */ 86 int wb_tcand_id; /* this foreign candidate wb id */ 87 size_t wb_bytes; /* bytes written by current wb */ 107 ((wbc)->wb ? (wbc)->wb->blkcg_css : blkcg_root_css) 113 * A wb_domain represents a domain that wb's (bdi_writeback's) belong to 115 * domain, global_wb_domain, that every wb i 313 struct bdi_writeback *wb; global() member [all...] |
| /linux/include/trace/events/ |
| H A D | writeback.h | 158 static inline ino_t __trace_wb_assign_cgroup(struct bdi_writeback *wb) 160 return cgroup_ino(wb->memcg_css->cgroup); in __trace_wb_assign_cgroup() 165 if (wbc->wb) in __trace_wbc_assign_cgroup() 166 return __trace_wb_assign_cgroup(wbc->wb); in __trace_wbc_assign_cgroup() 172 static inline ino_t __trace_wb_assign_cgroup(struct bdi_writeback *wb) 275 TP_PROTO(struct folio *folio, struct bdi_writeback *wb), 277 TP_ARGS(folio, wb), 292 strscpy_pad(__entry->name, bdi_dev_name(wb->bdi), 32); 293 __entry->bdi_id = wb->bdi->id; 295 __entry->memcg_id = wb 159 __trace_wb_assign_cgroup(struct bdi_writeback * wb) __trace_wb_assign_cgroup() argument 173 __trace_wb_assign_cgroup(struct bdi_writeback * wb) __trace_wb_assign_cgroup() argument [all...] |
| /linux/security/apparmor/ |
| H A D | match.c | 682 #define inc_wb_pos(wb) \ argument 685 wb->pos = (wb->pos + 1) & (WB_HISTORY_SIZE - 1); \ 686 wb->len = (wb->len + 1) > WB_HISTORY_SIZE ? WB_HISTORY_SIZE : \ 687 wb->len + 1; \ 692 static bool is_loop(struct match_workbuf *wb, aa_state_t state, in is_loop() argument 695 int pos = wb->pos; in is_loop() 698 if (wb->history[pos] < state) in is_loop() 701 for (i = 0; i < wb->len; i++) { in is_loop() 702 if (wb->history[pos] == state) { in is_loop() 714 const char *str, struct match_workbuf *wb, in leftmatch_fb() argument [all …]
|
| /linux/drivers/gpu/drm/radeon/ |
| H A D | r600_dma.c | 55 if (rdev->wb.enabled) in r600_dma_get_rptr() 56 rptr = rdev->wb.wb[ring->rptr_offs/4]; in r600_dma_get_rptr() 143 upper_32_bits(rdev->wb.gpu_addr + R600_WB_DMA_RPTR_OFFSET) & 0xFF); in r600_dma_resume() 145 ((rdev->wb.gpu_addr + R600_WB_DMA_RPTR_OFFSET) & 0xFFFFFFFC)); in r600_dma_resume() 147 if (rdev->wb.enabled) in r600_dma_resume() 243 gpu_addr = rdev->wb.gpu_addr + index; in r600_dma_ring_test() 246 rdev->wb.wb[index/4] = cpu_to_le32(tmp); in r600_dma_ring_test() 260 tmp = le32_to_cpu(rdev->wb.wb[index/4]); in r600_dma_ring_test() 350 gpu_addr = rdev->wb.gpu_addr + index; in r600_dma_ib_test() 381 tmp = le32_to_cpu(rdev->wb.wb[index/4]); in r600_dma_ib_test() [all …]
|
| H A D | radeon_device.c | 424 rdev->wb.enabled = false; in radeon_wb_disable() 438 if (rdev->wb.wb_obj) { in radeon_wb_fini() 439 if (!radeon_bo_reserve(rdev->wb.wb_obj, false)) { in radeon_wb_fini() 440 radeon_bo_kunmap(rdev->wb.wb_obj); in radeon_wb_fini() 441 radeon_bo_unpin(rdev->wb.wb_obj); in radeon_wb_fini() 442 radeon_bo_unreserve(rdev->wb.wb_obj); in radeon_wb_fini() 444 radeon_bo_unref(&rdev->wb.wb_obj); in radeon_wb_fini() 445 rdev->wb.wb = NULL; in radeon_wb_fini() 446 rdev->wb.wb_obj = NULL; in radeon_wb_fini() 463 if (rdev->wb.wb_obj == NULL) { in radeon_wb_init() [all …]
|
| H A D | cik_sdma.c | 67 if (rdev->wb.enabled) { in cik_sdma_get_rptr() 68 rptr = rdev->wb.wb[ring->rptr_offs/4]; in cik_sdma_get_rptr() 138 if (rdev->wb.enabled) { in cik_sdma_ring_ib_execute() 400 upper_32_bits(rdev->wb.gpu_addr + wb_offset) & 0xFFFFFFFF); in cik_sdma_gfx_resume() 402 ((rdev->wb.gpu_addr + wb_offset) & 0xFFFFFFFC)); in cik_sdma_gfx_resume() 404 if (rdev->wb.enabled) in cik_sdma_gfx_resume() 658 gpu_addr = rdev->wb.gpu_addr + index; in cik_sdma_ring_test() 661 rdev->wb.wb[index/4] = cpu_to_le32(tmp); in cik_sdma_ring_test() 676 tmp = le32_to_cpu(rdev->wb.wb[index/4]); in cik_sdma_ring_test() 715 gpu_addr = rdev->wb.gpu_addr + index; in cik_sdma_ib_test() [all …]
|
| H A D | ni_dma.c | 57 if (rdev->wb.enabled) { in cayman_dma_get_rptr() 58 rptr = rdev->wb.wb[ring->rptr_offs/4]; in cayman_dma_get_rptr() 127 if (rdev->wb.enabled) { in cayman_dma_ring_ib_execute() 222 upper_32_bits(rdev->wb.gpu_addr + wb_offset) & 0xFF); in cayman_dma_resume() 224 ((rdev->wb.gpu_addr + wb_offset) & 0xFFFFFFFC)); in cayman_dma_resume() 226 if (rdev->wb.enabled) in cayman_dma_resume()
|
| H A D | radeon_fence.c | 71 if (likely(rdev->wb.enabled || !drv->scratch_reg)) { in radeon_fence_write() 93 if (likely(rdev->wb.enabled || !drv->scratch_reg)) { in radeon_fence_read() 765 if (rdev->wb.use_event || !radeon_ring_supports_scratch_reg(rdev, &rdev->ring[ring])) { in radeon_fence_driver_start_ring() 769 rdev->fence_drv[ring].cpu_addr = &rdev->wb.wb[index/4]; in radeon_fence_driver_start_ring() 770 rdev->fence_drv[ring].gpu_addr = rdev->wb.gpu_addr + in radeon_fence_driver_start_ring() 789 rdev->fence_drv[ring].cpu_addr = &rdev->wb.wb[index/4]; in radeon_fence_driver_start_ring() 790 rdev->fence_drv[ring].gpu_addr = rdev->wb.gpu_addr + index; in radeon_fence_driver_start_ring()
|
| H A D | radeon_ring.c | 308 else if (rdev->wb.enabled) in radeon_ring_backup() 422 if (rdev->wb.enabled) { in radeon_ring_init() 424 ring->next_rptr_gpu_addr = rdev->wb.gpu_addr + index; in radeon_ring_init() 425 ring->next_rptr_cpu_addr = &rdev->wb.wb[index/4]; in radeon_ring_init()
|
| /linux/drivers/usb/class/ |
| H A D | cdc-acm.c | 157 usb_poison_urb(acm->wb[i].urb); in acm_poison_urbs() 169 usb_unpoison_urb(acm->wb[i].urb); in acm_unpoison_urbs() 182 struct acm_wb *wb; in acm_wb_alloc() local 187 wb = &acm->wb[wbn]; in acm_wb_alloc() 188 if (!wb->use) { in acm_wb_alloc() 189 wb->use = true; in acm_wb_alloc() 190 wb->len = 0; in acm_wb_alloc() 207 if(acm->wb[i].use) in acm_wb_is_avail() 216 static void acm_write_done(struct acm *acm, struct acm_wb *wb) in acm_write_done() argument 218 wb->use = false; in acm_write_done() [all …]
|
| /linux/drivers/net/wireless/microchip/wilc1000/ |
| H A D | spi.c | 402 static int wilc_spi_tx_rx(struct wilc *wilc, u8 *wb, u8 *rb, u32 rlen) in wilc_spi_tx_rx() argument 411 .tx_buf = wb, in wilc_spi_tx_rx() 524 u8 wb[32], rb[32]; in wilc_spi_single_read() local 531 memset(wb, 0x0, sizeof(wb)); in wilc_spi_single_read() 533 c = (struct wilc_spi_cmd *)wb; in wilc_spi_single_read() 554 c->u.simple_cmd.crc[0] = wilc_get_crc7(wb, cmd_len); in wilc_spi_single_read() 559 if (cmd_len + resp_len > ARRAY_SIZE(wb)) { in wilc_spi_single_read() 562 cmd_len, resp_len, ARRAY_SIZE(wb)); in wilc_spi_single_read() 566 if (wilc_spi_tx_rx(wilc, wb, rb, cmd_len + resp_len)) { in wilc_spi_single_read() 619 u8 wb[32], rb[32]; in wilc_spi_write_cmd() local [all …]
|
| /linux/certs/ |
| H A D | extract-cert.c | 47 static BIO *wb; variable 55 if (!wb) { in write_cert() 56 wb = BIO_new_file(cert_dst, "wb"); in write_cert() 57 ERR(!wb, "%s", cert_dst); in write_cert() 60 ERR(!i2d_X509_bio(wb, x509), "%s", cert_dst); in write_cert() 166 if (wb && !x509) { in main() 179 BIO_free(wb); in main()
|
| /linux/drivers/gpu/drm/msm/disp/dpu1/ |
| H A D | dpu_hw_wb.h | 37 struct dpu_hw_wb_cfg *wb); 40 struct dpu_hw_wb_cfg *wb, 44 struct dpu_hw_wb_cfg *wb);
|
| H A D | dpu_encoder_phys_wb.c | 38 static bool _dpu_encoder_phys_wb_clk_force_ctrl(struct dpu_hw_wb *wb, in _dpu_encoder_phys_wb_clk_force_ctrl() argument 42 if (wb->ops.setup_clk_force_ctrl) { in _dpu_encoder_phys_wb_clk_force_ctrl() 43 *forced_on = wb->ops.setup_clk_force_ctrl(wb, enable); in _dpu_encoder_phys_wb_clk_force_ctrl() 48 *forced_on = mdp->ops.setup_clk_force_ctrl(mdp, wb->caps->clk_ctrl, enable); in _dpu_encoder_phys_wb_clk_force_ctrl() 241 intf_cfg.wb = hw_wb->idx; in dpu_encoder_phys_wb_setup_ctl() 262 intf_cfg.wb = hw_wb->idx; in dpu_encoder_phys_wb_setup_ctl()
|
| /linux/scripts/ |
| H A D | kallsyms.c | 650 int wa, wb; in compare_symbols() local 660 wb = (sb->sym[0] == 'w') || (sb->sym[0] == 'W'); in compare_symbols() 661 if (wa != wb) in compare_symbols() 662 return wa - wb; in compare_symbols() 666 wb = may_be_linker_script_provide_symbol(sb); in compare_symbols() 667 if (wa != wb) in compare_symbols() 668 return wa - wb; in compare_symbols() 672 wb = strspn(sym_name(sb), "_"); in compare_symbols() 673 if (wa != wb) in compare_symbols() 674 return wa - wb; in compare_symbols()
|
| /linux/tools/testing/selftests/cgroup/ |
| H A D | test_zswap.c | 322 static int test_zswap_writeback_one(const char *cgroup, bool wb) in test_zswap_writeback_one() argument 332 if (cg_run(cgroup, attempt_writeback, (void *) &wb)) in test_zswap_writeback_one() 340 if (wb != !!zswpwb_after) { in test_zswap_writeback_one() 342 zswpwb_after, wb ? "enabled" : "disabled"); in test_zswap_writeback_one() 350 static int test_zswap_writeback(const char *root, bool wb) in test_zswap_writeback() argument 363 if (cg_write(test_group, "memory.zswap.writeback", wb ? "1" : "0")) in test_zswap_writeback() 366 if (test_zswap_writeback_one(test_group, wb)) in test_zswap_writeback() 385 if (test_zswap_writeback_one(test_group_child, wb)) in test_zswap_writeback()
|
| /linux/drivers/media/platform/mediatek/vcodec/encoder/venc/ |
| H A D | venc_vp8_if.c | 155 struct venc_vp8_vpu_buf *wb = inst->vsi->work_bufs; in vp8_enc_alloc_work_buf() local 158 if (wb[i].size == 0) in vp8_enc_alloc_work_buf() 170 inst->work_bufs[i].size = wb[i].size; in vp8_enc_alloc_work_buf() 189 wb[i].vpua); in vp8_enc_alloc_work_buf() 190 memcpy(inst->work_bufs[i].va, tmp_va, wb[i].size); in vp8_enc_alloc_work_buf() 192 wb[i].iova = inst->work_bufs[i].dma_addr; in vp8_enc_alloc_work_buf()
|
| /linux/drivers/md/ |
| H A D | dm-writecache.c | 1656 struct writeback_struct *wb = container_of(bio, struct writeback_struct, bio); in writecache_writeback_endio() local 1657 struct dm_writecache *wc = wb->wc; in writecache_writeback_endio() 1663 list_add_tail(&wb->endio_entry, &wc->endio_list); in writecache_writeback_endio() 1684 struct writeback_struct *wb; in __writecache_endio_pmem() local 1689 wb = list_entry(list->next, struct writeback_struct, endio_entry); in __writecache_endio_pmem() 1690 list_del(&wb->endio_entry); in __writecache_endio_pmem() 1692 if (unlikely(wb->bio.bi_status != BLK_STS_OK)) in __writecache_endio_pmem() 1693 writecache_error(wc, blk_status_to_errno(wb->bio.bi_status), in __writecache_endio_pmem() 1694 "write error %d", wb->bio.bi_status); in __writecache_endio_pmem() 1697 e = wb->wc_list[i]; in __writecache_endio_pmem() [all …]
|
| /linux/drivers/gpu/drm/i915/ |
| H A D | i915_active.c | 638 struct wait_barrier *wb = container_of(wq, typeof(*wb), base); in barrier_wake() local 640 if (i915_active_is_idle(wb->ref)) { in barrier_wake() 651 struct wait_barrier *wb; in __await_barrier() local 653 wb = kmalloc_obj(*wb); in __await_barrier() 654 if (unlikely(!wb)) in __await_barrier() 659 kfree(wb); in __await_barrier() 663 wb->base.flags = 0; in __await_barrier() 664 wb->base.func = barrier_wake; in __await_barrier() 665 wb->base.private = fence; in __await_barrier() 666 wb->ref = ref; in __await_barrier() [all …]
|
| /linux/drivers/net/ethernet/intel/igc/ |
| H A D | igc_base.h | 24 } wb; member 86 } wb; /* writeback */ member
|
| /linux/drivers/gpu/drm/vkms/ |
| H A D | vkms_writeback.c | 170 struct drm_writeback_connector *wb = &vkms_output->wb_connector; in vkms_enable_writeback_connector() local 181 drm_connector_helper_add(&wb->base, &vkms_wb_conn_helper_funcs); in vkms_enable_writeback_connector() 183 return drmm_writeback_connector_init(&vkmsdev->drm, wb, in vkms_enable_writeback_connector()
|