| /linux/block/ |
| H A D | blk-settings.c | 121 min_not_zero(lim->max_hw_zone_append_sectors, in blk_validate_zoned_limits() 416 max_hw_sectors = min_not_zero(lim->max_hw_sectors, in blk_validate_limits() 814 t->max_sectors = min_not_zero(t->max_sectors, b->max_sectors); in blk_stack_limits() 815 t->max_user_sectors = min_not_zero(t->max_user_sectors, in blk_stack_limits() 817 t->max_hw_sectors = min_not_zero(t->max_hw_sectors, b->max_hw_sectors); in blk_stack_limits() 818 t->max_dev_sectors = min_not_zero(t->max_dev_sectors, b->max_dev_sectors); in blk_stack_limits() 831 t->seg_boundary_mask = min_not_zero(t->seg_boundary_mask, in blk_stack_limits() 833 t->virt_boundary_mask = min_not_zero(t->virt_boundary_mask, in blk_stack_limits() 836 t->max_segments = min_not_zero(t->max_segments, b->max_segments); in blk_stack_limits() 837 t->max_discard_segments = min_not_zero(t->max_discard_segments, in blk_stack_limits() [all …]
|
| H A D | blk-mq-cpumap.c | 26 return min_not_zero(num, max_queues); in blk_mq_num_queues()
|
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | bpf_dctcp.c | 16 #define min_not_zero(x, y) ({ \ 131 alpha -= min_not_zero(alpha, alpha >> dctcp_shift_g); in BPF_PROG() 18 #define min_not_zero( global() macro
|
| /linux/drivers/md/ |
| H A D | dm-zone.c | 328 min_not_zero(max_active_zones, zlim->lim->max_active_zones); in device_get_zone_resource_limits() 334 min_not_zero(max_open_zones, zlim->lim->max_open_zones); in device_get_zone_resource_limits()
|
| H A D | dm-delay.c | 259 min_delay = min_not_zero(min_delay, dc->write.delay); in delay_ctr() 272 min_delay = min_not_zero(min_delay, dc->flush.delay); in delay_ctr()
|
| /linux/drivers/nvme/target/ |
| H A D | passthru.c | 105 max_hw_sectors = min_not_zero(pctrl->max_segments << PAGE_SECTORS_SHIFT, in nvmet_passthru_override_id_ctrl() 112 max_hw_sectors = min_not_zero(BIO_MAX_VECS << PAGE_SECTORS_SHIFT, in nvmet_passthru_override_id_ctrl()
|
| /linux/kernel/dma/ |
| H A D | direct.c | 50 u64 dma_limit = min_not_zero( in dma_direct_optimal_gfp_mask() 77 min_not_zero(dev->coherent_dma_mask, dev->bus_dma_limit); in dma_coherent_ok()
|
| H A D | coherent.c | 351 min_not_zero(dev->coherent_dma_mask, dev->bus_dma_limit)) in rmem_dma_device_init()
|
| H A D | mapping.c | 131 return min_not_zero(mask, dev->bus_dma_limit) >= in dma_go_direct() 941 if (min_not_zero(dma_get_mask(dev), dev->bus_dma_limit) < in __dma_addressing_limited()
|
| H A D | swiotlb.c | 1208 phys_limit = min_not_zero(*dev->dma_mask, dev->bus_dma_limit); in swiotlb_find_slots()
|
| /linux/net/ipv4/ |
| H A D | tcp_dctcp.c | 154 alpha -= min_not_zero(alpha, alpha >> dctcp_shift_g); in dctcp_update_alpha()
|
| /linux/include/linux/ |
| H A D | minmax.h | 176 #define min_not_zero(x, y) ({ \ macro
|
| /linux/drivers/scsi/ |
| H A D | sd.c | 893 max_blocks = min_not_zero(sdkp->max_unmap_blocks, in sd_config_discard() 903 max_blocks = min_not_zero(max_blocks, (u32)SD_MAX_WS16_BLOCKS); in sd_config_discard() 912 max_blocks = min_not_zero(max_blocks, (u32)SD_MAX_WS10_BLOCKS); in sd_config_discard() 916 max_blocks = min_not_zero(sdkp->max_ws_blocks, in sd_config_discard() 1132 sdkp->max_ws_blocks = min_not_zero(sdkp->max_ws_blocks, in sd_config_write_same() 1135 sdkp->max_ws_blocks = min_not_zero(sdkp->max_ws_blocks, in sd_config_write_same() 3818 dev_max = min_not_zero(dev_max, sdkp->max_xfer_blocks); in sd_revalidate_disk() 3833 lim->io_opt = min_not_zero(lim->io_opt, in sd_revalidate_disk()
|
| /linux/drivers/nvdimm/ |
| H A D | region_devs.c | 121 _num_flush = min_not_zero(_num_flush, nvdimm->num_flush); in get_flush_data() 397 avail = min_not_zero(avail, nd_pmem_max_contiguous_dpa( in nd_region_allocatable_dpa()
|
| /linux/fs/btrfs/ |
| H A D | zoned.c | 425 max_active_zones = min_not_zero(bdev_max_active_zones(bdev), in btrfs_get_dev_zone_info() 768 fs_info->max_extent_size = min_not_zero(fs_info->max_extent_size, in btrfs_check_zoned_mode() 1437 bg->zone_capacity = min_not_zero(zone_info[0].capacity, zone_info[1].capacity); in btrfs_load_block_group_dup() 1491 bg->zone_capacity = min_not_zero(zone_info[0].capacity, zone_info[1].capacity); in btrfs_load_block_group_raid1()
|
| /linux/drivers/pci/controller/dwc/ |
| H A D | pcie-rcar-gen4.c | 147 changes = min_not_zero(dw->max_link_speed, RCAR_MAX_LINK_SPEED) - 1; in rcar_gen4_pcie_start_link()
|
| /linux/io_uring/ |
| H A D | net.c | 592 .max_len = min_not_zero(sr->len, INT_MAX), in io_send_select_buffer() 1121 arg.max_len = min_not_zero(sel->val, (ssize_t) kmsg->msg.msg_inq); in io_recv_buf_select() 1125 arg.max_len = min_not_zero(arg.max_len, sr->mshot_total_len); in io_recv_buf_select()
|
| H A D | kbuf.c | 258 needed = min_not_zero(needed, (size_t) PEEK_MAX_IMPORT); in io_ring_buffers_peek()
|
| /linux/drivers/acpi/arm64/ |
| H A D | iort.c | 2030 limit = min_not_zero(limit, local_limit); in acpi_iort_dma_get_max_cpu_address() 2039 limit = min_not_zero(limit, local_limit); in acpi_iort_dma_get_max_cpu_address()
|
| /linux/fs/iomap/ |
| H A D | buffered-io.c | 133 unsigned end_blk = min_not_zero( in ifs_find_dirty_range() 917 len = min_not_zero(len, *plen); in iomap_write_begin()
|
| /linux/drivers/thunderbolt/ |
| H A D | tunnel.c | 138 spare = min_not_zero(sw->max_dma_credits, dma_credits); in tb_available_credits() 1904 credits = min_not_zero(dma_credits, nhi->sw->max_dma_credits); in tb_tunnel_alloc_dma()
|
| /linux/drivers/infiniband/ulp/rtrs/ |
| H A D | rtrs-clt.c | 1467 max_pages_per_mr = min_not_zero((u32)max_pages_per_mr, U32_MAX); in query_fast_reg_mode() 1877 clt_path->s.signal_interval = min_not_zero(queue_depth, in rtrs_rdma_conn_established() 1894 clt->max_io_size = min_not_zero(clt_path->max_io_size, in rtrs_rdma_conn_established()
|
| /linux/drivers/dma/ |
| H A D | nbpfaxi.c | 441 max_burst = min_not_zero(nbpf->max_burst_mem_read, in nbpf_xfer_ds()
|
| /linux/drivers/block/drbd/ |
| H A D | drbd_req.c | 1727 et = min_not_zero(dt, ent); in request_timer_fn()
|
| /linux/fs/ |
| H A D | exec.c | 559 min_not_zero(offset_in_page(pos), PAGE_SIZE)); in copy_string_kernel()
|