/linux/drivers/media/i2c/ |
H A D | aptina-pll.c | 16 const struct aptina_pll_limits *limits, in aptina_pll_calculate() argument 29 if (pll->ext_clock < limits->ext_clock_min || in aptina_pll_calculate() 30 pll->ext_clock > limits->ext_clock_max) { in aptina_pll_calculate() 35 if (pll->pix_clock == 0 || pll->pix_clock > limits->pix_clock_max) { in aptina_pll_calculate() 55 mf_min = DIV_ROUND_UP(limits->m_min, pll->m); in aptina_pll_calculate() 56 mf_min = max(mf_min, limits->out_clock_min / in aptina_pll_calculate() 57 (pll->ext_clock / limits->n_min * pll->m)); in aptina_pll_calculate() 58 mf_min = max(mf_min, limits->n_min * limits->p1_min / div); in aptina_pll_calculate() 59 mf_max = limits->m_max / pll->m; in aptina_pll_calculate() 60 mf_max = min(mf_max, limits->out_clock_max / in aptina_pll_calculate() [all …]
|
/linux/drivers/net/ipa/ |
H A D | ipa_resource.c | 35 /* We program at most 8 source or destination resource group limits */ in ipa_resource_limits_valid() 50 if (resource->limits[j].min || resource->limits[j].max) in ipa_resource_limits_valid() 63 if (resource->limits[j].min || resource->limits[j].max) in ipa_resource_limits_valid() 99 ylimits = group_count == 1 ? NULL : &resource->limits[1]; in ipa_resource_config_src() 101 &resource->limits[0], ylimits); in ipa_resource_config_src() 106 ylimits = group_count == 3 ? NULL : &resource->limits[3]; in ipa_resource_config_src() 108 &resource->limits[2], ylimits); in ipa_resource_config_src() 113 ylimits = group_count == 5 ? NULL : &resource->limits[5]; in ipa_resource_config_src() 115 &resource->limits[4], ylimits); in ipa_resource_config_src() 120 ylimits = group_count == 7 ? NULL : &resource->limits[7]; in ipa_resource_config_src() [all …]
|
/linux/drivers/infiniband/hw/mthca/ |
H A D | mthca_main.c | 168 mdev->limits.mtt_seg_size = (1 << log_mtts_per_seg) * 8; in mthca_dev_lim() 196 mdev->limits.num_ports = dev_lim->num_ports; in mthca_dev_lim() 197 mdev->limits.vl_cap = dev_lim->max_vl; in mthca_dev_lim() 198 mdev->limits.mtu_cap = dev_lim->max_mtu; in mthca_dev_lim() 199 mdev->limits.gid_table_len = dev_lim->max_gids; in mthca_dev_lim() 200 mdev->limits.pkey_table_len = dev_lim->max_pkeys; in mthca_dev_lim() 201 mdev->limits.local_ca_ack_delay = dev_lim->local_ca_ack_delay; in mthca_dev_lim() 207 mdev->limits.max_sg = min_t(int, dev_lim->max_sg, in mthca_dev_lim() 214 mdev->limits.max_wqes = dev_lim->max_qp_sz; in mthca_dev_lim() 215 mdev->limits.max_qp_init_rdma = dev_lim->max_requester_per_qp; in mthca_dev_lim() [all …]
|
H A D | mthca_profile.c | 94 profile[MTHCA_RES_MTT].size = dev->limits.mtt_seg_size; in mthca_make_profile() 173 dev->limits.num_qps = profile[i].num; in mthca_make_profile() 178 dev->limits.num_eecs = profile[i].num; in mthca_make_profile() 183 dev->limits.num_srqs = profile[i].num; in mthca_make_profile() 188 dev->limits.num_cqs = profile[i].num; in mthca_make_profile() 199 dev->limits.num_eqs = profile[i].num; in mthca_make_profile() 212 dev->limits.num_mgms = profile[i].num >> 1; in mthca_make_profile() 213 dev->limits.num_amgms = profile[i].num >> 1; in mthca_make_profile() 220 dev->limits.num_mpts = profile[i].num; in mthca_make_profile() 226 dev->limits.num_mtt_segs = profile[i].num; in mthca_make_profile() [all …]
|
H A D | mthca_mr.c | 221 for (i = dev->limits.mtt_seg_size / 8; i < size; i <<= 1) in __mthca_alloc_mtt() 267 mtt->first_seg * dev->limits.mtt_seg_size + in __mthca_write_mtt() 320 mtts = dev->mr_table.tavor_fmr.mtt_base + mtt->first_seg * dev->limits.mtt_seg_size + in mthca_tavor_write_mtt_seg() 339 BUG_ON(s % dev->limits.mtt_seg_size); in mthca_arbel_write_mtt_seg() 342 s / dev->limits.mtt_seg_size, &dma_handle); in mthca_arbel_write_mtt_seg() 475 mr->mtt->first_seg * dev->limits.mtt_seg_size); in mthca_mr_alloc() 489 key & (dev->limits.num_mpts - 1)); in mthca_mr_alloc() 556 (dev->limits.num_mpts - 1)); in mthca_free_mr() 570 dev->limits.num_mpts, in mthca_init_mr_table() 571 ~0, dev->limits.reserved_mrws); in mthca_init_mr_table() [all …]
|
H A D | mthca_srq.c | 215 if (attr->max_wr > dev->limits.max_srq_wqes || in mthca_alloc_srq() 216 attr->max_sge > dev->limits.max_srq_sge) in mthca_alloc_srq() 232 if (!mthca_is_memfree(dev) && (ds > dev->limits.max_desc_sz)) in mthca_alloc_srq() 285 srq->srqn & (dev->limits.num_srqs - 1), in mthca_alloc_srq() 355 srq->srqn & (dev->limits.num_srqs - 1)); in mthca_free_srq() 437 srq = mthca_array_get(&dev->srq_table.srq, srqn & (dev->limits.num_srqs - 1)); in mthca_srq_event() 645 return dev->limits.max_sg; in mthca_max_srq_sge() 661 return min_t(int, dev->limits.max_sg, in mthca_max_srq_sge() 662 ((1 << (fls(dev->limits.max_desc_sz) - 1)) - in mthca_max_srq_sge() 677 dev->limits.num_srqs, in mthca_init_srq_table() [all …]
|
/linux/drivers/clk/ |
H A D | clk-axi-clkgen.c | 66 struct axi_clkgen_limits limits; member 132 static void axi_clkgen_calc_params(const struct axi_clkgen_limits *limits, in axi_clkgen_calc_params() argument 151 d_min = max(DIV_ROUND_UP(fin, limits->fpfd_max), 1); in axi_clkgen_calc_params() 152 d_max = min(fin / limits->fpfd_min, 80); in axi_clkgen_calc_params() 155 fvco_min_fract = limits->fvco_min << fract_shift; in axi_clkgen_calc_params() 156 fvco_max_fract = limits->fvco_max << fract_shift; in axi_clkgen_calc_params() 350 const struct axi_clkgen_limits *limits = &axi_clkgen->limits; in axi_clkgen_set_rate() local 358 axi_clkgen_calc_params(limits, parent_rate, rate, &d, &m, &dout); in axi_clkgen_set_rate() 399 const struct axi_clkgen_limits *limits = &axi_clkgen->limits; in axi_clkgen_determine_rate() local 403 axi_clkgen_calc_params(limits, req->best_parent_rate, req->rate, in axi_clkgen_determine_rate() [all …]
|
/linux/include/linux/ |
H A D | blkdev.h | 350 * Flags automatically inherited when stacking limits. 401 /* atomic write limits */ 518 struct queue_limits limits; member 595 * Protects queue limits and also sysfs attribute read_ahead_kb. 674 #define blk_queue_nonrot(q) (!((q)->limits.features & BLK_FEAT_ROTATIONAL)) 675 #define blk_queue_io_stat(q) ((q)->limits.features & BLK_FEAT_IO_STAT) 677 ((q)->limits.flags & BLK_FLAG_IOSTATS_PASSTHROUGH) 678 #define blk_queue_dax(q) ((q)->limits.features & BLK_FEAT_DAX) 679 #define blk_queue_pci_p2pdma(q) ((q)->limits.features & BLK_FEAT_PCI_P2PDMA) 695 ((q)->limits.features & BLK_FEAT_SKIP_TAGSET_QUIESCE) [all …]
|
H A D | t10-pi.h | 45 rq->q->limits.integrity.interval_exp) in t10_pi_ref_tag() 46 shift = rq->q->limits.integrity.interval_exp; in t10_pi_ref_tag() 70 rq->q->limits.integrity.interval_exp) in ext_pi_ref_tag() 71 shift = rq->q->limits.integrity.interval_exp; in ext_pi_ref_tag()
|
/linux/security/apparmor/include/ |
H A D | resource.h | 5 * This file contains AppArmor resource limits function definitions. 22 * @mask: which hard limits to set 23 * @limits: rlimit values that override task limits 26 * limits specified in @mask will be controlled by apparmor. 30 struct rlimit limits[RLIM_NLIMITS]; member
|
/linux/drivers/md/ |
H A D | dm-table.c | 234 struct queue_limits *limits = data; in device_area_is_invalid() local 238 limits->logical_block_size >> SECTOR_SHIFT; in device_area_is_invalid() 292 limits->logical_block_size, bdev); in device_area_is_invalid() 300 limits->logical_block_size, bdev); in device_area_is_invalid() 423 struct queue_limits *limits = data; in dm_set_device_limits() local 428 DMWARN("%s: Cannot set limits for nonexistent device %pg", in dm_set_device_limits() 438 limits->features |= (q->limits.features & BLK_FEAT_ATOMIC_WRITES); in dm_set_device_limits() 440 if (blk_stack_limits(limits, &q->limits, in dm_set_device_limits() 446 q->limits.physical_block_size, in dm_set_device_limits() 447 q->limits.logical_block_size, in dm_set_device_limits() [all …]
|
H A D | dm-zone.c | 288 * sharing may happen and the zone limits will not be reliable. in device_get_zone_resource_limits() 298 * any zone resource limits. in device_get_zone_resource_limits() 304 * If the target does not map all sequential zones, the limits in device_get_zone_resource_limits() 314 * we do not have limits for this target. in device_get_zone_resource_limits() 316 max_active_zones = disk->queue->limits.max_active_zones; in device_get_zone_resource_limits() 322 max_open_zones = disk->queue->limits.max_open_zones; in device_get_zone_resource_limits() 361 * Determine the max open and max active zone limits for the mapped in dm_set_zones_restrictions() 362 * device by inspecting the zone resource limits and the zones mapped in dm_set_zones_restrictions() 379 DMERR("Could not determine %s zone resource limits", in dm_set_zones_restrictions() 401 if (q->limits.chunk_sectors != lim->chunk_sectors) { in dm_set_zones_restrictions() [all …]
|
/linux/Documentation/admin-guide/sysctl/ |
H A D | user.rst | 15 limits on the number of namespaces and other objects that have 16 per user per user namespace limits. 18 The primary purpose of these limits is to stop programs that 21 intention that the defaults of these limits are set high enough that 22 no program in normal operation should run into these limits. 31 limits in the user namespaces of those users. 34 user namespace does not allow a user to escape their current limits.
|
/linux/drivers/mfd/ |
H A D | max14577.c | 43 * @limits: constraints for charger, matching the MBCICHWRC register 54 * - is always between <limits.min, limits.max>; 60 * is outside of given charger limits) and 'dst' is not set. 62 int maxim_charger_calc_reg_current(const struct maxim_charger_current *limits, in maxim_charger_calc_reg_current() argument 70 if (min_ua > limits->max || max_ua < limits->min) in maxim_charger_calc_reg_current() 73 if (max_ua < limits->high_start) { in maxim_charger_calc_reg_current() 82 /* max_ua is in range: <high_start, infinite>, cut it to limits.max */ in maxim_charger_calc_reg_current() 83 max_ua = min(limits->max, max_ua); in maxim_charger_calc_reg_current() 84 max_ua -= limits->high_start; in maxim_charger_calc_reg_current() 87 * - max_ua >= limits.high_start in maxim_charger_calc_reg_current() [all …]
|
/linux/Documentation/hwmon/ |
H A D | lm77.rst | 27 The LM77 implements 3 limits: low (temp1_min), high (temp1_max) and 29 applies to all 3 limits. The relative difference is stored in a single 31 the limit and its hysteresis is always the same for all 3 limits. 44 hysteresis applies automatically to the low and high limits. 45 * The limits should be set before the hysteresis.
|
H A D | emc1403.rst | 72 The chips implement three limits for each sensor: low (tempX_min), high 74 hysteresis mechanism which applies to all limits. The relative difference 77 all three limits. 90 hysteresis applies automatically to all other limits. 91 * The limits should be set before the hysteresis.
|
/linux/drivers/macintosh/ |
H A D | therm_adt746x.c | 80 u8 limits[3]; member 216 " limits: %d,%d,%d;" in display_stats() 219 th->limits[0], th->limits[1], th->limits[2], in display_stats() 237 int var = th->temps[i] - th->limits[i]; in update_fans_speed() 314 th->limits[i] = default_limits_chip[i] + limit_adjust; in set_limit() 315 write_reg(th, LIMIT_REG[i], th->limits[i]); in set_limit() 317 /* set our limits to normal */ in set_limit() 318 th->limits[i] = default_limits_local[i] + limit_adjust; in set_limit() 357 printk(KERN_INFO "Adjusting limits by %d degrees\n", val); \ 378 BUILD_SHOW_FUNC_INT(sensor1_limit, th->limits[1]) [all …]
|
/linux/drivers/clk/sophgo/ |
H A D | clk-sg2044-pll.c | 73 const struct sg2044_pll_limit *limits; member 165 static int sg2042_pll_compute_postdiv(const struct sg2044_pll_limit *limits, in sg2042_pll_compute_postdiv() argument 177 for_each_pll_limit_range(div2, &limits[PLL_LIMIT_POSTDIV2]) { in sg2042_pll_compute_postdiv() 178 for_each_pll_limit_range(div1, &limits[PLL_LIMIT_POSTDIV1]) { in sg2042_pll_compute_postdiv() 207 static int sg2044_compute_pll_setting(const struct sg2044_pll_limit *limits, in sg2044_compute_pll_setting() argument 217 for_each_pll_limit_range(fbdiv, &limits[PLL_LIMIT_FBDIV]) { in sg2044_compute_pll_setting() 218 for_each_pll_limit_range(refdiv, &limits[PLL_LIMIT_REFDIV]) { in sg2044_compute_pll_setting() 221 if (!sg2044_clk_fit_limit(vco, &limits[PLL_LIMIT_FOUTVCO])) in sg2044_compute_pll_setting() 224 ret = sg2042_pll_compute_postdiv(limits, in sg2044_compute_pll_setting() 268 target = clamp(req->rate, pll->pll.limits[PLL_LIMIT_FOUT].min, in sg2044_pll_determine_rate() [all …]
|
/linux/Documentation/admin-guide/cgroup-v1/ |
H A D | rdma.rst | 47 But nothing limits to have it more. At present hundreds of RDMA devices per 70 (b) No resource limits were configured, but IB/RDMA stack tries to 72 running without limits and later on when limits are enforced during uncharging, 75 Resource pool is destroyed if all the resource limits are set to max and 81 IB stack honors limits enforced by the rdma controller. When application 82 query about maximum resource limits of IB device, it returns minimum of
|
/linux/drivers/gpu/drm/exynos/ |
H A D | exynos_drm_ipp.c | 205 * exynos_drm_ipp_get_limits_ioctl - get ipp module limits 244 if (copy_to_user((void __user *)ptr, format->limits, in exynos_drm_ipp_get_limits_ioctl() 245 sizeof(*format->limits) * format->num_limits)) in exynos_drm_ipp_get_limits_ioctl() 430 static void __get_size_limit(const struct drm_exynos_ipp_limit *limits, in __get_size_limit() argument 434 const struct drm_exynos_ipp_limit *l = limits; in __get_size_limit() 439 for (l = limits; l - limits < num_limits; l++) { in __get_size_limit() 457 DRM_DEBUG_DRIVER("Value %d exceeds HW limits (align %d)\n", in __align_check() 468 DRM_DEBUG_DRIVER("Value %d exceeds HW limits (min %d, max %d)\n", in __size_limit_check() 476 const struct drm_exynos_ipp_limit *limits, unsigned int num_limits, in exynos_drm_ipp_check_size_limits() argument 484 if (!limits) in exynos_drm_ipp_check_size_limits() [all …]
|
/linux/drivers/platform/x86/ |
H A D | intel_ips.c | 246 /* Per-SKU limits */ 321 struct ips_mcp_limits *limits; member 400 * as low as the platform limits will allow (though we could go lower there 590 * mcp_exceeded - check whether we're outside our thermal & power limits 620 * cpu_exceeded - check whether a CPU core is outside its limits 636 if (avg > (ips->limits->core_temp_limit * 100)) in cpu_exceeded() 662 if (ips->mch_avg_temp > (ips->limits->mch_temp_limit * 100)) in mch_exceeded() 672 * verify_limits - verify BIOS provided limits 675 * BIOS can optionally provide non-default limits for power and temp. Check 681 if (ips->mcp_power_limit < ips->limits->mcp_power_limit || in verify_limits() [all …]
|
/linux/include/uapi/linux/ |
H A D | dqblk_xfs.h | 37 #define Q_XGETQUOTA XQM_CMD(3) /* get disk limits and usage */ 38 #define Q_XSETQLIM XQM_CMD(4) /* set disk limits */ 43 #define Q_XGETNEXTQUOTA XQM_CMD(9) /* get disk limits and usage >= ID */ 64 __s32 d_itimer; /* Zero if within inode limits. If 96 * automatically started and stopped. Superusers timer values set the limits 109 * below the soft limit). Superusers warning values set the warning limits 138 #define FS_QUOTA_UDQ_ENFD (1<<1) /* user quota limits enforcement */ 140 #define FS_QUOTA_GDQ_ENFD (1<<3) /* group quota limits enforcement */ 142 #define FS_QUOTA_PDQ_ENFD (1<<5) /* project quota limits enforcement */
|
/linux/Documentation/userspace-api/media/v4l/ |
H A D | selection-api-configuration.rst | 9 settings and hardware limits. 16 the source and the sink may have arbitrary upper and lower size limits. 49 The composing targets refer to a memory buffer. The limits of composing 60 to the bounding limits. Moreover, the driver can perform other 89 to be inserted into a video signal or graphical screen. The limits of 111 driver must adjust the area to fit to the bounding limits. Moreover, the
|
/linux/mm/ |
H A D | shmem_quota.c | 8 * can lose valuable information, user provided limits, because there is 39 * before the soft limits are treated as hard limits (usually resulting 63 * set limits 161 * Load dquot with limits from existing entry, or create the new entry if 214 /* Load the stored limits from the tree */ in shmem_acquire_dquot() 261 * Store limits from dquot in the tree unless it's fake. If it is fake 303 /* Store the limits in the tree */ in shmem_release_dquot()
|
/linux/drivers/regulator/ |
H A D | max14577-regulator.c | 41 const struct maxim_charger_current *limits = in max14577_reg_get_current_limit() local 53 return limits->min; in max14577_reg_get_current_limit() 57 return limits->high_start + reg_data * limits->high_step; in max14577_reg_get_current_limit() 66 const struct maxim_charger_current *limits = in max14577_reg_set_current_limit() local 72 ret = maxim_charger_calc_reg_current(limits, min_uA, max_uA, ®_data); in max14577_reg_set_current_limit()
|