| /linux/tools/testing/selftests/mm/ |
| H A D | charge_reserved_hugetlb.sh | 170 local reserve="$9" 196 "$cgroup" "$path" "$method" "$private" "-l" "$reserve" 2>&1 | tee $output & 218 elif [[ "$reserve" != "-n" ]]; then 229 "$cgroup" "$path" "$method" "$private" "$reserve" 232 if [[ "$reserve" != "-n" ]]; then 288 local reserve="${10}" 306 "$reserve" 336 local reserve="${15}" 360 "$expect_failure" "$reserve" 377 "$expect_failure" "$reserve" [all...] |
| H A D | write_hugetlb_memory.sh | 14 reserve=$9 23 "$private" "$want_sleep" "$reserve"
|
| H A D | uffd-common.c | 146 void *reserve = mmap(NULL, region_size, PROT_NONE, MAP_PRIVATE | MAP_ANONYMOUS, in shmem_allocate_area() local 148 if (reserve == MAP_FAILED) { in shmem_allocate_area() 153 p = reserve; in shmem_allocate_area() 162 munmap(reserve, region_size); in shmem_allocate_area() 173 munmap(reserve, region_size); in shmem_allocate_area()
|
| /linux/arch/arm/mach-omap2/ |
| H A D | board-generic.c | 56 .reserve = omap_reserve, 73 .reserve = omap_reserve, 119 .reserve = rx51_reserve, 137 .reserve = omap_reserve, 154 .reserve = omap_reserve, 171 .reserve = omap_reserve, 187 .reserve = omap_reserve, 206 .reserve = omap_reserve, 223 .reserve = omap_reserve, 241 .reserve = omap_reserve, [all …]
|
| /linux/arch/um/kernel/ |
| H A D | physmem.c | 65 unsigned long reserve = reserve_end - start; in setup_physmem() local 66 unsigned long map_size = len - reserve; in setup_physmem() 69 if (len <= reserve) { in setup_physmem() 71 reserve, len); in setup_physmem() 77 err = os_map_memory((void *) reserve_end, physmem_fd, reserve, in setup_physmem() 94 memblock_reserve(__pa(start), reserve); in setup_physmem()
|
| /linux/arch/x86/include/asm/trace/ |
| H A D | irq_vectors.h | 270 bool reserve), 272 TP_ARGS(irq, is_managed, can_reserve, reserve), 278 __field( bool, reserve ) 285 __entry->reserve = reserve; 290 __entry->reserve) 296 bool can_reserve, bool reserve), \ 297 TP_ARGS(irq, is_managed, can_reserve, reserve), NULL, NULL); \
|
| /linux/drivers/gpu/drm/amd/pm/swsmu/inc/ |
| H A D | smu_v14_0_2_pptable.h | 126 …uint8_t reserve[3]; // Zero filled field r… member 165 …uint8_t reserve[143]; // Zero filled field reserved for future u… member 184 uint8_t reserve[3]; member 198 uint32_t reserve[8]; member
|
| H A D | smu_v11_0_pptable.h | 106 …uint8_t reserve[3]; //Zero filled field reserved for future … member 131 …uint8_t reserve[3]; //Zero filled field reserved for future … member 155 uint16_t reserve[6]; //Zero filled field reserved for future use member
|
| H A D | smu_v13_0_pptable.h | 108 …uint8_t reserve[3]; //Zero filled field reserved for future … member 133 …uint8_t reserve[3]; //Zero filled field reserved for future … member 157 uint16_t reserve[6]; //Zero filled field reserved for future use member
|
| H A D | smu_v11_0_7_pptable.h | 136 …uint8_t reserve[3]; //Zero filled field reserved for future … member 166 …uint8_t reserve[3]; //Zero filled field reserved for future … member 189 uint16_t reserve[8]; //Zero filled field reserved for future use member
|
| H A D | smu_v13_0_7_pptable.h | 145 uint8_t reserve[3]; //Zero filled field reserved for future use member 189 uint32_t reserve[45]; member
|
| /linux/drivers/md/bcache/ |
| H A D | alloc.c | 376 long bch_bucket_alloc(struct cache *ca, unsigned int reserve, bool wait) in bch_bucket_alloc() argument 389 fifo_pop(&ca->free[reserve], r)) in bch_bucket_alloc() 393 trace_bcache_alloc_fail(ca, reserve); in bch_bucket_alloc() 409 !fifo_pop(&ca->free[reserve], r)); in bch_bucket_alloc() 416 trace_bcache_alloc(ca, reserve); in bch_bucket_alloc() 439 if (reserve <= RESERVE_PRIO) { in bch_bucket_alloc() 476 int __bch_bucket_alloc_set(struct cache_set *c, unsigned int reserve, in __bch_bucket_alloc_set() argument 491 b = bch_bucket_alloc(ca, reserve, wait); in __bch_bucket_alloc_set() 504 int bch_bucket_alloc_set(struct cache_set *c, unsigned int reserve, in bch_bucket_alloc_set() argument 510 ret = __bch_bucket_alloc_set(c, reserve, k, wait); in bch_bucket_alloc_set()
|
| /linux/arch/powerpc/platforms/pseries/ |
| H A D | vio.c | 94 struct vio_cmo_pool reserve; member 221 vio_cmo.reserve.size += tmp; in vio_cmo_dealloc() 249 if (excess_freed && (vio_cmo.desired > vio_cmo.reserve.size)) { in vio_cmo_dealloc() 250 tmp = min(excess_freed, (vio_cmo.desired - vio_cmo.reserve.size)); in vio_cmo_dealloc() 253 vio_cmo.reserve.size += tmp; in vio_cmo_dealloc() 295 vio_cmo.reserve.size += tmp; in vio_cmo_entitlement_update() 461 cmo->reserve.size = cmo->min; in vio_cmo_balance() 469 cmo->reserve.size += (viodev->cmo.entitled - in vio_cmo_balance() 475 cmo->excess.size = cmo->entitled - cmo->reserve.size; in vio_cmo_balance() 673 vio_cmo.reserve.size -= viodev->cmo.entitled - desired; in vio_cmo_set_dev_desired() [all …]
|
| /linux/fs/jffs2/ |
| H A D | compr_rubin.c | 34 unsigned int reserve; member 49 unsigned reserve) in init_pushpull() argument 54 pp->reserve = reserve; in init_pushpull() 59 if (pp->ofs >= pp->buflen - (use_reserved?0:pp->reserve)) in pushbit()
|
| /linux/drivers/pinctrl/ |
| H A D | pinctrl-utils.c | 22 unsigned int *num_maps, unsigned int reserve) in pinctrl_utils_reserve_map() argument 25 unsigned int new_num = *num_maps + reserve; in pinctrl_utils_reserve_map()
|
| H A D | pinctrl-utils.h | 19 unsigned int *num_maps, unsigned int reserve);
|
| /linux/Documentation/filesystems/ |
| H A D | devpts.rst | 28 kernel.pty.reserve = 1024 - reserved for filesystems mounted from the initial mount namespace 34 ``sysctl kernel.pty.reserve``.
|
| /linux/drivers/nvdimm/ |
| H A D | pfn_devs.c | 665 unsigned long reserve = nd_info_block_reserve() >> PAGE_SHIFT; in init_altmap_reserve() local 668 reserve += base_pfn - SUBSECTION_ALIGN_DOWN(base_pfn); in init_altmap_reserve() 669 return reserve; in init_altmap_reserve() 680 u32 reserve = nd_info_block_reserve(); in __nvdimm_setup_pfn() local 687 .reserve = init_altmap_reserve(base), in __nvdimm_setup_pfn() 697 if (offset < reserve) in __nvdimm_setup_pfn() 708 altmap->free = PHYS_PFN(offset - reserve); in __nvdimm_setup_pfn()
|
| /linux/net/rose/ |
| H A D | rose_subr.c | 107 int reserve; in rose_write_internal() local 109 reserve = AX25_BPQ_HEADER_LEN + AX25_MAX_HEADER_LEN + 1; in rose_write_internal() 124 skb = alloc_skb(reserve + len + maxfaclen, GFP_ATOMIC); in rose_write_internal() 131 skb_reserve(skb, reserve); in rose_write_internal()
|
| /linux/drivers/scsi/ |
| H A D | sg.c | 148 Sg_scatter_hold reserve; /* buffer held for this file descriptor */ member 742 if (hp->dxfer_len > sfp->reserve.bufflen) { in sg_new_write() 1029 if (val != sfp->reserve.bufflen) { in sg_ioctl_common() 1036 sg_remove_scat(sfp, &sfp->reserve); in sg_ioctl_common() 1042 val = min_t(int, sfp->reserve.bufflen, in sg_ioctl_common() 1227 rsv_schp = &sfp->reserve; in sg_vma_fault() 1273 rsv_schp = &sfp->reserve; in sg_mmap() 1732 Sg_scatter_hold *rsv_schp = &sfp->reserve; in sg_start_req() 2001 Sg_scatter_hold *schp = &sfp->reserve; in sg_build_reserve() 2020 Sg_scatter_hold *rsv_schp = &sfp->reserve; in sg_link_reserve() [all …]
|
| /linux/drivers/net/ethernet/netronome/nfp/ |
| H A D | ccm_mbox.c | 154 int reserve, i, cnt; in nfp_ccm_mbox_copy_in() local 182 reserve = nfp_ccm_mbox_maxlen(skb) - len; in nfp_ccm_mbox_copy_in() 183 if (reserve > 0) { in nfp_ccm_mbox_copy_in() 186 reserve); in nfp_ccm_mbox_copy_in() 187 off += 4 + reserve; in nfp_ccm_mbox_copy_in()
|
| /linux/rust/kernel/alloc/ |
| H A D | kvec.rs | 326 self.reserve(1, flags)?; in push() 504 v.reserve(capacity, flags)?; in with_capacity() 622 pub fn reserve(&mut self, additional: usize, flags: Flags) -> Result<(), AllocError> { in reserve() function 745 self.reserve(n, flags)?; in extend_with() 780 self.reserve(other.len(), flags)?; in extend_from_slice()
|
| /linux/Documentation/mm/ |
| H A D | hugetlbfs_reserv.rst | 19 'reserve' huge pages at mmap() time to ensure that huge pages would be 21 describe how huge page reserve processing is done in the v4.10 kernel. 39 A reserve map is described by the structure:: 50 There is one reserve map for each huge page mapping in the system. 90 Specifically, vma->vm_private_data. This reserve map is created at the 239 The routine vma_commit_reservation() is then called to adjust the reserve 243 in the reserve map already existed so no change is made. However, if there 247 It is possible that the reserve map could have been changed between the call 254 a race is detected, the subpool and global reserve counts are adjusted to 279 reserve count must be restored. [all …]
|
| /linux/arch/arm/mach-realtek/ |
| H A D | rtd1195.c | 37 .reserve = rtd1195_reserve,
|
| /linux/include/uapi/linux/ |
| H A D | um_timetravel.h | 245 char reserve[128]; /* reserved for future usage */ member
|