| /linux/kernel/bpf/ |
| H A D | bpf_local_storage.c | 21 select_bucket(struct bpf_local_storage_map *smap, in select_bucket() argument 24 return &smap->buckets[hash_ptr(local_storage, smap->bucket_log)]; in select_bucket() 27 static int mem_charge(struct bpf_local_storage_map *smap, void *owner, u32 size) in mem_charge() argument 29 struct bpf_map *map = &smap->map; in mem_charge() 34 return map->ops->map_local_storage_charge(smap, owner, size); in mem_charge() 37 static void mem_uncharge(struct bpf_local_storage_map *smap, void *owner, in mem_uncharge() argument 40 struct bpf_map *map = &smap->map; in mem_uncharge() 43 map->ops->map_local_storage_uncharge(smap, owner, size); in mem_uncharge() 47 owner_storage(struct bpf_local_storage_map *smap, voi argument 75 bpf_selem_alloc(struct bpf_local_storage_map * smap,void * owner,void * value,bool charge_mem,bool swap_uptrs,gfp_t gfp_flags) bpf_selem_alloc() argument 158 bpf_local_storage_free(struct bpf_local_storage * local_storage,struct bpf_local_storage_map * smap,bool bpf_ma,bool reuse_now) bpf_local_storage_free() argument 210 struct bpf_local_storage_map *smap; bpf_selem_free_rcu() local 231 bpf_selem_free(struct bpf_local_storage_elem * selem,struct bpf_local_storage_map * smap,bool reuse_now) bpf_selem_free() argument 267 struct bpf_local_storage_map *smap; bpf_selem_free_list() local 289 struct bpf_local_storage_map *smap; bpf_selem_unlink_storage_nolock() local 414 struct bpf_local_storage_map *smap; bpf_selem_unlink_map() local 430 bpf_selem_link_map(struct bpf_local_storage_map * smap,struct bpf_local_storage_elem * selem) bpf_selem_link_map() argument 453 __bpf_local_storage_insert_cache(struct bpf_local_storage * local_storage,struct bpf_local_storage_map * smap,struct bpf_local_storage_elem * selem) __bpf_local_storage_insert_cache() argument 484 bpf_local_storage_alloc(void * owner,struct bpf_local_storage_map * smap,struct bpf_local_storage_elem * first_selem,gfp_t gfp_flags) bpf_local_storage_alloc() argument 557 bpf_local_storage_update(void * owner,struct bpf_local_storage_map * smap,void * value,u64 map_flags,bool swap_uptrs,gfp_t gfp_flags) bpf_local_storage_update() argument 777 struct bpf_local_storage_map *smap = (struct bpf_local_storage_map *)map; bpf_local_storage_map_mem_usage() local 800 struct bpf_local_storage_map *smap; bpf_local_storage_map_alloc() local 862 struct bpf_local_storage_map *smap; bpf_local_storage_map_free() local [all...] |
| /linux/tools/testing/selftests/mm/ |
| H A D | map_populate.c | 35 static void parent_f(int sock, unsigned long *smap, int child) in parent_f() argument 42 *smap = 0x22222BAD; in parent_f() 43 ret = msync(smap, MMAP_SZ, MS_SYNC); in parent_f() 56 static int child_f(int sock, unsigned long *smap, int fd) in child_f() argument 60 smap = mmap(0, MMAP_SZ, PROT_READ | PROT_WRITE, in child_f() 62 BUG_ON(smap == MAP_FAILED, "mmap()"); in child_f() 64 BUG_ON(*smap != 0xdeadbabe, "MAP_PRIVATE | MAP_POPULATE changed file"); in child_f() 72 ksft_test_result(*smap != 0x22222BAD, "MAP_POPULATE COW private page\n"); in child_f() 73 ksft_test_result(*smap == 0xdeadbabe, "The mapping state\n"); in child_f() 83 unsigned long *smap; in main() local [all …]
|
| /linux/net/core/ |
| H A D | bpf_sk_storage.c | 24 struct bpf_local_storage_map *smap; in bpf_sk_storage_lookup() local 31 smap = (struct bpf_local_storage_map *)map; in bpf_sk_storage_lookup() 32 return bpf_local_storage_lookup(sk_storage, smap, cacheit_lockit); in bpf_sk_storage_lookup() 135 struct bpf_local_storage_map *smap, in bpf_sk_storage_clone_elem() argument 140 copy_selem = bpf_selem_alloc(smap, newsk, NULL, false, GFP_ATOMIC); in bpf_sk_storage_clone_elem() 144 if (btf_record_has_field(smap->map.record, BPF_SPIN_LOCK)) in bpf_sk_storage_clone_elem() 145 copy_map_value_locked(&smap->map, SDATA(copy_selem)->data, in bpf_sk_storage_clone_elem() 148 copy_map_value(&smap->map, SDATA(copy_selem)->data, in bpf_sk_storage_clone_elem() 171 struct bpf_local_storage_map *smap; in bpf_sk_storage_clone() local 174 smap = rcu_dereference(SDATA(selem)->smap); in bpf_sk_storage_clone() [all …]
|
| /linux/fs/btrfs/ |
| H A D | bio.c | 561 struct btrfs_io_stripe *smap, int mirror_num) in btrfs_submit_bio() argument 566 bio->bi_iter.bi_sector = smap->physical >> SECTOR_SHIFT; in btrfs_submit_bio() 568 btrfs_bio(bio)->orig_physical = smap->physical; in btrfs_submit_bio() 569 bio->bi_private = smap->dev; in btrfs_submit_bio() 571 btrfs_submit_dev_bio(smap->dev, bio); in btrfs_submit_bio() 608 struct btrfs_io_stripe smap; member 665 btrfs_submit_bio(bio, async->bioc, &async->smap, async->mirror_num); in run_one_async_done() 707 struct btrfs_io_stripe *smap, int mirror_num) in btrfs_wq_submit_bio() argument 718 async->smap = *smap; in btrfs_wq_submit_bio() 755 struct btrfs_io_stripe smap; in btrfs_submit_chunk() local [all …]
|
| H A D | volumes.h | 713 struct btrfs_io_stripe *smap, int *mirror_num_ret); 715 struct btrfs_io_stripe *smap, u64 logical,
|
| H A D | volumes.c | 6693 const struct btrfs_io_stripe *smap, in is_single_device_io() argument 6698 if (!smap) in is_single_device_io() 6881 struct btrfs_io_stripe *smap, int *mirror_num_ret) in btrfs_map_block() argument 7000 if (is_single_device_io(fs_info, smap, map, num_alloc_stripes, &io_geom)) { in btrfs_map_block() 7001 ret = set_io_stripe(fs_info, logical, length, smap, map, &io_geom); in btrfs_map_block() 8570 struct btrfs_io_stripe *smap, in map_raid56_repair_block() argument 8585 smap->dev = bioc->stripes[i].dev; in map_raid56_repair_block() 8586 smap->physical = bioc->stripes[i].physical + in map_raid56_repair_block() 8605 struct btrfs_io_stripe *smap, u64 logical, in btrfs_map_repair_block() argument 8616 &bioc, smap, &mirror_ret); in btrfs_map_repair_block() [all …]
|
| /linux/fs/nilfs2/ |
| H A D | page.c | 244 struct address_space *smap) in nilfs_copy_dirty_pages() argument 253 if (!filemap_get_folios_tag(smap, &index, (pgoff_t)-1, in nilfs_copy_dirty_pages() 299 struct address_space *smap) in nilfs_copy_back_pages() argument 307 n = filemap_get_folios(smap, &start, ~0UL, &fbatch); in nilfs_copy_back_pages() 328 xa_lock_irq(&smap->i_pages); in nilfs_copy_back_pages() 329 f = __xa_erase(&smap->i_pages, index); in nilfs_copy_back_pages() 331 smap->nrpages--; in nilfs_copy_back_pages() 332 xa_unlock_irq(&smap->i_pages); in nilfs_copy_back_pages()
|
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | local_storage.c | |
| /linux/drivers/net/ethernet/mellanox/mlx5/core/ |
| H A D | hwmon.c | 209 DECLARE_BITMAP(smap, BITS_PER_TYPE(sensor_map)); in mlx5_hwmon_init_sensors_indexes() 214 bitmap_from_u64(smap, sensor_map); in mlx5_hwmon_init_sensors_indexes() 216 for_each_set_bit(bit_pos, smap, BITS_PER_TYPE(sensor_map)) { in mlx5_hwmon_init_sensors_indexes()
|
| /linux/fs/xfs/ |
| H A D | xfs_iomap.c | 1647 struct xfs_bmbt_irec smap; in xfs_zoned_buffered_write_iomap_begin() local 1651 &smap)) in xfs_zoned_buffered_write_iomap_begin() 1652 smap.br_startoff = end_fsb; /* fake hole until EOF */ in xfs_zoned_buffered_write_iomap_begin() 1653 if (smap.br_startoff > offset_fsb) { in xfs_zoned_buffered_write_iomap_begin() 1659 smap.br_startoff); in xfs_zoned_buffered_write_iomap_begin() 1662 end_fsb = min(end_fsb, smap.br_startoff); in xfs_zoned_buffered_write_iomap_begin() 1665 smap.br_startoff + smap.br_blockcount); in xfs_zoned_buffered_write_iomap_begin() 1666 xfs_trim_extent(&smap, offset_fsb, in xfs_zoned_buffered_write_iomap_begin() 1668 error = xfs_bmbt_to_iomap(ip, srcmap, &smap, flags, 0, in xfs_zoned_buffered_write_iomap_begin()
|
| /linux/lib/zstd/common/ |
| H A D | cpu.h | 176 B(smap, 20)
|
| /linux/fs/ |
| H A D | dax.c | 2159 const struct iomap *smap = &it_src->iomap; in dax_range_compare_iter() local 2165 len = min(len, min(smap->length, dmap->length)); in dax_range_compare_iter() 2167 if (smap->type == IOMAP_HOLE && dmap->type == IOMAP_HOLE) { in dax_range_compare_iter() 2172 if (smap->type == IOMAP_HOLE || dmap->type == IOMAP_HOLE) { in dax_range_compare_iter() 2178 ret = dax_iomap_direct_access(smap, pos1, ALIGN(pos1 + len, PAGE_SIZE), in dax_range_compare_iter()
|
| /linux/fs/ocfs2/dlm/ |
| H A D | dlmdomain.c | 49 static inline void byte_copymap(u8 dmap[], unsigned long smap[], in byte_copymap() argument 59 if (test_bit(nn, smap)) in byte_copymap()
|
| /linux/arch/x86/kvm/svm/ |
| H A D | svm.c | 4914 bool smep, smap, is_user; in svm_check_emulate_instruction() local 5023 smap = kvm_is_cr4_bit_set(vcpu, X86_CR4_SMAP); in svm_check_emulate_instruction() 5025 if (smap && (!smep || is_user)) { in svm_check_emulate_instruction()
|
| /linux/include/linux/ |
| H A D | bpf.h | 151 int (*map_local_storage_charge)(struct bpf_local_storage_map *smap, 153 void (*map_local_storage_uncharge)(struct bpf_local_storage_map *smap,
|
| /linux/Documentation/virt/kvm/x86/ |
| H A D | mmu.rst | 192 Contains the value of cr4.smap && !cr0.wp for which the page is valid
|
| /linux/arch/x86/kvm/mmu/ |
| H A D | mmu.c | 211 BUILD_MMU_ROLE_REGS_ACCESSOR(cr4, smap, X86_CR4_SMAP); 231 BUILD_MMU_ROLE_ACCESSOR(ext, cr4, smap);
|
| /linux/tools/arch/x86/kcpuid/ |
| H A D | cpuid.csv | 209 0x7, 0, ebx, 20, smap , Supervisor mode access prevention
|