| /linux/mm/kmsan/ |
| H A D | instrumentation.c | 140 depot_stack_handle_t *origin) in get_param0_metadata() argument 145 *origin = ctx->cstate.param_origin_tls[0]; in get_param0_metadata() 148 static inline void set_retval_metadata(u64 shadow, depot_stack_handle_t origin) in set_retval_metadata() argument 153 ctx->cstate.retval_origin_tls = origin; in set_retval_metadata() 160 depot_stack_handle_t origin; in __msan_memmove() local 164 get_param0_metadata(&shadow, &origin); in __msan_memmove() 176 set_retval_metadata(shadow, origin); in __msan_memmove() 185 depot_stack_handle_t origin; in __msan_memcpy() local 189 get_param0_metadata(&shadow, &origin); in __msan_memcpy() 203 set_retval_metadata(shadow, origin); in __msan_memcpy() [all …]
|
| H A D | init.c | 99 struct page *shadow, *origin; member 118 struct page *shadow, *origin; in kmsan_memblock_free_pages() local 124 if (!held_back[order].origin) { in kmsan_memblock_free_pages() 125 held_back[order].origin = page; in kmsan_memblock_free_pages() 129 origin = held_back[order].origin; in kmsan_memblock_free_pages() 130 kmsan_setup_meta(page, shadow, origin, order); in kmsan_memblock_free_pages() 133 held_back[order].origin = NULL; in kmsan_memblock_free_pages() 170 struct page *page, *shadow, *origin; in do_collection() local 175 origin = smallstack_pop(&collect); in do_collection() 176 kmsan_setup_meta(page, shadow, origin, collect.order); in do_collection() [all …]
|
| H A D | shadow.c | 102 ret.origin = kmsan_get_metadata(address, KMSAN_META_ORIGIN); in kmsan_get_shadow_origin_ptr() 109 ret.origin = dummy_store_page; in kmsan_get_shadow_origin_ptr() 113 ret.origin = dummy_load_page; in kmsan_get_shadow_origin_ptr() 173 struct page *shadow, *origin; in kmsan_alloc_page() local 181 origin = origin_page_for(page); in kmsan_alloc_page() 185 __memset(page_address(origin), 0, PAGE_SIZE * pages); in kmsan_alloc_page() 202 ((depot_stack_handle_t *)page_address(origin))[i] = handle; in kmsan_alloc_page() 278 void *shadow, *origin; in kmsan_init_alloc_meta_for_range() local 285 origin = memblock_alloc_or_panic(size, PAGE_SIZE); in kmsan_init_alloc_meta_for_range() 292 origin_p = virt_to_page((char *)origin + addr); in kmsan_init_alloc_meta_for_range() [all …]
|
| H A D | report.c | 86 void kmsan_print_origin(depot_stack_handle_t origin) in kmsan_print_origin() argument 96 if (!origin) in kmsan_print_origin() 100 nr_entries = stack_depot_fetch(origin, &entries); in kmsan_print_origin() 101 depth = kmsan_depth_from_eb(stack_depot_get_extra_bits(origin)); in kmsan_print_origin() 123 origin = entries[2]; in kmsan_print_origin() 150 void kmsan_report(depot_stack_handle_t origin, void *address, int size, in kmsan_report() argument 164 if (!origin) in kmsan_report() 171 is_uaf = kmsan_uaf_from_eb(stack_depot_get_extra_bits(origin)); in kmsan_report() 196 kmsan_print_origin(origin); in kmsan_report()
|
| H A D | hooks.c | 154 struct page *shadow, *origin; in kmsan_ioremap_page_range() local 165 origin = alloc_pages(gfp_mask, 1); in kmsan_ioremap_page_range() 166 if (!shadow || !origin) { in kmsan_ioremap_page_range() 181 vmalloc_origin(start + off + PAGE_SIZE), prot, &origin, in kmsan_ioremap_page_range() 190 origin = NULL; in kmsan_ioremap_page_range() 204 if (origin) in kmsan_ioremap_page_range() 205 __free_pages(origin, 1); in kmsan_ioremap_page_range() 222 struct page *shadow, *origin; in kmsan_iounmap_page_range() local 235 origin = kmsan_vmalloc_to_page_or_null((void *)v_origin); in kmsan_iounmap_page_range() 240 if (origin) in kmsan_iounmap_page_range() [all …]
|
| H A D | kmsan.h | 42 void *shadow, *origin; member 55 void kmsan_print_origin(depot_stack_handle_t origin); 75 void kmsan_report(depot_stack_handle_t origin, void *address, int size, 159 u32 origin, bool checked); 170 struct page *origin, int order);
|
| H A D | core.c | 192 u32 origin, bool checked) in kmsan_internal_set_shadow_origin() argument 234 if (origin || !aligned_shadow[i]) in kmsan_internal_set_shadow_origin() 235 origin_start[i] = origin; in kmsan_internal_set_shadow_origin() 258 depot_stack_handle_t *origin = NULL; in kmsan_internal_check_memory() local 302 origin = kmsan_get_metadata((void *)(addr64 + pos + i), in kmsan_internal_check_memory() 304 KMSAN_WARN_ON(!origin); in kmsan_internal_check_memory() 305 new_origin = *origin; in kmsan_internal_check_memory()
|
| /linux/drivers/gpu/drm/i915/display/ |
| H A D | intel_frontbuffer.c | 85 enum fb_op_origin origin) in frontbuffer_flush() argument 95 trace_intel_frontbuffer_flush(display, frontbuffer_bits, origin); in frontbuffer_flush() 100 intel_psr_flush(display, frontbuffer_bits, origin); in frontbuffer_flush() 101 intel_fbc_flush(display, frontbuffer_bits, origin); in frontbuffer_flush() 127 enum fb_op_origin origin, in __intel_fb_invalidate() argument 132 if (origin == ORIGIN_CS) { in __intel_fb_invalidate() 138 trace_intel_frontbuffer_invalidate(display, frontbuffer_bits, origin); in __intel_fb_invalidate() 141 intel_psr_invalidate(display, frontbuffer_bits, origin); in __intel_fb_invalidate() 143 intel_fbc_invalidate(display, frontbuffer_bits, origin); in __intel_fb_invalidate() 147 enum fb_op_origin origin, in __intel_fb_flush() argument [all …]
|
| /linux/Documentation/admin-guide/device-mapper/ |
| H A D | snapshot.rst | 12 - To merge a snapshot of a block device back into the snapshot's origin 20 the origin device. 24 snapshot, snapshot-origin, and snapshot-merge. 26 - snapshot-origin <origin> 34 - snapshot <origin> <COW device> <persistent?> <chunksize> 37 A snapshot of the <origin> block device is created. Changed chunks of 40 from <origin> for unchanged data. <COW device> will often be 41 smaller than the origin and if it fills up the snapshot will become 55 snapshot-origin or snapshot-merge target must be suspended. A failure to 56 suspend the origin target could result in data corruption. [all …]
|
| H A D | cache.rst | 39 The origin device always contains a copy of the logical block, which 52 1. An origin device - the big, slow one. 66 The origin is divided up into blocks of a fixed size. This block size 89 complete until it has hit both the origin and cache devices. Clean 93 to be coherent with the origin device, then all reads are served from 94 the origin device (all reads miss the cache) and all writes are 95 forwarded to the origin device; additionally, write hits cause cache 121 Migrating data between the origin and cache device uses bandwidth. 170 state for all of the origin device (compare with the dirty bitset 181 cache <metadata dev> <cache dev> <origin dev> <block size> [all …]
|
| /linux/drivers/gpu/drm/i915/gem/ |
| H A D | i915_gem_object_frontbuffer.h | 24 enum fb_op_origin origin); 26 enum fb_op_origin origin); 30 enum fb_op_origin origin) in i915_gem_object_flush_frontbuffer() argument 33 __i915_gem_object_flush_frontbuffer(obj, origin); in i915_gem_object_flush_frontbuffer() 38 enum fb_op_origin origin) in i915_gem_object_invalidate_frontbuffer() argument 41 __i915_gem_object_invalidate_frontbuffer(obj, origin); in i915_gem_object_invalidate_frontbuffer()
|
| /linux/drivers/infiniband/hw/hns/ |
| H A D | hns_roce_common.h | 42 #define roce_get_field(origin, mask, shift) \ argument 43 ((le32_to_cpu(origin) & (mask)) >> (u32)(shift)) 45 #define roce_get_bit(origin, shift) \ argument 46 roce_get_field((origin), (1ul << (shift)), (shift)) 48 #define roce_set_field(origin, mask, shift, val) \ argument 50 (origin) &= ~cpu_to_le32(mask); \ 51 (origin) |= \ 55 #define roce_set_bit(origin, shift, val) \ argument 56 roce_set_field((origin), (1ul << (shift)), (shift), (val))
|
| /linux/fs/netfs/ |
| H A D | objects.c | 21 enum netfs_io_origin origin) in netfs_alloc_request() argument 42 rreq->origin = origin; in netfs_alloc_request() 57 if (origin == NETFS_READAHEAD || in netfs_alloc_request() 58 origin == NETFS_READPAGE || in netfs_alloc_request() 59 origin == NETFS_READ_GAPS || in netfs_alloc_request() 60 origin == NETFS_READ_SINGLE || in netfs_alloc_request() 61 origin == NETFS_READ_FOR_WRITE || in netfs_alloc_request() 62 origin == NETFS_UNBUFFERED_READ || in netfs_alloc_request() 63 origin == NETFS_DIO_READ) { in netfs_alloc_request()
|
| H A D | read_collect.c | 196 if (rreq->origin == NETFS_READAHEAD || in netfs_collect_read_results() 197 rreq->origin == NETFS_READPAGE || in netfs_collect_read_results() 198 rreq->origin == NETFS_READ_FOR_WRITE) in netfs_collect_read_results() 347 if (rreq->origin == NETFS_UNBUFFERED_READ || in netfs_rreq_assess_dio() 348 rreq->origin == NETFS_DIO_READ) { in netfs_rreq_assess_dio() 368 if (rreq->origin == NETFS_UNBUFFERED_READ || in netfs_rreq_assess_dio() 369 rreq->origin == NETFS_DIO_READ) in netfs_rreq_assess_dio() 426 switch (rreq->origin) { in netfs_read_collection() 487 (rreq->origin == NETFS_READAHEAD || in netfs_read_subreq_progress() 488 rreq->origin == NETFS_READPAGE || in netfs_read_subreq_progress() [all …]
|
| H A D | write_collect.c | 28 rreq->origin, rreq->error); in netfs_dump_request() 124 if (wreq->origin == NETFS_PGPRIV2_COPY_TO_CACHE) { in netfs_writeback_unlock_folios() 213 if (wreq->origin == NETFS_WRITEBACK || in netfs_collect_write_results() 214 wreq->origin == NETFS_WRITETHROUGH || in netfs_collect_write_results() 215 wreq->origin == NETFS_PGPRIV2_COPY_TO_CACHE) in netfs_collect_write_results() 402 if ((wreq->origin == NETFS_UNBUFFERED_WRITE || in netfs_write_collection() 403 wreq->origin == NETFS_DIO_WRITE) && in netfs_write_collection() 407 if (wreq->origin == NETFS_DIO_WRITE && in netfs_write_collection() 420 if (wreq->origin == NETFS_DIO_WRITE) in netfs_write_collection()
|
| /linux/security/loadpin/ |
| H A D | loadpin.c | 27 static void report_load(const char *origin, struct file *file, char *operation) in report_load() argument 35 origin, operation, in report_load() 132 const char *origin = kernel_read_file_id_str(id); in loadpin_check() local 139 report_load(origin, file, "pinning-excluded"); in loadpin_check() 146 report_load(origin, NULL, "old-api-pinning-ignored"); in loadpin_check() 150 report_load(origin, NULL, "old-api-denied"); in loadpin_check() 173 report_load(origin, file, "pinned"); in loadpin_check() 179 report_load(origin, file, "pinning-ignored"); in loadpin_check() 183 report_load(origin, file, "denied"); in loadpin_check()
|
| /linux/drivers/net/ethernet/hisilicon/hns/ |
| H A D | hns_dsaf_reg.h | 1043 #define dsaf_set_field(origin, mask, shift, val) \ argument 1045 (origin) &= (~(mask)); \ 1046 (origin) |= (((val) << (shift)) & (mask)); \ 1049 #define dsaf_set_bit(origin, shift, val) \ argument 1050 dsaf_set_field((origin), (1ull << (shift)), (shift), (val)) 1055 u32 origin = dsaf_read_reg(base, reg); in dsaf_set_reg_field() local 1057 dsaf_set_field(origin, mask, shift, val); in dsaf_set_reg_field() 1058 dsaf_write_reg(base, reg, origin); in dsaf_set_reg_field() 1067 #define dsaf_get_field(origin, mask, shift) (((origin) & (mask)) >> (shift)) argument 1069 #define dsaf_get_bit(origin, shift) \ argument [all …]
|
| /linux/drivers/md/ |
| H A D | dm-snap.c | 57 struct dm_dev *origin; member 179 return s->origin; in dm_snap_origin() 319 struct origin { struct 393 static struct origin *__lookup_origin(struct block_device *origin) in __lookup_origin() argument 396 struct origin *o; in __lookup_origin() 398 ol = &_origins[origin_hash(origin)]; in __lookup_origin() 400 if (bdev_equal(o->bdev, origin)) in __lookup_origin() 406 static void __insert_origin(struct origin *o) in __insert_origin() 413 static struct dm_origin *__lookup_dm_origin(struct block_device *origin) in __lookup_dm_origin() argument 418 ol = &_dm_origins[origin_hash(origin)]; in __lookup_dm_origin() [all …]
|
| /linux/fs/ |
| H A D | pnode.c | 178 struct mount *origin) in __propagation_next() argument 183 if (master == origin->mnt_master) { in __propagation_next() 185 return (next == origin) ? NULL : next; in __propagation_next() 205 struct mount *origin) in propagation_next() argument 211 return __propagation_next(m, origin); in propagation_next() 215 struct mount *origin) in skip_propagation_subtree() argument 220 struct mount *p = __propagation_next(m, origin); in skip_propagation_subtree() 223 p = __propagation_next(p, origin); in skip_propagation_subtree() 228 static struct mount *next_group(struct mount *m, struct mount *origin) in next_group() argument 236 if (m->mnt_group_id == origin->mnt_group_id) { in next_group() [all …]
|
| /linux/block/partitions/ |
| H A D | msdos.c | 251 sector_t offset, sector_t size, int origin) in parse_solaris_x86() argument 269 snprintf(tmp, sizeof(tmp), " %s%d: <solaris:", state->name, origin); in parse_solaris_x86() 356 sector_t offset, sector_t size, int origin, char *flavour, in parse_bsd() argument 372 snprintf(tmp, sizeof(tmp), " %s%d: <%s:", state->name, origin, flavour); in parse_bsd() 410 sector_t offset, sector_t size, int origin) in parse_freebsd() argument 413 parse_bsd(state, offset, size, origin, "bsd", BSD_MAXPARTITIONS); in parse_freebsd() 418 sector_t offset, sector_t size, int origin) in parse_netbsd() argument 421 parse_bsd(state, offset, size, origin, "netbsd", BSD_MAXPARTITIONS); in parse_netbsd() 426 sector_t offset, sector_t size, int origin) in parse_openbsd() argument 429 parse_bsd(state, offset, size, origin, "openbsd", in parse_openbsd() [all …]
|
| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_gt_sriov_pf_monitor.c | 46 char origin[8]; in pf_handle_vf_threshold_event() local 50 xe_sriov_function_name(vfid, origin, sizeof(origin)); in pf_handle_vf_threshold_event() 55 threshold, origin); in pf_handle_vf_threshold_event() 60 origin, xe_gt_sriov_pf_config_get_threshold(gt, vfid, e), in pf_handle_vf_threshold_event()
|
| H A D | xe_gt_sriov_pf_service.h | 21 int xe_gt_sriov_pf_service_process_request(struct xe_gt *gt, u32 origin, 26 xe_gt_sriov_pf_service_process_request(struct xe_gt *gt, u32 origin, in xe_gt_sriov_pf_service_process_request() argument
|
| /linux/drivers/net/ethernet/hisilicon/ |
| H A D | hns_mdio.c | 108 #define mdio_set_field(origin, mask, shift, val) \ argument 110 (origin) &= (~((mask) << (shift))); \ 111 (origin) |= (((val) & (mask)) << (shift)); \ 114 #define mdio_get_field(origin, mask, shift) (((origin) >> (shift)) & (mask)) argument 119 u32 origin = mdio_read_reg(base, reg); in mdio_set_reg_field() local 121 mdio_set_field(origin, mask, shift, val); in mdio_set_reg_field() 122 mdio_write_reg(base, reg, origin); in mdio_set_reg_field() 130 u32 origin; in mdio_get_reg_field() local 132 origin = mdio_read_reg(base, reg); in mdio_get_reg_field() 133 return mdio_get_field(origin, mask, shift); in mdio_get_reg_field()
|
| /linux/arch/s390/kvm/ |
| H A D | interrupt.c | 240 word = READ_ONCE(gi->origin->u64.word[0]); in gisa_get_ipm_or_restore_iam() 247 } while (!try_cmpxchg(&gi->origin->u64.word[0], &word, _word)); in gisa_get_ipm_or_restore_iam() 282 if (gi->origin) in pending_irqs() 283 pending_mask |= gisa_get_ipm(gi->origin) << IRQ_PEND_IO_ISC_7; in pending_irqs() 1154 if (gi->origin && gisa_tac_ipm_gisc(gi->origin, isc)) { in __deliver_io() 1258 if (gi->origin && in kvm_s390_handle_wait() 1701 if (!gi->origin) in get_top_gisa_isc() 1704 active_mask = (isc_mask & gisa_get_ipm(gi->origin) << 24) << 32; in get_top_gisa_isc() 1707 if (gisa_tac_ipm_gisc(gi->origin, isc)) in get_top_gisa_isc() 1748 gisa_set_ipm_gisc(gi->origin, isc); in kvm_s390_get_io_int() [all …]
|
| /linux/arch/um/drivers/ |
| H A D | mconsole_user.c | 48 msg.msg_name = &(req->origin); in mconsole_reply_v0() 85 req->originlen = sizeof(req->origin); in mconsole_get_request() 87 (struct sockaddr *) req->origin, &req->originlen); in mconsole_get_request() 159 (struct sockaddr *) req->origin, req->originlen); in mconsole_reply_len()
|