Home
last modified time | relevance | path

Searched full:origin (Results 1 – 25 of 294) sorted by relevance

12345678910>>...12

/linux/Documentation/admin-guide/device-mapper/
H A Dsnapshot.rst12 - To merge a snapshot of a block device back into the snapshot's origin
20 the origin device.
24 snapshot, snapshot-origin, and snapshot-merge.
26 - snapshot-origin <origin>
34 - snapshot <origin> <COW device> <persistent?> <chunksize>
37 A snapshot of the <origin> block device is created. Changed chunks of
40 from <origin> for unchanged data. <COW device> will often be
41 smaller than the origin and if it fills up the snapshot will become
55 snapshot-origin or snapshot-merge target must be suspended. A failure to
56 suspend the origin target could result in data corruption.
[all …]
H A Dcache.rst39 The origin device always contains a copy of the logical block, which
52 1. An origin device - the big, slow one.
66 The origin is divided up into blocks of a fixed size. This block size
89 complete until it has hit both the origin and cache devices. Clean
93 to be coherent with the origin device, then all reads are served from
94 the origin device (all reads miss the cache) and all writes are
95 forwarded to the origin device; additionally, write hits cause cache
121 Migrating data between the origin and cache device uses bandwidth.
170 state for all of the origin device (compare with the dirty bitset
181 cache <metadata dev> <cache dev> <origin dev> <block size>
[all …]
H A Dthin-provisioning.rst186 N.B. If the origin device that you wish to snapshot is active, you
197 identifier for the origin device.
202 between the origin and the snapshot. Indeed the snapshot is no
216 You can use an external **read only** device as an origin for a
218 thin device will be passed through to the origin. Writes trigger
225 You must not write to the origin device if you use this technique!
232 You don't mention the origin at this stage.
240 Append an extra parameter to the thin target specifying the origin::
245 same extra origin parameter.
364 create_snap <dev id> <origin id>
[all …]
/linux/mm/kmsan/
H A Dinstrumentation.c49 /* Get shadow and origin pointers for a memory load with non-standard size. */
59 /* Get shadow and origin pointers for a memory store with non-standard size. */
70 * Declare functions that obtain shadow/origin pointers for loads and stores
135 * get_param0_metadata() and set_retval_metadata() to store the shadow/origin
140 depot_stack_handle_t *origin) in get_param0_metadata() argument
145 *origin = ctx->cstate.param_origin_tls[0]; in get_param0_metadata()
148 static inline void set_retval_metadata(u64 shadow, depot_stack_handle_t origin) in set_retval_metadata() argument
153 ctx->cstate.retval_origin_tls = origin; in set_retval_metadata()
160 depot_stack_handle_t origin; in __msan_memmove() local
164 get_param0_metadata(&shadow, &origin); in __msan_memmove()
[all …]
H A Dinit.c99 struct page *shadow, *origin; member
109 * time - as an origin block. On the third time the incoming block receives its
110 * shadow and origin ranges from the previously saved shadow and origin blocks,
118 struct page *shadow, *origin; in kmsan_memblock_free_pages() local
124 if (!held_back[order].origin) { in kmsan_memblock_free_pages()
125 held_back[order].origin = page; in kmsan_memblock_free_pages()
129 origin = held_back[order].origin; in kmsan_memblock_free_pages()
130 kmsan_setup_meta(page, shadow, origin, order); in kmsan_memblock_free_pages()
133 held_back[order].origin = NULL; in kmsan_memblock_free_pages()
170 struct page *page, *shadow, *origin; in do_collection() local
[all …]
H A Dcore.c133 * origin value. in kmsan_internal_memmove_metadata()
169 * are copied around many times. Origin chains for such structures are in kmsan_internal_chain_origin()
192 u32 origin, bool checked) in kmsan_internal_set_shadow_origin() argument
204 * and origin pages are NULL, or all are non-NULL. in kmsan_internal_set_shadow_origin()
228 * If the new origin is non-zero, assume that the shadow byte is also non-zero, in kmsan_internal_set_shadow_origin()
229 * and unconditionally overwrite the old origin slot. in kmsan_internal_set_shadow_origin()
230 * If the new origin is zero, overwrite the old origin slot iff the in kmsan_internal_set_shadow_origin()
234 if (origin || !aligned_shadow[i]) in kmsan_internal_set_shadow_origin()
235 origin_start[i] = origin; in kmsan_internal_set_shadow_origin()
258 depot_stack_handle_t *origin = NULL; in kmsan_internal_check_memory() local
[all …]
H A Dshadow.c102 ret.origin = kmsan_get_metadata(address, KMSAN_META_ORIGIN); in kmsan_get_shadow_origin_ptr()
109 ret.origin = dummy_store_page; in kmsan_get_shadow_origin_ptr()
113 ret.origin = dummy_load_page; in kmsan_get_shadow_origin_ptr()
119 * Obtain the shadow or origin pointer for the given address, or NULL if there's
173 struct page *shadow, *origin; in kmsan_alloc_page() local
181 origin = origin_page_for(page); in kmsan_alloc_page()
185 __memset(page_address(origin), 0, PAGE_SIZE * pages); in kmsan_alloc_page()
199 * to just fill the origin pages with @handle. in kmsan_alloc_page()
202 ((depot_stack_handle_t *)page_address(origin))[i] = handle; in kmsan_alloc_page()
278 void *shadow, *origin; in kmsan_init_alloc_meta_for_range() local
[all …]
H A Dhooks.c145 * This function creates new shadow/origin pages for the physical pages mapped
146 * into the virtual memory. If those physical pages already had shadow/origin,
154 struct page *shadow, *origin; in kmsan_ioremap_page_range() local
165 origin = alloc_pages(gfp_mask, 1); in kmsan_ioremap_page_range()
166 if (!shadow || !origin) { in kmsan_ioremap_page_range()
181 vmalloc_origin(start + off + PAGE_SIZE), prot, &origin, in kmsan_ioremap_page_range()
190 origin = NULL; in kmsan_ioremap_page_range()
198 * Something went wrong. Clean up shadow/origin pages allocated in kmsan_ioremap_page_range()
204 if (origin) in kmsan_ioremap_page_range()
205 __free_pages(origin, 1); in kmsan_ioremap_page_range()
[all …]
H A Dreport.c86 void kmsan_print_origin(depot_stack_handle_t origin) in kmsan_print_origin() argument
96 if (!origin) in kmsan_print_origin()
100 nr_entries = stack_depot_fetch(origin, &entries); in kmsan_print_origin()
101 depth = kmsan_depth_from_eb(stack_depot_get_extra_bits(origin)); in kmsan_print_origin()
117 * Origin chains deeper than KMSAN_MAX_ORIGIN_DEPTH are in kmsan_print_origin()
123 origin = entries[2]; in kmsan_print_origin()
150 void kmsan_report(depot_stack_handle_t origin, void *address, int size, in kmsan_report() argument
164 if (!origin) in kmsan_report()
171 is_uaf = kmsan_uaf_from_eb(stack_depot_get_extra_bits(origin)); in kmsan_report()
196 kmsan_print_origin(origin); in kmsan_report()
H A Dkmsan.h42 void *shadow, *origin; member
55 void kmsan_print_origin(depot_stack_handle_t origin);
59 * @origin: Stack ID of the uninitialized value.
69 * sharing the same origin. If an uninitialized value is used in a comparison,
73 * @off_last corresponding to different @origin values.
75 void kmsan_report(depot_stack_handle_t origin, void *address, int size,
128 * Pack and unpack the origin chain depth and UAF flag to/from the extra bits
159 u32 origin, bool checked);
170 struct page *origin, int order);
/linux/drivers/gpu/drm/i915/display/
H A Dintel_frontbuffer.c75 * @origin: which operation caused the flush
85 enum fb_op_origin origin) in frontbuffer_flush() argument
95 trace_intel_frontbuffer_flush(display, frontbuffer_bits, origin); in frontbuffer_flush()
100 intel_psr_flush(display, frontbuffer_bits, origin); in frontbuffer_flush()
101 intel_fbc_flush(display, frontbuffer_bits, origin); in frontbuffer_flush()
127 enum fb_op_origin origin, in __intel_fb_invalidate() argument
132 if (origin == ORIGIN_CS) { in __intel_fb_invalidate()
138 trace_intel_frontbuffer_invalidate(display, frontbuffer_bits, origin); in __intel_fb_invalidate()
141 intel_psr_invalidate(display, frontbuffer_bits, origin); in __intel_fb_invalidate()
143 intel_fbc_invalidate(display, frontbuffer_bits, origin); in __intel_fb_invalidate()
[all …]
/linux/Documentation/devicetree/bindings/input/touchscreen/
H A Dsitronix,st1232.yaml54 x-origin = <0>;
56 y-origin = <40>;
63 x-origin = <40>;
65 y-origin = <0>;
72 x-origin = <160>;
74 y-origin = <0>;
/linux/drivers/gpu/drm/i915/gem/
H A Di915_gem_object_frontbuffer.h24 enum fb_op_origin origin);
26 enum fb_op_origin origin);
30 enum fb_op_origin origin) in i915_gem_object_flush_frontbuffer() argument
33 __i915_gem_object_flush_frontbuffer(obj, origin); in i915_gem_object_flush_frontbuffer()
38 enum fb_op_origin origin) in i915_gem_object_invalidate_frontbuffer() argument
41 __i915_gem_object_invalidate_frontbuffer(obj, origin); in i915_gem_object_invalidate_frontbuffer()
/linux/fs/netfs/
H A Dobjects.c21 enum netfs_io_origin origin) in netfs_alloc_request() argument
42 rreq->origin = origin; in netfs_alloc_request()
57 if (origin == NETFS_READAHEAD || in netfs_alloc_request()
58 origin == NETFS_READPAGE || in netfs_alloc_request()
59 origin == NETFS_READ_GAPS || in netfs_alloc_request()
60 origin == NETFS_READ_SINGLE || in netfs_alloc_request()
61 origin == NETFS_READ_FOR_WRITE || in netfs_alloc_request()
62 origin == NETFS_UNBUFFERED_READ || in netfs_alloc_request()
63 origin == NETFS_DIO_READ) { in netfs_alloc_request()
H A Dread_collect.c196 if (rreq->origin == NETFS_READAHEAD || in netfs_collect_read_results()
197 rreq->origin == NETFS_READPAGE || in netfs_collect_read_results()
198 rreq->origin == NETFS_READ_FOR_WRITE) in netfs_collect_read_results()
347 if (rreq->origin == NETFS_UNBUFFERED_READ || in netfs_rreq_assess_dio()
348 rreq->origin == NETFS_DIO_READ) { in netfs_rreq_assess_dio()
368 if (rreq->origin == NETFS_UNBUFFERED_READ || in netfs_rreq_assess_dio()
369 rreq->origin == NETFS_DIO_READ) in netfs_rreq_assess_dio()
426 switch (rreq->origin) { in netfs_read_collection()
487 (rreq->origin == NETFS_READAHEAD || in netfs_read_subreq_progress()
488 rreq->origin == NETFS_READPAGE || in netfs_read_subreq_progress()
[all …]
/linux/drivers/infiniband/hw/hns/
H A Dhns_roce_common.h42 #define roce_get_field(origin, mask, shift) \ argument
43 ((le32_to_cpu(origin) & (mask)) >> (u32)(shift))
45 #define roce_get_bit(origin, shift) \ argument
46 roce_get_field((origin), (1ul << (shift)), (shift))
48 #define roce_set_field(origin, mask, shift, val) \ argument
50 (origin) &= ~cpu_to_le32(mask); \
51 (origin) |= \
55 #define roce_set_bit(origin, shift, val) \ argument
56 roce_set_field((origin), (1ul << (shift)), (shift), (val))
/linux/fs/
H A Dpnode.c178 struct mount *origin) in __propagation_next() argument
183 if (master == origin->mnt_master) { in __propagation_next()
185 return (next == origin) ? NULL : next; in __propagation_next()
197 * @origin: the original mount from where the tree walk initiated
205 struct mount *origin) in propagation_next() argument
211 return __propagation_next(m, origin); in propagation_next()
215 struct mount *origin) in skip_propagation_subtree() argument
220 struct mount *p = __propagation_next(m, origin); in skip_propagation_subtree()
223 p = __propagation_next(p, origin); in skip_propagation_subtree()
228 static struct mount *next_group(struct mount *m, struct mount *origin) in next_group() argument
[all …]
/linux/security/loadpin/
H A Dloadpin.c27 static void report_load(const char *origin, struct file *file, char *operation) in report_load() argument
35 origin, operation, in report_load()
132 const char *origin = kernel_read_file_id_str(id); in loadpin_check() local
139 report_load(origin, file, "pinning-excluded"); in loadpin_check()
146 report_load(origin, NULL, "old-api-pinning-ignored"); in loadpin_check()
150 report_load(origin, NULL, "old-api-denied"); in loadpin_check()
173 report_load(origin, file, "pinned"); in loadpin_check()
179 report_load(origin, file, "pinning-ignored"); in loadpin_check()
183 report_load(origin, file, "denied"); in loadpin_check()
/linux/drivers/gpu/drm/xe/
H A Dxe_gt_sriov_pf_monitor.c46 char origin[8]; in pf_handle_vf_threshold_event() local
50 xe_sriov_function_name(vfid, origin, sizeof(origin)); in pf_handle_vf_threshold_event()
55 threshold, origin); in pf_handle_vf_threshold_event()
60 origin, xe_gt_sriov_pf_config_get_threshold(gt, vfid, e), in pf_handle_vf_threshold_event()
H A Dxe_gt_sriov_pf_service.c248 static int pf_process_handshake_msg(struct xe_gt *gt, u32 origin, in pf_process_handshake_msg()
266 err = xe_sriov_pf_service_handshake_vf(gt_to_xe(gt), origin, wanted_major, wanted_minor,
317 static int pf_process_runtime_query_msg(struct xe_gt *gt, u32 origin, in pf_process_runtime_query_msg()
327 if (!xe_sriov_pf_service_is_negotiated(gt_to_xe(gt), origin, 1, 0)) in pf_process_runtime_query_msg()
360 * @origin: VF number that is requesting the service in xe_gt_sriov_pf_service_process_request()
370 int xe_gt_sriov_pf_service_process_request(struct xe_gt *gt, u32 origin, in xe_gt_sriov_pf_service_process_request()
383 action, data, origin);
387 ret = pf_process_handshake_msg(gt, origin, msg, msg_len, response, resp_size);
390 ret = pf_process_runtime_query_msg(gt, origin, msg, msg_len, response, resp_size); in xe_gt_sriov_pf_service_print_runtime()
229 pf_process_handshake_msg(struct xe_gt * gt,u32 origin,const u32 * request,u32 len,u32 * response,u32 size) pf_process_handshake_msg() argument
298 pf_process_runtime_query_msg(struct xe_gt * gt,u32 origin,const u32 * msg,u32 msg_len,u32 * response,u32 resp_size) pf_process_runtime_query_msg() argument
351 xe_gt_sriov_pf_service_process_request(struct xe_gt * gt,u32 origin,const u32 * msg,u32 msg_len,u32 * response,u32 resp_size) xe_gt_sriov_pf_service_process_request() argument
/linux/drivers/net/ethernet/hisilicon/hns/
H A Dhns_dsaf_reg.h1043 #define dsaf_set_field(origin, mask, shift, val) \ argument
1045 (origin) &= (~(mask)); \
1046 (origin) |= (((val) << (shift)) & (mask)); \
1049 #define dsaf_set_bit(origin, shift, val) \ argument
1050 dsaf_set_field((origin), (1ull << (shift)), (shift), (val))
1055 u32 origin = dsaf_read_reg(base, reg); in dsaf_set_reg_field() local
1057 dsaf_set_field(origin, mask, shift, val); in dsaf_set_reg_field()
1058 dsaf_write_reg(base, reg, origin); in dsaf_set_reg_field()
1067 #define dsaf_get_field(origin, mask, shift) (((origin) & (mask)) >> (shift)) argument
1069 #define dsaf_get_bit(origin, shift) \ argument
[all …]
/linux/drivers/net/ethernet/hisilicon/
H A Dhns_mdio.c108 #define mdio_set_field(origin, mask, shift, val) \ argument
110 (origin) &= (~((mask) << (shift))); \
111 (origin) |= (((val) & (mask)) << (shift)); \
114 #define mdio_get_field(origin, mask, shift) (((origin) >> (shift)) & (mask)) argument
119 u32 origin = mdio_read_reg(base, reg); in mdio_set_reg_field() local
121 mdio_set_field(origin, mask, shift, val); in mdio_set_reg_field()
122 mdio_write_reg(base, reg, origin); in mdio_set_reg_field()
130 u32 origin; in mdio_get_reg_field() local
132 origin = mdio_read_reg(base, reg); in mdio_get_reg_field()
133 return mdio_get_field(origin, mask, shift); in mdio_get_reg_field()
/linux/drivers/crypto/intel/qat/qat_common/
H A Dadf_pfvf_msg.h34 * | | Message Origin
45 * | | Message Origin
49 * Message Origin (Should always be 1)
51 * by this driver; these had a Msg Origin of 0 and are ignored by this driver.
66 * Both the Interrupt bit and the Message Origin bit retain the same position
77 * | | Message Origin
/linux/block/partitions/
H A Dmsdos.c251 sector_t offset, sector_t size, int origin) in parse_solaris_x86() argument
269 snprintf(tmp, sizeof(tmp), " %s%d: <solaris:", state->name, origin); in parse_solaris_x86()
356 sector_t offset, sector_t size, int origin, char *flavour, in parse_bsd() argument
372 snprintf(tmp, sizeof(tmp), " %s%d: <%s:", state->name, origin, flavour); in parse_bsd()
410 sector_t offset, sector_t size, int origin) in parse_freebsd() argument
413 parse_bsd(state, offset, size, origin, "bsd", BSD_MAXPARTITIONS); in parse_freebsd()
418 sector_t offset, sector_t size, int origin) in parse_netbsd() argument
421 parse_bsd(state, offset, size, origin, "netbsd", BSD_MAXPARTITIONS); in parse_netbsd()
426 sector_t offset, sector_t size, int origin) in parse_openbsd() argument
429 parse_bsd(state, offset, size, origin, "openbsd", in parse_openbsd()
[all …]
/linux/arch/s390/kvm/
H A Dinterrupt.c240 word = READ_ONCE(gi->origin->u64.word[0]); in gisa_get_ipm_or_restore_iam()
247 } while (!try_cmpxchg(&gi->origin->u64.word[0], &word, _word)); in gisa_get_ipm_or_restore_iam()
282 if (gi->origin) in pending_irqs()
283 pending_mask |= gisa_get_ipm(gi->origin) << IRQ_PEND_IO_ISC_7; in pending_irqs()
1154 if (gi->origin && gisa_tac_ipm_gisc(gi->origin, isc)) { in __deliver_io()
1258 if (gi->origin && in kvm_s390_handle_wait()
1701 if (!gi->origin) in get_top_gisa_isc()
1704 active_mask = (isc_mask & gisa_get_ipm(gi->origin) << 24) << 32; in get_top_gisa_isc()
1707 if (gisa_tac_ipm_gisc(gi->origin, isc)) in get_top_gisa_isc()
1748 gisa_set_ipm_gisc(gi->origin, isc); in kvm_s390_get_io_int()
[all …]

12345678910>>...12