/linux/arch/parisc/math-emu/ |
H A D | float.h | 48 #define Sall(object) (object) argument 49 #define Ssign(object) Bitfield_extract( 0, 1,object) argument 50 #define Ssignedsign(object) Bitfield_signed_extract( 0, 1,object) argument 51 #define Sexponent(object) Bitfield_extract( 1, 8,object) argument 52 #define Smantissa(object) Bitfield_mask( 9, 23,object) argument 53 #define Ssignaling(object) Bitfield_extract( 9, 1,object) argument 54 #define Ssignalingnan(object) Bitfield_extract( 1, 9,object) argument 55 #define Shigh2mantissa(object) Bitfield_extract( 9, 2,object) argument 56 #define Sexponentmantissa(object) Bitfield_mask( 1, 31,object) argument 57 #define Ssignexponent(object) Bitfield_extract( 0, 9,object) argument [all …]
|
/linux/drivers/gpu/drm/nouveau/nvkm/core/ |
H A D | object.c | 24 #include <core/object.h> 32 struct nvkm_object *object; in nvkm_object_search() local 39 object = rb_entry(node, typeof(*object), node); in nvkm_object_search() 40 if (handle < object->object) in nvkm_object_search() 43 if (handle > object->object) in nvkm_object_search() 53 object = &client->object; in nvkm_object_search() 57 if (unlikely(func && object->func != func)) in nvkm_object_search() 59 return object; in nvkm_object_search() 63 nvkm_object_remove(struct nvkm_object *object) in nvkm_object_remove() argument 67 spin_lock_irqsave(&object->client->obj_lock, flags); in nvkm_object_remove() [all …]
|
H A D | ioctl.c | 34 struct nvkm_object *object, void *data, u32 size) in nvkm_ioctl_nop() argument 42 nvkm_ioctl_sclass_(struct nvkm_object *object, int index, struct nvkm_oclass *oclass) in nvkm_ioctl_sclass_() argument 44 if ( object->func->uevent && in nvkm_ioctl_sclass_() 45 !object->func->uevent(object, NULL, 0, NULL) && index-- == 0) { in nvkm_ioctl_sclass_() 53 if (object->func->sclass) in nvkm_ioctl_sclass_() 54 return object->func->sclass(object, index, oclass); in nvkm_ioctl_sclass_() 61 struct nvkm_object *object, void *data, u32 size) in nvkm_ioctl_sclass() argument 69 nvif_ioctl(object, "sclass size %d\n", size); in nvkm_ioctl_sclass() 71 nvif_ioctl(object, "sclass vers %d count %d\n", in nvkm_ioctl_sclass() 76 while (nvkm_ioctl_sclass_(object, i, &oclass) >= 0) { in nvkm_ioctl_sclass() [all …]
|
H A D | oproxy.c | 27 nvkm_oproxy_mthd(struct nvkm_object *object, u32 mthd, void *data, u32 size) in nvkm_oproxy_mthd() argument 29 return nvkm_object_mthd(nvkm_oproxy(object)->object, mthd, data, size); in nvkm_oproxy_mthd() 33 nvkm_oproxy_ntfy(struct nvkm_object *object, u32 mthd, in nvkm_oproxy_ntfy() argument 36 return nvkm_object_ntfy(nvkm_oproxy(object)->object, mthd, pevent); in nvkm_oproxy_ntfy() 40 nvkm_oproxy_map(struct nvkm_object *object, void *argv, u32 argc, in nvkm_oproxy_map() argument 43 struct nvkm_oproxy *oproxy = nvkm_oproxy(object); in nvkm_oproxy_map() 44 return nvkm_object_map(oproxy->object, argv, argc, type, addr, size); in nvkm_oproxy_map() 48 nvkm_oproxy_unmap(struct nvkm_object *object) in nvkm_oproxy_unmap() argument 50 struct nvkm_oproxy *oproxy = nvkm_oproxy(object); in nvkm_oproxy_unmap() 52 if (unlikely(!oproxy->object)) in nvkm_oproxy_unmap() [all …]
|
/linux/mm/ |
H A D | kmemleak.c | 17 * del_state modifications and accesses to the object trees 20 * kmemleak_object) for the allocated memory blocks. The object trees are 23 * the object_list and the object tree root in the create_object() function 38 * Note that the kmemleak_object.use_count is incremented when an object is 46 * scan_mutex [-> object->lock] -> kmemleak_lock -> other_object->lock (SINGLE_DEPTH_NESTING) 48 * No kmemleak_lock and object->lock nesting is allowed outside scan_mutex 110 #define MSECS_MIN_AGE 5000 /* minimum object age for reporting */ 130 * object->lock. Insertions or deletions from object_list, gray_list or 137 unsigned int flags; /* object status flags */ 142 /* object usage count; object freed when use_count == 0 */ [all …]
|
/linux/drivers/acpi/acpica/ |
H A D | utdelete.c | 4 * Module Name: utdelete - object deletion and reference count utilities 18 static void acpi_ut_delete_internal_obj(union acpi_operand_object *object); 21 acpi_ut_update_ref_count(union acpi_operand_object *object, u32 action); 27 * PARAMETERS: object - Object to be deleted 31 * DESCRIPTION: Low level object deletion, after reference counts have been 36 static void acpi_ut_delete_internal_obj(union acpi_operand_object *object) in acpi_ut_delete_internal_obj() argument 45 ACPI_FUNCTION_TRACE_PTR(ut_delete_internal_obj, object); in acpi_ut_delete_internal_obj() 47 if (!object) { in acpi_ut_delete_internal_obj() 52 * Must delete or free any pointers within the object that are not in acpi_ut_delete_internal_obj() 55 switch (object->common.type) { in acpi_ut_delete_internal_obj() [all …]
|
H A D | nsobject.c | 21 * object - Object to be attached 22 * type - Type of object, or ACPI_TYPE_ANY if not 27 * DESCRIPTION: Record the given object as the value associated with the 28 * name whose acpi_handle is passed. If Object is NULL 38 union acpi_operand_object *object, acpi_object_type type) in acpi_ns_attach_object() argument 57 if (!object && (ACPI_TYPE_ANY != type)) { in acpi_ns_attach_object() 59 /* Null object */ in acpi_ns_attach_object() 62 "Null object, but type not ACPI_TYPE_ANY")); in acpi_ns_attach_object() 75 /* Check if this object is already attached */ in acpi_ns_attach_object() 77 if (node->object == object) { in acpi_ns_attach_object() [all …]
|
H A D | utobject.c | 4 * Module Name: utobject - ACPI object create/delete/size/cache routines 39 * type - ACPI Type of the new object 41 * RETURN: A new internal object, null on failure 43 * DESCRIPTION: Create and initialize a new internal object. 45 * NOTE: We always allocate the worst-case object descriptor because 48 * the most memory efficient, but the efficiency of the object 60 union acpi_operand_object *object; in acpi_ut_create_internal_object_dbg() local 66 /* Allocate the raw object descriptor */ in acpi_ut_create_internal_object_dbg() 68 object = in acpi_ut_create_internal_object_dbg() 71 if (!object) { in acpi_ut_create_internal_object_dbg() [all …]
|
H A D | utcache.c | 22 * object_size - Size of each cached object 24 * return_cache - Where the new cache object is returned 28 * DESCRIPTION: Create a cache object 44 /* Create the cache object */ in acpi_os_create_cache() 51 /* Populate the cache object and return it */ in acpi_os_create_cache() 66 * PARAMETERS: cache - Handle to cache object 94 /* Delete and unlink one cached state object */ in acpi_os_purge_cache() 111 * PARAMETERS: cache - Handle to cache object 116 * cache object. 133 /* Now we can delete the cache object */ in acpi_os_delete_cache() [all …]
|
H A D | dsmthdat.c | 25 union acpi_operand_object *object, 38 * PARAMETERS: walk_state - Current walk state object 93 * PARAMETERS: walk_state - Current walk state object 111 if (walk_state->local_variables[index].object) { in acpi_ds_method_data_delete_all() 115 object)); in acpi_ds_method_data_delete_all() 117 /* Detach object (if present) and remove a reference */ in acpi_ds_method_data_delete_all() 127 if (walk_state->arguments[index].object) { in acpi_ds_method_data_delete_all() 130 walk_state->arguments[index].object)); in acpi_ds_method_data_delete_all() 132 /* Detach object (if present) and remove a reference */ in acpi_ds_method_data_delete_all() 147 * walk_state - Current walk state object [all …]
|
H A D | nsrepair.c | 23 * predefined methods to an object type that is expected, as per the ACPI 28 * than the internal object conversion routines used for implicit object 46 * An incorrect standalone object is wrapped with required outer package 85 /* Object reference conversions */ 105 * expected_btypes - Object types expected 106 * package_index - Index of object within parent package (if 109 * return_object_ptr - Pointer to the object returned from the 110 * evaluation of a method or object 114 * DESCRIPTION: Attempt to repair/convert a return object of a type that was 153 "During return object analysis")); in acpi_ns_simple_repair() [all …]
|
H A D | utcopy.c | 4 * Module Name: utcopy - Internal to external object translation utilities 61 * PARAMETERS: internal_object - Source object to be copied 62 * external_object - Where to return the copied object 63 * data_space - Where object data is returned (such as 69 * DESCRIPTION: This function is called to copy a simple internal object to 70 * an external object. 73 * the object. 89 * Check for NULL object case (could be an uninitialized in acpi_ut_copy_isimple_to_esimple() 96 /* Always clear the external object */ in acpi_ut_copy_isimple_to_esimple() 101 * In general, the external object will be the same type as in acpi_ut_copy_isimple_to_esimple() [all …]
|
/linux/Documentation/core-api/ |
H A D | debug-objects.rst | 2 The object-lifetime debugging infrastructure 21 debugobjects is not changing the data structure of the real object so it 29 object type and add calls into the debug code at appropriate places. The 30 data structure to describe the object type needs at minimum the name of 31 the object type. Optional functions can and should be provided to fixup 53 Each of these functions takes the address of the real object and a 54 pointer to the object type specific debug description structure. 71 object is called. 73 When the real object is already tracked by debugobjects it is checked, 74 whether the object can be initialized. Initializing is not allowed for [all …]
|
/linux/mm/kasan/ |
H A D | common.c | 156 void __kasan_unpoison_new_object(struct kmem_cache *cache, void *object) in __kasan_unpoison_new_object() argument 158 kasan_unpoison(object, cache->object_size, false); in __kasan_unpoison_new_object() 161 void __kasan_poison_new_object(struct kmem_cache *cache, void *object) in __kasan_poison_new_object() argument 163 kasan_poison(object, round_up(cache->object_size, KASAN_GRANULE_SIZE), in __kasan_poison_new_object() 168 * This function assigns a tag to an object considering the following: 170 * object somewhere (e.g. in the object itself). We preassign a tag for 171 * each object in caches with constructors during slab creation and reuse 172 * the same tag each time a particular object is allocated. 178 const void *object, bool init) in assign_tag() argument 185 * set, assign a tag when the object is being allocated (init == false). in assign_tag() [all …]
|
/linux/drivers/gpu/drm/nouveau/nvif/ |
H A D | object.c | 25 #include <nvif/object.h> 31 nvif_object_ioctl(struct nvif_object *object, void *data, u32 size, void **hack) in nvif_object_ioctl() argument 33 struct nvif_client *client = object->client; in nvif_object_ioctl() 39 if (object != &client->object) in nvif_object_ioctl() 40 args->v0.object = nvif_handle(object); in nvif_object_ioctl() 42 args->v0.object = 0; in nvif_object_ioctl() 46 return client->driver->ioctl(client->object.priv, data, size, hack); in nvif_object_ioctl() 57 nvif_object_sclass_get(struct nvif_object *object, struct nvif_sclass **psclass) in nvif_object_sclass_get() argument 75 ret = nvif_object_ioctl(object, args, size, NULL); in nvif_object_sclass_get() 101 nvif_object_mthd(struct nvif_object *object, u32 mthd, void *data, u32 size) in nvif_object_mthd() argument [all …]
|
H A D | outp.c | 42 ret = nvif_object_mthd(&outp->object, NVIF_OUTP_V0_DP_MST_VCPI, &args, sizeof(args)); in nvif_outp_dp_mst_vcpi() 43 NVIF_ERRON(ret, &outp->object, in nvif_outp_dp_mst_vcpi() 57 ret = nvif_object_mthd(&outp->object, NVIF_OUTP_V0_DP_MST_ID_PUT, &args, sizeof(args)); in nvif_outp_dp_mst_id_put() 58 NVIF_ERRON(ret, &outp->object, "[DP_MST_ID_PUT id:%08x]", args.id); in nvif_outp_dp_mst_id_put() 69 ret = nvif_object_mthd(&outp->object, NVIF_OUTP_V0_DP_MST_ID_GET, &args, sizeof(args)); in nvif_outp_dp_mst_id_get() 70 NVIF_ERRON(ret, &outp->object, "[DP_MST_ID_GET] id:%08x", args.id); in nvif_outp_dp_mst_id_get() 89 ret = nvif_object_mthd(&outp->object, NVIF_OUTP_V0_DP_SST, &args, sizeof(args)); in nvif_outp_dp_sst() 90 NVIF_ERRON(ret, &outp->object, in nvif_outp_dp_sst() 107 ret = nvif_object_mthd(&outp->object, NVIF_OUTP_V0_DP_DRIVE, &args, sizeof(args)); in nvif_outp_dp_drive() 108 NVIF_ERRON(ret, &outp->object, "[DP_DRIVE lanes:%d]", args.lanes); in nvif_outp_dp_drive() [all …]
|
/linux/tools/lib/perf/include/internal/ |
H A D | rc_check.h | 40 * Interpose the indirection. Result will hold the indirection and object is the 43 #define ADD_RC_CHK(result, object) (result = object, object) argument 46 #define RC_CHK_ACCESS(object) object argument 48 /* Frees the object and the indirection layer. */ 49 #define RC_CHK_FREE(object) free(object) argument 52 #define RC_CHK_GET(result, object) ADD_RC_CHK(result, object) argument 55 #define RC_CHK_PUT(object) {} argument 74 * Interpose the indirection. Result will hold the indirection and object is the 77 #define ADD_RC_CHK(result, object) \ argument 79 object ? (result = malloc(sizeof(*result)), \ [all …]
|
/linux/sound/pci/asihpi/ |
H A D | hpimsginit.c | 17 /* The actual message size for each object type */ 19 /* The actual response size for each object type */ 27 static void hpi_init_message(struct hpi_message *phm, u16 object, in hpi_init_message() argument 32 if ((object > 0) && (object <= HPI_OBJ_MAXINDEX)) { in hpi_init_message() 33 object = array_index_nospec(object, HPI_OBJ_MAXINDEX + 1); in hpi_init_message() 34 size = msg_size[object]; in hpi_init_message() 46 phm->object = object; in hpi_init_message() 56 void hpi_init_response(struct hpi_response *phr, u16 object, u16 function, in hpi_init_response() argument 61 if ((object > 0) && (object <= HPI_OBJ_MAXINDEX)) { in hpi_init_response() 62 object = array_index_nospec(object, HPI_OBJ_MAXINDEX + 1); in hpi_init_response() [all …]
|
/linux/drivers/gpu/drm/vmwgfx/ |
H A D | ttm_object.h | 32 * Base- and reference object implementation for the various 50 * One entry per ttm object type. 73 * @hash: hash entry for the per-device object hash. 74 * @type: derived type this object is base class for. 75 * @shareable: Other ttm_object_files can access this object. 78 * NULL if the object was not created by a user request. 79 * (kernel object). 81 * @refcount: Number of references to this object, not 82 * including the hash entry. A reference to a base object can 83 * only be held by a ref object. [all …]
|
/linux/drivers/gpu/drm/nouveau/include/nvkm/subdev/ |
H A D | gsp.h | 158 } object; member 166 struct nvkm_gsp_object object; member 292 nvkm_gsp_rm_ctrl_get(struct nvkm_gsp_object *object, u32 cmd, u32 argc) in nvkm_gsp_rm_ctrl_get() argument 294 return object->client->gsp->rm->rm_ctrl_get(object, cmd, argc); in nvkm_gsp_rm_ctrl_get() 298 nvkm_gsp_rm_ctrl_push(struct nvkm_gsp_object *object, void *argv, u32 repc) in nvkm_gsp_rm_ctrl_push() argument 300 return object->client->gsp->rm->rm_ctrl_push(object, argv, repc); in nvkm_gsp_rm_ctrl_push() 304 nvkm_gsp_rm_ctrl_rd(struct nvkm_gsp_object *object, u32 cmd, u32 repc) in nvkm_gsp_rm_ctrl_rd() argument 306 void *argv = nvkm_gsp_rm_ctrl_get(object, cmd, repc); in nvkm_gsp_rm_ctrl_rd() 312 ret = nvkm_gsp_rm_ctrl_push(object, &argv, repc); in nvkm_gsp_rm_ctrl_rd() 319 nvkm_gsp_rm_ctrl_wr(struct nvkm_gsp_object *object, void *argv) in nvkm_gsp_rm_ctrl_wr() argument [all …]
|
/linux/security/landlock/ |
H A D | object.c | 3 * Landlock LSM - Object management 18 #include "object.h" 39 * The caller must own the object (i.e. thanks to object->usage) to safely put 42 void landlock_put_object(struct landlock_object *const object) in landlock_put_object() argument 45 * The call to @object->underops->release(object) might sleep, e.g. in landlock_put_object() 49 if (!object) in landlock_put_object() 53 * If the @object's refcount cannot drop to zero, we can just decrement in landlock_put_object() 55 * happen under @object->lock for synchronization with things like in landlock_put_object() 58 if (refcount_dec_and_lock(&object->usage, &object->lock)) { in landlock_put_object() 59 __acquire(&object->lock); in landlock_put_object() [all …]
|
H A D | object.h | 3 * Landlock LSM - Object management 19 * struct landlock_object_underops - Operations on an underlying object 23 * @release: Releases the underlying object (e.g. iput() for an inode). 25 void (*release)(struct landlock_object *const object) 26 __releases(object->lock); 30 * struct landlock_object - Security blob tied to a kernel object 33 * rights (pertaining to different domains) to a kernel object (e.g an inode) 41 * @usage: This counter is used to tie an object to the rules matching 45 * adding a new rule to an object with a usage counter of zero, we must 46 * wait until the pointer to this object is set to NULL (or recycled). [all …]
|
/linux/include/linux/ |
H A D | kfence.h | 38 * object range. 44 * an object requires specific handling. 91 * which include stack traces to the user of the object, the original allocation 97 * Allocate a KFENCE object. Allocators must not call this function directly, 103 * kfence_alloc() - allocate a KFENCE object with a low probability 104 * @s: struct kmem_cache with object requirements 105 * @size: exact size of the object to allocate (can be less than @s->size 111 * * non-NULL - pointer to a KFENCE object. 133 * kfence_ksize() - get actual amount of memory allocated for a KFENCE object 134 * @addr: pointer to a heap object [all …]
|
/linux/drivers/gpu/drm/nouveau/nvkm/engine/disp/ |
H A D | chan.c | 30 nvkm_disp_chan_ntfy(struct nvkm_object *object, u32 type, struct nvkm_event **pevent) in nvkm_disp_chan_ntfy() argument 32 struct nvkm_disp_chan *chan = nvkm_disp_chan(object); in nvkm_disp_chan_ntfy() 47 nvkm_disp_chan_map(struct nvkm_object *object, void *argv, u32 argc, in nvkm_disp_chan_map() argument 50 struct nvkm_disp_chan *chan = nvkm_disp_chan(object); in nvkm_disp_chan_map() 68 struct nvkm_disp_chan_object *object = container_of(base, typeof(*object), oproxy); in nvkm_disp_chan_child_del_() local 70 nvkm_ramht_remove(object->disp->ramht, object->hash); in nvkm_disp_chan_child_del_() 86 struct nvkm_disp_chan_object *object; in nvkm_disp_chan_child_new() local 89 if (!(object = kzalloc(sizeof(*object), GFP_KERNEL))) in nvkm_disp_chan_child_new() 91 nvkm_oproxy_ctor(&nvkm_disp_chan_child_func_, oclass, &object->oproxy); in nvkm_disp_chan_child_new() 92 object->disp = disp; in nvkm_disp_chan_child_new() [all …]
|
/linux/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_object.c | 98 * A gem object is embedded both in a struct ttm_buffer_object :/ and in i915_gem_object_init() 133 * __i915_gem_object_fini - Clean up a GEM object initialization 134 * @obj: The gem object to cleanup 136 * This function cleans up gem object fields that are set up by 139 * clean up the gem object in separate steps. 149 * i915_gem_object_set_cache_coherency - Mark up the object's coherency levels 225 * kernel, as per the given object cache_level. This is troublesome in i915_gem_object_can_bypass_llc() 227 * skipped if the kernel thinks the object is coherent with the GPU. As in i915_gem_object_can_bypass_llc() 334 * __i915_gem_object_pages_fini - Clean up pages use of a gem object 335 * @obj: The gem object to clean up [all …]
|