| /linux/arch/parisc/math-emu/ |
| H A D | float.h | 48 #define Sall(object) (object) argument 49 #define Ssign(object) Bitfield_extract( 0, 1,object) argument 50 #define Ssignedsign(object) Bitfield_signed_extract( 0, 1,object) argument 51 #define Sexponent(object) Bitfield_extract( 1, 8,object) argument 52 #define Smantissa(object) Bitfield_mask( 9, 23,object) argument 53 #define Ssignaling(object) Bitfield_extract( 9, 1,object) argument 54 #define Ssignalingnan(object) Bitfield_extract( 1, 9,object) argument 55 #define Shigh2mantissa(object) Bitfield_extract( 9, 2,object) argument 56 #define Sexponentmantissa(object) Bitfield_mask( 1, 31,object) argument 57 #define Ssignexponent(object) Bitfield_extract( 0, 9,object) argument [all …]
|
| /linux/drivers/gpu/drm/nouveau/nvkm/core/ |
| H A D | object.c | 24 #include <core/object.h> 32 struct nvkm_object *object; in nvkm_object_search() local 39 object = rb_entry(node, typeof(*object), node); in nvkm_object_search() 40 if (handle < object->object) in nvkm_object_search() 43 if (handle > object->object) in nvkm_object_search() 53 object = &client->object; in nvkm_object_search() 57 if (unlikely(func && object->func != func)) in nvkm_object_search() 59 return object; in nvkm_object_search() 63 nvkm_object_remove(struct nvkm_object *object) in nvkm_object_remove() argument 67 spin_lock_irqsave(&object->client->obj_lock, flags); in nvkm_object_remove() [all …]
|
| H A D | ioctl.c | 34 struct nvkm_object *object, void *data, u32 size) in nvkm_ioctl_nop() argument 42 nvkm_ioctl_sclass_(struct nvkm_object *object, int index, struct nvkm_oclass *oclass) in nvkm_ioctl_sclass_() argument 44 if ( object->func->uevent && in nvkm_ioctl_sclass_() 45 !object->func->uevent(object, NULL, 0, NULL) && index-- == 0) { in nvkm_ioctl_sclass_() 53 if (object->func->sclass) in nvkm_ioctl_sclass_() 54 return object->func->sclass(object, index, oclass); in nvkm_ioctl_sclass_() 61 struct nvkm_object *object, void *data, u32 size) in nvkm_ioctl_sclass() argument 69 nvif_ioctl(object, "sclass size %d\n", size); in nvkm_ioctl_sclass() 71 nvif_ioctl(object, "sclass vers %d count %d\n", in nvkm_ioctl_sclass() 76 while (nvkm_ioctl_sclass_(object, i, &oclass) >= 0) { in nvkm_ioctl_sclass() [all …]
|
| H A D | oproxy.c | 27 nvkm_oproxy_mthd(struct nvkm_object *object, u32 mthd, void *data, u32 size) in nvkm_oproxy_mthd() argument 29 return nvkm_object_mthd(nvkm_oproxy(object)->object, mthd, data, size); in nvkm_oproxy_mthd() 33 nvkm_oproxy_ntfy(struct nvkm_object *object, u32 mthd, in nvkm_oproxy_ntfy() argument 36 return nvkm_object_ntfy(nvkm_oproxy(object)->object, mthd, pevent); in nvkm_oproxy_ntfy() 40 nvkm_oproxy_map(struct nvkm_object *object, void *argv, u32 argc, in nvkm_oproxy_map() argument 43 struct nvkm_oproxy *oproxy = nvkm_oproxy(object); in nvkm_oproxy_map() 44 return nvkm_object_map(oproxy->object, argv, argc, type, addr, size); in nvkm_oproxy_map() 48 nvkm_oproxy_unmap(struct nvkm_object *object) in nvkm_oproxy_unmap() argument 50 struct nvkm_oproxy *oproxy = nvkm_oproxy(object); in nvkm_oproxy_unmap() 52 if (unlikely(!oproxy->object)) in nvkm_oproxy_unmap() [all …]
|
| /linux/fs/cachefiles/ |
| H A D | interface.c | 20 * Allocate a cache object record. 27 struct cachefiles_object *object; in cachefiles_alloc_object() local 31 object = kmem_cache_zalloc(cachefiles_object_jar, GFP_KERNEL); in cachefiles_alloc_object() 32 if (!object) in cachefiles_alloc_object() 35 if (cachefiles_ondemand_init_obj_info(object, volume)) { in cachefiles_alloc_object() 36 kmem_cache_free(cachefiles_object_jar, object); in cachefiles_alloc_object() 40 refcount_set(&object->ref, 1); in cachefiles_alloc_object() 42 spin_lock_init(&object->lock); in cachefiles_alloc_object() 43 INIT_LIST_HEAD(&object->cache_link); in cachefiles_alloc_object() 44 object->volume = volume; in cachefiles_alloc_object() [all …]
|
| H A D | ondemand.c | 20 struct cachefiles_object *object = file->private_data; in cachefiles_ondemand_fd_release() local 27 if (!object) in cachefiles_ondemand_fd_release() 30 info = object->ondemand; in cachefiles_ondemand_fd_release() 31 cache = object->volume->cache; in cachefiles_ondemand_fd_release() 38 cachefiles_ondemand_set_object_close(object); in cachefiles_ondemand_fd_release() 52 trace_cachefiles_ondemand_fd_release(object, object_id); in cachefiles_ondemand_fd_release() 53 cachefiles_put_object(object, cachefiles_obj_put_ondemand_fd); in cachefiles_ondemand_fd_release() 61 struct cachefiles_object *object = kiocb->ki_filp->private_data; in cachefiles_ondemand_fd_write_iter() local 62 struct cachefiles_cache *cache = object->volume->cache; in cachefiles_ondemand_fd_write_iter() 69 spin_lock(&object->lock); in cachefiles_ondemand_fd_write_iter() [all …]
|
| H A D | namei.c | 17 static bool __cachefiles_mark_inode_in_use(struct cachefiles_object *object, in __cachefiles_mark_inode_in_use() argument 24 trace_cachefiles_mark_active(object, inode); in __cachefiles_mark_inode_in_use() 27 trace_cachefiles_mark_failed(object, inode); in __cachefiles_mark_inode_in_use() 33 static bool cachefiles_mark_inode_in_use(struct cachefiles_object *object, in cachefiles_mark_inode_in_use() argument 39 can_use = __cachefiles_mark_inode_in_use(object, inode); in cachefiles_mark_inode_in_use() 47 static void __cachefiles_unmark_inode_in_use(struct cachefiles_object *object, in __cachefiles_unmark_inode_in_use() argument 51 trace_cachefiles_mark_inactive(object, inode); in __cachefiles_unmark_inode_in_use() 54 static void cachefiles_do_unmark_inode_in_use(struct cachefiles_object *object, in cachefiles_do_unmark_inode_in_use() argument 58 __cachefiles_unmark_inode_in_use(object, inode); in cachefiles_do_unmark_inode_in_use() 66 void cachefiles_unmark_inode_in_use(struct cachefiles_object *object, in cachefiles_unmark_inode_in_use() argument [all …]
|
| /linux/mm/ |
| H A D | kmemleak.c | 17 * del_state modifications and accesses to the object trees 20 * kmemleak_object) for the allocated memory blocks. The object trees are 23 * the object_list and the object tree root in the create_object() function 38 * Note that the kmemleak_object.use_count is incremented when an object is 46 * scan_mutex [-> object->lock] -> kmemleak_lock -> other_object->lock (SINGLE_DEPTH_NESTING) 48 * No kmemleak_lock and object->lock nesting is allowed outside scan_mutex 110 #define MSECS_MIN_AGE 5000 /* minimum object age for reporting */ 130 * object->lock. Insertions or deletions from object_list, gray_list or 137 unsigned int flags; /* object status flags */ 142 /* object usage count; object freed when use_count == 0 */ [all …]
|
| /linux/drivers/acpi/acpica/ |
| H A D | utdelete.c | 4 * Module Name: utdelete - object deletion and reference count utilities 18 static void acpi_ut_delete_internal_obj(union acpi_operand_object *object); 21 acpi_ut_update_ref_count(union acpi_operand_object *object, u32 action); 27 * PARAMETERS: object - Object to be deleted 31 * DESCRIPTION: Low level object deletion, after reference counts have been 36 static void acpi_ut_delete_internal_obj(union acpi_operand_object *object) in acpi_ut_delete_internal_obj() argument 45 ACPI_FUNCTION_TRACE_PTR(ut_delete_internal_obj, object); in acpi_ut_delete_internal_obj() 47 if (!object) { in acpi_ut_delete_internal_obj() 52 * Must delete or free any pointers within the object that are not in acpi_ut_delete_internal_obj() 55 switch (object->common.type) { in acpi_ut_delete_internal_obj() [all …]
|
| H A D | nsobject.c | 21 * object - Object to be attached 22 * type - Type of object, or ACPI_TYPE_ANY if not 27 * DESCRIPTION: Record the given object as the value associated with the 28 * name whose acpi_handle is passed. If Object is NULL 38 union acpi_operand_object *object, acpi_object_type type) in acpi_ns_attach_object() argument 57 if (!object && (ACPI_TYPE_ANY != type)) { in acpi_ns_attach_object() 59 /* Null object */ in acpi_ns_attach_object() 62 "Null object, but type not ACPI_TYPE_ANY")); in acpi_ns_attach_object() 75 /* Check if this object is already attached */ in acpi_ns_attach_object() 77 if (node->object == object) { in acpi_ns_attach_object() [all …]
|
| H A D | utobject.c | 4 * Module Name: utobject - ACPI object create/delete/size/cache routines 39 * type - ACPI Type of the new object 41 * RETURN: A new internal object, null on failure 43 * DESCRIPTION: Create and initialize a new internal object. 45 * NOTE: We always allocate the worst-case object descriptor because 48 * the most memory efficient, but the efficiency of the object 60 union acpi_operand_object *object; in acpi_ut_create_internal_object_dbg() local 66 /* Allocate the raw object descriptor */ in acpi_ut_create_internal_object_dbg() 68 object = in acpi_ut_create_internal_object_dbg() 71 if (!object) { in acpi_ut_create_internal_object_dbg() [all …]
|
| H A D | dsmthdat.c | 25 union acpi_operand_object *object, 38 * PARAMETERS: walk_state - Current walk state object 93 * PARAMETERS: walk_state - Current walk state object 111 if (walk_state->local_variables[index].object) { in acpi_ds_method_data_delete_all() 115 object)); in acpi_ds_method_data_delete_all() 117 /* Detach object (if present) and remove a reference */ in acpi_ds_method_data_delete_all() 127 if (walk_state->arguments[index].object) { in acpi_ds_method_data_delete_all() 130 walk_state->arguments[index].object)); in acpi_ds_method_data_delete_all() 132 /* Detach object (if present) and remove a reference */ in acpi_ds_method_data_delete_all() 147 * walk_state - Current walk state object [all …]
|
| /linux/Documentation/core-api/ |
| H A D | debug-objects.rst | 2 The object-lifetime debugging infrastructure 21 debugobjects is not changing the data structure of the real object so it 29 object type and add calls into the debug code at appropriate places. The 30 data structure to describe the object type needs at minimum the name of 31 the object type. Optional functions can and should be provided to fixup 53 Each of these functions takes the address of the real object and a 54 pointer to the object type specific debug description structure. 71 object is called. 73 When the real object is already tracked by debugobjects it is checked, 74 whether the object can be initialized. Initializing is not allowed for [all …]
|
| /linux/drivers/tee/qcomtee/ |
| H A D | core.c | 17 /* QTEE root object. */ 36 * set is reserved for the primordial object. 55 struct qcomtee_object *object; in qcomtee_qtee_object_alloc() local 57 object = kzalloc_obj(*object); in qcomtee_qtee_object_alloc() 58 if (!object) in qcomtee_qtee_object_alloc() 62 object->name = kasprintf(GFP_KERNEL, "qcomtee-%u", object_id); in qcomtee_qtee_object_alloc() 63 QCOMTEE_OBJECT_SET(object, QCOMTEE_OBJECT_TYPE_TEE, object_id); in qcomtee_qtee_object_alloc() 64 kref_init(&object->refcount); in qcomtee_qtee_object_alloc() 65 /* A QTEE object requires a context for async operations. */ in qcomtee_qtee_object_alloc() 66 object->info.qcomtee_async_ctx = qcomtee->ctx; in qcomtee_qtee_object_alloc() [all …]
|
| H A D | qcomtee_object.h | 19 * qcomtee_object provides object refcounting, ID allocation for objects hosted 22 * To invoke an object in QTEE, the user calls qcomtee_object_do_invoke() 26 * After boot, QTEE provides a static object %ROOT_QCOMTEE_OBJECT (type of 27 * %QCOMTEE_OBJECT_TYPE_ROOT). The root object is invoked to pass the user's 39 * QTEE (this is called a callback object). It issues qcomtee_object_user_init() 40 * to set the dispatch() operation for the callback object and set its type 43 * core.c holds an object table for callback objects. An object ID is assigned 44 * to each callback object, which is an index to the object table. QTEE uses 47 * If QTEE invokes a callback object in the kernel, the dispatch() operation is 53 * enum qcomtee_object_type - Object types. [all …]
|
| H A D | mem_obj.c | 17 * A memory object can be a standard dma_buf or a contiguous memory range, 18 * e.g., tee_shm. A memory object should support one operation: map. When 19 * invoked by QTEE, a mapping object is generated. A mapping object supports 22 * (1) To map a memory object, QTEE invokes the primordial object with 25 * (2) To unmap a memory object, QTEE releases the mapping object which 28 * The map operation is implemented in the primordial object as a privileged 31 * kernel memory by passing a user object as a memory object and returning a 36 struct qcomtee_object object; member 44 container_of((o), struct qcomtee_mem_object, object) 48 /* Is it a memory object using tee_shm? */ [all …]
|
| H A D | qcomtee.h | 15 /* Flags relating to object reference. */ 52 * qcomtee_object_do_invoke_internal() - Submit an invocation for an object. 54 * @object: object being invoked. 55 * @op: requested operation on the object. 60 * object, including @object. On return, the caller loses ownership of all 66 struct qcomtee_object *object, u32 op, 95 struct qcomtee_object *object, 97 int qcomtee_context_find_qtee_object(struct qcomtee_object **object, 110 /* (1) User Object API. */ 112 int is_qcomtee_user_object(struct qcomtee_object *object); [all …]
|
| /linux/drivers/gpu/drm/nouveau/nvif/ |
| H A D | object.c | 25 #include <nvif/object.h> 31 nvif_object_ioctl(struct nvif_object *object, void *data, u32 size, void **hack) in nvif_object_ioctl() argument 33 struct nvif_client *client = object->client; in nvif_object_ioctl() 39 if (object != &client->object) in nvif_object_ioctl() 40 args->v0.object = nvif_handle(object); in nvif_object_ioctl() 42 args->v0.object = 0; in nvif_object_ioctl() 46 return client->driver->ioctl(client->object.priv, data, size, hack); in nvif_object_ioctl() 57 nvif_object_sclass_get(struct nvif_object *object, struct nvif_sclass **psclass) in nvif_object_sclass_get() argument 75 ret = nvif_object_ioctl(object, args, size, NULL); in nvif_object_sclass_get() 101 nvif_object_mthd(struct nvif_object *object, u32 mthd, void *data, u32 size) in nvif_object_mthd() argument [all …]
|
| /linux/mm/kasan/ |
| H A D | common.c | 166 void __kasan_unpoison_new_object(struct kmem_cache *cache, void *object) in __kasan_unpoison_new_object() argument 168 kasan_unpoison(object, cache->object_size, false); in __kasan_unpoison_new_object() 171 void __kasan_poison_new_object(struct kmem_cache *cache, void *object) in __kasan_poison_new_object() argument 173 kasan_poison(object, round_up(cache->object_size, KASAN_GRANULE_SIZE), in __kasan_poison_new_object() 178 * This function assigns a tag to an object considering the following: 180 * object somewhere (e.g. in the object itself). We preassign a tag for 181 * each object in caches with constructors during slab creation and reuse 182 * the same tag each time a particular object is allocated. 188 const void *object, bool init) in assign_tag() argument 195 * set, assign a tag when the object is being allocated (init == false). in assign_tag() [all …]
|
| /linux/scripts/kconfig/ |
| H A D | gconf.ui | 5 <object class="GtkWindow" id="window1"> 22 <object class="GtkBox" id="vbox1"> 29 <object class="GtkMenuBar" id="menubar1"> 33 <object class="GtkMenuItem" id="file1"> 39 <object class="GtkMenu" id="file1_menu"> 42 <object class="GtkMenuItem" id="load1"> 48 </object> 52 <object class="GtkMenuItem" id="save1"> 58 </object> 62 <object class="GtkMenuItem" id="save_as1"> [all …]
|
| /linux/include/media/ |
| H A D | media-request.h | 131 * object in it. A reference to the request must be held during the update. This 252 * function will WARN and the object completion will be delayed until 276 * struct media_request_object_ops - Media request object operations 277 * @prepare: Validate and prepare the request object, optional. 278 * @unprepare: Unprepare the request object, optional. 279 * @queue: Queue the request object, optional. 280 * @unbind: Unbind the request object, optional. 281 * @release: Release the request object, required. 284 int (*prepare)(struct media_request_object *object); 285 void (*unprepare)(struct media_request_object *object); [all...] |
| /linux/sound/pci/asihpi/ |
| H A D | hpimsginit.c | 17 /* The actual message size for each object type */ 19 /* The actual response size for each object type */ 27 static void hpi_init_message(struct hpi_message *phm, u16 object, in hpi_init_message() argument 32 if ((object > 0) && (object <= HPI_OBJ_MAXINDEX)) { in hpi_init_message() 33 object = array_index_nospec(object, HPI_OBJ_MAXINDEX + 1); in hpi_init_message() 34 size = msg_size[object]; in hpi_init_message() 46 phm->object = object; in hpi_init_message() 56 void hpi_init_response(struct hpi_response *phr, u16 object, u16 function, in hpi_init_response() argument 61 if ((object > 0) && (object <= HPI_OBJ_MAXINDEX)) { in hpi_init_response() 62 object = array_index_nospec(object, HPI_OBJ_MAXINDEX + 1); in hpi_init_response() [all …]
|
| /linux/tools/lib/perf/include/internal/ |
| H A D | rc_check.h | 40 * Interpose the indirection. Result will hold the indirection and object is the 43 #define ADD_RC_CHK(result, object) (result = object, object) argument 46 #define RC_CHK_ACCESS(object) object argument 48 /* Frees the object and the indirection layer. */ 49 #define RC_CHK_FREE(object) free(object) argument 52 #define RC_CHK_GET(result, object) ADD_RC_CHK(result, object) argument 55 #define RC_CHK_PUT(object) {} argument 74 * Interpose the indirection. Result will hold the indirection and object is the 77 #define ADD_RC_CHK(result, object) \ argument 79 object ? (result = malloc(sizeof(*result)), \ [all …]
|
| /linux/drivers/gpu/drm/vmwgfx/ |
| H A D | ttm_object.h | 32 * Base- and reference object implementation for the various 50 * One entry per ttm object type. 73 * @hash: hash entry for the per-device object hash. 74 * @type: derived type this object is base class for. 75 * @shareable: Other ttm_object_files can access this object. 78 * NULL if the object was not created by a user request. 79 * (kernel object). 81 * @refcount: Number of references to this object, not 82 * including the hash entry. A reference to a base object can 83 * only be held by a ref object. [all …]
|
| /linux/include/linux/ |
| H A D | kasan.h | 137 void __kasan_unpoison_new_object(struct kmem_cache *cache, void *object); 139 * kasan_unpoison_new_object - Temporarily unpoison a new slab object. 140 * @cache: Cache the object belong to. 141 * @object: Pointer to the object. 144 * temporarily unpoisons an object from a newly allocated slab without doing 145 * anything else. The object must later be repoisoned by 149 void *object) in kasan_unpoison_new_object() argument 152 __kasan_unpoison_new_object(cache, object); in kasan_unpoison_new_object() 155 void __kasan_poison_new_object(struct kmem_cache *cache, void *object); 157 * kasan_poison_new_object - Repoison a new slab object. [all …]
|