/linux-5.10/drivers/misc/vmw_vmci/ |
D | vmci_context.c | 30 * These, along with context lookup, are protected by the 35 spinlock_t lock; /* Spinlock for context list operations */ 44 static void ctx_signal_notify(struct vmci_ctx *context) in ctx_signal_notify() argument 46 *context->notify = true; in ctx_signal_notify() 49 static void ctx_clear_notify(struct vmci_ctx *context) in ctx_clear_notify() argument 51 *context->notify = false; in ctx_clear_notify() 58 static void ctx_clear_notify_call(struct vmci_ctx *context) in ctx_clear_notify_call() argument 60 if (context->pending_datagrams == 0 && in ctx_clear_notify_call() 61 vmci_handle_arr_get_size(context->pending_doorbell_array) == 0) in ctx_clear_notify_call() 62 ctx_clear_notify(context); in ctx_clear_notify_call() [all …]
|
D | vmci_route.c | 18 * devices. Will set the source context if it is invalid. 41 /* Must have a valid destination context. */ in vmci_route() 42 if (VMCI_INVALID_ID == dst->context) in vmci_route() 46 if (VMCI_HYPERVISOR_CONTEXT_ID == dst->context) { in vmci_route() 63 /* And we cannot send if the source is the host context. */ in vmci_route() 64 if (VMCI_HOST_CONTEXT_ID == src->context) in vmci_route() 69 * respect it (both context and resource are invalid). in vmci_route() 70 * However, if they passed only an invalid context, in vmci_route() 72 * should set the real context here before passing it in vmci_route() 75 if (VMCI_INVALID_ID == src->context && in vmci_route() [all …]
|
/linux-5.10/security/selinux/ss/ |
D | mls.h | 27 #include "context.h" 31 int mls_compute_context_len(struct policydb *p, struct context *context); 32 void mls_sid_to_context(struct policydb *p, struct context *context, 34 int mls_context_isvalid(struct policydb *p, struct context *c); 41 struct context *context, 45 int mls_from_string(struct policydb *p, char *str, struct context *context, 48 int mls_range_set(struct context *context, struct mls_range *range); 52 struct context *oldc, 53 struct context *newc); 56 struct context *scontext, [all …]
|
D | mls.c | 34 * security context string representation of `context'. 36 int mls_compute_context_len(struct policydb *p, struct context *context) in mls_compute_context_len() argument 48 int index_sens = context->range.level[l].sens; in mls_compute_context_len() 54 e = &context->range.level[l].cat; in mls_compute_context_len() 73 if (mls_level_eq(&context->range.level[0], in mls_compute_context_len() 74 &context->range.level[1])) in mls_compute_context_len() 85 * Write the security context string representation of 86 * the MLS fields of `context' into the string `*scontext'. 90 struct context *context, in mls_sid_to_context() argument 108 context->range.level[l].sens - 1)); in mls_sid_to_context() [all …]
|
/linux-5.10/drivers/gpu/drm/etnaviv/ |
D | etnaviv_mmu.c | 16 static void etnaviv_context_unmap(struct etnaviv_iommu_context *context, in etnaviv_context_unmap() argument 29 unmapped_page = context->global->ops->unmap(context, iova, in etnaviv_context_unmap() 39 static int etnaviv_context_map(struct etnaviv_iommu_context *context, in etnaviv_context_map() argument 55 ret = context->global->ops->map(context, iova, paddr, pgsize, in etnaviv_context_map() 67 etnaviv_context_unmap(context, orig_iova, orig_size - size); in etnaviv_context_map() 72 static int etnaviv_iommu_map(struct etnaviv_iommu_context *context, u32 iova, in etnaviv_iommu_map() argument 79 if (!context || !sgt) in etnaviv_iommu_map() 88 ret = etnaviv_context_map(context, da, pa, bytes, prot); in etnaviv_iommu_map() 98 etnaviv_context_unmap(context, iova, da - iova); in etnaviv_iommu_map() 102 static void etnaviv_iommu_unmap(struct etnaviv_iommu_context *context, u32 iova, in etnaviv_iommu_unmap() argument [all …]
|
D | etnaviv_iommu.c | 28 to_v1_context(struct etnaviv_iommu_context *context) in to_v1_context() argument 30 return container_of(context, struct etnaviv_iommuv1_context, base); in to_v1_context() 33 static void etnaviv_iommuv1_free(struct etnaviv_iommu_context *context) in etnaviv_iommuv1_free() argument 35 struct etnaviv_iommuv1_context *v1_context = to_v1_context(context); in etnaviv_iommuv1_free() 37 drm_mm_takedown(&context->mm); in etnaviv_iommuv1_free() 39 dma_free_wc(context->global->dev, PT_SIZE, v1_context->pgtable_cpu, in etnaviv_iommuv1_free() 42 context->global->v1.shared_context = NULL; in etnaviv_iommuv1_free() 47 static int etnaviv_iommuv1_map(struct etnaviv_iommu_context *context, in etnaviv_iommuv1_map() argument 51 struct etnaviv_iommuv1_context *v1_context = to_v1_context(context); in etnaviv_iommuv1_map() 62 static size_t etnaviv_iommuv1_unmap(struct etnaviv_iommu_context *context, in etnaviv_iommuv1_unmap() argument [all …]
|
D | etnaviv_iommu_v2.c | 42 to_v2_context(struct etnaviv_iommu_context *context) in to_v2_context() argument 44 return container_of(context, struct etnaviv_iommuv2_context, base); in to_v2_context() 47 static void etnaviv_iommuv2_free(struct etnaviv_iommu_context *context) in etnaviv_iommuv2_free() argument 49 struct etnaviv_iommuv2_context *v2_context = to_v2_context(context); in etnaviv_iommuv2_free() 52 drm_mm_takedown(&context->mm); in etnaviv_iommuv2_free() 56 dma_free_wc(context->global->dev, SZ_4K, in etnaviv_iommuv2_free() 61 dma_free_wc(context->global->dev, SZ_4K, v2_context->mtlb_cpu, in etnaviv_iommuv2_free() 64 clear_bit(v2_context->id, context->global->v2.pta_alloc); in etnaviv_iommuv2_free() 92 static int etnaviv_iommuv2_map(struct etnaviv_iommu_context *context, in etnaviv_iommuv2_map() argument 96 struct etnaviv_iommuv2_context *v2_context = to_v2_context(context); in etnaviv_iommuv2_map() [all …]
|
/linux-5.10/kernel/ |
D | auditsc.c | 41 * Subject and object context labeling support added by <danjones@us.ibm.com> 228 * it's going to remain 1-element for almost any setup) until we free context itself. 459 /* Determine if any context name data matches a rule's watch data */ 876 static inline void audit_proctitle_free(struct audit_context *context) in audit_proctitle_free() argument 878 kfree(context->proctitle.value); in audit_proctitle_free() 879 context->proctitle.value = NULL; in audit_proctitle_free() 880 context->proctitle.len = 0; in audit_proctitle_free() 883 static inline void audit_free_module(struct audit_context *context) in audit_free_module() argument 885 if (context->type == AUDIT_KERN_MODULE) { in audit_free_module() 886 kfree(context->module.name); in audit_free_module() [all …]
|
/linux-5.10/fs/xfs/ |
D | xfs_attr_list.c | 53 struct xfs_attr_list_context *context) in xfs_attr_shortform_list() argument 55 struct xfs_attrlist_cursor_kern *cursor = &context->cursor; in xfs_attr_shortform_list() 56 struct xfs_inode *dp = context->dp; in xfs_attr_shortform_list() 69 trace_xfs_attr_list_sf(context); in xfs_attr_shortform_list() 80 if (context->bufsize == 0 || in xfs_attr_shortform_list() 82 (dp->i_afp->if_bytes + sf->hdr.count * 16) < context->bufsize)) { in xfs_attr_shortform_list() 84 if (XFS_IS_CORRUPT(context->dp->i_mount, in xfs_attr_shortform_list() 88 context->put_listent(context, in xfs_attr_shortform_list() 97 if (context->seen_enough) in xfs_attr_shortform_list() 101 trace_xfs_attr_list_sf_all(context); in xfs_attr_shortform_list() [all …]
|
D | xfs_xattr.c | 96 struct xfs_attr_list_context *context, in __xfs_xattr_put_listent() argument 105 if (context->count < 0 || context->seen_enough) in __xfs_xattr_put_listent() 108 if (!context->buffer) in __xfs_xattr_put_listent() 111 arraytop = context->count + prefix_len + namelen + 1; in __xfs_xattr_put_listent() 112 if (arraytop > context->firstu) { in __xfs_xattr_put_listent() 113 context->count = -1; /* insufficient space */ in __xfs_xattr_put_listent() 114 context->seen_enough = 1; in __xfs_xattr_put_listent() 117 offset = context->buffer + context->count; in __xfs_xattr_put_listent() 125 context->count += prefix_len + namelen + 1; in __xfs_xattr_put_listent() 131 struct xfs_attr_list_context *context, in xfs_xattr_put_listent() argument [all …]
|
/linux-5.10/arch/s390/include/asm/ |
D | mmu_context.h | 23 spin_lock_init(&mm->context.lock); in init_new_context() 24 INIT_LIST_HEAD(&mm->context.pgtable_list); in init_new_context() 25 INIT_LIST_HEAD(&mm->context.gmap_list); in init_new_context() 26 cpumask_clear(&mm->context.cpu_attach_mask); in init_new_context() 27 atomic_set(&mm->context.flush_count, 0); in init_new_context() 28 atomic_set(&mm->context.is_protected, 0); in init_new_context() 29 mm->context.gmap_asce = 0; in init_new_context() 30 mm->context.flush_mm = 0; in init_new_context() 32 mm->context.alloc_pgste = page_table_allocate_pgste || in init_new_context() 34 (current->mm && current->mm->context.alloc_pgste); in init_new_context() [all …]
|
/linux-5.10/drivers/net/ethernet/mellanox/mlx4/ |
D | en_resources.c | 42 int user_prio, struct mlx4_qp_context *context) in mlx4_en_fill_qp_context() argument 47 memset(context, 0, sizeof(*context)); in mlx4_en_fill_qp_context() 48 context->flags = cpu_to_be32(7 << 16 | rss << MLX4_RSS_QPC_FLAG_OFFSET); in mlx4_en_fill_qp_context() 49 context->pd = cpu_to_be32(mdev->priv_pdn); in mlx4_en_fill_qp_context() 50 context->mtu_msgmax = 0xff; in mlx4_en_fill_qp_context() 52 context->rq_size_stride = ilog2(size) << 3 | (ilog2(stride) - 4); in mlx4_en_fill_qp_context() 54 context->sq_size_stride = ilog2(size) << 3 | (ilog2(stride) - 4); in mlx4_en_fill_qp_context() 56 context->params2 |= cpu_to_be32(MLX4_QP_BIT_FPP); in mlx4_en_fill_qp_context() 59 context->sq_size_stride = ilog2(TXBB_SIZE) - 4; in mlx4_en_fill_qp_context() 61 context->usr_page = cpu_to_be32(mlx4_to_hw_uar_index(mdev->dev, in mlx4_en_fill_qp_context() [all …]
|
/linux-5.10/arch/ia64/include/asm/ |
D | mmu_context.h | 11 * Routines to manage the allocation of task context numbers. Task context 13 * due to context switches. Context numbers are implemented using ia-64 38 unsigned int next; /* next context number to use */ 40 unsigned int max_ctx; /* max. context value supported by all CPUs */ 58 * When the context counter wraps around all TLBs need to be flushed because 59 * an old context number might have been reused. This is signalled by the 83 nv_mm_context_t context = mm->context; in get_mmu_context() local 85 if (likely(context)) in get_mmu_context() 90 context = mm->context; in get_mmu_context() 91 if (context == 0) { in get_mmu_context() [all …]
|
/linux-5.10/arch/powerpc/mm/nohash/ |
D | mmu_context.c | 15 * - The global context lock will not scale very well 18 * - Implement flush_tlb_mm() by making the context stale and picking 93 /* Steal a context from a task that has one at the moment. 98 * This isn't an LRU system, it just frees up each context in 103 * For context stealing, we use a slightly different approach for 124 if (mm->context.active) { in steal_context_smp() 132 /* Mark this mm has having no context anymore */ in steal_context_smp() 133 mm->context.id = MMU_NO_CONTEXT; in steal_context_smp() 177 /* Mark this mm as having no context anymore */ in steal_all_contexts() 178 mm->context.id = MMU_NO_CONTEXT; in steal_all_contexts() [all …]
|
/linux-5.10/include/misc/ |
D | cxl.h | 31 * Context lifetime overview: 33 * An AFU context may be inited and then started and stoppped multiple times 42 * Once released, a context can't be started again. 44 * One context is inited by the cxl driver for every pci_dev. This is to be 45 * used as a default kernel context. cxl_get_context() will get this 46 * context. This context will be released by PCI hot unplug, so doesn't need to 52 * Once a context has been inited, IRQs may be configured. Firstly these IRQs 65 * On pci_enabled_device(), the cxl driver will init a single cxl context for 66 * use by the driver. It doesn't start this context (as that will likely 69 * This gets the default context associated with this pci_dev. This context [all …]
|
/linux-5.10/arch/nios2/mm/ |
D | mmu_context.c | 2 * MMU context handling. 19 /* The pids position and mask in context */ 24 /* The versions position and mask in context */ 29 /* Return the version part of a context */ 32 /* Return the pid part of a context */ 35 /* Value of the first context (version 1, pid 0) */ 41 * Initialize MMU context management stuff. 51 * Set new context (pid), keep way 53 static void set_context(mm_context_t context) in set_context() argument 55 set_mmu_pid(CTX_PID(context)); in set_context() [all …]
|
/linux-5.10/tools/perf/scripts/perl/Perf-Trace-Util/ |
D | Context.c | 4 * contents of Context.xs. Do not edit this file, edit Context.xs instead. 12 #line 1 "Context.xs" 14 * Context.xs. XS interfaces for perf script. 28 #line 42 "Context.c" 39 Perl_croak(aTHX_ "Usage: %s(%s)", "Perf::Trace::Context::common_pc", "context"); in XS() 42 struct scripting_context * context = INT2PTR(struct scripting_context *,SvIV(ST(0))); in XS() local 46 RETVAL = common_pc(context); in XS() 62 Perl_croak(aTHX_ "Usage: %s(%s)", "Perf::Trace::Context::common_flags", "context"); in XS() 65 struct scripting_context * context = INT2PTR(struct scripting_context *,SvIV(ST(0))); in XS() local 69 RETVAL = common_flags(context); in XS() [all …]
|
/linux-5.10/drivers/media/usb/as102/ |
D | as10x_cmd_cfg.c | 16 * as10x_cmd_get_context - Send get context command to AS10x 18 * @tag: context tag 19 * @pvalue: pointer where to store context value read 34 sizeof(pcmd->body.context.req)); in as10x_cmd_get_context() 37 pcmd->body.context.req.proc_id = cpu_to_le16(CONTROL_PROC_CONTEXT); in as10x_cmd_get_context() 38 pcmd->body.context.req.tag = cpu_to_le16(tag); in as10x_cmd_get_context() 39 pcmd->body.context.req.type = cpu_to_le16(GET_CONTEXT_DATA); in as10x_cmd_get_context() 45 sizeof(pcmd->body.context.req) in as10x_cmd_get_context() 48 sizeof(prsp->body.context.rsp) in as10x_cmd_get_context() 57 /* parse response: context command do not follow the common response */ in as10x_cmd_get_context() [all …]
|
/linux-5.10/drivers/infiniband/hw/hns/ |
D | hns_roce_restrack.c | 13 struct hns_roce_v2_cq_context *context) in hns_roce_fill_cq() argument 16 roce_get_field(context->byte_4_pg_ceqn, in hns_roce_fill_cq() 22 roce_get_field(context->byte_4_pg_ceqn, in hns_roce_fill_cq() 28 roce_get_field(context->byte_8_cqn, in hns_roce_fill_cq() 34 roce_get_field(context->byte_16_hop_addr, in hns_roce_fill_cq() 41 roce_get_field(context->byte_28_cq_pi, in hns_roce_fill_cq() 48 roce_get_field(context->byte_32_cq_ci, in hns_roce_fill_cq() 55 roce_get_field(context->byte_56_cqe_period_maxcnt, in hns_roce_fill_cq() 62 roce_get_field(context->byte_56_cqe_period_maxcnt, in hns_roce_fill_cq() 68 roce_get_field(context->byte_52_cqe_cnt, in hns_roce_fill_cq() [all …]
|
/linux-5.10/drivers/gpu/drm/i915/selftests/ |
D | i915_syncmap.c | 163 static int check_one(struct i915_syncmap **sync, u64 context, u32 seqno) in check_one() argument 167 err = i915_syncmap_set(sync, context, seqno); in check_one() 172 pr_err("Inserting first context=%llx did not return leaf (height=%d, prefix=%llx\n", in check_one() 173 context, (*sync)->height, (*sync)->prefix); in check_one() 178 pr_err("Inserting first context=%llx created branches!\n", in check_one() 179 context); in check_one() 193 if (!i915_syncmap_is_later(sync, context, seqno)) { in check_one() 194 pr_err("Lookup of first context=%llx/seqno=%x failed!\n", in check_one() 195 context, seqno); in check_one() 217 u64 context = i915_prandom_u64_state(&prng); in igt_syncmap_one() local [all …]
|
/linux-5.10/arch/mips/include/uapi/asm/ |
D | ucontext.h | 6 * struct extcontext - extended context header structure 7 * @magic: magic value identifying the type of extended context 10 * Extended context structures provide context which does not fit within struct 12 * ucontext and struct sigframe, with each extended context structure beginning 13 * with a header defined by this struct. The type of context represented is 14 * indicated by the magic field. Userland may check each extended context 16 * unrecognised context to be skipped, allowing for future expansion. The end 17 * of the extended context data is indicated by the magic value 26 * struct msa_extcontext - MSA extended context structure 27 * @ext: the extended context header, with magic == MSA_EXTCONTEXT_MAGIC [all …]
|
/linux-5.10/fs/ocfs2/ |
D | move_extents.c | 52 struct ocfs2_move_extents_context *context, in __ocfs2_move_extent() argument 57 struct inode *inode = context->inode; in __ocfs2_move_extent() 62 u64 ino = ocfs2_metadata_cache_owner(context->et.et_ci); in __ocfs2_move_extent() 78 path = ocfs2_new_path_from_et(&context->et); in __ocfs2_move_extent() 111 context->et.et_root_bh, in __ocfs2_move_extent() 118 ret = ocfs2_split_extent(handle, &context->et, path, index, in __ocfs2_move_extent() 119 &replace_rec, context->meta_ac, in __ocfs2_move_extent() 120 &context->dealloc); in __ocfs2_move_extent() 126 ocfs2_journal_dirty(handle, context->et.et_root_bh); in __ocfs2_move_extent() 128 context->new_phys_cpos = new_p_cpos; in __ocfs2_move_extent() [all …]
|
/linux-5.10/drivers/pci/hotplug/ |
D | acpiphp_glue.c | 50 static void hotplug_event(u32 type, struct acpiphp_context *context); 54 * acpiphp_init_context - Create hotplug context and grab a reference to it. 55 * @adev: ACPI device object to create the context for. 61 struct acpiphp_context *context; in acpiphp_init_context() local 63 context = kzalloc(sizeof(*context), GFP_KERNEL); in acpiphp_init_context() 64 if (!context) in acpiphp_init_context() 67 context->refcount = 1; in acpiphp_init_context() 68 context->hp.notify = acpiphp_hotplug_notify; in acpiphp_init_context() 69 context->hp.fixup = acpiphp_post_dock_fixup; in acpiphp_init_context() 70 acpi_set_hp_context(adev, &context->hp); in acpiphp_init_context() [all …]
|
/linux-5.10/arch/powerpc/mm/book3s64/ |
D | mmu_context.c | 3 * MMU context allocation for 64-bit kernels. 38 WARN(result != id, "mmu: Failed to reserve context id %d (rc %d)\n", id, result); in hash__reserve_context_id() 96 mm->context.hash_context = kmalloc(sizeof(struct hash_mm_context), in hash__init_new_context() 98 if (!mm->context.hash_context) in hash__init_new_context() 107 * explicitly against context.id == 0. This ensures that we properly in hash__init_new_context() 108 * initialize context slice details for newly allocated mm's (which will in hash__init_new_context() 109 * have id == 0) and don't alter context slice inherited via fork (which in hash__init_new_context() 115 if (mm->context.id == 0) { in hash__init_new_context() 116 memset(mm->context.hash_context, 0, sizeof(struct hash_mm_context)); in hash__init_new_context() 120 …memcpy(mm->context.hash_context, current->mm->context.hash_context, sizeof(struct hash_mm_context)… in hash__init_new_context() [all …]
|
/linux-5.10/arch/sparc/mm/ |
D | tsb.c | 124 spin_lock_irqsave(&mm->context.lock, flags); in flush_tsb_user() 127 base = (unsigned long) mm->context.tsb_block[MM_TSB_BASE].tsb; in flush_tsb_user() 128 nentries = mm->context.tsb_block[MM_TSB_BASE].tsb_nentries; in flush_tsb_user() 140 else if (mm->context.tsb_block[MM_TSB_HUGE].tsb) { in flush_tsb_user() 141 base = (unsigned long) mm->context.tsb_block[MM_TSB_HUGE].tsb; in flush_tsb_user() 142 nentries = mm->context.tsb_block[MM_TSB_HUGE].tsb_nentries; in flush_tsb_user() 149 spin_unlock_irqrestore(&mm->context.lock, flags); in flush_tsb_user() 157 spin_lock_irqsave(&mm->context.lock, flags); in flush_tsb_user_page() 160 base = (unsigned long) mm->context.tsb_block[MM_TSB_BASE].tsb; in flush_tsb_user_page() 161 nentries = mm->context.tsb_block[MM_TSB_BASE].tsb_nentries; in flush_tsb_user_page() [all …]
|