Home
last modified time | relevance | path

Searched refs:context (Results 1 – 25 of 2576) sorted by relevance

12345678910>>...104

/linux/tools/tracing/rtla/src/
H A Dosnoise.c28 char *osnoise_get_cpus(struct osnoise_context *context) in osnoise_get_cpus() argument
30 if (context->curr_cpus) in osnoise_get_cpus()
31 return context->curr_cpus; in osnoise_get_cpus()
33 if (context->orig_cpus) in osnoise_get_cpus()
34 return context->orig_cpus; in osnoise_get_cpus()
36 context->orig_cpus = tracefs_instance_file_read(NULL, "osnoise/cpus", NULL); in osnoise_get_cpus()
42 return context->orig_cpus; in osnoise_get_cpus()
52 int osnoise_set_cpus(struct osnoise_context *context, char *cpus) in osnoise_set_cpus() argument
54 char *orig_cpus = osnoise_get_cpus(context); in osnoise_set_cpus()
61 context->curr_cpus = strdup(cpus); in osnoise_set_cpus()
[all …]
/linux/drivers/accel/qaic/
H A Dsahara.c226 static bool is_streaming(struct sahara_context *context) in is_streaming() argument
228 return !context->non_streaming; in is_streaming()
231 static int sahara_find_image(struct sahara_context *context, u32 image_id) in sahara_find_image() argument
235 if (image_id == context->active_image_id) in sahara_find_image()
238 if (context->active_image_id != SAHARA_IMAGE_ID_NONE) { in sahara_find_image()
239 dev_err(&context->mhi_dev->dev, "image id %d is not valid as %d is active\n", in sahara_find_image()
240 image_id, context->active_image_id); in sahara_find_image()
244 if (image_id >= context->table_size || !context->image_table[image_id]) { in sahara_find_image()
245 dev_err(&context->mhi_dev->dev, "request for unknown image: %d\n", image_id); in sahara_find_image()
254 ret = firmware_request_nowarn(&context->firmware, in sahara_find_image()
[all …]
/linux/drivers/misc/vmw_vmci/
H A Dvmci_context.c44 static void ctx_signal_notify(struct vmci_ctx *context) in ctx_signal_notify() argument
46 *context->notify = true; in ctx_signal_notify()
49 static void ctx_clear_notify(struct vmci_ctx *context) in ctx_clear_notify() argument
51 *context->notify = false; in ctx_clear_notify()
58 static void ctx_clear_notify_call(struct vmci_ctx *context) in ctx_clear_notify_call() argument
60 if (context->pending_datagrams == 0 && in ctx_clear_notify_call()
61 vmci_handle_arr_get_size(context->pending_doorbell_array) == 0) in ctx_clear_notify_call()
62 ctx_clear_notify(context); in ctx_clear_notify_call()
69 void vmci_ctx_check_signal_notify(struct vmci_ctx *context) in vmci_ctx_check_signal_notify() argument
71 spin_lock(&context->lock); in vmci_ctx_check_signal_notify()
[all …]
H A Dvmci_route.c42 if (VMCI_INVALID_ID == dst->context) in vmci_route()
46 if (VMCI_HYPERVISOR_CONTEXT_ID == dst->context) { in vmci_route()
64 if (VMCI_HOST_CONTEXT_ID == src->context) in vmci_route()
75 if (VMCI_INVALID_ID == src->context && in vmci_route()
77 src->context = vmci_get_context_id(); in vmci_route()
85 if (VMCI_HOST_CONTEXT_ID == dst->context) { in vmci_route()
94 if (src->context == VMCI_HYPERVISOR_CONTEXT_ID) { in vmci_route()
113 if (VMCI_INVALID_ID == src->context) in vmci_route()
114 src->context = vmci_get_context_id(); in vmci_route()
130 if (VMCI_INVALID_ID == src->context) { in vmci_route()
[all …]
/linux/drivers/gpu/drm/etnaviv/
H A Detnaviv_mmu.c18 static void etnaviv_context_unmap(struct etnaviv_iommu_context *context, in etnaviv_context_unmap() argument
25 unmapped_page = context->global->ops->unmap(context, iova, in etnaviv_context_unmap()
35 static int etnaviv_context_map(struct etnaviv_iommu_context *context, in etnaviv_context_map() argument
45 ret = context->global->ops->map(context, iova, paddr, pgsize, in etnaviv_context_map()
57 etnaviv_context_unmap(context, orig_iova, orig_size - size); in etnaviv_context_map()
62 static int etnaviv_iommu_map(struct etnaviv_iommu_context *context, in etnaviv_iommu_map() argument
71 if (!context || !sgt) in etnaviv_iommu_map()
82 dev_err(context->global->dev, in etnaviv_iommu_map()
89 ret = etnaviv_context_map(context, da, pa, bytes, prot); in etnaviv_iommu_map()
97 context->flush_seq++; in etnaviv_iommu_map()
[all …]
H A Detnaviv_iommu.c28 to_v1_context(struct etnaviv_iommu_context *context) in to_v1_context() argument
30 return container_of(context, struct etnaviv_iommuv1_context, base); in to_v1_context()
33 static void etnaviv_iommuv1_free(struct etnaviv_iommu_context *context) in etnaviv_iommuv1_free() argument
35 struct etnaviv_iommuv1_context *v1_context = to_v1_context(context); in etnaviv_iommuv1_free()
37 drm_mm_takedown(&context->mm); in etnaviv_iommuv1_free()
39 dma_free_wc(context->global->dev, PT_SIZE, v1_context->pgtable_cpu, in etnaviv_iommuv1_free()
42 context->global->v1.shared_context = NULL; in etnaviv_iommuv1_free()
47 static int etnaviv_iommuv1_map(struct etnaviv_iommu_context *context, in etnaviv_iommuv1_map() argument
51 struct etnaviv_iommuv1_context *v1_context = to_v1_context(context); in etnaviv_iommuv1_map()
62 static size_t etnaviv_iommuv1_unmap(struct etnaviv_iommu_context *context, in etnaviv_iommuv1_unmap() argument
[all …]
H A Detnaviv_iommu_v2.c42 to_v2_context(struct etnaviv_iommu_context *context) in to_v2_context() argument
44 return container_of(context, struct etnaviv_iommuv2_context, base); in to_v2_context()
47 static void etnaviv_iommuv2_free(struct etnaviv_iommu_context *context) in etnaviv_iommuv2_free() argument
49 struct etnaviv_iommuv2_context *v2_context = to_v2_context(context); in etnaviv_iommuv2_free()
52 drm_mm_takedown(&context->mm); in etnaviv_iommuv2_free()
56 dma_free_wc(context->global->dev, SZ_4K, in etnaviv_iommuv2_free()
61 dma_free_wc(context->global->dev, SZ_4K, v2_context->mtlb_cpu, in etnaviv_iommuv2_free()
64 clear_bit(v2_context->id, context->global->v2.pta_alloc); in etnaviv_iommuv2_free()
92 static int etnaviv_iommuv2_map(struct etnaviv_iommu_context *context, in etnaviv_iommuv2_map() argument
96 struct etnaviv_iommuv2_context *v2_context = to_v2_context(context); in etnaviv_iommuv2_map()
[all …]
/linux/security/selinux/ss/
H A Dmls.h27 int mls_compute_context_len(struct policydb *p, struct context *context);
28 void mls_sid_to_context(struct policydb *p, struct context *context,
30 int mls_context_isvalid(struct policydb *p, struct context *c);
35 struct context *context, struct sidtab *s, u32 def_sid);
37 int mls_from_string(struct policydb *p, char *str, struct context *context,
40 int mls_range_set(struct context *context, struct mls_range *range);
43 struct context *oldc, struct context *newc);
45 int mls_compute_sid(struct policydb *p, struct context *scontext,
46 struct context *tcontext, u16 tclass, u32 specified,
47 struct context *newcontext, bool sock);
[all …]
H A Dmls.c32 int mls_compute_context_len(struct policydb *p, struct context *context) in mls_compute_context_len() argument
44 u32 index_sens = context->range.level[l].sens; in mls_compute_context_len()
50 e = &context->range.level[l].cat; in mls_compute_context_len()
70 if (mls_level_eq(&context->range.level[0], in mls_compute_context_len()
71 &context->range.level[1])) in mls_compute_context_len()
86 void mls_sid_to_context(struct policydb *p, struct context *context, in mls_sid_to_context() argument
104 context->range.level[l].sens - 1)); in mls_sid_to_context()
110 e = &context->range.level[l].cat; in mls_sid_to_context()
147 if (mls_level_eq(&context->range.level[0], in mls_sid_to_context()
148 &context->range.level[1])) in mls_sid_to_context()
[all …]
/linux/kernel/
H A Dauditsc.c916 static inline void audit_proctitle_free(struct audit_context *context) in audit_proctitle_free() argument
918 kfree(context->proctitle.value); in audit_proctitle_free()
919 context->proctitle.value = NULL; in audit_proctitle_free()
920 context->proctitle.len = 0; in audit_proctitle_free()
923 static inline void audit_free_module(struct audit_context *context) in audit_free_module() argument
925 if (context->type == AUDIT_KERN_MODULE) { in audit_free_module()
926 kfree(context->module.name); in audit_free_module()
927 context->module.name = NULL; in audit_free_module()
930 static inline void audit_free_names(struct audit_context *context) in audit_free_names() argument
934 list_for_each_entry_safe(n, next, &context->names_list, list) { in audit_free_names()
[all …]
/linux/fs/xfs/
H A Dxfs_attr_list.c55 struct xfs_attr_list_context *context) in xfs_attr_shortform_list() argument
57 struct xfs_attrlist_cursor_kern *cursor = &context->cursor; in xfs_attr_shortform_list()
58 struct xfs_inode *dp = context->dp; in xfs_attr_shortform_list()
69 trace_xfs_attr_list_sf(context); in xfs_attr_shortform_list()
80 if (context->bufsize == 0 || in xfs_attr_shortform_list()
82 (dp->i_af.if_bytes + sf->count * 16) < context->bufsize)) { in xfs_attr_shortform_list()
84 if (XFS_IS_CORRUPT(context->dp->i_mount, in xfs_attr_shortform_list()
88 xfs_dirattr_mark_sick(context->dp, XFS_ATTR_FORK); in xfs_attr_shortform_list()
91 context->put_listent(context, in xfs_attr_shortform_list()
101 if (context->seen_enough) in xfs_attr_shortform_list()
[all …]
H A Dxfs_handle.c148 void *context, in xfs_handle_acceptable() argument
329 struct xfs_attr_list_context *context, in xfs_ioc_attr_put_listent() argument
336 struct xfs_attrlist *alist = context->buffer; in xfs_ioc_attr_put_listent()
340 ASSERT(!context->seen_enough); in xfs_ioc_attr_put_listent()
341 ASSERT(context->count >= 0); in xfs_ioc_attr_put_listent()
342 ASSERT(context->count < (ATTR_MAX_VALUELEN/8)); in xfs_ioc_attr_put_listent()
343 ASSERT(context->firstu >= sizeof(*alist)); in xfs_ioc_attr_put_listent()
344 ASSERT(context->firstu <= context->bufsize); in xfs_ioc_attr_put_listent()
349 if (context->attr_filter != (flags & XFS_ATTR_NSP_ONDISK_MASK)) in xfs_ioc_attr_put_listent()
353 context->count * sizeof(alist->al_offset[0]); in xfs_ioc_attr_put_listent()
[all …]
H A Dxfs_xattr.c222 struct xfs_attr_list_context *context, in __xfs_xattr_put_listent() argument
231 if (context->count < 0 || context->seen_enough) in __xfs_xattr_put_listent()
234 if (!context->buffer) in __xfs_xattr_put_listent()
237 arraytop = context->count + prefix_len + namelen + 1; in __xfs_xattr_put_listent()
238 if (arraytop > context->firstu) { in __xfs_xattr_put_listent()
239 context->count = -1; /* insufficient space */ in __xfs_xattr_put_listent()
240 context->seen_enough = 1; in __xfs_xattr_put_listent()
243 offset = context->buffer + context->count; in __xfs_xattr_put_listent()
251 context->count += prefix_len + namelen + 1; in __xfs_xattr_put_listent()
257 struct xfs_attr_list_context *context, in xfs_xattr_put_listent() argument
[all …]
/linux/drivers/gpu/drm/tegra/
H A Dsubmit.c26 #define SUBMIT_ERR(context, fmt, ...) \ argument
27 dev_err_ratelimited(context->client->base.dev, \
146 tegra_drm_mapping_get(struct tegra_drm_context *context, u32 id) in tegra_drm_mapping_get() argument
150 xa_lock(&context->mappings); in tegra_drm_mapping_get()
152 mapping = xa_load(&context->mappings, id); in tegra_drm_mapping_get()
156 xa_unlock(&context->mappings); in tegra_drm_mapping_get()
180 struct tegra_drm_context *context, in submit_copy_gather_data() argument
187 SUBMIT_ERR(context, "gather_data_words cannot be zero"); in submit_copy_gather_data()
192 SUBMIT_ERR(context, "gather_data_words is too large"); in submit_copy_gather_data()
198 SUBMIT_ERR(context, "failed to allocate memory for bo info"); in submit_copy_gather_data()
[all …]
H A Duapi.c31 static void tegra_drm_channel_context_close(struct tegra_drm_context *context) in tegra_drm_channel_context_close() argument
36 if (context->memory_context) in tegra_drm_channel_context_close()
37 host1x_memory_context_put(context->memory_context); in tegra_drm_channel_context_close()
39 xa_for_each(&context->mappings, id, mapping) in tegra_drm_channel_context_close()
42 xa_destroy(&context->mappings); in tegra_drm_channel_context_close()
44 host1x_channel_put(context->channel); in tegra_drm_channel_context_close()
46 kfree(context); in tegra_drm_channel_context_close()
51 struct tegra_drm_context *context; in tegra_drm_uapi_close_file() local
55 xa_for_each(&file->contexts, id, context) in tegra_drm_uapi_close_file()
56 tegra_drm_channel_context_close(context); in tegra_drm_uapi_close_file()
[all …]
/linux/drivers/platform/x86/intel/int1092/
H A Dintel_sar.c38 static void update_sar_data(struct wwan_sar_context *context) in update_sar_data() argument
41 &context->config_data[context->reg_value]; in update_sar_data()
44 context->sar_data.device_mode < config->total_dev_mode) { in update_sar_data()
48 if (context->sar_data.device_mode == in update_sar_data()
53 context->sar_data.antennatable_index = dev_mode->antennatable_index; in update_sar_data()
54 context->sar_data.bandtable_index = dev_mode->bandtable_index; in update_sar_data()
55 context->sar_data.sartable_index = dev_mode->sartable_index; in update_sar_data()
77 static acpi_status parse_package(struct wwan_sar_context *context, union acpi_object *item) in parse_package() argument
89 data = &context->config_data[reg]; in parse_package()
130 struct wwan_sar_context *context = dev_get_drvdata(&device->dev); in sar_get_device_mode() local
[all …]
/linux/drivers/gpu/drm/amd/display/dc/dml/dcn32/
H A Ddcn32_fpu.c183 static bool dcn32_apply_merge_split_flags_helper(struct dc *dc, struct dc_state *context,
276 struct dc_state *context, in dcn32_find_dummy_latency_index_for_fw_based_mclk_switch() argument
282 struct vba_vars_st *vba = &context->bw_ctx.dml.vba; in dcn32_find_dummy_latency_index_for_fw_based_mclk_switch()
284 …enum clock_change_support temp_clock_change_support = vba->DRAMClockChangeSupport[vlevel][context-… in dcn32_find_dummy_latency_index_for_fw_based_mclk_switch()
290 …vba->DRAMClockChangeSupport[vlevel][context->bw_ctx.dml.vba.maxMpcComb] = temp_clock_change_suppor… in dcn32_find_dummy_latency_index_for_fw_based_mclk_switch()
291 context->bw_ctx.dml.soc.dram_clock_change_latency_us = in dcn32_find_dummy_latency_index_for_fw_based_mclk_switch()
293 …dcn32_internal_validate_bw(dc, context, pipes, &pipe_cnt, &vlevel, DC_VALIDATE_MODE_AND_PROGRAMMIN… in dcn32_find_dummy_latency_index_for_fw_based_mclk_switch()
297 dcn32_subvp_in_use(dc, context)) in dcn32_find_dummy_latency_index_for_fw_based_mclk_switch()
298 …vba->DRAMClockChangeSupport[vlevel][context->bw_ctx.dml.vba.maxMpcComb] = temp_clock_change_suppor… in dcn32_find_dummy_latency_index_for_fw_based_mclk_switch()
300 if (vlevel < context->bw_ctx.dml.vba.soc.num_states && in dcn32_find_dummy_latency_index_for_fw_based_mclk_switch()
[all …]
/linux/drivers/gpu/drm/amd/display/dc/dml/dcn30/
H A Ddcn30_fpu.c289 void dcn30_fpu_update_soc_for_wm_a(struct dc *dc, struct dc_state *context) in dcn30_fpu_update_soc_for_wm_a() argument
295 if (!context->bw_ctx.bw.dcn.clk.fw_based_mclk_switching || in dcn30_fpu_update_soc_for_wm_a()
296 context->bw_ctx.dml.soc.dram_clock_change_latency_us == 0) in dcn30_fpu_update_soc_for_wm_a()
297context->bw_ctx.dml.soc.dram_clock_change_latency_us = dc->clk_mgr->bw_params->wm_table.nv_entries… in dcn30_fpu_update_soc_for_wm_a()
298context->bw_ctx.dml.soc.sr_enter_plus_exit_time_us = dc->clk_mgr->bw_params->wm_table.nv_entries[W… in dcn30_fpu_update_soc_for_wm_a()
299context->bw_ctx.dml.soc.sr_exit_time_us = dc->clk_mgr->bw_params->wm_table.nv_entries[WM_A].dml_in… in dcn30_fpu_update_soc_for_wm_a()
304 struct dc *dc, struct dc_state *context, in dcn30_fpu_calculate_wm_and_dlg() argument
309 int maxMpcComb = context->bw_ctx.dml.vba.maxMpcComb; in dcn30_fpu_calculate_wm_and_dlg()
311 double dcfclk = context->bw_ctx.dml.vba.DCFCLKState[vlevel][maxMpcComb]; in dcn30_fpu_calculate_wm_and_dlg()
312 …bool pstate_en = context->bw_ctx.dml.vba.DRAMClockChangeSupport[vlevel][maxMpcComb] != dm_dram_clo… in dcn30_fpu_calculate_wm_and_dlg()
[all …]
/linux/drivers/net/ethernet/mellanox/mlx4/
H A Den_resources.c42 int user_prio, struct mlx4_qp_context *context) in mlx4_en_fill_qp_context() argument
47 memset(context, 0, sizeof(*context)); in mlx4_en_fill_qp_context()
48 context->flags = cpu_to_be32(7 << 16 | rss << MLX4_RSS_QPC_FLAG_OFFSET); in mlx4_en_fill_qp_context()
49 context->pd = cpu_to_be32(mdev->priv_pdn); in mlx4_en_fill_qp_context()
50 context->mtu_msgmax = 0xff; in mlx4_en_fill_qp_context()
52 context->rq_size_stride = ilog2(size) << 3 | (ilog2(stride) - 4); in mlx4_en_fill_qp_context()
54 context->sq_size_stride = ilog2(size) << 3 | (ilog2(stride) - 4); in mlx4_en_fill_qp_context()
56 context->params2 |= cpu_to_be32(MLX4_QP_BIT_FPP); in mlx4_en_fill_qp_context()
59 context->sq_size_stride = ilog2(TXBB_SIZE) - 4; in mlx4_en_fill_qp_context()
61 context->usr_page = cpu_to_be32(mlx4_to_hw_uar_index(mdev->dev, in mlx4_en_fill_qp_context()
[all …]
/linux/drivers/net/ethernet/qlogic/qed/
H A Dqed_nvmetcp_fw_funcs.c129 init_nvmetcp_task_params(struct e5_nvmetcp_task_context *context, in init_nvmetcp_task_params() argument
133 context->ystorm_st_context.state.cccid = task_params->host_cccid; in init_nvmetcp_task_params()
134 SET_FIELD(context->ustorm_st_context.error_flags, USTORM_NVMETCP_TASK_ST_CTX_NVME_TCP, 1); in init_nvmetcp_task_params()
135 context->ustorm_st_context.nvme_tcp_opaque_lo = cpu_to_le32(task_params->opq.lo); in init_nvmetcp_task_params()
136 context->ustorm_st_context.nvme_tcp_opaque_hi = cpu_to_le32(task_params->opq.hi); in init_nvmetcp_task_params()
145 struct e5_nvmetcp_task_context *context = task_params->context; in init_default_nvmetcp_task() local
146 const u8 val_byte = context->mstorm_ag_context.cdu_validation; in init_default_nvmetcp_task()
149 memset(context, 0, sizeof(*context)); in init_default_nvmetcp_task()
150 init_nvmetcp_task_params(context, task_params, in init_default_nvmetcp_task()
159 context->ystorm_st_context.pdu_hdr.task_hdr.reg[dw_index] = in init_default_nvmetcp_task()
[all …]
/linux/drivers/gpu/drm/amd/display/dc/dml2_0/dml21/
H A Ddml21_wrapper.c87 static void dml21_calculate_rq_and_dlg_params(const struct dc *dc, struct dc_state *context, struct… in dml21_calculate_rq_and_dlg_params() argument
98 context->bw_ctx.bw.dcn.clk.dppclk_khz = 0; in dml21_calculate_rq_and_dlg_params()
101 …memcpy(&context->bw_ctx.bw.dcn.arb_regs, &in_ctx->v21.mode_programming.programming->global_regs.ar… in dml21_calculate_rq_and_dlg_params()
104context->bw_ctx.bw.dcn.compbuf_size_kb = (int)in_ctx->v21.mode_programming.programming->global_reg… in dml21_calculate_rq_and_dlg_params()
106 context->bw_ctx.bw.dcn.mall_ss_size_bytes = 0; in dml21_calculate_rq_and_dlg_params()
107 context->bw_ctx.bw.dcn.mall_ss_psr_active_size_bytes = 0; in dml21_calculate_rq_and_dlg_params()
108 context->bw_ctx.bw.dcn.mall_subvp_size_bytes = 0; in dml21_calculate_rq_and_dlg_params()
125 …num_pipes = dml21_find_dc_pipes_for_plane(dc, context, in_ctx, dc_main_pipes, dc_phantom_pipes, dm… in dml21_calculate_rq_and_dlg_params()
132 dml21_program_dc_pipe(in_ctx, context, dc_main_pipes[dc_pipe_index], pln_prog, stream_prog); in dml21_calculate_rq_and_dlg_params()
135 dml21_program_dc_pipe(in_ctx, context, dc_phantom_pipes[dc_pipe_index], pln_prog, stream_prog); in dml21_calculate_rq_and_dlg_params()
[all …]
/linux/drivers/gpu/drm/amd/display/dc/resource/dcn32/
H A Ddcn32_resource_helpers.c92 struct dc_state *context) in dcn32_helper_calculate_num_ways_for_subvp() argument
94 if (context->bw_ctx.bw.dcn.mall_subvp_size_bytes > 0) { in dcn32_helper_calculate_num_ways_for_subvp()
98 …return dc->res_pool->funcs->calculate_mall_ways_from_bytes(dc, context->bw_ctx.bw.dcn.mall_subvp_s… in dcn32_helper_calculate_num_ways_for_subvp()
108 struct dc_state *context) in dcn32_merge_pipes_for_subvp() argument
114 struct pipe_ctx *pipe = &context->res_ctx.pipe_ctx[i]; in dcn32_merge_pipes_for_subvp()
132 dcn20_release_dsc(&context->res_ctx, dc->res_pool, &pipe->stream_res.dsc); in dcn32_merge_pipes_for_subvp()
154 struct dc_state *context) in dcn32_all_pipes_have_stream_and_plane() argument
159 struct pipe_ctx *pipe = &context->res_ctx.pipe_ctx[i]; in dcn32_all_pipes_have_stream_and_plane()
171 struct dc_state *context) in dcn32_subvp_in_use() argument
176 struct pipe_ctx *pipe = &context->res_ctx.pipe_ctx[i]; in dcn32_subvp_in_use()
[all …]
/linux/arch/sparc/mm/
H A Dtsb.c124 spin_lock_irqsave(&mm->context.lock, flags); in flush_tsb_user()
127 base = (unsigned long) mm->context.tsb_block[MM_TSB_BASE].tsb; in flush_tsb_user()
128 nentries = mm->context.tsb_block[MM_TSB_BASE].tsb_nentries; in flush_tsb_user()
140 else if (mm->context.tsb_block[MM_TSB_HUGE].tsb) { in flush_tsb_user()
141 base = (unsigned long) mm->context.tsb_block[MM_TSB_HUGE].tsb; in flush_tsb_user()
142 nentries = mm->context.tsb_block[MM_TSB_HUGE].tsb_nentries; in flush_tsb_user()
149 spin_unlock_irqrestore(&mm->context.lock, flags); in flush_tsb_user()
157 spin_lock_irqsave(&mm->context.lock, flags); in flush_tsb_user_page()
160 base = (unsigned long) mm->context.tsb_block[MM_TSB_BASE].tsb; in flush_tsb_user_page()
161 nentries = mm->context.tsb_block[MM_TSB_BASE].tsb_nentries; in flush_tsb_user_page()
[all …]
/linux/drivers/usb/image/
H A Dmicrotek.c191 MTS_DEBUG("transfer = 0x%x context = 0x%x\n",(int)transfer,(int)context ); \
192 …MTS_DEBUG("status = 0x%x data-length = 0x%x sent = 0x%x\n",transfer->status,(int)context->data_len…
193 mts_debug_dump(context->instance);\
208 struct mts_transfer_context* context = (struct mts_transfer_context*)transfer->context; \
379 context->instance->usb_dev, in mts_int_submit_urb()
384 context in mts_int_submit_urb()
390 set_host_byte(context->srb, DID_ERROR); in mts_int_submit_urb()
401 if ( likely(context->final_callback != NULL) ) in mts_transfer_cleanup()
402 context->final_callback(context->srb); in mts_transfer_cleanup()
409 context->srb->result &= MTS_SCSI_ERR_MASK; in mts_transfer_done()
[all …]
/linux/arch/powerpc/mm/book3s64/
H A Dmmu_context.c99 mm->context.hash_context = kmalloc_obj(struct hash_mm_context); in hash__init_new_context()
100 if (!mm->context.hash_context) in hash__init_new_context()
117 if (mm->context.id == 0) { in hash__init_new_context()
118 memset(mm->context.hash_context, 0, sizeof(struct hash_mm_context)); in hash__init_new_context()
122 …memcpy(mm->context.hash_context, current->mm->context.hash_context, sizeof(struct hash_mm_context)… in hash__init_new_context()
125 if (current->mm->context.hash_context->spt) { in hash__init_new_context()
126 mm->context.hash_context->spt = kmalloc_obj(struct subpage_prot_table); in hash__init_new_context()
127 if (!mm->context.hash_context->spt) { in hash__init_new_context()
128 kfree(mm->context.hash_context); in hash__init_new_context()
135 index = realloc_context_ids(&mm->context); in hash__init_new_context()
[all …]

12345678910>>...104