Home
last modified time | relevance | path

Searched refs:xe_gt_assert (Results 1 – 22 of 22) sorted by relevance

/linux/drivers/gpu/drm/xe/
H A Dxe_gt_sriov_pf_monitor.c28 xe_gt_assert(gt, IS_SRIOV_PF(gt_to_xe(gt))); in xe_gt_sriov_pf_monitor_flr()
38 xe_gt_assert(gt, IS_SRIOV_PF(gt_to_xe(gt))); in pf_update_event_counter()
39 xe_gt_assert(gt, e < XE_GUC_KLV_NUM_THRESHOLDS); in pf_update_event_counter()
84 xe_gt_assert(gt, len >= GUC_HXG_MSG_MIN_LEN); in xe_gt_sriov_pf_monitor_process_guc2pf()
85 xe_gt_assert(gt, FIELD_GET(GUC_HXG_MSG_0_ORIGIN, msg[0]) == GUC_HXG_ORIGIN_GUC); in xe_gt_sriov_pf_monitor_process_guc2pf()
86 xe_gt_assert(gt, FIELD_GET(GUC_HXG_MSG_0_TYPE, msg[0]) == GUC_HXG_TYPE_EVENT); in xe_gt_sriov_pf_monitor_process_guc2pf()
87 xe_gt_assert(gt, FIELD_GET(GUC_HXG_EVENT_MSG_0_ACTION, msg[0]) == in xe_gt_sriov_pf_monitor_process_guc2pf()
124 xe_gt_assert(gt, IS_SRIOV_PF(gt_to_xe(gt))); in xe_gt_sriov_pf_monitor_print_events()
H A Dxe_gt_sriov_vf.c206 xe_gt_assert(gt, IS_SRIOV_VF(gt_to_xe(gt))); in vf_handshake_with_guc()
214 xe_gt_assert(gt, wanted.major != GUC_VERSION_MAJOR_ANY); in vf_handshake_with_guc()
255 xe_gt_assert(gt, wanted.major != GUC_VERSION_MAJOR_ANY); in vf_handshake_with_guc()
323 xe_gt_assert(gt, IS_SRIOV_VF(gt_to_xe(gt))); in xe_gt_sriov_vf_guc_versions()
324 xe_gt_assert(gt, gt->sriov.vf.guc_version.major); in xe_gt_sriov_vf_guc_versions()
352 xe_gt_assert(gt, IS_SRIOV_VF(gt_to_xe(gt))); in vf_resfix_start()
380 xe_gt_assert(gt, IS_SRIOV_VF(gt_to_xe(gt))); in vf_resfix_done()
473 xe_gt_assert(gt, IS_SRIOV_VF(gt_to_xe(gt))); in xe_gt_sriov_vf_gmdid()
474 xe_gt_assert(gt, !GRAPHICS_VERx100(gt_to_xe(gt)) || has_gmdid(gt_to_xe(gt))); in xe_gt_sriov_vf_gmdid()
475 xe_gt_assert(gt, gt->sriov.vf.guc_version.major > 1 || gt->sriov.vf.guc_version.minor >= 2); in xe_gt_sriov_vf_gmdid()
[all …]
H A Dxe_guc_submit.c257 xe_gt_assert(gt, ret); in guc_submit_sw_fini()
532 xe_gt_assert(guc_to_gt(guc), exec_queue_registered(q) && in init_policies()
584 xe_gt_assert(guc_to_gt(exec_queue_to_guc(q)), in xe_guc_exec_queue_group_trigger_cleanup()
632 xe_gt_assert(guc_to_gt(exec_queue_to_guc(q)), in set_exec_queue_group_banned()
807 xe_gt_assert(guc_to_gt(guc), len <= MAX_MULTI_QUEUE_REG_SIZE); in __register_exec_queue_group()
825 xe_gt_assert(guc_to_gt(guc), xe_exec_queue_is_multi_queue_secondary(q)); in xe_guc_exec_queue_group_add()
830 xe_gt_assert(guc_to_gt(guc), len <= MAX_MULTI_QUEUE_CGP_SYNC_SIZE); in xe_guc_exec_queue_group_add()
850 xe_gt_assert(guc_to_gt(guc), xe_exec_queue_is_parallel(q)); in __register_mlrc_exec_queue()
874 xe_gt_assert(guc_to_gt(guc), info->wq_desc_lo == in __register_mlrc_exec_queue()
876 xe_gt_assert(guc_to_gt(guc), info->wq_base_lo == in __register_mlrc_exec_queue()
[all …]
H A Dxe_bb.c105 xe_gt_assert(q->gt, bb->len * 4 + bb_prefetch(q->gt) <= size); in __xe_bb_create_job()
123 xe_gt_assert(q->gt, second_idx <= bb->len); in xe_bb_create_migration_job()
124 xe_gt_assert(q->gt, xe_sched_job_is_migration(q)); in xe_bb_create_migration_job()
125 xe_gt_assert(q->gt, q->width == 1); in xe_bb_create_migration_job()
135 xe_gt_assert(q->gt, !xe_sched_job_is_migration(q)); in xe_bb_create_job()
136 xe_gt_assert(q->gt, q->width == 1); in xe_bb_create_job()
H A Dxe_gt_sriov_pf_debugfs.c150 xe_gt_assert(gt, gt == extract_gt(parent)); in pf_add_policy_attrs()
151 xe_gt_assert(gt, PFID == extract_vfid(parent)); in pf_add_policy_attrs()
410 xe_gt_assert(gt, gt == extract_gt(parent)); in pf_add_sched_groups()
411 xe_gt_assert(gt, vfid == extract_vfid(parent)); in pf_add_sched_groups()
582 xe_gt_assert(gt, gt == extract_gt(parent)); in MAKE_XE_GUC_KLV_THRESHOLDS_SET()
583 xe_gt_assert(gt, vfid == extract_vfid(parent)); in MAKE_XE_GUC_KLV_THRESHOLDS_SET()
635 xe_gt_assert(gt, vfid); in control_write()
650 xe_gt_assert(gt, sizeof(cmd) > strlen(control_cmds[n].cmd)); in control_write()
852 xe_gt_assert(gt, IS_SRIOV_PF(gt_to_xe(gt))); in xe_gt_sriov_pf_debugfs_populate()
853 xe_gt_assert(gt, extract_priv(parent) == gt->tile); in xe_gt_sriov_pf_debugfs_populate()
[all …]
H A Dxe_guc_ct.c313 xe_gt_assert(gt, !(guc_ct_size() % PAGE_SIZE)); in xe_guc_ct_init_noalloc()
338 xe_gt_assert(gt, ct->state == XE_GUC_CT_STATE_NOT_INITIALIZED); in xe_guc_ct_init_noalloc()
527 xe_gt_assert(ct_to_gt(ct), ct->g2h_outstanding == 0 || in guc_ct_change_state()
605 xe_gt_assert(gt, !xe_guc_ct_enabled(ct)); in __xe_guc_ct_start()
743 xe_gt_assert(ct_to_gt(ct), g2h->info.space == credits); in xe_guc_ct_runtime_suspend()
823 xe_gt_assert(ct_to_gt(ct), g2h_len <= ct->ctbs.g2h.info.space); in __g2h_reserve_space()
824 xe_gt_assert(ct_to_gt(ct), (!g2h_len && !num_g2h) || in __g2h_reserve_space()
930 xe_gt_assert(gt, full_len <= GUC_CTB_MSG_MAX_LEN); in h2g_write()
1027 xe_gt_assert(gt, xe_guc_ct_initialized(ct)); in __guc_ct_send_locked()
1028 xe_gt_assert(gt, !g2h_len || !g2h_fence); in __guc_ct_send_locked()
[all …]
H A Dxe_gt_sriov_pf_control.c65 xe_gt_assert(gt, vfid != PFID); in pf_send_vf_control_cmd()
234 xe_gt_assert(gt, IS_SRIOV_PF(gt_to_xe(gt))); in pf_pick_vf_control()
235 xe_gt_assert(gt, vfid <= xe_gt_sriov_pf_get_totalvfs(gt)); in pf_pick_vf_control()
360 xe_gt_assert(gt, IS_SRIOV_PF(xe)); in pf_queue_control_worker()
369 xe_gt_assert(gt, IS_SRIOV_PF(gt_to_xe(gt))); in pf_queue_vf()
1964 xe_gt_assert(gt, len); in xe_gt_sriov_pf_control_process_guc2pf()
1965 xe_gt_assert(gt, FIELD_GET(GUC_HXG_MSG_0_ORIGIN, msg[0]) == GUC_HXG_ORIGIN_GUC); in xe_gt_sriov_pf_control_process_guc2pf()
1966 xe_gt_assert(gt, FIELD_GET(GUC_HXG_MSG_0_TYPE, msg[0]) == GUC_HXG_TYPE_EVENT); in xe_gt_sriov_pf_control_process_guc2pf()
1967 xe_gt_assert(gt, FIELD_GET(GUC_HXG_EVENT_MSG_0_ACTION, msg[0]) == in xe_gt_sriov_pf_control_process_guc2pf()
2076 xe_gt_assert(gt, vfid <= xe_gt_sriov_pf_get_totalvfs(gt)); in pf_worker_find_work()
[all …]
H A Dxe_gt_topology.c26 xe_gt_assert(gt, numregs <= ARRAY_SIZE(fuse_val)); in load_dss_mask()
249 xe_gt_assert(gt, num_geometry_regs <= ARRAY_SIZE(geometry_regs)); in xe_gt_topology_init()
250 xe_gt_assert(gt, num_compute_regs <= ARRAY_SIZE(compute_regs)); in xe_gt_topology_init()
H A Dxe_ring_ops.c297 xe_gt_assert(gt, i <= MAX_JOB_SIZE_DW); in __emit_job_gen12_simple()
365 xe_gt_assert(gt, i <= MAX_JOB_SIZE_DW); in __emit_job_gen12_video()
420 xe_gt_assert(gt, i <= MAX_JOB_SIZE_DW); in __emit_job_gen12_render_compute()
456 xe_gt_assert(job->q->gt, i <= MAX_JOB_SIZE_DW); in emit_migration_job_gen12()
465 xe_gt_assert(gt, job->q->width <= 1); /* no parallel submission for GSCCS */ in emit_job_gen12_gsc()
H A Dxe_guc_ads.c428 xe_gt_assert(gt, ads->bo); in xe_guc_ads_init_post_hwconfig()
435 xe_gt_assert(gt, ads->golden_lrc_size + in xe_guc_ads_init_post_hwconfig()
815 xe_gt_assert(gt, regset_used <= ads->regset_size); in guc_mmio_reg_state_init()
874 xe_gt_assert(gt, ads->bo); in xe_guc_ads_populate_minimal()
898 xe_gt_assert(gt, ads->bo); in xe_guc_ads_populate()
950 xe_gt_assert(gt, gt->default_lrc[class]); in guc_golden_lrc_populate()
962 xe_gt_assert(gt, total_size == ads->golden_lrc_size); in guc_golden_lrc_populate()
H A Dxe_migrate.c679 xe_gt_assert(gt, FIELD_FIT(XE2_CCS_SIZE_MASK, num_pages - 1)); in emit_copy_ccs()
687 xe_gt_assert(gt, FIELD_FIT(CCS_SIZE_MASK, num_ccs_blks - 1)); in emit_copy_ccs()
714 xe_gt_assert(gt, !(pitch & 3)); in emit_xy_fast_copy()
715 xe_gt_assert(gt, size / pitch <= S16_MAX); in emit_xy_fast_copy()
716 xe_gt_assert(gt, pitch / 4 <= S16_MAX); in emit_xy_fast_copy()
717 xe_gt_assert(gt, pitch <= U16_MAX); in emit_xy_fast_copy()
743 xe_gt_assert(gt, IS_ALIGNED(size, pitch)); in emit_mem_copy()
744 xe_gt_assert(gt, pitch <= U16_MAX); in emit_mem_copy()
745 xe_gt_assert(gt, pitch); in emit_mem_copy()
746 xe_gt_assert(gt, size); in emit_mem_copy()
[all …]
H A Dxe_hw_engine.c298 xe_gt_assert(hwe->gt, !(reg.addr & hwe->mmio_base)); in xe_hw_engine_mmio_write32()
318 xe_gt_assert(hwe->gt, !(reg.addr & hwe->mmio_base)); in xe_hw_engine_mmio_read32()
497 xe_gt_assert(gt, info); in get_msix_irq_offset()
515 xe_gt_assert(gt, !hwe->gt); in hw_engine_init_early()
613 xe_gt_assert(gt, id < ARRAY_SIZE(engine_infos) && engine_infos[id].name); in hw_engine_init()
614 xe_gt_assert(gt, gt->info.engine_mask & BIT(id)); in hw_engine_init()
H A Dxe_wa.h33 xe_gt_assert(gt__, (gt__)->wa_active.oob_initialized); \
H A Dxe_assert.h169 #define xe_gt_assert(gt, condition) xe_gt_assert_msg((gt), condition, "") macro
H A Dxe_execlist.c55 xe_gt_assert(hwe->gt, FIELD_FIT(XEHP_SW_CTX_ID, ctx_id)); in __start_lrc()
58 xe_gt_assert(hwe->gt, FIELD_FIT(SW_CTX_ID, ctx_id)); in __start_lrc()
H A Dxe_guc_id_mgr.c32 #define idm_assert(idm, cond) xe_gt_assert(idm_to_gt(idm), cond)
H A Dxe_guc_db_mgr.c46 #define dbm_assert(_dbm, _cond) xe_gt_assert(dbm_to_gt(_dbm), _cond)
H A Dxe_lrc.c208 xe_gt_assert(hwe->gt, count); in set_offsets()
1238 xe_gt_assert(state->hwe->gt, state->buffer); in setup_bo()
2245 xe_gt_assert(gt, (instr & XE_INSTR_CMD_TYPE) == XE_INSTR_GFXPIPE); in xe_lrc_emit_hwe_state_instructions()
2246 xe_gt_assert(gt, num_dw != 0); in xe_lrc_emit_hwe_state_instructions()
2247 xe_gt_assert(gt, is_single_dw ^ (num_dw > 1)); in xe_lrc_emit_hwe_state_instructions()
H A Dxe_gt.c764 xe_gt_assert(gt, (gt->user_engines.mask | gt->info.engine_mask) in xe_gt_record_user_engines()
1138 xe_gt_assert(gt, gt_to_xe(gt)->wedged.mode); in xe_gt_declare_wedged()
H A Dxe_guc_engine_activity.c308 xe_gt_assert(guc_to_gt(guc), index < engine_activity->num_activity_group); in engine_activity_set_cpu_ts()
H A Dxe_oa.c1037 xe_gt_assert(stream->gt, !err || err == -ENOENT); in xe_oa_emit_oa_config()
2650 xe_gt_assert(gt, xe_device_has_mert(gt_to_xe(gt))); in __xe_oa_init_oa_units()
2651 xe_gt_assert(gt, gt == xe_root_mmio_gt(gt_to_xe(gt))); in __xe_oa_init_oa_units()
2656 xe_gt_assert(gt, GRAPHICS_VERx100(gt_to_xe(gt)) >= 1270); in __xe_oa_init_oa_units()
H A Dxe_guc.c534 xe_gt_assert(gt, (gt->info.type == XE_GT_TYPE_MAIN) || (gt->info.type == XE_GT_TYPE_MEDIA)); in guc_g2g_start()
1736 xe_gt_assert(guc_to_gt(guc), guc_to_xe(guc)->wedged.mode); in xe_guc_declare_wedged()