| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_gt_sriov_pf_monitor.c | 28 xe_gt_assert(gt, IS_SRIOV_PF(gt_to_xe(gt))); in xe_gt_sriov_pf_monitor_flr() 38 xe_gt_assert(gt, IS_SRIOV_PF(gt_to_xe(gt))); in pf_update_event_counter() 39 xe_gt_assert(gt, e < XE_GUC_KLV_NUM_THRESHOLDS); in pf_update_event_counter() 84 xe_gt_assert(gt, len >= GUC_HXG_MSG_MIN_LEN); in xe_gt_sriov_pf_monitor_process_guc2pf() 85 xe_gt_assert(gt, FIELD_GET(GUC_HXG_MSG_0_ORIGIN, msg[0]) == GUC_HXG_ORIGIN_GUC); in xe_gt_sriov_pf_monitor_process_guc2pf() 86 xe_gt_assert(gt, FIELD_GET(GUC_HXG_MSG_0_TYPE, msg[0]) == GUC_HXG_TYPE_EVENT); in xe_gt_sriov_pf_monitor_process_guc2pf() 87 xe_gt_assert(gt, FIELD_GET(GUC_HXG_EVENT_MSG_0_ACTION, msg[0]) == in xe_gt_sriov_pf_monitor_process_guc2pf() 124 xe_gt_assert(gt, IS_SRIOV_PF(gt_to_xe(gt))); in xe_gt_sriov_pf_monitor_print_events()
|
| H A D | xe_gt_sriov_vf.c | 206 xe_gt_assert(gt, IS_SRIOV_VF(gt_to_xe(gt))); in vf_handshake_with_guc() 214 xe_gt_assert(gt, wanted.major != GUC_VERSION_MAJOR_ANY); in vf_handshake_with_guc() 255 xe_gt_assert(gt, wanted.major != GUC_VERSION_MAJOR_ANY); in vf_handshake_with_guc() 323 xe_gt_assert(gt, IS_SRIOV_VF(gt_to_xe(gt))); in xe_gt_sriov_vf_guc_versions() 324 xe_gt_assert(gt, gt->sriov.vf.guc_version.major); in xe_gt_sriov_vf_guc_versions() 352 xe_gt_assert(gt, IS_SRIOV_VF(gt_to_xe(gt))); in vf_resfix_start() 380 xe_gt_assert(gt, IS_SRIOV_VF(gt_to_xe(gt))); in vf_resfix_done() 473 xe_gt_assert(gt, IS_SRIOV_VF(gt_to_xe(gt))); in xe_gt_sriov_vf_gmdid() 474 xe_gt_assert(gt, !GRAPHICS_VERx100(gt_to_xe(gt)) || has_gmdid(gt_to_xe(gt))); in xe_gt_sriov_vf_gmdid() 475 xe_gt_assert(gt, gt->sriov.vf.guc_version.major > 1 || gt->sriov.vf.guc_version.minor >= 2); in xe_gt_sriov_vf_gmdid() [all …]
|
| H A D | xe_guc_submit.c | 257 xe_gt_assert(gt, ret); in guc_submit_sw_fini() 532 xe_gt_assert(guc_to_gt(guc), exec_queue_registered(q) && in init_policies() 584 xe_gt_assert(guc_to_gt(exec_queue_to_guc(q)), in xe_guc_exec_queue_group_trigger_cleanup() 632 xe_gt_assert(guc_to_gt(exec_queue_to_guc(q)), in set_exec_queue_group_banned() 807 xe_gt_assert(guc_to_gt(guc), len <= MAX_MULTI_QUEUE_REG_SIZE); in __register_exec_queue_group() 825 xe_gt_assert(guc_to_gt(guc), xe_exec_queue_is_multi_queue_secondary(q)); in xe_guc_exec_queue_group_add() 830 xe_gt_assert(guc_to_gt(guc), len <= MAX_MULTI_QUEUE_CGP_SYNC_SIZE); in xe_guc_exec_queue_group_add() 850 xe_gt_assert(guc_to_gt(guc), xe_exec_queue_is_parallel(q)); in __register_mlrc_exec_queue() 874 xe_gt_assert(guc_to_gt(guc), info->wq_desc_lo == in __register_mlrc_exec_queue() 876 xe_gt_assert(guc_to_gt(guc), info->wq_base_lo == in __register_mlrc_exec_queue() [all …]
|
| H A D | xe_bb.c | 105 xe_gt_assert(q->gt, bb->len * 4 + bb_prefetch(q->gt) <= size); in __xe_bb_create_job() 123 xe_gt_assert(q->gt, second_idx <= bb->len); in xe_bb_create_migration_job() 124 xe_gt_assert(q->gt, xe_sched_job_is_migration(q)); in xe_bb_create_migration_job() 125 xe_gt_assert(q->gt, q->width == 1); in xe_bb_create_migration_job() 135 xe_gt_assert(q->gt, !xe_sched_job_is_migration(q)); in xe_bb_create_job() 136 xe_gt_assert(q->gt, q->width == 1); in xe_bb_create_job()
|
| H A D | xe_gt_sriov_pf_debugfs.c | 150 xe_gt_assert(gt, gt == extract_gt(parent)); in pf_add_policy_attrs() 151 xe_gt_assert(gt, PFID == extract_vfid(parent)); in pf_add_policy_attrs() 410 xe_gt_assert(gt, gt == extract_gt(parent)); in pf_add_sched_groups() 411 xe_gt_assert(gt, vfid == extract_vfid(parent)); in pf_add_sched_groups() 582 xe_gt_assert(gt, gt == extract_gt(parent)); in MAKE_XE_GUC_KLV_THRESHOLDS_SET() 583 xe_gt_assert(gt, vfid == extract_vfid(parent)); in MAKE_XE_GUC_KLV_THRESHOLDS_SET() 635 xe_gt_assert(gt, vfid); in control_write() 650 xe_gt_assert(gt, sizeof(cmd) > strlen(control_cmds[n].cmd)); in control_write() 852 xe_gt_assert(gt, IS_SRIOV_PF(gt_to_xe(gt))); in xe_gt_sriov_pf_debugfs_populate() 853 xe_gt_assert(gt, extract_priv(parent) == gt->tile); in xe_gt_sriov_pf_debugfs_populate() [all …]
|
| H A D | xe_guc_ct.c | 313 xe_gt_assert(gt, !(guc_ct_size() % PAGE_SIZE)); in xe_guc_ct_init_noalloc() 338 xe_gt_assert(gt, ct->state == XE_GUC_CT_STATE_NOT_INITIALIZED); in xe_guc_ct_init_noalloc() 527 xe_gt_assert(ct_to_gt(ct), ct->g2h_outstanding == 0 || in guc_ct_change_state() 605 xe_gt_assert(gt, !xe_guc_ct_enabled(ct)); in __xe_guc_ct_start() 743 xe_gt_assert(ct_to_gt(ct), g2h->info.space == credits); in xe_guc_ct_runtime_suspend() 823 xe_gt_assert(ct_to_gt(ct), g2h_len <= ct->ctbs.g2h.info.space); in __g2h_reserve_space() 824 xe_gt_assert(ct_to_gt(ct), (!g2h_len && !num_g2h) || in __g2h_reserve_space() 930 xe_gt_assert(gt, full_len <= GUC_CTB_MSG_MAX_LEN); in h2g_write() 1027 xe_gt_assert(gt, xe_guc_ct_initialized(ct)); in __guc_ct_send_locked() 1028 xe_gt_assert(gt, !g2h_len || !g2h_fence); in __guc_ct_send_locked() [all …]
|
| H A D | xe_gt_sriov_pf_control.c | 65 xe_gt_assert(gt, vfid != PFID); in pf_send_vf_control_cmd() 234 xe_gt_assert(gt, IS_SRIOV_PF(gt_to_xe(gt))); in pf_pick_vf_control() 235 xe_gt_assert(gt, vfid <= xe_gt_sriov_pf_get_totalvfs(gt)); in pf_pick_vf_control() 360 xe_gt_assert(gt, IS_SRIOV_PF(xe)); in pf_queue_control_worker() 369 xe_gt_assert(gt, IS_SRIOV_PF(gt_to_xe(gt))); in pf_queue_vf() 1964 xe_gt_assert(gt, len); in xe_gt_sriov_pf_control_process_guc2pf() 1965 xe_gt_assert(gt, FIELD_GET(GUC_HXG_MSG_0_ORIGIN, msg[0]) == GUC_HXG_ORIGIN_GUC); in xe_gt_sriov_pf_control_process_guc2pf() 1966 xe_gt_assert(gt, FIELD_GET(GUC_HXG_MSG_0_TYPE, msg[0]) == GUC_HXG_TYPE_EVENT); in xe_gt_sriov_pf_control_process_guc2pf() 1967 xe_gt_assert(gt, FIELD_GET(GUC_HXG_EVENT_MSG_0_ACTION, msg[0]) == in xe_gt_sriov_pf_control_process_guc2pf() 2076 xe_gt_assert(gt, vfid <= xe_gt_sriov_pf_get_totalvfs(gt)); in pf_worker_find_work() [all …]
|
| H A D | xe_gt_topology.c | 26 xe_gt_assert(gt, numregs <= ARRAY_SIZE(fuse_val)); in load_dss_mask() 249 xe_gt_assert(gt, num_geometry_regs <= ARRAY_SIZE(geometry_regs)); in xe_gt_topology_init() 250 xe_gt_assert(gt, num_compute_regs <= ARRAY_SIZE(compute_regs)); in xe_gt_topology_init()
|
| H A D | xe_ring_ops.c | 297 xe_gt_assert(gt, i <= MAX_JOB_SIZE_DW); in __emit_job_gen12_simple() 365 xe_gt_assert(gt, i <= MAX_JOB_SIZE_DW); in __emit_job_gen12_video() 420 xe_gt_assert(gt, i <= MAX_JOB_SIZE_DW); in __emit_job_gen12_render_compute() 456 xe_gt_assert(job->q->gt, i <= MAX_JOB_SIZE_DW); in emit_migration_job_gen12() 465 xe_gt_assert(gt, job->q->width <= 1); /* no parallel submission for GSCCS */ in emit_job_gen12_gsc()
|
| H A D | xe_guc_ads.c | 428 xe_gt_assert(gt, ads->bo); in xe_guc_ads_init_post_hwconfig() 435 xe_gt_assert(gt, ads->golden_lrc_size + in xe_guc_ads_init_post_hwconfig() 815 xe_gt_assert(gt, regset_used <= ads->regset_size); in guc_mmio_reg_state_init() 874 xe_gt_assert(gt, ads->bo); in xe_guc_ads_populate_minimal() 898 xe_gt_assert(gt, ads->bo); in xe_guc_ads_populate() 950 xe_gt_assert(gt, gt->default_lrc[class]); in guc_golden_lrc_populate() 962 xe_gt_assert(gt, total_size == ads->golden_lrc_size); in guc_golden_lrc_populate()
|
| H A D | xe_migrate.c | 679 xe_gt_assert(gt, FIELD_FIT(XE2_CCS_SIZE_MASK, num_pages - 1)); in emit_copy_ccs() 687 xe_gt_assert(gt, FIELD_FIT(CCS_SIZE_MASK, num_ccs_blks - 1)); in emit_copy_ccs() 714 xe_gt_assert(gt, !(pitch & 3)); in emit_xy_fast_copy() 715 xe_gt_assert(gt, size / pitch <= S16_MAX); in emit_xy_fast_copy() 716 xe_gt_assert(gt, pitch / 4 <= S16_MAX); in emit_xy_fast_copy() 717 xe_gt_assert(gt, pitch <= U16_MAX); in emit_xy_fast_copy() 743 xe_gt_assert(gt, IS_ALIGNED(size, pitch)); in emit_mem_copy() 744 xe_gt_assert(gt, pitch <= U16_MAX); in emit_mem_copy() 745 xe_gt_assert(gt, pitch); in emit_mem_copy() 746 xe_gt_assert(gt, size); in emit_mem_copy() [all …]
|
| H A D | xe_hw_engine.c | 298 xe_gt_assert(hwe->gt, !(reg.addr & hwe->mmio_base)); in xe_hw_engine_mmio_write32() 318 xe_gt_assert(hwe->gt, !(reg.addr & hwe->mmio_base)); in xe_hw_engine_mmio_read32() 497 xe_gt_assert(gt, info); in get_msix_irq_offset() 515 xe_gt_assert(gt, !hwe->gt); in hw_engine_init_early() 613 xe_gt_assert(gt, id < ARRAY_SIZE(engine_infos) && engine_infos[id].name); in hw_engine_init() 614 xe_gt_assert(gt, gt->info.engine_mask & BIT(id)); in hw_engine_init()
|
| H A D | xe_wa.h | 33 xe_gt_assert(gt__, (gt__)->wa_active.oob_initialized); \
|
| H A D | xe_assert.h | 169 #define xe_gt_assert(gt, condition) xe_gt_assert_msg((gt), condition, "") macro
|
| H A D | xe_execlist.c | 55 xe_gt_assert(hwe->gt, FIELD_FIT(XEHP_SW_CTX_ID, ctx_id)); in __start_lrc() 58 xe_gt_assert(hwe->gt, FIELD_FIT(SW_CTX_ID, ctx_id)); in __start_lrc()
|
| H A D | xe_guc_id_mgr.c | 32 #define idm_assert(idm, cond) xe_gt_assert(idm_to_gt(idm), cond)
|
| H A D | xe_guc_db_mgr.c | 46 #define dbm_assert(_dbm, _cond) xe_gt_assert(dbm_to_gt(_dbm), _cond)
|
| H A D | xe_lrc.c | 208 xe_gt_assert(hwe->gt, count); in set_offsets() 1238 xe_gt_assert(state->hwe->gt, state->buffer); in setup_bo() 2245 xe_gt_assert(gt, (instr & XE_INSTR_CMD_TYPE) == XE_INSTR_GFXPIPE); in xe_lrc_emit_hwe_state_instructions() 2246 xe_gt_assert(gt, num_dw != 0); in xe_lrc_emit_hwe_state_instructions() 2247 xe_gt_assert(gt, is_single_dw ^ (num_dw > 1)); in xe_lrc_emit_hwe_state_instructions()
|
| H A D | xe_gt.c | 764 xe_gt_assert(gt, (gt->user_engines.mask | gt->info.engine_mask) in xe_gt_record_user_engines() 1138 xe_gt_assert(gt, gt_to_xe(gt)->wedged.mode); in xe_gt_declare_wedged()
|
| H A D | xe_guc_engine_activity.c | 308 xe_gt_assert(guc_to_gt(guc), index < engine_activity->num_activity_group); in engine_activity_set_cpu_ts()
|
| H A D | xe_oa.c | 1037 xe_gt_assert(stream->gt, !err || err == -ENOENT); in xe_oa_emit_oa_config() 2650 xe_gt_assert(gt, xe_device_has_mert(gt_to_xe(gt))); in __xe_oa_init_oa_units() 2651 xe_gt_assert(gt, gt == xe_root_mmio_gt(gt_to_xe(gt))); in __xe_oa_init_oa_units() 2656 xe_gt_assert(gt, GRAPHICS_VERx100(gt_to_xe(gt)) >= 1270); in __xe_oa_init_oa_units()
|
| H A D | xe_guc.c | 534 xe_gt_assert(gt, (gt->info.type == XE_GT_TYPE_MAIN) || (gt->info.type == XE_GT_TYPE_MEDIA)); in guc_g2g_start() 1736 xe_gt_assert(guc_to_gt(guc), guc_to_xe(guc)->wedged.mode); in xe_guc_declare_wedged()
|