| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_sriov_pf_migration.c | 21 xe_assert(xe, IS_SRIOV_PF(xe)); in pf_pick_migration() 22 xe_assert(xe, vfid <= xe_sriov_pf_get_totalvfs(xe)); in pf_pick_migration() 47 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_migration_supported() 62 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_migration_disable() 102 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_migration_init() 184 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_migration_save_consume() 253 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_migration_restore_produce() 287 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_migration_read() 321 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_migration_write() 353 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_migration_size() [all …]
|
| H A D | xe_sriov_pf_service.c | 29 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_service_init() 48 xe_assert(xe, IS_SRIOV_PF(xe)); in pf_negotiate_version() 49 xe_assert(xe, base.major); in pf_negotiate_version() 50 xe_assert(xe, base.major <= latest.major); in pf_negotiate_version() 51 xe_assert(xe, (base.major < latest.major) || (base.minor <= latest.minor)); in pf_negotiate_version() 77 xe_assert(xe, base.major == latest.major); in pf_negotiate_version() 90 xe_assert(xe, major || minor); in pf_connect() 200 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_service_print_versions()
|
| H A D | xe_sriov_packet.c | 17 xe_assert(xe, IS_SRIOV_PF(xe)); in pf_migration_mutex() 18 xe_assert(xe, vfid <= xe_sriov_pf_get_totalvfs(xe)); in pf_migration_mutex() 25 xe_assert(xe, IS_SRIOV_PF(xe)); in pf_pick_pending() 26 xe_assert(xe, vfid <= xe_sriov_pf_get_totalvfs(xe)); in pf_pick_pending() 35 xe_assert(xe, IS_SRIOV_PF(xe)); in pf_pick_descriptor() 36 xe_assert(xe, vfid <= xe_sriov_pf_get_totalvfs(xe)); in pf_pick_descriptor() 44 xe_assert(xe, IS_SRIOV_PF(xe)); in pf_pick_trailer() 45 xe_assert(xe, vfid <= xe_sriov_pf_get_totalvfs(xe)); in pf_pick_trailer() 190 xe_assert(data->xe, !data->hdr_remaining); in xe_sriov_packet_init_from_hdr() 389 xe_assert(xe, len == MIGRATION_DESCRIPTOR_DWORDS); in pf_descriptor_init() [all …]
|
| H A D | xe_sriov_pf_helpers.h | 26 xe_assert((xe), (vfid) <= xe_sriov_pf_get_totalvfs(xe)) 36 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_get_totalvfs() 64 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_master_mutex()
|
| H A D | xe_bo.c | 171 xe_assert(xe, mem_type == XE_PL_STOLEN || mem_type_is_vram(mem_type)); in mem_type_to_migrate() 182 xe_assert(xe, resource_is_vram(res)); in res_to_mem_region() 193 xe_assert(xe, *c < ARRAY_SIZE(bo->placements)); in try_add_system() 224 xe_assert(xe, vram_bo_flag & XE_BO_FLAG_VRAM_MASK); in vram_bo_flag_to_tile_id() 225 xe_assert(xe, (vram_bo_flag & (vram_bo_flag - 1)) == 0); in vram_bo_flag_to_tile_id() 235 xe_assert(xe, tile_id < xe->info.tile_count); in bo_vram_flags_to_vram_placement() 253 xe_assert(xe, *c < ARRAY_SIZE(bo->placements)); in add_vram() 256 xe_assert(xe, vram && vram->usable_size); in add_vram() 290 xe_assert(xe, *c < ARRAY_SIZE(bo->placements)); in try_add_stolen() 528 xe_assert(xe, bo->cpu_caching == 0); in xe_ttm_tt_create() [all …]
|
| H A D | xe_guc.c | 65 xe_assert(xe, addr >= xe_wopcm_size(guc_to_xe(guc))); in guc_bo_ggtt_addr() 66 xe_assert(xe, addr < GUC_GGTT_TOP); in guc_bo_ggtt_addr() 67 xe_assert(xe, xe_bo_size(bo) <= GUC_GGTT_TOP - addr); in guc_bo_ggtt_addr() 307 xe_assert(xe, (type == XE_G2G_TYPE_IN) || (type == XE_G2G_TYPE_OUT)); in guc_action_register_g2g_buffer() 308 xe_assert(xe, !(size % SZ_4K)); in guc_action_register_g2g_buffer() 324 xe_assert(xe, (type == XE_G2G_TYPE_IN) || (type == XE_G2G_TYPE_OUT)); in guc_action_deregister_g2g_buffer() 417 xe_assert(xe, xe == gt_to_xe(far_gt)); in guc_g2g_register() 420 xe_assert(xe, g2g_bo); in guc_g2g_register() 423 xe_assert(xe, slot >= 0); in guc_g2g_register() 429 xe_assert(xe, (desc - base + G2G_DESC_SIZE) <= G2G_DESC_AREA_SIZE); in guc_g2g_register() [all …]
|
| H A D | xe_sriov_vf_ccs.c | 257 xe_assert(xe, IS_VF_CCS_READY(xe)); in xe_sriov_vf_ccs_register_context() 279 xe_assert(xe, IS_SRIOV_VF(xe)); in vf_migration_ccs_bb_needed() 339 xe_assert(xe, IS_SRIOV_VF(xe)); in xe_sriov_vf_ccs_init() 419 xe_assert(xe, IS_VF_CCS_READY(xe)); in xe_sriov_vf_ccs_attach_bo() 426 xe_assert(xe, !bb); in xe_sriov_vf_ccs_attach_bo() 450 xe_assert(xe, IS_VF_CCS_READY(xe)); in xe_sriov_vf_ccs_detach_bo()
|
| H A D | xe_sriov_pf_sysfs.c | 246 xe_assert(xe, priority < num_priorities); in xe_sriov_vf_attr_sched_priority_show() 495 xe_assert(xe, IS_SRIOV_PF(xe)); in pf_setup_root() 496 xe_assert(xe, !xe->sriov.pf.sysfs.root); in pf_setup_root() 508 xe_assert(xe, IS_SRIOV_PF(xe)); in pf_setup_tree() 523 xe_assert(xe, !xe->sriov.pf.vfs[n].kobj); in pf_setup_tree() 601 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_sysfs_link_vfs() 602 xe_assert(xe, num_vfs <= totalvfs); in xe_sriov_pf_sysfs_link_vfs() 632 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_sysfs_unlink_vfs() 633 xe_assert(xe, num_vfs <= xe_sriov_pf_get_totalvfs(xe)); in xe_sriov_pf_sysfs_unlink_vfs()
|
| H A D | xe_sriov_pf_provision.c | 30 xe_assert(xe, IS_SRIOV_PF(xe)); in pf_auto_provisioning_mode() 90 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_provision_vfs() 109 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_unprovision_vfs() 136 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_provision_set_mode() 371 xe_assert(xe, prio < GUC_SCHED_PRIORITY_HIGH); in xe_sriov_pf_provision_bulk_apply_priority()
|
| H A D | xe_sriov.c | 81 xe_assert(xe, !xe->sriov.__mode); in xe_sriov_probe_early() 83 xe_assert(xe, xe->sriov.__mode); in xe_sriov_probe_early() 122 xe_assert(xe, !xe->sriov.wq); in xe_sriov_init()
|
| H A D | xe_tlb_inval_job.c | 108 xe_assert(vm->xe, type == XE_EXEC_QUEUE_TLB_INVAL_MEDIA_GT || in xe_tlb_inval_job_create() 212 xe_assert(gt_to_xe(job->q->gt), !xa_load(&job->dep.drm.dependencies, 0)); in xe_tlb_inval_job_alloc_dep() 248 xe_assert(gt_to_xe(job->q->gt), in xe_tlb_inval_job_push() 255 xe_assert(gt_to_xe(job->q->gt), !xa_is_err(ptr)); in xe_tlb_inval_job_push()
|
| H A D | xe_gt_ccs_mode.c | 26 xe_assert(xe, xe_gt_ccs_mode_enabled(gt)); in __xe_gt_apply_ccs_mode() 28 xe_assert(xe, num_engines && num_engines <= num_slices); in __xe_gt_apply_ccs_mode() 29 xe_assert(xe, !(num_slices % num_engines)); in __xe_gt_apply_ccs_mode()
|
| H A D | xe_wa.h | 44 xe_assert(xe__, (xe__)->wa_active.oob_initialized); \ 49 xe_assert(xe__, (xe__)->wa_active.oob_initialized); \
|
| H A D | xe_exec_queue.c | 346 …xe_assert(xe, !vm || (!!(vm->flags & XE_VM_FLAG_GSC) == !!(hwe->engine_id == XE_HW_ENGINE_GSCCS0))… in xe_exec_queue_create() 477 xe_assert(gt_to_xe(q->gt), atomic_read(&q->job_cnt) == 0); in xe_exec_queue_destroy() 763 xe_assert(xe, xe_exec_queue_is_multi_queue_secondary(q)); in xe_exec_queue_group_add() 788 xe_assert(xe, xe_exec_queue_is_multi_queue_secondary(q)); in xe_exec_queue_group_delete() 791 xe_assert(xe, lrc); in xe_exec_queue_group_delete() 1543 xe_assert(vm->xe, !dma_fence_is_container(fence)); in xe_exec_queue_last_fence_set() 1560 xe_assert(vm->xe, type == XE_EXEC_QUEUE_TLB_INVAL_MEDIA_GT || in xe_exec_queue_tlb_inval_last_fence_put() 1577 xe_assert(q->vm->xe, type == XE_EXEC_QUEUE_TLB_INVAL_MEDIA_GT || in xe_exec_queue_tlb_inval_last_fence_put_unlocked() 1601 xe_assert(vm->xe, type == XE_EXEC_QUEUE_TLB_INVAL_MEDIA_GT || in xe_exec_queue_tlb_inval_last_fence_get() 1603 xe_assert(vm->xe, q->flags & (EXEC_QUEUE_FLAG_VM | in xe_exec_queue_tlb_inval_last_fence_get() [all …]
|
| H A D | xe_mmio.c | 397 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_mmio_init_vf_view() 398 xe_assert(xe, vfid); in xe_mmio_init_vf_view() 399 xe_assert(xe, !base->sriov_vf_gt); in xe_mmio_init_vf_view() 400 xe_assert(xe, base->regs_size > offset); in xe_mmio_init_vf_view()
|
| H A D | xe_shrinker.c | 90 xe_assert(xe, !IS_ERR(ttm_bo)); in __xe_shrinker_walk() 270 xe_assert(shrinker->xe, !shrinker->shrinkable_pages); in xe_shrinker_fini() 271 xe_assert(shrinker->xe, !shrinker->purgeable_pages); in xe_shrinker_fini()
|
| H A D | xe_migrate.c | 156 xe_assert(xe, IS_ALIGNED(xe_vram_region_usable_size(vram), SZ_2M)); in xe_migrate_program_identity() 183 xe_assert(xe, pos == vram_limit); in xe_migrate_program_identity() 321 xe_assert(xe, actual_phy_size <= (MAX_NUM_PTE - IDENTITY_OFFSET) * SZ_1G); in xe_migrate_prepare_vm() 333 xe_assert(xe, actual_phy_size <= (MAX_NUM_PTE - IDENTITY_OFFSET - in xe_migrate_prepare_vm() 642 xe_assert(xe, (va & (SZ_64K - 1)) == in emit_pte() 945 xe_assert(xe, type_device); in xe_migrate_copy() 951 xe_assert(xe, IS_ALIGNED(ccs_it.start, PAGE_SIZE)); in xe_migrate_copy() 1143 xe_assert(xe, IS_ALIGNED(ccs_it.start, PAGE_SIZE)); in xe_migrate_ccs_rw_copy() 1189 xe_assert(xe, IS_ALIGNED(ccs_it.start, PAGE_SIZE)); in xe_migrate_ccs_rw_copy() 1205 xe_assert(xe, (batch_size_allocated == bb->len)); in xe_migrate_ccs_rw_copy() [all …]
|
| H A D | xe_vm.c | 126 xe_assert(vm->xe, vf_migration); in wait_for_existing_preempt_fences() 164 xe_assert(vm->xe, link != list); in arm_preempt_fences() 227 xe_assert(vm->xe, xe_vm_in_preempt_fence_mode(vm)); in xe_vm_add_compute_exec_queue() 464 xe_assert(vm->xe, xe_vm_in_preempt_fence_mode(vm)); in preempt_rebind_work_func() 706 xe_assert(vm->xe, vma->tile_present); in xe_vm_rebind() 753 xe_assert(vm->xe, xe_vm_in_fault_mode(vm)); in xe_vma_rebind() 843 xe_assert(vm->xe, xe_vm_in_fault_mode(vm)); in xe_vm_range_rebind() 844 xe_assert(vm->xe, xe_vma_is_cpu_addr_mirror(vma)); in xe_vm_range_rebind() 925 xe_assert(vm->xe, xe_vm_in_fault_mode(vm)); in xe_vm_range_unbind() 1003 xe_assert(vm->xe, start < end); in xe_vma_create() [all …]
|
| H A D | xe_pxp.c | 492 xe_assert(pxp->xe, type == DRM_XE_PXP_TYPE_HWDRM); in xe_pxp_exec_queue_set_type() 533 xe_assert(pxp->xe, type == DRM_XE_PXP_TYPE_HWDRM); in pxp_start() 767 xe_assert(pxp->xe, !bo->pxp_key_instance); in xe_pxp_key_assign() 801 xe_assert(pxp->xe, bo->pxp_key_instance); in xe_pxp_bo_key_check() 930 xe_assert(pxp->xe, pxp->status == XE_PXP_SUSPENDED); in xe_pxp_pm_resume()
|
| H A D | xe_vm.h | 193 xe_assert(xe_vma_vm(vma)->xe, xe_vma_is_userptr(vma)); in to_userptr_vma() 245 xe_assert(vm->xe, xe_vm_in_preempt_fence_mode(vm)); in xe_vm_queue_rebind_worker() 380 xe_assert(vm->xe, !!exec ^ !!vm->validation._exec); in xe_vm_set_validation_exec()
|
| H A D | xe_vm_madvise.c | 95 xe_assert(vm->xe, op->type == DRM_XE_MEM_RANGE_ATTR_PREFERRED_LOC); in madvise_preferred_mem_loc() 129 xe_assert(vm->xe, op->type == DRM_XE_MEM_RANGE_ATTR_ATOMIC); in madvise_atomic() 130 xe_assert(vm->xe, op->atomic.val <= DRM_XE_ATOMIC_CPU); in madvise_atomic() 170 xe_assert(vm->xe, op->type == DRM_XE_MEM_RANGE_ATTR_PAT); in madvise_pat_index()
|
| H A D | xe_sriov.h | 25 xe_assert(xe, xe->sriov.__mode); in xe_device_sriov_mode()
|
| H A D | xe_drm_client.c | 158 xe_assert(xe, !kref_read(&bo->ttm.base.refcount)); in xe_drm_client_remove_bo() 246 xe_assert(xef->xe, !list_empty(&bo->client_link)); in show_meminfo()
|
| H A D | xe_gt_topology.c | 114 xe_assert(xe, find_last_bit(pattern, XE_MAX_L3_BANK_MASK_BITS) < patternbits || in gen_l3_mask_from_pattern() 116 xe_assert(xe, !mask || patternbits * (__fls(mask) + 1) <= XE_MAX_L3_BANK_MASK_BITS); in gen_l3_mask_from_pattern()
|
| /linux/Documentation/gpu/xe/ |
| H A D | xe_debugging.rst | 7 .. kernel-doc:: drivers/gpu/drm/xe/xe_assert.h
|