Home
last modified time | relevance | path

Searched refs:xe_assert (Results 1 – 25 of 49) sorted by relevance

12

/linux/drivers/gpu/drm/xe/
H A Dxe_sriov_pf_service.c29 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_service_init()
48 xe_assert(xe, IS_SRIOV_PF(xe)); in pf_negotiate_version()
49 xe_assert(xe, base.major); in pf_negotiate_version()
50 xe_assert(xe, base.major <= latest.major); in pf_negotiate_version()
51 xe_assert(xe, (base.major < latest.major) || (base.minor <= latest.minor)); in pf_negotiate_version()
77 xe_assert(xe, base.major == latest.major); in pf_negotiate_version()
90 xe_assert(xe, major || minor); in pf_connect()
200 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_service_print_versions()
H A Dxe_sriov_pf_helpers.h9 #include "xe_assert.h"
26 xe_assert((xe), (vfid) <= xe_sriov_pf_get_totalvfs(xe))
36 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_get_totalvfs()
42 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_master_mutex()
H A Dxe_hmm.c44 xe_assert(xe, hmm_pfn & HMM_PFN_VALID); in xe_alloc_sg()
111 xe_assert(xe, !is_device_private_page(page)); in xe_build_sg()
118 xe_assert(xe, i >= npages); in xe_build_sg()
124 xe_assert(xe, i < npages); in xe_build_sg()
143 xe_assert(vm->xe, !userptr->mapped); in xe_hmm_userptr_set_mapped()
185 xe_assert(xe_vma_vm(&uvma->vma)->xe, userptr->sg); in xe_hmm_userptr_free_sg()
H A Dxe_pci_sriov.c76 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_pci_pf_get_vf_dev()
153 xe_assert(xe, IS_SRIOV_PF(xe)); in pf_enable_vfs()
154 xe_assert(xe, num_vfs > 0); in pf_enable_vfs()
155 xe_assert(xe, num_vfs <= total_vfs); in pf_enable_vfs()
211 xe_assert(xe, IS_SRIOV_PF(xe)); in pf_disable_vfs()
H A Dxe_sriov.c12 #include "xe_assert.h"
80 xe_assert(xe, !xe->sriov.__mode); in xe_sriov_probe_early()
82 xe_assert(xe, xe->sriov.__mode); in xe_sriov_probe_early()
121 xe_assert(xe, !xe->sriov.wq); in xe_sriov_init()
H A Dxe_hw_engine_group.c8 #include "xe_assert.h"
127 xe_assert(xe, group); in xe_hw_engine_group_add_exec_queue()
128 xe_assert(xe, !(q->flags & EXEC_QUEUE_FLAG_VM)); in xe_hw_engine_group_add_exec_queue()
129 xe_assert(xe, q->vm); in xe_hw_engine_group_add_exec_queue()
167 xe_assert(xe, group); in xe_hw_engine_group_del_exec_queue()
168 xe_assert(xe, q->vm); in xe_hw_engine_group_del_exec_queue()
H A Dxe_guc.c61 xe_assert(xe, addr >= xe_wopcm_size(guc_to_xe(guc))); in guc_bo_ggtt_addr()
62 xe_assert(xe, addr < GUC_GGTT_TOP); in guc_bo_ggtt_addr()
63 xe_assert(xe, xe_bo_size(bo) <= GUC_GGTT_TOP - addr); in guc_bo_ggtt_addr()
301 xe_assert(xe, (type == XE_G2G_TYPE_IN) || (type == XE_G2G_TYPE_OUT)); in guc_action_register_g2g_buffer()
302 xe_assert(xe, !(size % SZ_4K)); in guc_action_register_g2g_buffer()
318 xe_assert(xe, (type == XE_G2G_TYPE_IN) || (type == XE_G2G_TYPE_OUT)); in guc_action_deregister_g2g_buffer()
411 xe_assert(xe, xe == gt_to_xe(far_gt)); in guc_g2g_register()
414 xe_assert(xe, g2g_bo); in guc_g2g_register()
417 xe_assert(xe, slot >= 0); in guc_g2g_register()
423 xe_assert(x in guc_g2g_register()
[all...]
H A Dxe_gt_sriov_vf_debugfs.c58 xe_assert(xe, IS_SRIOV_VF(xe)); in xe_gt_sriov_vf_debugfs_register()
59 xe_assert(xe, root->d_inode->i_private == gt); in xe_gt_sriov_vf_debugfs_register()
H A Dxe_bo.c163 xe_assert(xe, mem_type == XE_PL_STOLEN || mem_type_is_vram(mem_type)); in mem_type_to_migrate()
174 xe_assert(xe, resource_is_vram(res)); in res_to_mem_region()
185 xe_assert(xe, *c < ARRAY_SIZE(bo->placements)); in try_add_system()
220 xe_assert(xe, *c < ARRAY_SIZE(bo->placements)); in add_vram()
223 xe_assert(xe, vram && vram->usable_size); in add_vram()
254 xe_assert(xe, *c < ARRAY_SIZE(bo->placements)); in try_add_stolen()
493 xe_assert(xe, bo->cpu_caching == 0); in xe_ttm_tt_create()
695 xe_assert(xe, attach); in xe_bo_move_dmabuf()
696 xe_assert(xe, ttm_bo->ttm); in xe_bo_move_dmabuf()
914 xe_assert(x in xe_bo_move()
[all...]
H A Dxe_sriov_pf.c10 #include "xe_assert.h"
61 xe_assert(xe, totalvfs <= U16_MAX); in xe_sriov_pf_readiness()
90 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_init_early()
143 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_print_vfs_summary()
H A Dxe_wa.h9 #include "xe_assert.h"
44 xe_assert(xe__, (xe__)->wa_active.oob_initialized); \
49 xe_assert(xe__, (xe__)->wa_active.oob_initialized); \
H A Dxe_gt_ccs_mode.c9 #include "xe_assert.h"
25 xe_assert(xe, xe_gt_ccs_mode_enabled(gt)); in __xe_gt_apply_ccs_mode()
27 xe_assert(xe, num_engines && num_engines <= num_slices); in __xe_gt_apply_ccs_mode()
28 xe_assert(xe, !(num_slices % num_engines)); in __xe_gt_apply_ccs_mode()
H A Dxe_shrinker.c90 xe_assert(xe, !IS_ERR(ttm_bo)); in __xe_shrinker_walk()
270 xe_assert(shrinker->xe, !shrinker->shrinkable_pages); in xe_shrinker_fini()
271 xe_assert(shrinker->xe, !shrinker->purgeable_pages); in xe_shrinker_fini()
H A Dxe_vm.c25 #include "xe_assert.h"
76 xe_assert(xe, xe_vma_is_userptr(vma)); in xe_vma_userptr_pin_pages()
172 xe_assert(vm->xe, link != list); in arm_preempt_fences()
234 xe_assert(vm->xe, xe_vm_in_preempt_fence_mode(vm)); in xe_vm_add_compute_exec_queue()
493 xe_assert(vm->xe, xe_vm_in_preempt_fence_mode(vm)); in preempt_rebind_work_func()
639 xe_assert(vm->xe, xe_vma_is_userptr(vma)); in vma_userptr_invalidate()
696 xe_assert(vm->xe, !xe_vm_in_fault_mode(vm)); in xe_vm_userptr_pin()
701 xe_assert(vm->xe, list_empty(&vm->userptr.repin_list)); in xe_vm_userptr_pin()
903 xe_assert(vm->xe, vma->tile_present); in xe_vm_rebind()
950 xe_assert(v in xe_vma_rebind()
[all...]
H A Dxe_pxp.c500 xe_assert(pxp->xe, type == DRM_XE_PXP_TYPE_HWDRM); in xe_pxp_exec_queue_set_type()
541 xe_assert(pxp->xe, type == DRM_XE_PXP_TYPE_HWDRM); in pxp_start()
771 xe_assert(pxp->xe, !bo->pxp_key_instance); in xe_pxp_key_assign()
805 xe_assert(pxp->xe, bo->pxp_key_instance); in xe_pxp_bo_key_check()
940 xe_assert(pxp->xe, pxp->status == XE_PXP_SUSPENDED); in xe_pxp_pm_resume()
H A Dxe_sriov.h9 #include "xe_assert.h"
24 xe_assert(xe, xe->sriov.__mode); in xe_device_sriov_mode()
H A Dxe_gt_topology.c13 #include "xe_assert.h"
113 xe_assert(xe, find_last_bit(pattern, XE_MAX_L3_BANK_MASK_BITS) < patternbits || in gen_l3_mask_from_pattern()
115 xe_assert(xe, !mask || patternbits * (__fls(mask) + 1) <= XE_MAX_L3_BANK_MASK_BITS); in gen_l3_mask_from_pattern()
H A Dxe_drm_client.c13 #include "xe_assert.h"
158 xe_assert(xe, !kref_read(&bo->ttm.base.refcount)); in xe_drm_client_remove_bo()
246 xe_assert(xef->xe, !list_empty(&bo->client_link)); in show_meminfo()
H A Dxe_pat.c13 #include "xe_assert.h"
447 xe_assert(xe, !xe->pat.ops || xe->pat.ops->dump); in xe_pat_init_early()
448 xe_assert(xe, !xe->pat.ops || xe->pat.ops->program_graphics); in xe_pat_init_early()
449 xe_assert(xe, !xe->pat.ops || MEDIA_VER(xe) < 13 || xe->pat.ops->program_media); in xe_pat_init_early()
H A Dxe_vm.h9 #include "xe_assert.h"
182 xe_assert(xe_vma_vm(vma)->xe, xe_vma_is_userptr(vma)); in to_userptr_vma()
240 xe_assert(vm->xe, xe_vm_in_preempt_fence_mode(vm)); in xe_vm_queue_rebind_worker()
H A Dxe_uc_fw.c312 xe_assert(xe, !(size % 4)); in xe_uc_fw_copy_rsa()
313 xe_assert(xe, xe_uc_fw_is_available(uc_fw)); in xe_uc_fw_copy_rsa()
490 xe_assert(xe, manifest_entry); in parse_cpd_header()
542 xe_assert(xe, xe->info.platform != XE_DG2); in parse_cpd_header()
898 xe_assert(xe, !xe_uc_fw_is_loaded(uc_fw)); in xe_uc_fw_upload()
H A Dxe_svm.c596 xe_assert(vm->xe, xe_vm_is_closed(vm)); in xe_svm_close()
608 xe_assert(vm->xe, xe_vm_is_closed(vm)); in xe_svm_fini()
778 xe_assert(vm->xe, IS_DGFX(vm->xe)); in xe_svm_range_needs_migrate_to_vram()
830 xe_assert(vm->xe, xe_vma_is_cpu_addr_mirror(vma)); in xe_svm_handle_pagefault()
1016 xe_assert(tile_to_xe(tile), range->base.flags.migrate_devmem); in xe_svm_alloc_vram()
H A Dxe_mocs.c652 xe_assert(xe, info->unused_entries_index != 0); in get_mocs_settings()
654 xe_assert(xe, info->ops && info->ops->dump); in get_mocs_settings()
655 xe_assert(xe, info->table_size <= info->num_mocs_regs); in get_mocs_settings()
H A Dxe_sriov_vf.c8 #include "xe_assert.h"
300 xe_assert(xe, IS_SRIOV_VF(xe)); in xe_sriov_vf_start_migration_recovery()
/linux/Documentation/gpu/xe/
H A Dxe_debugging.rst7 .. kernel-doc:: drivers/gpu/drm/xe/xe_assert.h

12