Lines Matching full:remap
2168 vma = gpuva_to_vma(op->remap.unmap->va); in print_op()
2169 vm_dbg(&xe->drm, "REMAP:UNMAP: addr=0x%016llx, range=0x%016llx, keep=%d", in print_op()
2171 op->remap.unmap->keep ? 1 : 0); in print_op()
2172 if (op->remap.prev) in print_op()
2174 "REMAP:PREV: addr=0x%016llx, range=0x%016llx", in print_op()
2175 (ULL)op->remap.prev->va.addr, in print_op()
2176 (ULL)op->remap.prev->va.range); in print_op()
2177 if (op->remap.next) in print_op()
2179 "REMAP:NEXT: addr=0x%016llx, range=0x%016llx", in print_op()
2180 (ULL)op->remap.next->va.addr, in print_op()
2181 (ULL)op->remap.next->va.range); in print_op()
2386 gpuva_to_vma(op->base.remap.unmap->va)->tile_present; in xe_vma_op_commit()
2388 prep_vma_destroy(vm, gpuva_to_vma(op->base.remap.unmap->va), in xe_vma_op_commit()
2392 if (op->remap.prev) { in xe_vma_op_commit()
2393 err |= xe_vm_insert_vma(vm, op->remap.prev); in xe_vma_op_commit()
2396 if (!err && op->remap.skip_prev) { in xe_vma_op_commit()
2397 op->remap.prev->tile_present = in xe_vma_op_commit()
2399 op->remap.prev = NULL; in xe_vma_op_commit()
2402 if (op->remap.next) { in xe_vma_op_commit()
2403 err |= xe_vm_insert_vma(vm, op->remap.next); in xe_vma_op_commit()
2406 if (!err && op->remap.skip_next) { in xe_vma_op_commit()
2407 op->remap.next->tile_present = in xe_vma_op_commit()
2409 op->remap.next = NULL; in xe_vma_op_commit()
2415 op->base.remap.unmap->va->va.addr = op->remap.start; in xe_vma_op_commit()
2416 op->base.remap.unmap->va->va.range = op->remap.range; in xe_vma_op_commit()
2484 gpuva_to_vma(op->base.remap.unmap->va); in vm_bind_ioctl_ops_parse()
2488 if (op->base.remap.prev) in vm_bind_ioctl_ops_parse()
2489 start = op->base.remap.prev->va.addr + in vm_bind_ioctl_ops_parse()
2490 op->base.remap.prev->va.range; in vm_bind_ioctl_ops_parse()
2491 if (op->base.remap.next) in vm_bind_ioctl_ops_parse()
2492 end = op->base.remap.next->va.addr; in vm_bind_ioctl_ops_parse()
2498 op->remap.start = xe_vma_start(old); in vm_bind_ioctl_ops_parse()
2499 op->remap.range = xe_vma_size(old); in vm_bind_ioctl_ops_parse()
2501 flags |= op->base.remap.unmap->va->flags & in vm_bind_ioctl_ops_parse()
2504 flags |= op->base.remap.unmap->va->flags & in vm_bind_ioctl_ops_parse()
2507 flags |= op->base.remap.unmap->va->flags & in vm_bind_ioctl_ops_parse()
2513 if (op->base.remap.prev) { in vm_bind_ioctl_ops_parse()
2514 vma = new_vma(vm, op->base.remap.prev, in vm_bind_ioctl_ops_parse()
2519 op->remap.prev = vma; in vm_bind_ioctl_ops_parse()
2525 op->remap.skip_prev = skip || in vm_bind_ioctl_ops_parse()
2529 if (op->remap.skip_prev) { in vm_bind_ioctl_ops_parse()
2531 op->remap.range -= in vm_bind_ioctl_ops_parse()
2534 op->remap.start = xe_vma_end(vma); in vm_bind_ioctl_ops_parse()
2535 vm_dbg(&xe->drm, "REMAP:SKIP_PREV: addr=0x%016llx, range=0x%016llx", in vm_bind_ioctl_ops_parse()
2536 (ULL)op->remap.start, in vm_bind_ioctl_ops_parse()
2537 (ULL)op->remap.range); in vm_bind_ioctl_ops_parse()
2543 if (op->base.remap.next) { in vm_bind_ioctl_ops_parse()
2544 vma = new_vma(vm, op->base.remap.next, in vm_bind_ioctl_ops_parse()
2549 op->remap.next = vma; in vm_bind_ioctl_ops_parse()
2555 op->remap.skip_next = skip || in vm_bind_ioctl_ops_parse()
2559 if (op->remap.skip_next) { in vm_bind_ioctl_ops_parse()
2561 op->remap.range -= in vm_bind_ioctl_ops_parse()
2564 vm_dbg(&xe->drm, "REMAP:SKIP_NEXT: addr=0x%016llx, range=0x%016llx", in vm_bind_ioctl_ops_parse()
2565 (ULL)op->remap.start, in vm_bind_ioctl_ops_parse()
2566 (ULL)op->remap.range); in vm_bind_ioctl_ops_parse()
2638 struct xe_vma *vma = gpuva_to_vma(op->base.remap.unmap->va); in xe_vma_op_unwind()
2640 if (op->remap.prev) { in xe_vma_op_unwind()
2641 prep_vma_destroy(vm, op->remap.prev, prev_post_commit); in xe_vma_op_unwind()
2642 xe_vma_destroy_unlocked(op->remap.prev); in xe_vma_op_unwind()
2644 if (op->remap.next) { in xe_vma_op_unwind()
2645 prep_vma_destroy(vm, op->remap.next, next_post_commit); in xe_vma_op_unwind()
2646 xe_vma_destroy_unlocked(op->remap.next); in xe_vma_op_unwind()
2734 err = check_ufence(gpuva_to_vma(op->base.remap.unmap->va)); in op_lock_and_prep()
2739 gpuva_to_vma(op->base.remap.unmap->va), in op_lock_and_prep()
2741 if (!err && op->remap.prev) in op_lock_and_prep()
2742 err = vma_lock_and_validate(exec, op->remap.prev, true); in op_lock_and_prep()
2743 if (!err && op->remap.next) in op_lock_and_prep()
2744 err = vma_lock_and_validate(exec, op->remap.next, true); in op_lock_and_prep()
2810 trace_xe_vma_unbind(gpuva_to_vma(op->base.remap.unmap->va)); in op_trace()
2811 if (op->remap.prev) in op_trace()
2812 trace_xe_vma_bind(op->remap.prev); in op_trace()
2813 if (op->remap.next) in op_trace()
2814 trace_xe_vma_bind(op->remap.next); in op_trace()
2965 if (op->remap.prev) in op_add_ufence()
2966 vma_add_ufence(op->remap.prev, ufence); in op_add_ufence()
2967 if (op->remap.next) in op_add_ufence()
2968 vma_add_ufence(op->remap.next, ufence); in op_add_ufence()
2996 xe_vma_destroy(gpuva_to_vma(op->base.remap.unmap->va), in vm_bind_ioctl_ops_fini()