| /linux/arch/powerpc/platforms/pseries/ |
| H A D | plpks.c | 176 virt_to_phys(password), maxpwsize); in plpks_gen_password() 296 rc = plpar_hcall(H_PKS_GET_CONFIG, retbuf, virt_to_phys(config), size); in _plpks_get_config() 576 virt_to_phys(auth), virt_to_phys(label), in plpks_confirm_object_flushed() 665 virt_to_phys(auth), virt_to_phys(label), in plpks_signed_update_var() 667 virt_to_phys(var->data), var->datalen, in plpks_signed_update_var() 744 rc = plpar_hcall(H_PKS_WRITE_OBJECT, retbuf, virt_to_phys(auth), in plpks_write_var() 745 virt_to_phys(label), label->size, var.policy, in plpks_write_var() 746 virt_to_phys(var.data), var.datalen); in plpks_write_var() 803 rc = plpar_hcall(H_PKS_REMOVE_OBJECT, retbuf, virt_to_phys(auth), in plpks_remove_var() 804 virt_to_phys(label), label->size); in plpks_remove_var() [all …]
|
| H A D | htmdump.c | 103 htmtype, H_HTM_OP_DUMP_DATA, virt_to_phys(htm_buf), in htmdump_read() 241 htmtype, H_HTM_OP_STATUS, virt_to_phys(htm_status_buf), in htmstatus_read() 287 htmtype, H_HTM_OP_DUMP_SYSPROC_CONF, virt_to_phys(htm_info_buf), in htminfo_read() 322 htmtype, H_HTM_OP_CAPABILITIES, virt_to_phys(htm_caps_buf), in htmcaps_read()
|
| /linux/arch/arm/kernel/ |
| H A D | suspend.c | 87 *save_ptr = virt_to_phys(ptr); in __cpu_suspend_save() 90 *ptr++ = virt_to_phys(idmap_pgd); in __cpu_suspend_save() 92 *ptr++ = virt_to_phys(cpu_do_resume); in __cpu_suspend_save() 111 outer_clean_range(virt_to_phys(save_ptr), in __cpu_suspend_save() 112 virt_to_phys(save_ptr) + sizeof(*save_ptr)); in __cpu_suspend_save() 126 sleep_save_sp.save_ptr_stash_phys = virt_to_phys(ctx_ptr); in cpu_suspend_alloc_sp()
|
| /linux/arch/m68k/include/asm/ |
| H A D | virtconvert.h | 19 #define virt_to_phys virt_to_phys macro 20 static inline unsigned long virt_to_phys(void *address) in virt_to_phys() function 36 #define virt_to_bus virt_to_phys
|
| /linux/drivers/virt/acrn/ |
| H A D | hsm.c | 56 ret = hcall_get_cpu_state(cmd, virt_to_phys(pm_info)); in pmcmd_ioctl() 71 ret = hcall_get_cpu_state(cmd, virt_to_phys(px_data)); in pmcmd_ioctl() 86 ret = hcall_get_cpu_state(cmd, virt_to_phys(cx_data)); in pmcmd_ioctl() 211 ret = hcall_set_vcpu_regs(vm->vmid, virt_to_phys(cpu_regs)); in acrn_dev_ioctl() 238 ret = hcall_assign_mmiodev(vm->vmid, virt_to_phys(mmiodev)); in acrn_dev_ioctl() 250 ret = hcall_deassign_mmiodev(vm->vmid, virt_to_phys(mmiodev)); in acrn_dev_ioctl() 262 ret = hcall_assign_pcidev(vm->vmid, virt_to_phys(pcidev)); in acrn_dev_ioctl() 274 ret = hcall_deassign_pcidev(vm->vmid, virt_to_phys(pcidev)); in acrn_dev_ioctl() 286 ret = hcall_create_vdev(vm->vmid, virt_to_phys(vdev)); in acrn_dev_ioctl() 297 ret = hcall_destroy_vdev(vm->vmid, virt_to_phys(vdev)); in acrn_dev_ioctl() [all …]
|
| H A D | mm.c | 30 regions->regions_gpa = virt_to_phys(region); in modify_region() 32 ret = hcall_set_memory_regions(virt_to_phys(regions)); in modify_region() 298 regions_info->regions_gpa = virt_to_phys(vm_region); in acrn_vm_ram_map() 319 ret = hcall_set_memory_regions(virt_to_phys(regions_info)); in acrn_vm_ram_map()
|
| /linux/arch/sh/kernel/ |
| H A D | setup.c | 295 code_resource.start = virt_to_phys(_text); in setup_arch() 296 code_resource.end = virt_to_phys(_etext)-1; in setup_arch() 297 data_resource.start = virt_to_phys(_etext); in setup_arch() 298 data_resource.end = virt_to_phys(_edata)-1; in setup_arch() 299 bss_resource.start = virt_to_phys(__bss_start); in setup_arch() 300 bss_resource.end = virt_to_phys(__bss_stop)-1; in setup_arch()
|
| /linux/tools/include/linux/ |
| H A D | mm.h | 29 #define virt_to_phys virt_to_phys macro 30 static inline phys_addr_t virt_to_phys(volatile void *address) in virt_to_phys() function
|
| /linux/arch/m68k/emu/ |
| H A D | natfeat.c | 51 return nf_get_id_phys(virt_to_phys(name_copy)); in nf_get_id() 62 nf_call(nf_get_id("NF_STDERR"), virt_to_phys(buf)); in nfprint() 87 nf_call(id, virt_to_phys(buf), 256); in nf_init()
|
| H A D | nfeth.c | 118 nf_call(nfEtherID + XIF_READBLOCK, priv->ethX, virt_to_phys(skb->data), in recv_packet() 160 nf_call(nfEtherID + XIF_WRITEBLOCK, priv->ethX, virt_to_phys(data), in nfeth_xmit() 192 if (!nf_call(nfEtherID + XIF_GET_MAC, unit, virt_to_phys(mac), in nfeth_probe() 215 virt_to_phys(host_ip), sizeof(host_ip)); in nfeth_probe() 217 virt_to_phys(local_ip), sizeof(local_ip)); in nfeth_probe()
|
| /linux/arch/s390/include/asm/ |
| H A D | appldata.h | 58 parm_list->product_id_addr = virt_to_phys(id); in appldata_asm() 59 parm_list->buffer_addr = virt_to_phys(buffer); in appldata_asm() 64 : "d" (virt_to_phys(parm_list)), "m" (*parm_list), "m" (*id) in appldata_asm()
|
| /linux/arch/hexagon/include/asm/ |
| H A D | io.h | 22 static inline unsigned long virt_to_phys(volatile void *address) in virt_to_phys() function 129 #define virt_to_phys virt_to_phys macro
|
| /linux/arch/riscv/mm/ |
| H A D | pmem.c | 16 noncoherent_cache_ops.wback(virt_to_phys(addr), size); in arch_wb_cache_pmem() 28 noncoherent_cache_ops.inv(virt_to_phys(addr), size); in arch_invalidate_pmem()
|
| /linux/lib/ |
| H A D | test_debug_virtual.c | 28 pa = virt_to_phys(va); in test_debug_virtual_init() 36 pa = virt_to_phys(foo); in test_debug_virtual_init()
|
| /linux/arch/s390/kvm/ |
| H A D | pci.c | 76 zpci_aipb->aipb.faisb = virt_to_phys(aift->sbv->vector); in zpci_setup_aipb() 77 zpci_aipb->aipb.gait = virt_to_phys(aift->gait); in zpci_setup_aipb() 167 fib.fmt0.aibv = virt_to_phys(zdev->aibv->vector); in kvm_zpci_set_airq() 169 fib.fmt0.aisb = virt_to_phys(aift->sbv->vector + (zdev->aisb / 64) * 8); in kvm_zpci_set_airq() 298 gaite->gisa = (u32)virt_to_phys(&kvm->arch.sie_page2->gisa); in kvm_s390_pci_aif_enable() 305 gaite->aisb = virt_to_phys(page_address(aisb_page) + (fib->fmt0.aisb & in kvm_s390_pci_aif_enable() 312 fib->fmt0.aisb = virt_to_phys(aift->sbv->vector + (zdev->aisb / 64) * 8); in kvm_s390_pci_aif_enable() 479 zdev->gisa = (u32)virt_to_phys(&kvm->arch.sie_page2->gisa); in kvm_s390_pci_register_kvm()
|
| /linux/arch/s390/mm/ |
| H A D | maccess.c | 137 lc = virt_to_phys(lowcore_ptr[cpu]); in get_swapped_owner() 176 ptr = (void *)(addr - virt_to_phys(lowcore_ptr[cpu])); in xlate_dev_mem_ptr() 192 if (addr != virt_to_phys(ptr)) in unxlate_dev_mem_ptr()
|
| /linux/drivers/virt/ |
| H A D | fsl_hypervisor.c | 291 virt_to_phys(sg_list), num_pages); in ioctl_memcpy() 376 virt_to_phys(path), in ioctl_dtprop() 377 virt_to_phys(propname), in ioctl_dtprop() 378 virt_to_phys(propval), in ioctl_dtprop() 382 virt_to_phys(path), in ioctl_dtprop() 383 virt_to_phys(propname), in ioctl_dtprop() 384 virt_to_phys(propval), in ioctl_dtprop()
|
| /linux/arch/arm64/include/asm/ |
| H A D | memory.h | 366 #define virt_to_phys virt_to_phys macro 367 static inline phys_addr_t virt_to_phys(const volatile void *x) in virt_to_phys() function 383 return __phys_to_pfn(virt_to_phys(kaddr)); in virt_to_pfn()
|
| /linux/arch/s390/kernel/diag/ |
| H A D | diag.c | 159 diag_amode31_ops.diag0c(virt_to_phys(data)); in diag0c() 182 rx = virt_to_phys((void *)rx); in diag14() 307 return diag_amode31_ops.diag26c(virt_to_phys(req), virt_to_phys(resp), subcode); in diag26c()
|
| /linux/drivers/iommu/ |
| H A D | iommu-pages.c | 168 if (WARN_ON(dma != virt_to_phys(virt))) { in iommu_pages_start_incoherent() 227 dma_unmap_single(dma_dev, virt_to_phys(folio_address(folio)), in iommu_pages_stop_incoherent_list() 246 dma_unmap_single(dma_dev, virt_to_phys(virt), in iommu_pages_free_incoherent()
|
| /linux/arch/x86/include/asm/ |
| H A D | io.h | 129 static inline phys_addr_t virt_to_phys(volatile void *address) in virt_to_phys() function 133 #define virt_to_phys virt_to_phys macro 161 return (unsigned int)virt_to_phys(address); in isa_virt_to_bus()
|
| /linux/drivers/staging/media/ipu3/ |
| H A D | ipu3-mmu.c | 203 pteval = IPU3_ADDR2PTE(virt_to_phys(new_l2pt)); in imgu_mmu_get_l2pt() 450 pteval = IPU3_ADDR2PTE(virt_to_phys(mmu->dummy_page)); in imgu_mmu_init() 460 pteval = IPU3_ADDR2PTE(virt_to_phys(mmu->dummy_l2pt)); in imgu_mmu_init() 476 pteval = IPU3_ADDR2PTE(virt_to_phys(mmu->l1pt)); in imgu_mmu_init() 532 pteval = IPU3_ADDR2PTE(virt_to_phys(mmu->l1pt)); in imgu_mmu_resume()
|
| /linux/arch/nios2/include/asm/ |
| H A D | page.h | 85 # define virt_to_page(vaddr) pfn_to_page(PFN_DOWN(virt_to_phys(vaddr))) 86 # define virt_addr_valid(vaddr) pfn_valid(PFN_DOWN(virt_to_phys(vaddr)))
|
| /linux/arch/arm/include/asm/ |
| H A D | memory.h | 312 #define virt_to_phys virt_to_phys macro 313 static inline phys_addr_t virt_to_phys(const volatile void *x) in virt_to_phys() function
|
| /linux/drivers/hv/ |
| H A D | hv.c | 86 virt_to_phys(aligned_msg), 0); in hv_post_message() 297 simp.base_simp_gpa = virt_to_phys(hv_cpu->hyp_synic_message_page) in hv_hyp_synic_enable_regs() 316 siefp.base_siefp_gpa = virt_to_phys(hv_cpu->hyp_synic_event_page) in hv_hyp_synic_enable_regs() 355 simp.base_simp_gpa = virt_to_phys(hv_cpu->para_synic_message_page) in hv_para_synic_enable_regs() 362 siefp.base_siefp_gpa = virt_to_phys(hv_cpu->para_synic_event_page) in hv_para_synic_enable_regs()
|