Lines Matching refs:gva
44 target_ulong gva; member
89 index = gpt_entry(pt->gva, level, pae); in get_pt_entry()
163 return (pt->pte[2] & PAE_PTE_SUPER_PAGE_MASK) | (pt->gva & 0x3fffffff); in large_page_gpa()
170 return (pt->pte[1] & PAE_PTE_LARGE_PAGE_MASK) | (pt->gva & 0x1fffff); in large_page_gpa()
174 return pse_pte_to_page(pt->pte[1]) | (pt->gva & 0x3fffff); in large_page_gpa()
191 pt->gva = addr; in walk_gpt()
209 pt->gpa = (pt->pte[0] & page_mask) | (pt->gva & 0xfff); in walk_gpt()
218 bool mmu_gva_to_gpa(CPUState *cpu, target_ulong gva, uint64_t *gpa) in mmu_gva_to_gpa() argument
225 *gpa = gva; in mmu_gva_to_gpa()
229 res = walk_gpt(cpu, gva, err_code, &pt, x86_is_pae_enabled(cpu)); in mmu_gva_to_gpa()
238 void vmx_write_mem(CPUState *cpu, target_ulong gva, void *data, int bytes) in vmx_write_mem() argument
244 int copy = MIN(bytes, 0x1000 - (gva & 0xfff)); in vmx_write_mem()
246 if (!mmu_gva_to_gpa(cpu, gva, &gpa)) { in vmx_write_mem()
247 VM_PANIC_EX("%s: mmu_gva_to_gpa %llx failed\n", __func__, gva); in vmx_write_mem()
254 gva += copy; in vmx_write_mem()
259 void vmx_read_mem(CPUState *cpu, void *data, target_ulong gva, int bytes) in vmx_read_mem() argument
265 int copy = MIN(bytes, 0x1000 - (gva & 0xfff)); in vmx_read_mem()
267 if (!mmu_gva_to_gpa(cpu, gva, &gpa)) { in vmx_read_mem()
268 VM_PANIC_EX("%s: mmu_gva_to_gpa %llx failed\n", __func__, gva); in vmx_read_mem()
274 gva += copy; in vmx_read_mem()