/kvm-unit-tests/lib/s390x/ |
H A D | mmu.h | 25 * @vaddr address whose page tables are to split 34 * @vaddr address whose protection bits are to be changed 44 * @vaddr address whose protection bits are to be changed 53 * @start starting address whose protection bits are to be changed 62 * @start starting address whose protection bits are to be changed 68 /* Similar to install_page, maps the virtual address to the physical address 72 * @phys physical address to map, must be 1MB aligned! 73 * @vaddr virtual address to map, must be 1MB aligned! 77 /* Similar to install_page, maps the virtual address to the physical address 81 * @phys physical address to map, must be 2GB aligned! [all …]
|
H A D | fault.c | 80 * address is unpredictable. in print_decode_teid() 85 printf("Address: %lx, unpredictable\n ", raw_teid & PAGE_MASK); in print_decode_teid() 89 printf("Address: %lx\n\n", raw_teid & PAGE_MASK); in print_decode_teid()
|
H A D | mmu.c | 50 * Get the pud (region 3) DAT table entry for the given address and root, 63 * Get the pmd (segment) DAT table entry for the given address and pud, 77 * Get the pte (page) DAT table entry for the given address and pmd, 94 * @va the virtual address corresponding to this pmd. 116 * @va the virtual address corresponding to this pud. 186 * Get the DAT table entry of the given level for the given address, 190 * @vaddr virtual address 242 * Set the DAT entry for the given level of the given virtual address. If a 248 * @vaddr the virtual address
|
/kvm-unit-tests/s390x/ |
H A D | ex.c | 10 * address and the instruction-length code. 26 * Saves the next instruction address (address from PSW + length of instruction) 29 * BASR does *not* perform a relative address calculation with an intermediate. 47 report(ret_addr == after_ex, "return address after EX"); in test_basr() 53 * According to PoP (Branch-Address Generation), the address calculated relative 54 * to the instruction address is relative to BRAS when it is the target of an 73 * In case the address calculation is correct, we jump by the relative offset 1b-0b from 0b to 1b. in test_bras() 74 * In case the address calculation is relative to the exrl (i.e. a test failure), in test_bras() 99 report(after_target == branch_addr, "address calculated relative to BRAS"); in test_bras() 100 report(ret_addr == after_ex, "return address after EX"); in test_bras() [all …]
|
H A D | diag258.c | 94 * Verify that the refbk pointer is a real address and not a virtual 95 * address. This is tested by enabling DAT and establishing a mapping 97 * address space. 156 * address lowcore_offset_for_refbk (reverse prefixing). in test_refbk_prefixing() 162 * address &prefix_buf[0] + lowcore_offset_for_refbk (forward prefixing). in test_refbk_prefixing() 171 * refbk_in_reverse_prefix correctly, it would access absolute address in test_refbk_prefixing() 172 * refbk_in_reverse_prefix (which to us is accessible at real address in test_refbk_prefixing() 185 report(!ry, "real address refbk accessed"); in test_refbk_prefixing()
|
H A D | smp.c | 30 { SIGP_STOP, "stop with invalid CPU address" }, 31 { SIGP_START, "start with invalid CPU address" }, 32 { SIGP_CPU_RESET, "reset with invalid CPU address" }, 33 { SIGP_COND_EMERGENCY_SIGNAL, "conditional emcall with invalid CPU address" }, 34 { SIGP_EMERGENCY_SIGNAL, "emcall with invalid CPU address" }, 35 { SIGP_EXTERNAL_CALL, "ecall with invalid CPU address" }, 36 { INVALID_ORDER_CODE, "invalid order code and CPU address" }, 37 { SIGP_SENSE, "sense with invalid CPU address" }, 38 { SIGP_STOP_AND_STORE_STATUS, "stop and store status with invalid CPU address" }, 202 report_prefix_push("store status at address"); in test_store_status() [all …]
|
H A D | diag308.c | 24 * Check that diag308 with subcode 0 and 1 loads the PSW at address 0, i.e. 25 * that we can put a pointer into address 4 which then gets executed. 51 /* Expect a specification exception when using an unaligned address */ 57 report_prefix_push("unaligned address"); in test_unaligned_address()
|
H A D | sclp.c | 55 * Perform one test at the given address, optionally using the SCCB template, 59 * should be copied to the test address, and should be 0 when the test 60 * address is invalid, in which case nothing is copied. 72 /* Copy the template to the test address if needed */ in test_one_sccb() 80 report_info("SCLP not ready (command %#x, address %p, cc %d)", cmd, addr, res); in test_one_sccb() 101 * copied and the memory address should not be touched. 113 * template that need to be copied to the actual test address. In many cases 155 * Test SCCBs whose address is in the lowcore or prefix area. 191 * SCCBs all over the two pages starting at absolute address 0, thus in test_sccb_prefix() 369 report_skip("Invalid SCCB address"); in test_addressing() [all …]
|
H A D | sie-dat.c | 32 /* guest will tell us the guest physical address of the test buffer */ in test_sie_dat() 55 …/* the guest will now write to an unmapped address and we check that this causes a segment transla… in test_sie_dat() 60 …report((lowcore.trans_exc_id & PAGE_MASK) == (GUEST_TOTAL_PAGE_COUNT * PAGE_SIZE), "TEID address m… in test_sie_dat()
|
/kvm-unit-tests/lib/ |
H A D | acpi.h | 35 u32 rsdt_physical_address; /* 32-bit physical address of RSDT */ 37 u64 xsdt_physical_address; /* 64-bit physical address of XSDT */ 69 u8 space_id; /* Address space where struct or register exists */ 73 u64 address; /* 64-bit address of struct or register */ member 78 u32 firmware_ctrl; /* Physical address of FACS */ 79 u32 dsdt; /* Physical address of DSDT */ 83 u32 smi_cmd; /* Port address of SMI command port */ 88 u32 pm1a_evt_blk; /* Port address of Power Mgt 1a acpi_event Reg Blk */ 89 u32 pm1b_evt_blk; /* Port address of Power Mgt 1b acpi_event Reg Blk */ 90 u32 pm1a_cnt_blk; /* Port address of Power Mgt 1a Control Reg Blk */ [all …]
|
H A D | vmalloc.h | 13 /* Set the top of the virtual address space */ 24 /* Walk the page table and resolve the virtual address to a physical address */ 26 /* Map the virtual address to the physical address for the given page tables */
|
H A D | devicetree.h | 61 * translate the @regidx'th "address size" tuple of 74 /* the processor bus (pbus) address type and register tuple */ 92 * processor bus using the parent node's #address-cells 166 * #address-cells and #size-cells properties of @fdtnode 223 * dt_get_initrd gets the physical address of the initrd and its 234 * storing each memory region ("address size" tuple) in consecutive entries
|
/kvm-unit-tests/x86/ |
H A D | lam.c | 49 static bool get_lam_mask(u64 address, u64* lam_mask) in get_lam_mask() argument 52 * Use LAM57_MASK as mask to construct non-canonical address if LAM is in get_lam_mask() 58 * Bit 63 determines if the address should be treated as a user address in get_lam_mask() 59 * or a supervisor address. in get_lam_mask() 61 if (address & BIT_ULL(63)) { in get_lam_mask() 90 report(!fault, "Expected access to untagged address for %s to succeed", in test_ptr() 95 report(fault != lam_active, "Expected access to tagged address for %s %s LAM to %s", in test_ptr() 106 report(fault, "Expected access to non-LAM-canonical address for %s to #GP", in test_ptr() 111 /* invlpg with tagged address is same as NOP, no #GP expected. */ 131 /* LAM doesn't apply to the linear address in the descriptor of invpcid */ [all …]
|
H A D | hypercall.c | 34 "popq test_rip(%rip)\n\t" // pop return address 35 "pushq %rsi\n\t" // new return address 45 "leaq 1f(%%rip), %%rsi\n\t" // save return address for gp_tss in test_edge() 85 "VMCALL on edge of canonical address space (intel)"); in main() 92 "VMMCALL on edge of canonical address space (AMD)"); in main()
|
/kvm-unit-tests/lib/riscv/ |
H A D | mmu.c | 201 phys_addr_t virt_to_phys(volatile void *address) in virt_to_phys() argument 207 return __pa(address); in virt_to_phys() 209 return virt_to_pte_phys(pgtable, (void *)address); in virt_to_phys() 212 void *phys_to_virt(phys_addr_t address) in phys_to_virt() argument 214 /* @address must have an identity mapping for this to work. */ in phys_to_virt() 215 assert(address == __pa(address)); in phys_to_virt() 216 assert(virt_to_phys(__va(address)) == address); in phys_to_virt() 217 return __va(address); in phys_to_virt()
|
/kvm-unit-tests/lib/s390x/asm/ |
H A D | cpacf.h | 272 * @param: address of parameter block; see POP for details on each func 273 * @dest: address of destination memory area 274 * @src: address of source memory area 302 * @param: address of parameter block; see POP for details on each func 303 * @dest: address of destination memory area 304 * @src: address of source memory area 333 * @param: address of parameter block; see POP for details on each func 334 * @src: address of source memory area 356 * @param: address of parameter block; see POP for details on each func 357 * @src: address of source memory area [all …]
|
/kvm-unit-tests/lib/x86/ |
H A D | intel-iommu.h | 48 #define DMAR_FEUADDR_REG 0x44 /* Upper address */ 69 #define DMAR_IEADDR_REG 0xa8 /* Invalidation event address */ 70 #define DMAR_IEUADDR_REG 0xac /* Invalidation event address */ 75 #define DMAR_PQA_REG 0xd0 /* Page request queue address */ 80 #define DMAR_PEADDR_REG 0xe8 /* Page request event address */ 81 #define DMAR_PEUADDR_REG 0xec /* Page event upper address */ 97 /* Supported Adjusted Guest Address Widths */
|
H A D | intel-iommu.c | 20 * VT-d in QEMU currently only support 39 bits address width, which is 130 printf("DMAR table address: %#018lx\n", vtd_root_table()); in vtd_setup_root_table() 140 printf("IR table address: %#018lx\n", vtd_ir_table()); in vtd_setup_ir_table() 169 * vtd_map_range: setup IO address mapping for specific memory range 172 * @iova: start IO virtual address 173 * @pa: start physical address
|
/kvm-unit-tests/powerpc/ |
H A D | cstart64.S | 32 * We were loaded at QEMU's kernel load address, but we're not 36 * pseries load addresses are not the same, so find the address 41 subi r31, r31, 0b - start /* QEMU's kernel load address */ 66 /* compute address of call_handler */ 71 /* relocate vector table to base address 0x0 (MSR_IP = 0) */ 142 subi r31, r31, 0b - start /* QEMU's kernel load address */ 303 mfsprg0 r1 /* get struct cpu address */ 304 ld r1,CPU_EXCEPTION_STACK(r1) /* get exception stack address */ 320 mfsprg0 r1 /* get struct cpu address */ 321 ld r1,CPU_EXCEPTION_STACK(r1) /* get exception stack address */
|
/kvm-unit-tests/scripts/ |
H A D | get_maintainer.pl | 91 # rfc822 email address - preloaded methods go here. 429 # Read mail address map 461 my $address = $2; 464 ($real_name, $address) = parse_email("$real_name <$address>"); 465 $mailmap->{names}->{$address} = $real_name; 685 print("$x->{file}:$x->{linenr}: warning: section without email address\t$x->{line}\n"); 773 my ($address) = @_; 776 return 1 if ($ignore eq $address); 951 my ($name, $address) = parse_email($email); 953 my $tmp_email = format_email($name, $address, $email_usename); [all …]
|
H A D | pretty_print_stacks.py | 19 # caller address so line numbers are more intuitive. Thus we subtract 1 20 # from the address to get the call code. 31 cmd = [config.get('ADDR2LINE', 'addr2line'), '-e', binary, '-i', '-f', '--pretty', '--address']
|
/kvm-unit-tests/lib/arm64/ |
H A D | stack.c | 46 * address (it deducts 1 to find the call address, but we want in arch_backtrace_frame() 47 * the actual address). in arch_backtrace_frame()
|
/kvm-unit-tests/lib/libfdt/ |
H A D | fdt_addresses.c | 38 val = fdt_cells(fdt, nodeoffset, "#address-cells"); in fdt_address_cells() 56 /* This function assumes that [address|size]_cells is 1 or 2 */ 73 /* check validity of address */ in fdt_appendprop_addrrange()
|
/kvm-unit-tests/lib/arm/ |
H A D | io.c | 25 * base address that we read from the device tree later. This is 26 * the address we expect the virtual machine manager to put in 37 * the TX register at offset 0 from the base address, so there is no in uart0_init_fdt() 83 uart0_base = ioremap(spcr->serial_port.address, spcr->serial_port.bit_width); in uart0_init_acpi()
|
/kvm-unit-tests/s390x/snippets/c/ |
H A D | sie-dat.c | 23 /* tell the host the page's physical address (we're running DAT off) */ in main() 33 /* the first unmapped address */ in main()
|