/linux/tools/testing/selftests/sgx/ |
H A D | main.c | 172 struct sgx_enclave_run run; in FIXTURE() local 262 #define ENCL_CALL(op, run, clobbered) \ argument 267 EENTER, 0, 0, (run)); \ 270 (run)); \ 274 #define EXPECT_EEXIT(run) \ argument 276 EXPECT_EQ((run)->function, EEXIT); \ 277 if ((run)->function != EEXIT) \ 278 TH_LOG("0x%02x 0x%02x 0x%016llx", (run)->exception_vector, \ 279 (run)->exception_error_code, (run) 533 test_handler(long rdi,long rsi,long rdx,long ursp,long r8,long r9,struct sgx_enclave_run * run) test_handler() argument [all...] |
/linux/tools/testing/selftests/kvm/s390/ |
H A D | sync_regs_test.c | 76 struct kvm_run *run = vcpu->run; in test_read_invalid() local 80 run->kvm_valid_regs = INVALID_SYNC_FIELD; in test_read_invalid() 85 run->kvm_valid_regs = 0; in test_read_invalid() 87 run->kvm_valid_regs = INVALID_SYNC_FIELD | TEST_SYNC_FIELDS; in test_read_invalid() 92 run->kvm_valid_regs = 0; in test_read_invalid() 97 struct kvm_run *run = vcpu->run; in test_set_invalid() local 101 run->kvm_dirty_regs = INVALID_SYNC_FIELD; in test_set_invalid() 106 run in test_set_invalid() 118 struct kvm_run *run = vcpu->run; test_req_and_verify_all_valid_regs() local 144 struct kvm_run *run = vcpu->run; test_set_and_verify_various_reg_values() local 183 struct kvm_run *run = vcpu->run; test_clear_kvm_dirty_regs_bits() local [all...] |
H A D | ucontrol_test.c | 108 struct kvm_run *run; in FIXTURE() local 145 self->run = (struct kvm_run *)mmap(NULL, self->kvm_run_size, in FIXTURE_SETUP() 147 ASSERT_NE(self->run, MAP_FAILED); in FIXTURE_SETUP() 159 TH_LOG("VM created %p %p", self->run, self->sie_block); in FIXTURE_SETUP() 190 munmap(self->run, self->kvm_run_size); in FIXTURE_TEARDOWN() 308 struct kvm_run *run = self->run; in uc_handle_exit_ucontrol() local 312 TEST_ASSERT_EQ(KVM_EXIT_S390_UCONTROL, run->exit_reason); in uc_handle_exit_ucontrol() 313 switch (run->s390_ucontrol.pgm_code) { in uc_handle_exit_ucontrol() 315 seg_addr = run in uc_handle_exit_ucontrol() 350 struct kvm_run *run = self->run; uc_handle_insn_ic() local 375 struct kvm_run *run = self->run; uc_handle_sieic() local 402 struct kvm_run *run = self->run; uc_handle_exit() local 471 struct kvm_run *run = self->run; TEST_F() local 532 struct kvm_run *run = self->run; TEST_F() local 577 struct kvm_run *run = self->run; TEST_F() local [all...] |
/linux/tools/testing/selftests/kvm/x86/ |
H A D | sync_regs_test.c | 92 struct kvm_run *run = (struct kvm_run *)arg; in race_events_inj_pen() local 93 struct kvm_vcpu_events *events = &run->s.regs.events; in race_events_inj_pen() 98 WRITE_ONCE(run->kvm_dirty_regs, KVM_SYNC_X86_EVENTS); in race_events_inj_pen() 115 struct kvm_run *run = (struct kvm_run *)arg; in race_events_exc() local 116 struct kvm_vcpu_events *events = &run->s.regs.events; in race_events_exc() 119 WRITE_ONCE(run->kvm_dirty_regs, KVM_SYNC_X86_EVENTS); in race_events_exc() 137 struct kvm_run *run = (struct kvm_run *)arg; in race_sregs_cr4() local 138 __u64 *cr4 = &run->s.regs.sregs.cr4; in race_sregs_cr4() 143 WRITE_ONCE(run->kvm_dirty_regs, KVM_SYNC_X86_SREGS); in race_sregs_cr4() 161 struct kvm_run *run; in race_sync_regs() local 210 struct kvm_run *run = vcpu->run; KVM_ONE_VCPU_TEST() local 231 struct kvm_run *run = vcpu->run; KVM_ONE_VCPU_TEST() local 252 struct kvm_run *run = vcpu->run; KVM_ONE_VCPU_TEST() local 275 struct kvm_run *run = vcpu->run; KVM_ONE_VCPU_TEST() local 313 struct kvm_run *run = vcpu->run; KVM_ONE_VCPU_TEST() local 330 struct kvm_run *run = vcpu->run; KVM_ONE_VCPU_TEST() local 361 struct kvm_run *run = vcpu->run; KVM_ONE_VCPU_TEST() local [all...] |
H A D | debug_regs.c | 86 struct kvm_run *run; in main() local 104 run = vcpu->run; in main() 111 TEST_ASSERT(run->exit_reason == KVM_EXIT_DEBUG && in main() 112 run->debug.arch.exception == BP_VECTOR && in main() 113 run->debug.arch.pc == CAST_TO_RIP(sw_bp), in main() 115 run->exit_reason, run->debug.arch.exception, in main() 116 run->debug.arch.pc, CAST_TO_RIP(sw_bp)); in main() 128 TEST_ASSERT(run in main() [all...] |
H A D | userspace_msr_exit_test.c | 262 * Now run the same tests with the instruction emulator. in guest_code_filter_allow() 388 if (vcpu->run->exit_reason == KVM_EXIT_IO && in check_for_guest_assert() 396 struct kvm_run *run = vcpu->run; in process_rdmsr() local 401 TEST_ASSERT(run->msr.index == msr_index, in process_rdmsr() 403 run->msr.index, msr_index); in process_rdmsr() 405 switch (run->msr.index) { in process_rdmsr() 407 run->msr.data = 0; in process_rdmsr() 410 run->msr.error = 1; in process_rdmsr() 413 run in process_rdmsr() 428 struct kvm_run *run = vcpu->run; process_wrmsr() local 595 handle_rdmsr(struct kvm_run * run) handle_rdmsr() argument 612 handle_wrmsr(struct kvm_run * run) handle_wrmsr() argument 635 struct kvm_run *run = vcpu->run; KVM_ONE_VCPU_TEST() local [all...] |
H A D | flds_emulation.h | 22 struct kvm_run *run = vcpu->run; in handle_flds_emulation_failure_exit() local 29 TEST_ASSERT(run->emulation_failure.suberror == KVM_INTERNAL_ERROR_EMULATION, in handle_flds_emulation_failure_exit() 31 run->emulation_failure.suberror); in handle_flds_emulation_failure_exit() 33 flags = run->emulation_failure.flags; in handle_flds_emulation_failure_exit() 34 TEST_ASSERT(run->emulation_failure.ndata >= 3 && in handle_flds_emulation_failure_exit() 36 "run->emulation_failure is missing instruction bytes"); in handle_flds_emulation_failure_exit() 38 TEST_ASSERT(run->emulation_failure.insn_size >= 2, in handle_flds_emulation_failure_exit() 40 run->emulation_failure.insn_size); in handle_flds_emulation_failure_exit() 42 insn_bytes = run in handle_flds_emulation_failure_exit() [all...] |
H A D | xen_vmcall_test.c | 106 volatile struct kvm_run *run = vcpu->run; in main() local 111 if (run->exit_reason == KVM_EXIT_XEN) { in main() 112 TEST_ASSERT_EQ(run->xen.type, KVM_EXIT_XEN_HCALL); in main() 113 TEST_ASSERT_EQ(run->xen.u.hcall.cpl, 0); in main() 114 TEST_ASSERT_EQ(run->xen.u.hcall.longmode, 1); in main() 115 TEST_ASSERT_EQ(run->xen.u.hcall.input, INPUTVALUE); in main() 116 TEST_ASSERT_EQ(run->xen.u.hcall.params[0], ARGVALUE(1)); in main() 117 TEST_ASSERT_EQ(run->xen.u.hcall.params[1], ARGVALUE(2)); in main() 118 TEST_ASSERT_EQ(run in main() [all...] |
H A D | hyperv_extended_hypercalls.c | 41 struct kvm_run *run; in main() local 56 run = vcpu->run; in main() 72 TEST_ASSERT(run->exit_reason == KVM_EXIT_HYPERV, in main() 74 run->exit_reason, exit_reason_str(run->exit_reason)); in main() 76 outval = addr_gpa2hva(vm, run->hyperv.u.hcall.params[1]); in main() 78 run->hyperv.u.hcall.result = HV_STATUS_SUCCESS; in main() 82 TEST_ASSERT(run->exit_reason == KVM_EXIT_IO, in main() 84 run in main() [all...] |
/linux/arch/loongarch/kvm/ |
H A D | exit.c | 157 int kvm_emu_iocsr(larch_inst inst, struct kvm_run *run, struct kvm_vcpu *vcpu) in kvm_emu_iocsr() argument 170 run->iocsr_io.phys_addr = addr; in kvm_emu_iocsr() 171 run->iocsr_io.is_write = 0; in kvm_emu_iocsr() 177 run->iocsr_io.len = 1; in kvm_emu_iocsr() 180 run->iocsr_io.len = 2; in kvm_emu_iocsr() 183 run->iocsr_io.len = 4; in kvm_emu_iocsr() 186 run->iocsr_io.len = 8; in kvm_emu_iocsr() 189 run->iocsr_io.len = 1; in kvm_emu_iocsr() 190 run->iocsr_io.is_write = 1; in kvm_emu_iocsr() 193 run in kvm_emu_iocsr() 237 kvm_complete_iocsr_read(struct kvm_vcpu * vcpu,struct kvm_run * run) kvm_complete_iocsr_read() argument 281 struct kvm_run *run = vcpu->run; kvm_trap_handle_gspr() local 373 struct kvm_run *run = vcpu->run; kvm_emu_mmio_read() local 501 kvm_complete_mmio_read(struct kvm_vcpu * vcpu,struct kvm_run * run) kvm_complete_mmio_read() argument 548 struct kvm_run *run = vcpu->run; kvm_emu_mmio_write() local 671 struct kvm_run *run = vcpu->run; kvm_handle_rdwr_fault() local 720 kvm_complete_user_service(struct kvm_vcpu * vcpu,struct kvm_run * run) kvm_complete_user_service() argument 738 struct kvm_run *run = vcpu->run; kvm_handle_fpu_disabled() local [all...] |
/linux/drivers/staging/media/sunxi/cedrus/ |
H A D | cedrus_dec.c | 29 struct cedrus_run run = {}; in cedrus_device_run() local 33 run.src = v4l2_m2m_next_src_buf(ctx->fh.m2m_ctx); in cedrus_device_run() 34 run.dst = v4l2_m2m_next_dst_buf(ctx->fh.m2m_ctx); in cedrus_device_run() 37 src_req = run.src->vb2_buf.req_obj.req; in cedrus_device_run() 44 run.mpeg2.sequence = cedrus_find_control_data(ctx, in cedrus_device_run() 46 run.mpeg2.picture = cedrus_find_control_data(ctx, in cedrus_device_run() 48 run.mpeg2.quantisation = cedrus_find_control_data(ctx, in cedrus_device_run() 53 run.h264.decode_params = cedrus_find_control_data(ctx, in cedrus_device_run() 55 run.h264.pps = cedrus_find_control_data(ctx, in cedrus_device_run() 57 run in cedrus_device_run() [all...] |
/linux/fs/ntfs3/ |
H A D | run.c | 34 static bool run_lookup(const struct runs_tree *run, CLST vcn, size_t *index) in run_lookup() argument 39 if (!run->count) { in run_lookup() 45 max_idx = run->count - 1; in run_lookup() 48 r = run->runs; in run_lookup() 61 *index = run->count; in run_lookup() 72 r = run->runs + mid_idx; in run_lookup() 93 static void run_consolidate(struct runs_tree *run, size_t index) in run_consolidate() argument 96 struct ntfs_run *r = run->runs + index; in run_consolidate() 98 while (index + 1 < run->count) { in run_consolidate() 100 * I should merge current run wit in run_consolidate() 167 run_is_mapped_full(const struct runs_tree * run,CLST svcn,CLST evcn) run_is_mapped_full() argument 192 run_lookup_entry(const struct runs_tree * run,CLST vcn,CLST * lcn,CLST * len,size_t * index) run_lookup_entry() argument 228 run_truncate_head(struct runs_tree * run,CLST vcn) run_truncate_head() argument 263 run_truncate(struct runs_tree * run,CLST vcn) run_truncate() argument 300 run_truncate_around(struct runs_tree * run,CLST vcn) run_truncate_around() argument 316 run_add_entry(struct runs_tree * run,CLST vcn,CLST lcn,CLST len,bool is_mft) run_add_entry() argument 489 run_collapse_range(struct runs_tree * run,CLST vcn,CLST len) run_collapse_range() argument 555 run_insert_range(struct runs_tree * run,CLST vcn,CLST len) run_insert_range() argument 595 run_get_entry(const struct runs_tree * run,size_t index,CLST * vcn,CLST * lcn,CLST * len) run_get_entry() argument 817 run_pack(const struct runs_tree * run,CLST svcn,CLST len,u8 * run_buf,u32 run_buf_size,CLST * packed_vcns) run_pack() argument 920 run_unpack(struct runs_tree * run,struct ntfs_sb_info * sbi,CLST ino,CLST svcn,CLST evcn,CLST vcn,const u8 * run_buf,int run_buf_size) run_unpack() argument 1052 run_unpack_ex(struct runs_tree * run,struct ntfs_sb_info * sbi,CLST ino,CLST svcn,CLST evcn,CLST vcn,const u8 * run_buf,int run_buf_size) run_unpack_ex() argument 1173 run_clone(const struct runs_tree * run,struct runs_tree * new_run) run_clone() argument [all...] |
/linux/tools/testing/selftests/kvm/ |
H A D | coalesced_io_test.c | 78 struct kvm_run *run = vcpu->run; in vcpu_run_and_verify_io_exit() local 90 if (run->exit_reason == KVM_EXIT_IO) in vcpu_run_and_verify_io_exit() 91 pio_value = *(uint32_t *)((void *)run + run->io.data_offset); in vcpu_run_and_verify_io_exit() 95 TEST_ASSERT((!want_pio && (run->exit_reason == KVM_EXIT_MMIO && run->mmio.is_write && in vcpu_run_and_verify_io_exit() 96 run->mmio.phys_addr == io->mmio_gpa && run->mmio.len == 8 && in vcpu_run_and_verify_io_exit() 97 *(uint64_t *)run in vcpu_run_and_verify_io_exit() [all...] |
/linux/arch/mips/kvm/ |
H A D | emulate.c | 962 vcpu->run->exit_reason = KVM_EXIT_IRQ_WINDOW_OPEN; in kvm_mips_emul_wait() 975 struct kvm_run *run = vcpu->run; in kvm_mips_emulate_store() local 976 void *data = run->mmio.data; in kvm_mips_emulate_store() 991 run->mmio.phys_addr = kvm_mips_callbacks->gva_to_gpa( in kvm_mips_emulate_store() 993 if (run->mmio.phys_addr == KVM_INVALID_ADDR) in kvm_mips_emulate_store() 999 run->mmio.len = 8; in kvm_mips_emulate_store() 1009 run->mmio.len = 4; in kvm_mips_emulate_store() 1018 run->mmio.len = 2; in kvm_mips_emulate_store() 1027 run in kvm_mips_emulate_store() 1271 struct kvm_run *run = vcpu->run; kvm_mips_emulate_load() local 1511 struct kvm_run *run = vcpu->run; kvm_mips_complete_mmio_load() local [all...] |
/linux/arch/s390/kvm/ |
H A D | diag.c | 43 start = vcpu->run->s.regs.gprs[(vcpu->arch.sie_block->ipa & 0xf0) >> 4]; in diag_release_pages() 44 end = vcpu->run->s.regs.gprs[vcpu->arch.sie_block->ipa & 0xf] + PAGE_SIZE; in diag_release_pages() 96 vcpu->run->s.regs.gprs[rx]); in __diag_page_ref_service() 98 if (vcpu->run->s.regs.gprs[rx] & 7) in __diag_page_ref_service() 100 rc = read_guest_real(vcpu, vcpu->run->s.regs.gprs[rx], &parm, sizeof(parm)); in __diag_page_ref_service() 117 vcpu->run->s.regs.gprs[ry] = 8; in __diag_page_ref_service() 131 vcpu->run->s.regs.gprs[ry] = 0; in __diag_page_ref_service() 145 vcpu->run->s.regs.gprs[ry] = 0; in __diag_page_ref_service() 151 vcpu->run->s.regs.gprs[ry] = 4; in __diag_page_ref_service() 192 tid = vcpu->run in __diag_time_slice_end_directed() [all...] |
/linux/tools/testing/selftests/kvm/include/s390/ |
H A D | debug_print.h | 43 static inline void print_psw(struct kvm_run *run, struct kvm_s390_sie_block *sie_block) in print_psw() argument 46 run->flags, in print_psw() 47 run->psw_mask, run->psw_addr, in print_psw() 48 run->exit_reason, exit_reason_str(run->exit_reason)); in print_psw() 53 static inline void print_run(struct kvm_run *run, struct kvm_s390_sie_block *sie_block) in print_run() argument 55 print_hex_bytes("run", (u64)run, 0x150); in print_run() 57 print_psw(run, sie_bloc in print_run() 60 print_regs(struct kvm_run * run) print_regs() argument [all...] |
/linux/tools/perf/scripts/python/ |
H A D | stat-cpi.py | 23 def store(time, event, cpu, thread, val, ena, run): argument 24 #print("event %s cpu %d, thread %d, time %d, val %d, ena %d, run %d" % 25 # (event, cpu, thread, time, val, ena, run)) 29 data[key] = [ val, ena, run] 35 def stat__cycles_k(cpu, thread, time, val, ena, run): argument 36 store(time, "cycles", cpu, thread, val, ena, run); 38 def stat__instructions_k(cpu, thread, time, val, ena, run): argument 39 store(time, "instructions", cpu, thread, val, ena, run); 41 def stat__cycles_u(cpu, thread, time, val, ena, run): argument 42 store(time, "cycles", cpu, thread, val, ena, run); 44 stat__instructions_u(cpu, thread, time, val, ena, run) global() argument 47 stat__cycles(cpu, thread, time, val, ena, run) global() argument 50 stat__instructions(cpu, thread, time, val, ena, run) global() argument [all...] |
/linux/arch/riscv/kvm/ |
H A D | vcpu_insn.c | 153 int (*func)(struct kvm_vcpu *vcpu, struct kvm_run *run, ulong insn); 156 static int truly_illegal_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, in truly_illegal_insn() argument 172 static int truly_virtual_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, in truly_virtual_insn() argument 202 static int wfi_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, ulong insn) in wfi_insn() argument 209 static int wrs_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, ulong insn) in wrs_insn() argument 249 * @run: The VCPU run struct containing the CSR data 253 int kvm_riscv_vcpu_csr_return(struct kvm_vcpu *vcpu, struct kvm_run *run) in kvm_riscv_vcpu_csr_return() argument 265 run->riscv_csr.ret_value); in kvm_riscv_vcpu_csr_return() 273 static int csr_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, ulon argument 398 system_opcode_insn(struct kvm_vcpu * vcpu,struct kvm_run * run,ulong insn) system_opcode_insn() argument 438 kvm_riscv_vcpu_virtual_insn(struct kvm_vcpu * vcpu,struct kvm_run * run,struct kvm_cpu_trap * trap) kvm_riscv_vcpu_virtual_insn() argument 481 kvm_riscv_vcpu_mmio_load(struct kvm_vcpu * vcpu,struct kvm_run * run,unsigned long fault_addr,unsigned long htinst) kvm_riscv_vcpu_mmio_load() argument 603 kvm_riscv_vcpu_mmio_store(struct kvm_vcpu * vcpu,struct kvm_run * run,unsigned long fault_addr,unsigned long htinst) kvm_riscv_vcpu_mmio_store() argument 731 kvm_riscv_vcpu_mmio_return(struct kvm_vcpu * vcpu,struct kvm_run * run) kvm_riscv_vcpu_mmio_return() argument [all...] |
/linux/fs/befs/ |
H A D | endian.h | 74 befs_block_run run; in fsrun_to_cpu() local 77 run.allocation_group = le32_to_cpu((__force __le32)n.allocation_group); in fsrun_to_cpu() 78 run.start = le16_to_cpu((__force __le16)n.start); in fsrun_to_cpu() 79 run.len = le16_to_cpu((__force __le16)n.len); in fsrun_to_cpu() 81 run.allocation_group = be32_to_cpu((__force __be32)n.allocation_group); in fsrun_to_cpu() 82 run.start = be16_to_cpu((__force __be16)n.start); in fsrun_to_cpu() 83 run.len = be16_to_cpu((__force __be16)n.len); in fsrun_to_cpu() 85 return run; in fsrun_to_cpu() 91 befs_disk_block_run run; in cpu_to_fsrun() local 94 run in cpu_to_fsrun() [all...] |
H A D | datastream.c | 26 befs_blocknr_t blockno, befs_block_run *run); 31 befs_block_run *run); 36 befs_block_run *run); 53 befs_block_run run; in befs_read_datastream() local 61 if (befs_fblock2brun(sb, ds, block, &run) != BEFS_OK) { in befs_read_datastream() 67 bh = befs_bread_iaddr(sb, run); in befs_read_datastream() 80 * befs_fblock2brun - give back block run for fblock 84 * @run: The found run is passed back through this pointer 96 befs_blocknr_t fblock, befs_block_run *run) in befs_fblock2brun() argument 252 befs_find_brun_direct(struct super_block * sb,const befs_data_stream * data,befs_blocknr_t blockno,befs_block_run * run) befs_find_brun_direct() argument 308 befs_find_brun_indirect(struct super_block * sb,const befs_data_stream * data,befs_blocknr_t blockno,befs_block_run * run) befs_find_brun_indirect() argument 417 befs_find_brun_dblindirect(struct super_block * sb,const befs_data_stream * data,befs_blocknr_t blockno,befs_block_run * run) befs_find_brun_dblindirect() argument [all...] |
/linux/tools/testing/selftests/kvm/lib/s390/ |
H A D | ucall.c | 11 struct kvm_run *run = vcpu->run; in ucall_arch_get_ucall() local 13 if (run->exit_reason == KVM_EXIT_S390_SIEIC && in ucall_arch_get_ucall() 14 run->s390_sieic.icptcode == 4 && in ucall_arch_get_ucall() 15 (run->s390_sieic.ipa >> 8) == 0x83 && /* 0x83 means DIAGNOSE */ in ucall_arch_get_ucall() 16 (run->s390_sieic.ipb >> 16) == 0x501) { in ucall_arch_get_ucall() 17 int reg = run->s390_sieic.ipa & 0xf; in ucall_arch_get_ucall() 19 return (void *)run->s.regs.gprs[reg]; in ucall_arch_get_ucall()
|
H A D | diag318_test_handler.c | 30 struct kvm_run *run; in diag318_handler() local 36 run = vcpu->run; in diag318_handler() 39 TEST_ASSERT(run->s390_sieic.icptcode == ICPT_INSTRUCTION, in diag318_handler() 40 "Unexpected intercept code: 0x%x", run->s390_sieic.icptcode); in diag318_handler() 41 TEST_ASSERT((run->s390_sieic.ipa & 0xff00) == IPA0_DIAG, in diag318_handler() 42 "Unexpected IPA0 code: 0x%x", (run->s390_sieic.ipa & 0xff00)); in diag318_handler() 44 reg = (run->s390_sieic.ipa & 0x00f0) >> 4; in diag318_handler() 45 diag318_info = run->s.regs.gprs[reg]; in diag318_handler()
|
/linux/tools/testing/selftests/arm64/mte/ |
H A D | check_child_memory.c | 87 int run, result; in check_child_memory_mapping() local 92 for (run = 0; run < item; run++) { in check_child_memory_mapping() 93 ptr = (char *)mte_allocate_memory_tag_range(sizes[run], mem_type, mapping, in check_child_memory_mapping() 95 if (check_allocated_memory_range(ptr, sizes[run], mem_type, in check_child_memory_mapping() 98 result = check_child_tag_inheritance(ptr, sizes[run], mode); in check_child_memory_mapping() 99 mte_free_memory_tag_range((void *)ptr, sizes[run], mem_type, UNDERFLOW, OVERFLOW); in check_child_memory_mapping() 109 int run, fd, map_size, result = KSFT_PASS; in check_child_file_mapping() local 113 for (run in check_child_file_mapping() [all...] |
H A D | check_tags_inclusion.c | 52 int tag, run, ret, result = KSFT_PASS; in check_single_included_tags() local 64 for (run = 0; (run < RUNS) && (result == KSFT_PASS); run++) { in check_single_included_tags() 84 int tag, run, result = KSFT_PASS; in check_multiple_included_tags() local 96 for (run = 0; (run < RUNS) && (result == KSFT_PASS); run++) { in check_multiple_included_tags() 116 int run, ret, result = KSFT_PASS; in check_all_included_tags() local 127 for (run in check_all_included_tags() 142 int run, ret; check_none_included_tags() local [all...] |
/linux/arch/arm64/kvm/ |
H A D | mmio.c | 124 struct kvm_run *run = vcpu->run; in kvm_handle_mmio_return() local 127 data = kvm_mmio_read_buf(run->mmio.data, len); in kvm_handle_mmio_return() 138 trace_kvm_mmio(KVM_TRACE_MMIO_READ, len, run->mmio.phys_addr, in kvm_handle_mmio_return() 155 struct kvm_run *run = vcpu->run; in io_mem_abort() local 179 run->exit_reason = KVM_EXIT_ARM_NISV; in io_mem_abort() 180 run->arm_nisv.esr_iss = kvm_vcpu_dabt_iss_nisv_sanitized(vcpu); in io_mem_abort() 181 run->arm_nisv.fault_ipa = fault_ipa; in io_mem_abort() 215 run in io_mem_abort() [all...] |