| /linux/tools/testing/selftests/sgx/ |
| H A D | main.c | 172 struct sgx_enclave_run run; in FIXTURE() local 262 #define ENCL_CALL(op, run, clobbered) \ argument 267 EENTER, 0, 0, (run)); \ 270 (run)); \ 274 #define EXPECT_EEXIT(run) \ argument 276 EXPECT_EQ((run)->function, EEXIT); \ 277 if ((run)->function != EEXIT) \ 278 TH_LOG("0x%02x 0x%02x 0x%016llx", (run)->exception_vector, \ 279 (run)->exception_error_code, (run)->exception_addr); \ 289 memset(&self->run, 0, sizeof(self->run)); in TEST_F() [all …]
|
| /linux/tools/perf/tests/ |
| H A D | make | 10 # run only specific test over 'Makefile' 127 # $(run) contains all available tests 128 run := make_pure 129 # Targets 'clean all' can be run together only through top level 133 run += make_clean_all 138 run += make_python_perf_so 139 run += make_debug 140 run += make_nondistro 141 run += make_extra_tests 142 run += make_jevents_all [all …]
|
| /linux/arch/riscv/kvm/ |
| H A D | vcpu_insn.c | 27 int (*func)(struct kvm_vcpu *vcpu, struct kvm_run *run, ulong insn); 30 static int truly_illegal_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, in truly_illegal_insn() argument 46 static int truly_virtual_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, in truly_virtual_insn() argument 76 static int wfi_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, ulong insn) in wfi_insn() argument 83 static int wrs_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, ulong insn) in wrs_insn() argument 127 int kvm_riscv_vcpu_csr_return(struct kvm_vcpu *vcpu, struct kvm_run *run) in kvm_riscv_vcpu_csr_return() argument 139 run->riscv_csr.ret_value); in kvm_riscv_vcpu_csr_return() 147 static int csr_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, ulong insn) in csr_insn() argument 191 run->riscv_csr.csr_num = csr_num; in csr_insn() 192 run->riscv_csr.new_value = new_val; in csr_insn() [all …]
|
| H A D | vcpu_sbi.c | 128 struct kvm_run *run, in kvm_riscv_vcpu_sbi_forward_handler() argument 135 run->exit_reason = KVM_EXIT_RISCV_SBI; in kvm_riscv_vcpu_sbi_forward_handler() 136 run->riscv_sbi.extension_id = cp->a7; in kvm_riscv_vcpu_sbi_forward_handler() 137 run->riscv_sbi.function_id = cp->a6; in kvm_riscv_vcpu_sbi_forward_handler() 138 run->riscv_sbi.args[0] = cp->a0; in kvm_riscv_vcpu_sbi_forward_handler() 139 run->riscv_sbi.args[1] = cp->a1; in kvm_riscv_vcpu_sbi_forward_handler() 140 run->riscv_sbi.args[2] = cp->a2; in kvm_riscv_vcpu_sbi_forward_handler() 141 run->riscv_sbi.args[3] = cp->a3; in kvm_riscv_vcpu_sbi_forward_handler() 142 run->riscv_sbi.args[4] = cp->a4; in kvm_riscv_vcpu_sbi_forward_handler() 143 run->riscv_sbi.args[5] = cp->a5; in kvm_riscv_vcpu_sbi_forward_handler() [all …]
|
| /linux/tools/testing/selftests/kvm/ |
| H A D | coalesced_io_test.c | 78 struct kvm_run *run = vcpu->run; in vcpu_run_and_verify_io_exit() local 90 if (run->exit_reason == KVM_EXIT_IO) in vcpu_run_and_verify_io_exit() 91 pio_value = *(uint32_t *)((void *)run + run->io.data_offset); in vcpu_run_and_verify_io_exit() 95 TEST_ASSERT((!want_pio && (run->exit_reason == KVM_EXIT_MMIO && run->mmio.is_write && in vcpu_run_and_verify_io_exit() 96 run->mmio.phys_addr == io->mmio_gpa && run->mmio.len == 8 && in vcpu_run_and_verify_io_exit() 97 *(uint64_t *)run->mmio.data == io->mmio_gpa + io->ring_size - 1)) || in vcpu_run_and_verify_io_exit() 98 (want_pio && (run->exit_reason == KVM_EXIT_IO && run->io.port == io->pio_port && in vcpu_run_and_verify_io_exit() 99 run->io.direction == KVM_EXIT_IO_OUT && run->io.count == 1 && in vcpu_run_and_verify_io_exit() 106 (want_pio ? io->pio_port : io->mmio_gpa) + io->ring_size - 1, run->exit_reason, in vcpu_run_and_verify_io_exit() 107 … run->exit_reason == KVM_EXIT_MMIO ? "MMIO" : run->exit_reason == KVM_EXIT_IO ? "PIO" : "other", in vcpu_run_and_verify_io_exit() [all …]
|
| H A D | set_memory_region_test.c | 55 struct kvm_run *run = vcpu->run; in vcpu_worker() local 67 if (run->exit_reason == KVM_EXIT_IO) { in vcpu_worker() 76 if (run->exit_reason != KVM_EXIT_MMIO) in vcpu_worker() 79 TEST_ASSERT(!run->mmio.is_write, "Unexpected exit mmio write"); in vcpu_worker() 80 TEST_ASSERT(run->mmio.len == 8, in vcpu_worker() 81 "Unexpected exit mmio size = %u", run->mmio.len); in vcpu_worker() 83 TEST_ASSERT(run->mmio.phys_addr == MEM_REGION_GPA, in vcpu_worker() 85 run->mmio.phys_addr); in vcpu_worker() 86 memcpy(run->mmio.data, &MMIO_VAL, 8); in vcpu_worker() 89 if (run->exit_reason == KVM_EXIT_IO && cmd == UCALL_ABORT) in vcpu_worker() [all …]
|
| H A D | guest_print_test.c | 128 struct kvm_run *run = vcpu->run; in run_test() local 134 TEST_ASSERT(run->exit_reason == UCALL_EXIT_REASON, in run_test() 136 run->exit_reason, exit_reason_str(run->exit_reason)); in run_test() 171 struct kvm_run *run; in test_limits() local 176 run = vcpu->run; in test_limits() 179 TEST_ASSERT(run->exit_reason == UCALL_EXIT_REASON, in test_limits() 181 run->exit_reason, exit_reason_str(run->exit_reason)); in test_limits()
|
| /linux/tools/testing/selftests/kvm/s390/ |
| H A D | ucontrol_test.c | 108 struct kvm_run *run; in FIXTURE() local 145 self->run = kvm_mmap(self->kvm_run_size, PROT_READ | PROT_WRITE, in FIXTURE_SETUP() 157 TH_LOG("VM created %p %p", self->run, self->sie_block); in FIXTURE_SETUP() 188 kvm_munmap(self->run, self->kvm_run_size); in FIXTURE_TEARDOWN() 306 struct kvm_run *run = self->run; in uc_handle_exit_ucontrol() local 310 TEST_ASSERT_EQ(KVM_EXIT_S390_UCONTROL, run->exit_reason); in uc_handle_exit_ucontrol() 311 switch (run->s390_ucontrol.pgm_code) { in uc_handle_exit_ucontrol() 313 seg_addr = run->s390_ucontrol.trans_exc_code & ~(SZ_1M - 1); in uc_handle_exit_ucontrol() 315 run->s390_ucontrol.trans_exc_code, seg_addr); in uc_handle_exit_ucontrol() 321 TEST_FAIL("UNEXPECTED PGM CODE %d", run->s390_ucontrol.pgm_code); in uc_handle_exit_ucontrol() [all …]
|
| H A D | user_operexec.c | 33 TEST_ASSERT_EQ(vcpu->run->s390_sieic.icptcode, ICPT_OPEREXC); in test_user_instr0() 34 TEST_ASSERT_EQ(vcpu->run->s390_sieic.ipa, 0); in test_user_instr0() 56 TEST_ASSERT_EQ(vcpu->run->s390_sieic.icptcode, ICPT_OPEREXC); in test_user_operexec() 57 TEST_ASSERT_EQ(vcpu->run->s390_sieic.ipa, 0x0807); in test_user_operexec() 71 TEST_ASSERT_EQ(vcpu->run->s390_sieic.icptcode, ICPT_OPEREXC); in test_user_operexec() 72 TEST_ASSERT_EQ(vcpu->run->s390_sieic.ipa, 0); in test_user_operexec() 92 TEST_ASSERT_EQ(vcpu->run->s390_sieic.icptcode, ICPT_OPEREXC); in test_user_operexec_combined() 93 TEST_ASSERT_EQ(vcpu->run->s390_sieic.ipa, 0x0807); in test_user_operexec_combined() 106 TEST_ASSERT_EQ(vcpu->run->s390_sieic.icptcode, ICPT_OPEREXC); in test_user_operexec_combined() 107 TEST_ASSERT_EQ(vcpu->run->s390_sieic.ipa, 0x0807); in test_user_operexec_combined()
|
| /linux/arch/mips/kvm/ |
| H A D | emulate.c | 962 vcpu->run->exit_reason = KVM_EXIT_IRQ_WINDOW_OPEN; in kvm_mips_emul_wait() 975 struct kvm_run *run = vcpu->run; in kvm_mips_emulate_store() local 976 void *data = run->mmio.data; in kvm_mips_emulate_store() 991 run->mmio.phys_addr = kvm_mips_callbacks->gva_to_gpa( in kvm_mips_emulate_store() 993 if (run->mmio.phys_addr == KVM_INVALID_ADDR) in kvm_mips_emulate_store() 999 run->mmio.len = 8; in kvm_mips_emulate_store() 1009 run->mmio.len = 4; in kvm_mips_emulate_store() 1018 run->mmio.len = 2; in kvm_mips_emulate_store() 1027 run->mmio.len = 1; in kvm_mips_emulate_store() 1036 run->mmio.phys_addr = kvm_mips_callbacks->gva_to_gpa( in kvm_mips_emulate_store() [all …]
|
| /linux/tools/testing/selftests/kvm/x86/ |
| H A D | userspace_msr_exit_test.c | 388 if (vcpu->run->exit_reason == KVM_EXIT_IO && in check_for_guest_assert() 396 struct kvm_run *run = vcpu->run; in process_rdmsr() local 401 TEST_ASSERT(run->msr.index == msr_index, in process_rdmsr() 403 run->msr.index, msr_index); in process_rdmsr() 405 switch (run->msr.index) { in process_rdmsr() 407 run->msr.data = 0; in process_rdmsr() 410 run->msr.error = 1; in process_rdmsr() 413 run->msr.data = msr_non_existent_data; in process_rdmsr() 416 run->msr.data = MSR_FS_BASE; in process_rdmsr() 419 run->msr.data = MSR_GS_BASE; in process_rdmsr() [all …]
|
| H A D | sev_smoke_test.c | 111 TEST_ASSERT(vcpu->run->exit_reason == KVM_EXIT_SYSTEM_EVENT, in test_sync_vmsa() 113 exit_reason_str(vcpu->run->exit_reason)); in test_sync_vmsa() 114 TEST_ASSERT_EQ(vcpu->run->system_event.type, KVM_SYSTEM_EVENT_SEV_TERM); in test_sync_vmsa() 115 TEST_ASSERT_EQ(vcpu->run->system_event.ndata, 1); in test_sync_vmsa() 116 TEST_ASSERT_EQ(vcpu->run->system_event.data[0], GHCB_MSR_TERM_REQ); in test_sync_vmsa() 138 TEST_ASSERT(vcpu->run->exit_reason == KVM_EXIT_SYSTEM_EVENT, in test_sev() 140 exit_reason_str(vcpu->run->exit_reason)); in test_sev() 141 TEST_ASSERT_EQ(vcpu->run->system_event.type, KVM_SYSTEM_EVENT_SEV_TERM); in test_sev() 142 TEST_ASSERT_EQ(vcpu->run->system_event.ndata, 1); in test_sev() 143 TEST_ASSERT_EQ(vcpu->run->system_event.data[0], GHCB_MSR_TERM_REQ); in test_sev() [all …]
|
| H A D | userspace_io_test.c | 57 struct kvm_run *run; in main() local 62 run = vcpu->run; in main() 73 TEST_ASSERT(run->io.port == 0x80, in main() 74 "Expected I/O at port 0x80, got port 0x%x", run->io.port); in main() 88 memset((void *)run + run->io.data_offset, 0xaa, PAGE_SIZE); in main()
|
| H A D | private_mem_conversions_test.c | 291 struct kvm_run *run = vcpu->run; in handle_exit_hypercall() local 292 uint64_t gpa = run->hypercall.args[0]; in handle_exit_hypercall() 293 uint64_t size = run->hypercall.args[1] * PAGE_SIZE; in handle_exit_hypercall() 294 bool set_attributes = run->hypercall.args[2] & MAP_GPA_SET_ATTRIBUTES; in handle_exit_hypercall() 295 bool map_shared = run->hypercall.args[2] & MAP_GPA_SHARED; in handle_exit_hypercall() 296 bool do_fallocate = run->hypercall.args[2] & MAP_GPA_DO_FALLOCATE; in handle_exit_hypercall() 299 TEST_ASSERT(run->hypercall.nr == KVM_HC_MAP_GPA_RANGE, in handle_exit_hypercall() 301 KVM_HC_MAP_GPA_RANGE, run->hypercall.nr); in handle_exit_hypercall() 309 run->hypercall.ret = 0; in handle_exit_hypercall() 317 struct kvm_run *run = vcpu->run; in __test_mem_conversions() local [all …]
|
| /linux/tools/perf/scripts/python/ |
| H A D | stat-cpi.py | 23 def store(time, event, cpu, thread, val, ena, run): argument 29 data[key] = [ val, ena, run] 35 def stat__cycles_k(cpu, thread, time, val, ena, run): argument 36 store(time, "cycles", cpu, thread, val, ena, run); 38 def stat__instructions_k(cpu, thread, time, val, ena, run): argument 39 store(time, "instructions", cpu, thread, val, ena, run); 41 def stat__cycles_u(cpu, thread, time, val, ena, run): argument 42 store(time, "cycles", cpu, thread, val, ena, run); 44 def stat__instructions_u(cpu, thread, time, val, ena, run): argument 45 store(time, "instructions", cpu, thread, val, ena, run); [all …]
|
| /linux/fs/befs/ |
| H A D | endian.h | 74 befs_block_run run; in fsrun_to_cpu() local 77 run.allocation_group = le32_to_cpu((__force __le32)n.allocation_group); in fsrun_to_cpu() 78 run.start = le16_to_cpu((__force __le16)n.start); in fsrun_to_cpu() 79 run.len = le16_to_cpu((__force __le16)n.len); in fsrun_to_cpu() 81 run.allocation_group = be32_to_cpu((__force __be32)n.allocation_group); in fsrun_to_cpu() 82 run.start = be16_to_cpu((__force __be16)n.start); in fsrun_to_cpu() 83 run.len = be16_to_cpu((__force __be16)n.len); in fsrun_to_cpu() 85 return run; in fsrun_to_cpu() 91 befs_disk_block_run run; in cpu_to_fsrun() local 94 run.allocation_group = cpu_to_le32(n.allocation_group); in cpu_to_fsrun() [all …]
|
| H A D | datastream.c | 26 befs_blocknr_t blockno, befs_block_run *run); 31 befs_block_run *run); 36 befs_block_run *run); 53 befs_block_run run; in befs_read_datastream() local 61 if (befs_fblock2brun(sb, ds, block, &run) != BEFS_OK) { in befs_read_datastream() 67 bh = befs_bread_iaddr(sb, run); in befs_read_datastream() 96 befs_blocknr_t fblock, befs_block_run *run) in befs_fblock2brun() argument 102 err = befs_find_brun_direct(sb, data, fblock, run); in befs_fblock2brun() 105 err = befs_find_brun_indirect(sb, data, fblock, run); in befs_fblock2brun() 108 err = befs_find_brun_dblindirect(sb, data, fblock, run); in befs_fblock2brun() [all …]
|
| /linux/tools/testing/selftests/arm64/mte/ |
| H A D | check_child_memory.c | 87 int run, result; in check_child_memory_mapping() local 92 for (run = 0; run < item; run++) { in check_child_memory_mapping() 93 ptr = (char *)mte_allocate_memory_tag_range(sizes[run], mem_type, mapping, in check_child_memory_mapping() 95 if (check_allocated_memory_range(ptr, sizes[run], mem_type, in check_child_memory_mapping() 98 result = check_child_tag_inheritance(ptr, sizes[run], mode); in check_child_memory_mapping() 99 mte_free_memory_tag_range((void *)ptr, sizes[run], mem_type, UNDERFLOW, OVERFLOW); in check_child_memory_mapping() 109 int run, fd, map_size, result = KSFT_PASS; in check_child_file_mapping() local 113 for (run = 0; run < total; run++) { in check_child_file_mapping() 118 map_size = sizes[run] + OVERFLOW + UNDERFLOW; in check_child_file_mapping() 125 mte_initialize_current_context(mode, (uintptr_t)ptr, sizes[run]); in check_child_file_mapping() [all …]
|
| H A D | check_tags_inclusion.c | 52 int tag, run, ret, result = KSFT_PASS; in check_single_included_tags() local 64 for (run = 0; (run < RUNS) && (result == KSFT_PASS); run++) { in check_single_included_tags() 84 int tag, run, result = KSFT_PASS; in check_multiple_included_tags() local 96 for (run = 0; (run < RUNS) && (result == KSFT_PASS); run++) { in check_multiple_included_tags() 116 int run, ret, result = KSFT_PASS; in check_all_included_tags() local 127 for (run = 0; (run < RUNS) && (result == KSFT_PASS); run++) { in check_all_included_tags() 142 int run, ret; in check_none_included_tags() local 152 for (run = 0; run < RUNS; run++) { in check_none_included_tags()
|
| H A D | check_mmap_options.c | 117 int run, result, map_size; in check_anonymous_memory_mapping() local 124 for (run = 0; run < item; run++) { in check_anonymous_memory_mapping() 125 map_size = sizes[run] + OVERFLOW + UNDERFLOW; in check_anonymous_memory_mapping() 131 mte_initialize_current_context(mode, (uintptr_t)ptr, sizes[run]); in check_anonymous_memory_mapping() 133 ptr = mte_insert_tags((void *)ptr, sizes[run]); in check_anonymous_memory_mapping() 139 result = check_mte_memory(ptr, sizes[run], mode, tag_check, atag_check, tag_op); in check_anonymous_memory_mapping() 140 mte_clear_tags((void *)ptr, sizes[run]); in check_anonymous_memory_mapping() 152 int run, fd, map_size; in check_file_memory_mapping() local 160 for (run = 0; run < total; run++) { in check_file_memory_mapping() 165 map_size = sizes[run] + UNDERFLOW + OVERFLOW; in check_file_memory_mapping() [all …]
|
| /linux/arch/arm64/kvm/ |
| H A D | mmio.c | 124 struct kvm_run *run = vcpu->run; in kvm_handle_mmio_return() local 127 data = kvm_mmio_read_buf(run->mmio.data, len); in kvm_handle_mmio_return() 138 trace_kvm_mmio(KVM_TRACE_MMIO_READ, len, run->mmio.phys_addr, in kvm_handle_mmio_return() 155 struct kvm_run *run = vcpu->run; in io_mem_abort() local 179 run->exit_reason = KVM_EXIT_ARM_NISV; in io_mem_abort() 180 run->arm_nisv.esr_iss = kvm_vcpu_dabt_iss_nisv_sanitized(vcpu); in io_mem_abort() 181 run->arm_nisv.fault_ipa = fault_ipa; in io_mem_abort() 215 run->mmio.is_write = is_write; in io_mem_abort() 216 run->mmio.phys_addr = fault_ipa; in io_mem_abort() 217 run->mmio.len = len; in io_mem_abort() [all …]
|
| /linux/tools/testing/selftests/kvm/lib/loongarch/ |
| H A D | ucall.c | 27 struct kvm_run *run = vcpu->run; in ucall_arch_get_ucall() local 29 if (run->exit_reason == KVM_EXIT_MMIO && in ucall_arch_get_ucall() 30 run->mmio.phys_addr == vcpu->vm->ucall_mmio_addr) { in ucall_arch_get_ucall() 31 TEST_ASSERT(run->mmio.is_write && run->mmio.len == sizeof(uint64_t), in ucall_arch_get_ucall() 34 return (void *)(*((uint64_t *)run->mmio.data)); in ucall_arch_get_ucall()
|
| /linux/tools/testing/selftests/kvm/lib/riscv/ |
| H A D | ucall.c | 16 struct kvm_run *run = vcpu->run; in ucall_arch_get_ucall() local 18 if (run->exit_reason == KVM_EXIT_RISCV_SBI && in ucall_arch_get_ucall() 19 run->riscv_sbi.extension_id == KVM_RISCV_SELFTESTS_SBI_EXT) { in ucall_arch_get_ucall() 20 switch (run->riscv_sbi.function_id) { in ucall_arch_get_ucall() 22 return (void *)run->riscv_sbi.args[0]; in ucall_arch_get_ucall()
|
| /linux/drivers/staging/media/sunxi/cedrus/ |
| H A D | cedrus_h264.c | 87 struct cedrus_run *run) in cedrus_write_frame_list() argument 90 const struct v4l2_ctrl_h264_decode_params *decode = run->h264.decode_params; in cedrus_write_frame_list() 91 const struct v4l2_ctrl_h264_sps *sps = run->h264.sps; in cedrus_write_frame_list() 120 if (run->dst->vb2_buf.timestamp == dpb->reference_ts) { in cedrus_write_frame_list() 139 output_buf = vb2_to_cedrus_buffer(&run->dst->vb2_buf); in cedrus_write_frame_list() 143 const struct v4l2_ctrl_h264_sps *sps = run->h264.sps; in cedrus_write_frame_list() 190 struct cedrus_run *run, in _cedrus_write_ref_list() argument 194 const struct v4l2_ctrl_h264_decode_params *decode = run->h264.decode_params; in _cedrus_write_ref_list() 235 struct cedrus_run *run) in cedrus_write_ref_list0() argument 237 const struct v4l2_ctrl_h264_slice_params *slice = run->h264.slice_params; in cedrus_write_ref_list0() [all …]
|
| /linux/Documentation/dev-tools/kunit/ |
| H A D | start.rst | 8 teaching how to run existing tests and then how to write a simple test case, 14 build the kernel, you can run KUnit. 20 can run kunit_tool: 24 ./tools/testing/kunit/kunit.py run 28 "The source tree is not clean, please run 'make ARCH=um mrproper'" 59 Selecting which tests to run 64 you can select which tests to run by: 67 - `Filtering tests by name`_ to select specifically which compiled tests to run. 72 If you didn't run ``kunit.py run`` yet, you can generate it by running: 91 options required to run the desired tests, including their dependencies. [all …]
|