Home
last modified time | relevance | path

Searched refs:run (Results 1 – 25 of 1245) sorted by relevance

12345678910>>...50

/linux/tools/testing/selftests/sgx/
H A Dmain.c172 struct sgx_enclave_run run; in FIXTURE() local
262 #define ENCL_CALL(op, run, clobbered) \ argument
267 EENTER, 0, 0, (run)); \
270 (run)); \
274 #define EXPECT_EEXIT(run) \ argument
276 EXPECT_EQ((run)->function, EEXIT); \
277 if ((run)->function != EEXIT) \
278 TH_LOG("0x%02x 0x%02x 0x%016llx", (run)->exception_vector, \
279 (run)->exception_error_code, (run)->exception_addr); \
289 memset(&self->run, 0, sizeof(self->run)); in TEST_F()
[all …]
/linux/tools/perf/tests/
H A Dmake10 # run only specific test over 'Makefile'
127 # $(run) contains all available tests
128 run := make_pure
129 # Targets 'clean all' can be run together only through top level
133 run += make_clean_all
138 run += make_python_perf_so
139 run += make_debug
140 run += make_nondistro
141 run += make_extra_tests
142 run += make_jevents_all
[all …]
/linux/arch/riscv/kvm/
H A Dvcpu_insn.c27 int (*func)(struct kvm_vcpu *vcpu, struct kvm_run *run, ulong insn);
30 static int truly_illegal_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, in truly_illegal_insn() argument
46 static int truly_virtual_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, in truly_virtual_insn() argument
76 static int wfi_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, ulong insn) in wfi_insn() argument
83 static int wrs_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, ulong insn) in wrs_insn() argument
127 int kvm_riscv_vcpu_csr_return(struct kvm_vcpu *vcpu, struct kvm_run *run) in kvm_riscv_vcpu_csr_return() argument
139 run->riscv_csr.ret_value); in kvm_riscv_vcpu_csr_return()
147 static int csr_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, ulong insn) in csr_insn() argument
191 run->riscv_csr.csr_num = csr_num; in csr_insn()
192 run->riscv_csr.new_value = new_val; in csr_insn()
[all …]
H A Dvcpu_sbi.c128 struct kvm_run *run, in kvm_riscv_vcpu_sbi_forward_handler() argument
135 run->exit_reason = KVM_EXIT_RISCV_SBI; in kvm_riscv_vcpu_sbi_forward_handler()
136 run->riscv_sbi.extension_id = cp->a7; in kvm_riscv_vcpu_sbi_forward_handler()
137 run->riscv_sbi.function_id = cp->a6; in kvm_riscv_vcpu_sbi_forward_handler()
138 run->riscv_sbi.args[0] = cp->a0; in kvm_riscv_vcpu_sbi_forward_handler()
139 run->riscv_sbi.args[1] = cp->a1; in kvm_riscv_vcpu_sbi_forward_handler()
140 run->riscv_sbi.args[2] = cp->a2; in kvm_riscv_vcpu_sbi_forward_handler()
141 run->riscv_sbi.args[3] = cp->a3; in kvm_riscv_vcpu_sbi_forward_handler()
142 run->riscv_sbi.args[4] = cp->a4; in kvm_riscv_vcpu_sbi_forward_handler()
143 run->riscv_sbi.args[5] = cp->a5; in kvm_riscv_vcpu_sbi_forward_handler()
[all …]
/linux/tools/testing/selftests/kvm/
H A Dcoalesced_io_test.c78 struct kvm_run *run = vcpu->run; in vcpu_run_and_verify_io_exit() local
90 if (run->exit_reason == KVM_EXIT_IO) in vcpu_run_and_verify_io_exit()
91 pio_value = *(uint32_t *)((void *)run + run->io.data_offset); in vcpu_run_and_verify_io_exit()
95 TEST_ASSERT((!want_pio && (run->exit_reason == KVM_EXIT_MMIO && run->mmio.is_write && in vcpu_run_and_verify_io_exit()
96 run->mmio.phys_addr == io->mmio_gpa && run->mmio.len == 8 && in vcpu_run_and_verify_io_exit()
97 *(uint64_t *)run->mmio.data == io->mmio_gpa + io->ring_size - 1)) || in vcpu_run_and_verify_io_exit()
98 (want_pio && (run->exit_reason == KVM_EXIT_IO && run->io.port == io->pio_port && in vcpu_run_and_verify_io_exit()
99 run->io.direction == KVM_EXIT_IO_OUT && run->io.count == 1 && in vcpu_run_and_verify_io_exit()
106 (want_pio ? io->pio_port : io->mmio_gpa) + io->ring_size - 1, run->exit_reason, in vcpu_run_and_verify_io_exit()
107run->exit_reason == KVM_EXIT_MMIO ? "MMIO" : run->exit_reason == KVM_EXIT_IO ? "PIO" : "other", in vcpu_run_and_verify_io_exit()
[all …]
H A Dset_memory_region_test.c55 struct kvm_run *run = vcpu->run; in vcpu_worker() local
67 if (run->exit_reason == KVM_EXIT_IO) { in vcpu_worker()
76 if (run->exit_reason != KVM_EXIT_MMIO) in vcpu_worker()
79 TEST_ASSERT(!run->mmio.is_write, "Unexpected exit mmio write"); in vcpu_worker()
80 TEST_ASSERT(run->mmio.len == 8, in vcpu_worker()
81 "Unexpected exit mmio size = %u", run->mmio.len); in vcpu_worker()
83 TEST_ASSERT(run->mmio.phys_addr == MEM_REGION_GPA, in vcpu_worker()
85 run->mmio.phys_addr); in vcpu_worker()
86 memcpy(run->mmio.data, &MMIO_VAL, 8); in vcpu_worker()
89 if (run->exit_reason == KVM_EXIT_IO && cmd == UCALL_ABORT) in vcpu_worker()
[all …]
H A Dguest_print_test.c128 struct kvm_run *run = vcpu->run; in run_test() local
134 TEST_ASSERT(run->exit_reason == UCALL_EXIT_REASON, in run_test()
136 run->exit_reason, exit_reason_str(run->exit_reason)); in run_test()
171 struct kvm_run *run; in test_limits() local
176 run = vcpu->run; in test_limits()
179 TEST_ASSERT(run->exit_reason == UCALL_EXIT_REASON, in test_limits()
181 run->exit_reason, exit_reason_str(run->exit_reason)); in test_limits()
/linux/tools/testing/selftests/kvm/s390/
H A Ducontrol_test.c108 struct kvm_run *run; in FIXTURE() local
145 self->run = kvm_mmap(self->kvm_run_size, PROT_READ | PROT_WRITE, in FIXTURE_SETUP()
157 TH_LOG("VM created %p %p", self->run, self->sie_block); in FIXTURE_SETUP()
188 kvm_munmap(self->run, self->kvm_run_size); in FIXTURE_TEARDOWN()
306 struct kvm_run *run = self->run; in uc_handle_exit_ucontrol() local
310 TEST_ASSERT_EQ(KVM_EXIT_S390_UCONTROL, run->exit_reason); in uc_handle_exit_ucontrol()
311 switch (run->s390_ucontrol.pgm_code) { in uc_handle_exit_ucontrol()
313 seg_addr = run->s390_ucontrol.trans_exc_code & ~(SZ_1M - 1); in uc_handle_exit_ucontrol()
315 run->s390_ucontrol.trans_exc_code, seg_addr); in uc_handle_exit_ucontrol()
321 TEST_FAIL("UNEXPECTED PGM CODE %d", run->s390_ucontrol.pgm_code); in uc_handle_exit_ucontrol()
[all …]
H A Duser_operexec.c33 TEST_ASSERT_EQ(vcpu->run->s390_sieic.icptcode, ICPT_OPEREXC); in test_user_instr0()
34 TEST_ASSERT_EQ(vcpu->run->s390_sieic.ipa, 0); in test_user_instr0()
56 TEST_ASSERT_EQ(vcpu->run->s390_sieic.icptcode, ICPT_OPEREXC); in test_user_operexec()
57 TEST_ASSERT_EQ(vcpu->run->s390_sieic.ipa, 0x0807); in test_user_operexec()
71 TEST_ASSERT_EQ(vcpu->run->s390_sieic.icptcode, ICPT_OPEREXC); in test_user_operexec()
72 TEST_ASSERT_EQ(vcpu->run->s390_sieic.ipa, 0); in test_user_operexec()
92 TEST_ASSERT_EQ(vcpu->run->s390_sieic.icptcode, ICPT_OPEREXC); in test_user_operexec_combined()
93 TEST_ASSERT_EQ(vcpu->run->s390_sieic.ipa, 0x0807); in test_user_operexec_combined()
106 TEST_ASSERT_EQ(vcpu->run->s390_sieic.icptcode, ICPT_OPEREXC); in test_user_operexec_combined()
107 TEST_ASSERT_EQ(vcpu->run->s390_sieic.ipa, 0x0807); in test_user_operexec_combined()
/linux/arch/mips/kvm/
H A Demulate.c962 vcpu->run->exit_reason = KVM_EXIT_IRQ_WINDOW_OPEN; in kvm_mips_emul_wait()
975 struct kvm_run *run = vcpu->run; in kvm_mips_emulate_store() local
976 void *data = run->mmio.data; in kvm_mips_emulate_store()
991 run->mmio.phys_addr = kvm_mips_callbacks->gva_to_gpa( in kvm_mips_emulate_store()
993 if (run->mmio.phys_addr == KVM_INVALID_ADDR) in kvm_mips_emulate_store()
999 run->mmio.len = 8; in kvm_mips_emulate_store()
1009 run->mmio.len = 4; in kvm_mips_emulate_store()
1018 run->mmio.len = 2; in kvm_mips_emulate_store()
1027 run->mmio.len = 1; in kvm_mips_emulate_store()
1036 run->mmio.phys_addr = kvm_mips_callbacks->gva_to_gpa( in kvm_mips_emulate_store()
[all …]
/linux/tools/testing/selftests/kvm/x86/
H A Duserspace_msr_exit_test.c388 if (vcpu->run->exit_reason == KVM_EXIT_IO && in check_for_guest_assert()
396 struct kvm_run *run = vcpu->run; in process_rdmsr() local
401 TEST_ASSERT(run->msr.index == msr_index, in process_rdmsr()
403 run->msr.index, msr_index); in process_rdmsr()
405 switch (run->msr.index) { in process_rdmsr()
407 run->msr.data = 0; in process_rdmsr()
410 run->msr.error = 1; in process_rdmsr()
413 run->msr.data = msr_non_existent_data; in process_rdmsr()
416 run->msr.data = MSR_FS_BASE; in process_rdmsr()
419 run->msr.data = MSR_GS_BASE; in process_rdmsr()
[all …]
H A Dsev_smoke_test.c111 TEST_ASSERT(vcpu->run->exit_reason == KVM_EXIT_SYSTEM_EVENT, in test_sync_vmsa()
113 exit_reason_str(vcpu->run->exit_reason)); in test_sync_vmsa()
114 TEST_ASSERT_EQ(vcpu->run->system_event.type, KVM_SYSTEM_EVENT_SEV_TERM); in test_sync_vmsa()
115 TEST_ASSERT_EQ(vcpu->run->system_event.ndata, 1); in test_sync_vmsa()
116 TEST_ASSERT_EQ(vcpu->run->system_event.data[0], GHCB_MSR_TERM_REQ); in test_sync_vmsa()
138 TEST_ASSERT(vcpu->run->exit_reason == KVM_EXIT_SYSTEM_EVENT, in test_sev()
140 exit_reason_str(vcpu->run->exit_reason)); in test_sev()
141 TEST_ASSERT_EQ(vcpu->run->system_event.type, KVM_SYSTEM_EVENT_SEV_TERM); in test_sev()
142 TEST_ASSERT_EQ(vcpu->run->system_event.ndata, 1); in test_sev()
143 TEST_ASSERT_EQ(vcpu->run->system_event.data[0], GHCB_MSR_TERM_REQ); in test_sev()
[all …]
H A Duserspace_io_test.c57 struct kvm_run *run; in main() local
62 run = vcpu->run; in main()
73 TEST_ASSERT(run->io.port == 0x80, in main()
74 "Expected I/O at port 0x80, got port 0x%x", run->io.port); in main()
88 memset((void *)run + run->io.data_offset, 0xaa, PAGE_SIZE); in main()
H A Dprivate_mem_conversions_test.c291 struct kvm_run *run = vcpu->run; in handle_exit_hypercall() local
292 uint64_t gpa = run->hypercall.args[0]; in handle_exit_hypercall()
293 uint64_t size = run->hypercall.args[1] * PAGE_SIZE; in handle_exit_hypercall()
294 bool set_attributes = run->hypercall.args[2] & MAP_GPA_SET_ATTRIBUTES; in handle_exit_hypercall()
295 bool map_shared = run->hypercall.args[2] & MAP_GPA_SHARED; in handle_exit_hypercall()
296 bool do_fallocate = run->hypercall.args[2] & MAP_GPA_DO_FALLOCATE; in handle_exit_hypercall()
299 TEST_ASSERT(run->hypercall.nr == KVM_HC_MAP_GPA_RANGE, in handle_exit_hypercall()
301 KVM_HC_MAP_GPA_RANGE, run->hypercall.nr); in handle_exit_hypercall()
309 run->hypercall.ret = 0; in handle_exit_hypercall()
317 struct kvm_run *run = vcpu->run; in __test_mem_conversions() local
[all …]
/linux/tools/perf/scripts/python/
H A Dstat-cpi.py23 def store(time, event, cpu, thread, val, ena, run): argument
29 data[key] = [ val, ena, run]
35 def stat__cycles_k(cpu, thread, time, val, ena, run): argument
36 store(time, "cycles", cpu, thread, val, ena, run);
38 def stat__instructions_k(cpu, thread, time, val, ena, run): argument
39 store(time, "instructions", cpu, thread, val, ena, run);
41 def stat__cycles_u(cpu, thread, time, val, ena, run): argument
42 store(time, "cycles", cpu, thread, val, ena, run);
44 def stat__instructions_u(cpu, thread, time, val, ena, run): argument
45 store(time, "instructions", cpu, thread, val, ena, run);
[all …]
/linux/fs/befs/
H A Dendian.h74 befs_block_run run; in fsrun_to_cpu() local
77 run.allocation_group = le32_to_cpu((__force __le32)n.allocation_group); in fsrun_to_cpu()
78 run.start = le16_to_cpu((__force __le16)n.start); in fsrun_to_cpu()
79 run.len = le16_to_cpu((__force __le16)n.len); in fsrun_to_cpu()
81 run.allocation_group = be32_to_cpu((__force __be32)n.allocation_group); in fsrun_to_cpu()
82 run.start = be16_to_cpu((__force __be16)n.start); in fsrun_to_cpu()
83 run.len = be16_to_cpu((__force __be16)n.len); in fsrun_to_cpu()
85 return run; in fsrun_to_cpu()
91 befs_disk_block_run run; in cpu_to_fsrun() local
94 run.allocation_group = cpu_to_le32(n.allocation_group); in cpu_to_fsrun()
[all …]
H A Ddatastream.c26 befs_blocknr_t blockno, befs_block_run *run);
31 befs_block_run *run);
36 befs_block_run *run);
53 befs_block_run run; in befs_read_datastream() local
61 if (befs_fblock2brun(sb, ds, block, &run) != BEFS_OK) { in befs_read_datastream()
67 bh = befs_bread_iaddr(sb, run); in befs_read_datastream()
96 befs_blocknr_t fblock, befs_block_run *run) in befs_fblock2brun() argument
102 err = befs_find_brun_direct(sb, data, fblock, run); in befs_fblock2brun()
105 err = befs_find_brun_indirect(sb, data, fblock, run); in befs_fblock2brun()
108 err = befs_find_brun_dblindirect(sb, data, fblock, run); in befs_fblock2brun()
[all …]
/linux/tools/testing/selftests/arm64/mte/
H A Dcheck_child_memory.c87 int run, result; in check_child_memory_mapping() local
92 for (run = 0; run < item; run++) { in check_child_memory_mapping()
93 ptr = (char *)mte_allocate_memory_tag_range(sizes[run], mem_type, mapping, in check_child_memory_mapping()
95 if (check_allocated_memory_range(ptr, sizes[run], mem_type, in check_child_memory_mapping()
98 result = check_child_tag_inheritance(ptr, sizes[run], mode); in check_child_memory_mapping()
99 mte_free_memory_tag_range((void *)ptr, sizes[run], mem_type, UNDERFLOW, OVERFLOW); in check_child_memory_mapping()
109 int run, fd, map_size, result = KSFT_PASS; in check_child_file_mapping() local
113 for (run = 0; run < total; run++) { in check_child_file_mapping()
118 map_size = sizes[run] + OVERFLOW + UNDERFLOW; in check_child_file_mapping()
125 mte_initialize_current_context(mode, (uintptr_t)ptr, sizes[run]); in check_child_file_mapping()
[all …]
H A Dcheck_tags_inclusion.c52 int tag, run, ret, result = KSFT_PASS; in check_single_included_tags() local
64 for (run = 0; (run < RUNS) && (result == KSFT_PASS); run++) { in check_single_included_tags()
84 int tag, run, result = KSFT_PASS; in check_multiple_included_tags() local
96 for (run = 0; (run < RUNS) && (result == KSFT_PASS); run++) { in check_multiple_included_tags()
116 int run, ret, result = KSFT_PASS; in check_all_included_tags() local
127 for (run = 0; (run < RUNS) && (result == KSFT_PASS); run++) { in check_all_included_tags()
142 int run, ret; in check_none_included_tags() local
152 for (run = 0; run < RUNS; run++) { in check_none_included_tags()
H A Dcheck_mmap_options.c117 int run, result, map_size; in check_anonymous_memory_mapping() local
124 for (run = 0; run < item; run++) { in check_anonymous_memory_mapping()
125 map_size = sizes[run] + OVERFLOW + UNDERFLOW; in check_anonymous_memory_mapping()
131 mte_initialize_current_context(mode, (uintptr_t)ptr, sizes[run]); in check_anonymous_memory_mapping()
133 ptr = mte_insert_tags((void *)ptr, sizes[run]); in check_anonymous_memory_mapping()
139 result = check_mte_memory(ptr, sizes[run], mode, tag_check, atag_check, tag_op); in check_anonymous_memory_mapping()
140 mte_clear_tags((void *)ptr, sizes[run]); in check_anonymous_memory_mapping()
152 int run, fd, map_size; in check_file_memory_mapping() local
160 for (run = 0; run < total; run++) { in check_file_memory_mapping()
165 map_size = sizes[run] + UNDERFLOW + OVERFLOW; in check_file_memory_mapping()
[all …]
/linux/arch/arm64/kvm/
H A Dmmio.c124 struct kvm_run *run = vcpu->run; in kvm_handle_mmio_return() local
127 data = kvm_mmio_read_buf(run->mmio.data, len); in kvm_handle_mmio_return()
138 trace_kvm_mmio(KVM_TRACE_MMIO_READ, len, run->mmio.phys_addr, in kvm_handle_mmio_return()
155 struct kvm_run *run = vcpu->run; in io_mem_abort() local
179 run->exit_reason = KVM_EXIT_ARM_NISV; in io_mem_abort()
180 run->arm_nisv.esr_iss = kvm_vcpu_dabt_iss_nisv_sanitized(vcpu); in io_mem_abort()
181 run->arm_nisv.fault_ipa = fault_ipa; in io_mem_abort()
215 run->mmio.is_write = is_write; in io_mem_abort()
216 run->mmio.phys_addr = fault_ipa; in io_mem_abort()
217 run->mmio.len = len; in io_mem_abort()
[all …]
/linux/tools/testing/selftests/kvm/lib/loongarch/
H A Ducall.c27 struct kvm_run *run = vcpu->run; in ucall_arch_get_ucall() local
29 if (run->exit_reason == KVM_EXIT_MMIO && in ucall_arch_get_ucall()
30 run->mmio.phys_addr == vcpu->vm->ucall_mmio_addr) { in ucall_arch_get_ucall()
31 TEST_ASSERT(run->mmio.is_write && run->mmio.len == sizeof(uint64_t), in ucall_arch_get_ucall()
34 return (void *)(*((uint64_t *)run->mmio.data)); in ucall_arch_get_ucall()
/linux/tools/testing/selftests/kvm/lib/riscv/
H A Ducall.c16 struct kvm_run *run = vcpu->run; in ucall_arch_get_ucall() local
18 if (run->exit_reason == KVM_EXIT_RISCV_SBI && in ucall_arch_get_ucall()
19 run->riscv_sbi.extension_id == KVM_RISCV_SELFTESTS_SBI_EXT) { in ucall_arch_get_ucall()
20 switch (run->riscv_sbi.function_id) { in ucall_arch_get_ucall()
22 return (void *)run->riscv_sbi.args[0]; in ucall_arch_get_ucall()
/linux/drivers/staging/media/sunxi/cedrus/
H A Dcedrus_h264.c87 struct cedrus_run *run) in cedrus_write_frame_list() argument
90 const struct v4l2_ctrl_h264_decode_params *decode = run->h264.decode_params; in cedrus_write_frame_list()
91 const struct v4l2_ctrl_h264_sps *sps = run->h264.sps; in cedrus_write_frame_list()
120 if (run->dst->vb2_buf.timestamp == dpb->reference_ts) { in cedrus_write_frame_list()
139 output_buf = vb2_to_cedrus_buffer(&run->dst->vb2_buf); in cedrus_write_frame_list()
143 const struct v4l2_ctrl_h264_sps *sps = run->h264.sps; in cedrus_write_frame_list()
190 struct cedrus_run *run, in _cedrus_write_ref_list() argument
194 const struct v4l2_ctrl_h264_decode_params *decode = run->h264.decode_params; in _cedrus_write_ref_list()
235 struct cedrus_run *run) in cedrus_write_ref_list0() argument
237 const struct v4l2_ctrl_h264_slice_params *slice = run->h264.slice_params; in cedrus_write_ref_list0()
[all …]
/linux/Documentation/dev-tools/kunit/
H A Dstart.rst8 teaching how to run existing tests and then how to write a simple test case,
14 build the kernel, you can run KUnit.
20 can run kunit_tool:
24 ./tools/testing/kunit/kunit.py run
28 "The source tree is not clean, please run 'make ARCH=um mrproper'"
59 Selecting which tests to run
64 you can select which tests to run by:
67 - `Filtering tests by name`_ to select specifically which compiled tests to run.
72 If you didn't run ``kunit.py run`` yet, you can generate it by running:
91 options required to run the desired tests, including their dependencies.
[all …]

12345678910>>...50