/Linux-v5.15/tools/testing/selftests/kvm/s390x/ |
D | sync_regs_test.c | 80 struct kvm_run *run; in main() local 97 run = vcpu_state(vm, VCPU_ID); in main() 100 run->kvm_valid_regs = INVALID_SYNC_FIELD; in main() 107 run->kvm_valid_regs = INVALID_SYNC_FIELD | TEST_SYNC_FIELDS; in main() 115 run->kvm_dirty_regs = INVALID_SYNC_FIELD; in main() 122 run->kvm_dirty_regs = INVALID_SYNC_FIELD | TEST_SYNC_FIELDS; in main() 130 run->kvm_valid_regs = TEST_SYNC_FIELDS; in main() 133 TEST_ASSERT(run->exit_reason == KVM_EXIT_S390_SIEIC, in main() 135 run->exit_reason, in main() 136 exit_reason_str(run->exit_reason)); in main() [all …]
|
/Linux-v5.15/tools/testing/selftests/kvm/x86_64/ |
D | sync_regs_test.c | 88 struct kvm_run *run; in main() local 110 run = vcpu_state(vm, VCPU_ID); in main() 113 run->kvm_valid_regs = INVALID_SYNC_FIELD; in main() 120 run->kvm_valid_regs = INVALID_SYNC_FIELD | TEST_SYNC_FIELDS; in main() 128 run->kvm_dirty_regs = INVALID_SYNC_FIELD; in main() 135 run->kvm_dirty_regs = INVALID_SYNC_FIELD | TEST_SYNC_FIELDS; in main() 144 run->kvm_valid_regs = TEST_SYNC_FIELDS; in main() 146 TEST_ASSERT(run->exit_reason == KVM_EXIT_IO, in main() 148 run->exit_reason, in main() 149 exit_reason_str(run->exit_reason)); in main() [all …]
|
D | debug_regs.c | 84 struct kvm_run *run; in main() local 105 run = vcpu_state(vm, VCPU_ID); in main() 112 TEST_ASSERT(run->exit_reason == KVM_EXIT_DEBUG && in main() 113 run->debug.arch.exception == BP_VECTOR && in main() 114 run->debug.arch.pc == CAST_TO_RIP(sw_bp), in main() 116 run->exit_reason, run->debug.arch.exception, in main() 117 run->debug.arch.pc, CAST_TO_RIP(sw_bp)); in main() 129 TEST_ASSERT(run->exit_reason == KVM_EXIT_DEBUG && in main() 130 run->debug.arch.exception == DB_VECTOR && in main() 131 run->debug.arch.pc == CAST_TO_RIP(hw_bp) && in main() [all …]
|
D | userspace_msr_exit_test.c | 408 struct kvm_run *run = vcpu_state(vm, VCPU_ID); in check_for_guest_assert() local 411 if (run->exit_reason == KVM_EXIT_IO && in check_for_guest_assert() 420 struct kvm_run *run = vcpu_state(vm, VCPU_ID); in process_rdmsr() local 424 TEST_ASSERT(run->exit_reason == KVM_EXIT_X86_RDMSR, in process_rdmsr() 426 run->exit_reason, in process_rdmsr() 427 exit_reason_str(run->exit_reason)); in process_rdmsr() 428 TEST_ASSERT(run->msr.index == msr_index, in process_rdmsr() 430 run->msr.index, msr_index); in process_rdmsr() 432 switch (run->msr.index) { in process_rdmsr() 434 run->msr.data = 0; in process_rdmsr() [all …]
|
D | emulator_error_test.c | 62 struct kvm_run *run = vcpu_state(vm, VCPU_ID); in process_exit_on_emulation_error() local 68 TEST_ASSERT(run->exit_reason == KVM_EXIT_INTERNAL_ERROR, in process_exit_on_emulation_error() 70 run->exit_reason, in process_exit_on_emulation_error() 71 exit_reason_str(run->exit_reason)); in process_exit_on_emulation_error() 73 TEST_ASSERT(run->emulation_failure.suberror == KVM_INTERNAL_ERROR_EMULATION, in process_exit_on_emulation_error() 75 run->emulation_failure.suberror); in process_exit_on_emulation_error() 77 if (run->emulation_failure.ndata >= 1) { in process_exit_on_emulation_error() 78 flags = run->emulation_failure.flags; in process_exit_on_emulation_error() 80 run->emulation_failure.ndata >= 3) { in process_exit_on_emulation_error() 81 insn_size = run->emulation_failure.insn_size; in process_exit_on_emulation_error() [all …]
|
D | xen_vmcall_test.c | 109 volatile struct kvm_run *run = vcpu_state(vm, VCPU_ID); in main() local 114 if (run->exit_reason == KVM_EXIT_XEN) { in main() 115 ASSERT_EQ(run->xen.type, KVM_EXIT_XEN_HCALL); in main() 116 ASSERT_EQ(run->xen.u.hcall.cpl, 0); in main() 117 ASSERT_EQ(run->xen.u.hcall.longmode, 1); in main() 118 ASSERT_EQ(run->xen.u.hcall.input, INPUTVALUE); in main() 119 ASSERT_EQ(run->xen.u.hcall.params[0], ARGVALUE(1)); in main() 120 ASSERT_EQ(run->xen.u.hcall.params[1], ARGVALUE(2)); in main() 121 ASSERT_EQ(run->xen.u.hcall.params[2], ARGVALUE(3)); in main() 122 ASSERT_EQ(run->xen.u.hcall.params[3], ARGVALUE(4)); in main() [all …]
|
D | mmu_role_test.c | 28 struct kvm_run *run; in mmu_role_test() local 35 run = vcpu_state(vm, VCPU_ID); in mmu_role_test() 44 TEST_ASSERT(run->exit_reason == KVM_EXIT_MMIO, in mmu_role_test() 46 run->exit_reason, exit_reason_str(run->exit_reason)); in mmu_role_test() 48 TEST_ASSERT(run->mmio.len == 8, "Unexpected exit mmio size = %u", run->mmio.len); in mmu_role_test() 50 TEST_ASSERT(run->mmio.phys_addr == MMIO_GPA, in mmu_role_test() 51 "Unexpected exit mmio address = 0x%llx", run->mmio.phys_addr); in mmu_role_test() 82 exit_reason_str(run->exit_reason), cmd); in mmu_role_test()
|
D | mmio_warning_test.c | 36 struct kvm_run *run; member 44 struct kvm_run *run = tc->run; in thr() local 48 res, run->exit_reason, run->internal.suberror); in thr() 57 struct kvm_run *run; in test() local 66 run = (struct kvm_run *)mmap(0, 4096, PROT_READ|PROT_WRITE, MAP_SHARED, in test() 69 tc.run = run; in test()
|
D | vmx_apic_access_test.c | 103 volatile struct kvm_run *run = vcpu_state(vm, VCPU_ID); in main() local 108 TEST_ASSERT(run->exit_reason == in main() 111 run->exit_reason, in main() 112 exit_reason_str(run->exit_reason)); in main() 113 TEST_ASSERT(run->internal.suberror == in main() 116 run->internal.suberror); in main() 119 TEST_ASSERT(run->exit_reason == KVM_EXIT_IO, in main() 121 run->exit_reason, in main() 122 exit_reason_str(run->exit_reason)); in main()
|
/Linux-v5.15/drivers/staging/media/sunxi/cedrus/ |
D | cedrus_dec.c | 29 struct cedrus_run run = {}; in cedrus_device_run() local 32 run.src = v4l2_m2m_next_src_buf(ctx->fh.m2m_ctx); in cedrus_device_run() 33 run.dst = v4l2_m2m_next_dst_buf(ctx->fh.m2m_ctx); in cedrus_device_run() 36 src_req = run.src->vb2_buf.req_obj.req; in cedrus_device_run() 43 run.mpeg2.sequence = cedrus_find_control_data(ctx, in cedrus_device_run() 45 run.mpeg2.picture = cedrus_find_control_data(ctx, in cedrus_device_run() 47 run.mpeg2.quantisation = cedrus_find_control_data(ctx, in cedrus_device_run() 52 run.h264.decode_params = cedrus_find_control_data(ctx, in cedrus_device_run() 54 run.h264.pps = cedrus_find_control_data(ctx, in cedrus_device_run() 56 run.h264.scaling_matrix = cedrus_find_control_data(ctx, in cedrus_device_run() [all …]
|
/Linux-v5.15/fs/ntfs3/ |
D | run.c | 34 bool run_lookup(const struct runs_tree *run, CLST vcn, size_t *index) in run_lookup() argument 39 if (!run->count) { in run_lookup() 45 max_idx = run->count - 1; in run_lookup() 48 r = run->runs; in run_lookup() 61 *index = run->count; in run_lookup() 72 r = run->runs + mid_idx; in run_lookup() 93 static void run_consolidate(struct runs_tree *run, size_t index) in run_consolidate() argument 96 struct ntfs_run *r = run->runs + index; in run_consolidate() 98 while (index + 1 < run->count) { in run_consolidate() 154 i = run->count - (index + 1); in run_consolidate() [all …]
|
D | attrib.c | 88 struct runs_tree *run, const CLST *vcn) in attr_load_runs() argument 96 if (svcn >= evcn + 1 || run_is_mapped_full(run, svcn, evcn)) in attr_load_runs() 104 err = run_unpack_ex(run, ni->mi.sbi, ni->mi.rno, svcn, evcn, in attr_load_runs() 116 static int run_deallocate_ex(struct ntfs_sb_info *sbi, struct runs_tree *run, in run_deallocate_ex() argument 126 if (!run_lookup_entry(run, vcn, &lcn, &clen, &idx)) { in run_deallocate_ex() 128 run_truncate(run, vcn0); in run_deallocate_ex() 152 if (!run_get_entry(run, ++idx, &vcn, &lcn, &clen) || in run_deallocate_ex() 169 int attr_allocate_clusters(struct ntfs_sb_info *sbi, struct runs_tree *run, in attr_allocate_clusters() argument 177 size_t cnt = run->count; in attr_allocate_clusters() 197 if (!run_add_entry(run, vcn, lcn, flen, opt == ALLOCATE_MFT)) { in attr_allocate_clusters() [all …]
|
/Linux-v5.15/tools/testing/selftests/sgx/ |
D | main.c | 112 struct sgx_enclave_run run; in FIXTURE() local 169 memset(&self->run, 0, sizeof(self->run)); in FIXTURE_SETUP() 170 self->run.tcs = self->encl.encl_base; in FIXTURE_SETUP() 196 #define ENCL_CALL(op, run, clobbered) \ argument 201 EENTER, 0, 0, (run)); \ 204 (run)); \ 208 #define EXPECT_EEXIT(run) \ argument 210 EXPECT_EQ((run)->function, EEXIT); \ 211 if ((run)->function != EEXIT) \ 212 TH_LOG("0x%02x 0x%02x 0x%016llx", (run)->exception_vector, \ [all …]
|
/Linux-v5.15/tools/perf/tests/ |
D | make | 10 # run only specific test over 'Makefile' 123 # $(run) contains all available tests 124 run := make_pure 125 # Targets 'clean all' can be run together only through top level 129 run += make_clean_all 134 run += make_python_perf_so 135 run += make_debug 136 run += make_no_libperl 137 run += make_no_libpython 138 run += make_no_scripts [all …]
|
/Linux-v5.15/tools/testing/selftests/arm64/mte/ |
D | check_mmap_options.c | 63 int run, result, map_size; in check_anonymous_memory_mapping() local 68 for (run = 0; run < item; run++) { in check_anonymous_memory_mapping() 69 map_size = sizes[run] + OVERFLOW + UNDERFLOW; in check_anonymous_memory_mapping() 75 mte_initialize_current_context(mode, (uintptr_t)ptr, sizes[run]); in check_anonymous_memory_mapping() 77 ptr = mte_insert_tags((void *)ptr, sizes[run]); in check_anonymous_memory_mapping() 83 result = check_mte_memory(ptr, sizes[run], mode, tag_check); in check_anonymous_memory_mapping() 84 mte_clear_tags((void *)ptr, sizes[run]); in check_anonymous_memory_mapping() 95 int run, fd, map_size; in check_file_memory_mapping() local 100 for (run = 0; run < total; run++) { in check_file_memory_mapping() 105 map_size = sizes[run] + UNDERFLOW + OVERFLOW; in check_file_memory_mapping() [all …]
|
D | check_tags_inclusion.c | 46 int tag, run, result = KSFT_PASS; in check_single_included_tags() local 56 for (run = 0; (run < RUNS) && (result == KSFT_PASS); run++) { in check_single_included_tags() 76 int tag, run, result = KSFT_PASS; in check_multiple_included_tags() local 88 for (run = 0; (run < RUNS) && (result == KSFT_PASS); run++) { in check_multiple_included_tags() 108 int run, result = KSFT_PASS; in check_all_included_tags() local 117 for (run = 0; (run < RUNS) && (result == KSFT_PASS); run++) { in check_all_included_tags() 132 int run; in check_none_included_tags() local 140 for (run = 0; run < RUNS; run++) { in check_none_included_tags()
|
D | check_child_memory.c | 87 int run, result; in check_child_memory_mapping() local 92 for (run = 0; run < item; run++) { in check_child_memory_mapping() 93 ptr = (char *)mte_allocate_memory_tag_range(sizes[run], mem_type, mapping, in check_child_memory_mapping() 95 if (check_allocated_memory_range(ptr, sizes[run], mem_type, in check_child_memory_mapping() 98 result = check_child_tag_inheritance(ptr, sizes[run], mode); in check_child_memory_mapping() 99 mte_free_memory_tag_range((void *)ptr, sizes[run], mem_type, UNDERFLOW, OVERFLOW); in check_child_memory_mapping() 109 int run, fd, map_size, result = KSFT_PASS; in check_child_file_mapping() local 113 for (run = 0; run < total; run++) { in check_child_file_mapping() 118 map_size = sizes[run] + OVERFLOW + UNDERFLOW; in check_child_file_mapping() 125 mte_initialize_current_context(mode, (uintptr_t)ptr, sizes[run]); in check_child_file_mapping() [all …]
|
/Linux-v5.15/arch/s390/kvm/ |
D | diag.c | 25 start = vcpu->run->s.regs.gprs[(vcpu->arch.sie_block->ipa & 0xf0) >> 4]; in diag_release_pages() 26 end = vcpu->run->s.regs.gprs[vcpu->arch.sie_block->ipa & 0xf] + PAGE_SIZE; in diag_release_pages() 76 vcpu->run->s.regs.gprs[rx]); in __diag_page_ref_service() 78 if (vcpu->run->s.regs.gprs[rx] & 7) in __diag_page_ref_service() 80 rc = read_guest(vcpu, vcpu->run->s.regs.gprs[rx], rx, &parm, sizeof(parm)); in __diag_page_ref_service() 97 vcpu->run->s.regs.gprs[ry] = 8; in __diag_page_ref_service() 111 vcpu->run->s.regs.gprs[ry] = 0; in __diag_page_ref_service() 125 vcpu->run->s.regs.gprs[ry] = 0; in __diag_page_ref_service() 131 vcpu->run->s.regs.gprs[ry] = 4; in __diag_page_ref_service() 171 tid = vcpu->run->s.regs.gprs[(vcpu->arch.sie_block->ipa & 0xf0) >> 4]; in __diag_time_slice_end_directed() [all …]
|
/Linux-v5.15/arch/mips/kvm/ |
D | emulate.c | 963 vcpu->run->exit_reason = KVM_EXIT_IRQ_WINDOW_OPEN; in kvm_mips_emul_wait() 977 struct kvm_run *run = vcpu->run; in kvm_mips_emulate_store() local 978 void *data = run->mmio.data; in kvm_mips_emulate_store() 993 run->mmio.phys_addr = kvm_mips_callbacks->gva_to_gpa( in kvm_mips_emulate_store() 995 if (run->mmio.phys_addr == KVM_INVALID_ADDR) in kvm_mips_emulate_store() 1001 run->mmio.len = 8; in kvm_mips_emulate_store() 1011 run->mmio.len = 4; in kvm_mips_emulate_store() 1020 run->mmio.len = 2; in kvm_mips_emulate_store() 1029 run->mmio.len = 1; in kvm_mips_emulate_store() 1038 run->mmio.phys_addr = kvm_mips_callbacks->gva_to_gpa( in kvm_mips_emulate_store() [all …]
|
/Linux-v5.15/tools/perf/scripts/python/ |
D | stat-cpi.py | 23 def store(time, event, cpu, thread, val, ena, run): argument 29 data[key] = [ val, ena, run] 35 def stat__cycles_k(cpu, thread, time, val, ena, run): argument 36 store(time, "cycles", cpu, thread, val, ena, run); 38 def stat__instructions_k(cpu, thread, time, val, ena, run): argument 39 store(time, "instructions", cpu, thread, val, ena, run); 41 def stat__cycles_u(cpu, thread, time, val, ena, run): argument 42 store(time, "cycles", cpu, thread, val, ena, run); 44 def stat__instructions_u(cpu, thread, time, val, ena, run): argument 45 store(time, "instructions", cpu, thread, val, ena, run); [all …]
|
/Linux-v5.15/fs/befs/ |
D | endian.h | 74 befs_block_run run; in fsrun_to_cpu() local 77 run.allocation_group = le32_to_cpu((__force __le32)n.allocation_group); in fsrun_to_cpu() 78 run.start = le16_to_cpu((__force __le16)n.start); in fsrun_to_cpu() 79 run.len = le16_to_cpu((__force __le16)n.len); in fsrun_to_cpu() 81 run.allocation_group = be32_to_cpu((__force __be32)n.allocation_group); in fsrun_to_cpu() 82 run.start = be16_to_cpu((__force __be16)n.start); in fsrun_to_cpu() 83 run.len = be16_to_cpu((__force __be16)n.len); in fsrun_to_cpu() 85 return run; in fsrun_to_cpu() 91 befs_disk_block_run run; in cpu_to_fsrun() local 94 run.allocation_group = cpu_to_le32(n.allocation_group); in cpu_to_fsrun() [all …]
|
D | datastream.c | 26 befs_blocknr_t blockno, befs_block_run *run); 31 befs_block_run *run); 36 befs_block_run *run); 53 befs_block_run run; in befs_read_datastream() local 61 if (befs_fblock2brun(sb, ds, block, &run) != BEFS_OK) { in befs_read_datastream() 67 bh = befs_bread_iaddr(sb, run); in befs_read_datastream() 96 befs_blocknr_t fblock, befs_block_run *run) in befs_fblock2brun() argument 102 err = befs_find_brun_direct(sb, data, fblock, run); in befs_fblock2brun() 105 err = befs_find_brun_indirect(sb, data, fblock, run); in befs_fblock2brun() 108 err = befs_find_brun_dblindirect(sb, data, fblock, run); in befs_fblock2brun() [all …]
|
/Linux-v5.15/arch/arm64/kvm/ |
D | mmio.c | 94 struct kvm_run *run = vcpu->run; in kvm_handle_mmio_return() local 97 data = kvm_mmio_read_buf(run->mmio.data, len); in kvm_handle_mmio_return() 108 trace_kvm_mmio(KVM_TRACE_MMIO_READ, len, run->mmio.phys_addr, in kvm_handle_mmio_return() 125 struct kvm_run *run = vcpu->run; in io_mem_abort() local 139 run->exit_reason = KVM_EXIT_ARM_NISV; in io_mem_abort() 140 run->arm_nisv.esr_iss = kvm_vcpu_dabt_iss_nisv_sanitized(vcpu); in io_mem_abort() 141 run->arm_nisv.fault_ipa = fault_ipa; in io_mem_abort() 176 run->mmio.is_write = is_write; in io_mem_abort() 177 run->mmio.phys_addr = fault_ipa; in io_mem_abort() 178 run->mmio.len = len; in io_mem_abort() [all …]
|
/Linux-v5.15/tools/testing/selftests/kvm/lib/s390x/ |
D | diag318_test_handler.c | 31 struct kvm_run *run; in diag318_handler() local 37 run = vcpu_state(vm, VCPU_ID); in diag318_handler() 39 TEST_ASSERT(run->exit_reason == KVM_EXIT_S390_SIEIC, in diag318_handler() 41 TEST_ASSERT(run->s390_sieic.icptcode == ICPT_INSTRUCTION, in diag318_handler() 42 "Unexpected intercept code: 0x%x", run->s390_sieic.icptcode); in diag318_handler() 43 TEST_ASSERT((run->s390_sieic.ipa & 0xff00) == IPA0_DIAG, in diag318_handler() 44 "Unexpected IPA0 code: 0x%x", (run->s390_sieic.ipa & 0xff00)); in diag318_handler() 46 reg = (run->s390_sieic.ipa & 0x00f0) >> 4; in diag318_handler() 47 diag318_info = run->s.regs.gprs[reg]; in diag318_handler()
|
/Linux-v5.15/tools/testing/selftests/kvm/ |
D | set_memory_region_test.c | 58 struct kvm_run *run; in vcpu_worker() local 67 run = vcpu_state(vm, VCPU_ID); in vcpu_worker() 72 if (run->exit_reason == KVM_EXIT_IO) { in vcpu_worker() 81 if (run->exit_reason != KVM_EXIT_MMIO) in vcpu_worker() 84 TEST_ASSERT(!run->mmio.is_write, "Unexpected exit mmio write"); in vcpu_worker() 85 TEST_ASSERT(run->mmio.len == 8, in vcpu_worker() 86 "Unexpected exit mmio size = %u", run->mmio.len); in vcpu_worker() 88 TEST_ASSERT(run->mmio.phys_addr == MEM_REGION_GPA, in vcpu_worker() 90 run->mmio.phys_addr); in vcpu_worker() 91 memcpy(run->mmio.data, &MMIO_VAL, 8); in vcpu_worker() [all …]
|