/Linux-v5.10/tools/testing/selftests/kvm/x86_64/ |
D | sync_regs_test.c | 83 struct kvm_run *run; in main() local 105 run = vcpu_state(vm, VCPU_ID); in main() 108 run->kvm_valid_regs = INVALID_SYNC_FIELD; in main() 115 run->kvm_valid_regs = INVALID_SYNC_FIELD | TEST_SYNC_FIELDS; in main() 123 run->kvm_dirty_regs = INVALID_SYNC_FIELD; in main() 130 run->kvm_dirty_regs = INVALID_SYNC_FIELD | TEST_SYNC_FIELDS; in main() 139 run->kvm_valid_regs = TEST_SYNC_FIELDS; in main() 141 TEST_ASSERT(run->exit_reason == KVM_EXIT_IO, in main() 143 run->exit_reason, in main() 144 exit_reason_str(run->exit_reason)); in main() [all …]
|
D | debug_regs.c | 69 struct kvm_run *run; in main() local 89 run = vcpu_state(vm, VCPU_ID); in main() 96 TEST_ASSERT(run->exit_reason == KVM_EXIT_DEBUG && in main() 97 run->debug.arch.exception == BP_VECTOR && in main() 98 run->debug.arch.pc == CAST_TO_RIP(sw_bp), in main() 100 run->exit_reason, run->debug.arch.exception, in main() 101 run->debug.arch.pc, CAST_TO_RIP(sw_bp)); in main() 113 TEST_ASSERT(run->exit_reason == KVM_EXIT_DEBUG && in main() 114 run->debug.arch.exception == DB_VECTOR && in main() 115 run->debug.arch.pc == CAST_TO_RIP(hw_bp) && in main() [all …]
|
D | user_msr_test.c | 154 static void handle_rdmsr(struct kvm_run *run) in handle_rdmsr() argument 156 run->msr.data = run->msr.index; in handle_rdmsr() 159 if (run->msr.index == MSR_SYSCALL_MASK || in handle_rdmsr() 160 run->msr.index == MSR_GS_BASE) { in handle_rdmsr() 161 TEST_ASSERT(run->msr.reason == KVM_MSR_EXIT_REASON_FILTER, in handle_rdmsr() 165 if (run->msr.index == 0xdeadbeef) { in handle_rdmsr() 166 TEST_ASSERT(run->msr.reason == KVM_MSR_EXIT_REASON_UNKNOWN, in handle_rdmsr() 171 static void handle_wrmsr(struct kvm_run *run) in handle_wrmsr() argument 176 if (run->msr.index == MSR_IA32_POWER_CTL) { in handle_wrmsr() 177 TEST_ASSERT(run->msr.data == 0x1234, in handle_wrmsr() [all …]
|
D | mmio_warning_test.c | 36 struct kvm_run *run; member 44 struct kvm_run *run = tc->run; in thr() local 48 res, run->exit_reason, run->internal.suberror); in thr() 57 struct kvm_run *run; in test() local 66 run = (struct kvm_run *)mmap(0, 4096, PROT_READ|PROT_WRITE, MAP_SHARED, in test() 69 tc.run = run; in test()
|
D | vmx_apic_access_test.c | 104 volatile struct kvm_run *run = vcpu_state(vm, VCPU_ID); in main() local 109 TEST_ASSERT(run->exit_reason == in main() 112 run->exit_reason, in main() 113 exit_reason_str(run->exit_reason)); in main() 114 TEST_ASSERT(run->internal.suberror == in main() 117 run->internal.suberror); in main() 120 TEST_ASSERT(run->exit_reason == KVM_EXIT_IO, in main() 122 run->exit_reason, in main() 123 exit_reason_str(run->exit_reason)); in main()
|
D | platform_info_test.c | 50 struct kvm_run *run = vcpu_state(vm, VCPU_ID); in test_msr_platform_info_enabled() local 55 TEST_ASSERT(run->exit_reason == KVM_EXIT_IO, in test_msr_platform_info_enabled() 57 run->exit_reason, in test_msr_platform_info_enabled() 58 exit_reason_str(run->exit_reason)); in test_msr_platform_info_enabled() 70 struct kvm_run *run = vcpu_state(vm, VCPU_ID); in test_msr_platform_info_disabled() local 74 TEST_ASSERT(run->exit_reason == KVM_EXIT_SHUTDOWN, in test_msr_platform_info_disabled() 76 run->exit_reason, in test_msr_platform_info_disabled() 77 exit_reason_str(run->exit_reason)); in test_msr_platform_info_disabled()
|
/Linux-v5.10/tools/testing/selftests/kvm/s390x/ |
D | sync_regs_test.c | 79 struct kvm_run *run; in main() local 96 run = vcpu_state(vm, VCPU_ID); in main() 99 run->kvm_valid_regs = INVALID_SYNC_FIELD; in main() 106 run->kvm_valid_regs = INVALID_SYNC_FIELD | TEST_SYNC_FIELDS; in main() 114 run->kvm_dirty_regs = INVALID_SYNC_FIELD; in main() 121 run->kvm_dirty_regs = INVALID_SYNC_FIELD | TEST_SYNC_FIELDS; in main() 129 run->kvm_valid_regs = TEST_SYNC_FIELDS; in main() 132 TEST_ASSERT(run->exit_reason == KVM_EXIT_S390_SIEIC, in main() 134 run->exit_reason, in main() 135 exit_reason_str(run->exit_reason)); in main() [all …]
|
/Linux-v5.10/drivers/staging/media/sunxi/cedrus/ |
D | cedrus_dec.c | 29 struct cedrus_run run = {}; in cedrus_device_run() local 32 run.src = v4l2_m2m_next_src_buf(ctx->fh.m2m_ctx); in cedrus_device_run() 33 run.dst = v4l2_m2m_next_dst_buf(ctx->fh.m2m_ctx); in cedrus_device_run() 36 src_req = run.src->vb2_buf.req_obj.req; in cedrus_device_run() 43 run.mpeg2.slice_params = cedrus_find_control_data(ctx, in cedrus_device_run() 45 run.mpeg2.quantization = cedrus_find_control_data(ctx, in cedrus_device_run() 50 run.h264.decode_params = cedrus_find_control_data(ctx, in cedrus_device_run() 52 run.h264.pps = cedrus_find_control_data(ctx, in cedrus_device_run() 54 run.h264.scaling_matrix = cedrus_find_control_data(ctx, in cedrus_device_run() 56 run.h264.slice_params = cedrus_find_control_data(ctx, in cedrus_device_run() [all …]
|
D | cedrus_h264.c | 94 struct cedrus_run *run) in cedrus_write_frame_list() argument 97 const struct v4l2_ctrl_h264_decode_params *decode = run->h264.decode_params; in cedrus_write_frame_list() 98 const struct v4l2_ctrl_h264_sps *sps = run->h264.sps; in cedrus_write_frame_list() 127 if (run->dst->vb2_buf.timestamp == dpb->reference_ts) { in cedrus_write_frame_list() 146 output_buf = vb2_to_cedrus_buffer(&run->dst->vb2_buf); in cedrus_write_frame_list() 170 struct cedrus_run *run, in _cedrus_write_ref_list() argument 174 const struct v4l2_ctrl_h264_decode_params *decode = run->h264.decode_params; in _cedrus_write_ref_list() 215 struct cedrus_run *run) in cedrus_write_ref_list0() argument 217 const struct v4l2_ctrl_h264_slice_params *slice = run->h264.slice_params; in cedrus_write_ref_list0() 219 _cedrus_write_ref_list(ctx, run, in cedrus_write_ref_list0() [all …]
|
/Linux-v5.10/tools/perf/tests/ |
D | make | 10 # run only specific test over 'Makefile' 121 # $(run) contains all available tests 122 run := make_pure 123 # Targets 'clean all' can be run together only through top level 127 run += make_clean_all 132 run += make_python_perf_so 133 run += make_debug 134 run += make_no_libperl 135 run += make_no_libpython 136 run += make_no_scripts [all …]
|
/Linux-v5.10/tools/testing/selftests/arm64/mte/ |
D | check_mmap_options.c | 63 int run, result, map_size; in check_anonymous_memory_mapping() local 68 for (run = 0; run < item; run++) { in check_anonymous_memory_mapping() 69 map_size = sizes[run] + OVERFLOW + UNDERFLOW; in check_anonymous_memory_mapping() 75 mte_initialize_current_context(mode, (uintptr_t)ptr, sizes[run]); in check_anonymous_memory_mapping() 77 ptr = mte_insert_tags((void *)ptr, sizes[run]); in check_anonymous_memory_mapping() 83 result = check_mte_memory(ptr, sizes[run], mode, tag_check); in check_anonymous_memory_mapping() 84 mte_clear_tags((void *)ptr, sizes[run]); in check_anonymous_memory_mapping() 95 int run, fd, map_size; in check_file_memory_mapping() local 100 for (run = 0; run < total; run++) { in check_file_memory_mapping() 105 map_size = sizes[run] + UNDERFLOW + OVERFLOW; in check_file_memory_mapping() [all …]
|
D | check_tags_inclusion.c | 46 int tag, run, result = KSFT_PASS; in check_single_included_tags() local 56 for (run = 0; (run < RUNS) && (result == KSFT_PASS); run++) { in check_single_included_tags() 76 int tag, run, result = KSFT_PASS; in check_multiple_included_tags() local 88 for (run = 0; (run < RUNS) && (result == KSFT_PASS); run++) { in check_multiple_included_tags() 108 int run, result = KSFT_PASS; in check_all_included_tags() local 117 for (run = 0; (run < RUNS) && (result == KSFT_PASS); run++) { in check_all_included_tags() 132 int run; in check_none_included_tags() local 140 for (run = 0; run < RUNS; run++) { in check_none_included_tags()
|
D | check_child_memory.c | 87 int run, result; in check_child_memory_mapping() local 92 for (run = 0; run < item; run++) { in check_child_memory_mapping() 93 ptr = (char *)mte_allocate_memory_tag_range(sizes[run], mem_type, mapping, in check_child_memory_mapping() 95 if (check_allocated_memory_range(ptr, sizes[run], mem_type, in check_child_memory_mapping() 98 result = check_child_tag_inheritance(ptr, sizes[run], mode); in check_child_memory_mapping() 99 mte_free_memory_tag_range((void *)ptr, sizes[run], mem_type, UNDERFLOW, OVERFLOW); in check_child_memory_mapping() 109 int run, fd, map_size, result = KSFT_PASS; in check_child_file_mapping() local 113 for (run = 0; run < total; run++) { in check_child_file_mapping() 118 map_size = sizes[run] + OVERFLOW + UNDERFLOW; in check_child_file_mapping() 125 mte_initialize_current_context(mode, (uintptr_t)ptr, sizes[run]); in check_child_file_mapping() [all …]
|
/Linux-v5.10/arch/s390/kvm/ |
D | diag.c | 25 start = vcpu->run->s.regs.gprs[(vcpu->arch.sie_block->ipa & 0xf0) >> 4]; in diag_release_pages() 26 end = vcpu->run->s.regs.gprs[vcpu->arch.sie_block->ipa & 0xf] + PAGE_SIZE; in diag_release_pages() 76 vcpu->run->s.regs.gprs[rx]); in __diag_page_ref_service() 78 if (vcpu->run->s.regs.gprs[rx] & 7) in __diag_page_ref_service() 80 rc = read_guest(vcpu, vcpu->run->s.regs.gprs[rx], rx, &parm, sizeof(parm)); in __diag_page_ref_service() 97 vcpu->run->s.regs.gprs[ry] = 8; in __diag_page_ref_service() 111 vcpu->run->s.regs.gprs[ry] = 0; in __diag_page_ref_service() 125 vcpu->run->s.regs.gprs[ry] = 0; in __diag_page_ref_service() 131 vcpu->run->s.regs.gprs[ry] = 4; in __diag_page_ref_service() 158 tid = vcpu->run->s.regs.gprs[(vcpu->arch.sie_block->ipa & 0xf0) >> 4]; in __diag_time_slice_end_directed() [all …]
|
D | priv.c | 63 current->thread.gs_cb = (struct gs_cb *)&vcpu->run->s.regs.gscb; in handle_gs() 264 gaddr = vcpu->run->s.regs.gprs[reg2] & PAGE_MASK; in handle_iske() 288 vcpu->run->s.regs.gprs[reg1] &= ~0xff; in handle_iske() 289 vcpu->run->s.regs.gprs[reg1] |= key; in handle_iske() 311 gaddr = vcpu->run->s.regs.gprs[reg2] & PAGE_MASK; in handle_rrbe() 369 key = vcpu->run->s.regs.gprs[reg1] & 0xfe; in handle_sske() 370 start = vcpu->run->s.regs.gprs[reg2] & PAGE_MASK; in handle_sske() 411 vcpu->run->s.regs.gprs[reg1] &= ~0xff00UL; in handle_sske() 412 vcpu->run->s.regs.gprs[reg1] |= (u64) oldkey << 8; in handle_sske() 417 vcpu->run->s.regs.gprs[reg2] &= ~PAGE_MASK; in handle_sske() [all …]
|
/Linux-v5.10/tools/perf/scripts/python/ |
D | stat-cpi.py | 23 def store(time, event, cpu, thread, val, ena, run): argument 29 data[key] = [ val, ena, run] 35 def stat__cycles_k(cpu, thread, time, val, ena, run): argument 36 store(time, "cycles", cpu, thread, val, ena, run); 38 def stat__instructions_k(cpu, thread, time, val, ena, run): argument 39 store(time, "instructions", cpu, thread, val, ena, run); 41 def stat__cycles_u(cpu, thread, time, val, ena, run): argument 42 store(time, "cycles", cpu, thread, val, ena, run); 44 def stat__instructions_u(cpu, thread, time, val, ena, run): argument 45 store(time, "instructions", cpu, thread, val, ena, run); [all …]
|
/Linux-v5.10/fs/befs/ |
D | endian.h | 74 befs_block_run run; in fsrun_to_cpu() local 77 run.allocation_group = le32_to_cpu((__force __le32)n.allocation_group); in fsrun_to_cpu() 78 run.start = le16_to_cpu((__force __le16)n.start); in fsrun_to_cpu() 79 run.len = le16_to_cpu((__force __le16)n.len); in fsrun_to_cpu() 81 run.allocation_group = be32_to_cpu((__force __be32)n.allocation_group); in fsrun_to_cpu() 82 run.start = be16_to_cpu((__force __be16)n.start); in fsrun_to_cpu() 83 run.len = be16_to_cpu((__force __be16)n.len); in fsrun_to_cpu() 85 return run; in fsrun_to_cpu() 91 befs_disk_block_run run; in cpu_to_fsrun() local 94 run.allocation_group = cpu_to_le32(n.allocation_group); in cpu_to_fsrun() [all …]
|
D | datastream.c | 26 befs_blocknr_t blockno, befs_block_run *run); 31 befs_block_run *run); 36 befs_block_run *run); 53 befs_block_run run; in befs_read_datastream() local 61 if (befs_fblock2brun(sb, ds, block, &run) != BEFS_OK) { in befs_read_datastream() 67 bh = befs_bread_iaddr(sb, run); in befs_read_datastream() 96 befs_blocknr_t fblock, befs_block_run *run) in befs_fblock2brun() argument 102 err = befs_find_brun_direct(sb, data, fblock, run); in befs_fblock2brun() 105 err = befs_find_brun_indirect(sb, data, fblock, run); in befs_fblock2brun() 108 err = befs_find_brun_dblindirect(sb, data, fblock, run); in befs_fblock2brun() [all …]
|
/Linux-v5.10/tools/testing/selftests/kvm/ |
D | set_memory_region_test.c | 58 struct kvm_run *run; in vcpu_worker() local 67 run = vcpu_state(vm, VCPU_ID); in vcpu_worker() 72 if (run->exit_reason == KVM_EXIT_IO) { in vcpu_worker() 81 if (run->exit_reason != KVM_EXIT_MMIO) in vcpu_worker() 84 TEST_ASSERT(!run->mmio.is_write, "Unexpected exit mmio write"); in vcpu_worker() 85 TEST_ASSERT(run->mmio.len == 8, in vcpu_worker() 86 "Unexpected exit mmio size = %u", run->mmio.len); in vcpu_worker() 88 TEST_ASSERT(run->mmio.phys_addr == MEM_REGION_GPA, in vcpu_worker() 90 run->mmio.phys_addr); in vcpu_worker() 91 memcpy(run->mmio.data, &MMIO_VAL, 8); in vcpu_worker() [all …]
|
/Linux-v5.10/arch/arm64/kvm/ |
D | mmio.c | 94 struct kvm_run *run = vcpu->run; in kvm_handle_mmio_return() local 97 data = kvm_mmio_read_buf(run->mmio.data, len); in kvm_handle_mmio_return() 108 trace_kvm_mmio(KVM_TRACE_MMIO_READ, len, run->mmio.phys_addr, in kvm_handle_mmio_return() 125 struct kvm_run *run = vcpu->run; in io_mem_abort() local 139 run->exit_reason = KVM_EXIT_ARM_NISV; in io_mem_abort() 140 run->arm_nisv.esr_iss = kvm_vcpu_dabt_iss_nisv_sanitized(vcpu); in io_mem_abort() 141 run->arm_nisv.fault_ipa = fault_ipa; in io_mem_abort() 176 run->mmio.is_write = is_write; in io_mem_abort() 177 run->mmio.phys_addr = fault_ipa; in io_mem_abort() 178 run->mmio.len = len; in io_mem_abort() [all …]
|
D | handle_exit.c | 121 struct kvm_run *run = vcpu->run; in kvm_handle_guest_debug() local 125 run->exit_reason = KVM_EXIT_DEBUG; in kvm_handle_guest_debug() 126 run->debug.arch.hsr = esr; in kvm_handle_guest_debug() 130 run->debug.arch.far = vcpu->arch.fault.far_el2; in kvm_handle_guest_debug() 242 struct kvm_run *run = vcpu->run; in handle_exit() local 275 run->exit_reason = KVM_EXIT_FAIL_ENTRY; in handle_exit() 282 run->exit_reason = KVM_EXIT_FAIL_ENTRY; in handle_exit() 287 run->exit_reason = KVM_EXIT_INTERNAL_ERROR; in handle_exit()
|
/Linux-v5.10/tools/testing/selftests/kvm/lib/s390x/ |
D | ucall.c | 38 struct kvm_run *run = vcpu_state(vm, vcpu_id); in get_ucall() local 44 if (run->exit_reason == KVM_EXIT_S390_SIEIC && in get_ucall() 45 run->s390_sieic.icptcode == 4 && in get_ucall() 46 (run->s390_sieic.ipa >> 8) == 0x83 && /* 0x83 means DIAGNOSE */ in get_ucall() 47 (run->s390_sieic.ipb >> 16) == 0x501) { in get_ucall() 48 int reg = run->s390_sieic.ipa & 0xf; in get_ucall() 50 memcpy(&ucall, addr_gva2hva(vm, run->s.regs.gprs[reg]), in get_ucall()
|
/Linux-v5.10/fs/jbd2/ |
D | commit.c | 477 stats.run.rs_wait = commit_transaction->t_max_wait; in jbd2_journal_commit_transaction() 478 stats.run.rs_request_delay = 0; in jbd2_journal_commit_transaction() 479 stats.run.rs_locked = jiffies; in jbd2_journal_commit_transaction() 481 stats.run.rs_request_delay = in jbd2_journal_commit_transaction() 483 stats.run.rs_locked); in jbd2_journal_commit_transaction() 484 stats.run.rs_running = jbd2_time_diff(commit_transaction->t_start, in jbd2_journal_commit_transaction() 485 stats.run.rs_locked); in jbd2_journal_commit_transaction() 573 stats.run.rs_flushing = jiffies; in jbd2_journal_commit_transaction() 574 stats.run.rs_locked = jbd2_time_diff(stats.run.rs_locked, in jbd2_journal_commit_transaction() 575 stats.run.rs_flushing); in jbd2_journal_commit_transaction() [all …]
|
/Linux-v5.10/drivers/gpu/ipu-v3/ |
D | ipu-image-convert.c | 1257 struct ipu_image_convert_run *run; in get_run_count() local 1262 list_for_each_entry(run, q, list) { in get_run_count() 1263 if (run->ctx == ctx) in get_run_count() 1270 static void convert_stop(struct ipu_image_convert_run *run) in convert_stop() argument 1272 struct ipu_image_convert_ctx *ctx = run->ctx; in convert_stop() 1277 __func__, chan->ic_task, ctx, run); in convert_stop() 1383 static int convert_start(struct ipu_image_convert_run *run, unsigned int tile) in convert_start() argument 1385 struct ipu_image_convert_ctx *ctx = run->ctx; in convert_start() 1397 __func__, chan->ic_task, ctx, run, tile, dst_tile); in convert_start() 1498 static int do_run(struct ipu_image_convert_run *run) in do_run() argument [all …]
|
/Linux-v5.10/drivers/staging/media/rkvdec/ |
D | rkvdec-h264.c | 636 struct rkvdec_h264_run *run) in assemble_hw_pps() argument 639 const struct v4l2_ctrl_h264_sps *sps = run->sps; in assemble_hw_pps() 640 const struct v4l2_ctrl_h264_pps *pps = run->pps; in assemble_hw_pps() 641 const struct v4l2_ctrl_h264_decode_params *dec_params = run->decode_params; in assemble_hw_pps() 729 struct rkvdec_h264_run *run) in assemble_hw_rps() argument 731 const struct v4l2_ctrl_h264_decode_params *dec_params = run->decode_params; in assemble_hw_rps() 734 const struct v4l2_ctrl_h264_sps *sps = run->sps; in assemble_hw_rps() 792 struct rkvdec_h264_run *run) in assemble_hw_scaling_list() argument 794 const struct v4l2_ctrl_h264_scaling_matrix *scaling = run->scaling_matrix; in assemble_hw_scaling_list() 795 const struct v4l2_ctrl_h264_pps *pps = run->pps; in assemble_hw_scaling_list() [all …]
|