Home
last modified time | relevance | path

Searched refs:arch (Results 1 – 25 of 875) sorted by relevance

12345678910>>...35

/Linux-v5.4/arch/powerpc/kvm/
Dbooke_emulate.c26 vcpu->arch.regs.nip = vcpu->arch.shared->srr0; in kvmppc_emul_rfi()
27 kvmppc_set_msr(vcpu, vcpu->arch.shared->srr1); in kvmppc_emul_rfi()
32 vcpu->arch.regs.nip = vcpu->arch.dsrr0; in kvmppc_emul_rfdi()
33 kvmppc_set_msr(vcpu, vcpu->arch.dsrr1); in kvmppc_emul_rfdi()
38 vcpu->arch.regs.nip = vcpu->arch.csrr0; in kvmppc_emul_rfci()
39 kvmppc_set_msr(vcpu, vcpu->arch.csrr1); in kvmppc_emul_rfci()
80 kvmppc_set_gpr(vcpu, rt, vcpu->arch.shared->msr); in kvmppc_booke_emulate_op()
90 vcpu->arch.shared->msr = (vcpu->arch.shared->msr & ~MSR_EE) in kvmppc_booke_emulate_op()
96 vcpu->arch.shared->msr = (vcpu->arch.shared->msr & ~MSR_EE) in kvmppc_booke_emulate_op()
127 vcpu->arch.shared->dar = spr_val; in kvmppc_booke_emulate_mtspr()
[all …]
Dbook3s_hv_tm.c17 u64 msr = vcpu->arch.shregs.msr; in emulate_tx_failure()
19 tfiar = vcpu->arch.regs.nip & ~0x3ull; in emulate_tx_failure()
21 if (MSR_TM_SUSPENDED(vcpu->arch.shregs.msr)) in emulate_tx_failure()
27 vcpu->arch.tfiar = tfiar; in emulate_tx_failure()
29 vcpu->arch.texasr = (vcpu->arch.texasr & 0x3ffffff) | texasr; in emulate_tx_failure()
42 u32 instr = vcpu->arch.emul_inst; in kvmhv_p9_tm_emulation()
43 u64 msr = vcpu->arch.shregs.msr; in kvmhv_p9_tm_emulation()
50 newmsr = vcpu->arch.shregs.srr1; in kvmhv_p9_tm_emulation()
56 vcpu->arch.shregs.msr = newmsr; in kvmhv_p9_tm_emulation()
57 vcpu->arch.cfar = vcpu->arch.regs.nip - 4; in kvmhv_p9_tm_emulation()
[all …]
Dbook3s_hv.c130 return kvm->arch.nested_enable && kvm_is_radix(kvm); in nesting_enabled()
240 cpu = READ_ONCE(vcpu->arch.thread_cpu); in kvmppc_fast_vcpu_kick_hv()
306 struct kvmppc_vcore *vc = vcpu->arch.vcore; in kvmppc_core_vcpu_load_hv()
318 spin_lock_irqsave(&vcpu->arch.tbacct_lock, flags); in kvmppc_core_vcpu_load_hv()
319 if (vcpu->arch.state == KVMPPC_VCPU_BUSY_IN_HOST && in kvmppc_core_vcpu_load_hv()
320 vcpu->arch.busy_preempt != TB_NIL) { in kvmppc_core_vcpu_load_hv()
321 vcpu->arch.busy_stolen += mftb() - vcpu->arch.busy_preempt; in kvmppc_core_vcpu_load_hv()
322 vcpu->arch.busy_preempt = TB_NIL; in kvmppc_core_vcpu_load_hv()
324 spin_unlock_irqrestore(&vcpu->arch.tbacct_lock, flags); in kvmppc_core_vcpu_load_hv()
329 struct kvmppc_vcore *vc = vcpu->arch.vcore; in kvmppc_core_vcpu_put_hv()
[all …]
Dtiming.c27 mutex_lock(&vcpu->arch.exit_timing_lock); in kvmppc_init_timing_stats()
29 vcpu->arch.last_exit_type = 0xDEAD; in kvmppc_init_timing_stats()
31 vcpu->arch.timing_count_type[i] = 0; in kvmppc_init_timing_stats()
32 vcpu->arch.timing_max_duration[i] = 0; in kvmppc_init_timing_stats()
33 vcpu->arch.timing_min_duration[i] = 0xFFFFFFFF; in kvmppc_init_timing_stats()
34 vcpu->arch.timing_sum_duration[i] = 0; in kvmppc_init_timing_stats()
35 vcpu->arch.timing_sum_quad_duration[i] = 0; in kvmppc_init_timing_stats()
37 vcpu->arch.timing_last_exit = 0; in kvmppc_init_timing_stats()
38 vcpu->arch.timing_exit.tv64 = 0; in kvmppc_init_timing_stats()
39 vcpu->arch.timing_last_enter.tv64 = 0; in kvmppc_init_timing_stats()
[all …]
Dbooke.c69 printk("pc: %08lx msr: %08llx\n", vcpu->arch.regs.nip, in kvmppc_dump_vcpu()
70 vcpu->arch.shared->msr); in kvmppc_dump_vcpu()
71 printk("lr: %08lx ctr: %08lx\n", vcpu->arch.regs.link, in kvmppc_dump_vcpu()
72 vcpu->arch.regs.ctr); in kvmppc_dump_vcpu()
73 printk("srr0: %08llx srr1: %08llx\n", vcpu->arch.shared->srr0, in kvmppc_dump_vcpu()
74 vcpu->arch.shared->srr1); in kvmppc_dump_vcpu()
76 printk("exceptions: %08lx\n", vcpu->arch.pending_exceptions); in kvmppc_dump_vcpu()
94 vcpu->arch.shadow_msr &= ~MSR_SPE; in kvmppc_vcpu_disable_spe()
104 vcpu->arch.shadow_msr |= MSR_SPE; in kvmppc_vcpu_enable_spe()
110 if (vcpu->arch.shared->msr & MSR_SPE) { in kvmppc_vcpu_sync_spe()
[all …]
Demulate_loadstore.c92 vcpu->arch.mmio_vsx_copy_nums = 0; in kvmppc_emulate_loadstore()
93 vcpu->arch.mmio_vsx_offset = 0; in kvmppc_emulate_loadstore()
94 vcpu->arch.mmio_copy_type = KVMPPC_VSX_COPY_NONE; in kvmppc_emulate_loadstore()
95 vcpu->arch.mmio_sp64_extend = 0; in kvmppc_emulate_loadstore()
96 vcpu->arch.mmio_sign_extend = 0; in kvmppc_emulate_loadstore()
97 vcpu->arch.mmio_vmx_copy_nums = 0; in kvmppc_emulate_loadstore()
98 vcpu->arch.mmio_vmx_offset = 0; in kvmppc_emulate_loadstore()
99 vcpu->arch.mmio_host_swabbed = 0; in kvmppc_emulate_loadstore()
102 vcpu->arch.regs.msr = vcpu->arch.shared->msr; in kvmppc_emulate_loadstore()
103 if (analyse_instr(&op, &vcpu->arch.regs, inst) == 0) { in kvmppc_emulate_loadstore()
[all …]
Dbook3s_hv_tm_builtin.c22 u32 instr = vcpu->arch.emul_inst; in kvmhv_p9_tm_emulation_early()
29 newmsr = vcpu->arch.shregs.srr1; in kvmhv_p9_tm_emulation_early()
34 vcpu->arch.shregs.msr = newmsr; in kvmhv_p9_tm_emulation_early()
35 vcpu->arch.cfar = vcpu->arch.regs.nip - 4; in kvmhv_p9_tm_emulation_early()
36 vcpu->arch.regs.nip = vcpu->arch.shregs.srr0; in kvmhv_p9_tm_emulation_early()
41 msr = vcpu->arch.shregs.msr; in kvmhv_p9_tm_emulation_early()
42 if ((msr & MSR_PR) && (vcpu->arch.vcore->pcr & PCR_ARCH_206)) in kvmhv_p9_tm_emulation_early()
45 if (!(vcpu->arch.hfscr & HFSCR_EBB) || in kvmhv_p9_tm_emulation_early()
57 vcpu->arch.shregs.msr = msr; in kvmhv_p9_tm_emulation_early()
58 vcpu->arch.cfar = vcpu->arch.regs.nip - 4; in kvmhv_p9_tm_emulation_early()
[all …]
Dbook3s_emulate.c77 if (vcpu->arch.papr_enabled && (level > PRIV_SUPER)) in spr_allowed()
90 memcpy(&vcpu->arch.gpr_tm[0], &vcpu->arch.regs.gpr[0], in kvmppc_copyto_vcpu_tm()
91 sizeof(vcpu->arch.gpr_tm)); in kvmppc_copyto_vcpu_tm()
92 memcpy(&vcpu->arch.fp_tm, &vcpu->arch.fp, in kvmppc_copyto_vcpu_tm()
94 memcpy(&vcpu->arch.vr_tm, &vcpu->arch.vr, in kvmppc_copyto_vcpu_tm()
96 vcpu->arch.ppr_tm = vcpu->arch.ppr; in kvmppc_copyto_vcpu_tm()
97 vcpu->arch.dscr_tm = vcpu->arch.dscr; in kvmppc_copyto_vcpu_tm()
98 vcpu->arch.amr_tm = vcpu->arch.amr; in kvmppc_copyto_vcpu_tm()
99 vcpu->arch.ctr_tm = vcpu->arch.regs.ctr; in kvmppc_copyto_vcpu_tm()
100 vcpu->arch.tar_tm = vcpu->arch.tar; in kvmppc_copyto_vcpu_tm()
[all …]
De500_emulate.c53 ulong param = vcpu->arch.regs.gpr[rb]; in kvmppc_e500_emul_msgclr()
59 clear_bit(prio, &vcpu->arch.pending_exceptions); in kvmppc_e500_emul_msgclr()
65 ulong param = vcpu->arch.regs.gpr[rb]; in kvmppc_e500_emul_msgsnd()
75 int cpir = cvcpu->arch.shared->pir; in kvmppc_e500_emul_msgsnd()
77 set_bit(prio, &cvcpu->arch.pending_exceptions); in kvmppc_e500_emul_msgsnd()
94 run->debug.arch.address = vcpu->arch.regs.nip; in kvmppc_e500_emul_ehpriv()
95 run->debug.arch.status = 0; in kvmppc_e500_emul_ehpriv()
226 vcpu->arch.shared->mas0 = spr_val; in kvmppc_core_emulate_mtspr_e500()
229 vcpu->arch.shared->mas1 = spr_val; in kvmppc_core_emulate_mtspr_e500()
232 vcpu->arch.shared->mas2 = spr_val; in kvmppc_core_emulate_mtspr_e500()
[all …]
Dbook3s_pr.c82 if (vcpu->arch.hflags & BOOK3S_HFLAG_SPLIT_HACK) in kvmppc_fixup_split_real()
89 vcpu->arch.hflags |= BOOK3S_HFLAG_SPLIT_HACK; in kvmppc_fixup_split_real()
112 current->thread.kvm_shadow_vcpu = vcpu->arch.shadow_vcpu; in kvmppc_core_vcpu_load_pr()
153 svcpu->gpr[0] = vcpu->arch.regs.gpr[0]; in kvmppc_copy_to_svcpu()
154 svcpu->gpr[1] = vcpu->arch.regs.gpr[1]; in kvmppc_copy_to_svcpu()
155 svcpu->gpr[2] = vcpu->arch.regs.gpr[2]; in kvmppc_copy_to_svcpu()
156 svcpu->gpr[3] = vcpu->arch.regs.gpr[3]; in kvmppc_copy_to_svcpu()
157 svcpu->gpr[4] = vcpu->arch.regs.gpr[4]; in kvmppc_copy_to_svcpu()
158 svcpu->gpr[5] = vcpu->arch.regs.gpr[5]; in kvmppc_copy_to_svcpu()
159 svcpu->gpr[6] = vcpu->arch.regs.gpr[6]; in kvmppc_copy_to_svcpu()
[all …]
De500mc.c102 vcpu->arch.pid = pid; in kvmppc_set_pid()
119 mtspr(SPRN_EPCR, vcpu->arch.shadow_epcr); in kvmppc_core_vcpu_load_e500mc()
121 mtspr(SPRN_MSRP, vcpu->arch.shadow_msrp); in kvmppc_core_vcpu_load_e500mc()
122 vcpu->arch.eplc = EPC_EGS | (get_lpid(vcpu) << EPC_ELPID_SHIFT); in kvmppc_core_vcpu_load_e500mc()
123 vcpu->arch.epsc = vcpu->arch.eplc; in kvmppc_core_vcpu_load_e500mc()
124 mtspr(SPRN_EPLC, vcpu->arch.eplc); in kvmppc_core_vcpu_load_e500mc()
125 mtspr(SPRN_EPSC, vcpu->arch.epsc); in kvmppc_core_vcpu_load_e500mc()
127 mtspr(SPRN_GIVPR, vcpu->arch.ivpr); in kvmppc_core_vcpu_load_e500mc()
128 mtspr(SPRN_GIVOR2, vcpu->arch.ivor[BOOKE_IRQPRIO_DATA_STORAGE]); in kvmppc_core_vcpu_load_e500mc()
129 mtspr(SPRN_GIVOR8, vcpu->arch.ivor[BOOKE_IRQPRIO_SYSCALL]); in kvmppc_core_vcpu_load_e500mc()
[all …]
Dbook3s_hv_nested.c30 struct kvmppc_vcore *vc = vcpu->arch.vcore; in kvmhv_save_hv_regs()
34 hr->hfscr = vcpu->arch.hfscr; in kvmhv_save_hv_regs()
36 hr->dawr0 = vcpu->arch.dawr; in kvmhv_save_hv_regs()
37 hr->dawrx0 = vcpu->arch.dawrx; in kvmhv_save_hv_regs()
38 hr->ciabr = vcpu->arch.ciabr; in kvmhv_save_hv_regs()
39 hr->purr = vcpu->arch.purr; in kvmhv_save_hv_regs()
40 hr->spurr = vcpu->arch.spurr; in kvmhv_save_hv_regs()
41 hr->ic = vcpu->arch.ic; in kvmhv_save_hv_regs()
43 hr->srr0 = vcpu->arch.shregs.srr0; in kvmhv_save_hv_regs()
44 hr->srr1 = vcpu->arch.shregs.srr1; in kvmhv_save_hv_regs()
[all …]
De500_mmu.c71 esel += gtlb0_set_base(vcpu_e500, vcpu->arch.shared->mas2); in get_tlb_esel()
134 tlbsel = (vcpu->arch.shared->mas4 >> 28) & 0x1; in kvmppc_e500_deliver_tlb_miss()
136 tsized = (vcpu->arch.shared->mas4 >> 7) & 0x1f; in kvmppc_e500_deliver_tlb_miss()
138 vcpu->arch.shared->mas0 = MAS0_TLBSEL(tlbsel) | MAS0_ESEL(victim) in kvmppc_e500_deliver_tlb_miss()
140 vcpu->arch.shared->mas1 = MAS1_VALID | (as ? MAS1_TS : 0) in kvmppc_e500_deliver_tlb_miss()
143 vcpu->arch.shared->mas2 = (eaddr & MAS2_EPN) in kvmppc_e500_deliver_tlb_miss()
144 | (vcpu->arch.shared->mas4 & MAS2_ATTRIB_MASK); in kvmppc_e500_deliver_tlb_miss()
145 vcpu->arch.shared->mas7_3 &= MAS3_U0 | MAS3_U1 | MAS3_U2 | MAS3_U3; in kvmppc_e500_deliver_tlb_miss()
146 vcpu->arch.shared->mas6 = (vcpu->arch.shared->mas6 & MAS6_SPID1) in kvmppc_e500_deliver_tlb_miss()
332 vcpu->arch.shared->mas0 &= ~MAS0_NV(~0); in kvmppc_e500_emul_tlbre()
[all …]
/Linux-v5.4/arch/s390/kvm/
Dguestdbg.c62 u64 *cr9 = &vcpu->arch.sie_block->gcr[9]; in enable_all_hw_bp()
63 u64 *cr10 = &vcpu->arch.sie_block->gcr[10]; in enable_all_hw_bp()
64 u64 *cr11 = &vcpu->arch.sie_block->gcr[11]; in enable_all_hw_bp()
67 if (vcpu->arch.guestdbg.nr_hw_bp <= 0 || in enable_all_hw_bp()
68 vcpu->arch.guestdbg.hw_bp_info == NULL) in enable_all_hw_bp()
79 for (i = 0; i < vcpu->arch.guestdbg.nr_hw_bp; i++) { in enable_all_hw_bp()
80 start = vcpu->arch.guestdbg.hw_bp_info[i].addr; in enable_all_hw_bp()
81 len = vcpu->arch.guestdbg.hw_bp_info[i].len; in enable_all_hw_bp()
102 u64 *cr9 = &vcpu->arch.sie_block->gcr[9]; in enable_all_hw_wp()
103 u64 *cr10 = &vcpu->arch.sie_block->gcr[10]; in enable_all_hw_wp()
[all …]
Dkvm-s390.c276 kvm_clock_sync_scb(vcpu->arch.sie_block, *delta); in kvm_clock_sync()
278 kvm->arch.epoch = vcpu->arch.sie_block->epoch; in kvm_clock_sync()
279 kvm->arch.epdx = vcpu->arch.sie_block->epdx; in kvm_clock_sync()
281 if (vcpu->arch.cputm_enabled) in kvm_clock_sync()
282 vcpu->arch.cputm_start += *delta; in kvm_clock_sync()
283 if (vcpu->arch.vsie_block) in kvm_clock_sync()
284 kvm_clock_sync_scb(vcpu->arch.vsie_block, in kvm_clock_sync()
581 struct gmap *gmap = kvm->arch.gmap; in kvm_s390_sync_dirty_log()
672 kvm->arch.use_irqchip = 1; in kvm_vm_ioctl_enable_cap()
677 kvm->arch.user_sigp = 1; in kvm_vm_ioctl_enable_cap()
[all …]
Dintercept.c27 struct kvm_s390_sie_block *sie_block = vcpu->arch.sie_block; in kvm_s390_get_ilen()
30 switch (vcpu->arch.sie_block->icptcode) { in kvm_s390_get_ilen()
37 ilen = insn_length(vcpu->arch.sie_block->ipa >> 8); in kvm_s390_get_ilen()
47 ilen = vcpu->arch.sie_block->pgmilc & 0x6; in kvm_s390_get_ilen()
55 struct kvm_s390_local_interrupt *li = &vcpu->arch.local_int; in handle_stop()
89 int viwhy = vcpu->arch.sie_block->ipb >> 16; in handle_validity()
106 vcpu->arch.sie_block->ipa, in handle_instruction()
107 vcpu->arch.sie_block->ipb); in handle_instruction()
109 switch (vcpu->arch.sie_block->ipa >> 8) { in handle_instruction()
142 .code = vcpu->arch.sie_block->iprcc, in inject_prog_on_prog_intercept()
[all …]
/Linux-v5.4/arch/mips/kvm/
Demulate.c46 struct kvm_vcpu_arch *arch = &vcpu->arch; in kvm_compute_return_epc() local
66 arch->gprs[insn.r_format.rd] = epc + 8; in kvm_compute_return_epc()
69 nextpc = arch->gprs[insn.r_format.rs]; in kvm_compute_return_epc()
85 if ((long)arch->gprs[insn.i_format.rs] < 0) in kvm_compute_return_epc()
94 if ((long)arch->gprs[insn.i_format.rs] >= 0) in kvm_compute_return_epc()
103 arch->gprs[31] = epc + 8; in kvm_compute_return_epc()
104 if ((long)arch->gprs[insn.i_format.rs] < 0) in kvm_compute_return_epc()
113 arch->gprs[31] = epc + 8; in kvm_compute_return_epc()
114 if ((long)arch->gprs[insn.i_format.rs] >= 0) in kvm_compute_return_epc()
142 arch->gprs[31] = instpc + 8; in kvm_compute_return_epc()
[all …]
/Linux-v5.4/scripts/
Dcheckstack.pl40 my $arch = shift;
41 if ($arch eq "") {
42 $arch = `uname -m`;
43 chomp($arch);
49 if ($arch =~ '^(aarch|arm)64$') {
54 } elsif ($arch eq 'arm') {
57 } elsif ($arch =~ /^x86(_64)?$/ || $arch =~ /^i[3456]86$/) {
63 } elsif ($arch eq 'ia64') {
66 } elsif ($arch eq 'm68k') {
70 } elsif ($arch eq 'mips64') {
[all …]
/Linux-v5.4/arch/powerpc/kernel/
Dasm-offsets.c434 OFFSET(VCPU_HOST_STACK, kvm_vcpu, arch.host_stack); in main()
435 OFFSET(VCPU_HOST_PID, kvm_vcpu, arch.host_pid); in main()
436 OFFSET(VCPU_GUEST_PID, kvm_vcpu, arch.pid); in main()
437 OFFSET(VCPU_GPRS, kvm_vcpu, arch.regs.gpr); in main()
438 OFFSET(VCPU_VRSAVE, kvm_vcpu, arch.vrsave); in main()
439 OFFSET(VCPU_FPRS, kvm_vcpu, arch.fp.fpr); in main()
441 OFFSET(VCPU_VRS, kvm_vcpu, arch.vr.vr); in main()
443 OFFSET(VCPU_XER, kvm_vcpu, arch.regs.xer); in main()
444 OFFSET(VCPU_CTR, kvm_vcpu, arch.regs.ctr); in main()
445 OFFSET(VCPU_LR, kvm_vcpu, arch.regs.link); in main()
[all …]
/Linux-v5.4/arch/arm64/kvm/
Ddebug.c40 vcpu->arch.guest_debug_preserved.mdscr_el1 = val; in save_guest_debug_regs()
43 vcpu->arch.guest_debug_preserved.mdscr_el1); in save_guest_debug_regs()
48 u64 val = vcpu->arch.guest_debug_preserved.mdscr_el1; in restore_guest_debug_regs()
77 vcpu->arch.debug_ptr = &vcpu->arch.vcpu_debug_state; in kvm_arm_reset_debug_ptr()
103 bool trap_debug = !(vcpu->arch.flags & KVM_ARM64_DEBUG_DIRTY); in kvm_arm_setup_debug()
112 vcpu->arch.mdcr_el2 = __this_cpu_read(mdcr_el2) & MDCR_EL2_HPMN_MASK; in kvm_arm_setup_debug()
113 vcpu->arch.mdcr_el2 |= (MDCR_EL2_TPM | in kvm_arm_setup_debug()
122 vcpu->arch.mdcr_el2 |= MDCR_EL2_TDE; in kvm_arm_setup_debug()
175 vcpu->arch.debug_ptr = &vcpu->arch.external_debug_state; in kvm_arm_setup_debug()
176 vcpu->arch.flags |= KVM_ARM64_DEBUG_DIRTY; in kvm_arm_setup_debug()
[all …]
/Linux-v5.4/arch/powerpc/include/asm/
Dkvm_book3s_64.h140 return kvm->arch.radix; in kvm_is_radix()
147 if (vcpu->arch.nested) in kvmhv_vcpu_is_radix()
148 radix = vcpu->arch.nested->radix; in kvmhv_vcpu_is_radix()
526 if (atomic_read(&kvm->arch.hpte_mod_interest)) in note_hpte_modification()
590 vcpu->arch.regs.ccr = vcpu->arch.cr_tm; in copy_from_checkpoint()
591 vcpu->arch.regs.xer = vcpu->arch.xer_tm; in copy_from_checkpoint()
592 vcpu->arch.regs.link = vcpu->arch.lr_tm; in copy_from_checkpoint()
593 vcpu->arch.regs.ctr = vcpu->arch.ctr_tm; in copy_from_checkpoint()
594 vcpu->arch.amr = vcpu->arch.amr_tm; in copy_from_checkpoint()
595 vcpu->arch.ppr = vcpu->arch.ppr_tm; in copy_from_checkpoint()
[all …]
/Linux-v5.4/tools/perf/trace/beauty/
Darch_errno_names.sh20 local arch="$1"
23 header="$toolsdir/arch/$arch/include/uapi/asm/errno.h"
33 local arch=$(arch_string "$1")
37 static const char *errno_to_name__$arch(int err)
57 local arch="$1"
58 local asm_errno=$(asm_errno_file "$arch")
64 |IFS=, create_errno_lookup_func "$arch"
71 local arch
75 for arch in $archlist; do
76 printf '\tif (!strcmp(arch, "%s"))\n' $(arch_string "$arch")
[all …]
/Linux-v5.4/arch/x86/tools/
DMakefile16 reformatter = $(srctree)/arch/x86/tools/objdump_reformat.awk
17 chkobjdump = $(srctree)/arch/x86/tools/chkobjdump.awk
32 …st.o := -Wall -I$(objtree)/arch/x86/lib/ -I$(srctree)/arch/x86/include/uapi/ -I$(srctree)/arch/x86…
34 …STCFLAGS_insn_sanity.o := -Wall -I$(objtree)/arch/x86/lib/ -I$(srctree)/arch/x86/include/ -I$(srct…
37arch/x86/lib/insn.c $(srctree)/arch/x86/lib/inat.c $(srctree)/arch/x86/include/asm/inat_types.h $(…
39arch/x86/lib/insn.c $(srctree)/arch/x86/lib/inat.c $(srctree)/arch/x86/include/asm/inat_types.h $(…
/Linux-v5.4/arch/x86/kernel/
Dmachine_kexec_32.c57 free_pages((unsigned long)image->arch.pgd, PGD_ALLOCATION_ORDER); in machine_kexec_free_page_tables()
58 image->arch.pgd = NULL; in machine_kexec_free_page_tables()
60 free_page((unsigned long)image->arch.pmd0); in machine_kexec_free_page_tables()
61 image->arch.pmd0 = NULL; in machine_kexec_free_page_tables()
62 free_page((unsigned long)image->arch.pmd1); in machine_kexec_free_page_tables()
63 image->arch.pmd1 = NULL; in machine_kexec_free_page_tables()
65 free_page((unsigned long)image->arch.pte0); in machine_kexec_free_page_tables()
66 image->arch.pte0 = NULL; in machine_kexec_free_page_tables()
67 free_page((unsigned long)image->arch.pte1); in machine_kexec_free_page_tables()
68 image->arch.pte1 = NULL; in machine_kexec_free_page_tables()
[all …]
/Linux-v5.4/arch/riscv/kernel/
Dmodule-sections.c15 struct mod_section *got_sec = &mod->arch.got; in module_emit_got_entry()
34 struct mod_section *got_plt_sec = &mod->arch.got_plt; in module_emit_plt_entry()
36 struct mod_section *plt_sec = &mod->arch.plt; in module_emit_plt_entry()
102 mod->arch.plt.shdr = sechdrs + i; in module_frob_arch_sections()
104 mod->arch.got.shdr = sechdrs + i; in module_frob_arch_sections()
106 mod->arch.got_plt.shdr = sechdrs + i; in module_frob_arch_sections()
109 if (!mod->arch.plt.shdr) { in module_frob_arch_sections()
113 if (!mod->arch.got.shdr) { in module_frob_arch_sections()
117 if (!mod->arch.got_plt.shdr) { in module_frob_arch_sections()
138 mod->arch.plt.shdr->sh_type = SHT_NOBITS; in module_frob_arch_sections()
[all …]

12345678910>>...35