Searched refs:xmm (Results 1 – 16 of 16) sorted by relevance
/Linux-v5.4/arch/x86/include/asm/ |
D | inst.h | 126 .macro XMM_NUM opd xmm 128 .ifc \xmm,%xmm0 131 .ifc \xmm,%xmm1 134 .ifc \xmm,%xmm2 137 .ifc \xmm,%xmm3 140 .ifc \xmm,%xmm4 143 .ifc \xmm,%xmm5 146 .ifc \xmm,%xmm6 149 .ifc \xmm,%xmm7 152 .ifc \xmm,%xmm8 [all …]
|
D | perf_event.h | 211 u64 xmm[16*2]; /* two entries for each register */ member
|
D | kvm_emulate.h | 252 unsigned xmm; member
|
/Linux-v5.4/arch/x86/crypto/ |
D | aesni-intel_asm.S | 183 # states of %xmm registers %xmm6:%xmm15 not saved 184 # all %xmm registers are clobbered 199 # clobbers r12, and tmp xmm registers. 649 movups (\PLAIN_CYPH_IN), %xmm1 # If more than 16 bytes, just fill xmm 796 movdqu AadHash(%arg2), %xmm\i # XMM0 = Y0 809 movdqa \XMM0, %xmm\index 811 MOVADQ \XMM0, %xmm\index 813 PSHUFB_XMM %xmm14, %xmm\index # perform a 16 byte swap 814 pxor \TMP2, %xmm\index 824 AESENC \TMP1, %xmm\index [all …]
|
D | aes_ctrby8_avx-x86_64.S | 136 var_xdata = %xmm\n
|
D | sha256-avx-asm.S | 72 # COPY_XMM_AND_BSWAP xmm, [mem], byte_flip_mask 73 # Load xmm with mem and byte swap each dword
|
D | sha256-ssse3-asm.S | 65 # COPY_XMM_AND_BSWAP xmm, [mem], byte_flip_mask 66 # Load xmm with mem and byte swap each dword
|
D | aesni-intel_avx-x86_64.S | 78 ## padded AAD in xmm register = {A1 A0 0 0} 94 ## padded AAD in xmm register = {A2 A1 A0 0} 231 reg_\r = %xmm\n 297 # clobbering all xmm registers 758 vmovdqu (\PLAIN_CYPH_IN), %xmm1 # If more than 16 bytes, just fill xmm
|
/Linux-v5.4/tools/perf/util/intel-pt-decoder/ |
D | intel-pt-decoder.h | 194 uint64_t xmm[INTEL_PT_BLK_ITEM_ID_CNT]; member
|
/Linux-v5.4/arch/x86/include/uapi/asm/ |
D | kvm.h | 170 __u8 xmm[16][16]; member
|
/Linux-v5.4/tools/arch/x86/include/uapi/asm/ |
D | kvm.h | 170 __u8 xmm[16][16]; member
|
/Linux-v5.4/arch/x86/events/intel/ |
D | ds.c | 1590 struct pebs_xmm *xmm = next_record; in setup_pebs_adaptive_sample_data() local 1592 next_record = xmm + 1; in setup_pebs_adaptive_sample_data() 1593 perf_regs->xmm_regs = xmm->xmm; in setup_pebs_adaptive_sample_data()
|
/Linux-v5.4/tools/perf/util/ |
D | intel-pt.c | 1632 const u64 *xmm = items->xmm; in intel_pt_add_xmm() local 1642 for (; mask; mask >>= 1, xmm++) { in intel_pt_add_xmm() 1644 *pos++ = *xmm; in intel_pt_add_xmm()
|
/Linux-v5.4/arch/x86/kvm/ |
D | emulate.c | 1207 op->addr.xmm = reg; in decode_register_operand() 1258 op->addr.xmm = ctxt->modrm_rm; in decode_modrm() 1836 write_sse_reg(ctxt, &op->vec_val, op->addr.xmm); in writeback()
|
D | x86.c | 8965 memcpy(fpu->xmm, fxsave->xmm_space, sizeof(fxsave->xmm_space)); in kvm_arch_vcpu_ioctl_get_fpu() 8986 memcpy(fxsave->xmm_space, fpu->xmm, sizeof(fxsave->xmm_space)); in kvm_arch_vcpu_ioctl_set_fpu()
|
/Linux-v5.4/Documentation/virt/kvm/ |
D | api.txt | 676 __u8 xmm[16][16]; 702 __u8 xmm[16][16];
|