Lines Matching +full:11 +full:n
14 asm(".include \"asm/vx-insn.h\"\n");
42 asm volatile("std 11,%0" : "=Q" (state->fprs[11])); in __kernel_fpu_begin()
57 " la 1,%[vxrs]\n" /* load save area */ in __kernel_fpu_begin()
58 " tmll %[m],30\n" /* KERNEL_VXR */ in __kernel_fpu_begin()
59 " jz 7f\n" /* no work -> done */ in __kernel_fpu_begin()
60 " jo 5f\n" /* -> save V0..V31 */ in __kernel_fpu_begin()
65 " chi %[m],12\n" /* KERNEL_VXR_MID */ in __kernel_fpu_begin()
66 " jne 0f\n" /* -> save V8..V23 */ in __kernel_fpu_begin()
67 " VSTM 8,23,128,1\n" /* vstm %v8,%v23,128(%r1) */ in __kernel_fpu_begin()
68 " j 7f\n" in __kernel_fpu_begin()
70 "0: tmll %[m],6\n" /* KERNEL_VXR_LOW */ in __kernel_fpu_begin()
71 " jz 3f\n" /* -> KERNEL_VXR_HIGH */ in __kernel_fpu_begin()
72 " jo 2f\n" /* 11 -> save V0..V15 */ in __kernel_fpu_begin()
73 " brc 2,1f\n" /* 10 -> save V8..V15 */ in __kernel_fpu_begin()
74 " VSTM 0,7,0,1\n" /* vstm %v0,%v7,0(%r1) */ in __kernel_fpu_begin()
75 " j 3f\n" in __kernel_fpu_begin()
76 "1: VSTM 8,15,128,1\n" /* vstm %v8,%v15,128(%r1) */ in __kernel_fpu_begin()
77 " j 3f\n" in __kernel_fpu_begin()
78 "2: VSTM 0,15,0,1\n" /* vstm %v0,%v15,0(%r1) */ in __kernel_fpu_begin()
80 "3: tmll %[m],24\n" /* KERNEL_VXR_HIGH */ in __kernel_fpu_begin()
81 " jz 7f\n" in __kernel_fpu_begin()
82 " jo 6f\n" /* 11 -> save V16..V31 */ in __kernel_fpu_begin()
83 " brc 2,4f\n" /* 10 -> save V24..V31 */ in __kernel_fpu_begin()
84 " VSTM 16,23,256,1\n" /* vstm %v16,%v23,256(%r1) */ in __kernel_fpu_begin()
85 " j 7f\n" in __kernel_fpu_begin()
86 "4: VSTM 24,31,384,1\n" /* vstm %v24,%v31,384(%r1) */ in __kernel_fpu_begin()
87 " j 7f\n" in __kernel_fpu_begin()
88 "5: VSTM 0,15,0,1\n" /* vstm %v0,%v15,0(%r1) */ in __kernel_fpu_begin()
89 "6: VSTM 16,31,256,1\n" /* vstm %v16,%v31,256(%r1) */ in __kernel_fpu_begin()
124 asm volatile("ld 11,%0" : : "Q" (state->fprs[11])); in __kernel_fpu_end()
139 " la 1,%[vxrs]\n" /* load restore area */ in __kernel_fpu_end()
140 " tmll %[m],30\n" /* KERNEL_VXR */ in __kernel_fpu_end()
141 " jz 7f\n" /* no work -> done */ in __kernel_fpu_end()
142 " jo 5f\n" /* -> restore V0..V31 */ in __kernel_fpu_end()
147 " chi %[m],12\n" /* KERNEL_VXR_MID */ in __kernel_fpu_end()
148 " jne 0f\n" /* -> restore V8..V23 */ in __kernel_fpu_end()
149 " VLM 8,23,128,1\n" /* vlm %v8,%v23,128(%r1) */ in __kernel_fpu_end()
150 " j 7f\n" in __kernel_fpu_end()
152 "0: tmll %[m],6\n" /* KERNEL_VXR_LOW */ in __kernel_fpu_end()
153 " jz 3f\n" /* -> KERNEL_VXR_HIGH */ in __kernel_fpu_end()
154 " jo 2f\n" /* 11 -> restore V0..V15 */ in __kernel_fpu_end()
155 " brc 2,1f\n" /* 10 -> restore V8..V15 */ in __kernel_fpu_end()
156 " VLM 0,7,0,1\n" /* vlm %v0,%v7,0(%r1) */ in __kernel_fpu_end()
157 " j 3f\n" in __kernel_fpu_end()
158 "1: VLM 8,15,128,1\n" /* vlm %v8,%v15,128(%r1) */ in __kernel_fpu_end()
159 " j 3f\n" in __kernel_fpu_end()
160 "2: VLM 0,15,0,1\n" /* vlm %v0,%v15,0(%r1) */ in __kernel_fpu_end()
162 "3: tmll %[m],24\n" /* KERNEL_VXR_HIGH */ in __kernel_fpu_end()
163 " jz 7f\n" in __kernel_fpu_end()
164 " jo 6f\n" /* 11 -> restore V16..V31 */ in __kernel_fpu_end()
165 " brc 2,4f\n" /* 10 -> restore V24..V31 */ in __kernel_fpu_end()
166 " VLM 16,23,256,1\n" /* vlm %v16,%v23,256(%r1) */ in __kernel_fpu_end()
167 " j 7f\n" in __kernel_fpu_end()
168 "4: VLM 24,31,384,1\n" /* vlm %v24,%v31,384(%r1) */ in __kernel_fpu_end()
169 " j 7f\n" in __kernel_fpu_end()
170 "5: VLM 0,15,0,1\n" /* vlm %v0,%v15,0(%r1) */ in __kernel_fpu_end()
171 "6: VLM 16,31,256,1\n" /* vlm %v16,%v31,256(%r1) */ in __kernel_fpu_end()
186 asm volatile("lgr 1,%0\n" in __load_fpu_regs()
187 "VLM 0,15,0,1\n" in __load_fpu_regs()
188 "VLM 16,31,256,1\n" in __load_fpu_regs()
204 asm volatile("ld 11,%0" : : "Q" (regs[11])); in __load_fpu_regs()
237 asm volatile("lgr 1,%0\n" in save_fpu_regs()
238 "VSTM 0,15,0,1\n" in save_fpu_regs()
239 "VSTM 16,31,256,1\n" in save_fpu_regs()
255 asm volatile("std 11,%0" : "=Q" (regs[11])); in save_fpu_regs()